Troubleshoot NVIDIA NIM for LLMs#
Use this documentation to troubleshoot issues that arise when you work with NVIDIA NIM for large language models (LLMs).
Guided decoding fails to recognize whitespace pattern#
Using the outlines
backend for regular expressions might cause the regular expressions to fail to compile.
You might see a message such as "guided_whitespace_pattern": "*invalid*pattern*"
.
In this case, use the xgrammar
backend instead, which supports a wider range of regular expressions than outlines
.
For more information, refer to Structured Generation with NVIDIA NIM for LLMs.
vLLM profile fails to deploy#
If the vLLM profile fails to deploy, it is likely because of GPU resource problems. You can try the following troubleshooting options:
Allocate more GPU resources.
Reduce the value of NIM_MAX_MODEL_LEN. Start by setting it to 70,000. You can lower the value if this does not work. For more information, refer to Configure Your NIM with NVIDIA NIM for LLMs.