Related Software#
NVIDIA NIM for Large Language Models fits into a broader inference and platform ecosystem. The following software products are highly relevant when you are deploying, operating, or extending LLM workloads.
Product |
Relation to NIM LLM |
|---|---|
NVIDIA NIM for Large Language Models packages vLLM as its inference backend, so many request semantics and tuning concepts come directly from vLLM. |
|
The operator manages NIM deployments by using Kubernetes custom resources and is especially useful for repeatable, production-scale rollouts. |
|
NIM for VLMs follows the same NIM operational model, but targets vision-language workloads rather than text-only LLM inference. |
Usage Guidance#
To help you choose the right tool for your specific use case, consider the following recommendations:
Use NIM LLM when your workload is text-only and you want a curated, enterprise-ready container for production inference.
Use vLLM documentation alongside NIM documentation when you need deeper, backend-specific context for passthrough arguments or upstream model-serving behavior.
Use the NIM Operator when your primary deployment target is Kubernetes and you want lifecycle automation around NIM services.
Use NVIDIA NIM for Vision Language Models when your application must process images and text together instead of text alone.