NVIDIA NIM for Large Language Models Documentation# About NVIDIA NIM for LLMs Overview Enterprise-Grade Inference Software Stack Release Notes Get Started About Get Started Prerequisites Configuration Installation Quickstart Deployment Model Profiles and Selection Model Download Model-Free NIM Kubernetes Deployment Cloud Service Provider (CSP) Deployment Air-Gap Deployment Multi-Node Deployment vGPU Deployment Advanced Use Cases Fine-Tuning with LoRA Tool Calling and MCP Integration Custom Logits Processing Prompt Embeddings Reference Architecture Environment Variables API Reference CLI Reference Advanced Configuration Logging and Observability Model Signature Verification 1.x Migration Guide Support Matrix Archived Versions Troubleshooting GPU Memory (OOM) Errors CUDA Driver Initialization Resources Support and FAQ Related Products Legal