NVIDIA NIM for Large Language Models Documentation# About NVIDIA NIM for LLMs Overview Enterprise-Grade Inference Software Stack Release Notes Get Started About Get Started Prerequisites Installation Configuration Quickstart Deployment Model Profiles and Selection Model Download Model-Free NIM Kubernetes Deployment Cloud Service Provider (CSP) Deployment Air-Gap Deployment Multi-Node Deployment vGPU Deployment Advanced Use Cases Fine-Tuning with LoRA Custom Logits Processing Prompt Embeddings Reference Architecture Environment Variables API Reference CLI Reference Advanced Configuration Logging and Observability 1.x Migration Guide Support Matrix Resources Support and FAQ Related Products Legal