Skip to main content
Ctrl+K
NVIDIA NIM for Large Language Models (LLMs) - Home

NVIDIA NIM for Large Language Models (LLMs)

  • Documentation Home
NVIDIA NIM for Large Language Models (LLMs) - Home

NVIDIA NIM for Large Language Models (LLMs)

  • Documentation Home

Table of Contents

About NVIDIA NIM for LLMs

  • Overview
  • Benchmarking
  • Release Notes
  • Models

Get Started

  • Get Started with NIM
  • Tutorials

Deploy NIM

  • Deployment Guide
  • Deploy with Helm
  • Air Gap Deployment
  • Multi-Node Deployment
  • Deploy Behind a Proxy

Work with Models

  • Model Profiles
  • Supported Models
  • Fine-Tuned Models
  • Reward Models
  • Reasoning Models

Use Key Features

  • Custom Guided Decoding Backends
  • Function (Tool) Calling
  • Message Roles
  • Observability
  • Parameter-Efficient Fine-Tuning
  • Structured Generation

Configure Your NIM

  • Configure Your NIM
  • Deterministic Generation Mode
  • KV Cache Reuse
  • Repository Override

Reference

  • API Reference
  • Llama Stack API
  • Utilities
  • Troubleshoot

Resources

  • Acknowledgements
  • EULA

Index

Symbols | C

Symbols

  • --all
    • command line option
  • --lora
    • command line option
  • --model-store
    • command line option
  • --profile
    • command line option
  • --profiles
    • command line option
  • -m
    • command line option
  • -p
    • command line option, [1]

C

  • command line option
    • --all
    • --lora
    • --model-store
    • --profile
    • --profiles
    • -m
    • -p, [1]
NVIDIA NVIDIA
Privacy Policy | Manage My Privacy | Do Not Sell or Share My Data | Terms of Service | Accessibility | Corporate Policies | Product Security | Contact

Copyright © 2024-2025, NVIDIA Corporation.

Last updated on Jul 10, 2025.