Skip to main content
Ctrl+K
NVIDIA NIM for Large Language Models (LLMs) - Home

NVIDIA NIM for Large Language Models (LLMs)

  • Documentation Home
NVIDIA NIM for Large Language Models (LLMs) - Home

NVIDIA NIM for Large Language Models (LLMs)

  • Documentation Home

Table of Contents

NVIDIA NIM for LLMs

  • Introduction
  • Release Notes
  • Getting Started
  • Deployment Guide
  • Air Gap Deployment
  • Multi-Node Deployment
  • Deploying with Helm
  • Tutorials
  • Configuring a NIM
  • Model Profiles
  • Overview
  • Benchmarking
  • Models
  • Supported Models
  • Examples with system role
  • API Reference
  • Function Calling
  • Using Reward Models
  • Using Reasoning Models
  • Llama Stack API (Experimental)
  • Utilities
  • Fine-tuned model support
  • Observability
  • Structured Generation
  • Custom Guided Decoding Backend (Experimental)
  • Parameter-Efficient Fine-Tuning
  • KV Cache Reuse (a.k.a. prefix caching)
  • Acknowledgements
  • Eula
  • Eula

Eula#

By using this NIM, you acknowledge that you have read and agreed to the NVIDIA AI PRODUCT AGREEMENT.

previous

Acknowledgements

NVIDIA NVIDIA
Privacy Policy | Manage My Privacy | Do Not Sell or Share My Data | Terms of Service | Accessibility | Corporate Policies | Product Security | Contact

Copyright © 2024-2025, NVIDIA Corporation.

Last updated on May 09, 2025.