Skip to main content
Ctrl+K
NVIDIA NIM for Large Language Models (LLMs) - Home

NVIDIA NIM for Large Language Models (LLMs)

  • Documentation Home
NVIDIA NIM for Large Language Models (LLMs) - Home

NVIDIA NIM for Large Language Models (LLMs)

  • Documentation Home

Table of Contents

About NVIDIA NIM for LLMs

  • Overview
  • Benchmarking
  • Release Notes
  • Models

Get Started

  • Get Started with NIM
  • Tutorials

Deploy NIM

  • Deployment Guide
  • Deploy with Helm
  • Air Gap Deployment
  • Multi-Node Deployment
  • Deploy Behind a Proxy

Work with Models

  • Model Profiles
  • Supported Models
  • Fine-Tuned Models
  • Reward Models
  • Reasoning Models

Use Key Features

  • Custom Guided Decoding Backends
  • Function (Tool) Calling
  • Message Roles
  • Observability
  • Parameter-Efficient Fine-Tuning
  • Structured Generation

Configure Your NIM

  • Configure Your NIM
  • Deterministic Generation Mode
  • KV Cache Reuse
  • Repository Override

Reference

  • API Reference
  • Llama Stack API
  • Utilities
  • Troubleshoot

Resources

  • Acknowledgements
  • EULA
  • Eula

Eula#

By using this NIM, you acknowledge that you have read and agreed to the NVIDIA AI PRODUCT AGREEMENT.

previous

Acknowledgements

NVIDIA NVIDIA
Privacy Policy | Manage My Privacy | Do Not Sell or Share My Data | Terms of Service | Accessibility | Corporate Policies | Product Security | Contact

Copyright © 2024-2025, NVIDIA Corporation.

Last updated on Jul 10, 2025.