Skip to main content
Ctrl+K
NVIDIA NIM for Vision Language Models (VLMs) - Home NVIDIA NIM for Vision Language Models (VLMs) - Home

NVIDIA NIM for Vision Language Models (VLMs)

NVIDIA NIM for Vision Language Models (VLMs) - Home NVIDIA NIM for Vision Language Models (VLMs) - Home

NVIDIA NIM for Vision Language Models (VLMs)

Table of Contents

User Guide

  • Overview
  • Release Notes
  • Getting Started
  • Sampling Control
  • Configuring a NIM
  • Support Matrix
  • Performance
  • API Reference
  • Observability
  • Utilities
  • Model Profiles
  • KV Cache Reuse (a.k.a. prefix caching)
  • Structured Generation
  • Deploying with Helm

Examples

  • Llama 3.2 Vision
    • Llama 3.2 Vision Overview
    • Llama 3.2 Vision API
  • nemoretriever-parse
    • nemoretriever-parse Overview
    • nemoretriever-parse API

Notices

  • Acknowledgements
  • EULA

NVIDIA NIM for Vision Language Models (VLMs)#

User Guide

  • Overview
  • Release Notes
  • Getting Started
  • Sampling Control
  • Configuring a NIM
  • Support Matrix
  • Performance
  • API Reference
  • Observability
  • Utilities
  • Model Profiles
  • KV Cache Reuse (a.k.a. prefix caching)
  • Structured Generation
  • Deploying with Helm

Examples

  • Llama 3.2 Vision
  • nemoretriever-parse

Notices

  • Acknowledgements
  • EULA

next

Introduction

NVIDIA NVIDIA
Privacy Policy | Manage My Privacy | Do Not Sell or Share My Data | Terms of Service | Accessibility | Corporate Policies | Product Security | Contact

Copyright © 2024-2025, NVIDIA Corporation.

Last updated on May 09, 2025.