Release Notes#

These release notes describe NVIDIA AI Enterprise Infrastructure Release 7.5. Use them to identify the supported infrastructure software components and versions in this release, review compatibility and support information, and locate the per-component release notes for new features, fixed issues, and known limitations.

Note

NVIDIA AI Enterprise Infra 7.5 is a Long-Term Support Branch (LTSB), providing extended support for stable production deployments. For branch lifecycle, support windows, and cross-branch migration guidance, refer to the NVIDIA AI Enterprise Lifecycle Policy.

For deployment guidance, refer to the Quick Start Guide. For a full list of supported platforms, hypervisors, operating systems, and orchestration software, refer to the Support Matrix or the interactive support matrix linked under Compatibility and Support. If you are upgrading from 7.4, refer to Upgrading from 7.4 to 7.5.

Latest Release Highlights

  • NVIDIA Run:ai SaaS Now Included — In addition to NVIDIA Run:ai self-hosted, the NVIDIA-managed NVIDIA Run:ai SaaS offering is now included in the NVIDIA AI Enterprise license under the same enterprise SLA. Refer to the NVIDIA Run:ai SaaS Documentation for the NVIDIA-managed cloud-service option, or choose the deployment that fits your environment.

  • NVIDIA Run:ai 2.25 — Updated from 2.24 in 7.4. The same 2.25 release applies to both NVIDIA Run:ai self-hosted and NVIDIA Run:ai SaaS. Refer to the NVIDIA Run:ai release notes for scheduling, GPU-utilization, and platform updates in this version.

  • NVIDIA Data Center GPU Driver 580.159.03 — Maintenance update within the R580 production driver branch (from 580.126.09 in 7.4). NVIDIA Fabric Manager also updates to 580.159.03 in lockstep with the driver. Refer to the 580.159.03 release notes for fixes and platform-support details.

  • NVIDIA vGPU Software 19.5 — NVIDIA Virtual GPU Manager and the NVIDIA vGPU for Compute Guest Driver are both updated from 19.4 to 19.5, refreshing the full vGPU stack in a coordinated release.

  • NVIDIA DOCA 3.3.0 — NVIDIA DOCA Driver for Networking is updated from 3.2.0 to 3.3.0 and NVIDIA DOCA Microservices is updated from 3.2.1 to 3.3.0, advancing the full DOCA stack for NVIDIA BlueField DPUs and SuperNICs in a coordinated release.

  • Kubernetes Operator Major Updates — NVIDIA GPU Operator 26.3.1 (from 25.10.1 in 7.4) and NVIDIA Network Operator 26.1.1 (from 25.10.0 in 7.4) are both major-version bumps. NVIDIA NIM Operator 3.1.0 (from 3.0.2) and NVIDIA Container Toolkit 1.19.0 (from 1.18.1) also updated. NVIDIA DPU Operator (DPF) 25.10.1 carries forward unchanged from 7.4.

  • NVIDIA Base Command Manager 11.32.1 — Updated from 11.31.0 in 7.4. Refer to the BCM 11.32.1 release notes for cluster-management and provisioning updates.

What is Included in NVIDIA AI Enterprise Infra 7.5#

Complete list of infrastructure components with versions and documentation links. For platform-support detail (x86 / ARM / Government Ready), refer to the Support Matrix—this release-notes table mirrors the canonical component versions defined there.

Table 1 Supported Infrastructure Software#

Component

Description

Version

NVIDIA Data Center GPU Driver

Provides hardware support for NVIDIA GPUs.

580.159.03

NVIDIA DOCA Driver for Networking

Provides hardware support for NVIDIA BlueField DPUs and SuperNICs. Installing DOCA on the host provides all necessary drivers and tools to manage BlueField and ConnectX devices.

3.3.0

NVIDIA Fabric Manager

Manages NVSwitch fabric to enable high-bandwidth, low-latency GPU-to-GPU communication for multi-GPU AI workloads

580.159.03

NVIDIA DOCA Microservices

Infrastructure acceleration and offload services for NVIDIA BlueField, enabling accelerated networking, storage, and security workloads.

3.3.0

NVIDIA Virtual GPU Manager

GPU driver deployed alongside the hypervisor in virtualized environments. Enables multi-tenant GPU sharing, live migration, and monitoring.

19.5

NVIDIA vGPU for Compute Guest Driver [1]

GPU driver deployed in the guest VM to enable multiple VMs to have simultaneous, direct access to a single physical GPU.

19.5

NVIDIA Container Toolkit

Enables GPU-accelerated containers by providing runtime components and utilities for container engines (Docker, containerd, CRI-O)

1.19.0

NVIDIA Run:ai

Provides a Kubernetes-native orchestration and management platform that maximizes GPU utilization for AI workloads through advanced scheduling and resource management.

2.25

NVIDIA DPU Operator (DPF)

Enables cluster administrators to automate provisioning, orchestration, and lifecycle management of BlueField DPUs and DOCA Microservices to enable DPU-accelerated North-South networking in Kubernetes.

25.10.1

NVIDIA GPU Operator

Simplifies deployment of NVIDIA AI Enterprise by automating management of all NVIDIA software components needed to provision GPUs in Kubernetes.

26.3.1

NVIDIA Network Operator

Simplifies deployment of high-speed networking by automating management of NVIDIA ConnectX NICs and SuperNICs required to optimize East-West traffic and RDMA transfers in Kubernetes.

26.1.1

NVIDIA NIM Operator

Enables cluster administrators to operate the software components and services required to run LLM, embedding, and other models using NVIDIA NIM microservices in Kubernetes.

3.1.0

NVIDIA Base Command Manager (BCM)

Cluster management and provisioning tool for NVIDIA DGX systems.

11.32.1

Upgrading from 7.4 to 7.5#

Use this checklist when upgrading existing 7.4 deployments to 7.5. Run each step in a maintenance window and validate before proceeding.

Table 2 7.4 → 7.5 Upgrade Sequence#

Step

Action

Reference

  1. Review

Read the component version delta in the table above and review per-component release notes for breaking changes, deprecations, and feature additions.

Component release note links above

  1. Validate

Confirm hardware, hypervisor (for virtualized deployments), and operating-system compatibility against the 7.5 support matrix.

Support Matrix

  1. Plan rollback

In virtualized deployments, snapshot or back up VMs and capture current driver, operator, and licensing configuration before upgrading. Document the rollback path for each component.

N/A

  1. Upgrade host components

Depending on your deployment type i.e. bare metal or virtualized, upgrade the NVIDIA Data Center GPU Driver (applies to bare metal deployments), Virtual GPU Manager (only applies to vGPU for Compute deployments), and DOCA Driver. Fabric Manager is included in the NVIDIA AI Enterprise drivers and updates with the GPU Driver package.

vGPU Installation

  1. Upgrade Kubernetes operators

Upgrade NVIDIA GPU Operator, Network Operator, NIM Operator, DPU Operator (DPF), and NVIDIA Run:ai (self-hosted) following each operator’s documented upgrade path. NVIDIA Run:ai SaaS upgrades are managed by NVIDIA and require no customer-side upgrade.

Operator release note links above

  1. Upgrade guest VMs

For virtualized deployments, upgrade NVIDIA vGPU for Compute Guest Driver and NVIDIA Container Toolkit in tenant VMs.

vGPU Installation

  1. Verify licensing

Confirm each licensed vGPU VM reaches the NVIDIA License System and shows Licensed status.

Verifying License Configuration

  1. Validate workloads

Run a representative CUDA or AI/ML workload to confirm performance, feature parity, and operator-managed scheduling behavior.

N/A

For lifecycle policy, branch support windows, and migration windows beyond 7.x, refer to the NVIDIA AI Enterprise Lifecycle Policy.

Compatibility and Support#

Support Matrix

Use the Interactive Support Matrix to compare NVIDIA AI Enterprise infrastructure compatibility across releases 4.4 through 7.5. The web tool lets you:

  • Filter by deployment type, operating system, hypervisor, or orchestration platform.

  • Search by GPU architecture, platform, Kubernetes distribution, or cloud provider.

  • Inspect per-configuration release badges and footnote tooltips.

For deep linking, printing, or offline reference, the same information is also available as the static Support Matrix. Both forms cover supported GPU architectures, operating systems, hypervisor and orchestration platform versions, cloud provider instance types, and networking hardware.

Lifecycle and Compatibility Explorer

Use the Interactive Lifecycle and Compatibility Explorer on the NVIDIA AI Enterprise Lifecycle Policy documentation to:

  • Query by Infrastructure Branch, by release, or by component type and version.

  • Run a full-stack check from a GPU driver version to validate compatibility and plan upgrades.

Footnotes