NVIDIA vGPU for Compute Features#
NVIDIA vGPU (Virtual GPU) for Compute virtualizes NVIDIA GPUs for AI, machine learning, and high-performance computing. The subsections below describe MIG (Multi-Instance GPU) partitioning, provisioning, data paths, migration, multi-GPU guests, interconnects, scheduling, power state, and unified memory.
Feature |
Description |
|---|---|
Hardware-level GPU partitioning with spatial isolation for multi-tenant workloads |
|
Automated topology-aware device provisioning for multi-GPU and GPU-NIC pairs |
|
Direct memory access and storage I/O bypass between GPUs and network/storage devices |
|
Mixed vGPU profiles with different framebuffer sizes on a single GPU |
|
Zero-downtime VM migration between physical hosts |
|
Multiple vGPUs per VM with peer-to-peer NVLink communication |
|
High-bandwidth GPU-to-GPU interconnect fabric for HGX systems |
|
Efficient one-to-many data distribution across NVLink-connected GPUs |
|
Workload-specific GPU scheduling algorithms (Best Effort, Equal Share, Fixed Share) |
|
VM state preservation and resumption without losing GPU context |
|
Single memory address space across CPU and GPU for simplified CUDA programming |