Infrastructure-as-a-Service

View as Markdown

The Infrastructure-as-a-Service (IaaS) layer provides scalable bare metal and virtual machines (both called instances) to users, allowing tenants to run their applications without having to manage the underlying physical infrastructure. The IaaS layer should provide a range of offerings that can be customized by the NCP, configuring dimensions like general-purpose compute size, memory, storage, and GPU capacity.

IaaS Architecture

The following diagram defines the major components of the IaaS system that are considered part of the Data Center Management (DCM) functionality. This includes the Cloud Control Plane (CCP), Software Defined Networking (SDN), Compute Service, and Software Defined Storage (SDS).

Infrastructure as a Service Components

These components work together to build a CSP-like infrastructure-as-a-service system.

Component Descriptions

Infrastructure as a Service Components

ComponentDescription
Cloud Control Plane (CCP)Public-facing control plane providing API/UI to provision and manage compute, networking, and storage. Handles authentication, authorization, quota enforcement, and metering.
Software Defined Network (SDN)Enables centralized, programmable network management for automation, scalability, and multi-tenant isolation.
Compute ServiceManages bare metal and VM instance lifecycle, including provisioning, placement, monitoring, and sanitization between tenants.
Software Defined Storage (SDS)Manages provisioning, pooling, and scaling of block, file, and object storage.
Identity and Access Management (IAM)Authentication and authorization for end-users.

Each IaaS component requires specific capabilities to operate an AI-optimized, multi-tenant data center. Building a complete IaaS layer requires integrating software across all these components. NVIDIA provides a range of infrastructure software for network management, compute management, and storage acceleration. NCPs and ISVs integrate the software components with their choice of operating system, hypervisor, and storage vendor.

Network Management

To recap from the Data Center View, an AI data center operates multiple network fabrics:

Data Centre Networks

NetworkPurposeTechnology
Tenant Access Network (TAN)North-South traffic: Tenant access to storage, external servicesEthernet (Spectrum)
Cluster Interconnect Network (CIN)East-West traffic: GPU-to-GPU communication across nodesInfiniBand (Quantum) or Ethernet (SpectrumX)
NVLink NetworkScale-up: GPU-to-GPU communication within a rackNVLink (GB200/GB300)
Secure Management Network (SMN)Out-of-band (OOB) infrastructure managementEthernet (1GbE)
Shared ServicesTenant access to shared services (registry, IAM)Logical overlay on TAN

Network management encompasses provisioning, configuring, and isolating these fabrics to support multi-tenant AI workloads.

Capabilities Required

  • Ethernet Fabric Provisioning: Configure leaf-spine fabric for TAN and Ethernet-based CIN
  • Ethernet visibility & telemetry: Monitor fabric health, validate changes, troubleshoot issues
  • InfiniBand Fabric Management: Configure IB fabric for high-performance CIN
  • NVLink Domain Management: Configure rack-scale NVLink (GB200/GB300)
  • Tenant Network Isolation: Create isolated VPCs via overlays (VXLAN) or partition keys (PKeys)

Software Defined Networking (SDN)

With Software Defined Networking, the network control plane is separated from the data plane to enable centralized and programmable management for the network behavior. Thus, rather than having a system administrator manually go to switches and routers to update behavior, it can all be updated programmatically by the SDN controller.

The picture below shows the SDN system built on top of NVIDIA technologies. While this is not the only way to construct the SDN, it is shown this way to explain the relevant technologies.

Software Defined Network built on NVIDIA software components

The DCM SDN Layer contains the Network Manager and SDN Controller — these are typically ISV or NCP-built components that handle the control plane logic.

On the right side, the Compute Node runs an SDN Agent, with the BlueField DPU running DOCA/HBN to terminate tenant overlays at the edge.

At the bottom, NVIDIA software manages each network fabric: NMX manages NVSwitch-based NVLink domains, UFM manages InfiniBand switches running MLNX-OS, and NetQ provides visibility for Spectrum switches running Cumulus.

Compute Management

The Compute Service is responsible for managing bare metal (BM) and virtual machine (VM) instances and their full lifecycle. This applies to all the compute in the data center, including GPU nodes, general-purpose nodes, and potentially software-defined storage nodes. The Compute Service deals with:

  • Instance Lifecycle (create, start/stop, terminate, reboot, placement)
  • Instance Management (for example, Get VM Status, Attach Volume)
  • Monitor Instance (get logs, get metrics)
  • Instance Networking (Assign IP to Instance, attach virtual NIC to instance)
  • Sanitize compute resources between tenants
  • Tenant Isolation

Capabilities Required

Compute Management Capability Matrix

CapabilityPurpose
Machine DiscoveryDiscover and ingest physical machines into inventory
Bare Metal ProvisioningPXE boot, firmware updates, OS installation at scale
Instance LifecycleCreate, start/stop, terminate, reboot, placement
Instance ManagementGet status, attach volumes, assign IPs, attach vNICs
Hardware AttestationVerify node and firmware integrity prior to allocation
Node Health CheckStress test to verify performance before allocation
Node SanitizationSecurely wipe nodes between tenants
GPU VirtualizationShare GPUs across VMs (time-sliced or MIG)
GPU ObservabilityMonitor GPU health, metrics, diagnostics
Rack-Scale LifecyclePower sequencing, cooling, coordinated firmware

The following describes logical components that make up the Compute Service. A specific implementation may partition differently, but the capabilities should be present.

  • Machine Lifecycle Manager — Discovery, ingestion, placement, provisioning, sanitization, firmware updates, BMC management, IP allocation, OS installs
  • Instance Database — Maintains state for all physical machines and instances
  • VM Control Plane — Receives provisioned BM nodes, handles VM requests, bin-packing, event monitoring. This applies at the node level; for rack-level and larger deployments (e.g. scalable unit (SU) delineations), the same behavior and placement decisions should be considered to achieve optimal network bandwidth and latency. NVIDIA recommends KubeVirt on Kubernetes.
  • PXE Boot Server — The PXE (Pre-boot eXecution Environment) boot server responds to a request from a booting node and loads a minimal OS and a provisioning agent. The provisioning agent can be used to test the node, update all the firmware on the platform, and install a Bare Metal image (tenant OS and cloud-init configuration) or Virtual Machine image (base OS/hypervisor + core components). This may logically be part of the Lifecycle Manager.

Storage Management

Bare metal servers and virtual machines need access to storage. There are two classes of storage — ephemeral and persistent. Ephemeral is provided by access to (typically local) drives that are erased between tenants. Persistent provides access to a mix of block, file, and object storage across the network. For the latter, the SDS is responsible for managing the provisioning, pooling, and scaling of these three types of storage. This is accessed through the Cloud Control Plane.

Storage Systems in the NVIDIA Software Reference Architecture

The SDS system has two primary components:

SDS Controller

This is the tenant-facing control plane:

  • Provides the northbound APIs (for example, CreateVolume, DeleteVolume, SetPolicy)
  • Provides the Policy Engine to enforce quotas, QoS, and security
  • Namespace Manager to track the logical view of storage
  • Southbound driver to translate commands to the correct storage target mechanics

Cluster Orchestrator

This is the infrastructure-facing control plane:

  • Node Manager manages the physical node lifecycle
  • Placement Engine applies physical constraints to placement (for example, replicas across power domains)
  • Detect failures and manage recovery, manage upgrades, and more.

Beyond SDS controller and cluster orchestrator, there are the actual storage targets (like VAST, WEKA, Lustre, NFS) and potential data plane services (such as snapshots, data imports).

Capabilities Required

Storage Management Capability Matrix

CapabilityPurpose
Block StorageAttach persistent volumes to instances
File StorageShared filesystems for datasets and checkpoints
Object StorageS3-compatible storage for data lakes and artifacts
Storage Tenant IsolationVolume access control, namespace separation, per-tenant encryption
Storage QoSEnforce bandwidth/IOPS limits per tenant to prevent noisy neighbors
High-Performance Data PathDirect GPU-to-storage data movement, bypassing CPU
Storage Network IntegrationConnect to NVMe-oF, RDMA-capable storage fabrics

NVIDIA Software for Infrastructure-as-a-Service

NVIDIA provides software to address specific IaaS capabilities described above. The use of NVIDIA-provided software is optional and depends on architectural decisions made by the NCP. NCPs can work with an ecosystem of ISV partners to integrate specific software components provided.

NVIDIA Software Components for IaaS

Functional AreaNVIDIA Software Implementation
Network ManagementCumulus, NetQ, UFM, NMX, DOCA/HBN, NVIDIA Infra Controller1
Compute ManagementNVIDIA Infra Controller
Storage ConnectivityGPUDirect Storage, GPUDirect RDMA
OperationsDCGM

For detailed descriptions of each component, see Part 2: NVIDIA Software for Infrastructure as a Service.

Footnotes

  1. NVIDIA Infra Controller provides integrated multi-tenant infrastructure management, orchestrating both bare metal lifecycle and tenant network isolation across all three network domains (Ethernet via HBN, InfiniBand via UFM, NVLink via NMX).