Infrastructure-as-a-Service
Infrastructure-as-a-Service
The Infrastructure-as-a-Service (IaaS) layer provides scalable bare metal and virtual machines (both called instances) to users, allowing tenants to run their applications without having to manage the underlying physical infrastructure. The IaaS layer should provide a range of offerings that can be customized by the NCP, configuring dimensions like general-purpose compute size, memory, storage, and GPU capacity.
IaaS Architecture
The following diagram defines the major components of the IaaS system that are considered part of the Data Center Management (DCM) functionality. This includes the Cloud Control Plane (CCP), Software Defined Networking (SDN), Compute Service, and Software Defined Storage (SDS).

These components work together to build a CSP-like infrastructure-as-a-service system.
Component Descriptions
Infrastructure as a Service Components
Each IaaS component requires specific capabilities to operate an AI-optimized, multi-tenant data center. Building a complete IaaS layer requires integrating software across all these components. NVIDIA provides a range of infrastructure software for network management, compute management, and storage acceleration. NCPs and ISVs integrate the software components with their choice of operating system, hypervisor, and storage vendor.
Network Management
To recap from the Data Center View, an AI data center operates multiple network fabrics:
Data Centre Networks
Network management encompasses provisioning, configuring, and isolating these fabrics to support multi-tenant AI workloads.
Capabilities Required
- Ethernet Fabric Provisioning: Configure leaf-spine fabric for TAN and Ethernet-based CIN
- Ethernet visibility & telemetry: Monitor fabric health, validate changes, troubleshoot issues
- InfiniBand Fabric Management: Configure IB fabric for high-performance CIN
- NVLink Domain Management: Configure rack-scale NVLink (GB200/GB300)
- Tenant Network Isolation: Create isolated VPCs via overlays (VXLAN) or partition keys (PKeys)
Software Defined Networking (SDN)
With Software Defined Networking, the network control plane is separated from the data plane to enable centralized and programmable management for the network behavior. Thus, rather than having a system administrator manually go to switches and routers to update behavior, it can all be updated programmatically by the SDN controller.
The picture below shows the SDN system built on top of NVIDIA technologies. While this is not the only way to construct the SDN, it is shown this way to explain the relevant technologies.

The DCM SDN Layer contains the Network Manager and SDN Controller — these are typically ISV or NCP-built components that handle the control plane logic.
On the right side, the Compute Node runs an SDN Agent, with the BlueField DPU running DOCA/HBN to terminate tenant overlays at the edge.
At the bottom, NVIDIA software manages each network fabric: NMX manages NVSwitch-based NVLink domains, UFM manages InfiniBand switches running MLNX-OS, and NetQ provides visibility for Spectrum switches running Cumulus.
Compute Management
The Compute Service is responsible for managing bare metal (BM) and virtual machine (VM) instances and their full lifecycle. This applies to all the compute in the data center, including GPU nodes, general-purpose nodes, and potentially software-defined storage nodes. The Compute Service deals with:
- Instance Lifecycle (create, start/stop, terminate, reboot, placement)
- Instance Management (for example, Get VM Status, Attach Volume)
- Monitor Instance (get logs, get metrics)
- Instance Networking (Assign IP to Instance, attach virtual NIC to instance)
- Sanitize compute resources between tenants
- Tenant Isolation
Capabilities Required
Compute Management Capability Matrix
The following describes logical components that make up the Compute Service. A specific implementation may partition differently, but the capabilities should be present.
- Machine Lifecycle Manager — Discovery, ingestion, placement, provisioning, sanitization, firmware updates, BMC management, IP allocation, OS installs
- Instance Database — Maintains state for all physical machines and instances
- VM Control Plane — Receives provisioned BM nodes, handles VM requests, bin-packing, event monitoring. This applies at the node level; for rack-level and larger deployments (e.g. scalable unit (SU) delineations), the same behavior and placement decisions should be considered to achieve optimal network bandwidth and latency. NVIDIA recommends KubeVirt on Kubernetes.
- PXE Boot Server — The PXE (Pre-boot eXecution Environment) boot server responds to a request from a booting node and loads a minimal OS and a provisioning agent. The provisioning agent can be used to test the node, update all the firmware on the platform, and install a Bare Metal image (tenant OS and cloud-init configuration) or Virtual Machine image (base OS/hypervisor + core components). This may logically be part of the Lifecycle Manager.
Storage Management
Bare metal servers and virtual machines need access to storage. There are two classes of storage — ephemeral and persistent. Ephemeral is provided by access to (typically local) drives that are erased between tenants. Persistent provides access to a mix of block, file, and object storage across the network. For the latter, the SDS is responsible for managing the provisioning, pooling, and scaling of these three types of storage. This is accessed through the Cloud Control Plane.

The SDS system has two primary components:
SDS Controller
This is the tenant-facing control plane:
- Provides the northbound APIs (for example, CreateVolume, DeleteVolume, SetPolicy)
- Provides the Policy Engine to enforce quotas, QoS, and security
- Namespace Manager to track the logical view of storage
- Southbound driver to translate commands to the correct storage target mechanics
Cluster Orchestrator
This is the infrastructure-facing control plane:
- Node Manager manages the physical node lifecycle
- Placement Engine applies physical constraints to placement (for example, replicas across power domains)
- Detect failures and manage recovery, manage upgrades, and more.
Beyond SDS controller and cluster orchestrator, there are the actual storage targets (like VAST, WEKA, Lustre, NFS) and potential data plane services (such as snapshots, data imports).
Capabilities Required
Storage Management Capability Matrix
NVIDIA Software for Infrastructure-as-a-Service
NVIDIA provides software to address specific IaaS capabilities described above. The use of NVIDIA-provided software is optional and depends on architectural decisions made by the NCP. NCPs can work with an ecosystem of ISV partners to integrate specific software components provided.
NVIDIA Software Components for IaaS
For detailed descriptions of each component, see Part 2: NVIDIA Software for Infrastructure as a Service.
Footnotes
-
NVIDIA Infra Controller provides integrated multi-tenant infrastructure management, orchestrating both bare metal lifecycle and tenant network isolation across all three network domains (Ethernet via HBN, InfiniBand via UFM, NVLink via NMX). ↩