NVIDIA Enterprise Reference Architectures

Build AI factories that scale. Turn your data center into a high-performance AI factory with NVIDIA Enterprise Reference Architectures.

Parter Endorsed Designs Based on NVIDIA Enterprise Reference Architectures

OEMSolution (linked if applicable)ServerGPUEnterprise RA PatternMin Size - Max SizeInfrastructure EndorsementSpectrum-X EndorsementNetworking Logical Architecture Endorsement
CiscoCisco Nexus Hyperfabric AI Enterprise Reference ArchitectureUCS C885A Rack ServerHGX H2002-8-9-400 (adapted to 2-8-10-400)4 - 16YESYESYES
CiscoCisco Enterprise Reference Architecture (ERA) is based on Cisco NexusUCS C885A Rack ServerHGX H2002-8-9-400 (adapted to 2-8-10-400)4 - 16YESYESYES
CiscoCisco AI POD Infrastructure for the NVIDIA 2-8-9-400 Enterprise Reference ArchitectureUCS C885A M8HGX H2002-8-9-4004 - 16YESYES
Dell TechnologiesDell AI Factory with NVIDIAPowerEdge XE7745RTX PRO 6000 Blackwell Server Edition, H200 NVL2-8-5-2004, 16YESYESYES
Dell TechnologiesDell AI Factory with NVIDIAPowerEdge XE7740RTX PRO 6000 Blackwell Server Edition, H200 NVL2-8-5-2004, 16YESYESYES
Dell TechnologiesDell AI Factory with NVIDIAPowerEdge XE9680HGX H2002-8-9-4004, 12YESYESYES
Dell FederalDell AI Factory with NVIDIA for GovernmentPowerEdge XE7740RTX PRO 6000 Blackwell Server Edition2-8-5-2004 - 32YESYES
Dell FederalDell AI Factory with NVIDIA for GovernmentPowerEdge XE7745RTX PRO 6000 Blackwell Server Edition2-8-5-2004 - 32YESYES
HPEHPE AI Factory With NVIDIA Enterprise Reference ArchitecturesProLiant DL380a Gen12RTX PRO 6000 Blackwell Server Edition2-8-5-20016YES
LenovoLenovo Hybrid AI 289ThinkSystem SR680a V3HGX B200, HGX H2002-8-9-4004 - 32YESYESYES
SMCAI Factory SolutionsSYS-A22GA-NBRT-G1HGX B2002-8-9-4004 - 32YESYES
SMCAI Factory SolutionsSYS-522GA-NRT, SYS-422GL-NR, AS-5126GS-TNRT2RTX PRO 6000 Blackwell Server Edition2-8-5-2004 - 32YESYES
SMCAI Factory SolutionsSYS-822GS-NB3RTHGX B300 Single Plane2-8-9-8004, 8, 32YESYES

Minimum Criteria for Endorsement by NVIDIA's Design Review Board

- Nodes have passed NVIDIA-Certified for Reference Configuration

- Network design and topology aligned to Enterprise RA design points

- BOM for server vetted by NVIDIA

- BOM for cluster at max scale configuration vetted by NVIDIA

This whitepaper introduces NVIDIA Enterprise Reference Architectures (Enterprise RAs), which provide recommendations for building AI Factories for enterprise-class deployments, ranging from 32 to 256 GPUs. These architectures aim to simplify the deployment of AI infrastructure, reduce complexity, and accelerate time to value.
The NVIDIA RTX PRO AI Factory supports a range of enterprise workloads, including agentic AI inference, physical and industrial AI, visual computing, and high-performance computing for data analytics and simulation. This document outlines the hardware components that define this scalable and modular architecture. This includes guidance regarding the SU design and specifics of Ethernet fabric topologies.
The NVIDIA HGX AI Factory supports a range of enterprise workloads, including AI inference, AI training and fine‑tuning, and large‑scale GPU‑accelerated data analytics. It outlines the hardware components that define a scalable, modular architecture, including SU‑based design guidance and the specifics of the underlying network fabric topologies used to interconnect the cluster.
The NVIDIA NVL72 AI Factory supports the most intensive enterprise AI workloads, including large‑scale foundation model training, fine‑tuning, real‑time reasoning, and complex agentic AI pipelines. This document outlines the GB300 NVL72‑based rack‑scale building block, including guidance on node composition, NVLink domain design, and the network interfaces required to scale this architecture across the data center.
Presents the necessary components, including integrations from our ecosystem partners, automation tools, and deployment strategies. This design can be used by our enterprise partners for integrating accelerated computing, high-performance networking, and AI software for successfully building single tenant enterprise ready AI factories.
Provides an example infrastructure stack build that is geared towards OEMs and NVIDIA partners who intend to build systems that are ready for single-tenant production-grade AI workloads. While hardware components of the infrastructure stack can be modular, the software components of the infrastructure stack are consistent for various workloads, e.g. Inference, Finetuning, & Retrieval Augmented Generation.
This paper helps guide enterprises on how to pack more Inference models on a given set of NVIDIA GPUs using NVIDIA Run:ai, through intelligent scheduling, fractional GPUs, and dynamic resource management. We also explore the impact on performance with the Run:ai scheduler on utilizing fractional GPUs for NIM LLMs.
In this paper, we look at the NVIDIA AI-Q Research Agent blueprint, an agentic system that can generate detailed reports based on both internal and external data. We walk through how to deploy, how to scale and provide sizing guidance.
Offers a standardized and production-ready reference for implementing observability in enterprise AI or HPC environments. Built on top of NVIDIA’s AI infrastructure and Kubernetes-native platforms, this version of the guide specifically focuses on establishing advanced custom dashboard solutions for AI factories, providing administrators and enterprise customers with actionable insights into GPU, CPU, Kubernetes and applications.

Coming Soon.