Networking#

The NVIDIA DGX GB200 system leverages a sophisticated hybrid networking approach to enable seamless, high-performance communication both within the rack and across multiple racks. This model combines the unparalleled speed and coherence of NVLink with the traditional, highly scalable InfiniBand and Ethernet technologies.

  • NVLink forms the ultra-fast, memory-coherent scale-up fabric within each rack, making 72 GPUs act as one.

  • InfiniBand provides the high-bandwidth, low-latency scale-out compute fabric between racks, enabling massive multi-node AI training clusters.

  • Ethernet handles storage, management (in-band and out-of-band), and external connectivity, ensuring all components are integrated and manageable.

Network Overview#

The DGX GB200 rack installation uses several discrete networks:

  1. externalnet - This network provides external communication for the headnode to the enterprise network or to the Internet if allowed.

  2. internalnet - This network provides provisioning capabilities as well as system management to the control plane nodes.

  3. dgxnet - This network is used to provision and manage the compute trays.

  4. ipminet - This network provides Base Command Manager and the headnode access to the out-of-band interfaces of all the components in the rack, including the compute trays, NVLink switch trays, the power shelves, the top-of-rack switches, and the control plane nodes.

  5. computenet - This network is used for node communication across racks (east/west traffic).

  6. storagenet - This network provides each node access to storage.

  7. failovernet - This network is only used by the headnodes when configured in redundant failover configuration.

The following images identify the network ports used to communicate to these networks on the compute and switch trays. Note that the number corresponds to the network they are connected to in reference to the numbered list above.

Image showing port locations for each of the networks

DGX GB200 Compute and Switch Trays - Identification of network port locations#

The following image describes the port types available in the compute tray, and how DGX OS identifies each of the ports. Note that the port names in black are the default configurations. Since the BlueField-3 card ports can be switched to InfiniBand mode, those network port names are called out in light gray.

Image showing network port type for each port

DGX GB200 Compute and Switch Trays - Identification of network port types#

Compute Tray Network Ports#

The following table describes the function and type of each of the compute tray network ports.

Compute Tray Network Ports#

Port

Port Type

Network / switch

ConnectX-7 port

OSFP

computenet (InfiniBand)

BlueField-3 left port

QSFP

storagenet (Ethernet)

BlueField-3 right port

QSFP

dgxnet (Ethernet)

Bluefield-3 RJ-45 port

RJ45

ipminet (Ethernet)

1 GbEthernet port LAN

RJ45

Not Connected

1 GbEthernet Port BMC

RJ45

ipminet (Ethernet)

Network Interface Names#

The following table describes the network interface names for each of the ports on the compute tray.

Network Interface Names#

Port

PCIe Bus

Interface name

RDMA

Left Bay OSFP P3

00:03:00.0

ibp3s0

mlx5_0

Left Bay OSFP P4

02:03:00.0

ibp2p3s0

mlx5_1

Right Bay OSFP P3

10:03:00.0

ibp16p3s0

mlx5_4

Right Bay OSFP P4

12:03:00.0

ibp11p3s0

mlx5_5

Left Bay BF3 P1

06:03:00.0

enP6p3s0f0np0

mlx5_2

Left Bay BF3 P2

06:03:00.1

enP6p3s0f1np1

mlx5_3

Right Bay BF3 P1

16:03:00.0

enP22p3s0f0np0

mlx5_6

Right Bay BF3 P2

16:03:00.1

enP22p3s0f1np1

mlx5_7

LAN

enP5p9s0