Initial Cluster Setup

This document details how to deploy NVIDIA Base Command™ Manager (BCM) on NVIDIA DGX BasePOD™ configurations.

Physical installation and network switch configuration must be completed before deploying BCM. In addition, information about the intended deployment should be recorded in a site survey.

The deployment stage of a DGX BasePOD consists of using BCM to provision and manage the Kubernetes cluster.

  1. Configure the DGX systems to PXE boot by default.

    1. Using either KVM or a crash cart, connect to the DGX system, enter the BIOS menu, and configure Boot Option #1 to be [NETWORK].

      _images/cluster-setup-01.png
    2. Ensure that other Boot Options are [Disabled] and navigate to the UEFI NETWORK Drives BBS Priorities menu.

    3. Set Boot Option #1 and Boot Option #2 to use IPv4 for Storage 4-2 and Storage 5-2.

      _images/cluster-setup-02.png
    4. Ensure that other Boot Options are [Disabled].

    5. Select Save & Exit.

  2. On the failover head node and the management nodes, ensure that Network boot is configured as the primary option. Ensure that the Mellanox ports connected to the network on the head and CPU nodes are set to Ethernet mode as well.

    This is an example of a system that will boot from the network with Slot 1 Port 2 and Slot 2 Port 2.

    _images/cluster-setup-49.png
  3. Download a BCM ISO from the Bright Download site. Select Base Command Manager 10, RHEL 9.2 and check the Include NVIDIA CUDA Packages and MOFED packages checkbox.

    _images/cluster-setup-26.png
  4. Burn the ISO to a DVD or a bootable USB device.

    It can also be mounted as virtual media and installed using the BMC. The specific mechanism for the latter will vary by vendor.

  5. Ensure that the BIOS of the target head node is configured in UEFI mode and that its boot order is configured to boot the media containing the BCM installer image.

  6. Boot the installation media.

  7. At the grub menu, choose Start Base Command Manager Graphical Installer.

    _images/cluster-setup-04.png
  8. Select Start installation on the splash screen.

    _images/cluster-setup-05.png
  9. Accept the terms of the NVIDIA EULA by checking I agree and then select Next.

    _images/cluster-setup-27.png
  10. Accept the terms of the Ubuntu Server UELA by checking I agree and then select Next.

    _images/cluster-setup-28.png
  11. Unless instructed otherwise, select Next without modifying the kernel modules to be loaded at boot time.

    _images/cluster-setup-29.png
  12. Verify the Hardware info is correct and then select Next.

    For example, the target storage device and the cabled host network interfaces are present (in this case three NVMe drives are the target storage device, and ens1np0 and ens2np01 are the cabled host network interfaces).

    _images/cluster-setup-30.png
  13. On the Installation source screen, choose the appropriate source and then select Next.

    Running a media integrity check is optional.

    _images/cluster-setup-31.png
  14. On the Cluster settings screen, enter the required information and then select Next.

    _images/cluster-setup-32.png
  15. On the Workload manager screen, choose None and then select Next.

    _images/cluster-setup-33.png
  16. On the Network topology screen, choose the network type for the data center environment and then select Next.

    _images/cluster-setup-34.png
  17. On the Head node screen, enter the Hostname, and Administrator password, choose Other for Hardware manufacturer, and then select Next.

    _images/cluster-setup-35.png
  18. Adjust the node digits to 2 in the Compute Node Settings and then select Next.

    Ensure that the Node base name is node. Other values will be updated later in the installation.

    _images/cluster-setup-36.png
  19. On the BMC Configuration screen, choose Yes for both Head Node and Compute Nodes. Follow the screenshot below for the rest of the configuration. Then select Next.

    _images/cluster-setup-37.png
  20. On the Networks screen, enter the required information for externalnet, and then select Next.

    _images/cluster-setup-38.png
  21. Navigate to the internalnet tab and enter the required information for internalnet, leaving the MTU values at their defaults, and then select Next.

    _images/cluster-setup-39.png
  22. Navigate to the ipminet tab, enter the required information for ipminet, and then select Next.

    _images/cluster-setup-40.png
  23. On the Head node interfaces screen, ensure that one interface is configured with the head node’s target internalnet IP, and then select Next.

    _images/cluster-setup-41.png
  24. On the Compute node interfaces screen, change the offset to 0.0.0.3, and then select Next.

    These will be updated post-installation.

    _images/cluster-setup-42.png
  25. On the Disk layout screen, select the target install location and then select Next.

    _images/cluster-setup-43.png
  26. On the Disk layout settings screen, select One Big Partition for the Headnode, and Default Standard Layout for the compute nodes. Then select Next.

    _images/cluster-setup-44.png
  27. In the Additional software screen, select CUDA and MOFED 23.10, then select Next.

    _images/cluster-setup-46.png
  28. Confirm the information on the Summary screen and then select Next.

    The Summary screen provides an opportunity to confirm the Head node and basic cluster configuration before deployment begins. This configuration will be updated/modified for DGX BasePOD after deployment is complete. If values do not match expectations, use the Back button to navigate to the appropriate screen to correct any mistake.

    _images/cluster-setup-50.png
  29. Once the deployment is complete, select Reboot.

    _images/cluster-setup-48.png