Quick Start Guide#

This Quick Start Guide will help you rapidly deploy and configure NVIDIA AI Enterprise, enabling AI workloads on bare-metal, public cloud, or virtualized environments. Follow these steps to seamlessly set up and optimize your infrastructure for accelerated AI and deep learning tasks.

If you need complete instructions for installing and configuring NVIDIA AI Enterprise, are using NVIDIA AI Enterprise in an NVIDIA vGPU deployment, or are using multiple nodes, refer to the NVIDIA AI Enterprise Deployment Guide.

Note

These instructions do not apply to NVIDIA DGX systems. For information about how to use these systems, refer to NVIDIA DGX Systems.

Attention

If you have previously purchased NVIDIA AI Enterprise, have an NVIDIA Enterprise Account, and an NGC account, you can directly go to Installing NVIDIA AI Enterprise Software Components.

Activating the Accounts for Getting NVIDIA AI Enterprise#

After your order for NVIDIA AI Enterprise is processed, you will receive an order confirmation message. This message contains information that you need for getting NVIDIA AI Enterprise and technical support from NVIDIA. To get NVIDIA AI Enterprise and technical support from NVIDIA, you must have an NVIDIA Enterprise Account, which provides login access to the following NVIDIA software components:

  • NVIDIA NGC - provides access to all enterprise software, services, and management tools included in NVIDIA AI Enterprise

  • NVIDIA Enterprise Support Portal - provides access to to support services for NVIDIA AI Enterprise

  • NVIDIA Licensing Portal - provides access to your entitlements and options for managing your NVIDIA AI Enterprise license servers. Note that installing and managing NVIDIA AI Enterprise license servers are a prerequisite for deploying vGPU software.

These components can be accessed from the NVIDIA Application Hub. To activate your account and access NVIDIA AI Enterprise, follow these steps:

Before You Begin#

Before you attempt the procedures in this guide, ensure that the following prerequisites are met:

For information about supported hardware and software, and any known issues for this release of NVIDIA AI Enterprise, refer to the NVIDIA AI Enterprise Release Notes.

Your Order Confirmation Message#

After your order for NVIDIA AI Enterprise is processed, you will receive an order confirmation message to which your NVIDIA Entitlement Certificate is attached. Your NVIDIA Entitlement Certificate contains your product activation keys and provides instructions for using the certificate.

If you are a data center administrator, follow the instructions in the NVIDIA Entitlement Certificate to use the certificate. Otherwise, forward your order confirmation message, including the attached NVIDIA Entitlement Certificate, to a data center administrator in your organization.

NVIDIA Enterprise Account Requirements#

You must have a suitable NVIDIA Enterprise Account to get access to NVIDIA AI Enterprise and technical support from NVIDIA.

Whether or not you have a suitable NVIDIA Enterprise Account depends on whether you have previously purchased NVIDIA AI Enterprise.

  • If you have previously purchased NVIDIA AI Enterprise, you already have a suitable NVIDIA Enterprise Account.

To use this account to get NVIDIA AI Enterprise, download the software assets that you require from the NVIDIA AI Enterprise Infra Release 5 collection on NVIDIA NGC. For details, refer to Accessing the NVIDIA AI Enterprise Software Suite.

  • If you have obtained an evaluation license but have not previously purchased NVIDIA AI Enterprise, you do not have a suitable NVIDIA Enterprise Account. To create a suitable NVIDIA Enterprise Account, follow the Register link in the instructions for using the certificate to create an account for your purchased licenses. You can choose to create a separate account for your purchased licenses or link your existing account for an evaluation license to the account for your purchased licenses.

  • If you have not previously purchased NVIDIA AI Enterprise, you do not have a suitable NVIDIA Enterprise Account.

To create a suitable NVIDIA Enterprise Account, follow the Register link in the instructions for using the certificate to create your account. For details, refer to Creating your NVIDIA Enterprise Account.

Creating your NVIDIA Enterprise Account#

If you do not have an NVIDIA Enterprise Account, you must create an account to get access to the NVIDIA AI Enterprise software components and technical support from NVIDIA. For details on these software components, refer to Activating the Accounts for Getting NVIDIA AI Enterprise.

If you have an account that was created for an evaluation license and you want to access licenses that you purchased, you must repeat the registration process when you receive your purchased licenses. You can choose to create a separate account for your purchased licenses or link your existing account for an evaluation license to the account for your purchased licenses.

  • To create a separate account for your purchased licenses, perform this task, specifying a different email address than the address with which you created your existing account.

  • To link your existing account for an evaluation license to the account for your purchased licenses, follow the instructions in Linking an Evaluation Account to an NVIDIA Enterprise Account for Purchased Licenses, specifying the email address with which you created your existing account.

Before you begin, ensure that you have your order confirmation message.

  1. In the instructions for using your NVIDIA Entitlement Certificate, follow the Register link.

  2. Fill out the form on the NVIDIA Enterprise Account Registration page and click REGISTER. A message confirming that an account has been created appears. An email instructing you to log in to your account on the NVIDIA Application Hub is sent to the email address you provided.

  3. Open the email instructing you to log in to your account and click Log In.

  4. On the NVIDIA Application Hub Login page that opens, in the text-entry field, type the email address you provided and click Sign In.

  5. On the Create Your Account page that opens, provide and confirm a password for the account and click Create Account. A message prompting you to verify your email address appears. An email instructing you to verify your email address is sent to the email address you provided.

  6. Open the email instructing you to verify your email address and click Verify Email Address. A message confirming that your email address is confirmed appears.

From the NVIDIA Application Hub page, you can now log in to the components that are listed in Activating the Accounts for Getting NVIDIA AI Enterprise.

Linking an Evaluation Account to an NVIDIA Enterprise Account for Purchased Licenses#

If you have an account that was created for an evaluation license, you must repeat the registration process when you receive your purchased licenses. To link your existing account for an evaluation license to the account for your purchased licenses, register for an NVIDIA Enterprise Account with the email address with which you created your existing account.

If you want to create a separate account for your purchased licenses, follow the instructions in Creating your NVIDIA Enterprise Account, specifying a different email address than the address with which you created your existing account.

  1. In the instructions for using the NVIDIA Entitlement Certificate for your purchased licenses, follow the Register link.

  2. Fill out the form on the NVIDIA Enterprise Account Registration page, specifying the email address with which you created your existing account, and click Register.

  3. When a message stating that your email address is already linked to an evaluation account is displayed, click LINK TO NEW ACCOUNT.

  4. Log in to the NVIDIA Licensing Portal with the credentials for your existing account.

Installing NVIDIA AI Enterprise Software Components#

The NVIDIA NGC Catalog#

NVIDIA AI Enterprise components are distributed through the NVIDIA NGC Catalog. Infrastructure and workload management components are distributed as resources in the NVIDIA AI Enterprise Infra Release 5 collection. Tools for AI development and use cases are available from the NVIDIA AI Enterprise Software Suite.

Accessing the NVIDIA AI Enterprise Infrastructure Resources#

Infrastructure and workload management components of NVIDIA AI Enterprise are distributed as resources in the NVIDIA AI Enterprise Infra Release 5 collection.

The NVIDIA AI Enterprise Infra Release 5 collection offers infrastructure management and orchestration software bundled together to efficiently manage and scale AI workloads, and contains the following resources:

  • GPU Operator

  • Network Operator

  • NVIDIA NIM Operator

  • vGPU Host Driver

  • Containerized vGPU Guest Driver

  • Containerized NVIDIA GPU Datacenter Driver

  • NVIDIA Base Command Manager Essentials

Before downloading any NVIDIA AI Enterprise software assets, ensure that you have signed in to NVIDIA NGC from the NVIDIA NGC Sign In page.

  1. Go to the NVIDIA AI Enterprise Infra Release 5 collection on NVIDIA NGC.

  2. Click the Entities tab and select the resource that you are interested in.

  3. Click Download and, from the menu that opens, choose to download the resource by using a direct download in the browser, the displayed wget command, or the CLI.

Accessing the NVIDIA AI Enterprise Software Suite#

Tools for AI development and use cases are available from the NVIDIA AI Enterprise Software Suite and are distributed through the NVIDIA NGC Catalog.

Before downloading any NVIDIA AI Enterprise software assets, ensure that you have signed in to NVIDIA NGC from the NVIDIA NGC Sign In page.

  1. View the NVIDIA AI Enterprise Software Suite on NVIDIA NGC.

  2. Browse the NVIDIA AI Enterprise Software Suite to find software assets that you are interested in.

  3. For each software asset that you are interested in, click the asset to learn more about or download the asset.

NVIDIA AI Enterprise Deployment Options#

The following table outlines the most common approaches used to deploy NVIDIA AI Enterprise.

Installing NVIDIA AI Enterprise on Bare Metal Ubuntu 22.04#

This section of the NVIDIA AI Enterprise Quick Start Guide provides minimal instructions for a bare-metal, single-node deployment of NVIDIA AI Enterprise using Docker on a third-party NVIDIA-certified system.

NVIDIA AI Enterprise Software Prerequisites#

To enable NVIDIA GPU acceleration for compute and AI workloads running in data centers:

  1. Download the NVIDIA GPU data center drivers from this location.

  2. Select Linux 64 bits, Ubuntu 22.04 as the Operating System to download the .run file.

Installing NVIDIA AI Enterprise using the TRD Driver on Ubuntu 22.04 from a .run File#

Installation of the NVIDIA AI Enterprise software driver for Linux requires:

  • Compiler toolchain

  • Kernel headers

Prerequisites

Ensure you follow the pre-installation steps.

Note

If you prefer to use a Debian package, refer to the Debian instructions.

Steps

  1. Log into the system and check for updates.

    sudo apt-get update
    
  2. Install the GCC compiler and the make tool in the terminal.

    sudo apt-get install build-essential
    
  3. Copy the NVIDIA AI Enterprise Linux driver package, for example, NVIDIA-Linux-x86_64-550.90.12.run, to the host machine where you are installing the driver.

    Where x_x:xxx.xx.xx is the current NVIDIA AI Enterprise version and driver version.

  4. Navigate to the directory containing the NVIDIA Driver .run file. Then, add the Executable permission to the NVIDIA Driver file using the chmod command.

    sudo chmod +x NVIDIA-Linux-x86_64-xxx.xx.xx-grid.run
    
  5. From a console shell, run the driver installer as the root user, and accept the defaults.

    sudo sh ./NVIDIA-Linux-x86_64-xxx.xx.xx-grid.run
    
  6. Reboot the system.

    sudo reboot
    
  7. After the system has rebooted, confirm that you can see your NVIDIA vGPU device in the output from nvidia-smi.

    nvidia-smi
    

Installing the NVIDIA Container Toolkit#

The NVIDIA Container Toolkit allows users to build and run GPU accelerated Docker containers. The toolkit includes a container runtime library and utilities to configure containers to leverage NVIDIA GPUs automatically. Complete documentation and frequently asked questions are available on the repository wiki.

NVIDIA Container Toolkit
  1. Install Docker - Refer to Install Docker Engine on Ubuntu | Docker Documentation for the installation procedure for Ubuntu.

  2. Install the NVIDIA Container Toolkit - Refer to Installing the NVIDIA Container Toolkit | NVIDIA Documentation for the installation procedure to enable the Docker repository and install the NVIDIA Container Toolkit.

  3. After the NVIDIA Container Toolkit is installed, to configure the Docker container runtime, refer to Configuration | NVIDIA Documentation.

Verifying the Installation of NVIDIA Container Toolkit#

  1. Run the nvidia-smi command contained in the latest official NVIDIA CUDA Toolkit image that is compatible with the release of the NVIDIA CUDA Toolkit driver that is running on your machine.

    Note

    Do not use a release of the NVIDIA CUDA Toolkit image later than the release of the NVIDIA CUDA Toolkit driver that is running on your machine. For a list of all NVIDIA CUDA Toolkit images, refer to nvidia/cuda on Docker Hub.

    $ sudo docker run --rm --runtime=nvidia --gpus all ubuntu nvidia-smi
    
  2. Start a GPU-enabled container on any two available GPUs.

    $ docker run --runtime=nvidia --gpus 2 nvidia/cuda:12.4.0-base-ubuntu22.04 nvidia-smi
    
  3. Start a GPU-enabled container on two specific GPUs identified by their index numbers.

    $ docker run --runtime=nvidia --gpus '"device=1,2"' nvidia/cuda:12.4.0-base-ubuntu22.04 nvidia-smi
    
  4. Start a GPU-enabled container on two specific GPUs with one GPU identified by its UUID and the other GPU identified by its index number.

    $ docker run --runtime=nvidia --gpus '"device=UUID-ABCDEF,1"' nvidia/cuda:12.4.0-base-ubuntu22.04 nvidia-smi
    
  5. Specify a GPU capability for the container.

    $ docker run --runtime=nvidia --gpus all,capabilities=utility nvidia/cuda:12.4.0-base-ubuntu22.04 nvidia-smi
    

Installing Software Distributed as Container Images#

The NGC container images accessed through the NVIDIA NGC Catalog include the AI and data science applications and frameworks. Each container image for an AI and data science application or framework contains the entire user-space software stack that is required to run the application or framework, namely, the CUDA libraries, cuDNN, any required Magnum IO components, TensorRT, and the framework.

Ensure that you have completed the following tasks in the NGC Private Registry User Guide:

Perform this task from the host machine. Obtain the Docker pull command to download each of the following applications and deep learning framework components from the listing for the application or component in the NGC Public Catalog.

  • Applications

    • NVIDIA NIMs

    • NVIDIA Clara Parabricks

    • NVIDIA DeepStream

    • NVIDIA Riva

    • MONAI - Medical Open Network for Artificial Intelligence

    • RAPIDS

    • RAPIDS Accelerator for Apache Spark

    • TAO

  • Deep learning framework components

    • NVIDIA TensorRT

    • NVIDIA Triton Inference Server

    • PyTorch

    • TensorFlow 2

Running ResNet-50 with TensorFlow#

  1. Launch the TensorFlow 1 container image on all GPUs in interactive mode, specifying that the container will be deleted when stopped.

    $ sudo docker run --gpus all -it --rm \
    nvcr.io/nvidia/tensorflow:21.07-tf1-py3
    
  2. From within the container runtime, change to the directory that contains test data for cnn example.

    # cd /workspace/nvidia-examples/cnn
    
  3. Run the ResNet-50 training test with FP16 precision.

    # python resnet.py --layers 50 -b 64 -i 200 -u batch --precision fp16
    
  4. Confirm that all operations on the application are performed correctly and that a set of results is reported when the test is completed.

  5. Press Ctrl+P, Ctrl+Q to exit the container runtime and return to the Linux command shell.

Running ResNet-50 with TensorRT#

  1. Launch the NVIDIA TensorRT container image on all GPUs in interactive mode, specifying that the container will be deleted when stopped.

    $ sudo docker run --gpus all -it --rm nvcr.io/nvidia/tensorrt:21.07-py3
    
  2. From within the container runtime, change to the directory that contains test data for the ResNet-50 convolutional neural network.

    # cd /workspace/tensorrt/data/resnet50
    
  3. Run the ResNet-50 convolutional neural network with FP32, FP16, and INT8 precision and confirm that each test is completed with the result PASSED.

    1. To run ResNet-50 with the default FP32 precision, run this command:

      # trtexec --duration=90 --workspace=1024 --percentile=99 --avgRuns=100 \
      --deploy=ResNet50_N2.prototxt --batch=1 --output=prob
      
    2. To run ResNet-50 with FP16 precision, add the --fp16 option:

      # trtexec --duration=90 --workspace=1024 --percentile=99 --avgRuns=100 \
      --deploy=ResNet50_N2.prototxt --batch=1 --output=prob --fp16
      
    3. To run ResNet-50 with INT8 precision, add the --int8 option:

      # trtexec --duration=90 --workspace=1024 --percentile=99 --avgRuns=100 \
      --deploy=ResNet50_N2.prototxt --batch=1 --output=prob --int8
      
  4. Press Ctrl+P, Ctrl+Q to exit the container runtime and return to the Linux command shell.

Running NVIDIA NIM on Bare Metal Ubuntu 22.04#

NVIDIA Inference Microservices (NIMs) provide a streamlined path for developing AI-powered enterprise applications and deploying AI models in production.You can download and run the NIM of your choice from the NVIDIA NGC Catalog. Follow the instructions below to deploy a Llama3 8B Instruct NIM on your bare metal host machine setup and run inference.

  1. Pull and run meta/llama3-8b using Docker (this will download the full model and run it in your local environment).

    $ docker login nvcr.io
    Username: $oauthtoken
    Password: <PASTE_API_KEY_HERE>
    
  2. Pull and run NVIDIA NIM. This will download the optimized model for your infrastructure.

    export NGC_API_KEY=<PASTE_API_KEY_HERE>
    export LOCAL_NIM_CACHE=~/.cache/nim
    mkdir -p "$LOCAL_NIM_CACHE"
    docker run -it --rm \
        --gpus all \
        --shm-size=16GB \
        -e NGC_API_KEY=$NGC_API_KEY \
        -v "$LOCAL_NIM_CACHE:/opt/nim/.cache" \
        -u $(id -u) \
        -p 8000:8000 \
        nvcr.io/nim/meta/llama3-8b-instruct:1.0.0
    
  3. Make a local API call.

    curl -X 'POST' \
    'http://0.0.0.0:8000/v1/chat/completions' \
    -H 'accept: application/json' \
    -H 'Content-Type: application/json' \
    -d '{
        "model": "meta/llama3-8b-instruct",
        "messages": [{"role":"user", "content":"Write a limerick about the wonders of GPU computing."}],
        "max_tokens": 64
    }'
    

For more information about running inference on this locally deployed LLM NIM, refer to Launch NVIDIA NIM for LLMs.

Installing NVIDIA AI Enterprise on Public Cloud#

NVIDIA AI Enterprise can be run on Amazon Web Services (AWS), Google Cloud, Microsoft Azure, Oracle Cloud Infrastructure (OCI), Alibaba Cloud, and Tencent Cloud.

This section of the NVIDIA AI Enterprise Quick Start Guide provides minimal instructions for deploying NVIDIA AI Enterprise on Microsoft Azure using the NVIDIA AI Enterprise VMI.

Installing NVIDIA AI Enterprise on Microsoft Azure using the NVIDIA AI Enterprise VMI#

The NVIDIA AI Enterprise On-Demand VMI spins up a GPU-accelerated Compute Engine VM instance in minutes with pre-installed software for accelerating your Machine Learning, Deep Learning, Data Science, and HPC workloads.

The On-Demand VMI is preconfigured with the following software:

  • Ubuntu Operating System

  • NVIDIA GPU Data Center Driver

  • Docker-ce

  • NVIDIA Container Toolkit

  • CSP CLI, NGC CLI

  • Miniconda, JupyterLab, Git

  • Token Activation Script

Getting started with NVIDIA AI Enterprise in your Enterprise (On-Demand) VMI cloud instance can be broken down into two simple steps:

  1. Authorize the VMI cloud instance with NVIDIA NGC by copying over the provided instance ID token into the Activate Subscription page on NGC. There are four key steps to complete this part of the process:

    1. Get an identity token from the VMI.

    2. Activate your NVIDIA AI Enterprise subscription with the token.

    3. Generate an API key for accessing the catalog.

    4. Put the API key on the VMI.

    Follow the instructions outlined here to complete this step.

  2. Pulling and running NVIDIA AI Enterprise Containers. Refer to this section of the Cloud Deployment Guide for pulling and running NGC container images through the NVIDIA NGC Catalog.

Detailed instructions on how to install NVIDIA AI Enterprise on the public cloud can be found in the Microsoft Azure Overview.

Running an LLM NIM on Microsoft Azure using NVIDIA AI Enterprise#

Refer to the Appendix section of the Cloud Deployment Guide for pulling and running NGC container images through the NVIDIA NGC Catalog.

Installing NVIDIA AI Enterprise in Virtualized Environments using NVIDIA vGPU (C-Series)#

The NVIDIA AI Enterprise VMware Deployment Guide offers detailed information for deploying NVIDIA AI Enterprise on a third-party NVIDIA-certified system running VMware vSphere using NVIDIA vGPU (C-Series).

NVIDIA AI Enterprise Software Prerequisites#

  1. NVIDIA vGPU (C-Series) Host Driver

  2. NVIDIA vGPU (C-Series) Guest Driver

  3. NVIDIA License System

To download the NVIDIA vGPU (C-Series) software drivers, follow the instructions in Accessing the NVIDIA AI Enterprise Infrastructure Resources.

The NVIDIA AI Enterprise license entitles customers to download NVIDIA vGPU (C-Series) software through NVIDIA NGC. Deploying NVIDIA AI Enterprise using vGPU requires a valid license that is enforced through the installation and the NVIDIA License System. The NVIDIA License System is used to serve a pool of floating licenses to licensed NVIDIA software products and is configured with licenses obtained from the NVIDIA Licensing Portal.

NVIDIA License System supports the following types of service instances:

  • Cloud License Service (CLS) instance. A CLS instance is hosted on the NVIDIA Licensing Portal.

  • Delegated License Service (DLS) instance. A DLS instance is hosted on-premises at a location that is accessible from your private network, such as inside your data center.

An NVIDIA vGPU (C-Series) client VM with a network connection obtains a license by leasing it from an NVIDIA License System service instance. The service instance serves the license to the client over the network from a pool of floating licenses obtained from the NVIDIA Licensing Portal. The license is returned to the service instance when the licensed client no longer requires the license.

To activate an NVIDIA vGPU (C-Series), software licensing must be configured for the vGPU VM client when it is booted. NVIDIA vGPU (C-Series) VMs run at a reduced capability until a license is acquired.

Refer to the NVIDIA Licensing Quick Start Guide for instructions on configuring an express Cloud License Service (CLS) instance and verifying the license status of a licensed vGPU VM client.

Refer to the NVIDIA License System User Guide if you use Delegated License Service (DLS) instances to serve licenses. Detailed instructions on how to install, configure, and manage the NVIDIA License System can be found here.

Obtaining NVIDIA Base Command Manager Essentials#

NVIDIA Base Command Manager Essentials streamlines cluster provisioning, workload management, and infrastructure monitoring in the data center. In bare-metal deployments, NVIDIA Base Command Manager Essentials simplifies the installation of operating systems supported by NVIDIA Base Command Manager Essentials.

Before obtaining NVIDIA Base Command Manager Essentials, ensure that you have activated the accounts for getting NVIDIA AI Enterprise, as explained in Activating the Accounts for Getting NVIDIA AI Enterprise.

  1. Request your NVIDIA Base Command Manager Essentials product keys by sending an email with your entitlement certificate to sw-bright-sales-ops@NVIDIA.onmicrosoft.com. After your entitlement certificate has been reviewed, you will receive a product key from which you can generate a license key for the number of licenses that you purchased.

  2. Go to this page to download NVIDIA Base Command Manager Essentials for your operating system.

For detailed instructions on deploying and using Base Command Manager Essentials, refer to the Base Command Manager Essentials Product Manuals.

After obtaining NVIDIA Base Command Manager Essentials, follow the steps in the NVIDIA Base Command Manager Essentials Installation Manual to create and license your head node.