Step 2: Set Up Required Infrastructure

Digital Fingerprinting

NVIDIA AI Workflows can be deployed on-premise or using a cloud service provider (CSP). NVIDIA AI Workflows at minimum require a single GPU-enabled node for running the provided example workload. Production deployments should be performed in an HA environment.

The following hardware specification is recommended for this workflow:

  • 2x T4 (or newer) GPUs

  • 32 vCPU Cores

  • 128 GB RAM

  • 1 TB HDD

  • Ports 443, 22 allowed for application ingress and egress

  • DNS - A wildcard DNS A record must be created for the system along with the DNS A record for the system itself. Reverse lookup PTR records should also exist for both entries. Both DNS records should be resolvable within and outside of the system.


    If the DNS entries are only resolvable within a local network, such as within a corporate domain, and not directly resolvable by the VMI, a manual reverse lookup entry can be made in /etc/hosts on the systems for to point to various required hostnames as a workaround. An example hosts file is provided below.


Follow the steps below to set up the instances meeting the above requirements, before proceeding to the next section.

  1. We will start by deploying an On-Demand NVIDIA AI Enterprise VMI in the cloud, meeting the above specifications. You can find this VMI under the Marketplaces for major CSPs; follow the instructions from the listing to provision an instance.

  2. Once the instance has been provisioned, refer to the NVIDIA AI Enterprise Cloud Guide to authorize the instance and activate your subscription.

  3. Once your subscription has been activated, review the Prerequisites section to ensure you can access the Enterprise Catalog and create an NGC API Key if you do not already have one.

  4. After the Hardware Requirements and Prerequisites sections have been completed, move on to the Cloud Native Software Requirements section.

© Copyright 2022-2023, NVIDIA. Last updated on Mar 20, 2023.