Deployment Guide#
This section describes the various ways you can deploy NIM, including the following:
By using Docker, as described in Getting Started
By using a Helm chart, as described in Deploying with Helm
By using Kubernetes, as described in Kubernetes Installation
In multiple nodes, as described in Multi-Node Deployment
Locally, as described in Air Gap Deployment
Deploying on other platforms#
In addition to the NVIDIA deployment options, you can also deploy on other platforms:
NVIDIA NIM on WSL2 provides instructions on setting up and configuring to deploy on Windows PCs using Windows Subsystem for Linux (WSL). We recommend that you set
NIM_RELAX_MEM_CONSTRAINTS=1
when you deploy with Docker on RTX GPUs to avoid high memory usage.The NIM on Azure Kubernetes Service (AKS) deployment guide provides step-by-step instructions for deploying AKS.
The NIM on Azure Machine Learning (AzureML) deployment guide provides step-by-step instructions for deploying AzureML using Azure CLI and Jupyter Notebook.
The End to End LLM App development with Azure AI Studio, Prompt Flow and NIMs deployment guide provides end-to-end LLM App development with Azure AI Studio, Prompt Flow, and NIMs.
The NIM on AWS Elastic Kubernetes Service (EKS) deployment guide provides step-by-step instructions for deploying on AWS EKS.
The NIM on AWS SageMaker deployment guide provides step-by-step instructions for deploying on AWS SageMaker using Jupyter Notebooks, Python CLI, and the shell.
The NIM on KServe deployment guide provides step-by-step on how to deploy on KServe.