Support Matrix#
Models#
Model Name |
Model ID |
Publisher |
|
---|---|---|---|
FLUX.1-dev |
black-forest-labs/flux.1-dev |
Black Forest Labs |
base, canny, depth |
FLUX.1-Kontext-dev |
black-forest-labs/flux.1-kontext-dev |
Black Forest Labs |
base |
FLUX.1-schnell |
black-forest-labs/flux.1-schnell |
Black Forest Labs |
base |
Stable Diffusion 3.5 Large |
stabilityai/stable-diffusion-3.5-large |
Stability AI |
base, canny, depth |
TRELLIS |
microsoft/trellis |
Microsoft |
base:text, large:text, large:image |
Supported Hardware#
Black Forest Labs / FLUX.1-dev#
System Requirements |
GPU Memory |
RAM |
OS |
CPU |
---|---|---|---|---|
Minimal |
16GB |
40GB |
Linux/WSL2 |
x86_64 |
Recommended |
32GB |
64GB |
Linux/WSL2 |
x86_64 |
NVIDIA Virtual GPU (vGPU) technology is supported if vGPU configuration meets at least minimal GPU memory requirements.
NVIDIA supports the FLUX.1-dev model on the following GPUs with the optimized TensorRT engine.
GPU |
GPU Memory (GB) |
Precision |
---|---|---|
GeForce RTX 5090 |
32 |
FP4, FP8 |
GeForce RTX 5090 Laptop |
24 |
FP4, FP8 |
GeForce RTX 5080 |
16 |
FP4, FP8 |
GeForce RTX 5080 Laptop |
16 |
FP4, FP8 |
GeForce RTX 5070 TI |
16 |
FP4, FP8 |
GeForce RTX 4090 |
24 |
FP8 |
GeForce RTX 4090 Laptop |
16 |
FP8 |
GeForce RTX 4080 Super |
16 |
FP8 |
GeForce RTX 4080 |
16 |
FP8 |
NVIDIA RTX 6000 Ada Generation |
48 |
FP8 |
GeForce RTX 5090D |
32 |
FP4, FP8 |
GeForce RTX 4090D |
24 |
FP8 |
H100 SXM |
80 |
FP8 |
L40S |
48 |
FP8 |
L40 |
48 |
FP8 |
Black Forest Labs / FLUX.1-Kontext-dev#
System Requirements |
GPU Memory |
RAM |
OS |
CPU |
---|---|---|---|---|
Minimal |
16GB |
40GB |
Linux/WSL2 |
x86_64 |
Recommended |
32GB |
64GB |
Linux/WSL2 |
x86_64 |
NVIDIA Virtual GPU (vGPU) technology is supported if vGPU configuration meets at least minimal GPU memory requirements.
NVIDIA supports the FLUX.1-Kontext-dev model on the following GPUs with the optimized TensorRT engine.
GPU |
GPU Memory (GB) |
Precision |
---|---|---|
GeForce RTX 5090 |
32 |
FP4, FP8 |
GeForce RTX 5090 Laptop |
24 |
FP4, FP8 |
GeForce RTX 5080 |
16 |
FP4, FP8 |
GeForce RTX 5080 Laptop |
16 |
FP4, FP8 |
GeForce RTX 5070 TI |
16 |
FP4, FP8 |
GeForce RTX 4090 |
24 |
FP8 |
GeForce RTX 4090 Laptop |
16 |
FP8 |
GeForce RTX 4080 Super |
16 |
FP8 |
GeForce RTX 4080 |
16 |
FP8 |
NVIDIA RTX PRO 6000 Blackwell Workstation Edition |
96 |
FP4, FP8 |
NVIDIA RTX PRO 6000 Blackwell Server Edition |
96 |
FP4, FP8 |
NVIDIA RTX 6000 Ada Generation |
48 |
FP8 |
H100 SXM |
80 |
FP8 |
L40S |
48 |
FP8 |
Black Forest Labs / FLUX.1-schnell#
System Requirements |
GPU Memory |
RAM |
OS |
CPU |
---|---|---|---|---|
Minimal |
16GB |
40GB |
Linux/WSL2 |
x86_64 |
Recommended |
32GB |
40GB |
Linux/WSL2 |
x86_64 |
NVIDIA Virtual GPU (vGPU) technology is supported if vGPU configuration meets at least minimal GPU memory requirements.
NVIDIA supports the FLUX.1-schnell model on the following GPUs with the optimized TensorRT engine.
GPU |
GPU Memory (GB) |
Precision |
---|---|---|
GeForce RTX 5090 |
32 |
FP4, FP8 |
GeForce RTX 5090 Laptop |
24 |
FP4, FP8 |
GeForce RTX 5080 |
16 |
FP4, FP8 |
GeForce RTX 5080 Laptop |
16 |
FP4, FP8 |
GeForce RTX 5070 TI |
16 |
FP4, FP8 |
GeForce RTX 4090 |
24 |
FP8 |
GeForce RTX 4090 Laptop |
16 |
FP8 |
GeForce RTX 4080 Super |
16 |
FP8 |
GeForce RTX 4080 |
16 |
FP8 |
NVIDIA RTX 6000 Ada Generation |
48 |
FP8 |
GeForce RTX 5090D |
32 |
FP4, FP8 |
GeForce RTX 4090D |
24 |
FP8 |
H100 SXM |
80 |
FP8 |
L40S |
48 |
FP8 |
L40 |
48 |
FP8 |
Stability AI / Stable Diffusion 3.5 Large#
System Requirements |
GPU Memory |
RAM |
OS |
CPU |
---|---|---|---|---|
Minimal |
24GB |
48GB |
Linux/WSL2 |
x86_64 |
Recommended |
32GB |
64GB |
Linux/WSL2 |
x86_64 |
NVIDIA Virtual GPU (vGPU) technology is supported if vGPU configuration meets at least minimal GPU memory requirements.
NVIDIA supports the Stable Diffusion 3.5 Large model on the following GPUs with the optimized TensorRT engine.
GPU |
GPU Memory (GB) |
Precision |
---|---|---|
A100 SXM |
80 |
BF16 |
H100 SXM |
80 |
BF16 |
L40S |
48 |
BF16 |
Microsoft / TRELLIS#
System Requirements |
GPU Memory |
RAM |
OS |
CPU |
---|---|---|---|---|
Minimal |
12GB |
32GB |
Linux/WSL2 |
x86_64 |
Recommended |
24GB |
32GB |
Linux/WSL2 |
x86_64 |
NVIDIA Virtual GPU (vGPU) technology is supported if vGPU configuration meets at least minimal GPU memory requirements.
NVIDIA supports the TRELLIS model on any Ampere Architecture GPUs or newer with at least 12GB GPU Memory.
Software#
NVIDIA Driver#
NVIDIA NIM for Visual Generative AI is built on top of Triton Inference Server (25.06) which requires NVIDIA Driver release 560 or later.
Refer to the Release Notes for the detailed list of supported drivers.
NVIDIA Container Toolkit#
Your Docker environment must support NVIDIA GPUs. Refer to Installing the NVIDIA Container Toolkit for more information.
WSL2 Software Requirements#
A Windows 11 operating system (Build 23H2 and later) is supported via Windows Subsystem for Linux:
Minimum supported driver version is 570
Minimum linux distribution supported is Ubuntu 24.04
It is recommended to use Podman container management tools