TensorRT-RTX 1.0 Support Matrix#

GPU Architecture and Precision Support#

The following table lists NVIDIA hardware and the precision modes that each hardware supports.

Table 13 Supported Hardware#

CUDA Compute Capability

GPU Architecture

Example Devices

FP32

FP16

BF16

INT8 WoQ GEMM

INT4 WoQ GEMM [2]

FP8 GEMM

FP4 GEMM

12.0

NVIDIA Blackwell

NVIDIA RTX 5090

Yes

Yes

Yes

Yes

Yes

Yes

Yes

8.9

NVIDIA Ada

NVIDIA RTX 4090

Yes

Yes

Yes

Yes

Yes

Yes

No

8.6

NVIDIA Ampere

NVIDIA RTX 3090

Yes

Yes

Yes

Yes

Yes

No

No

7.5

NVIDIA Turing

NVIDIA RTX 2080Ti

Yes [1]

Yes

No

No

No

No

No

Note

  • Engines built with TensorRT-RTX are by default portable across devices with compute capabilities 8.6, 8.9, and 12.0, if the model precisions are supported on the target architecture.

  • By default, TensorRT-RTX does not create portable engines that are compatible with Turing when targeting other compute capabilities, as this can impact performance. Instead of relying on multi-device portable engines that include Turing support, developers are encouraged to build optimized, device-specific engines for Turing using appropriate API flags.

Supported Compute Capabilities#

The following table shows which GPU compute capabilities are supported on each platform.

Table 14 Supported Compute Capability per Platform#

Compute Capability

Platform

Linux x86-64

Windows x64

12.0

Yes

Yes

8.9

Yes

Yes

8.6

Yes

Yes

7.5

Yes

Yes

Compiler and Python Requirements#

The following table shows the required compiler and Python versions for building and using TensorRT on each supported platform.

Table 15 List of Supported Platforms per Software Version#

Platform

Compiler Version

Tested CUDA Version

TensorRT-RTX Python API [3]

TensorRT-RTX C++ API

  • Rocky Linux 8.9 x86-64

  • Ubuntu 22.04 x86-64

  • Ubuntu 24.04 x86-64

gcc 11.2.1

12.9

Yes

Yes

  • Windows 10 x64

  • Windows 11 x64

MSVC 2019 v16.9.2

12.9

Yes

Yes

Footnotes