Installing TensorRT-RTX#
There are several installation methods for TensorRT-RTX. This section covers the most common options using:
An SDK zip file (Windows), or
A tarball file (Linux), or
A standalone
pip
wheel file (all platforms).
Prerequisites#
Ensure you are a member of the NVIDIA Developer Program. If you need help, follow the prompts to gain access and download the package you want to install.
Go to this download link.
Click GET STARTED, then click Download Now.
Select the version of TensorRT that you are interested in.
Select the checkbox to agree to the license terms.
Click the package you want to install. Your download begins.
Install NVIDIA CUDA Toolkit 12.9 or later on the target system.
Windows SDK Installation#
Decompress the Windows zip package. The following can be directly decompressed to extract the:
Core library and ONNX parser DLLs and import libraries
Development headers
Source code samples
Documentation
Python bindings
The
tensorrt_rtx
executable for building engines and running inference from the command lineLicensing information and open-source acknowledgments
Note
The DLLs are signed and verified with the
signtool
utility. Remember to add the directories containing the DLLs and executable files to yourPATH
environment variable.Optionally, install the Python bindings.
$version = "1.0.0.14" # Replace with newest version $arch = "amd64" # Replace with your architecture $pyversion = "311" # For Python 3.11, replace with your # Python version $wheel = "TensorRT-RTX-$version\tensorrt_rtx-$version-cp$pyversion-none-win_$arch.whl" python3 -m pip install $wheel
Linux Tarball Installation#
TensorRT-RTX can be installed from a tarball package on Linux similar to the zip package installation on Windows. It is supported on the following distributions:
Ubuntu 22.04
Ubuntu 24.04
Rocky Linux 8
Prerequisites
Ensure that NVIDIA CUDA Toolkit 12.9 or later is installed on the target system.
Steps
Download the repo file that matches your operating system version and CPU architecture.
Unzip the tarball, optionally adding the path of the executable to your PATH variable and the path of the library to your
LD_LIBRARY_PATH
variable.version = "1.0.0.14" # Replace with newest version arch = "x86_64" # Replace with your architecture cuda = "12.9" # Replace with your CUDA version tarfile = "TensorRT-RTX-${version}.Linux.${arch}-gnu-${cuda}.tar.gz" tar -xzf $tarfile export PATH=$PATH:$PWD/TensorRT-RTX-${version}/bin export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$PWD/TensorRT-RTX-${version}/lib
Optionally, install the Python bindings.
pyversion = "311" # Assuming Python 3.11, else replace with your # Python version wheel = "tensorrt_rtx-${version}-cp${pyversion}-none-linux_${arch}.whl" python3 -m pip install TensorRT-RTX-${version}/python/${wheel}
Python Package Index Installation#
When installing TensorRT-RTX from the Python package Index, it is not necessary to install it first from a zip file or tar file package. All necessary libraries are included in the Python package and will be downloaded automatically.
The tensorrt_rtx
Python wheel files currently support Python versions 3.8 to 3.13 and will not work with other versions. We support Linux and Windows operating systems, specifically Ubuntu 20.04 or newer, Rocky Linux 8, and Windows 10 or newer.
We recommend installing inside a Python virtual environment. Alternatively, when not running as root
, the --user
flag may be appended to any of the pip
commands provided to install the wheels in a local path.
Prerequisites
If necessary, upgrade the
pip
module to the latest version and ensure thewheel
Python module is present.python3 -m pip install --upgrade pip python3 -m pip install wheel
Steps
Install the TensorRT-RTX Python wheel. The following
pip
command will pull in all the required CUDA libraries in Python wheel format from PyPI because they are dependencies of the TensorRT-RTX Python wheel. Also, it will upgradetensorrt_rtx
to the latest version if you have a previous version installed.python3 -m pip install --upgrade tensorrt_rtx
Verify that the installation is working.
python3 >>> import tensorrt_rtx >>> print(tensorrt_rtx.__version__) >>> assert(tensorrt_rtx.Builder(tensorrt_rtx.Logger())
Basic TensorRT-RTX Workflow#
Models can be specified via the TensorRT-RTX C++ or Python API, or read from the ONNX neural network exchange format. Popular model training frameworks like PyTorch or TensorFlow typically offer an ONNX export option. Alternatively, the native TensorRT-RTX API defines operators for manually assembling the network structure.
ONNX is a framework-agnostic option that works with models in TensorFlow, PyTorch, and more. TensorRT-RTX supports an automatic conversion from ONNX files using the TensorRT-RTX API or the tensorrt_rtx
executable, which we will use in this section. ONNX conversion is all-or-nothing, meaning all operations in your model must be supported by TensorRT-RTX.
For the most performance and customizability possible, you can manually construct TensorRT-RTX engines using the TensorRT-RTX network definition API. This involves building an identical network to your target model in TensorRT-RTX operation by operation, using only TensorRT-RTX operations. After a TensorRT-RTX network is created, you will export just the weights of your model from the framework and load them into your TensorRT-RTX network. For this approach, more information about constructing the model using TensorRT-RTX’s network definition API can be found here: