TRTEXEC with Multitask Classification

TAO Toolkit v5.2.0

The trtexec tool is a command-line wrapper included as part of the TensorRT samples. TAO 5.0.0 exposes the trtexec tool in the TAO Deploy container (or task group when run via launcher) for deploying the model with an x86-based CPU and discrete GPUs. To run trtexec on other platforms, such as Jetson devices, or with versions of TensorRT that are not used by default in the TAO containers, you can follow the official TensorRT documentation on how to get trtexec.

This section describes how to generate a TensorRT engine using trtexec, which allows you to deploy TAO-trained models on TensorRT, Triton, and Deepstream.

To generate an .onnx file for Multitask Classification, refer to the Multitask Classification documentation. You can also refer to the Multitask Classification TAO-Deploy documentation for instructions on generating an INT8 calibration file.


# Running using INT8 trtexec --onnx=/path/to/model.onnx \ --maxShapes=input_1:16x3x80x60 \ --minShapes=input_1:1x3x80x60 \ --optShapes=input_1:8x3x80x60 \ --calib=/path/to/cal.bin \ --fp16 \ --int8 \ --saveEngine=/path/to/model.plan # Running using FP16 trtexec --onnx=/path/to/model.onnx \ --maxShapes=input_1:16x3x80x60 \ --minShapes=input_1:1x3x80x60 \ --optShapes=input_1:8x3x80x60 \ --fp16 \ --saveEngine=/path/to/model.plan

Previous TRTEXEC with Mask RCNN
Next TRTEXEC with OCDNet
© Copyright 2024, NVIDIA. Last updated on Mar 18, 2024.