NVIDIA TensorRT Inference Server
1.6.0 -0000000
Version select:
Current release
master (unstable)
Older releases
Documentation home
User Guide
Quickstart
Prerequisites
Using A Prebuilt Docker Container
Building With Docker
Building With CMake
Run TensorRT Inference Server
Verify Inference Server Is Running Correctly
Getting The Client Examples
Running The Image Classification Example
Installing the Server
Installing Prebuilt Containers
Running the Server
Example Model Repository
Running The Inference Server
Running The Inference Server On A System Without A GPU
Running The Inference Server Without Docker
Checking Inference Server Status
Client Libraries
Getting the Client Libraries
Build Using Dockerfile
Build Using CMake
Ubuntu 16.04 / Ubuntu 18.04
Windows 10
Download From GitHub
Building Your Own Client
Client API
Shared Memory
String Datatype
Client API for Stateful Models
Client Examples
Getting the Client Examples
Build Using Dockerfile
Build Using CMake
Ubuntu 16.04 / Ubuntu 18.04
Windows 10
Download From GitHub
Image Classification Example Application
Ensemble Image Classification Example Application
Performance Example Application
Models And Schedulers
Stateless Models
Stateful Models
Ensemble Models
Model Repository
Modifying the Model Repository
Model Versions
Framework Model Definition
TensorRT Models
TensorFlow Models
TensorRT/TensorFlow Models
ONNX Models
PyTorch Models
Caffe2 Models
Custom Backends
Custom Backend API
Example Custom Backend
Ensemble Backends
Model Configuration
Generated Model Configuration
Datatypes
Reshape
Version Policy
Instance Groups
Scheduling And Batching
Default Scheduler
Dynamic Batcher
Sequence Batcher
Ensemble Scheduler
Optimization Policy
Model Management
Model Control Mode NONE
Model Control Mode POLL
Model Control Mode EXPLICIT
Metrics
Developer Guide
Architecture
Concurrent Model Execution
Custom Operations
TensorRT
TensorFlow
HTTP and GRPC API
Health
Status
Model Control
Inference
Stream Inference
Library API
Building
Building the Server with Docker
Incremental Builds with Docker
Building the Server with CMake
Dependencies
CUDA, cuBLAS, cuDNN
TensorRT
TensorFlow
ONNX Runtime
PyTorch and Caffe2
Configure Inference Server
Build Inference Server
Building A Custom Backend
Build Using CMake
Build Using Custom Backend SDK
Using the Custom Instance Wrapper Class
Building the Client Libraries and Examples
Build Using Dockerfile
Build Using CMake
Ubuntu 16.04 / Ubuntu 18.04
Windows 10
Building the Documentation
Testing
Generate QA Model Repositories
Build QA Container
Run QA Container
Contributing
Coding Convention
Reference
Capabilities
Protobuf API
HTTP/GRPC API
Model Configuration
Status
C++ API
Class Hierarchy
File Hierarchy
Full API
Namespaces
Namespace nvidia
Namespace nvidia::inferenceserver
Namespace nvidia::inferenceserver::custom
Classes and Structs
Struct custom_initdata_struct
Struct custom_payload_struct
Class CustomInstance
Enums
Enum custom_memorytype_enum
Enum custom_serverparamkind_enum
Enum trtserver_errorcode_enum
Enum trtserver_memorytype_enum
Enum trtserver_metricformat_enum
Enum trtserver_modelcontrolmode_enum
Functions
Function CustomErrorString
Function CustomExecute
Function CustomExecuteV2
Function CustomFinalize
Function CustomInitialize
Function CustomVersion
Function TRTSERVER_ErrorCode
Function TRTSERVER_ErrorCodeString
Function TRTSERVER_ErrorDelete
Function TRTSERVER_ErrorMessage
Function TRTSERVER_ErrorNew
Function TRTSERVER_InferenceRequestProviderDelete
Function TRTSERVER_InferenceRequestProviderInputBatchByteSize
Function TRTSERVER_InferenceRequestProviderNew
Function TRTSERVER_InferenceRequestProviderSetInputData
Function TRTSERVER_InferenceResponseDelete
Function TRTSERVER_InferenceResponseHeader
Function TRTSERVER_InferenceResponseOutputData
Function TRTSERVER_InferenceResponseStatus
Function TRTSERVER_MetricsDelete
Function TRTSERVER_MetricsFormatted
Function TRTSERVER_ProtobufDelete
Function TRTSERVER_ProtobufSerialize
Function TRTSERVER_ResponseAllocatorDelete
Function TRTSERVER_ResponseAllocatorNew
Function TRTSERVER_ServerDelete
Function TRTSERVER_ServerId
Function TRTSERVER_ServerInferAsync
Function TRTSERVER_ServerIsLive
Function TRTSERVER_ServerIsReady
Function TRTSERVER_ServerLoadModel
Function TRTSERVER_ServerMetrics
Function TRTSERVER_ServerModelStatus
Function TRTSERVER_ServerNew
Function TRTSERVER_ServerOptionsAddTensorFlowVgpuMemoryLimits
Function TRTSERVER_ServerOptionsDelete
Function TRTSERVER_ServerOptionsNew
Function TRTSERVER_ServerOptionsSetExitOnError
Function TRTSERVER_ServerOptionsSetExitTimeout
Function TRTSERVER_ServerOptionsSetGpuMetrics
Function TRTSERVER_ServerOptionsSetLogError
Function TRTSERVER_ServerOptionsSetLogInfo
Function TRTSERVER_ServerOptionsSetLogVerbose
Function TRTSERVER_ServerOptionsSetLogWarn
Function TRTSERVER_ServerOptionsSetMetrics
Function TRTSERVER_ServerOptionsSetModelControlMode
Function TRTSERVER_ServerOptionsSetModelRepositoryPath
Function TRTSERVER_ServerOptionsSetProfiling
Function TRTSERVER_ServerOptionsSetServerId
Function TRTSERVER_ServerOptionsSetStrictModelConfig
Function TRTSERVER_ServerOptionsSetStrictReadiness
Function TRTSERVER_ServerOptionsSetTensorFlowGpuMemoryFraction
Function TRTSERVER_ServerOptionsSetTensorFlowSoftPlacement
Function TRTSERVER_ServerPollModelRepository
Function TRTSERVER_ServerRegisterSharedMemory
Function TRTSERVER_ServerSharedMemoryAddress
Function TRTSERVER_ServerStatus
Function TRTSERVER_ServerStop
Function TRTSERVER_ServerUnloadModel
Function TRTSERVER_ServerUnregisterAllSharedMemory
Function TRTSERVER_ServerUnregisterSharedMemory
Function TRTSERVER_SharedMemoryBlockCpuNew
Function TRTSERVER_SharedMemoryBlockDelete
Defines
Define CUSTOM_NO_GPU_DEVICE
Define CUSTOM_SERVER_PARAMETER_CNT
Define TRTIS_CUSTOM_EXPORT
Define TRTSERVER_EXPORT
Typedefs
Typedef CustomErrorStringFn_t
Typedef CustomExecuteFn_t
Typedef CustomExecuteV2Fn_t
Typedef CustomFinalizeFn_t
Typedef CustomGetNextInputFn_t
Typedef CustomGetNextInputV2Fn_t
Typedef CustomGetOutputFn_t
Typedef CustomGetOutputV2Fn_t
Typedef CustomInitializeData
Typedef CustomInitializeFn_t
Typedef CustomMemoryType
Typedef CustomPayload
Typedef CustomServerParameter
Typedef CustomVersionFn_t
Typedef TRTSERVER_Error_Code
Typedef TRTSERVER_InferenceCompleteFn_t
Typedef TRTSERVER_Memory_Type
Typedef TRTSERVER_Metric_Format
Typedef TRTSERVER_Model_Control_Mode
Typedef TRTSERVER_ResponseAllocatorAllocFn_t
Typedef TRTSERVER_ResponseAllocatorReleaseFn_t
Python API
Client
NVIDIA TensorRT Inference Server
Docs
»
Full API
View page source
Full API
¶
Namespaces
¶
Namespace nvidia
Namespaces
Namespace nvidia::inferenceserver
Namespaces
Namespace nvidia::inferenceserver::custom
Classes
Classes and Structs
¶
Struct custom_initdata_struct
Struct Documentation
Struct custom_payload_struct
Struct Documentation
Class CustomInstance
Class Documentation
Enums
¶
Enum custom_memorytype_enum
Enum Documentation
Enum custom_serverparamkind_enum
Enum Documentation
Enum trtserver_errorcode_enum
Enum Documentation
Enum trtserver_memorytype_enum
Enum Documentation
Enum trtserver_metricformat_enum
Enum Documentation
Enum trtserver_modelcontrolmode_enum
Enum Documentation
Functions
¶
Function CustomErrorString
Function Documentation
Function CustomExecute
Function Documentation
Function CustomExecuteV2
Function Documentation
Function CustomFinalize
Function Documentation
Function CustomInitialize
Function Documentation
Function CustomVersion
Function Documentation
Function TRTSERVER_ErrorCode
Function Documentation
Function TRTSERVER_ErrorCodeString
Function Documentation
Function TRTSERVER_ErrorDelete
Function Documentation
Function TRTSERVER_ErrorMessage
Function Documentation
Function TRTSERVER_ErrorNew
Function Documentation
Function TRTSERVER_InferenceRequestProviderDelete
Function Documentation
Function TRTSERVER_InferenceRequestProviderInputBatchByteSize
Function Documentation
Function TRTSERVER_InferenceRequestProviderNew
Function Documentation
Function TRTSERVER_InferenceRequestProviderSetInputData
Function Documentation
Function TRTSERVER_InferenceResponseDelete
Function Documentation
Function TRTSERVER_InferenceResponseHeader
Function Documentation
Function TRTSERVER_InferenceResponseOutputData
Function Documentation
Function TRTSERVER_InferenceResponseStatus
Function Documentation
Function TRTSERVER_MetricsDelete
Function Documentation
Function TRTSERVER_MetricsFormatted
Function Documentation
Function TRTSERVER_ProtobufDelete
Function Documentation
Function TRTSERVER_ProtobufSerialize
Function Documentation
Function TRTSERVER_ResponseAllocatorDelete
Function Documentation
Function TRTSERVER_ResponseAllocatorNew
Function Documentation
Function TRTSERVER_ServerDelete
Function Documentation
Function TRTSERVER_ServerId
Function Documentation
Function TRTSERVER_ServerInferAsync
Function Documentation
Function TRTSERVER_ServerIsLive
Function Documentation
Function TRTSERVER_ServerIsReady
Function Documentation
Function TRTSERVER_ServerLoadModel
Function Documentation
Function TRTSERVER_ServerMetrics
Function Documentation
Function TRTSERVER_ServerModelStatus
Function Documentation
Function TRTSERVER_ServerNew
Function Documentation
Function TRTSERVER_ServerOptionsAddTensorFlowVgpuMemoryLimits
Function Documentation
Function TRTSERVER_ServerOptionsDelete
Function Documentation
Function TRTSERVER_ServerOptionsNew
Function Documentation
Function TRTSERVER_ServerOptionsSetExitOnError
Function Documentation
Function TRTSERVER_ServerOptionsSetExitTimeout
Function Documentation
Function TRTSERVER_ServerOptionsSetGpuMetrics
Function Documentation
Function TRTSERVER_ServerOptionsSetLogError
Function Documentation
Function TRTSERVER_ServerOptionsSetLogInfo
Function Documentation
Function TRTSERVER_ServerOptionsSetLogVerbose
Function Documentation
Function TRTSERVER_ServerOptionsSetLogWarn
Function Documentation
Function TRTSERVER_ServerOptionsSetMetrics
Function Documentation
Function TRTSERVER_ServerOptionsSetModelControlMode
Function Documentation
Function TRTSERVER_ServerOptionsSetModelRepositoryPath
Function Documentation
Function TRTSERVER_ServerOptionsSetProfiling
Function Documentation
Function TRTSERVER_ServerOptionsSetServerId
Function Documentation
Function TRTSERVER_ServerOptionsSetStrictModelConfig
Function Documentation
Function TRTSERVER_ServerOptionsSetStrictReadiness
Function Documentation
Function TRTSERVER_ServerOptionsSetTensorFlowGpuMemoryFraction
Function Documentation
Function TRTSERVER_ServerOptionsSetTensorFlowSoftPlacement
Function Documentation
Function TRTSERVER_ServerPollModelRepository
Function Documentation
Function TRTSERVER_ServerRegisterSharedMemory
Function Documentation
Function TRTSERVER_ServerSharedMemoryAddress
Function Documentation
Function TRTSERVER_ServerStatus
Function Documentation
Function TRTSERVER_ServerStop
Function Documentation
Function TRTSERVER_ServerUnloadModel
Function Documentation
Function TRTSERVER_ServerUnregisterAllSharedMemory
Function Documentation
Function TRTSERVER_ServerUnregisterSharedMemory
Function Documentation
Function TRTSERVER_SharedMemoryBlockCpuNew
Function Documentation
Function TRTSERVER_SharedMemoryBlockDelete
Function Documentation
Defines
¶
Define CUSTOM_NO_GPU_DEVICE
Define Documentation
Define CUSTOM_SERVER_PARAMETER_CNT
Define Documentation
Define TRTIS_CUSTOM_EXPORT
Define Documentation
Define TRTSERVER_EXPORT
Define Documentation
Typedefs
¶
Typedef CustomErrorStringFn_t
Typedef Documentation
Typedef CustomExecuteFn_t
Typedef Documentation
Typedef CustomExecuteV2Fn_t
Typedef Documentation
Typedef CustomFinalizeFn_t
Typedef Documentation
Typedef CustomGetNextInputFn_t
Typedef Documentation
Typedef CustomGetNextInputV2Fn_t
Typedef Documentation
Typedef CustomGetOutputFn_t
Typedef Documentation
Typedef CustomGetOutputV2Fn_t
Typedef Documentation
Typedef CustomInitializeData
Typedef Documentation
Typedef CustomInitializeFn_t
Typedef Documentation
Typedef CustomMemoryType
Typedef Documentation
Typedef CustomPayload
Typedef Documentation
Typedef CustomServerParameter
Typedef Documentation
Typedef CustomVersionFn_t
Typedef Documentation
Typedef TRTSERVER_Error_Code
Typedef Documentation
Typedef TRTSERVER_InferenceCompleteFn_t
Typedef Documentation
Typedef TRTSERVER_Memory_Type
Typedef Documentation
Typedef TRTSERVER_Metric_Format
Typedef Documentation
Typedef TRTSERVER_Model_Control_Mode
Typedef Documentation
Typedef TRTSERVER_ResponseAllocatorAllocFn_t
Typedef Documentation
Typedef TRTSERVER_ResponseAllocatorReleaseFn_t
Typedef Documentation