Inference context accessing Triton Inference Server in C-API mode.
Definition at line 33 of file infer_trtis_context.h.


Public Member Functions | |
| InferTrtISContext () | |
| Constructor, default. More... | |
| ~InferTrtISContext () override | |
| Destructor, default. More... | |
| NvDsInferStatus | deinit () override |
| Synchronize on the CUDA stream and call InferCudaContext::deinit(). More... | |
| SharedCuStream & | mainStream () override |
| Get the main processing CUDA event. More... | |
| SharedSysMem | acquireTensorHostBuf (const std::string &name, size_t bytes) |
| Allocator. More... | |
| SharedCuEvent | acquireTensorHostEvent () |
| Acquire a CUDA event from the events pool. More... | |
| NvDsInferStatus | initialize (const std::string &prototxt, InferLoggingFunc logFunc) final |
| NvDsInferStatus | run (SharedIBatchArray input, InferOutputCb outputCb) final |
Protected Member Functions | |
| NvDsInferStatus | fixateInferenceInfo (const ic::InferenceConfig &config, BaseBackend &backend) override |
| Check the tensor order, media format, and datatype for the input tensor. More... | |
| NvDsInferStatus | createPreprocessor (const ic::PreProcessParams ¶ms, std::vector< UniqPreprocessor > &processors) override |
| Create the surface converter and network preprocessor. More... | |
| NvDsInferStatus | createPostprocessor (const ic::PostProcessParams ¶ms, UniqPostprocessor &processor) override |
| Create the post-processor as per the network output type. More... | |
| NvDsInferStatus | preInference (SharedBatchArray &inputs, const ic::InferenceConfig &config) override |
| Initialize non-image input layers if the custom library has implemented the interface. More... | |
| NvDsInferStatus | extraOutputTensorCheck (SharedBatchArray &outputs, SharedOptions inOptions) override |
| Post inference steps for the custom processor and LSTM controller. More... | |
| void | notifyError (NvDsInferStatus status) override |
| In case of error, notify the waiting threads. More... | |
| void | getNetworkInputInfo (NvDsInferNetworkInfo &networkInfo) override |
| Get the network input layer information. More... | |
| int | tensorPoolSize () const |
| Get the size of the tensor pool. More... | |
| virtual void | backendConsumedInputs (SharedBatchArray inputs) |
| const ic::InferenceConfig & | config () const |
| int | maxBatchSize () const |
| int | uniqueId () const |
| BaseBackend * | backend () |
| const SharedDllHandle & | customLib () const |
| bool | needCopyInputToHost () const |
| void | print (NvDsInferLogLevel l, const char *msg) |
| bool | needPreprocess () const |
| nvdsinferserver::InferTrtISContext::InferTrtISContext | ( | ) |
Constructor, default.
|
override |
Destructor, default.
|
inherited |
Allocator.
Acquire a host buffer for the inference output.
| [in] | name | Name of the output layer. |
| [in] | bytes | Size of the buffer. |
|
inherited |
Acquire a CUDA event from the events pool.
|
inlineprotectedinherited |
Definition at line 101 of file infer_base_context.h.
|
inlineprotectedvirtualinherited |
Definition at line 93 of file infer_base_context.h.
|
inlineprotectedinherited |
Definition at line 98 of file infer_base_context.h.
|
overrideprotectedvirtualinherited |
Create the post-processor as per the network output type.
| [in] | params | The post processing configuration parameters. |
| [out] | processor | The handle to the created post processor. |
Implements nvdsinferserver::InferBaseContext.
|
overrideprotectedvirtualinherited |
Create the surface converter and network preprocessor.
| params | The preprocessor configuration. |
| processors | List of the created preprocessor handles. |
Implements nvdsinferserver::InferBaseContext.
|
inlineprotectedinherited |
Definition at line 102 of file infer_base_context.h.
|
override |
Synchronize on the CUDA stream and call InferCudaContext::deinit().
|
overrideprotectedvirtualinherited |
Post inference steps for the custom processor and LSTM controller.
| [in,out] | outputs | The output batch buffers array. |
| [in] | inOptions | The configuration options for the buffers. |
Reimplemented from nvdsinferserver::InferBaseContext.
|
overrideprotectedvirtualinherited |
Check the tensor order, media format, and datatype for the input tensor.
Initiate the extra processor and lstm controller if configured.
| [in] | config | The inference configuration protobuf message. |
| [in] | backend | The inference backend instance. |
Implements nvdsinferserver::InferBaseContext.
|
inlineoverrideprotectedinherited |
Get the network input layer information.
Definition at line 132 of file infer_cuda_context.h.
References nvdsinferserver::InferCudaContext::m_NetworkImageInfo.
|
finalinherited |
|
inlineoverridevirtual |
Get the main processing CUDA event.
Implements nvdsinferserver::InferBaseContext.
Definition at line 53 of file infer_trtis_context.h.
|
inlineprotectedinherited |
Definition at line 99 of file infer_base_context.h.
|
protectedinherited |
|
protectedinherited |
|
overrideprotectedvirtualinherited |
In case of error, notify the waiting threads.
Implements nvdsinferserver::InferBaseContext.
|
overrideprotectedvirtualinherited |
Initialize non-image input layers if the custom library has implemented the interface.
| [in,out] | inputs | Array of the input batch buffers. |
| [in] | config | The inference configuration settings. |
Reimplemented from nvdsinferserver::InferBaseContext.
|
protectedinherited |
|
finalinherited |
|
protectedinherited |
Get the size of the tensor pool.
|
inlineprotectedinherited |
Definition at line 100 of file infer_base_context.h.
|
protectedinherited |
Array of buffers of the additional inputs.
Definition at line 213 of file infer_cuda_context.h.
|
protectedinherited |
Extra and custom processing pre/post inference.
Definition at line 228 of file infer_cuda_context.h.
|
protectedinherited |
Definition at line 234 of file infer_cuda_context.h.
|
protectedinherited |
Pool of CUDA events for host tensor copy.
Definition at line 221 of file infer_cuda_context.h.
|
protectedinherited |
Map of pools for the output tensors.
Definition at line 217 of file infer_cuda_context.h.
|
protectedinherited |
The input layer datatype.
Definition at line 208 of file infer_cuda_context.h.
|
protectedinherited |
The input layer tensor order.
Definition at line 204 of file infer_cuda_context.h.
|
protectedinherited |
LSTM controller.
Definition at line 224 of file infer_cuda_context.h.
|
protectedinherited |
stream-id based management.
Definition at line 226 of file infer_cuda_context.h.
|
protectedinherited |
The input layer media format.
Definition at line 196 of file infer_cuda_context.h.
|
protectedinherited |
Network input height, width, channels for preprocessing.
Definition at line 192 of file infer_cuda_context.h.
Referenced by nvdsinferserver::InferCudaContext::getNetworkInputInfo().
|
protectedinherited |
The input layer name.
Definition at line 200 of file infer_cuda_context.h.
|
protectedinherited |
Definition at line 233 of file infer_cuda_context.h.
|
protectedinherited |
Preprocessor and post-processor handles.
Definition at line 232 of file infer_cuda_context.h.