|
NVIDIA DeepStream SDK API Reference
|
7.0 Release
|
Go to the documentation of this file.
13 #ifndef __NVDSINFER_BACKEND_H__
14 #define __NVDSINFER_BACKEND_H__
17 #include <condition_variable>
22 #include <cuda_runtime_api.h>
24 #include <NvCaffeParser.h>
26 #include <NvInferRuntime.h>
43 explicit CudaStream(uint flag = cudaStreamDefault,
int priority = 0);
52 m_Stream = o.m_Stream;
66 explicit CudaEvent(uint flag = cudaEventDefault);
68 operator cudaEvent_t() {
return m_Event; }
69 cudaEvent_t&
ptr() {
return m_Event; }
80 cudaEvent_t m_Event =
nullptr;
151 virtual NvDsInferBatchDims
getBatchDims(
int bindingIndex = 0)
const = 0;
180 virtual const NvDsInferBatchDimsLayerInfo&
getLayerInfo(
int bindingIdx) = 0;
182 virtual int getLayerIdx(
const std::string& bindingName) = 0;
187 int bindingIdx,
const NvDsInferBatchDims& batchDims) = 0;
196 const std::shared_ptr<InferBatchBuffer>& buffer,
CudaStream& stream,
216 std::shared_ptr<TrtEngine> engine);
218 int getLayerIdx(
const std::string& bindingName)
override;
221 const NvDsInferBatchDimsLayerInfo&
getLayerInfo(
int bindingIdx)
override
228 int bindingIdx,
const NvDsInferBatchDims& batchDims)
override;
233 return m_AllLayers[bindingIdx].profileDims[kSELECTOR_MAX];
238 return m_AllLayers[bindingIdx].profileDims[kSELECTOR_MIN];
243 return m_AllLayers[bindingIdx].profileDims[kSELECTOR_OPT];
264 std::shared_ptr<TrtEngine> engine);
270 const std::shared_ptr<InferBatchBuffer>& buffer,
CudaStream& stream,
275 int bindingIdx,
const NvDsInferBatchDims& batchDims)
override;
288 std::shared_ptr<TrtEngine> engine,
int profile = 0);
294 const std::shared_ptr<InferBatchBuffer>& buffer,
CudaStream& stream,
309 std::shared_ptr<TrtEngine> engine);
312 const std::shared_ptr<InferBatchBuffer>& buffer,
CudaStream& stream,
323 std::shared_ptr<TrtEngine> engine,
int profile = 0);
326 const std::shared_ptr<InferBatchBuffer>& buffer,
CudaStream& stream,
330 static std::mutex sExecutionMutex;
343 const std::shared_ptr<TrtEngine>& engine);
Backend context for full dimensions network.
virtual NvDsInferStatus enqueueBuffer(const std::shared_ptr< InferBatchBuffer > &buffer, CudaStream &stream, CudaEvent *consumeEvent)=0
Helper base class for managing Cuda allocated buffers.
CudaDeviceBuffer(size_t size)
virtual NvDsInferBatchDims getBatchDims(int bindingIndex=0) const =0
Helper class written on top of nvinfer1::ICudaEngine.
virtual NvDsInferBatchDims getMaxBatchDims(int bindingIdx)=0
InferBatchBuffer()=default
#define SIMPLE_MOVE_COPY(Cls)
const NvDsInferBatchDimsLayerInfo & getLayerInfo(int bindingIdx) override
Abstract interface for managing the actual inferencing implementation.
virtual NvDsInferBatchDims getOptBatchDims(int bindingIdx)=0
Helper class for managing Cuda Streams.
virtual ~BackendContext()=default
DlaFullDimTrtBackendContext(UniquePtrWDestroy< nvinfer1::IExecutionContext > &&ctx, std::shared_ptr< TrtEngine > engine, int profile=0)
bool canSupportBatchDims(int bindingIdx, const NvDsInferBatchDims &batchDims) override
Backend context for implicit batch dimension network.
Abstract interface to manage a batched buffer for inference.
virtual NvDsInferBatchDims getMinBatchDims(int bindingIdx)=0
NvDsInferDataType
Specifies the data type of a layer.
virtual std::vector< void * > & getDeviceBuffers()=0
virtual NvDsInferStatus initialize()=0
Helper class for managing Cuda Streams.
virtual const NvDsInferBatchDimsLayerInfo & getLayerInfo(int bindingIdx)=0
int getNumBoundLayers() override
Base class for implementations of the BackendContext interface.
FullDimTrtBackendContext(UniquePtrWDestroy< nvinfer1::IExecutionContext > &&ctx, std::shared_ptr< TrtEngine > engine, int profile=0)
virtual NvDsInferDataType getDataType(int bindingIndex=0) const =0
void move_copy(CudaBuffer &&o)
std::vector< NvDsInferBatchDimsLayerInfo > m_AllLayers
int getLayerIdx(const std::string &bindingName) override
CudaEvent(uint flag=cudaEventDefault)
DISABLE_CLASS_COPY(CudaBuffer)
virtual int getNumBoundLayers()=0
TrtBackendContext(UniquePtrWDestroy< nvinfer1::IExecutionContext > &&ctx, std::shared_ptr< TrtEngine > engine)
virtual NvDsInferBatchDims getOptBatchDims(int bindingIdx) override
NvDsInferStatus enqueueBuffer(const std::shared_ptr< InferBatchBuffer > &buffer, CudaStream &stream, CudaEvent *consumeEvent) override
Backend context for implicit batch dimension network inferencing on DLA.
CudaHostBuffer(size_t size)
virtual bool canSupportBatchDims(int bindingIdx, const NvDsInferBatchDims &batchDims)=0
ImplicitTrtBackendContext(UniquePtrWDestroy< nvinfer1::IExecutionContext > &&ctx, std::shared_ptr< TrtEngine > engine)
CudaStream(uint flag=cudaStreamDefault, int priority=0)
Backend context for implicit batch dimension network inferencing on DLA.
static std::mutex sDLAExecutionMutex
std::shared_ptr< TrtEngine > m_CudaEngine
virtual NvDsInferBatchDims getMaxBatchDims(int bindingIdx) override
virtual ~CudaBuffer()=default
DlaImplicitTrtBackendContext(UniquePtrWDestroy< nvinfer1::IExecutionContext > &&ctx, std::shared_ptr< TrtEngine > engine)
virtual NvDsInferBatchDims getMinBatchDims(int bindingIdx) override
NvDsInferStatus enqueueBuffer(const std::shared_ptr< InferBatchBuffer > &buffer, CudaStream &stream, CudaEvent *consumeEvent) override
virtual ~InferBatchBuffer()=default
virtual int getLayerIdx(const std::string &bindingName)=0
UniquePtrWDestroy< nvinfer1::IExecutionContext > m_Context
Helper class for managing Cuda events.
bool canSupportBatchDims(int bindingIdx, const NvDsInferBatchDims &batchDims) override
std::unique_ptr< TrtBackendContext > createBackendContext(const std::shared_ptr< TrtEngine > &engine)
Create an instance of a BackendContext.
NvDsInferStatus
Enum for the status codes returned by NvDsInferContext.