NVIDIA DeepStream SDK API Reference

6.4 Release
infer_simple_runtime.h
Go to the documentation of this file.
1 
12 #ifndef __NVDSINFERSERVER_SIMPLE_RUNTIME_H__
13 #define __NVDSINFERSERVER_SIMPLE_RUNTIME_H__
14 
15 #include "infer_common.h"
16 #include "infer_trtis_backend.h"
17 
18 namespace nvdsinferserver {
19 
21 public:
22  TritonSimpleRuntime(std::string model, int64_t version);
23  ~TritonSimpleRuntime() override;
24 
25  void setOutputs(const std::set<std::string>& names)
26  {
27  m_RequestOutputs = names;
28  }
29 
30  // derived functions
31  NvDsInferStatus initialize() override;
32 
33 protected:
34  NvDsInferStatus specifyInputDims(const InputShapes& shapes) override;
36  InputsConsumed bufConsumed, InferenceDone inferenceDone) override;
37  void requestTritonOutputNames(std::set<std::string>& names) override;
38 
39 private:
40  SharedSysMem allocateSimpleRes(
41  const std::string& tensor, size_t bytes, InferMemType memType, int64_t devId);
42  // Mark releaseCallback as static in case
43  static void releaseSimpleRes(const std::string& tensor, SharedSysMem mem);
44 
45 private:
46  std::set<std::string> m_RequestOutputs;
47 };
48 
49 }
50 
51 #endif
nvdsinferserver
Copyright (c) 2021, NVIDIA CORPORATION.
Definition: infer_custom_process.h:28
nvdsinferserver::TrtISBackend::model
const std::string & model() const
Definition: infer_trtis_backend.h:73
nvdsinferserver::TritonSimpleRuntime::enqueue
NvDsInferStatus enqueue(SharedBatchArray inputs, SharedCuStream stream, InputsConsumed bufConsumed, InferenceDone inferenceDone) override
nvdsinferserver::TritonSimpleRuntime::TritonSimpleRuntime
TritonSimpleRuntime(std::string model, int64_t version)
nvdsinferserver::IBackend::InputShapes
std::vector< InputShapeTuple > InputShapes
Definition: infer_ibackend.h:84
infer_trtis_backend.h
Header file of Triton Inference Server inference backend.
nvdsinferserver::SharedSysMem
std::shared_ptr< SysMem > SharedSysMem
Definition: infer_common.h:88
nvdsinferserver::IBackend::InputsConsumed
std::function< void(SharedBatchArray)> InputsConsumed
Function wrapper called after the input buffer is consumed.
Definition: infer_ibackend.h:70
nvdsinferserver::TrtISBackend
Triton backend processing class.
Definition: infer_trtis_backend.h:39
nvdsinferserver::TritonSimpleRuntime
Definition: infer_simple_runtime.h:20
infer_common.h
Header file of the common declarations for the nvinferserver library.
nvdsinferserver::IBackend::InferenceDone
std::function< void(NvDsInferStatus, SharedBatchArray)> InferenceDone
Function wrapper for post inference processing.
Definition: infer_ibackend.h:66
nvdsinferserver::TrtISBackend::version
int64_t version() const
Definition: infer_trtis_backend.h:74
nvdsinferserver::TritonSimpleRuntime::specifyInputDims
NvDsInferStatus specifyInputDims(const InputShapes &shapes) override
nvdsinferserver::TritonSimpleRuntime::~TritonSimpleRuntime
~TritonSimpleRuntime() override
nvdsinferserver::SharedCuStream
std::shared_ptr< CudaStream > SharedCuStream
Cuda based pointers.
Definition: infer_common.h:84
nvdsinferserver::TritonSimpleRuntime::setOutputs
void setOutputs(const std::set< std::string > &names)
Definition: infer_simple_runtime.h:25
nvdsinferserver::InferMemType
InferMemType
The memory types of inference buffers.
Definition: infer_datatypes.h:56
nvdsinferserver::TritonSimpleRuntime::initialize
NvDsInferStatus initialize() override
nvdsinferserver::SharedBatchArray
std::shared_ptr< BaseBatchArray > SharedBatchArray
Definition: infer_common.h:75
nvdsinferserver::TritonSimpleRuntime::requestTritonOutputNames
void requestTritonOutputNames(std::set< std::string > &names) override
NvDsInferStatus
NvDsInferStatus
Enum for the status codes returned by NvDsInferContext.
Definition: nvdsinfer.h:218