Class ManagerInfer
Defined in File infer_manager.hpp
- 
class ManagerInfer
 Manager class for inference.
Public Functions
- 
ManagerInfer()
 Default Constructor.
- 
~ManagerInfer()
 Destructor.
Create inference settings and memory.
- Parameters
 inference_specs – specifications for inference
- Returns
 InferStatus with appropriate code and message
Prepares and launches single/multiple inference.
The provided CUDA stream is used to prepare the input data and will be used to operate on the output data, any execution of CUDA work should be in sync with this stream.
- Parameters
 inference_specs – specifications for inference
cuda_stream – CUDA stream
- Returns
 InferStatus with appropriate code and message
- 
InferStatus run_core_inference(const std::string &model_name, const DataMap &permodel_preprocess_data, const DataMap &permodel_output_data, cudaStream_t cuda_stream)
 Executes Core inference for a particular model and generates inferred data The provided CUDA stream is used to prepare the input data and will be used to operate on the output data, any execution of CUDA work should be in sync with this stream.
- Parameters
 model_name – Input model to do the inference on
permodel_preprocess_data – Input DataMap with model name as key and DataBuffer as value
permodel_output_data – Output DataMap with tensor name as key and DataBuffer as value
cuda_stream – CUDA stream
- Returns
 InferStatus with appropriate code and message
- 
void cleanup()
 Cleans up internal context per model.
- 
DimType get_input_dimensions() const
 Get input dimension per model.
- Returns
 Map with model name as key and dimension as value
- 
DimType get_output_dimensions() const
 Get output dimension per tensor.
- Returns
 Map with tensor name as key and dimension as value
- 
ManagerInfer()