NVIDIA Holoscan SDK v0.4.0
Clara Holoscan v0.4.0

Program Listing for File process_manager.hpp

Return to documentation for file (modules/holoinfer/src/manager/process_manager.hpp)

Copy
Copied!
            

/* * SPDX-FileCopyrightText: Copyright (c) 2022 NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef _HOLOSCAN_PROCESS_MANAGER_H #define _HOLOSCAN_PROCESS_MANAGER_H #include <future> #include <iostream> #include <map> #include <memory> #include <string> #include <vector> #include <holoinfer.hpp> #include <process/data_processor.hpp> namespace holoscan { namespace inference { /* * @brief Manager class for multi ai processing */ class ManagerProcessor { public: /* * @brief Default Constructor */ ManagerProcessor() {} /* * @brief Destructor */ ~ManagerProcessor() {} /* * @brief Initializes the underlying contexts and checks the validity of operations * * @param process_operations Map where tensor name is the key, and operations to perform on * the tensor as vector of strings. Each value in the vector of strings is the supported * operation. * * @returns InferStatus with appropriate code and message */ InferStatus initialize(const MultiMappings& process_operations); /* * @brief Executes post processing operations and generates the result * * @param tensor_oper_map Map where tensor name is the key, and operations to perform on * the tensor as vector of strings. * @param in_out_tensor_map Map where input tensor name is the key, and generated output tensor * name is the value * @param inferred_result_map Map with output tensor name as key, and related DataBuffer as value * @param dimension_map Map with model name as key and related output dimension as value. One-one * mapping supported. * @returns InferStatus with appropriate code and message */ InferStatus process(const MultiMappings& tensor_oper_map, const Mappings& in_out_tensor_map, DataMap& inferred_result_map, const std::map<std::string, std::vector<int>>& dimension_map); /* * @brief Get processed data * * @returns DataMap with tensor name as key and related DataBuffer as value */ DataMap get_processed_data() const; /* * @brief Get processed data dimensions * * @returns DataMap with tensor name as key and related dimension as value */ DimType get_processed_data_dims() const; private: std::unique_ptr<DataProcessor> infer_data_; DataMap processed_data_map_; // TODO: make it generic per operation DimType processed_dims_map_; }; std::unique_ptr<ManagerProcessor> process_manager; } // namespace inference } // namespace holoscan #endif

© Copyright 2022, NVIDIA. Last updated on Jun 28, 2023.