What can I help you with?
NVIDIA Morpheus (25.02.01)

Program Listing for File inference_client_stage.hpp

Return to documentation for file (python/morpheus/morpheus/_lib/include/morpheus/stages/inference_client_stage.hpp)

Copy
Copied!
            

/* * SPDX-FileCopyrightText: Copyright (c) 2024-2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #pragma once #include "morpheus/export.h" // for MORPHEUS_EXPORT #include "morpheus/messages/control.hpp" // for ControlMessage #include "morpheus/types.hpp" // for TensorMap #include <mrc/coroutines/async_generator.hpp> // for AsyncGenerator #include <mrc/coroutines/scheduler.hpp> // for Scheduler #include <mrc/coroutines/task.hpp> // for Task #include <mrc/segment/builder.hpp> // for Builder #include <mrc/segment/object.hpp> // for Object #include <pybind11/pybind11.h> // for object_api::operator() #include <pymrc/asyncio_runnable.hpp> // for AsyncioRunnable #include <cstdint> // for int32_t #include <map> // for map #include <memory> // for shared_ptr, unique_ptr #include <mutex> // for mutex #include <string> // for string #include <vector> // for vector namespace morpheus { struct MORPHEUS_EXPORT TensorModelMapping { std::string model_field_name; std::string tensor_field_name; }; class MORPHEUS_EXPORT IInferenceClientSession { public: virtual ~IInferenceClientSession() = default; virtual std::vector<TensorModelMapping> get_input_mappings(std::vector<TensorModelMapping> input_map_overrides) = 0; virtual std::vector<TensorModelMapping> get_output_mappings( std::vector<TensorModelMapping> output_map_overrides) = 0; virtual mrc::coroutines::Task<TensorMap> infer(TensorMap&& inputs) = 0; }; class MORPHEUS_EXPORT IInferenceClient { public: virtual ~IInferenceClient() = default; virtual std::unique_ptr<IInferenceClientSession> create_session() = 0; }; class MORPHEUS_EXPORT InferenceClientStage : public mrc::pymrc::AsyncioRunnable<std::shared_ptr<ControlMessage>, std::shared_ptr<ControlMessage>> { public: using sink_type_t = std::shared_ptr<ControlMessage>; using source_type_t = std::shared_ptr<ControlMessage>; InferenceClientStage(std::unique_ptr<IInferenceClient>&& client, std::string model_name, bool needs_logits, std::vector<TensorModelMapping> input_mapping, std::vector<TensorModelMapping> output_mapping); mrc::coroutines::AsyncGenerator<std::shared_ptr<ControlMessage>> on_data( std::shared_ptr<ControlMessage>&& data, std::shared_ptr<mrc::coroutines::Scheduler> on) override; private: std::string m_model_name; std::shared_ptr<IInferenceClient> m_client; std::shared_ptr<IInferenceClientSession> m_session; bool m_needs_logits{true}; std::vector<TensorModelMapping> m_input_mapping; std::vector<TensorModelMapping> m_output_mapping; std::mutex m_session_mutex; int32_t m_retry_max = 10; }; /****** InferenceClientStageInferenceProxy******************/ struct MORPHEUS_EXPORT InferenceClientStageInterfaceProxy { static std::shared_ptr<mrc::segment::Object<InferenceClientStage>> init( mrc::segment::Builder& builder, const std::string& name, std::string model_name, std::string server_url, bool needs_logits, bool force_convert_inputs, std::map<std::string, std::string> input_mapping, std::map<std::string, std::string> output_mapping); }; // end of group } // namespace morpheus

© Copyright 2024, NVIDIA. Last updated on Mar 3, 2025.