Program Listing for File tritonserver.h¶
↰ Return to documentation for file (src/core/tritonserver.h
)
// Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved.
//
// Redistribution and use in source and binary forms, with or without
// modification, are permitted provided that the following conditions
// are met:
// * Redistributions of source code must retain the above copyright
// notice, this list of conditions and the following disclaimer.
// * Redistributions in binary form must reproduce the above copyright
// notice, this list of conditions and the following disclaimer in the
// documentation and/or other materials provided with the distribution.
// * Neither the name of NVIDIA CORPORATION nor the names of its
// contributors may be used to endorse or promote products derived
// from this software without specific prior written permission.
//
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ``AS IS'' AND ANY
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
#pragma once
#include <stdbool.h>
#include <stddef.h>
#include <stdint.h>
#ifdef __cplusplus
extern "C" {
#endif
#if defined(_MSC_VER)
#define TRITONSERVER_EXPORT __declspec(dllexport)
#elif defined(__GNUC__)
#define TRITONSERVER_EXPORT __attribute__((__visibility__("default")))
#else
#define TRITONSERVER_EXPORT
#endif
struct TRITONSERVER_Error;
struct TRITONSERVER_InferenceRequest;
struct TRITONSERVER_InferenceResponse;
struct TRITONSERVER_InferenceTrace;
struct TRITONSERVER_Message;
struct TRITONSERVER_Metrics;
struct TRITONSERVER_ResponseAllocator;
struct TRITONSERVER_Server;
struct TRITONSERVER_ServerOptions;
typedef enum TRITONSERVER_datatype_enum {
TRITONSERVER_TYPE_INVALID,
TRITONSERVER_TYPE_BOOL,
TRITONSERVER_TYPE_UINT8,
TRITONSERVER_TYPE_UINT16,
TRITONSERVER_TYPE_UINT32,
TRITONSERVER_TYPE_UINT64,
TRITONSERVER_TYPE_INT8,
TRITONSERVER_TYPE_INT16,
TRITONSERVER_TYPE_INT32,
TRITONSERVER_TYPE_INT64,
TRITONSERVER_TYPE_FP16,
TRITONSERVER_TYPE_FP32,
TRITONSERVER_TYPE_FP64,
TRITONSERVER_TYPE_BYTES
} TRITONSERVER_DataType;
TRITONSERVER_EXPORT const char* TRITONSERVER_DataTypeString(
TRITONSERVER_DataType datatype);
TRITONSERVER_EXPORT TRITONSERVER_DataType
TRITONSERVER_StringToDataType(const char* dtype);
TRITONSERVER_EXPORT uint32_t
TRITONSERVER_DataTypeByteSize(TRITONSERVER_DataType datatype);
typedef enum TRITONSERVER_memorytype_enum {
TRITONSERVER_MEMORY_CPU,
TRITONSERVER_MEMORY_CPU_PINNED,
TRITONSERVER_MEMORY_GPU
} TRITONSERVER_MemoryType;
TRITONSERVER_EXPORT const char* TRITONSERVER_MemoryTypeString(
TRITONSERVER_MemoryType memtype);
typedef enum TRITONSERVER_loglevel_enum {
TRITONSERVER_LOG_INFO,
TRITONSERVER_LOG_WARN,
TRITONSERVER_LOG_ERROR,
TRITONSERVER_LOG_VERBOSE
} TRITONSERVER_LogLevel;
TRITONSERVER_EXPORT bool TRITONSERVER_LogIsEnabled(TRITONSERVER_LogLevel level);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_LogMessage(
TRITONSERVER_LogLevel level, const char* filename, const int line,
const char* msg);
typedef enum TRITONSERVER_errorcode_enum {
TRITONSERVER_ERROR_UNKNOWN,
TRITONSERVER_ERROR_INTERNAL,
TRITONSERVER_ERROR_NOT_FOUND,
TRITONSERVER_ERROR_INVALID_ARG,
TRITONSERVER_ERROR_UNAVAILABLE,
TRITONSERVER_ERROR_UNSUPPORTED,
TRITONSERVER_ERROR_ALREADY_EXISTS
} TRITONSERVER_Error_Code;
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ErrorNew(
TRITONSERVER_Error_Code code, const char* msg);
TRITONSERVER_EXPORT void TRITONSERVER_ErrorDelete(TRITONSERVER_Error* error);
TRITONSERVER_EXPORT TRITONSERVER_Error_Code
TRITONSERVER_ErrorCode(TRITONSERVER_Error* error);
TRITONSERVER_EXPORT const char* TRITONSERVER_ErrorCodeString(
TRITONSERVER_Error* error);
TRITONSERVER_EXPORT const char* TRITONSERVER_ErrorMessage(
TRITONSERVER_Error* error);
typedef TRITONSERVER_Error* (*TRITONSERVER_ResponseAllocatorAllocFn_t)(
TRITONSERVER_ResponseAllocator* allocator, const char* tensor_name,
size_t byte_size, TRITONSERVER_MemoryType memory_type,
int64_t memory_type_id, void* userp, void** buffer, void** buffer_userp,
TRITONSERVER_MemoryType* actual_memory_type,
int64_t* actual_memory_type_id);
typedef TRITONSERVER_Error* (*TRITONSERVER_ResponseAllocatorReleaseFn_t)(
TRITONSERVER_ResponseAllocator* allocator, void* buffer, void* buffer_userp,
size_t byte_size, TRITONSERVER_MemoryType memory_type,
int64_t memory_type_id);
typedef TRITONSERVER_Error* (*TRITONSERVER_ResponseAllocatorStartFn_t)(
TRITONSERVER_ResponseAllocator* allocator, void* userp);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ResponseAllocatorNew(
TRITONSERVER_ResponseAllocator** allocator,
TRITONSERVER_ResponseAllocatorAllocFn_t alloc_fn,
TRITONSERVER_ResponseAllocatorReleaseFn_t release_fn,
TRITONSERVER_ResponseAllocatorStartFn_t start_fn);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ResponseAllocatorDelete(
TRITONSERVER_ResponseAllocator* allocator);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_MessageDelete(
TRITONSERVER_Message* message);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_MessageSerializeToJson(
TRITONSERVER_Message* message, const char** base, size_t* byte_size);
typedef enum tritonserver_metricformat_enum {
TRITONSERVER_METRIC_PROMETHEUS
} TRITONSERVER_MetricFormat;
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_MetricsDelete(
TRITONSERVER_Metrics* metrics);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_MetricsFormatted(
TRITONSERVER_Metrics* metrics, TRITONSERVER_MetricFormat format,
const char** base, size_t* byte_size);
typedef enum tritonserver_tracelevel_enum {
TRITONSERVER_TRACE_LEVEL_DISABLED,
TRITONSERVER_TRACE_LEVEL_MIN,
TRITONSERVER_TRACE_LEVEL_MAX
} TRITONSERVER_InferenceTraceLevel;
TRITONSERVER_EXPORT const char* TRITONSERVER_InferenceTraceLevelString(
TRITONSERVER_InferenceTraceLevel level);
// Trace activities
typedef enum tritonserver_traceactivity_enum {
TRITONSERVER_TRACE_REQUEST_START = 0,
TRITONSERVER_TRACE_QUEUE_START = 1,
TRITONSERVER_TRACE_COMPUTE_START = 2,
TRITONSERVER_TRACE_COMPUTE_INPUT_END = 3,
TRITONSERVER_TRACE_COMPUTE_OUTPUT_START = 4,
TRITONSERVER_TRACE_COMPUTE_END = 5,
TRITONSERVER_TRACE_REQUEST_END = 6
} TRITONSERVER_InferenceTraceActivity;
TRITONSERVER_EXPORT const char* TRITONSERVER_InferenceTraceActivityString(
TRITONSERVER_InferenceTraceActivity activity);
typedef void (*TRITONSERVER_InferenceTraceActivityFn_t)(
TRITONSERVER_InferenceTrace* trace,
TRITONSERVER_InferenceTraceActivity activity, uint64_t timestamp_ns,
void* userp);
typedef void (*TRITONSERVER_InferenceTraceReleaseFn_t)(
TRITONSERVER_InferenceTrace* trace, void* userp);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceTraceNew(
TRITONSERVER_InferenceTrace** trace, TRITONSERVER_InferenceTraceLevel level,
uint64_t parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn,
TRITONSERVER_InferenceTraceReleaseFn_t release_fn, void* trace_userp);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceTraceDelete(
TRITONSERVER_InferenceTrace* trace);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceTraceId(
TRITONSERVER_InferenceTrace* trace, uint64_t* id);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceTraceParentId(
TRITONSERVER_InferenceTrace* trace, uint64_t* parent_id);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceTraceModelName(
TRITONSERVER_InferenceTrace* trace, const char** model_name);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceTraceModelVersion(
TRITONSERVER_InferenceTrace* trace, int64_t* model_version);
typedef enum tritonserver_requestflag_enum {
TRITONSERVER_REQUEST_FLAG_SEQUENCE_START = 1,
TRITONSERVER_REQUEST_FLAG_SEQUENCE_END = 2
} TRITONSERVER_RequestFlag;
typedef enum tritonserver_requestreleaseflag_enum {
TRITONSERVER_REQUEST_RELEASE_ALL = 1
} TRITONSERVER_RequestReleaseFlag;
typedef enum tritonserver_responsecompleteflag_enum {
TRITONSERVER_RESPONSE_COMPLETE_FINAL = 1
} TRITONSERVER_ResponseCompleteFlag;
typedef void (*TRITONSERVER_InferenceRequestReleaseFn_t)(
TRITONSERVER_InferenceRequest* request, const uint32_t flags, void* userp);
typedef void (*TRITONSERVER_InferenceResponseCompleteFn_t)(
TRITONSERVER_InferenceResponse* response, const uint32_t flags,
void* userp);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceRequestNew(
TRITONSERVER_InferenceRequest** inference_request,
TRITONSERVER_Server* server, const char* model_name,
const int64_t model_version);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceRequestDelete(
TRITONSERVER_InferenceRequest* inference_request);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceRequestId(
TRITONSERVER_InferenceRequest* inference_request, const char** id);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceRequestSetId(
TRITONSERVER_InferenceRequest* inference_request, const char* id);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceRequestFlags(
TRITONSERVER_InferenceRequest* inference_request, uint32_t* flags);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceRequestSetFlags(
TRITONSERVER_InferenceRequest* inference_request, uint32_t flags);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestCorrelationId(
TRITONSERVER_InferenceRequest* inference_request, uint64_t* correlation_id);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestSetCorrelationId(
TRITONSERVER_InferenceRequest* inference_request, uint64_t correlation_id);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceRequestPriority(
TRITONSERVER_InferenceRequest* inference_request, uint32_t* priority);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestSetPriority(
TRITONSERVER_InferenceRequest* inference_request, uint32_t priority);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestTimeoutMicroseconds(
TRITONSERVER_InferenceRequest* inference_request, uint64_t* timeout_us);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestSetTimeoutMicroseconds(
TRITONSERVER_InferenceRequest* inference_request, uint64_t timeout_us);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceRequestAddInput(
TRITONSERVER_InferenceRequest* inference_request, const char* name,
const TRITONSERVER_DataType datatype, const int64_t* shape,
uint64_t dim_count);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestRemoveInput(
TRITONSERVER_InferenceRequest* inference_request, const char* name);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestRemoveAllInputs(
TRITONSERVER_InferenceRequest* inference_request);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestAppendInputData(
TRITONSERVER_InferenceRequest* inference_request, const char* name,
const void* base, size_t byte_size, TRITONSERVER_MemoryType memory_type,
int64_t memory_type_id);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestRemoveAllInputData(
TRITONSERVER_InferenceRequest* inference_request, const char* name);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestAddRequestedOutput(
TRITONSERVER_InferenceRequest* inference_request, const char* name);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestRemoveRequestedOutput(
TRITONSERVER_InferenceRequest* inference_request, const char* name);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestRemoveAllRequestedOutputs(
TRITONSERVER_InferenceRequest* inference_request);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestSetReleaseCallback(
TRITONSERVER_InferenceRequest* inference_request,
TRITONSERVER_InferenceRequestReleaseFn_t request_release_fn,
void* request_release_userp);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceRequestSetResponseCallback(
TRITONSERVER_InferenceRequest* inference_request,
TRITONSERVER_ResponseAllocator* response_allocator,
void* response_allocator_userp,
TRITONSERVER_InferenceResponseCompleteFn_t response_fn,
void* response_userp);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceResponseDelete(
TRITONSERVER_InferenceResponse* inference_response);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceResponseError(
TRITONSERVER_InferenceResponse* inference_response);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceResponseModel(
TRITONSERVER_InferenceResponse* inference_response, const char** model_name,
int64_t* model_version);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceResponseId(
TRITONSERVER_InferenceResponse* inference_response,
const char** request_id);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceResponseOutputCount(
TRITONSERVER_InferenceResponse* inference_response, uint32_t* count);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_InferenceResponseOutput(
TRITONSERVER_InferenceResponse* inference_response, const uint32_t index,
const char** name, TRITONSERVER_DataType* datatype, const int64_t** shape,
uint64_t* dim_count, const void** base, size_t* byte_size,
TRITONSERVER_MemoryType* memory_type, int64_t* memory_type_id,
void** userp);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_InferenceResponseOutputClassificationLabel(
TRITONSERVER_InferenceResponse* inference_response, const uint32_t index,
const size_t class_index, const char** label);
typedef enum tritonserver_modelcontrolmode_enum {
TRITONSERVER_MODEL_CONTROL_NONE,
TRITONSERVER_MODEL_CONTROL_POLL,
TRITONSERVER_MODEL_CONTROL_EXPLICIT
} TRITONSERVER_ModelControlMode;
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsNew(
TRITONSERVER_ServerOptions** options);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsDelete(
TRITONSERVER_ServerOptions* options);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsSetServerId(
TRITONSERVER_ServerOptions* options, const char* server_id);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetModelRepositoryPath(
TRITONSERVER_ServerOptions* options, const char* model_repository_path);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetModelControlMode(
TRITONSERVER_ServerOptions* options, TRITONSERVER_ModelControlMode mode);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetStartupModel(
TRITONSERVER_ServerOptions* options, const char* model_name);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetStrictModelConfig(
TRITONSERVER_ServerOptions* options, bool strict);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetPinnedMemoryPoolByteSize(
TRITONSERVER_ServerOptions* options, uint64_t size);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetCudaMemoryPoolByteSize(
TRITONSERVER_ServerOptions* options, int gpu_device, uint64_t size);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetMinSupportedComputeCapability(
TRITONSERVER_ServerOptions* options, double cc);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetExitOnError(
TRITONSERVER_ServerOptions* options, bool exit);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetStrictReadiness(
TRITONSERVER_ServerOptions* options, bool strict);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetExitTimeout(
TRITONSERVER_ServerOptions* options, unsigned int timeout);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsSetLogInfo(
TRITONSERVER_ServerOptions* options, bool log);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsSetLogWarn(
TRITONSERVER_ServerOptions* options, bool log);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsSetLogError(
TRITONSERVER_ServerOptions* options, bool log);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsSetLogVerbose(
TRITONSERVER_ServerOptions* options, int level);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsSetMetrics(
TRITONSERVER_ServerOptions* options, bool metrics);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerOptionsSetGpuMetrics(
TRITONSERVER_ServerOptions* options, bool gpu_metrics);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetTensorFlowSoftPlacement(
TRITONSERVER_ServerOptions* options, bool soft_placement);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsSetTensorFlowGpuMemoryFraction(
TRITONSERVER_ServerOptions* options, float fraction);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerOptionsAddTensorFlowVgpuMemoryLimits(
TRITONSERVER_ServerOptions* options, int gpu_device, int num_vgpus,
uint64_t per_vgpu_memory_mbytes);
typedef enum tritonserver_batchflag_enum {
TRITONSERVER_BATCH_UNKNOWN = 1,
TRITONSERVER_BATCH_FIRST_DIM = 2
} TRITONSERVER_ModelBatchFlag;
typedef enum tritonserver_modelindexflag_enum {
TRITONSERVER_INDEX_FLAG_READY = 1
} TRITONSERVER_ModelIndexFlag;
typedef enum tritonserver_txn_property_flag_enum {
TRITONSERVER_TXN_ONE_TO_ONE = 1,
TRITONSERVER_TXN_DECOUPLED = 2
} TRITONSERVER_ModelTxnPropertyFlag;
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerNew(
TRITONSERVER_Server** server, TRITONSERVER_ServerOptions* options);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerDelete(
TRITONSERVER_Server* server);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerStop(
TRITONSERVER_Server* server);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerPollModelRepository(
TRITONSERVER_Server* server);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerIsLive(
TRITONSERVER_Server* server, bool* live);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerIsReady(
TRITONSERVER_Server* server, bool* ready);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerModelIsReady(
TRITONSERVER_Server* server, const char* model_name,
const int64_t model_version, bool* ready);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerModelBatchProperties(
TRITONSERVER_Server* server, const char* model_name,
const int64_t model_version, uint32_t* flags, void** voidp);
TRITONSERVER_EXPORT TRITONSERVER_Error*
TRITONSERVER_ServerModelTransactionProperties(
TRITONSERVER_Server* server, const char* model_name,
const int64_t model_version, uint32_t* txn_flags, void** voidp);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerMetadata(
TRITONSERVER_Server* server, TRITONSERVER_Message** server_metadata);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerModelMetadata(
TRITONSERVER_Server* server, const char* model_name,
const int64_t model_version, TRITONSERVER_Message** model_metadata);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerModelStatistics(
TRITONSERVER_Server* server, const char* model_name,
const int64_t model_version, TRITONSERVER_Message** model_stats);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerModelConfig(
TRITONSERVER_Server* server, const char* model_name,
const int64_t model_version, const uint32_t config_version,
TRITONSERVER_Message** model_config);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerModelIndex(
TRITONSERVER_Server* server, uint32_t flags,
TRITONSERVER_Message** model_index);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerLoadModel(
TRITONSERVER_Server* server, const char* model_name);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerUnloadModel(
TRITONSERVER_Server* server, const char* model_name);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerMetrics(
TRITONSERVER_Server* server, TRITONSERVER_Metrics** metrics);
TRITONSERVER_EXPORT TRITONSERVER_Error* TRITONSERVER_ServerInferAsync(
TRITONSERVER_Server* server,
TRITONSERVER_InferenceRequest* inference_request,
TRITONSERVER_InferenceTrace* trace);
#ifdef __cplusplus
}
#endif