NVIDIA DeepStream SDK API Reference

6.4 Release
tritonserver.h
Go to the documentation of this file.
1 // Copyright 2020-2022, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
2 //
3 // Redistribution and use in source and binary forms, with or without
4 // modification, are permitted provided that the following conditions
5 // are met:
6 // * Redistributions of source code must retain the above copyright
7 // notice, this list of conditions and the following disclaimer.
8 // * Redistributions in binary form must reproduce the above copyright
9 // notice, this list of conditions and the following disclaimer in the
10 // documentation and/or other materials provided with the distribution.
11 // * Neither the name of NVIDIA CORPORATION nor the names of its
12 // contributors may be used to endorse or promote products derived
13 // from this software without specific prior written permission.
14 //
15 // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ``AS IS'' AND ANY
16 // EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17 // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
18 // PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
19 // CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
20 // EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
21 // PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
22 // PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
23 // OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
25 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26 #pragma once
27 
29 
30 #include <stdbool.h>
31 #include <stddef.h>
32 #include <stdint.h>
33 
34 #ifdef __cplusplus
35 extern "C" {
36 #endif
37 
38 #ifdef _COMPILING_TRITONSERVER
39 #if defined(_MSC_VER)
40 #define TRITONSERVER_DECLSPEC __declspec(dllexport)
41 #elif defined(__GNUC__)
42 #define TRITONSERVER_DECLSPEC __attribute__((__visibility__("default")))
43 #else
44 #define TRITONSERVER_DECLSPEC
45 #endif
46 #else
47 #if defined(_MSC_VER)
48 #define TRITONSERVER_DECLSPEC __declspec(dllimport)
49 #else
50 #define TRITONSERVER_DECLSPEC
51 #endif
52 #endif
53 
54 struct TRITONSERVER_BufferAttributes;
55 struct TRITONSERVER_Error;
56 struct TRITONSERVER_InferenceRequest;
57 struct TRITONSERVER_InferenceResponse;
58 struct TRITONSERVER_InferenceTrace;
59 struct TRITONSERVER_Message;
60 struct TRITONSERVER_Metrics;
61 struct TRITONSERVER_Parameter;
62 struct TRITONSERVER_ResponseAllocator;
63 struct TRITONSERVER_Server;
64 struct TRITONSERVER_ServerOptions;
65 struct TRITONSERVER_Metric;
66 struct TRITONSERVER_MetricFamily;
67 
93 #define TRITONSERVER_API_VERSION_MAJOR 1
94 #define TRITONSERVER_API_VERSION_MINOR 17
95 
108  uint32_t* major, uint32_t* minor);
109 
131 
138  TRITONSERVER_DataType datatype);
139 
147 TRITONSERVER_StringToDataType(const char* dtype);
148 
155 TRITONSERVER_DECLSPEC uint32_t
157 
167 
175  TRITONSERVER_MemoryType memtype);
176 
187 
195  TRITONSERVER_ParameterType paramtype);
196 
207 TRITONSERVER_DECLSPEC TRITONSERVER_Parameter* TRITONSERVER_ParameterNew(
208  const char* name, const TRITONSERVER_ParameterType type, const void* value);
209 
221  const char* name, const void* byte_ptr, const uint64_t size);
222 
227  TRITONSERVER_Parameter* parameter);
228 
239 
248 
259 
272 
278  TRITONSERVER_LogLevel level);
279 
288  TRITONSERVER_LogLevel level, const char* filename, const int line,
289  const char* msg);
290 
301 
312 
320 TRITONSERVER_DECLSPEC TRITONSERVER_Error* TRITONSERVER_ErrorNew(
321  TRITONSERVER_Error_Code code, const char* msg);
322 
326 TRITONSERVER_DECLSPEC void TRITONSERVER_ErrorDelete(TRITONSERVER_Error* error);
327 
333 TRITONSERVER_ErrorCode(TRITONSERVER_Error* error);
334 
343  TRITONSERVER_Error* error);
344 
353  TRITONSERVER_Error* error);
354 
360 
391 typedef TRITONSERVER_Error* (*TRITONSERVER_ResponseAllocatorAllocFn_t)(
392  TRITONSERVER_ResponseAllocator* allocator, const char* tensor_name,
393  size_t byte_size, TRITONSERVER_MemoryType memory_type,
394  int64_t memory_type_id, void* userp, void** buffer, void** buffer_userp,
395  TRITONSERVER_MemoryType* actual_memory_type,
396  int64_t* actual_memory_type_id);
397 
420 typedef TRITONSERVER_Error* (
422  TRITONSERVER_ResponseAllocator* allocator, const char* tensor_name,
423  TRITONSERVER_BufferAttributes* buffer_attributes, void* userp,
424  void* buffer_userp);
425 
450 typedef TRITONSERVER_Error* (*TRITONSERVER_ResponseAllocatorQueryFn_t)(
451  TRITONSERVER_ResponseAllocator* allocator, void* userp,
452  const char* tensor_name, size_t* byte_size,
453  TRITONSERVER_MemoryType* memory_type, int64_t* memory_type_id);
454 
472 typedef TRITONSERVER_Error* (*TRITONSERVER_ResponseAllocatorReleaseFn_t)(
473  TRITONSERVER_ResponseAllocator* allocator, void* buffer, void* buffer_userp,
474  size_t byte_size, TRITONSERVER_MemoryType memory_type,
475  int64_t memory_type_id);
476 
486 typedef TRITONSERVER_Error* (*TRITONSERVER_ResponseAllocatorStartFn_t)(
487  TRITONSERVER_ResponseAllocator* allocator, void* userp);
488 
539  TRITONSERVER_ResponseAllocator** allocator,
543 
555 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
557  TRITONSERVER_ResponseAllocator* allocator,
559 
572 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
574  TRITONSERVER_ResponseAllocator* allocator,
576 
582  TRITONSERVER_ResponseAllocator* allocator);
583 
588 
595 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
597  TRITONSERVER_Message** message, const char* base, size_t byte_size);
598 
604  TRITONSERVER_Message* message);
605 
618  TRITONSERVER_Message* message, const char** base, size_t* byte_size);
619 
624 
629 
635  TRITONSERVER_Metrics* metrics);
636 
658  TRITONSERVER_Metrics* metrics, TRITONSERVER_MetricFormat format,
659  const char** base, size_t* byte_size);
660 
665 
689 
698 
712 
721 
729  TRITONSERVER_InferenceTrace* trace,
730  TRITONSERVER_InferenceTraceActivity activity, uint64_t timestamp_ns,
731  void* userp);
732 
740  TRITONSERVER_InferenceTrace* trace,
741  TRITONSERVER_InferenceTraceActivity activity, const char* name,
742  TRITONSERVER_DataType datatype, const void* base, size_t byte_size,
743  const int64_t* shape, uint64_t dim_count,
744  TRITONSERVER_MemoryType memory_type, int64_t memory_type_id, void* userp);
745 
752  TRITONSERVER_InferenceTrace* trace, void* userp);
753 
778  TRITONSERVER_InferenceTrace** trace, TRITONSERVER_InferenceTraceLevel level,
779  uint64_t parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn,
780  TRITONSERVER_InferenceTraceReleaseFn_t release_fn, void* trace_userp);
781 
808  TRITONSERVER_InferenceTrace** trace, TRITONSERVER_InferenceTraceLevel level,
809  uint64_t parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn,
811  TRITONSERVER_InferenceTraceReleaseFn_t release_fn, void* trace_userp);
812 
818  TRITONSERVER_InferenceTrace* trace);
819 
827  TRITONSERVER_InferenceTrace* trace, uint64_t* id);
828 
837  TRITONSERVER_InferenceTrace* trace, uint64_t* parent_id);
838 
848  TRITONSERVER_InferenceTrace* trace, const char** model_name);
849 
856 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
858  TRITONSERVER_InferenceTrace* trace, int64_t* model_version);
859 
868 
874 
880 
886 
910  TRITONSERVER_InferenceRequest* request, const uint32_t flags, void* userp);
911 
930  TRITONSERVER_InferenceResponse* response, const uint32_t flags,
931  void* userp);
932 
943  TRITONSERVER_InferenceRequest** inference_request,
944  TRITONSERVER_Server* server, const char* model_name,
945  const int64_t model_version);
946 
952  TRITONSERVER_InferenceRequest* inference_request);
953 
962  TRITONSERVER_InferenceRequest* inference_request, const char** id);
963 
970  TRITONSERVER_InferenceRequest* inference_request, const char* id);
971 
980  TRITONSERVER_InferenceRequest* inference_request, uint32_t* flags);
981 
990  TRITONSERVER_InferenceRequest* inference_request, uint32_t flags);
991 
1003 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1005  TRITONSERVER_InferenceRequest* inference_request, uint64_t* correlation_id);
1006 
1018 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1020  TRITONSERVER_InferenceRequest* inference_request,
1021  const char** correlation_id);
1022 
1032 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1034  TRITONSERVER_InferenceRequest* inference_request, uint64_t correlation_id);
1035 
1045 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1047  TRITONSERVER_InferenceRequest* inference_request,
1048  const char* correlation_id);
1049 
1058  TRITONSERVER_InferenceRequest* inference_request, uint32_t* priority);
1059 
1067 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1069  TRITONSERVER_InferenceRequest* inference_request, uint32_t priority);
1070 
1077 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1079  TRITONSERVER_InferenceRequest* inference_request, uint64_t* timeout_us);
1080 
1087 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1089  TRITONSERVER_InferenceRequest* inference_request, uint64_t timeout_us);
1090 
1102  TRITONSERVER_InferenceRequest* inference_request, const char* name,
1103  const TRITONSERVER_DataType datatype, const int64_t* shape,
1104  uint64_t dim_count);
1105 
1116 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1118  TRITONSERVER_InferenceRequest* inference_request, const char* name);
1119 
1125 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1127  TRITONSERVER_InferenceRequest* inference_request, const char* name);
1128 
1133 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1135  TRITONSERVER_InferenceRequest* inference_request);
1136 
1151 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1153  TRITONSERVER_InferenceRequest* inference_request, const char* name,
1154  const void* base, size_t byte_size, TRITONSERVER_MemoryType memory_type,
1155  int64_t memory_type_id);
1156 
1176 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1178  TRITONSERVER_InferenceRequest* inference_request, const char* name,
1179  const void* base, size_t byte_size, TRITONSERVER_MemoryType memory_type,
1180  int64_t memory_type_id, const char* host_policy_name);
1181 
1194 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1196  TRITONSERVER_InferenceRequest* inference_request, const char* name,
1197  const void* base, TRITONSERVER_BufferAttributes* buffer_attributes);
1198 
1205 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1207  TRITONSERVER_InferenceRequest* inference_request, const char* name);
1208 
1214 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1216  TRITONSERVER_InferenceRequest* inference_request, const char* name);
1217 
1223 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1225  TRITONSERVER_InferenceRequest* inference_request, const char* name);
1226 
1231 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1233  TRITONSERVER_InferenceRequest* inference_request);
1234 
1245 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1247  TRITONSERVER_InferenceRequest* inference_request,
1248  TRITONSERVER_InferenceRequestReleaseFn_t request_release_fn,
1249  void* request_release_userp);
1250 
1267 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1269  TRITONSERVER_InferenceRequest* inference_request,
1270  TRITONSERVER_ResponseAllocator* response_allocator,
1271  void* response_allocator_userp,
1273  void* response_userp);
1274 
1281 
1287  TRITONSERVER_InferenceResponse* inference_response);
1288 
1298  TRITONSERVER_InferenceResponse* inference_response);
1299 
1311  TRITONSERVER_InferenceResponse* inference_response, const char** model_name,
1312  int64_t* model_version);
1313 
1324  TRITONSERVER_InferenceResponse* inference_response,
1325  const char** request_id);
1326 
1332 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1334  TRITONSERVER_InferenceResponse* inference_response, uint32_t* count);
1335 
1367 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1369  TRITONSERVER_InferenceResponse* inference_response, const uint32_t index,
1370  const char** name, TRITONSERVER_ParameterType* type, const void** vvalue);
1371 
1377 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1379  TRITONSERVER_InferenceResponse* inference_response, uint32_t* count);
1380 
1404  TRITONSERVER_InferenceResponse* inference_response, const uint32_t index,
1405  const char** name, TRITONSERVER_DataType* datatype, const int64_t** shape,
1406  uint64_t* dim_count, const void** base, size_t* byte_size,
1407  TRITONSERVER_MemoryType* memory_type, int64_t* memory_type_id,
1408  void** userp);
1409 
1423 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1425  TRITONSERVER_InferenceResponse* inference_response, const uint32_t index,
1426  const size_t class_index, const char** label);
1427 
1432 
1440  TRITONSERVER_BufferAttributes** buffer_attributes);
1441 
1447  TRITONSERVER_BufferAttributes* buffer_attributes);
1448 
1455 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1457  TRITONSERVER_BufferAttributes* buffer_attributes, int64_t memory_type_id);
1458 
1464 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1466  TRITONSERVER_BufferAttributes* buffer_attributes,
1467  TRITONSERVER_MemoryType memory_type);
1468 
1475 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1477  TRITONSERVER_BufferAttributes* buffer_attributes, void* cuda_ipc_handle);
1478 
1484 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1486  TRITONSERVER_BufferAttributes* buffer_attributes, size_t byte_size);
1487 
1494 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1496  TRITONSERVER_BufferAttributes* buffer_attributes, int64_t* memory_type_id);
1497 
1504 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1506  TRITONSERVER_BufferAttributes* buffer_attributes,
1507  TRITONSERVER_MemoryType* memory_type);
1508 
1516 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1518  TRITONSERVER_BufferAttributes* buffer_attributes, void** cuda_ipc_handle);
1519 
1527  TRITONSERVER_BufferAttributes* buffer_attributes, size_t* byte_size);
1528 
1529 
1534 
1541 
1547 
1555  TRITONSERVER_ServerOptions** options);
1556 
1562  TRITONSERVER_ServerOptions* options);
1563 
1571  TRITONSERVER_ServerOptions* options, const char* server_id);
1572 
1582 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1584  TRITONSERVER_ServerOptions* options, const char* model_repository_path);
1585 
1606 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1608  TRITONSERVER_ServerOptions* options, TRITONSERVER_ModelControlMode mode);
1609 
1619 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1621  TRITONSERVER_ServerOptions* options, const char* model_name);
1622 
1630 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1632  TRITONSERVER_ServerOptions* options, bool strict);
1633 
1648 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1650  TRITONSERVER_ServerOptions* options, TRITONSERVER_RateLimitMode mode);
1651 
1665 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1667  TRITONSERVER_ServerOptions* options, const char* resource_name,
1668  const size_t resource_count, const int device);
1669 
1678 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1680  TRITONSERVER_ServerOptions* options, uint64_t size);
1681 
1691 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1693  TRITONSERVER_ServerOptions* options, int gpu_device, uint64_t size);
1694 
1702 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1704  TRITONSERVER_ServerOptions* options, uint64_t size);
1705 
1712 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1714  TRITONSERVER_ServerOptions* options, double cc);
1715 
1722 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1724  TRITONSERVER_ServerOptions* options, bool exit);
1725 
1732 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1734  TRITONSERVER_ServerOptions* options, bool strict);
1735 
1742 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1744  TRITONSERVER_ServerOptions* options, unsigned int timeout);
1745 
1751 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1753  TRITONSERVER_ServerOptions* options, unsigned int thread_count);
1754 
1760 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1762  TRITONSERVER_ServerOptions* options, unsigned int thread_count);
1763 
1772  TRITONSERVER_ServerOptions* options, const char* file);
1773 
1780  TRITONSERVER_ServerOptions* options, bool log);
1781 
1788  TRITONSERVER_ServerOptions* options, bool log);
1789 
1796  TRITONSERVER_ServerOptions* options, bool log);
1797 
1803 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1805  TRITONSERVER_ServerOptions* options, const TRITONSERVER_LogFormat format);
1806 
1812 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1814  TRITONSERVER_ServerOptions* options, int level);
1815 
1822  TRITONSERVER_ServerOptions* options, bool metrics);
1823 
1831 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1833  TRITONSERVER_ServerOptions* options, bool gpu_metrics);
1834 
1842 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1844  TRITONSERVER_ServerOptions* options, bool cpu_metrics);
1845 
1853 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1855  TRITONSERVER_ServerOptions* options, uint64_t metrics_interval_ms);
1856 
1866 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1868  TRITONSERVER_ServerOptions* options, const char* backend_dir);
1869 
1878 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1880  TRITONSERVER_ServerOptions* options, const char* repoagent_dir);
1881 
1894 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1896  TRITONSERVER_ServerOptions* options,
1897  const TRITONSERVER_InstanceGroupKind kind, const int device_id,
1898  const double fraction);
1899 
1908 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1910  TRITONSERVER_ServerOptions* options, const char* backend_name,
1911  const char* setting, const char* value);
1912 
1920 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1922  TRITONSERVER_ServerOptions* options, const char* policy_name,
1923  const char* setting, const char* value);
1924 
1929 
1935 
1940 
1947 
1955 TRITONSERVER_DECLSPEC TRITONSERVER_Error* TRITONSERVER_ServerNew(
1956  TRITONSERVER_Server** server, TRITONSERVER_ServerOptions* options);
1957 
1964  TRITONSERVER_Server* server);
1965 
1972  TRITONSERVER_Server* server);
1973 
1982 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1984  TRITONSERVER_Server* server, const char* repository_path,
1985  const TRITONSERVER_Parameter** name_mapping, const uint32_t mapping_count);
1986 
1992 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
1994  TRITONSERVER_Server* server, const char* repository_path);
1995 
2001 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
2002 TRITONSERVER_ServerPollModelRepository(TRITONSERVER_Server* server);
2003 
2010  TRITONSERVER_Server* server, bool* live);
2011 
2018  TRITONSERVER_Server* server, bool* ready);
2019 
2030  TRITONSERVER_Server* server, const char* model_name,
2031  const int64_t model_version, bool* ready);
2032 
2057 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
2059  TRITONSERVER_Server* server, const char* model_name,
2060  const int64_t model_version, uint32_t* flags, void** voidp);
2061 
2079 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
2081  TRITONSERVER_Server* server, const char* model_name,
2082  const int64_t model_version, uint32_t* txn_flags, void** voidp);
2083 
2092  TRITONSERVER_Server* server, TRITONSERVER_Message** server_metadata);
2093 
2106  TRITONSERVER_Server* server, const char* model_name,
2107  const int64_t model_version, TRITONSERVER_Message** model_metadata);
2108 
2122  TRITONSERVER_Server* server, const char* model_name,
2123  const int64_t model_version, TRITONSERVER_Message** model_stats);
2124 
2140  TRITONSERVER_Server* server, const char* model_name,
2141  const int64_t model_version, const uint32_t config_version,
2142  TRITONSERVER_Message** model_config);
2143 
2160  TRITONSERVER_Server* server, uint32_t flags,
2161  TRITONSERVER_Message** model_index);
2162 
2172  TRITONSERVER_Server* server, const char* model_name);
2173 
2188 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
2190  TRITONSERVER_Server* server, const char* model_name,
2191  const TRITONSERVER_Parameter** parameters, const uint64_t parameter_count);
2192 
2203  TRITONSERVER_Server* server, const char* model_name);
2204 
2216 TRITONSERVER_DECLSPEC TRITONSERVER_Error*
2218  TRITONSERVER_Server* server, const char* model_name);
2219 
2228  TRITONSERVER_Server* server, TRITONSERVER_Metrics** metrics);
2229 
2252  TRITONSERVER_Server* server,
2253  TRITONSERVER_InferenceRequest* inference_request,
2254  TRITONSERVER_InferenceTrace* trace);
2255 
2264 
2277  TRITONSERVER_MetricFamily** family, const TRITONSERVER_MetricKind kind,
2278  const char* name, const char* description);
2279 
2288  TRITONSERVER_MetricFamily* family);
2289 
2302 TRITONSERVER_DECLSPEC TRITONSERVER_Error* TRITONSERVER_MetricNew(
2303  TRITONSERVER_Metric** metric, TRITONSERVER_MetricFamily* family,
2304  const TRITONSERVER_Parameter** labels, const uint64_t label_count);
2305 
2314  TRITONSERVER_Metric* metric);
2315 
2325  TRITONSERVER_Metric* metric, double* value);
2326 
2338  TRITONSERVER_Metric* metric, double value);
2339 
2347 TRITONSERVER_DECLSPEC TRITONSERVER_Error* TRITONSERVER_MetricSet(
2348  TRITONSERVER_Metric* metric, double value);
2349 
2356  TRITONSERVER_Metric* metric, TRITONSERVER_MetricKind* kind);
2357 
2358 #ifdef __cplusplus
2359 }
2360 #endif
TRITONSERVER_TYPE_INT64
@ TRITONSERVER_TYPE_INT64
Definition: tritonserver.h:124
tritonserver_traceactivity_enum
tritonserver_traceactivity_enum
Trace activities.
Definition: tritonserver.h:700
TRITONSERVER_ServerOptionsSetMinSupportedComputeCapability
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetMinSupportedComputeCapability(TRITONSERVER_ServerOptions *options, double cc)
Set the minimum support CUDA compute capability in a server options.
TRITONSERVER_ServerOptionsSetRepoAgentDirectory
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetRepoAgentDirectory(TRITONSERVER_ServerOptions *options, const char *repoagent_dir)
Set the directory containing repository agent shared libraries.
TRITONSERVER_InferenceTraceLevelString
const TRITONSERVER_DECLSPEC char * TRITONSERVER_InferenceTraceLevelString(TRITONSERVER_InferenceTraceLevel level)
Get the string representation of a trace level.
TRITONSERVER_TYPE_BF16
@ TRITONSERVER_TYPE_BF16
Definition: tritonserver.h:129
TRITONSERVER_InferenceTraceTensorActivityFn_t
void(* TRITONSERVER_InferenceTraceTensorActivityFn_t)(TRITONSERVER_InferenceTrace *trace, TRITONSERVER_InferenceTraceActivity activity, const char *name, TRITONSERVER_DataType datatype, const void *base, size_t byte_size, const int64_t *shape, uint64_t dim_count, TRITONSERVER_MemoryType memory_type, int64_t memory_type_id, void *userp)
Type for trace tensor activity callback function.
Definition: tritonserver.h:739
TRITONSERVER_ServerOptionsSetExitOnError
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetExitOnError(TRITONSERVER_ServerOptions *options, bool exit)
Enable or disable exit-on-error in a server options.
TRITONSERVER_InferenceRequestId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestId(TRITONSERVER_InferenceRequest *inference_request, const char **id)
Get the ID for a request.
TRITONSERVER_InferenceTraceLevel
enum tritonserver_tracelevel_enum TRITONSERVER_InferenceTraceLevel
TRITONSERVER_InferenceTrace.
TRITONSERVER_TRACE_LEVEL_DISABLED
@ TRITONSERVER_TRACE_LEVEL_DISABLED
Tracing disabled. No trace activities are reported.
Definition: tritonserver.h:679
TRITONSERVER_logformat_enum
TRITONSERVER_logformat_enum
Format of logging.
Definition: tritonserver.h:268
tritonserver_responsecompleteflag_enum
tritonserver_responsecompleteflag_enum
Inference response complete flags.
Definition: tritonserver.h:883
TRITONSERVER_ResponseAllocatorSetQueryFunction
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ResponseAllocatorSetQueryFunction(TRITONSERVER_ResponseAllocator *allocator, TRITONSERVER_ResponseAllocatorQueryFn_t query_fn)
Set the query function to a response allocator object.
TRITONSERVER_LogMessage
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_LogMessage(TRITONSERVER_LogLevel level, const char *filename, const int line, const char *msg)
Log a message at a given log level if that level is enabled.
TRITONSERVER_InferenceResponseId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseId(TRITONSERVER_InferenceResponse *inference_response, const char **request_id)
Get the ID of the request corresponding to a response.
TRITONSERVER_MEMORY_CPU
@ TRITONSERVER_MEMORY_CPU
Definition: tritonserver.h:163
TRITONSERVER_PARAMETER_BYTES
@ TRITONSERVER_PARAMETER_BYTES
Definition: tritonserver.h:185
TRITONSERVER_LOG_ISO8601
@ TRITONSERVER_LOG_ISO8601
Definition: tritonserver.h:270
TRITONSERVER_ServerOptionsSetMetrics
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetMetrics(TRITONSERVER_ServerOptions *options, bool metrics)
Enable or disable metrics collection in a server options.
TRITONSERVER_ServerModelIndex
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerModelIndex(TRITONSERVER_Server *server, uint32_t flags, TRITONSERVER_Message **model_index)
Get the index of all unique models in the model repositories as a TRITONSERVER_Message object.
TRITONSERVER_InferenceResponseError
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseError(TRITONSERVER_InferenceResponse *inference_response)
Return the error status of an inference response.
TRITONSERVER_MetricsFormatted
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricsFormatted(TRITONSERVER_Metrics *metrics, TRITONSERVER_MetricFormat format, const char **base, size_t *byte_size)
Get a buffer containing the metrics in the specified format.
TRITONSERVER_LOG_DEFAULT
@ TRITONSERVER_LOG_DEFAULT
Definition: tritonserver.h:269
TRITONSERVER_TYPE_UINT16
@ TRITONSERVER_TYPE_UINT16
Definition: tritonserver.h:118
TRITONSERVER_InferenceRequestAddInput
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestAddInput(TRITONSERVER_InferenceRequest *inference_request, const char *name, const TRITONSERVER_DataType datatype, const int64_t *shape, uint64_t dim_count)
Add an input to a request.
TRITONSERVER_InferenceTraceReleaseFn_t
void(* TRITONSERVER_InferenceTraceReleaseFn_t)(TRITONSERVER_InferenceTrace *trace, void *userp)
Type for trace release callback function.
Definition: tritonserver.h:751
TRITONSERVER_InferenceTraceDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceTraceDelete(TRITONSERVER_InferenceTrace *trace)
Delete a trace object.
TRITONSERVER_InferenceRequestSetFlags
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestSetFlags(TRITONSERVER_InferenceRequest *inference_request, uint32_t flags)
Set the flag(s) associated with a request.
TRITONSERVER_ServerNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerNew(TRITONSERVER_Server **server, TRITONSERVER_ServerOptions *options)
Create a new server object.
TRITONSERVER_TYPE_INT16
@ TRITONSERVER_TYPE_INT16
Definition: tritonserver.h:122
TRITONSERVER_TRACE_QUEUE_START
@ TRITONSERVER_TRACE_QUEUE_START
Definition: tritonserver.h:702
TRITONSERVER_TYPE_UINT64
@ TRITONSERVER_TYPE_UINT64
Definition: tritonserver.h:120
TRITONSERVER_ServerModelBatchProperties
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerModelBatchProperties(TRITONSERVER_Server *server, const char *model_name, const int64_t model_version, uint32_t *flags, void **voidp)
Get the batch properties of the model.
TRITONSERVER_ErrorDelete
TRITONSERVER_DECLSPEC void TRITONSERVER_ErrorDelete(TRITONSERVER_Error *error)
Delete an error object.
TRITONSERVER_ErrorMessage
const TRITONSERVER_DECLSPEC char * TRITONSERVER_ErrorMessage(TRITONSERVER_Error *error)
Get the error message.
TRITONSERVER_metrickind_enum
TRITONSERVER_metrickind_enum
TRITONSERVER_MetricKind.
Definition: tritonserver.h:2260
TRITONSERVER_ServerUnloadModelAndDependents
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerUnloadModelAndDependents(TRITONSERVER_Server *server, const char *model_name)
Unload the requested model, and also unload any dependent model that was loaded along with the reques...
TRITONSERVER_ModelControlMode
enum tritonserver_modelcontrolmode_enum TRITONSERVER_ModelControlMode
TRITONSERVER_ServerOptions.
TRITONSERVER_RATE_LIMIT_OFF
@ TRITONSERVER_RATE_LIMIT_OFF
Definition: tritonserver.h:1544
TRITONSERVER_ERROR_NOT_FOUND
@ TRITONSERVER_ERROR_NOT_FOUND
Definition: tritonserver.h:306
TRITONSERVER_MessageSerializeToJson
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MessageSerializeToJson(TRITONSERVER_Message *message, const char **base, size_t *byte_size)
Get the base and size of the buffer containing the serialized message in JSON format.
TRITONSERVER_InferenceRequestRemoveRequestedOutput
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestRemoveRequestedOutput(TRITONSERVER_InferenceRequest *inference_request, const char *name)
Remove an output request from an inference request.
TRITONSERVER_TRACE_COMPUTE_END
@ TRITONSERVER_TRACE_COMPUTE_END
Definition: tritonserver.h:706
TRITONSERVER_InferenceTraceId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceTraceId(TRITONSERVER_InferenceTrace *trace, uint64_t *id)
Get the id associated with a trace.
TRITONSERVER_ServerOptionsSetServerId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetServerId(TRITONSERVER_ServerOptions *options, const char *server_id)
Set the textual ID for the server in a server options.
TRITONSERVER_BufferAttributesSetCudaIpcHandle
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesSetCudaIpcHandle(TRITONSERVER_BufferAttributes *buffer_attributes, void *cuda_ipc_handle)
Set the CudaIpcHandle field of the buffer attributes.
TRITONSERVER_LOG_VERBOSE
@ TRITONSERVER_LOG_VERBOSE
Definition: tritonserver.h:257
TRITONSERVER_MemoryType
enum TRITONSERVER_memorytype_enum TRITONSERVER_MemoryType
TRITONSERVER_MemoryType.
TRITONSERVER_TRACE_TENSOR_QUEUE_INPUT
@ TRITONSERVER_TRACE_TENSOR_QUEUE_INPUT
Definition: tritonserver.h:708
TRITONSERVER_DataTypeByteSize
TRITONSERVER_DECLSPEC uint32_t TRITONSERVER_DataTypeByteSize(TRITONSERVER_DataType datatype)
Get the size of a Triton datatype in bytes.
TRITONSERVER_BATCH_UNKNOWN
@ TRITONSERVER_BATCH_UNKNOWN
Definition: tritonserver.h:1932
TRITONSERVER_ServerOptionsSetLogVerbose
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetLogVerbose(TRITONSERVER_ServerOptions *options, int level)
Set verbose logging level.
TRITONSERVER_METRIC_PROMETHEUS
@ TRITONSERVER_METRIC_PROMETHEUS
Definition: tritonserver.h:627
TRITONSERVER_ServerOptionsSetStartupModel
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetStartupModel(TRITONSERVER_ServerOptions *options, const char *model_name)
Set the model to be loaded at startup in a server options.
TRITONSERVER_ServerOptionsSetLogInfo
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetLogInfo(TRITONSERVER_ServerOptions *options, bool log)
Enable or disable info level logging.
TRITONSERVER_REQUEST_FLAG_SEQUENCE_END
@ TRITONSERVER_REQUEST_FLAG_SEQUENCE_END
Definition: tritonserver.h:872
TRITONSERVER_InferenceRequestReleaseFn_t
void(* TRITONSERVER_InferenceRequestReleaseFn_t)(TRITONSERVER_InferenceRequest *request, const uint32_t flags, void *userp)
Type for inference request release callback function.
Definition: tritonserver.h:909
TRITONSERVER_InferenceRequestSetTimeoutMicroseconds
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestSetTimeoutMicroseconds(TRITONSERVER_InferenceRequest *inference_request, uint64_t timeout_us)
Set the timeout for a request, in microseconds.
TRITONSERVER_ResponseAllocatorAllocFn_t
TRITONSERVER_Error *(* TRITONSERVER_ResponseAllocatorAllocFn_t)(TRITONSERVER_ResponseAllocator *allocator, const char *tensor_name, size_t byte_size, TRITONSERVER_MemoryType memory_type, int64_t memory_type_id, void *userp, void **buffer, void **buffer_userp, TRITONSERVER_MemoryType *actual_memory_type, int64_t *actual_memory_type_id)
TRITONSERVER_ResponseAllocator.
Definition: tritonserver.h:391
TRITONSERVER_ErrorCodeString
const TRITONSERVER_DECLSPEC char * TRITONSERVER_ErrorCodeString(TRITONSERVER_Error *error)
Get the string representation of an error code.
TRITONSERVER_MODEL_CONTROL_POLL
@ TRITONSERVER_MODEL_CONTROL_POLL
Definition: tritonserver.h:1538
TRITONSERVER_TRACE_COMPUTE_INPUT_END
@ TRITONSERVER_TRACE_COMPUTE_INPUT_END
Definition: tritonserver.h:704
TRITONSERVER_TXN_DECOUPLED
@ TRITONSERVER_TXN_DECOUPLED
Definition: tritonserver.h:1945
TRITONSERVER_ERROR_UNKNOWN
@ TRITONSERVER_ERROR_UNKNOWN
Definition: tritonserver.h:304
TRITONSERVER_InferenceRequestPriority
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestPriority(TRITONSERVER_InferenceRequest *inference_request, uint32_t *priority)
Get the priority for a request.
TRITONSERVER_InferenceTraceActivityString
const TRITONSERVER_DECLSPEC char * TRITONSERVER_InferenceTraceActivityString(TRITONSERVER_InferenceTraceActivity activity)
Get the string representation of a trace activity.
TRITONSERVER_MetricKind
enum TRITONSERVER_metrickind_enum TRITONSERVER_MetricKind
TRITONSERVER_MetricKind.
TRITONSERVER_TRACE_LEVEL_TIMESTAMPS
@ TRITONSERVER_TRACE_LEVEL_TIMESTAMPS
Record timestamps for the inference request.
Definition: tritonserver.h:685
TRITONSERVER_MessageDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MessageDelete(TRITONSERVER_Message *message)
Delete a message object.
TRITONSERVER_PARAMETER_INT
@ TRITONSERVER_PARAMETER_INT
Definition: tritonserver.h:183
TRITONSERVER_ServerOptionsSetModelRepositoryPath
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetModelRepositoryPath(TRITONSERVER_ServerOptions *options, const char *model_repository_path)
Set the model repository path in a server options.
TRITONSERVER_MemoryTypeString
const TRITONSERVER_DECLSPEC char * TRITONSERVER_MemoryTypeString(TRITONSERVER_MemoryType memtype)
Get the string representation of a memory type.
TRITONSERVER_ERROR_ALREADY_EXISTS
@ TRITONSERVER_ERROR_ALREADY_EXISTS
Definition: tritonserver.h:310
tritonserver_modelindexflag_enum
tritonserver_modelindexflag_enum
Model index flags. The enum values must be power-of-2 values.
Definition: tritonserver.h:1937
TRITONSERVER_ErrorCode
TRITONSERVER_DECLSPEC TRITONSERVER_Error_Code TRITONSERVER_ErrorCode(TRITONSERVER_Error *error)
Get the error code.
TRITONSERVER_ServerOptionsSetPinnedMemoryPoolByteSize
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetPinnedMemoryPoolByteSize(TRITONSERVER_ServerOptions *options, uint64_t size)
Set the total pinned memory byte size that the server can allocate in a server options.
TRITONSERVER_REQUEST_FLAG_SEQUENCE_START
@ TRITONSERVER_REQUEST_FLAG_SEQUENCE_START
Definition: tritonserver.h:871
TRITONSERVER_DECLSPEC
#define TRITONSERVER_DECLSPEC
Definition: tritonserver.h:50
TRITONSERVER_ResponseAllocatorSetBufferAttributesFunction
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ResponseAllocatorSetBufferAttributesFunction(TRITONSERVER_ResponseAllocator *allocator, TRITONSERVER_ResponseAllocatorBufferAttributesFn_t buffer_attributes_fn)
Set the buffer attributes function for a response allocator object.
TRITONSERVER_ServerOptionsSetBackendConfig
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetBackendConfig(TRITONSERVER_ServerOptions *options, const char *backend_name, const char *setting, const char *value)
Set a configuration setting for a named backend in a server options.
TRITONSERVER_InferenceRequestSetCorrelationIdString
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestSetCorrelationIdString(TRITONSERVER_InferenceRequest *inference_request, const char *correlation_id)
Set the correlation ID of the inference request to be a string.
TRITONSERVER_TRACE_LEVEL_TENSORS
@ TRITONSERVER_TRACE_LEVEL_TENSORS
Record input and output tensor values for the inference request.
Definition: tritonserver.h:687
TRITONSERVER_RequestReleaseFlag
enum tritonserver_requestreleaseflag_enum TRITONSERVER_RequestReleaseFlag
Inference request release flags.
TRITONSERVER_MetricsDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricsDelete(TRITONSERVER_Metrics *metrics)
Delete a metrics object.
TRITONSERVER_MetricValue
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricValue(TRITONSERVER_Metric *metric, double *value)
Get the current value of a metric object.
TRITONSERVER_METRIC_KIND_GAUGE
@ TRITONSERVER_METRIC_KIND_GAUGE
Definition: tritonserver.h:2262
TRITONSERVER_LOG_INFO
@ TRITONSERVER_LOG_INFO
Definition: tritonserver.h:254
TRITONSERVER_TYPE_INT8
@ TRITONSERVER_TYPE_INT8
Definition: tritonserver.h:121
TRITONSERVER_InferenceResponseCompleteFn_t
void(* TRITONSERVER_InferenceResponseCompleteFn_t)(TRITONSERVER_InferenceResponse *response, const uint32_t flags, void *userp)
Type for callback function indicating that an inference response has completed.
Definition: tritonserver.h:929
TRITONSERVER_ResponseAllocatorQueryFn_t
TRITONSERVER_Error *(* TRITONSERVER_ResponseAllocatorQueryFn_t)(TRITONSERVER_ResponseAllocator *allocator, void *userp, const char *tensor_name, size_t *byte_size, TRITONSERVER_MemoryType *memory_type, int64_t *memory_type_id)
Type for function that is called to query the allocator's preferred memory type and memory type ID.
Definition: tritonserver.h:450
TRITONSERVER_InferenceRequestSetPriority
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestSetPriority(TRITONSERVER_InferenceRequest *inference_request, uint32_t priority)
Set the priority for a request.
TRITONSERVER_InferenceResponseOutputClassificationLabel
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseOutputClassificationLabel(TRITONSERVER_InferenceResponse *inference_response, const uint32_t index, const size_t class_index, const char **label)
Get a classification label associated with an output for a given index.
TRITONSERVER_ServerOptionsSetModelLoadThreadCount
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetModelLoadThreadCount(TRITONSERVER_ServerOptions *options, unsigned int thread_count)
Set the number of threads to concurrently load models in a server options.
TRITONSERVER_BufferAttributesNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesNew(TRITONSERVER_BufferAttributes **buffer_attributes)
TRITONSERVER_BufferAttributes.
TRITONSERVER_TYPE_BYTES
@ TRITONSERVER_TYPE_BYTES
Definition: tritonserver.h:128
TRITONSERVER_InferenceRequestRemoveAllInputs
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestRemoveAllInputs(TRITONSERVER_InferenceRequest *inference_request)
Remove all inputs from a request.
TRITONSERVER_ResponseAllocatorReleaseFn_t
TRITONSERVER_Error *(* TRITONSERVER_ResponseAllocatorReleaseFn_t)(TRITONSERVER_ResponseAllocator *allocator, void *buffer, void *buffer_userp, size_t byte_size, TRITONSERVER_MemoryType memory_type, int64_t memory_type_id)
Type for function that is called when the server no longer holds any reference to a buffer allocated ...
Definition: tritonserver.h:472
TRITONSERVER_ServerModelIsReady
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerModelIsReady(TRITONSERVER_Server *server, const char *model_name, const int64_t model_version, bool *ready)
Is the model ready?
TRITONSERVER_ServerIsReady
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerIsReady(TRITONSERVER_Server *server, bool *ready)
Is the server ready?
TRITONSERVER_INSTANCEGROUPKIND_MODEL
@ TRITONSERVER_INSTANCEGROUPKIND_MODEL
Definition: tritonserver.h:237
TRITONSERVER_ServerRegisterModelRepository
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerRegisterModelRepository(TRITONSERVER_Server *server, const char *repository_path, const TRITONSERVER_Parameter **name_mapping, const uint32_t mapping_count)
Register a new model repository.
TRITONSERVER_ServerOptionsSetLogWarn
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetLogWarn(TRITONSERVER_ServerOptions *options, bool log)
Enable or disable warning level logging.
TRITONSERVER_Error_Code
enum TRITONSERVER_errorcode_enum TRITONSERVER_Error_Code
TRITONSERVER_Error.
TRITONSERVER_StringToDataType
TRITONSERVER_DECLSPEC TRITONSERVER_DataType TRITONSERVER_StringToDataType(const char *dtype)
Get the Triton datatype corresponding to a string representation of a datatype.
TRITONSERVER_MetricNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricNew(TRITONSERVER_Metric **metric, TRITONSERVER_MetricFamily *family, const TRITONSERVER_Parameter **labels, const uint64_t label_count)
Create a new metric object.
TRITONSERVER_LogFormat
enum TRITONSERVER_logformat_enum TRITONSERVER_LogFormat
Format of logging.
TRITONSERVER_REQUEST_RELEASE_ALL
@ TRITONSERVER_REQUEST_RELEASE_ALL
Definition: tritonserver.h:878
TRITONSERVER_ServerStop
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerStop(TRITONSERVER_Server *server)
Stop a server object.
TRITONSERVER_DataTypeString
const TRITONSERVER_DECLSPEC char * TRITONSERVER_DataTypeString(TRITONSERVER_DataType datatype)
Get the string representation of a data type.
TRITONSERVER_InferenceTraceActivityFn_t
void(* TRITONSERVER_InferenceTraceActivityFn_t)(TRITONSERVER_InferenceTrace *trace, TRITONSERVER_InferenceTraceActivity activity, uint64_t timestamp_ns, void *userp)
Type for trace timeline activity callback function.
Definition: tritonserver.h:728
TRITONSERVER_ServerOptionsSetMetricsInterval
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetMetricsInterval(TRITONSERVER_ServerOptions *options, uint64_t metrics_interval_ms)
Set the interval for metrics collection in a server options.
TRITONSERVER_InferenceRequestRemoveAllInputData
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestRemoveAllInputData(TRITONSERVER_InferenceRequest *inference_request, const char *name)
Clear all input data from an input, releasing ownership of the buffer(s) that were appended to the in...
TRITONSERVER_METRIC_KIND_COUNTER
@ TRITONSERVER_METRIC_KIND_COUNTER
Definition: tritonserver.h:2261
TRITONSERVER_ServerOptionsSetModelControlMode
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetModelControlMode(TRITONSERVER_ServerOptions *options, TRITONSERVER_ModelControlMode mode)
Set the model control mode in a server options.
TRITONSERVER_ServerOptionsSetStrictModelConfig
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetStrictModelConfig(TRITONSERVER_ServerOptions *options, bool strict)
Enable or disable strict model configuration handling in a server options.
TRITONSERVER_MetricDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricDelete(TRITONSERVER_Metric *metric)
Delete a metric object.
TRITONSERVER_InferenceRequestAppendInputDataWithHostPolicy
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestAppendInputDataWithHostPolicy(TRITONSERVER_InferenceRequest *inference_request, const char *name, const void *base, size_t byte_size, TRITONSERVER_MemoryType memory_type, int64_t memory_type_id, const char *host_policy_name)
Assign a buffer of data to an input for execution on all model instances with the specified host poli...
TRITONSERVER_InferenceRequestSetReleaseCallback
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestSetReleaseCallback(TRITONSERVER_InferenceRequest *inference_request, TRITONSERVER_InferenceRequestReleaseFn_t request_release_fn, void *request_release_userp)
Set the release callback for an inference request.
TRITONSERVER_TXN_ONE_TO_ONE
@ TRITONSERVER_TXN_ONE_TO_ONE
Definition: tritonserver.h:1944
TRITONSERVER_errorcode_enum
TRITONSERVER_errorcode_enum
TRITONSERVER_Error.
Definition: tritonserver.h:303
TRITONSERVER_RATE_LIMIT_EXEC_COUNT
@ TRITONSERVER_RATE_LIMIT_EXEC_COUNT
Definition: tritonserver.h:1545
TRITONSERVER_ServerInferAsync
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerInferAsync(TRITONSERVER_Server *server, TRITONSERVER_InferenceRequest *inference_request, TRITONSERVER_InferenceTrace *trace)
Perform inference using the meta-data and inputs supplied by the 'inference_request'.
TRITONSERVER_MessageNewFromSerializedJson
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MessageNewFromSerializedJson(TRITONSERVER_Message **message, const char *base, size_t byte_size)
TRITONSERVER_Message.
TRITONSERVER_INDEX_FLAG_READY
@ TRITONSERVER_INDEX_FLAG_READY
Definition: tritonserver.h:1938
TRITONSERVER_ModelTxnPropertyFlag
enum tritonserver_txn_property_flag_enum TRITONSERVER_ModelTxnPropertyFlag
Model transaction policy flags.
TRITONSERVER_ServerLoadModel
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerLoadModel(TRITONSERVER_Server *server, const char *model_name)
Load the requested model or reload the model if it is already loaded.
TRITONSERVER_ParameterTypeString
const TRITONSERVER_DECLSPEC char * TRITONSERVER_ParameterTypeString(TRITONSERVER_ParameterType paramtype)
Get the string representation of a parameter type.
TRITONSERVER_MetricIncrement
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricIncrement(TRITONSERVER_Metric *metric, double value)
Increment the current value of metric by value.
TRITONSERVER_InferenceResponseOutput
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseOutput(TRITONSERVER_InferenceResponse *inference_response, const uint32_t index, const char **name, TRITONSERVER_DataType *datatype, const int64_t **shape, uint64_t *dim_count, const void **base, size_t *byte_size, TRITONSERVER_MemoryType *memory_type, int64_t *memory_type_id, void **userp)
Get all information about an output tensor.
TRITONSERVER_memorytype_enum
TRITONSERVER_memorytype_enum
TRITONSERVER_MemoryType.
Definition: tritonserver.h:162
TRITONSERVER_TYPE_FP16
@ TRITONSERVER_TYPE_FP16
Definition: tritonserver.h:125
TRITONSERVER_TRACE_REQUEST_END
@ TRITONSERVER_TRACE_REQUEST_END
Definition: tritonserver.h:707
TRITONSERVER_RequestFlag
enum tritonserver_requestflag_enum TRITONSERVER_RequestFlag
TRITONSERVER_InferenceRequest.
TRITONSERVER_InferenceTraceActivity
enum tritonserver_traceactivity_enum TRITONSERVER_InferenceTraceActivity
Trace activities.
TRITONSERVER_InferenceRequestFlags
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestFlags(TRITONSERVER_InferenceRequest *inference_request, uint32_t *flags)
Get the flag(s) associated with a request.
TRITONSERVER_ServerDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerDelete(TRITONSERVER_Server *server)
Delete a server object.
TRITONSERVER_ServerUnregisterModelRepository
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerUnregisterModelRepository(TRITONSERVER_Server *server, const char *repository_path)
Unregister a model repository.
TRITONSERVER_ServerOptionsSetHostPolicy
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetHostPolicy(TRITONSERVER_ServerOptions *options, const char *policy_name, const char *setting, const char *value)
Set a host policy setting for a given policy name in a server options.
TRITONSERVER_MetricSet
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricSet(TRITONSERVER_Metric *metric, double value)
Set the current value of metric to value.
TRITONSERVER_BufferAttributesDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesDelete(TRITONSERVER_BufferAttributes *buffer_attributes)
Delete a buffer attributes object.
TRITONSERVER_ERROR_INTERNAL
@ TRITONSERVER_ERROR_INTERNAL
Definition: tritonserver.h:305
TRITONSERVER_MODEL_CONTROL_NONE
@ TRITONSERVER_MODEL_CONTROL_NONE
Definition: tritonserver.h:1537
TRITONSERVER_MODEL_CONTROL_EXPLICIT
@ TRITONSERVER_MODEL_CONTROL_EXPLICIT
Definition: tritonserver.h:1539
TRITONSERVER_MetricFamilyNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricFamilyNew(TRITONSERVER_MetricFamily **family, const TRITONSERVER_MetricKind kind, const char *name, const char *description)
Create a new metric family object.
TRITONSERVER_ServerOptionsSetBufferManagerThreadCount
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetBufferManagerThreadCount(TRITONSERVER_ServerOptions *options, unsigned int thread_count)
Set the number of threads used in buffer manager in a server options.
TRITONSERVER_ServerMetadata
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerMetadata(TRITONSERVER_Server *server, TRITONSERVER_Message **server_metadata)
Get the metadata of the server as a TRITONSERVER_Message object.
TRITONSERVER_InferenceRequestCorrelationId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestCorrelationId(TRITONSERVER_InferenceRequest *inference_request, uint64_t *correlation_id)
Get the correlation ID of the inference request as an unsigned integer.
TRITONSERVER_ParameterType
enum TRITONSERVER_parametertype_enum TRITONSERVER_ParameterType
TRITONSERVER_ParameterType.
TRITONSERVER_InferenceRequestRemoveAllRequestedOutputs
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestRemoveAllRequestedOutputs(TRITONSERVER_InferenceRequest *inference_request)
Remove all output requests from an inference request.
TRITONSERVER_InferenceRequestAppendInputDataWithBufferAttributes
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestAppendInputDataWithBufferAttributes(TRITONSERVER_InferenceRequest *inference_request, const char *name, const void *base, TRITONSERVER_BufferAttributes *buffer_attributes)
Assign a buffer of data to an input.
tritonserver_tracelevel_enum
tritonserver_tracelevel_enum
TRITONSERVER_InferenceTrace.
Definition: tritonserver.h:677
TRITONSERVER_ParameterBytesNew
TRITONSERVER_DECLSPEC TRITONSERVER_Parameter * TRITONSERVER_ParameterBytesNew(const char *name, const void *byte_ptr, const uint64_t size)
Create a new parameter object with type TRITONSERVER_PARAMETER_BYTES.
TRITONSERVER_parametertype_enum
TRITONSERVER_parametertype_enum
TRITONSERVER_ParameterType.
Definition: tritonserver.h:181
TRITONSERVER_TYPE_INVALID
@ TRITONSERVER_TYPE_INVALID
Definition: tritonserver.h:115
TRITONSERVER_PARAMETER_STRING
@ TRITONSERVER_PARAMETER_STRING
Definition: tritonserver.h:182
TRITONSERVER_TYPE_FP64
@ TRITONSERVER_TYPE_FP64
Definition: tritonserver.h:127
TRITONSERVER_ServerLoadModelWithParameters
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerLoadModelWithParameters(TRITONSERVER_Server *server, const char *model_name, const TRITONSERVER_Parameter **parameters, const uint64_t parameter_count)
Load the requested model or reload the model if it is already loaded, with load parameters provided.
TRITONSERVER_ServerMetrics
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerMetrics(TRITONSERVER_Server *server, TRITONSERVER_Metrics **metrics)
Get the current metrics for the server.
TRITONSERVER_ServerModelTransactionProperties
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerModelTransactionProperties(TRITONSERVER_Server *server, const char *model_name, const int64_t model_version, uint32_t *txn_flags, void **voidp)
Get the transaction policy of the model.
TRITONSERVER_ServerOptionsSetRateLimiterMode
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetRateLimiterMode(TRITONSERVER_ServerOptions *options, TRITONSERVER_RateLimitMode mode)
Set the rate limit mode in a server options.
TRITONSERVER_InferenceTraceModelVersion
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceTraceModelVersion(TRITONSERVER_InferenceTrace *trace, int64_t *model_version)
Get the version of the model associated with a trace.
TRITONSERVER_InferenceRequestSetId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestSetId(TRITONSERVER_InferenceRequest *inference_request, const char *id)
Set the ID for a request.
TRITONSERVER_LogLevel
enum TRITONSERVER_loglevel_enum TRITONSERVER_LogLevel
TRITONSERVER_Logging.
tritonserver_txn_property_flag_enum
tritonserver_txn_property_flag_enum
Model transaction policy flags.
Definition: tritonserver.h:1943
TRITONSERVER_ServerOptionsSetStrictReadiness
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetStrictReadiness(TRITONSERVER_ServerOptions *options, bool strict)
Enable or disable strict readiness handling in a server options.
TRITONSERVER_ServerIsLive
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerIsLive(TRITONSERVER_Server *server, bool *live)
Is the server live?
TRITONSERVER_DataType
enum TRITONSERVER_datatype_enum TRITONSERVER_DataType
TRITONSERVER_DataType.
TRITONSERVER_InferenceRequestAddRawInput
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestAddRawInput(TRITONSERVER_InferenceRequest *inference_request, const char *name)
Add a raw input to a request.
TRITONSERVER_ServerOptionsSetGpuMetrics
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetGpuMetrics(TRITONSERVER_ServerOptions *options, bool gpu_metrics)
Enable or disable GPU metrics collection in a server options.
TRITONSERVER_InferenceTraceModelName
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceTraceModelName(TRITONSERVER_InferenceTrace *trace, const char **model_name)
Get the name of the model associated with a trace.
TRITONSERVER_TRACE_COMPUTE_START
@ TRITONSERVER_TRACE_COMPUTE_START
Definition: tritonserver.h:703
tritonserver_batchflag_enum
tritonserver_batchflag_enum
TRITONSERVER_Server.
Definition: tritonserver.h:1931
TRITONSERVER_ERROR_INVALID_ARG
@ TRITONSERVER_ERROR_INVALID_ARG
Definition: tritonserver.h:307
TRITONSERVER_TYPE_BOOL
@ TRITONSERVER_TYPE_BOOL
Definition: tritonserver.h:116
TRITONSERVER_MEMORY_CPU_PINNED
@ TRITONSERVER_MEMORY_CPU_PINNED
Definition: tritonserver.h:164
TRITONSERVER_ResponseCompleteFlag
enum tritonserver_responsecompleteflag_enum TRITONSERVER_ResponseCompleteFlag
Inference response complete flags.
TRITONSERVER_TRACE_TENSOR_BACKEND_OUTPUT
@ TRITONSERVER_TRACE_TENSOR_BACKEND_OUTPUT
Definition: tritonserver.h:710
TRITONSERVER_TYPE_UINT8
@ TRITONSERVER_TYPE_UINT8
Definition: tritonserver.h:117
TRITONSERVER_ParameterNew
TRITONSERVER_DECLSPEC TRITONSERVER_Parameter * TRITONSERVER_ParameterNew(const char *name, const TRITONSERVER_ParameterType type, const void *value)
Create a new parameter object.
TRITONSERVER_TYPE_INT32
@ TRITONSERVER_TYPE_INT32
Definition: tritonserver.h:123
eMode::strict
@ strict
TRITONSERVER_BufferAttributesSetMemoryTypeId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesSetMemoryTypeId(TRITONSERVER_BufferAttributes *buffer_attributes, int64_t memory_type_id)
Set the memory type id field of the buffer attributes.
TRITONSERVER_ServerOptionsSetLogFile
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetLogFile(TRITONSERVER_ServerOptions *options, const char *file)
Provide a log output file.
TRITONSERVER_ApiVersion
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ApiVersion(uint32_t *major, uint32_t *minor)
Get the TRITONBACKEND API version supported by the Triton shared library.
TRITONSERVER_TRACE_LEVEL_MIN
@ TRITONSERVER_TRACE_LEVEL_MIN
Deprecated. Use TRITONSERVER_TRACE_LEVEL_TIMESTAMPS.
Definition: tritonserver.h:681
TRITONSERVER_ServerPollModelRepository
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerPollModelRepository(TRITONSERVER_Server *server)
Check the model repository for changes and update server state based on those changes.
TRITONSERVER_InferenceResponseParameter
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseParameter(TRITONSERVER_InferenceResponse *inference_response, const uint32_t index, const char **name, TRITONSERVER_ParameterType *type, const void **vvalue)
Get all information about a parameter.
TRITONSERVER_ServerModelStatistics
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerModelStatistics(TRITONSERVER_Server *server, const char *model_name, const int64_t model_version, TRITONSERVER_Message **model_stats)
Get the statistics of a model as a TRITONSERVER_Message object.
TRITONSERVER_ResponseAllocatorStartFn_t
TRITONSERVER_Error *(* TRITONSERVER_ResponseAllocatorStartFn_t)(TRITONSERVER_ResponseAllocator *allocator, void *userp)
Type for function that is called to indicate that subsequent allocation requests will refer to a new ...
Definition: tritonserver.h:486
TRITONSERVER_ERROR_UNAVAILABLE
@ TRITONSERVER_ERROR_UNAVAILABLE
Definition: tritonserver.h:308
TRITONSERVER_INSTANCEGROUPKIND_AUTO
@ TRITONSERVER_INSTANCEGROUPKIND_AUTO
Definition: tritonserver.h:234
TRITONSERVER_TYPE_FP32
@ TRITONSERVER_TYPE_FP32
Definition: tritonserver.h:126
TRITONSERVER_InferenceTraceTensorNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceTraceTensorNew(TRITONSERVER_InferenceTrace **trace, TRITONSERVER_InferenceTraceLevel level, uint64_t parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn, TRITONSERVER_InferenceTraceTensorActivityFn_t tensor_activity_fn, TRITONSERVER_InferenceTraceReleaseFn_t release_fn, void *trace_userp)
Create a new inference trace object.
TRITONSERVER_instancegroupkind_enum
TRITONSERVER_instancegroupkind_enum
TRITONSERVER_InstanceGroupKind.
Definition: tritonserver.h:233
TRITONSERVER_ServerOptionsSetExitTimeout
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetExitTimeout(TRITONSERVER_ServerOptions *options, unsigned int timeout)
Set the exit timeout, in seconds, for the server in a server options.
tritonserver_modelcontrolmode_enum
tritonserver_modelcontrolmode_enum
TRITONSERVER_ServerOptions.
Definition: tritonserver.h:1536
TRITONSERVER_TRACE_TENSOR_BACKEND_INPUT
@ TRITONSERVER_TRACE_TENSOR_BACKEND_INPUT
Definition: tritonserver.h:709
TRITONSERVER_ServerUnloadModel
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerUnloadModel(TRITONSERVER_Server *server, const char *model_name)
Unload the requested model.
TRITONSERVER_ResponseAllocatorBufferAttributesFn_t
TRITONSERVER_Error *(* TRITONSERVER_ResponseAllocatorBufferAttributesFn_t)(TRITONSERVER_ResponseAllocator *allocator, const char *tensor_name, TRITONSERVER_BufferAttributes *buffer_attributes, void *userp, void *buffer_userp)
Type for allocation function that allocates a buffer to hold an output tensor with buffer attributes.
Definition: tritonserver.h:421
TRITONSERVER_loglevel_enum
TRITONSERVER_loglevel_enum
TRITONSERVER_Logging.
Definition: tritonserver.h:253
TRITONSERVER_InferenceTraceParentId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceTraceParentId(TRITONSERVER_InferenceTrace *trace, uint64_t *parent_id)
Get the parent id associated with a trace.
TRITONSERVER_ERROR_UNSUPPORTED
@ TRITONSERVER_ERROR_UNSUPPORTED
Definition: tritonserver.h:309
TRITONSERVER_ErrorNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ErrorNew(TRITONSERVER_Error_Code code, const char *msg)
Create a new error object.
TRITONSERVER_GetMetricKind
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_GetMetricKind(TRITONSERVER_Metric *metric, TRITONSERVER_MetricKind *kind)
Get the TRITONSERVER_MetricKind of metric and its corresponding family.
TRITONSERVER_InferenceResponseOutputCount
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseOutputCount(TRITONSERVER_InferenceResponse *inference_response, uint32_t *count)
Get the number of outputs available in the response.
TRITONSERVER_InstanceGroupKindString
const TRITONSERVER_DECLSPEC char * TRITONSERVER_InstanceGroupKindString(TRITONSERVER_InstanceGroupKind kind)
Get the string representation of an instance-group kind.
TRITONSERVER_RateLimitMode
enum tritonserver_ratelimitmode_enum TRITONSERVER_RateLimitMode
Rate limit modes.
TRITONSERVER_BufferAttributesByteSize
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesByteSize(TRITONSERVER_BufferAttributes *buffer_attributes, size_t *byte_size)
Get the byte size field of the buffer attributes.
TRITONSERVER_ServerOptionsSetResponseCacheByteSize
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetResponseCacheByteSize(TRITONSERVER_ServerOptions *options, uint64_t size)
Set the total response cache byte size that the server can allocate in CPU memory.
TRITONSERVER_MetricFamilyDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_MetricFamilyDelete(TRITONSERVER_MetricFamily *family)
Delete a metric family object.
TRITONSERVER_LOG_ERROR
@ TRITONSERVER_LOG_ERROR
Definition: tritonserver.h:256
TRITONSERVER_BufferAttributesSetMemoryType
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesSetMemoryType(TRITONSERVER_BufferAttributes *buffer_attributes, TRITONSERVER_MemoryType memory_type)
Set the memory type field of the buffer attributes.
TRITONSERVER_MetricFormat
enum tritonserver_metricformat_enum TRITONSERVER_MetricFormat
TRITONSERVER_Metrics.
TRITONSERVER_BufferAttributesCudaIpcHandle
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesCudaIpcHandle(TRITONSERVER_BufferAttributes *buffer_attributes, void **cuda_ipc_handle)
Get the CudaIpcHandle field of the buffer attributes object.
TRITONSERVER_InferenceResponseModel
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseModel(TRITONSERVER_InferenceResponse *inference_response, const char **model_name, int64_t *model_version)
Get model used to produce a response.
TRITONSERVER_TRACE_COMPUTE_OUTPUT_START
@ TRITONSERVER_TRACE_COMPUTE_OUTPUT_START
Definition: tritonserver.h:705
TRITONSERVER_InferenceRequestSetCorrelationId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestSetCorrelationId(TRITONSERVER_InferenceRequest *inference_request, uint64_t correlation_id)
Set the correlation ID of the inference request to be an unsigned integer.
TRITONSERVER_ResponseAllocatorNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ResponseAllocatorNew(TRITONSERVER_ResponseAllocator **allocator, TRITONSERVER_ResponseAllocatorAllocFn_t alloc_fn, TRITONSERVER_ResponseAllocatorReleaseFn_t release_fn, TRITONSERVER_ResponseAllocatorStartFn_t start_fn)
Create a new response allocator object.
TRITONSERVER_ServerOptionsSetModelLoadDeviceLimit
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetModelLoadDeviceLimit(TRITONSERVER_ServerOptions *options, const TRITONSERVER_InstanceGroupKind kind, const int device_id, const double fraction)
Specify the limit on memory usage as a fraction on the device identified by 'kind' and 'device_id'.
TRITONSERVER_RESPONSE_COMPLETE_FINAL
@ TRITONSERVER_RESPONSE_COMPLETE_FINAL
Definition: tritonserver.h:884
TRITONSERVER_LOG_WARN
@ TRITONSERVER_LOG_WARN
Definition: tritonserver.h:255
TRITONSERVER_InferenceRequestAppendInputData
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestAppendInputData(TRITONSERVER_InferenceRequest *inference_request, const char *name, const void *base, size_t byte_size, TRITONSERVER_MemoryType memory_type, int64_t memory_type_id)
Assign a buffer of data to an input.
TRITONSERVER_datatype_enum
TRITONSERVER_datatype_enum
TRITONSERVER_DataType.
Definition: tritonserver.h:114
tritonserver_metricformat_enum
tritonserver_metricformat_enum
TRITONSERVER_Metrics.
Definition: tritonserver.h:626
TRITONSERVER_InferenceRequestCorrelationIdString
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestCorrelationIdString(TRITONSERVER_InferenceRequest *inference_request, const char **correlation_id)
Get the correlation ID of the inference request as a string.
TRITONSERVER_INSTANCEGROUPKIND_GPU
@ TRITONSERVER_INSTANCEGROUPKIND_GPU
Definition: tritonserver.h:236
TRITONSERVER_InstanceGroupKind
enum TRITONSERVER_instancegroupkind_enum TRITONSERVER_InstanceGroupKind
TRITONSERVER_InstanceGroupKind.
TRITONSERVER_LogIsEnabled
TRITONSERVER_DECLSPEC bool TRITONSERVER_LogIsEnabled(TRITONSERVER_LogLevel level)
Is a log level enabled?
tritonserver_requestflag_enum
tritonserver_requestflag_enum
TRITONSERVER_InferenceRequest.
Definition: tritonserver.h:870
TRITONSERVER_InferenceRequestAddRequestedOutput
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestAddRequestedOutput(TRITONSERVER_InferenceRequest *inference_request, const char *name)
Add an output request to an inference request.
TRITONSERVER_TYPE_UINT32
@ TRITONSERVER_TYPE_UINT32
Definition: tritonserver.h:119
TRITONSERVER_ServerModelMetadata
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerModelMetadata(TRITONSERVER_Server *server, const char *model_name, const int64_t model_version, TRITONSERVER_Message **model_metadata)
Get the metadata of a model as a TRITONSERVER_Message object.
TRITONSERVER_ServerOptionsSetLogError
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetLogError(TRITONSERVER_ServerOptions *options, bool log)
Enable or disable error level logging.
TRITONSERVER_ModelIndexFlag
enum tritonserver_modelindexflag_enum TRITONSERVER_ModelIndexFlag
Model index flags. The enum values must be power-of-2 values.
TRITONSERVER_TRACE_LEVEL_MAX
@ TRITONSERVER_TRACE_LEVEL_MAX
Deprecated. Use TRITONSERVER_TRACE_LEVEL_TIMESTAMPS.
Definition: tritonserver.h:683
TRITONSERVER_BufferAttributesSetByteSize
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesSetByteSize(TRITONSERVER_BufferAttributes *buffer_attributes, size_t byte_size)
Set the byte size field of the buffer attributes.
TRITONSERVER_TRACE_REQUEST_START
@ TRITONSERVER_TRACE_REQUEST_START
Definition: tritonserver.h:701
TRITONSERVER_InferenceTraceNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceTraceNew(TRITONSERVER_InferenceTrace **trace, TRITONSERVER_InferenceTraceLevel level, uint64_t parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn, TRITONSERVER_InferenceTraceReleaseFn_t release_fn, void *trace_userp)
Create a new inference trace object.
TRITONSERVER_PARAMETER_BOOL
@ TRITONSERVER_PARAMETER_BOOL
Definition: tritonserver.h:184
TRITONSERVER_ServerModelConfig
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerModelConfig(TRITONSERVER_Server *server, const char *model_name, const int64_t model_version, const uint32_t config_version, TRITONSERVER_Message **model_config)
Get the configuration of a model as a TRITONSERVER_Message object.
TRITONSERVER_ServerOptionsSetCpuMetrics
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetCpuMetrics(TRITONSERVER_ServerOptions *options, bool cpu_metrics)
Enable or disable CPU metrics collection in a server options.
TRITONSERVER_ServerOptionsSetBackendDirectory
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetBackendDirectory(TRITONSERVER_ServerOptions *options, const char *backend_dir)
Set the directory containing backend shared libraries.
TRITONSERVER_ResponseAllocatorDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ResponseAllocatorDelete(TRITONSERVER_ResponseAllocator *allocator)
Delete a response allocator.
TRITONSERVER_ParameterDelete
TRITONSERVER_DECLSPEC void TRITONSERVER_ParameterDelete(TRITONSERVER_Parameter *parameter)
Delete an parameter object.
TRITONSERVER_ModelBatchFlag
enum tritonserver_batchflag_enum TRITONSERVER_ModelBatchFlag
TRITONSERVER_Server.
TRITONSERVER_BufferAttributesMemoryType
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesMemoryType(TRITONSERVER_BufferAttributes *buffer_attributes, TRITONSERVER_MemoryType *memory_type)
Get the memory type field of the buffer attributes.
TRITONSERVER_InferenceRequestRemoveInput
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestRemoveInput(TRITONSERVER_InferenceRequest *inference_request, const char *name)
Remove an input from a request.
TRITONSERVER_ServerOptionsSetCudaMemoryPoolByteSize
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetCudaMemoryPoolByteSize(TRITONSERVER_ServerOptions *options, int gpu_device, uint64_t size)
Set the total CUDA memory byte size that the server can allocate on given GPU device in a server opti...
TRITONSERVER_ServerOptionsAddRateLimiterResource
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsAddRateLimiterResource(TRITONSERVER_ServerOptions *options, const char *resource_name, const size_t resource_count, const int device)
Add resource count for rate limiting.
TRITONSERVER_MEMORY_GPU
@ TRITONSERVER_MEMORY_GPU
Definition: tritonserver.h:165
tritonserver_requestreleaseflag_enum
tritonserver_requestreleaseflag_enum
Inference request release flags.
Definition: tritonserver.h:877
TRITONSERVER_InferenceRequestSetResponseCallback
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestSetResponseCallback(TRITONSERVER_InferenceRequest *inference_request, TRITONSERVER_ResponseAllocator *response_allocator, void *response_allocator_userp, TRITONSERVER_InferenceResponseCompleteFn_t response_fn, void *response_userp)
Set the allocator and response callback for an inference request.
TRITONSERVER_InferenceRequestTimeoutMicroseconds
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestTimeoutMicroseconds(TRITONSERVER_InferenceRequest *inference_request, uint64_t *timeout_us)
Get the timeout for a request, in microseconds.
TRITONSERVER_InferenceResponseDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseDelete(TRITONSERVER_InferenceResponse *inference_response)
TRITONSERVER_InferenceResponse.
TRITONSERVER_ServerOptionsSetLogFormat
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsSetLogFormat(TRITONSERVER_ServerOptions *options, const TRITONSERVER_LogFormat format)
Set the logging format.
tritonserver_ratelimitmode_enum
tritonserver_ratelimitmode_enum
Rate limit modes.
Definition: tritonserver.h:1543
TRITONSERVER_InferenceRequestDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestDelete(TRITONSERVER_InferenceRequest *inference_request)
Delete an inference request object.
TRITONSERVER_BATCH_FIRST_DIM
@ TRITONSERVER_BATCH_FIRST_DIM
Definition: tritonserver.h:1933
TRITONSERVER_ServerOptionsNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsNew(TRITONSERVER_ServerOptions **options)
Create a new server options object.
TRITONSERVER_BufferAttributesMemoryTypeId
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_BufferAttributesMemoryTypeId(TRITONSERVER_BufferAttributes *buffer_attributes, int64_t *memory_type_id)
Get the memory type id field of the buffer attributes.
TRITONSERVER_InferenceRequestNew
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceRequestNew(TRITONSERVER_InferenceRequest **inference_request, TRITONSERVER_Server *server, const char *model_name, const int64_t model_version)
Create a new inference request object.
TRITONSERVER_INSTANCEGROUPKIND_CPU
@ TRITONSERVER_INSTANCEGROUPKIND_CPU
Definition: tritonserver.h:235
TRITONSERVER_ServerOptionsDelete
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_ServerOptionsDelete(TRITONSERVER_ServerOptions *options)
Delete a server options object.
TRITONSERVER_InferenceResponseParameterCount
TRITONSERVER_DECLSPEC TRITONSERVER_Error * TRITONSERVER_InferenceResponseParameterCount(TRITONSERVER_InferenceResponse *inference_response, uint32_t *count)
Get the number of parameters available in the response.