tritonclient.http.aio#
Classes
|
This feature is currently in beta and may be subject to change. |
- class tritonclient.http.aio.InferenceServerClient(url, verbose=False, conn_limit=100, conn_timeout=60.0, ssl=False, ssl_context=None)#
This feature is currently in beta and may be subject to change.
An analogy of the
tritonclient.http.InferenceServerClientto enable calling via asyncio syntax. The object is intended to be used by a single thread and simultaneously calling methods with different threads is not supported and can cause undefined behavior.- _fix_header(headers)#
Returns a header that is valid for aiohttp.
- Parameters:
headers (dict (or None)) – HTTP headers to fix before processing the request.
- async _get(request_uri, headers, query_params)#
Issues the GET request to the server
- Parameters:
request_uri (str) – The request URI to be used in GET request.
headers (dict) – Additional HTTP headers to include in the request.
query_params (dict) – Optional url query parameters to use in network transaction.
- Returns:
The response from server.
- Return type:
aiohttp.ClientResponse
- async _post(request_uri, request_body, headers, query_params)#
Issues the POST request to the server
- Parameters:
request_uri (str) – The request URI to be used in POST request.
request_body (str) – The body of the request
headers (dict) – Additional HTTP headers to include in the request.
query_params (dict) – Optional url query parameters to use in network transaction.
- Returns:
The response from server.
- Return type:
aiohttp.ClientResponse
- _validate_headers(headers)#
Checks for any unsupported HTTP headers before processing a request.
- Parameters:
headers (dict) – HTTP headers to validate before processing the request.
- Raises:
InferenceServerException – If an unsupported HTTP header is included in a request.
- async close()#
Close the client. Any future calls to server will result in an Error.
- static generate_request_body(inputs, outputs=None, request_id='', sequence_id=0, sequence_start=False, sequence_end=False, priority=0, timeout=None, parameters=None)#
Refer to
tritonclient.http.InferenceServerClient.generate_request_body()
- async get_inference_statistics(model_name='', model_version='', headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.get_inference_statistics()
- async get_log_settings(headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.get_log_settings()
- async get_model_config(model_name, model_version='', headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.get_model_config()
- async get_model_metadata(model_name, model_version='', headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.get_model_metadata()
- async get_model_repository_index(headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.get_model_repository_index()
- async get_server_metadata(headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.get_server_metadata()
- async get_trace_settings(model_name=None, headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.get_trace_settings()
- async infer(model_name, inputs, model_version='', outputs=None, request_id='', sequence_id=0, sequence_start=False, sequence_end=False, priority=0, timeout=None, headers=None, query_params=None, request_compression_algorithm=None, response_compression_algorithm=None, parameters=None)#
- async is_model_ready(model_name, model_version='', headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.is_model_ready()
- async is_server_live(headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.is_server_live()
- async is_server_ready(headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.is_server_ready()
- async load_model(model_name, headers=None, query_params=None, config=None, files=None)#
Refer to
tritonclient.http.InferenceServerClient.load_model()
- static parse_response_body(response_body, verbose=False, header_length=None, content_encoding=None)#
Refer to
tritonclient.http.InferenceServerClient.parse_response_body()
- async unload_model(model_name, headers=None, query_params=None, unload_dependents=False)#
Refer to
tritonclient.http.InferenceServerClient.unload_model()
- async update_log_settings(settings, headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.update_log_settings()
- async update_trace_settings(model_name=None, settings={}, headers=None, query_params=None)#
Refer to
tritonclient.http.InferenceServerClient.update_trace_settings()
- async tritonclient.http.aio._get_error(response)#
Returns the
InferenceServerExceptionobject if response indicates the error. If no error then return None
- async tritonclient.http.aio._raise_if_error(response)#
Raise
InferenceServerExceptionif received non-Success response from the server
Modules