Environment Variables for NVIDIA NIM for Object Detection#
Use this documentation to learn about the environment variables for NVIDIA NIM for Object Detection.
Environment Variables#
Note
The following NIMs do not support NIM_SERVED_MODEL_NAME:
nemotron-graphic-elements-v1
nemoretriever-page-elements-v2
nemotron-page-elements-v3
nemotron-table-structure-v1
nemotron-ocr-v1
The following table identifies the environment variables that are used in the container.
Set environment variables with the -e command-line argument to the docker run command.
Name |
Description |
Default Value |
|---|---|---|
|
Set this variable to the value of your personal NGC API key. |
None |
|
Specifies the fully qualified path, in the container, for downloaded models. |
|
|
Specifies the network port number, in the container, for HTTP access to the microservice. Refer to Publishing ports in the Docker documentation for more information about host and container network ports. |
|
|
Specifies the number of worker threads to start for HTTP requests. |
|
|
Specifies the network port number, in the container, for NVIDIA Triton Inference Server. |
|
|
When set to |
|
|
When set to |
|
|
Specifies the logging level. The microservice supports the following values: DEBUG, INFO, WARNING, ERROR, and CRITICAL. |
|
|
Set to |
|
|
Specifies the fully qualified path, in the container, for the model manifest YAML file. |
|
|
Specifies the model profile ID to use with the container. By default, the container attempts to automatically match the host GPU model and GPU count with the optimal model profile. |
None |
|
If set to a non-empty string, the |
None |
|
Specifies the model names used in the API.
Specify multiple names in a comma-separated list.
If you specify multiple names, the server responds to any of the names.
The name in the model field of a response matches the name in the request.
By default, the model name is inferred from the For Prometheus metrics, this value is used in the |
None |
|
For the NVIDIA Triton Inference Server, specify the byte size for the CUDA memory pool for all GPUs visible to the container. |
By default, Image Retriever NIMs automatically set the CUDA memory pool based on the maximum input data size for the loaded TensorRT engine. However, you might want to increase the CUDA memory pool size when you enable dynamic batching or run highly concurrent workloads. A typical error message that indicates that you should increase the CUDA memory pool is |
|
The maximum tensor size used for pre-processing and inference. Increase this if pre-processing takes too long. Decrease this if worker VRAM consumption is too high. |
|
|
For the NVIDIA Triton Inference Server, sets the max queue delayed time to allow other requests to join the dynamic batch.
It takes effect only when |
|
|
True to enable the pipeline’s dynamic batching. We recommended that you only set this to False for debugging. |
|
|
True to enable concurrent model inference when the model’s max batch size is configured to be less than the data max batch size. If you enable this on a single-GPU deployment, you can experience GPU resource contention. |
|
|
By default, Image Retriever NIMs start Triton with |
None |
|
True to enable detailed pipeline timing instrumentation. Adds ‘cuda.synchronize()’ calls for accurate GPU timing, which can degrade performance. |
|
|
This option determines after how many requests the |
1 (After every request) |
|
The batch size threshold for switching between GPU and GPU decoding. |
|
|
Specifies the gRPC port number, for NVIDIA Triton Inference Server. |
|
|
The threshold for idle VRAM memory (bytes) after which the Torch CUDA cache is emptied and all inter-process communication (IPC) files are closed. |
1GB |
|
When set to |
|
|
Specify the maximum batch size that the underlying Triton instance can process. The value must be less than or equal to maximum batch size that was used to compile the engine. By default, the NIM uses the maximum possible batch size for a given model and GPU. To decrease the memory footprint of the server, choose a smaller maximum batch size. If the model uses the |
None |
|
Sets the max queue size for the underlying Triton instance. For more information, refer to the Triton User Guide. Triton returns an InferenceServerException on new requests if you exceed the max queue size. |
None |
|
The max queue delay for the BLS pipeline (Triton dynamic batching). Controls how many requests Triton aggregates before it calls the pipeline. Increase this value to improve throughput at the cost of increased latency. This is an alias for ‘NIM_TRITON_PIPELINE_MAX_QUEUE_DELAY_MICROSECONDS’ for backwards compatibility. |
|
|
The max batch size for the underlying model. For TensorRT, this value is validated against the engine profiles and is used to inform which TensorRT profile to load. This value is also used for inference chunking with the pipeline. |
|
|
The max queue delay for the underlying model’s dynamic batching. Controls how long Triton waits to batch inference calls to the model from the pipeline. For most scenarios, this should be kept at |
|
|
The max batch size for the BLS pipeline (Triton dynamic batching). Controls how many requests Triton aggregates before it calls the pipeline. |
|
|
Max queue delay for the BLS pipeline (Triton dynamic batching). Controls how many requests Triton aggregates before calling the pipeline. Increase this value to improve throughput at the cost of increased latency. |
‘10000’ |
|
The number of batches between pipeline timing summary prints. |
|
|
For the NVIDIA Triton Inference Server, specify the byte size for the pinned CPU memory pool that is used to transfer data from the host to the GPU device. |
By default, |
|
If set, this option configures Triton to rate limit the execution count throughout the ensemble model pipeline to the provided integer value. GPU-bound model inference is given priority, while other components of the ensemble model pipeline (pre-processors, post-processors, etc.) are given lower priority. |
None |
|
The number of pipeline worker instances to deploy. Assuming the pipeline’s dynamic batching parameters are set sufficiently high, this value should be increased to reduce request queuing at the cost of higher CPU, GPU, and VRAM resource use. For resource constrained deployments, keep this value set to |
|