Helm Chart Values for NVIDIA NeMo Retriever Embedding NIM#
Use this documentation to learn about the environment variables for NVIDIA NeMo Retriever Embedding NIM.
Deployment Parameters#
Name |
Description |
Value |
|---|---|---|
|
Affinity settings for deployment. Allows to constraint pods to nodes. |
|
|
Specify privilege and access control settings for Container(Only affects the main container). |
|
|
Adds arbitrary environment variables to the main container - Key Value Pairs. |
|
|
Adds arbitrary additional volumes to the deployment set definition. |
|
|
NIM-LLM Image Repository. |
|
|
Image tag. |
|
|
Image pull policy. |
|
|
Specify secret names that are needed for the main container and any init containers. Object keys are the names of the secrets. |
|
|
Specify labels to ensure that NeMo Inference is deployed only on certain nodes (likely best to set this to |
|
|
Specify additional annotation to the main deployment pods. |
|
|
Specify privilege and access control settings for pod (Only affects the main pod). |
|
|
Specify user UID for pod. |
|
|
Specify group ID for pod. |
|
|
Specify file system owner group id. |
|
|
Specify replica count for deployment. |
|
|
Specify resources limits and requests for the running service. |
|
|
Specify number of GPUs to present to the running service. |
|
|
Specifies whether a service account should be created. |
|
|
Specifies annotations to be added to the service account. |
|
|
Specifies whether to automatically mount the service account to the container. |
|
|
Specify name of the service account to use. If it is not set and create is true, a name is generated using a fullname template. |
|
|
Specify tolerations for pod assignment. Allows the scheduler to schedule pods with matching taints. |
Autoscaling Parameters#
Values used for autoscaling. If autoscaling is not enabled, these are ignored. They should be overridden on a per-model basis based on quality-of-service metrics as well as cost metrics. This isn’t recommended except with usage of the custom metrics API using something like the prometheus-adapter. Standard metrics of CPU and memory are of limited use in scaling NIM.
Name |
Description |
Value |
|---|---|---|
|
Enable horizontal pod autoscaler. |
|
|
Specify minimum replicas for autoscaling. |
|
|
Specify maximum replicas for autoscaling. |
|
|
Array of metrics for autoscaling. |
|
Ingress Parameters#
Name |
Description |
Value |
|---|---|---|
|
Enables ingress. |
|
|
Specify class name for Ingress. |
|
|
Specify additional annotations for ingress. |
|
|
Specify list of hosts each containing lists of paths. |
|
|
Specify name of host. |
|
|
Specify ingress path. |
|
|
Specify path type. |
|
|
Specify service type. It can be nemo or openai – make sure your model serves the appropriate port(s). |
|
|
Specify list of pairs of TLS secretName and hosts. |
|
Probe Parameters#
Name |
Description |
Value |
|---|---|---|
|
Enable livenessProbe. |
|
|
LivenessProbe http or script, but no script is currently provided. |
|
|
LivenessProbe endpoint path. |
|
|
Initial delay seconds for livenessProbe. |
|
|
Timeout seconds for livenessProbe. |
|
|
Period seconds for livenessProbe. |
|
|
Success threshold for livenessProbe. |
|
|
Failure threshold for livenessProbe. |
|
|
Enable readinessProbe. |
|
|
Readiness Endpoint Path. |
|
|
Initial delay seconds for readinessProbe. |
|
|
Timeout seconds for readinessProbe. |
|
|
Period seconds for readinessProbe. |
|
|
Success threshold for readinessProbe. |
|
|
Failure threshold for readinessProbe. |
|
|
Enable startupProbe. |
|
|
StartupProbe Endpoint Path. |
|
|
Initial delay seconds for startupProbe. |
|
|
Timeout seconds for startupProbe. |
|
|
Period seconds for startupProbe. |
|
|
Success threshold for startupProbe. |
|
|
Failure threshold for startupProbe. |
|
Storage Parameters#
Name |
Description |
Default Value |
|---|---|---|
|
Settings to modify the path |
– |
|
True to enable persistent volumes. |
|
|
An existing PersistentVolumeClaim to use. If you use |
|
|
A persistent volume storage class. If empty, the default provisioner is used. |
|
|
The PVC retention policy when the release is deleted. Only used with StatefulSet volume templates. |
|
|
The PVC retention policy when the release is scaled. Only used with StatefulSet volume templates. |
|
|
The access modes for the storage. If you are using an NFS or similar storage, you can use |
|
|
Specify the size of claim, for example |
|
|
Configures model cache on local disk on the nodes using hostPath. For special cases only. You should investigate and understand the security implications before you use this option. |
|
Service Parameters#
Name |
Description |
Value |
|---|---|---|
|
Specify service type for the deployment. |
|
|
Override the default service name. |
|
|
Specify HTTP Port for the service. |
|
|
Specify additional annotations to be added to service. |
|
OpenTelemetry Parameters#
Name |
Description |
Value |
|---|---|---|
|
Specify if this chart should deploy zipkin for metrics. |
|
|
Specify if this chart should deploy OpenTelemetry for metrics. |
|
|
Specify if this chart should sink metrics to OpenTelemetry. |
|
|
Env variables to configure OTEL in the container, sane defaults in chart. |
|
|
Log Level to set for the container and metrics collection. |
|
OpenTelemetry configurations can be found in the values section of the OpenTelemetry repository.
Note
Configure the OpenTelemetry exporters according to your needs. The provided helm chart provides sample configuration for exporting traces to Zipkin, and metrics to an OTLP-compatible receiver, stored at opentelemetry-collector.config.exporters.zipkin and opentelemetry-collector.config.exporters.otlp, respectively.
For example, if your metrics setup operates in a pull-based fashion and you want to expose NIM metrics in Prometheus format, you can do so by replacing the OTLP exporter with a Prometheus exporter.