Build.nvidia.com Model Target#
The following target references the NVIDIA Llama 3.1 Nemotron Nano 8B V1 model from build.nvidia.com.
Refer to garak.generators.nim.NVOpenAIChat for the parameters to specify in the options.nim field.
The options override the default values from the DEFAULT_PARAMS in the API reference.
Set the NMP_BASE_URL environment variable to the NeMo Auditor service endpoint.
Refer to Accessing the Microservice for more information.
With NeMo Platform running, follow the model provider instructions to set up a model provider that points at integrate.api.nvidia.com and the model to audit. Create the provider in the default workspace and name it build.
import os
from nemo_platform import NeMoPlatform
client = NeMoPlatform(
base_url=os.environ.get("NMP_BASE_URL", "http://localhost:8080"),
workspace="default",
)
target = client.audit.targets.create(
workspace="default",
name="demo-build-nvidia-com-target",
type="nim.NVOpenAIChat",
model="nvidia/llama-3.1-nemotron-nano-8b-v1",
options={
"nim": {
"skip_seq_start": "<think>",
"skip_seq_end": "</think>",
"max_tokens": 3200,
"nmp_uri_spec": {"inference_gateway": { "workspace": "default", "provider": "build" }}
}
}
)
print(target.model_dump_json(indent=2))
nmp audit targets create --workspace "default" \
--name "demo-build-nvidia-com-target" \
--type "nim.NVOpenAIChat" \
--model "nvidia/llama-3.1-nemotron-nano-8b-v1" \
--options '{"nim": {"skip_seq_start": "<think>", "skip_seq_end": "</think>", "max_tokens": 3200, "nmp_uri_spec": {"inference_gateway": {"workspace": "default", "provider": "build"}}}}' \
-f json
Example Output
{
"id": "audit-target-8wFuUtWXEFCZNBssheu8Y1",
"created_at": "2026-03-24T22:16:15.338486",
"created_by": "service:platform",
"entity_id": "audit-target-8wFuUtWXEFCZNBssheu8Y1",
"model": "nvidia/llama-3.1-nemotron-nano-8b-v1",
"parent": null,
"type": "nim.NVOpenAIChat",
"updated_at": "2026-03-24T22:16:15.338491",
"updated_by": "service:platform",
"workspace": "default",
"description": null,
"name": "demo-build-nvidia-com-target",
"options": {
"nim": {
"skip_seq_start": "<think>",
"skip_seq_end": "</think>",
"max_tokens": 3200,
"nmp_uri_spec": {
"inference_gateway": {
"workspace": "default",
"provider": "build"
}
}
}
},
"project": null
}