Release Notes for NeMo Microservices#
Check out the latest release notes for the NeMo microservices.
Tip
If you’ve installed one of the previous releases of the NeMo microservices using Helm and want to upgrade, choose one of the following options:
To upgrade to the latest release, follow the steps at Upgrade NeMo Microservices Helm Chart.
To uninstall and reinstall, follow the steps at Uninstall NeMo Microservices Helm Chart and Install NeMo Microservices Helm Chart.
Release 25.12.0#
This release includes the following key features and known issues.
Key Features#
Review the key features introduced to the NeMo microservices in this release.
NeMo Studio (Early Access)#
Added a Safe Synthesizer interface to create, monitor, and analyze privacy-protected synthetic data generation jobs
To learn more about NeMo Studio, refer to the NeMo Studio user guide
NeMo Auditor (Early Access)#
Updated the garak security scanner to v0.13.2. This update adds the following new probe.
fitd.FITD: Foot in the door probe: Transform queries from benign to harmful obtaining intermediate responses to get compliance from the model.
The web_injection group of probes are now part of the default/default audit configuration.
The packagehallucination.PackageHallucinationProbe probe is removed in this release.
For more information, refer to the probe reference summary.
Fixed an issue with the HTML report result. In last release, only the most-recently run probe was shown as the entire probe specification. In this release, all the probes from the audit config are shown in the probe specification.
Refer to Get an HTML Report for a sample.
NeMo Customizer#
Added support for nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16. This is a 30B parameter MoE model (3.5B active) supporting both LoRA and All Weights fine-tuning. See NVIDIA Nemotron 3 Nano 30B A3B for configuration details.
Note
Full SFT Limitation:
For All Weights fine-tuning, make sure you configure batch_size == data_parallel_size × micro_batch_size.
When deploying the fine-tuned model using the Deployment Management Service, the default NIM profile uses FP8 precision (requires H100+). For A100 GPUs, set NIM_TAGS_SELECTOR: “precision=bf16” in additional_envs to select the BF16 precision profile.
Added support for the mistralai/Ministral-3-3B-Reasoning-2512. This model supports both All Weights and LoRA fine-tuning methods.
Clarified the batch_size parameter in the documentation, distinguishing it from microbatch_size, and removed the invalid global_batch_size option.
Upgraded Automodel to October 2025 version, enabling support for more recent Hugging Face models.
NeMo Data Designer#
Support for passing extra params to model providers.
Open sourced the core Data Designer library
NeMo Evaluator#
Trajectory evaluation of NAT based agents in offline mode
Parallelization of LLM-as-a-Judge as a default
Usability improvements:
Allow job to run and return results even with some NaN outputs
Simplify agent evaluation user journey by consolidating evaluation types
Updated documentation on secret management and deletion
NeMo Guardrails#
Improved code examples throughout the documentation