Model Support#
Broad coverage for ð€Hugging Face models via NeMo AutoModel#
NeMo-RL support ð€Hugging Face models from the following classes
LLMs (AutoModelForCausalLM)
VLMs (AutoModelForImageTextToText)
for model sizes under 70B at up to 32k sequence length.
Optimal acceleration for top models via NeMo Megatron-bridge#
NeMo Megatron-Bridge provides acceleration recipes for the below models. Users can also leverage the on-line checkpoint conversion (i.e the âbridgeâ) by directly inputting a ð€Hugging Face checkpoint.
LLMs:
Qwen: Qwen2.5-1.5B/7B/32B, Qwen3-1.5B/8B/32B, Qwen3-30B-A3B, Qwen3-235B-A22B
Llama: Llama 3.1/3.3-8B, Llama 3.1/3.3-70B, Llama 3.2-1B
Deepseek: Deepseek-V3/R1-671B
Mistral: Mistral-NeMo-12B
Moonlight-16B-A3B
Gemma: Gemma-3-1B/27B
GPT-OSS: GPT-OSS-20B/120B
NeMotron: Llama-Nemotron-Super-49B, Nemotron-nano-v2-12B, Nemotron-Nano-v3-30A3B
VLMs:
Qwen: Qwen2.5VL-3B
In addition, please refer to our performance page for benchmarks and full reproducible yaml recipe configs.