Integrate RL Frameworks
These guides cover how to integrate NeMo Gym into a new RL training framework. Use them if you are:
- A training framework maintainer adding NeMo Gym support
- Contributing NeMo Gym integration for a training framework that does not have one yet
Just want to train models? See Training Tutorials for supported frameworks.
Prerequisites
Before integrating Gym into your training framework, ensure you have:
- An RL training framework with policy optimization support (PPO, GRPO, or similar)
- A generation backend (vLLM, SGLang, or equivalent)
- Familiarity with OpenAI-compatible HTTP server APIs
Integration Components
Gym integration requires implementing the following components in your training framework:
Generation Backend
OpenAI-compatible HTTP server requirements and existing implementations across RL frameworks.
prerequisiteOn-Policy Corrections
Fixes for on-policy training in multi-step and multi-turn scenarios to prevent train-generation mismatch.
prerequisiteIntegration Footprint
Implementation components, form factor, and reference implementations from NeMo RL.
implementationSuccess Criteria
Validation criteria and benchmarks to verify correct Gym integration.
validationIntegration Workflow
The typical integration workflow follows this sequence: