RL Training with Unsloth#
This tutorial demonstrates how to use Unsloth to fine-tune models with NeMo Gym environments.
Unsloth is a fast, memory-efficient library for fine-tuning large language models. It provides optimized implementations that significantly reduce memory usage and training time, making it possible to fine-tune larger models on consumer hardware.
Prerequisites#
A Google account (for Colab) or a local GPU with 16GB+ VRAM
Familiarity with NeMo Gym concepts (Quickstart)
Getting Started#
Follow these interactive notebooks to train models with Unsloth and NeMo Gym:
Check out Unsloth’s documentation for more details.