NVIDIA AI Ecosystem: Related Tools#
After preparing your data with NeMo Curator, you’ll likely want to use it to train models. NVIDIA provides an integrated ecosystem of AI tools that work seamlessly with data prepared by NeMo Curator. This guide outlines the related tools for your next steps.
NeMo Framework#
NVIDIA NeMo is an end-to-end framework for building, training, and fine-tuning GPU-accelerated language models. It provides:
Pretrained model checkpoints
Training and inference scripts
Optimization techniques for large-scale deployments
Training a Tokenizer#
Tokenizers transform text into tokens that language models can interpret. While NeMo Curator doesn’t handle tokenizer training or tokenization in general, NeMo does.
Learn how to train a tokenizer using NeMo in the tokenizer training documentation.
Training Large Language Models#
Pretraining a large language model involves running next-token prediction on large curated datasets, exactly the type that NeMo Curator helps you prepare. NeMo handles everything for pretraining large language models using your curated data.
Find comprehensive information on:
Pretraining methodologies
Model evaluation
Parameter-efficient fine-tuning (PEFT)
Distributed training
NeMo Aligner#
NVIDIA NeMo Aligner is a framework designed for aligning language models with human preferences.
After pretraining a large language model, aligning it allows you to interact with it in a chat-like setting. NeMo Aligner lets you take curated alignment data and use it to align a pretrained language model.
Learn about NeMo Aligner’s capabilities including:
Reinforcement Learning from Human Feedback (RLHF)
Direct Preference Optimization (DPO)
Proximal Policy Optimization (PPO)
Constitutional AI (CAI)
In the NeMo Aligner documentation.
NVIDIA AI Enterprise#
For organizations looking to deploy trained models to production, NVIDIA AI Enterprise provides a software platform that includes enterprise support for:
The complete NeMo framework
Pretrained foundation models
Deployment and inference tools
Enterprise-grade security and support
Complete Workflow#
A typical end-to-end workflow with NVIDIA’s AI tools includes:
Data Preparation: Use NeMo Curator to clean, filter, and prepare your dataset
Tokenization: Train or use a tokenizer with NeMo
Model Training: Pretrain or fine-tune models with NeMo
Alignment: Align models with human preferences using NeMo Aligner
Deployment: Deploy models using NVIDIA AI Enterprise or Triton Inference Server
This integrated ecosystem allows you to move from raw data to deployed, production-ready models with consistent tooling and optimized performance.