NeMo Framework is NVIDIA's GPU accelerated, end-to-end training framework for large language models (LLMs), multi-modal models and speech models. It enables seamless scaling of training (both pretraining and post-training) workloads from single GPU to thousand-node clusters for both 🤗Hugging Face/PyTorch and Megatron models. This GitHub organization includes a suite of libraries and recipe collections to help users train models from end to end.
NeMo Framework is also a part of the NVIDIA NeMo software suite for managing the AI agent lifecycle.
- [10/6/2025]Enabling PyTorch Native Pipeline Parallelism for 🤗 Hugging Face Transformer Models
- [9/22/2025]Fine-tune Hugging Face Models Instantly with Day-0 Support with NVIDIA NeMo AutoModel
- [9/18/2025]🚀 NeMo Framework Now Supports Google Gemma 3n: Efficient Multimodal Fine-tuning Made Simple
- [10/1/2025]On-policy Distillation (Qwen3-style) (Qwen3-style)
- [9/27/2025]FP8 Quantization in NeMo RL
- [8/15/2025]NeMo-RL: Journey of Optimizing Weight Transfer in Large MoE Models by 10x
- [7/31/2025]NeMo-RL V0.3: Scalable and Performant Post-training with Nemo-RL via Megatron-Core
- [5/15/2025]Reinforcement Learning with NVIDIA NeMo-RL: Reproducing a DeepScaleR Recipe Using GRPO
More to come and stay tuned!
Visit the individual repos to find out more 🔍, raise 🐛, contribute ✍️ and participate in discussion forums 🗣️!
Repo | Summary | Training Loop | Training Backends | Infernece Backends | Model Coverage | Container |
---|---|---|---|---|---|---|
NeMo Megatron-Bridge | Pretraining, LoRA, SFT | PyT native loop | Megatron-core | NA | LLM & VLM | NeMo Framework Container |
NeMo AutoModel | Pretraining, LoRA, SFT | PyT native loop | PyTorch | NA | LLM, VLM, Omni, VFM | NeMo AutoModel Container |
NeMo 1.x & 2.x (with Lightning)->will repurpose to focus on Speech | Pretraining,SFT | PyTorch Lightning Loop | Megatron-core & PyTorch | RIVA | Speech | NA |
NeMo RL | SFT, RL | PyT native loop | Megatron-core & PyTorch | vLLM | LLM, VLM | NeMo RL container |
NeMo Gym (WIP) | RL Environment, integrate with RL Framework | NA | NA | NA | NA | NeMo RL Container |
NeMo Aligner (deprecated) | SFT, RL | PyT Lightning Loop | Megatron-core | TRTLLM | LLM | NA |
NeMo Curator | Data curation | NA | NA | NA | Agnostic | NeMo Curator Container |
NeMo Evaluator | Model evaluation | NA | NA | Agnostic | NeMo Framework Container | |
NeMo Export-Deploy | Export to Production | NA | NA | vLLM, TRT, TRTLLM, ONNX | Agnostic | NeMo Framework Container |
NeMo Run | Experiment launcher | NA | NA | NA | Agnostic | NeMo Framework Container |
NeMo Guardrails | Guardrail model response | NA | NA | NA | NA | |
NeMo Skills | Reference pipeline for SDG & Eval | NA | NA | NA | Agnostic | NA |
NeMo Emerging Optimizers | Collection of Optimizers | NA | Agnostic | NA | NA | NA |
NeMo DFM (WIP) | Diffusion foundation model training | PyT native loop | Megatron-core and PyTorch | PyTorch | VFM, Diffusion | TBD |
The NeMo GitHub Org and its repo collections are created to address the following problems
- Need for composability: The Previous NeMo is monolithic and encompasses too many things, making it hard for users to find what they need. Container size is also an issue. Breaking down the Monolithic repo into a series of functional-focused repos to facilitate code discovery.
- Need for customizability: The Previous NeMo uses PyTorch Lighting as the default trainer loop, which provides some out of the box functionality but making it hard to customize. NeMo Megatron-Bridge, NeMo AutoModel, and NeMo RL have adopted pytorch native custom loop to improve flexibility and ease of use for developers.
To learn more about NVIDIA NeMo Framework and all of its component libraries, please refer to the NeMo Framework User Guide, which includes quick start guide, tutorials, model-specific recipes, best practice guides and performance benchmarks.
Apache 2.0 licensed with third-party attributions documented in each repository.