Stars
State-of-the-art framework for fast, large-scale training and inference of diffusion models
HuggingFace conversion and training library for Megatron-based models
A library for exporting models including NeMo and Hugging Face to optimized inference backends, and deploying them for efficient querying
Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face support
SWE-agent takes a GitHub issue and tries to automatically fix it, using your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challenges. [NeurIPS 2024]
Scalable toolkit for efficient model reinforcement
A Datacenter Scale Distributed Inference Serving Framework
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & TIS & vLLM & Ray & Dynamic Sampling & Async Agentic RL)
BioNeMo Framework: For building and adapting AI models in drug discovery at scale
NVIDIA Resiliency Extension is a python package for framework developers and users to implement fault-tolerant features. It improves the effective training time by minimizing the downtime due to fa…
Auto generate docs for typer and click commands using typer's rich console formatting.
A tool to configure, launch and manage your machine learning experiments.
A project to improve skills of large language models
PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.
Speed up model training by fixing data loading.
Scalable data pre processing and curation toolkit for LLMs
GPU Affinity is a package to automatically set the CPU process affinity to match the hardware architecture on a given platform
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
Provides end-to-end model development pipelines for LLMs and Multimodal models that can be launched on-prem or cloud-native.
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on Hopper, Ada and Blackwell GPUs, to provide better performance…
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
🤗 The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation tools
Hydra is a framework for elegantly configuring complex applications