Starred repositories
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!
Minimalistic large language model 3D-parallelism training
Minimalistic 4D-parallelism distributed training framework for education purpose
📚 Collection of awesome generation acceleration resources.
每个人都能看懂的大模型知识分享,LLMs春/秋招大模型面试前必看,让你和面试官侃侃而谈
Clean, Robust, and Unified PyTorch implementation of popular Deep Reinforcement Learning (DRL) algorithms (Q-learning, Duel DDQN, PER, C51, Noisy DQN, PPO, DDPG, TD3, SAC, ASL)
Best practices & guides on how to write distributed pytorch training code
An extremely fast Python package and project manager, written in Rust.
Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.
A curated list of awesome Multimodal studies.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
50+ mini web projects using HTML, CSS & JS
✨✨Beyond LLaVA-HD: Diving into High-Resolution Large Multimodal Models
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & vLLM & Ray & Dynamic Sampling & Async Agentic RL)
A curated list of awesome vision and language resources (still under construction... stay tuned!)
Supporting PyTorch models with the Google AI Edge TFLite runtime.
👁️ + 💬 + 🎧 = 🤖 Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]
Code and models for ICML 2024 paper, NExT-GPT: Any-to-Any Multimodal Large Language Model
✨✨Latest Advances on Multimodal Large Language Models
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)