-
Tsinghua University
- Beijing, China
- https://jason-huang03.github.io/
Stars
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga…
Clone a voice in 5 seconds to generate arbitrary speech in real-time
🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!
Your AI second brain. Self-hostable. Get answers from the web or your docs. Build custom agents, schedule automations, do deep research. Turn any online or local LLM into your personal, autonomous …
Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.
Interactive deep learning book with multi-framework code, math, and discussions. Adopted at 500 universities from 70 countries including Stanford, MIT, Harvard, and Cambridge.
Official inference framework for 1-bit LLMs
Fully open reproduction of DeepSeek-R1
GUI for a Vocal Remover that uses Deep Neural Networks.
SGLang is a high-performance serving framework for large language models and multimodal models.
Universal LLM Deployment Engine with ML Compilation
verl: Volcano Engine Reinforcement Learning for LLMs
Train transformer language models with reinforcement learning.
A Flexible Framework for Experiencing Heterogeneous LLM Inference/Fine-tune Optimizations
An open source implementation of CLIP.
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. Tensor…
Minimal reproduction of DeepSeek R1-Zero
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)
Easily fine-tune, evaluate and deploy gpt-oss, Qwen3, DeepSeek-R1, or any open source LLM / VLM!
Accessible large language models via k-bit quantization for PyTorch.
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.