Stars
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
A high-throughput and memory-efficient inference and serving engine for LLMs
Turn any PDF or image document into structured data for your AI. A powerful, lightweight OCR toolkit that bridges the gap between images/PDFs and LLMs. Supports 100+ languages.
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
🚀🚀 「大模型」2小时完全从0训练64M的小参数GPT!🌏 Train a 64M-parameter GPT from scratch in just 2h!
Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.
verl: Volcano Engine Reinforcement Learning for LLMs
A framework for few-shot evaluation of language models.
slime is an LLM post-training framework for RL Scaling.
[NeurIPS D&B '25] The one-stop repository for LLM unlearning
[CVPR2024] MMA-Diffusion: MultiModal Attack on Diffusion Models
[NeurIPS25] Official repo for "Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning"
tmlr-group / G-effect
Forked from QizhouWang/G-effect[ICLR 2025] "Rethinking LLM Unlearning Objectives: A Gradient Perspective and Go Beyond"