Starred repositories
Official repo for the paper "Scaling Synthetic Data Creation with 1,000,000,000 Personas"
Refine high-quality datasets and visual AI models
A Distributed Attention Towards Linear Scalability for Ultra-Long Context, Heterogeneous Data Training
A Lighting Pytorch Framework for Recommendation Models (PyTorch推荐算法框架), Easy-to-use and Easy-to-extend. https://datawhalechina.github.io/torch-rechub/
An Open Foundation Model and Benchmark to Accelerate Generative Recommendation
This repo contains the code for "VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks" [ICLR 2025]
A curated list of awesome platforms, tools, practices and resources that helps run LLMs locally
CaptionQA: Is Your Caption as Useful as the Image Itself?
The repository provides code for running inference and finetuning with the Meta Segment Anything Model 3 (SAM 3), links for downloading the trained model checkpoints, and example notebooks that sho…
The Open-Source Multimodal AI Agent Stack: Connecting Cutting-Edge AI Models and Agent Infra
BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.
🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using Agentic Retrieval 🔄.
[ICLR'26] Grasp Any Region: Towards Precise, Contextual Pixel Understanding for Multimodal LLMs
[NeurIPS 2025 Spotlight] A Token is Worth over 1,000 Tokens: Efficient Knowledge Distillation through Low-Rank Clone.
Tarsier -- a family of large-scale video-language models, which is designed to generate high-quality video descriptions , together with good capability of general video understanding.
The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.
Best practices for training DeepSeek, Mixtral, Qwen and other MoE models using Megatron Core.
A Next-Generation Training Engine Built for Ultra-Large MoE Models
ScalarLM - a unified training and inference stack
PyTorch building blocks for the OLMo ecosystem
Repository containing code and data for the paper "ArgCMV: An Argument Summarization Benchmark for the LLM-era", accepted at EMNLP 2025 Main Conference.
Lumina-DiMOO - An Open-Sourced Multi-Modal Large Diffusion Language Model
Awesome LLM pre-training resources, including data, frameworks, and methods.
Analyze the inference of Large Language Models (LLMs). Analyze aspects like computation, storage, transmission, and hardware roofline model in a user-friendly interface.
Latency and Memory Analysis of Transformer Models for Training and Inference