Stars
A PyTorch native platform for training generative AI models
We bring the spirit of nanogpt-speedrun into the omni-modal world
Official implementation of GDPO: Group reward-Decoupled Normalization Policy Optimization for Multi-reward RL Optimization
A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.
(CVPR 2025) From Slow Bidirectional to Fast Autoregressive Video Diffusion Models
Sharp Monocular View Synthesis in Less Than a Second
Official implementation of HPSv3: Towards Wide-Spectrum Human Preference Score (ICCV2025)
Ongoing research training transformer models at scale
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Joint Distillation for Fast Likelihood Evaluation and Sampling in Flow-based Models
cuTile is a programming model for writing parallel kernels for NVIDIA GPUs
An extremely fast Python type checker and language server, written in Rust.
Directly Aligning the Full Diffusion Trajectory with Fine-Grained Human Preference
Provide with pre-build flash-attention package wheels on Linux and Windows platforms using GitHub Actions
[ICLR 2026] Self-Representation Alignment for Diffusion Transformers (SRA)
Official inference repo for FLUX.2 models
DC-Gen: Post-Training Diffusion Acceleration with Deeply Compressed Latent Space
[ICLR 2026 Oral] DiffusionNFT: Online Diffusion Reinforcement with Forward Process
Zotero MCP: Connects your Zotero research library with Claude and other AI assistants via the Model Context Protocol to discuss papers, get summaries, analyze citations, and more.
HunyuanImage-3.0: A Powerful Native Multimodal Model for Image Generation
[NeurIPS 2025] An official implementation of Flow-GRPO: Training Flow Matching Models via Online RL