Stars
The most open diffusion language model for code generation — releasing pretraining, evaluation, inference, and checkpoints.
Official implementation of "Fast-dLLM: Training-free Acceleration of Diffusion LLM by Enabling KV Cache and Parallel Decoding"
The definitive Web UI for local AI, with powerful features and easy setup.
[Arxiv] Discrete Diffusion in Large Language and Multimodal Models: A Survey
Discrete Diffusion Forcing (D2F): dLLMs Can Do Faster-Than-AR Inference
SwiReasoning: Switch-Thinking in Latent and Explicit for Pareto-Superior Reasoning LLMs
[Open-source Project] UniMoCap: community implementation to unify the text-motion datasets (HumanML3D, KIT-ML, and BABEL) and whole-body motion dataset (Motion-X).
[NeurIPS 2023] Reflexion: Language Agents with Verbal Reinforcement Learning
[ICML 2025] LaCache: Ladder-Shaped KV Caching for Efficient Long-Context Modeling of Large Language Models
Pruna is a model optimization framework built for developers, enabling you to deliver faster, more efficient models with minimal overhead.
This repo contains the source code for the paper "Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement Learning"
Local-first AI Notepad for Private Meetings
[ICCV'23] Dual Aggregation Transformer for Image Super-Resolution
Code and project page for "3D-aware Face Swapping" in CVPR 2023
Retrieval and Retrieval-augmented LLMs
⚡️🧪 Fast LLM Tool Calling Experimentation, big and smol
Code for Paper: “Low-Resource” Text Classification: A Parameter-Free Classification Method with Compressors
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
[ICCV-2023] The official repo for the paper "LivelySpeaker: Towards Semantic-aware Co-Speech Gesture Generation".
X-modaler is a versatile and high-performance codebase for cross-modal analytics(e.g., image captioning, video captioning, vision-language pre-training, visual question answering, visual commonsens…
[ICCV 2023] ExposureDiffusion: Learning to Expose for Low-light Image Enhancement
[ICCV 2023] Lighting Every Darkness in Two Pairs: A Calibration-Free Pipeline for RAW Denoising && [Arxiv 2023] Make Explicit Calibration Implicit: Calibrate Denoiser Instead of the Noise Model
Adversarial Training of Denoising Diffusion Model Using Dual Discriminators for High-Fidelity Multi-Speaker TTS
[CVPR2022] PyTorch implementation of ''Background Activation Suppression for Weakly Supervised Object Localization''.
[NeurIPS 2022] TOIST: Task Oriented Instance Segmentation Transformer with Noun-Pronoun Distillation
CVPR2022, BatchFormer: Learning to Explore Sample Relationships for Robust Representation Learning, https://arxiv.org/abs/2203.01522