Stars
MoE training for Me and You and maybe other people
Fast and memory-efficient exact attention
Official PyTorch implementation of DistiLLM: Towards Streamlined Distillation for Large Language Models (ICML 2024)
Public release of the code for "Accelerating Vision Transformers with Adaptive Patches"
AMD-AGI / torchtitan-amd
Forked from pytorch/torchtitanA PyTorch native platform for training generative AI models
A PyTorch native platform for training generative AI models
Reference PyTorch implementation and models for DINOv3
State-of-the-art Image & Video CLIP, Multimodal Large Language Models, and More!
The official Python SDK for the Perceptron API
[ICLR 2023 Spotlight] Vision Transformer Adapter for Dense Predictions
Open source implementation of "A Self-Supervised Descriptor for Image Copy Detection" (SSCD).
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
PyTorch code for hierarchical k-means -- a data curation method for self-supervised learning
Official repository for "AM-RADIO: Reduce All Domains Into One"
[CVPR 2025] Mono-InternVL: Pushing the Boundaries of Monolithic Multimodal Large Language Models with Endogenous Visual Pre-training
A library for efficient similarity search and clustering of dense vectors.
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & TIS & vLLM & Ray & Dynamic Sampling & Async Agentic RL)
(CVPR 2025) PyramidDrop: Accelerating Your Large Vision-Language Models via Pyramid Visual Redundancy Reduction
A high-throughput and memory-efficient inference and serving engine for LLMs
The official repository of "Video assistant towards large language model makes everything easy"