Highlights
- Pro
Starred repositories
a teaching deep learning framework: the bridge from micrograd to tinygrad
JAX implementation of OpenAI's Whisper model for up to 70x speed-up on TPU.
Introduction to Machine Learning Systems
The simplest, fastest repository for training/finetuning small-sized VLMs.
MoE training for Me and You and maybe other people
Comprehensive guide, algorithms and tools on distributed systems
Writing custom Linear Algebra and ML kernels in CUDA to outperform pytorch, cuBLAS, numpy
Miles is an enterprise-facing reinforcement learning framework for LLM and VLM post-training, forked from and co-evolving with slime.
Tile primitives for speedy kernels
Continuous Thought Machines, because thought takes time and reasoning is a process.
Training framework with a goal to explore the frontier of sample efficiency of small language models
SGLang is a high-performance serving framework for large language models and multimodal models.
biasing the universal tokenizer and an attempt to optimize compression rates in multilingual compression
a parallel and minimal implementation of Byte Pair Encoding (BPE) from scratch in less than 200 lines of python.
custom flash attention kernel in cuda to benchmark it against torch and burn my rtx 3050
List of papers related to neural network quantization in recent AI conferences and journals.
The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
You like pytorch? You like micrograd? You love tinygrad! ❤️
Course 18.S191 at MIT, Fall 2022 - Introduction to computational thinking with Julia
A Complete Resource to Master Graduate-Level GenAI Mathematics
Learn ML engineering for free in 4 months! Register here 👇🏼
🍕 AI agent that calls to order pizza for you
Cleanlab's open-source library is the standard data-centric AI package for data quality and machine learning with messy, real-world data and labels.