- Denver, CO
-
06:04
(UTC -06:00) - https://www.aaronbatilo.dev
- @aaronbatilo
- https://sliceofexperiments.com
Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Starred repositories
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Robust Speech Recognition via Large-Scale Weak Supervision
Unsloth Studio is a web UI for training and running open models like Qwen, DeepSeek, gpt-oss and Gemma locally.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
aider is AI pair programming in your terminal
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
A toolkit for developing and comparing reinforcement learning algorithms.
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
Fully open reproduction of DeepSeek-R1
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
Train transformer language models with reinforcement learning.
Hackable and optimized Transformers building blocks, supporting a composable construction.
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, Slurm, 20+ clouds, on-prem).
Benchmarks of approximate nearest neighbor libraries in Python
A PyTorch native platform for training generative AI models
Our library for RL environments + evals
A compact implementation of SGLang, designed to demystify the complexities of modern LLM serving systems.
A library for mechanistic interpretability of GPT-style language models
An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Models
Post-training with Tinker
A simple, performant and scalable Jax LLM!
Minimalistic 4D-parallelism distributed training framework for education purpose
Distribute and run AI workloads on Kubernetes magically in Python, like PyTorch for ML infra.
Checkpoint-engine is a simple middleware to update model weights in LLM inference engines
A CLI to estimate inference memory requirements for Hugging Face models, written in Python.