Stars
Voice Activity Detector (VAD) : low-latency, high-performance and lightweight
Implementation of papers in 100 lines of code.
General plug-and-play inference library for Recursive Language Models (RLMs), supporting various sandboxes.
High-quality single file implementation of Deep Reinforcement Learning algorithms with research-friendly features (PPO, DQN, C51, DDPG, TD3, SAC, PPG)
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
Fully Open Framework for Democratized Multimodal Training
A Conversational Speech Generation Model
Minimal and annotated implementations of key ideas from modern deep learning research.
A Comprehensive Toolkit for High-Quality PDF Content Extraction
Fixed version of `torchrun` on Jülich Supercomputing Centre
verl: Volcano Engine Reinforcement Learning for LLMs
Muon is an optimizer for hidden layers in neural networks
Code for 'LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders'
LLaSA: Scaling Train-time and Inference-time Compute for LLaMA-based Speech Synthesis
Convert PDF to HTML without losing text or format.
Minimalistic 4D-parallelism distributed training framework for education purpose
Multipack distributed sampler for fast padding-free training of LLMs
Everything about the SmolLM and SmolVLM family of models
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
Tips for Writing a Research Paper using LaTeX