- New York
Stars
Learning Deep Representations of Data Distributions
Extremely fast Query Engine for DataFrames, written in Rust
A reactive notebook for Python — run reproducible experiments, query with SQL, execute as a script, deploy as an app, and version with git. Stored as pure Python. All in a modern, AI-native editor.
Our library for RL environments + evals
🌊 The leading agent orchestration platform for Claude. Deploy intelligent multi-agent swarms, coordinate autonomous workflows, and build conversational AI systems. Features enterprise-grade archite…
Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"
MiMo: Unlocking the Reasoning Potential of Language Model – From Pretraining to Posttraining
Darwin Gödel Machine: Open-Ended Evolution of Self-Improving Agents
Implementing DeepSeek R1's GRPO algorithm from scratch
FULL Augment Code, Claude Code, Cluely, CodeBuddy, Comet, Cursor, Devin AI, Junie, Kiro, Leap.new, Lovable, Manus, NotionAI, Orchids.app, Perplexity, Poke, Qoder, Replit, Same.dev, Trae, Traycer AI…
PyTorch building blocks for the OLMo ecosystem
Benchmarking Agentic LLM and VLM Reasoning On Games
A Collection of Competitive Text-Based Games for Language Model Evaluation and Reinforcement Learning
fast trainer for educational purposes
🌐 Make websites accessible for AI agents. Automate tasks online with ease.
A bunch of kernels that might make stuff slower 😉
Amazon Nova Act is an AWS service for building and deploying highly reliable AI agents that automate UI-based workflows at scale.
Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible.
Synthetic Data Generation Toolkit for LLMs
Pocket Flow: 100-line LLM framework. Let Agents build Agents!
No fortress, purely open ground. OpenManus is Coming.
verl: Volcano Engine Reinforcement Learning for LLMs
Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verified research papers.
LM engine is a library for pretraining/finetuning LLMs