Stars
A curated list of awesome Claude Skills, resources, and tools for customizing Claude AI workflows
My Python scripts to make high-quality figures for publications in top AI conferences and journals.
🚀🤖 Crawl4AI: Open-source LLM Friendly Web Crawler & Scraper. Don't be shy, join here: https://discord.gg/jP8KfhDhyN
Code and dataset for paper: DeepPlanner: Scaling Planning Capability for Deep Research Agents via Advantage Shaping
A version of verl to support diverse tool use
slime is an LLM post-training framework for RL Scaling.
An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Models
Recommend new arxiv papers of your interest daily according to your Zotero libarary.
[NeurIPS 2025] 🌐 WebThinker: Empowering Large Reasoning Models with Deep Research Capability
Search-R1: An Efficient, Scalable RL Training Framework for Reasoning & Search Engine Calling interleaved LLM based on veRL
The AI-native database built for LLM applications, providing incredibly fast hybrid search of dense vector, sparse vector, tensor (multi-vector), and full-text.
[EMNLP'25 Industry] Repo for "Z1: Efficient Test-time Scaling with Code"
yuanzhoulvpi2017 / nano_rl
Forked from verl-project/verl在verl上做reward的定制开发
🔍 Search-o1: Agentic Search-Enhanced Large Reasoning Models [EMNLP 2025]
SWE-bench: Can Language Models Resolve Real-world Github Issues?
Rigourous evaluation of LLM-synthesized code - NeurIPS 2023 & COLM 2024
A framework for the evaluation of autoregressive code generation language models.
A comprehensive collection of process reward models.
Scalable RL solution for advanced reasoning of language models
Curated, opinionated index of post-R1 LLM × Reinforcement Learning. Many deep-dive blog posts cross-linked to many papers — GRPO, DAPO, DPO, PPO, RLHF, GSPO, CISPO, VAPO, Reward Modeling, MoE RL st…