Starred repositories
A Multi-Dimensional Constraint Framework for Evaluating and Improving Instruction Following in Large Language Models
🚀🤖 Crawl4AI: Open-source LLM Friendly Web Crawler & Scraper. Don't be shy, join here: https://discord.gg/jP8KfhDhyN
📦 Repomix is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) or other AI tools lik…
Official Repository for SIGIR2024 Demo Paper "An Integrated Data Processing Framework for Pretraining Foundation Models"
Training Turn-by-Turn Verifiers for Dialogue Tutoring Agents: The Curious Case of LLMs as Your Coding Tutors (ACL Findings 2025)
🐳 Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"
[TMLR] A curated list of language modeling researches for code (and other software engineering activities), plus related datasets.
Fully open reproduction of DeepSeek-R1
APOLLO: SGD-like Memory, AdamW-level Performance; MLSys'25 Oustanding Paper Honorable Mention
[ACL 2025 Main] Official Repository for "Evaluating Language Models as Synthetic Data Generators"
[CSUR 2025] Continual Learning of Large Language Models: A Comprehensive Survey
Representation Engineering: A Top-Down Approach to AI Transparency
Rigourous evaluation of LLM-synthesized code - NeurIPS 2023 & COLM 2024
Efficient Triton Kernels for LLM Training
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
Tools for merging pretrained large language models.
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
🇨🇳 GitHub中文排行榜,各语言分设「软件 | 资料」榜单,精准定位中文好项目。各取所需,高效学习。
Minimalistic large language model 3D-parallelism training
心理健康大模型 (LLM x Mental Health), Pre & Post-training & Dataset & Evaluation & Depoly & RAG, with InternLM / Qwen / Baichuan / DeepSeek / Mixtral / LLama / GLM series models
Netease Youdao's open-source embedding and reranker models for RAG products.