Highlights
- Pro
Stars
Train the smallest LM you can that fits in 16MB. Best model wins!
Open Multi-Agent Interactive Classroom — Get an immersive, multi-agent learning experience in just one click
IndexCache: Accelerating Sparse Attention via Cross-Layer Index Reuse
The awesome collection of OpenClaw skills. 5,400+ skills filtered and categorized from the official OpenClaw Skills Registry.🦞
Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞
Ongoing research training transformer models at scale
Official Repository for "Glyph: Scaling Context Windows via Visual-Text Compression"
🌿 DeepPrune: Parallel Scaling without Inter-trace Redundancy
Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels
[SIGGRAPH Asia 2025] CHARM: Control-point-based 3D Anime Hairstyle Auto-Regressive Modeling
🚀🚀 Efficient implementations of Native Sparse Attention
Renderer for the harmony response format to be used with gpt-oss
GLM-SIMPLE-EVALS: The evaluation repository for the GLM-4.5 series of models by Z.ai.
GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models
[COLM 2024] A Survey on Deep Learning for Theorem Proving
An efficient implementation of the NSA (Native Sparse Attention) kernel
TradingAgents: Multi-Agents LLM Financial Trading Framework
slime is an LLM post-training framework for RL Scaling.
[SIGGRAPH 2025] PrimitiveAnything: Human-Crafted 3D Primitive Assembly Generation with Auto-Regressive Transformer
ReLE评测:中文AI大模型能力评测(持续更新):目前已囊括359个大模型,覆盖chatgpt、gpt-5.2、o4-mini、谷歌gemini-3-pro、Claude-4.6、文心ERNIE-X1.1、ERNIE-5.0、qwen3-max、qwen3.5-plus、百川、讯飞星火、商汤senseChat等商用模型, 以及step3.5-flash、kimi-k2.5、ernie4.5、…
[ICML 2025] XAttention: Block Sparse Attention with Antidiagonal Scoring