Stars
Building General-Purpose Robots Based on Embodied Foundation Model
Pytorch implementation of "Spatial As Deep: Spatial CNN for Traffic Scene Understanding"
RLinf is a flexible and scalable open-source infrastructure designed for post-training foundation models (LLMs, VLMs, VLAs) via reinforcement learning.
NVIDIA Isaac GR00T N1.6 - A Foundation Model for Generalist Robots.
SERL: A Software Suite for Sample-Efficient Robotic Reinforcement Learning
Official inference framework for 1-bit LLMs
本仓库包含对 Claude Code v1.0.33 进行逆向工程的完整研究和分析资料。包括对混淆源代码的深度技术分析、系统架构文档,以及重构 Claude Code agent 系统的实现蓝图。主要发现包括实时 Steering 机制、多 Agent 架构、智能上下文管理和工具执行管道。该项目为理解现代 AI agent 系统设计和实现提供技术参考。
[RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions
Embodied Reasoning Question Answer (ERQA) Benchmark
A collection of tabletop tasks in Mujoco
[RSS'25] This repository is the implementation of "NaVILA: Legged Robot Vision-Language-Action Model for Navigation"
Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels
FlashInfer: Kernel Library for LLM Serving
[HPCA 2026] A GPU-optimized system for efficient long-context LLMs decoding with low-bit KV cache.
[CVPR 2025 Oral]Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
QLoRA: Efficient Finetuning of Quantized LLMs
FlashMLA: Efficient Multi-head Latent Attention Kernels
A collection of resources and papers on Diffusion Models
The code repository of "MBQ: Modality-Balanced Quantization for Large Vision-Language Models"
The official repo of MiniMax-Text-01 and MiniMax-VL-01, large-language-model & vision-language-model based on Linear Attention