Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Stars
Repository for the paper "Large Language Model-Based Agents for Software Engineering: A Survey". Keep updating.
AIInfra(AI 基础设施)指AI系统从底层芯片等硬件,到上层软件栈支持AI大模型训练和推理。
Accelerating Long Context LLM Inference with Accuracy-Preserving Context Optimization in SGLang, vLLM, llama.cpp, OpenClaw, RAG, and Agentic AI.
project-numina / kimina-prover-rl
Forked from verl-project/verlKimina-Prover RL pipeline
Kimina Lean server (+ client SDK)
Major CS conference publication stats (including accepted and submitted) by year.
slime is an LLM post-training framework for RL Scaling.
Serverless LLM Serving for Everyone.
Official repository for the EMNLP 2025 paper "Slim-SC: Thought Pruning for Efficient Scaling with Self-Consistency".
Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models
The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.
[ICLR 2025🔥] SVD-LLM & [NAACL 2025🔥] SVD-LLM V2
A PyTorch native platform for training generative AI models
An interference-aware scheduler for fine-grained GPU sharing
NVIDIA Linux open GPU kernel module source
LM engine is a library for pretraining/finetuning LLMs
LaTeX Template for Statement of Purpose (SoP)
dInfer: An Efficient Inference Framework for Diffusion Language Models
[SIGMOD 2025] PQCache: Product Quantization-based KVCache for Long Context LLM Inference
An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Models
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.