Lists (9)
Sort Name ascending (A-Z)
Stars
Awsome of VLM-CL. Continual Learning for VLMs: A Survey and Taxonomy Beyond Forgetting
📖 This is a repository for organizing papers, codes and other resources related to Visual Reinforcement Learning.
Agent framework and applications built upon Qwen>=3.0, featuring Function Calling, MCP, Code Interpreter, RAG, Chrome extension, etc.
Tongyi Deep Research, the Leading Open-source Deep Research Agent
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, GLM4.5, InternLM3, DeepSeek-R1, ...) and 200+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, Llava, GLM4v, Ph…
SGLang is a fast serving framework for large language models and vision language models.
Use Claude Code as the foundation for coding infrastructure, allowing you to decide how to interact with the model while enjoying updates from Anthropic.
Use Kimi latest model(kimi-k2-0711-preview) to drive your Claude Code.
verl: Volcano Engine Reinforcement Learning for LLMs
Train transformer language models with reinforcement learning.
📚LeetCUDA: Modern CUDA Learn Notes with PyTorch for Beginners🐑, 200+ CUDA Kernels, Tensor Cores, HGEMM, FA-2 MMA.🎉
Universal LLM Deployment Engine with ML Compilation
CUDA Templates and Python DSLs for High-Performance Linear Algebra
FlashInfer: Kernel Library for LLM Serving
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
A framework for few-shot evaluation of language models.
Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM
This is the homepage of a new book entitled "Mathematical Foundations of Reinforcement Learning."
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!
Introduction to Machine Learning Systems
A high-throughput and memory-efficient inference and serving engine for LLMs
TinyChatEngine: On-Device LLM Inference Library
Open deep learning compiler stack for cpu, gpu and specialized accelerators
Development repository for the Triton language and compiler
Fast and memory-efficient exact attention
Speech-to-text, text-to-speech, speaker diarization, speech enhancement, source separation, and VAD using next-gen Kaldi with onnxruntime without Internet connection. Support embedded systems, Andr…
中文nlp解决方案(大模型、数据、模型、训练、推理)