-
SJTU & Alibaba Cloud
- Hangzhou, China
-
13:06
(UTC -12:00)
Highlights
- Pro
Stars
Large Language Model (LLM) Systems Paper List
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
Renderer for the harmony response format to be used with gpt-oss
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
Kimi K2 is the large language model series developed by Moonshot AI team
A fast communication-overlapping library for tensor/expert parallelism on GPUs.
Copilot Chat extension for VS Code
Python package built to ease deep learning on graph, on top of existing DL frameworks.
Aims to implement dual-port and multi-qp solutions in deepEP ibrc transport
Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.
hostCC is a congestion control architecture which handles host congestion, along with in-network congestion
SGLang is a fast serving framework for large language models and vision language models.
A high-throughput and memory-efficient inference and serving engine for LLMs
Set of datasets for the deep learning recommendation model (DLRM).
Efficient and easy multi-instance LLM serving
rocSHMEM intra-kernel networking runtime for AMD dGPUs on the ROCm platform.
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
verl: Volcano Engine Reinforcement Learning for LLMs
VeOmni: Scaling Any Modality Model Training with Model-Centric Distributed Recipe Zoo
Distributed Compiler based on Triton for Parallel Systems
DeepSeek-V3/R1 inference performance simulator
A Datacenter Scale Distributed Inference Serving Framework
A lightweight, powerful framework for multi-agent workflows
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)