🎯
Focusing
Focused on LLM Inference KV Cache & Storage System
- ChengDu, China
-
02:48
(UTC +08:00) - syaojun.github.io
- @jasonyao1024
- in/yaojun4096
Tom Turney
TheTom
Working on LLM inference systems, KV cache compression, and kernel-level optimizations (TurboQuant).
Texas
Shangming Cai
ShangmingCai
Currently working at Alibaba Cloud Apsara Lab.
Research Interests: Efficient LLM serving system.
Alibaba Cloud
Teng Ma
stmatengss
PhD, Tsinghua (16~21); Postdoc, Alibaba (21~23); Staff Engineer, Alibaba (23~present)
Alibaba Group
Kuntai Du
KuntaiDu
Chief Scientist & cofounder @ Tensormesh, Inc.
University of Chicago, Chicago, Illinois
Yihua Cheng
ApostaC
CTO @ TensorMesh;
Core developer at @LMCache and @vllm-project
TensorMesh United States
John
Thespica
PPMC of Apache GraphAr(incubating) | Committer of Apache HugeGraph
@apache Chengdu, China
张林伟
lewiszlw
Database engineer | Apache DataFusion committer | Hobbyist game dev | Creator of @systemxlabs and @NightsWatchGames
@apache Wuhan
Cancai Cai
caicancai
All our current efforts are for the purpose of creating
@aftership | @apache Shenzhen, China
roseduan
roseduan
Stay hungry, stay foolish.
@apache/cloudberry @rosedblabs @lotusdblabs Beijing, China
Alex Chi Z
skyzh
building database systems @databricks, previously @cmu-db; a bubble tea every day makes me happy right away
Databricks 🧱 Bellevue, WA, USA ⇌ Shanghai, China