-
Mahō Shōjo特別作戦班です
- Beijing
-
22:00
(UTC -12:00)
Lists (17)
Sort Name ascending (A-Z)
AI Agent
Alibaba
AMD YES!
Codeing
eBPF
Fault tolerance
FPGA
Inference LLM
Kernel SHM
LLM Inference
LLM Training System
LLM Training SystemMagic book
Magic bookNetWork Tech
NetWork TechOperation and maintenance
Operation and maintenanceRLHF
Robot
Tutorial
Stars
Open-source platform to build and deploy AI agent workflows.
My learning notes for ML SYS.
ncnn is a high-performance neural network inference framework optimized for the mobile platform
Fast and memory-efficient exact attention
A privacy-first, self-hosted, fully open source personal knowledge management software, written in typescript and golang.
eBPF-based Networking, Security, and Observability
A feedback-driven fault injection tool for reproducing distributed systems failures
Automated Testing and Adaptive Detection of **Slow Faults** in Distributed Systems
NVSHMEM‑Tutorial: Build a DeepEP‑like GPU Buffer
"AI-Trader: Can AI Beat the Market?" Live Trading Bench: https://ai4trade.ai Tech Report Link: https://arxiv.org/abs/2512.10971
Efficient Distributed GPU Programming for Exascale, an SC/ISC Tutorial
Zero instrucment LLM and AI agent (e.g. claude code, gemini-cli) observability in eBPF
Build your personal knowledge base with Trilium Notes
NVSentinel is a cross-platform fault remediation service designed to rapidly remediate runtime node-level issues in GPU-accelerated computing environments
Userspace/GPU eBPF VM with llvm JIT/AOT compiler
A high-performance agent for collecting NVIDIA GPU metrics and exporting them via OpenTelemetry Arrow protocol.
PyTorch native quantization and sparsity for training and inference
This is a tool for managing GPU partitions for NVIDIA Fabric Manager’s Shared NVSwitch.
Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible.
A service-aware RoCE network monitoring system based on end- to-end probing.
Standardized Distributed Generative and Predictive AI Inference Platform for Scalable, Multi-Framework Deployment on Kubernetes
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
Supercharge Your LLM with the Fastest KV Cache Layer
💖🧸 Self hosted, you owned Grok Companion, a container of souls of waifu, cyber livings to bring them into our worlds, wishing to achieve Neuro-sama's altitude. Capable of realtime voice chat, Minec…
AI Inference Operator for Kubernetes. The easiest way to serve ML models in production. Supports VLMs, LLMs, embeddings, and speech-to-text.
Achieve state of the art inference performance with modern accelerators on Kubernetes