OpenClaw-RL: Train any agent simply by talking
-
Updated
Apr 18, 2026 - Python
OpenClaw-RL: Train any agent simply by talking
A GPU cluster manager that configures and orchestrates inference engines like vLLM and SGLang for high-performance AI model deployment.
MOSS-TTSD is a spoken dialogue generation model designed for expressive multi-speaker synthesis. It features long-context modeling, flexible speaker control, and multilingual support, while enabling zero-shot voice cloning from short audio references.
LLM model quantization (compression) toolkit with HW acceleration support for Nvidia, AMD, Intel GPU and Intel/AMD/Apple CPU via HF, vLLM, and SGLang.
A SOTA quantization algorithm for high-accuracy low-bit LLM inference, seamlessly optimized for CPU/XPU/CUDA, with multi-datatype support and full compatibility with vLLM, SGLang, and Transformers.
MOVA: Towards Scalable and Synchronized Video–Audio Generation
Virtualized Elastic KV Cache for Dynamic GPU Sharing and Beyond
基于SparkTTS、OrpheusTTS等模型,提供高质量中文语音合成与声音克隆服务。
Efficient LLM inference on Slurm clusters.
A tool for benchmarking LLMs on Modal
SGLang model provider for Strands Agents for on-policy agentic RL training.
Automated system for LLM evaluation via agents.
Boosting GPU utilization for LLM serving via dynamic spatial-temporal prefill & decode orchestration
DeepSeek-V3, R1 671B on 8xH100 Throughput Benchmarks
Add a description, image, and links to the sglang topic page so that developers can more easily learn about it.
To associate your repository with the sglang topic, visit your repo's landing page and select "manage topics."