Stars
4
results
for sponsorable starred repositories
Clear filter
https://wavespeed.ai/ Context parallel attention that accelerates DiT model inference with dynamic caching
A high-throughput and memory-efficient inference and serving engine for LLMs