Lists (15)
Sort Name ascending (A-Z)
Starred repositories
Achieve state of the art inference performance with modern accelerators on Kubernetes
CUDA Agent: Large-Scale Agentic RL for High-Performance CUDA Kernel Generation
📚 《从零开始构建智能体》——从零开始的智能体原理与实践教程
You like pytorch? You like micrograd? You love tinygrad! ❤️
A list of free LLM inference resources accessible via API.
Write scalable load tests in plain Python 🚗💨
🌐 Make websites accessible for AI agents. Automate tasks online with ease.
A compact implementation of SGLang, designed to demystify the complexities of modern LLM serving systems.
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
cuTile is a programming model for writing parallel kernels for NVIDIA GPUs
A book for Learning the Foundations of LLMs
微舆:人人可用的多Agent舆情分析助手,打破信息茧房,还原舆情原貌,预测未来走向,辅助决策!从0实现,不依赖任何框架。
Official Implementation of EAGLE-1 (ICML'24), EAGLE-2 (EMNLP'24), and EAGLE-3 (NeurIPS'25).
A Datacenter Scale Distributed Inference Serving Framework
Qwen3-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
Move and resize windows on macOS with keyboard shortcuts and snap areas
The 500 AI Agents Projects is a curated collection of AI agent use cases across various industries. It showcases practical applications and provides links to open-source projects for implementation…
Infisical is the open-source platform for secrets, certificates, and privileged access management.
Sync notes between local and cloud with smart conflict: S3 (Amazon S3/Cloudflare R2/Backblaze B2/...), Dropbox, webdav (NextCloud/InfiniCLOUD/Synology/...), OneDrive, Google Drive (GDrive), Box, pC…
小红书笔记 | 评论爬虫、抖音视频 | 评论爬虫、快手视频 | 评论爬虫、B 站视频 | 评论爬虫、微博帖子 | 评论爬虫、百度贴吧帖子 | 百度贴吧评论回复爬虫 | 知乎问答文章|评论爬虫
Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.
A high-throughput and memory-efficient inference and serving engine for LLMs
SGLang is a high-performance serving framework for large language models and multimodal models.
zero-peak / ZeroOmega
Forked from FelisCatus/SwitchyOmegaManage and switch between multiple proxies quickly & easily.