Stars
A CLI to estimate inference memory requirements for Hugging Face models, written in Python.
A Claude Code plugin that automatically captures everything Claude does during your coding sessions, compresses it with AI (using Claude's agent-sdk), and injects relevant context back into future β¦
A bunch of links to blog posts, articles, videos, etc for learning Rust
A Datacenter Scale Distributed Inference Serving Framework
Magical Rust patterns laid out and simplified
A launch point for your personal nvim configuration
Curated list of project-based tutorials
A simple and easy-to-use library to enjoy videogames programming
π₯ Blazing fast terminal file manager written in Rust, based on async I/O.
Pretty fancy and modern terminal file manager
Produce redistributable builds of Python
A community-maintained Python framework for creating mathematical animations.
CTF framework and exploit development library
Exploit Development and Reverse Engineering with GDB & LLDB Made Easy
Triton CLI is an open source command line interface that enables users to create, deploy, and profile models served by the Triton Inference Server.
Seamless operability between C++11 and Python
An open-source cross-platform alternative to AirDrop
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
πA curated list of Awesome LLM/VLM Inference Papers with Codes: Flash-Attention, Paged-Attention, WINT8/4, Parallelism, etc.π
The Patterns of Scalable, Reliable, and Performant Large-Scale Systems
A curated list of awesome C++ (or C) frameworks, libraries, resources, and shiny things. Inspired by awesome-... stuff.
π Awesome lists about all kinds of interesting topics
Awesome-LLM: a curated list of Large Language Model