Starred repositories
BranchFS is a FUSE-based filesystem that provides lightweight, atomic speculative branching on top of any existing filesystem.
Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞
A Claude Code plugin that automatically captures everything Claude does during your coding sessions, compresses it with AI (using Claude's agent-sdk), and injects relevant context back into future …
Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models
WeDLM: The fastest diffusion language model with standard causal attention and native KV cache compatibility, delivering real speedups over vLLM-optimized baselines.
Awesome AI Memory | LLM Memory | A curated knowledge base on AI memory for LLMs and agents, covering long-term memory, reasoning, retrieval, and memory-native system design. Awesome-AI-Memory 是一个 集…
[NeurIPS'24] HippoRAG is a novel RAG framework inspired by human long-term memory that enables LLMs to continuously integrate knowledge across external documents. RAG + Knowledge Graphs + Personali…
A python module to repair invalid JSON from LLMs
Build Real-Time Knowledge Graphs for AI Agents
AI memory OS for LLM and Agent systems(moltbot,clawdbot,openclaw), enabling persistent Skill memory for cross-task skill reuse and evolution.
[EMNLP 2025 Oral] MemoryOS is designed to provide a memory operating system for personalized AI agents.
Universal LLM Deployment Engine with ML Compilation
A low-latency, billion-scale, and updatable graph-based vector store on SSD.
⏰ Agenticly track worldwide conference deadlines (Website, Python Cli, Wechat Applet)
A tool for creating and running Linux containers using lightweight virtual machines on a Mac. It is written in Swift, and optimized for Apple silicon.
Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search
A library for efficient similarity search and clustering of dense vectors.
Community maintained hardware plugin for vLLM on Ascend
Expert Kit is an efficient foundation of Expert Parallelism (EP) for MoE model Inference on heterogenous hardware