-
University of Oxford
Stars
Infer Ring is an iOS and macOS app that facilitates cross-device LLM inference using MLX
Archive: Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚
⚙️ A macOS DMG package builder with a native GUI and CLI support
Pretraining and inference code for a large-scale depth-recurrent language model
Artificial Neural Engine Machine Learning Library
Vane is an AI-powered answering engine.
A systematic reasoning MCP server implementation for Claude Desktop with beam search and thought evaluation.
EXO Gym is an open-source Python toolkit that facilitates distributed AI research.
exo-explore / llama98.c
Forked from karpathy/llama2.cInference Llama models in one file of pure C for Windows 98 running on 25-year-old hardware
AI wearables. Put it on, speak, transcribe, automatically
Sidecar is the AI brains for the Aide editor and works alongside it, locally on your machine
Official inference framework for 1-bit LLMs
A high-throughput and memory-efficient inference and serving engine for LLMs
The AI-native database built for LLM applications, providing incredibly fast hybrid search of dense vector, sparse vector, tensor (multi-vector), and full-text.
Galadriel TEE oracle configuration and verification code [Deprecated]
Generate accurate transcripts using Apple's MLX framework
You like pytorch? You like micrograd? You love tinygrad! ❤️
Action to enable running Vulkan apps on GitHub runners
llama3 implementation one matrix multiplication at a time