Use your locally running AI models to assist you in your web browsing
-
Updated
May 10, 2026 - TypeScript
Use your locally running AI models to assist you in your web browsing
Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc
A generalized information-seeking agent system with Large Language Models (LLMs).
Steno is the AI powered intelligence layer for all your confidential conversations. Capture beautiful notes whilst keeping your data confidential. Perfect for government, defence, legal and CXOs.
[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
[NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
A flexible, AI powered C2 framework built with operators in mind
Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs
Fenix Ai Trading Bot with LangGraph and ollama and multipe providers
Unified management and routing for llama.cpp, MLX and vLLM models with web dashboard.
A Cli, a webUI, and a MCP server for the Z-Image-Turbo text-to-image generation model (Tongyi-MAI/Z-Image-Turbo base model as well as quantized models)
The PyVisionAI Official Repo
MVP of an idea using multiple local LLM models to simulate and play D&D
Open-source AI IDE powered by local & cloud LLMs. A privacy-first alternative to Cursor.
Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on different ports and loading/unloading them on demand
Vesta macOS Distribution - Official releases and downloads.Vesta AI Chat Assistant for macOS - Built with SwiftUI, Swift MLX and Apple Intelligence using Apple's on device model on MacOs Tahoe (MacOS 26). Now with side-by-side Qwen3-VL for vison
Add a description, image, and links to the localllm topic page so that developers can more easily learn about it.
To associate your repository with the localllm topic, visit your repo's landing page and select "manage topics."