Smart HTTP router for local LLMs (Ollama, LM Studio, llama.cpp). Rule-based + LLM-powered routing, health checks, load balancing, Prometheus metrics. Rust-native, zero-overhead.
-
Updated
Nov 27, 2025 - Rust
Smart HTTP router for local LLMs (Ollama, LM Studio, llama.cpp). Rule-based + LLM-powered routing, health checks, load balancing, Prometheus metrics. Rust-native, zero-overhead.
A high-level Rust interface for language models powered by the Candle ML framework. It provides ergonomic and efficient APIs for intuitive language model interactions.
Experiments running offline LLMs in Python and Rust locally using Ollama and llama.cpp
ChatLogger is a Python package written in RUST for managing chat conversations with support for branching dialogue, user management, and integration with language models.
A terminal shell built with AI at the core
Nucleus is a modular Rust library for building advanced local AI tools and applications, built-in RAG for private documents, and an extensible plugin system for custom workflows.
Attempt to summarize text from `stdin`, using a large language model (locally and offline), to `stdout`
Add a description, image, and links to the local-llm topic page so that developers can more easily learn about it.
To associate your repository with the local-llm topic, visit your repo's landing page and select "manage topics."