Local LLM proxy, DevOps friendly
-
Updated
Dec 14, 2025 - Go
Local LLM proxy, DevOps friendly
Kubernetes operator for GPU-accelerated LLM inference - air-gapped, edge-native, production-ready
🚀 Enterprise-grade AI coding assistant with local AI processing, GitHub integration, and web scraping capabilities. Built with Go and powered by Ollama.
Stop paying for AI APIs during development. LocalCloud runs everything locally - GPT-level models, databases, all free.
A high-performance task queue management system for Ollama models, built in Go.
🤖 The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. Features: Generate Text, Audio, Video, Images, Voice Cloning, Distributed, P2P inference
AI-Native Autoscaler for Docker Compose — built with cagent + MCP + Model Runner.
Read PGS (Bluray) and VobSub (DVD) image subtitles and extract their text using external Vision Language Models.
Nous: A privacy-focused personal knowledge assistant using local LLMs to securely interact with your documents and enhance information retrieval.
Local agentic CLI coding assistant
EduSphere is an AI-powered academic assistant that turns raw transcripts into personalized insights, course paths, and scholarships — powered by local LLM inference. Built with a Golang Fiber backend and React (Vite) frontend, it integrates Generative AI reasoning, natural language interaction, and real-world data into a cohesive, production-grade.
Add a description, image, and links to the local-llm topic page so that developers can more easily learn about it.
To associate your repository with the local-llm topic, visit your repo's landing page and select "manage topics."