Local LLM proxy, DevOps friendly
-
Updated
Dec 14, 2025 - Go
Local LLM proxy, DevOps friendly
EduSphere is an AI-powered academic assistant that turns raw transcripts into personalized insights, course paths, and scholarships — powered by local LLM inference. Built with a Golang Fiber backend and React (Vite) frontend, it integrates Generative AI reasoning, natural language interaction, and real-world data into a cohesive, production-grade.
🚀 Enterprise-grade AI coding assistant with local AI processing, GitHub integration, and web scraping capabilities. Built with Go and powered by Ollama.
Local agentic CLI coding assistant
Kubernetes operator for GPU-accelerated LLM inference - air-gapped, edge-native, production-ready
AI-Native Autoscaler for Docker Compose — built with cagent + MCP + Model Runner.
Read PGS (Bluray) and VobSub (DVD) image subtitles and extract their text using external Vision Language Models.
A high-performance task queue management system for Ollama models, built in Go.
Stop paying for AI APIs during development. LocalCloud runs everything locally - GPT-level models, databases, all free.
🤖 The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. Features: Generate Text, Audio, Video, Images, Voice Cloning, Distributed, P2P inference
Nous: A privacy-focused personal knowledge assistant using local LLMs to securely interact with your documents and enhance information retrieval.
Add a description, image, and links to the local-llm topic page so that developers can more easily learn about it.
To associate your repository with the local-llm topic, visit your repo's landing page and select "manage topics."