Stop paying for AI APIs during development. LocalCloud runs everything locally - GPT-level models, databases, all free.
-
Updated
Jul 8, 2025 - Go
Stop paying for AI APIs during development. LocalCloud runs everything locally - GPT-level models, databases, all free.
Kubernetes operator for GPU-accelerated LLM inference - air-gapped, edge-native, production-ready
Local LLM proxy, DevOps friendly
Nous: A privacy-focused personal knowledge assistant using local LLMs to securely interact with your documents and enhance information retrieval.
AI-Native Autoscaler for Docker Compose — built with cagent + MCP + Model Runner.
Read PGS (Bluray) and VobSub (DVD) image subtitles and extract their text using external Vision Language Models.
A high-performance task queue management system for Ollama models, built in Go.
🤖 The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. Features: Generate Text, Audio, Video, Images, Voice Cloning, Distributed, P2P inference
🚀 Enterprise-grade AI coding assistant with local AI processing, GitHub integration, and web scraping capabilities. Built with Go and powered by Ollama.
Local agentic CLI coding assistant
EduSphere is an AI-powered academic assistant that turns raw transcripts into personalized insights, course paths, and scholarships — powered by local LLM inference. Built with a Golang Fiber backend and React (Vite) frontend, it integrates Generative AI reasoning, natural language interaction, and real-world data into a cohesive, production-grade.
Add a description, image, and links to the local-llm topic page so that developers can more easily learn about it.
To associate your repository with the local-llm topic, visit your repo's landing page and select "manage topics."