On-device Android AI chat using llama.cpp (Vulkan). Offline GGUF models, RAG search, streaming, Room DB, model catalog, Compose UI.
-
Updated
Dec 6, 2025 - C++
On-device Android AI chat using llama.cpp (Vulkan). Offline GGUF models, RAG search, streaming, Room DB, model catalog, Compose UI.
A portable, offline-first CLI AI assistant built with C++ and llama.cpp. Features dynamic file context awareness, memory management, and runs entirely from a USB drive without dependencies.
Add a description, image, and links to the offline-ai topic page so that developers can more easily learn about it.
To associate your repository with the offline-ai topic, visit your repo's landing page and select "manage topics."