Quantize TinyLlama-1.1B-Chat from PyTorch to CoreML (float16, int8, int4) for efficient on-device inference on iOS 18+.
-
Updated
Nov 12, 2025 - Python
Quantize TinyLlama-1.1B-Chat from PyTorch to CoreML (float16, int8, int4) for efficient on-device inference on iOS 18+.
On-device speech-to-text engine powered by deep learning
On-device wake word detection powered by deep learning
On-device Speech-to-Intent engine powered by deep learning
open-source healthcare ai
Face verification for Flutter using on-device FaceNet-style embeddings (offline TFLite), with local storage and cosine matching.
Real-world context insights SDK for React Native apps, enhancing user experience and engagement with a privacy-first approach
电子鹦鹉 / Toy Language Model
This tool helps you easily deploy ASR models on NPUs on AMD's Ryzen AI 300 series laptops
StratoSort is an AI-powered desktop file organization application built with Tauri and Rust, featuring local AI analysis via Ollama, semantic search across 91+ file types, smart folder automation, and comprehensive testing with 119 API commands.
A light-weight header-mostly Neural Network framework for on-device Inference
Precision genomics for everyone, everywhere. Powered by private AI.
On-device screen understanding for macOS
On-device voice activity detection (VAD) powered by deep learning
On-Device Learning for Human Activity Recognition on Low-Power Microcontrollers
A proof-of-concept app using KeenASR SDK on Android. WE ARE HIRING: https://keenresearch.com/careers.html
This Android app leverages a TensorFlow Lite model for on-device classification of social media posts into 11 categories, including technology, sports, finance, and more. Built with Kotlin and Jetpack Compose, it ensures efficient, privacy-focused inference without server dependencies.
Add a description, image, and links to the on-device topic page so that developers can more easily learn about it.
To associate your repository with the on-device topic, visit your repo's landing page and select "manage topics."