An AI-powered AR glasses development platform with real-time computer vision, object detection, and AI agent capabilities.
Location: snap/snap/Assets/CoordinateFetcher.ts
- File: TypeScript component for Lens Studio
- Function: Captures camera frames from AR glasses and sends to server
- To add new UI: Edit
CoordinateFetcher.tsin thedisplayLabel()method (lines 124-151) to modify visual markers and positioning
Location: Multiple files in root directory
ai_client.py: Gemini Live API and OpenRouter integrationstream_to_gemini_live(): Real-time streaming with Geminisend_to_openrouter(): Multi-model support via OpenRouter
ollama_client.py: Local Ollama integrationget_ollama_response(): Vision-capable models (llava, llama3.2-vision)
grounding.py: YOLO-based object detectionGrounding.detect(): Run detection on framesGrounding.update_prompt(): Modify detection classes dynamically
To add new AI tools: Create functions in ai_client.py or
ollama_client.py following existing patterns
serve.py: WebSocket server for AR glasses (port 5001)app.py: Webcam webapp (port 5001)- Camera feed display
- Frame analysis with AI models
# Install dependencies
uv sync
# Or with pip
pip install -r requirements.txtpython main.pypython app.py
# Visit http://localhost:5001- Python 3.13+
- Snap Lens Studio (for AR glasses development)
- Ollama (optional, for local AI models)
- OpenRouter API key (optional, for cloud AI models)
- Google API key (optional, for Gemini Live)
GOOGLE_API_KEY: For Gemini Live APIOPENROUTER_API_KEY: For OpenRouter multi-model access