Strix Halo local LLM guide: 63-97 t/s direct MoE on Ryzen AI MAX+ 395 / 128GB unified memory. Setup, model choices, benchmarks, and raw evidence.
-
Updated
May 10, 2026 - Python
Strix Halo local LLM guide: 63-97 t/s direct MoE on Ryzen AI MAX+ 395 / 128GB unified memory. Setup, model choices, benchmarks, and raw evidence.
A production-grade Python SDK for the Lemonade LLM backend. Featuring auto-discovery, port scanning, and a low-overhead client architecture. Powering the Sorana AI workspace.
Sample MCP Server for weather queries - Reference implementation for ASUS and OEM partners
Add a description, image, and links to the ai-pc topic page so that developers can more easily learn about it.
To associate your repository with the ai-pc topic, visit your repo's landing page and select "manage topics."