- Germany
-
00:04
(UTC +01:00) - https://www.thinktecture.com/christian-weyer
- @christianweyer
Stars
Docker configuration for running VLLM on dual DGX Sparks
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.
The TypeScript AI agent framework. ⚡ Assistants, RAG, observability. Supports any LLM: GPT-4, Claude, Gemini, Llama.
An open-source AI Voice Agent that integrates with Asterisk/FreePBX using Audiosocket/RTP technology
Efficient few-shot learning with Sentence Transformers
Real-time Visualizer for Neural Networks
Collection of step-by-step playbooks for setting up AI/ML workloads on NVIDIA DGX Spark devices with Blackwell architecture.
🌐 The open-source Agentic browser; privacy-first alternative to ChatGPT Atlas, Perplexity Comet, Dia.
Examples, end-2-end tutorials and apps built using Liquid AI Foundational Models (LFM) and the LEAP SDK
Learn to build and deploy local Visual Language Models for Edge AI
RamaLama is an open-source developer tool that simplifies the local serving of AI models from any source and facilitates their use for inference in production, all through the familiar language of …
Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.
FastAPI + MLX offline-first voice agent with <1s latency. Minimal UI
This course is designed to guide beginners through the exciting world of Edge AI, covering fundamental concepts, popular models, inference techniques, device-specific applications, model optimizati…
Awesome LLM speech-to-speech models and frameworks
Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audio codec.
Kyutai's Speech-To-Text and Text-To-Speech models based on the Delayed Streams Modeling framework.
This prototype shows how to build a local Retrieval-Augmented Generation (RAG)
CLI-based tester for verifying that MCP servers work correctly when called directory and by agents
AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.
LLM agents built for control. Designed for real-world use. Deployed in minutes.
ollama like cli tool for MLX models on huggingface (pull, rm, list, show, serve etc.)
A high-performance API server that provides OpenAI-compatible endpoints for MLX models. Developed using Python and powered by the FastAPI framework, it provides an efficient, scalable, and user-fri…
The official C# SDK for Model Context Protocol servers and clients. Maintained in collaboration with Microsoft.