Stars
🔊 Text-Prompted Generative Audio Model
A guidance language for controlling large language models.
Zero-Shot Speech Editing and Text-to-Speech in the Wild
AirLLM 70B inference with single 4GB GPU
Neo4j graph construction from unstructured data using LLMs
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
A library of data loaders for LLMs made by the community -- to be used with LlamaIndex and/or LangChain
Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads
Apache Hamilton helps data scientists and engineers define testable, modular, self-documenting dataflows, that encode lineage/tracing and metadata. Runs and scales everywhere python does.
A comprehensive guide to building RAG-based LLM applications for production.
A collection of open-source GPU accelerated Python tools and examples for quantitative analyst tasks and leverages RAPIDS AI project, Numba, cuDF, and Dask.
Training and inference on AWS Trainium and Inferentia chips.
Graph based retrieval + GenAI = Better RAG in production
Example code for AWS Neuron SDK developers building inference and training applications
Active WIP for experimenting with GraphRAG and Knowledge Graphs
Adding NeMo Guardrails to a LlamaIndex RAG pipeline
The code for LexDrafter framework: a framework that assists in drafting Definitions articles for legislative documents using retrieval augmented generation (RAG) and existing term definitions prese…