Highlights
- Pro
Stars
Master programming by recreating your favorite technologies from scratch.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
FastAPI framework, high performance, easy to learn, fast to code, ready for production
scikit-learn: machine learning in Python
A high-throughput and memory-efficient inference and serving engine for LLMs
Python ETL framework for stream processing, real-time analytics, LLM pipelines, and RAG.
Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more
Ready-to-run cloud templates for RAG, AI pipelines, and enterprise search with live data. 🐳Docker-friendly.⚡Always in sync with Sharepoint, Google Drive, S3, Kafka, PostgreSQL, real-time data APIs,…
TensorFlow code and pre-trained models for BERT
all of the workflows of n8n i could find (also from the site itself)
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
The fundamental package for scientific computing with Python.
Conductor is an event driven orchestration platform providing durable and highly resilient execution engine for your applications
Daytona is a Secure and Elastic Infrastructure for Running AI-Generated Code
Composio equips your AI agents & LLMs with 100+ high-quality integrations via function calling
In-depth tutorials on LLMs, RAGs and real-world AI agent applications.
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
Unified framework for building enterprise RAG pipelines with small, specialized models
Ongoing research training transformer models at scale
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. Tensor…
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
A Datacenter Scale Distributed Inference Serving Framework
A fast type checker and language server for Python
100+ Fine-tuning Tutorial Notebooks on Google Colab, Kaggle and more.