Lists (1)
Sort Name ascending (A-Z)
Stars
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
A high-throughput and memory-efficient inference and serving engine for LLMs
The Python micro framework for building web applications.
Unified web UI for training and running open models like Qwen, DeepSeek, gpt-oss and Gemma locally.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
TensorFlow code and pre-trained models for BERT
A community-maintained Python framework for creating mathematical animations.
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
You like pytorch? You like micrograd? You love tinygrad! ❤️
Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.
An autonomous agent that conducts deep research on any data using any LLM providers
SGLang is a high-performance serving framework for large language models and multimodal models.
Code for the paper "Language Models are Unsupervised Multitask Learners"
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Open standard for machine learning interoperability
An extremely fast Python type checker and language server, written in Rust.
An orchestration platform for the development, production, and observation of data assets.
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. Tensor…
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
The property-based testing library for Python
Accessible large language models via k-bit quantization for PyTorch.