Stars
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
PyTorch package for the discrete VAE used for DALL·E.
A framework for few-shot evaluation of language models.
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
A debugging and profiling tool that can trace and visualize python code execution
Model parallel transformers in JAX and Haiku
Athens is no longer maintainted. Athens was an open-source, collaborative knowledge graph, backed by YC W21
Transforms PDF, Documents and Images into Enriched Structured Data
Sacred is a tool to help you configure, organize, log and reproduce experiments developed at IDSIA.
Simple, elegant, Pythonic functional programming.
A simple and efficient tool to parallelize Pandas operations on all available CPUs
Python 3.9 to JavaScript compiler - Lean, fast, open!
Open-AI's DALL-E for large scale training in mesh-tensorflow.
Keeping language models honest by directly eliciting knowledge encoded in their activations.
EleutherAI / DeeperSpeed
Forked from deepspeedai/DeepSpeedDeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
A Python library for integrating model-based and judgmental forecasting
A dataset of alignment research and code to reproduce it