Stars
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
PyTorch package for the discrete VAE used for DALL·E.
A framework for few-shot evaluation of language models.
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
A debugging and profiling tool that can trace and visualize python code execution
Model parallel transformers in JAX and Haiku
Sacred is a tool to help you configure, organize, log and reproduce experiments developed at IDSIA.
Simple, elegant, Pythonic functional programming.
A simple and efficient tool to parallelize Pandas operations on all available CPUs
Python 3.9 to JavaScript compiler - Lean, fast, open!
Open-AI's DALL-E for large scale training in mesh-tensorflow.
Keeping language models honest by directly eliciting knowledge encoded in their activations.
Using queues, tqdm-multiprocess supports multiple worker processes, each with multiple tqdm progress bars, displaying them cleanly through the main process. It offers similar functionality for pyth…
Downloads 2020 English Wikipedia articles as plaintext
URL downloader supporting checkpointing and continuous checksumming.
downloads and parses subtitle dataset from opensubtitles.org
Download, parse, and filter data from Court Listener, part of the FreeLaw projects. Data-ready for The-Pile.
Evaluation of measurement tampering detection techniques on the datasets from Benchmarks for Detecting Measurement Tampering
a small library for combinatorial iters of dicts, useful for config/hyperparameter sweep management
Script/utility that tests sample cases from Kattis locally, built for UAPSC.