Stars
Complete GRPO implementation for training small language models to reason using reinforcement learning
zipzou / flash-attention
Forked from Dao-AILab/flash-attentionFast and memory-efficient exact attention
Convert PDF to markdown + JSON quickly with high accuracy
Tool to download Blinkist's free offerings, namely "Free Daily" and free curated lists
Wikipedia text corpus for self-supervised NLP model training
MinHash, LSH, LSH Forest, Weighted MinHash, HyperLogLog, HyperLogLog++, LSH Ensemble and HNSW
Convert all of libgen to high quality markdown
Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.
The official repository of the paper "On the Exploitability of Instruction Tuning".
Sea-Snell / JAX_llama
Forked from meta-llama/llamaInference code for LLaMA models in JAX
Deep Learning Hard (DL-HARD) is a new annotated dataset extending TREC Deep Learning benchmark.
Provides a common interface to many IR ranking datasets.
Unofficial JAX implementations of deep learning research papers
All about the fundamental blocks of TF and JAX!
π€ Machine Learning Summer School Guide
Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanations
Reference tables to introduce and organize evaluation methods and measures for explainable machine learning systems
Code for fast dpsgd implementations in JAX/TF
functorch is JAX-like composable function transforms for PyTorch.
π€ The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation tools
A repo to explore different NLP tasks which can be solved using T5
π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support