Highlights
- Pro
Stars
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
Code and documentation to train Stanford's Alpaca models, and generate the data.
Repository to track the progress in Natural Language Processing (NLP), including the datasets and the current state-of-the-art for the most common NLP tasks.
Simple Reinforcement learning tutorials, 莫烦Python 中文AI教学
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
提供同花顺客户端/miniqmt/雪球的股票量化交易,支持跟踪 joinquant /ricequant 模拟交易 和 实盘雪球组合
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
Accessible large language models via k-bit quantization for PyTorch.
The official repository for ERNIE 4.5 and ERNIEKit – its industrial-grade development toolkit based on PaddlePaddle.
Example models using DeepSpeed
arXiv LaTeX Cleaner: Easily clean the LaTeX code of your paper to submit to arXiv
Model parallel transformers in JAX and Haiku
XLNet: Generalized Autoregressive Pretraining for Language Understanding
An Open-Source Package for Neural Relation Extraction (NRE)
An Open-Source Package for Knowledge Embedding (KE)
🚀 Efficient implementations of state-of-the-art linear attention models
[EMNLP 2021] SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821
Code for "Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks"
Training RNNs as Fast as CNNs (https://arxiv.org/abs/1709.02755)
A Heterogeneous Benchmark for Information Retrieval. Easy to use, evaluate your models across 15+ diverse IR datasets.
Pyserini is a Python toolkit for reproducible information retrieval research with sparse and dense representations.
Shared repository for open-sourced projects from the Google AI Language team.
This repository contains the code for "Exploiting Cloze Questions for Few-Shot Text Classification and Natural Language Inference"
A powerful and flexible machine learning platform for drug discovery
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Source code and dataset for ACL 2019 paper "ERNIE: Enhanced Language Representation with Informative Entities"
A Data Streaming Library for Efficient Neural Network Training