Stars
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
All Algorithms implemented in Python
Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"
Underthesea - Vietnamese NLP Toolkit
The official repository for ERNIE 4.5 and ERNIEKit – its industrial-grade development toolkit based on PaddlePaddle.
Official Implementation of EAGLE-1 (ICML'24), EAGLE-2 (EMNLP'24), and EAGLE-3 (NeurIPS'25).
Automatically Discovering Fast Parallelization Strategies for Distributed Deep Neural Network Training
Accessible large language models via k-bit quantization for PyTorch.
Google Research
中文 NLP 预处理、解析工具包,准确、高效、易用 A Chinese NLP Preprocessing & Parsing Package www.jionlp.com
An open source library for deep learning end-to-end dialog systems and chatbots.
Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
Bolt is a deep learning library with high performance and heterogeneous flexibility.
Easy and Efficient Transformer : Scalable Inference Solution For Large NLP model
[EMNLP 2021] SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821
xk-time 是时间转换,时间计算,时间格式化,时间解析,日历,时间cron表达式和时间NLP等的工具,使用Java8(JSR-310),线程安全,简单易用,多达70几种常用日期格式化模板,支持Java8时间类和Date,轻量级,无第三方依赖。
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
TensorFlow Tutorial and Examples for Beginners (support TF v1 & v2)
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads
GraphVite: A General and High-performance Graph Embedding System
An annotated implementation of the Transformer paper.
Transformer related optimization, including BERT, GPT
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
My blogs and code for machine learning. http://cnblogs.com/pinard
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.