Stars
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
Data and code for paper "M3Exam: A Multilingual, Multimodal, Multilevel Benchmark for Examining Large Language Models"
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
LLM Zoo collects information of various open- and close-sourced LLMs
Panda项目是于2023年5月启动的开源海外中文大语言模型项目,致力于大模型时代探索整个技术栈,旨在推动中文自然语言处理领域的创新和合作。
Semantic cache for LLMs. Fully integrated with LangChain and llama_index.
Code and data accompanying our paper on arXiv "Faithful Chain-of-Thought Reasoning".
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RN…
ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.
Making large AI models cheaper, faster and more accessible
Code repository for supporting the paper "Atlas Few-shot Learning with Retrieval Augmented Language Models",(https//arxiv.org/abs/2208.03299)
Repo for Generating Flashbacks in Stories (NAACL'22)
A modular RL library to fine-tune language models to human preferences
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
This repository implements our EMNLP 2022 research paper A Dataset for Hyper-Relational Extraction and a Cube-Filling Approach.
Code and Data for NeurIPS2021 Paper "A Dataset for Answering Time-Sensitive Questions"
Implementation of Document-level Relation Extraction with Knowledge Distillation and Adaptive Focal Loss
Winner system (DAMO-NLP) of SemEval 2022 MultiCoNER shared task over 10 out of 13 tracks.
[ICLR 2020] Contrastive Representation Distillation (CRD), and benchmark of recent knowledge distillation methods
Code associated with the paper **Two are Better Than One: Joint Entity and Relation Extraction with Table-Sequence Encoders**, at EMNLP 2020
EasyTransfer is designed to make the development of transfer learning in NLP applications easier.
DeLighT: Very Deep and Light-Weight Transformers
Official implementation of the papers "GECToR – Grammatical Error Correction: Tag, Not Rewrite" (BEA-20) and "Text Simplification by Tagging" (BEA-21)
A single model that parses Universal Dependencies across 75 languages. Given a sentence, jointly predicts part-of-speech tags, morphology tags, lemmas, and dependency trees.