Starred repositories
The official implemention of "TreeRPO: Tree Relative Policy Optimization"
[AAAI'26, Oral] Code for "Teaching Large Language Models to Maintain Contextual Faithfulness via Synthetic Tasks and Reinforcement Learning"
[ACL'25] Code for "Aligning Large Language Models to Follow Instructions and Hallucinate Less via Effective Data Filtering"
[EMNLP'25, SAC Highlights Award] Code for "GATEAU: Selecting Influential Samples for Long Context Alignment"
[EMNLP 2025] Expanding before Inferring: Enhancing Factuality in Large Language Models through Premature Layers Interpolation
Agent-R1: Training Powerful LLM Agents with End-to-End Reinforcement Learning
verl-agent is an extension of veRL, designed for training LLM/VLM agents via RL. verl-agent is also the official code for paper "Group-in-Group Policy Optimization for LLM Agent Training"
🧑🚀 全世界最好的LLM资料总结(多模态生成、Agent、辅助编程、AI审稿、数据处理、模型训练、模型推理、o1 模型、MCP、小语言模型、视觉语言模型) | Summary of the world's best LLM resources.
Train your Agent model via our easy and efficient framework
verl: Volcano Engine Reinforcement Learning for LLMs
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & TIS & vLLM & Ray & Dynamic Sampling & Async Agentic RL)
《EasyOffer》(<大模型面经合集>)是针对LLM宝宝们量身打造的大模型暑期实习Offer指南,主要记录大模型暑期实习和秋招准备的一些常见大厂手撕代码、大厂面经经验、常见大厂思考题等;小白一个,正在学习ing......有问题各位大佬随时指正,希望大家都能拿到心仪Offer!
[NAACL'25 Oral] Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering
[ACL-2024]Enhancing Noise Robustness of Retrieval-Augmented Language Models with Adaptive Adversarial Training
《动手学大模型Dive into LLMs》系列编程实践教程
Mastering Transformers, published by Packt
[EMNLP 2024] To Forget or Not? Towards Practical Knowledge Unlearning for Large Language Models
Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"
A package to evaluate factuality of long-form generation. Original implementation of our EMNLP 2023 paper "FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation"
Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large Language Models"
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights i…
Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"
Official implementation of our paper "Separate the Wheat from the Chaff: Model Deficiency Unlearning via Parameter-Efficient Module Operation". A model merge method for deficiency unlearning, compi…
Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"