Stars
Stanford NLP Python library for understanding and improving PyTorch models via interventions
💪 Models' quality and performance metrics (R2, ICC, LOO, AIC, BF, ...)
Materials for my Mixed Model Workshop
OLMost every training recipe you need to perform data interventions with the OLMo family of models.
Attribute (or cite) statements generated by LLMs back to in-context information.
verl: Volcano Engine Reinforcement Learning for LLMs
ICLR 2025 Workshop & CHI 2025 SIG: "Bidirectional Human-AI Alignment"
EMNLP 2025 Two Papers - Value-Action Gap in LLMs (Main Track); ValueCompass (WiNLP Workshop)
(EMNLP 2025) Should I Share this Translation? Evaluating Quality Feedback for User Reliance on Machine Translation
GraphicBench: A Planning Benchmark for Graphic Design Generation with Language Agents
What would you do with 1000 H100s...
⚔️ Official code of "Search Arena: Analyzing Search-Augmented LLMs".
Skywork series models are pre-trained on 3.2TB of high-quality multilingual (mainly Chinese and English) and code data. We have open-sourced the model, training data, evaluation data, evaluation me…
Replication data and code for the paper: When LLMs are Reliable for Judging Empathic Communication
Micromodels -- A framework for accurate, explainable, data efficient, and reusable NLP models.
[ICML 2025] HypotheSAEs: Hypothesizing interpretable relationships in text datasets using sparse autoencoders. https://arxiv.org/abs/2502.04382
Code, data, and models for the ICWSM 2023 paper "Bridging Nations: Quantifying the Role of Multilinguals in Communication on Social Media"
Topic modeling on 100,000 r/AmItheAsshole threads.
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
This repository contains published parts test cross-lingual calls, collected and annotated as part of the InCroMin project (an FSTP under the EU project UTTER.
Code for the arXiv preprint "Answer, Assemble, Ace: Understanding How Transformers Answer Multiple Choice Questions"
LongCite: Enabling LLMs to Generate Fine-grained Citations in Long-context QA
Code for the ICML 2025 paper "SelfCite Self-Supervised Alignment for Context Attribution in Large Language Models"