-
01:34
(UTC +09:00)
Highlights
- Pro
Stars
This is the repository for llm jp membership inference attack.
A library for mechanistic interpretability of GPT-style language models
Interpretability for sequence generation models 🐛 🔍
Materials for EACL2024 tutorial: Transformer-specific Interpretability
[ICLR'25 Spotlight] Min-K%++: Improved baseline for detecting pre-training data of LLMs
Official Repository for Dataset Inference for LLMs
Python implementation of an extension of the Kolmogorov-Smirnov test for multivariate samples
Python package for measuring memorization in LLMs.
An open-source library for contamination detection in NLP datasets and Large Language Models (LLMs).
The official repository for the paper entitled "Data Contamination Quiz: A Tool to Detect and Estimate Contamination in Large Language Models."
The hub for EleutherAI's work on interpretability and learning dynamics
JFLEG (JHU FLuency-Extended GUG) corpus for Grammatical Error Correction Evaluation
Noise-robust de-duplication at scale
Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.
From Chain-of-Thought prompting to OpenAI o1 and DeepSeek-R1 🍓
A Python 3 compatible version of goose http://goose3.readthedocs.io/en/latest/index.html
This repository contains the files for Syntactic and Semantic Uniformity for Semantic Parsing and Task-Oriented Dialogue Systems
Codebase for SIGIR 2022 paper: Coarse-to-Fine Sparse Sequential Recommendation
Lark is a parsing toolkit for Python, built with a focus on ergonomics, performance and modularity.
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.