Skip to content
View chawins's full-sized avatar

Highlights

  • Pro

Organizations

@wagner-group

Block or report chawins

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Showing results

A research workbench for developing and testing attacks against large language models, with a focus on prompt injection vulnerabilities and defenses.

Python 47 17 Updated Apr 2, 2026
Python 32 9 Updated Sep 11, 2025

dInfer: An Efficient Inference Framework for Diffusion Language Models

Python 449 42 Updated Feb 11, 2026

Flexible and Pluggable Serving Engine for Diffusion LLMs

Python 68 13 Updated Mar 31, 2026
Python 10 Updated Nov 29, 2025

Gemma open-weight LLM library, from Google DeepMind

Python 4,189 708 Updated Apr 3, 2026

The best ChatGPT that $100 can buy.

Python 50,914 6,700 Updated Mar 27, 2026

[EMNLP 2025 Oral] IPIGuard: A Novel Tool Dependency Graph-Based Defense Against Indirect Prompt Injection in LLM Agents

Python 18 3 Updated Sep 16, 2025

Curated resources, research, and tools for securing AI systems

655 112 Updated Apr 1, 2026

gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI

Python 19,966 2,064 Updated Mar 27, 2026

Renderer for the harmony response format to be used with gpt-oss

Rust 4,251 264 Updated Mar 27, 2026

Repo for the paper "Meta SecAlign: A Secure Foundation LLM Against Prompt Injection Attacks".

Python 58 18 Updated Apr 1, 2026

Code for the paper "Defeating Prompt Injections by Design"

Jupyter Notebook 300 43 Updated Jun 20, 2025

[ICLR 2025] Dissecting adversarial robustness of multimodal language model agents

Python 136 11 Updated Feb 19, 2025

Open-source implementation of AlphaEvolve

Python 5,834 932 Updated Mar 18, 2026

Official PyTorch implementation for "Large Language Diffusion Models"

Python 3,709 254 Updated Nov 12, 2025

OO for LLMs

Python 902 72 Updated Apr 3, 2026

Dataset and code for "JailbreaksOverTime: Detecting Jailbreak Attacks Under Distribution Shift"

Jupyter Notebook 10 1 Updated Apr 24, 2025

Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)

Python 162 14 Updated Nov 30, 2024

Repo for the research paper "SecAlign: Defending Against Prompt Injection with Preference Optimization"

Python 94 9 Updated Jul 24, 2025
Python 11 Updated Mar 22, 2025

A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.

Python 510 130 Updated Mar 30, 2026

A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).

1,922 133 Updated Apr 2, 2026

A Survey on Jailbreak Attacks and Defenses against Multimodal Generative Models

314 13 Updated Jan 11, 2026

A data augmentations library for audio, image, text, and video.

Python 5,070 311 Updated Mar 31, 2026

Fast near-duplicate matching is a method for quickly finding near-duplicate spans in a document by utilizing the Rabin-Karp algorithm.

Rust 2 Updated Sep 22, 2024
Python 1 Updated Jun 7, 2024

The Security Toolkit for LLM Interactions

Python 2,771 368 Updated Dec 15, 2025

LLM Prompt Injection Detector

TypeScript 1,455 132 Updated Aug 7, 2024

Every practical and proposed defense against prompt injection.

668 49 Updated Feb 22, 2025
Next