Skip to content
View chawins's full-sized avatar

Highlights

  • Pro

Organizations

@wagner-group

Block or report chawins

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Showing results

A research workbench for developing and testing attacks against large language models, with a focus on prompt injection vulnerabilities and defenses.

Python 23 8 Updated Dec 22, 2025
Python 21 9 Updated Sep 11, 2025

dInfer: An Efficient Inference Framework for Diffusion Language Models

Python 369 35 Updated Dec 23, 2025

Flexible and Pluggable Serving Engine for Diffusion LLMs

Python 45 8 Updated Dec 22, 2025
Python 9 Updated Nov 29, 2025

Gemma open-weight LLM library, from Google DeepMind

Python 3,903 608 Updated Nov 18, 2025

The best ChatGPT that $100 can buy.

Python 39,130 4,954 Updated Dec 9, 2025

[EMNLP 2025 Oral] IPIGuard: A Novel Tool Dependency Graph-Based Defense Against Indirect Prompt Injection in LLM Agents

Python 16 1 Updated Sep 16, 2025

Curated resources, research, and tools for securing AI systems

276 41 Updated Dec 14, 2025

gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI

Python 19,461 1,999 Updated Nov 1, 2025

Renderer for the harmony response format to be used with gpt-oss

Rust 4,088 240 Updated Dec 15, 2025

Repo for the paper "Meta SecAlign: A Secure Foundation LLM Against Prompt Injection Attacks".

Python 37 10 Updated Dec 16, 2025

Code for the paper "Defeating Prompt Injections by Design"

Jupyter Notebook 185 26 Updated Jun 20, 2025

[ICLR 2025] Dissecting adversarial robustness of multimodal language model agents

Python 120 7 Updated Feb 19, 2025

Open-source implementation of AlphaEvolve

Python 4,952 763 Updated Dec 23, 2025

Official PyTorch implementation for "Large Language Diffusion Models"

Python 3,425 230 Updated Nov 12, 2025

OO for LLMs

Python 883 70 Updated Dec 18, 2025

Dataset and code for "JailbreaksOverTime: Detecting Jailbreak Attacks Under Distribution Shift"

Jupyter Notebook 8 1 Updated Apr 24, 2025

Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)

Python 156 12 Updated Nov 30, 2024

Repo for the research paper "SecAlign: Defending Against Prompt Injection with Preference Optimization"

Python 76 7 Updated Jul 24, 2025
Python 11 Updated Mar 22, 2025

A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.

Python 386 94 Updated Dec 1, 2025

A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).

1,804 114 Updated Dec 23, 2025

A Survey on Jailbreak Attacks and Defenses against Multimodal Generative Models

291 12 Updated Nov 25, 2025

A data augmentations library for audio, image, text, and video.

Python 5,067 310 Updated Oct 31, 2025

Fast near-duplicate matching is a method for quickly finding near-duplicate spans in a document by utilizing the Rabin-Karp algorithm.

Rust 2 Updated Sep 22, 2024
Python 1 Updated Jun 7, 2024

The Security Toolkit for LLM Interactions

Python 2,343 315 Updated Dec 15, 2025

LLM Prompt Injection Detector

TypeScript 1,390 121 Updated Aug 7, 2024

Every practical and proposed defense against prompt injection.

597 39 Updated Feb 22, 2025
Next