🐢 Open-Source Evaluation & Testing for AI & LLM systems
-
Updated
Feb 14, 2025 - Python
🐢 Open-Source Evaluation & Testing for AI & LLM systems
the LLM vulnerability scanner
The Security Toolkit for LLM Interactions
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
An easy-to-use Python framework to generate adversarial jailbreak prompts.
Papers and resources related to the security and privacy of LLMs 🤖
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses
The fastest && easiest LLM security guardrails for CX AI Agents and applications.
Framework for LLM evaluation, guardrails and security
Ultra-fast, low latency LLM prompt injection/jailbreak detection ⛓️
Framework for testing vulnerabilities of large language models (LLM).
An Execution Isolation Architecture for LLM-Based Agentic Systems
A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.
Experimental tools to backdoor large language models by re-writing their system prompts at a raw parameter level. This allows you to potentially execute offline remote code execution without running any actual code on the victim's machine or thwart LLM-based fraud/moderation systems.
Whispers in the Machine: Confidentiality in LLM-integrated Systems
Risks and targets for assessing LLMs & LLM vulnerabilities
Guard your LangChain applications against prompt injection with Lakera ChainGuard.
Vulnerable LLM Application
LLM Security Project with Llama Guard
Add a description, image, and links to the llm-security topic page so that developers can more easily learn about it.
To associate your repository with the llm-security topic, visit your repo's landing page and select "manage topics."