Security scanner for code and logs of AI-powered applications
-
Updated
Jul 14, 2025 - Python
Security scanner for code and logs of AI-powered applications
Test your LLM system prompts against 287 real-world attack vectors including prompt injection, jailbreaks, and data leaks.
Comprehensive LLM AI Model protection - cybersecurity toolset aligned to addressing OWASP vulnerabilities - https://genai.owasp.org/llm-top-10/
Stop prompt injections in 20ms. The safety toolkit every LLM app needs. No API keys, no complex setup, just `pip install llm-guard` and you're protected.
Sovereign AI Infrastructure for Enterprise RAG. Zero-Trust PII Sanitization, Local Inference (CPU-optimized), and Docker-ready architecture.
Trained Without My Consent (TraWiC): Detecting Code Inclusion In Language Models Trained on Code
Simulating prompt injection and guardrail bypass across chained LLMs in security decision pipelines.
Security scanner for LLM/RAG applications - Test for prompt injection, jailbreaks, PII leakage, hallucinations & more
PromptSniffer is a security auditing tool designed for authorized penetration testing and corporate DLP monitoring. It captures and monitors prompts sent to Large Language Models (ChatGPT, Claude, Gemini, etc.) across your entire network, providing real-time email alerts and comprehensive logging.
Open-source enforcement layer for LLM safety and governance — ingress/egress evaluation, policy packs, verifier support, and multimodal protection.
A research framework for implementing and evaluating poisoning attacks on Retrieval-Augmented Generation (RAG) systems, enabling the study of their security vulnerabilities.
Evaluating LLM Robustness with Manipulated Prompts
A research framework for simulating, detecting, and defending against backdoor loop attacks in LLM-based multi-agent systems.
PromptMe is an educational project that showcases security vulnerabilities in large language models (LLMs) and their web integrations. It includes 10 hands-on challenges inspired by the OWASP LLM Top 10, demonstrating how these vulnerabilities can be discovered and exploited in real-world scenarios.
CLI tool that uses the Lakera API to perform security checks in LLM inputs
This repo focus on how to deal with prompt injection problem faced by LLMs
MalPromptSentinel (MPS) is a Claude Code skill that detects malicious prompts in uploaded files before Claude processes them. It provides two-tier scanning to identify prompt injection attacks, role manipulation attempts, privilege escalation, and other adversarial techniques.
Add a description, image, and links to the llm-security topic page so that developers can more easily learn about it.
To associate your repository with the llm-security topic, visit your repo's landing page and select "manage topics."