Skip to content
#

llm-security

Here are 99 public repositories matching this topic...

MalPromptSentinel (MPS) is a Claude Code skill that detects malicious prompts in uploaded files before Claude processes them. It provides two-tier scanning to identify prompt injection attacks, role manipulation attempts, privilege escalation, and other adversarial techniques.

  • Updated Nov 27, 2025
  • Python

A cross-provider AI model security scanner that evaluates HuggingFace, OpenRouter, and Ollama models for malicious content, unsafe code, license issues, and known vulnerabilities. Includes automated reports and risk scoring.

  • Updated Nov 26, 2025
  • Python

Bidirectional LLM security firewall providing risk reduction (not complete protection) for human/LLM interfaces. Hexagonal architecture with multi-layer validation of inputs, outputs, memory and tool state. Beta status. ~528 KB wheel, optional ML guards.

  • Updated Dec 13, 2025
  • Python

Improve this page

Add a description, image, and links to the llm-security topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the llm-security topic, visit your repo's landing page and select "manage topics."

Learn more