Skip to content
#

llm-security

Here are 186 public repositories matching this topic...

🛡️ Explore tools for securing Large Language Models, uncovering their strengths and weaknesses in the realm of offensive and defensive security.

  • Updated Dec 14, 2025

MER is a software that identifies and highlights manipulative communication in text from human conversations and AI-generated responses. MER benchmarks language models for manipulative expressions, fostering development of transparency and safety in AI. It also supports manipulation victims by detecting manipulative patterns in human communication.

  • Updated Dec 13, 2025
  • HTML

Bidirectional LLM security firewall providing risk reduction (not complete protection) for human/LLM interfaces. Hexagonal architecture with multi-layer validation of inputs, outputs, memory and tool state. Beta status. ~528 KB wheel, optional ML guards.

  • Updated Dec 13, 2025
  • Python

🛡️ AI Firewall with 87 detection engines | Topological Data Analysis | Sheaf Theory | Post-Quantum | Protect LLMs from jailbreaks, injections & adversarial attacks

  • Updated Dec 13, 2025
  • HTML

Improve this page

Add a description, image, and links to the llm-security topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the llm-security topic, visit your repo's landing page and select "manage topics."

Learn more