Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Starred repositories
Post-training with Tinker
Working Group on Artificial Intelligence and Machine Learning (AI/ML) Security
HexStrike AI MCP Agents is an advanced MCP server that lets AI agents (Claude, GPT, Copilot, etc.) autonomously run 150+ cybersecurity tools for automated pentesting, vulnerability discovery, bug b…
Context retrieval for AI agents across apps and databases
The dataset and code for the ICLR 2024 paper "Can LLM-Generated Misinformation Be Detected?"
Paper list for the survey "Combating Misinformation in the Age of LLMs: Opportunities and Challenges" and the initiative "LLMs Meet Misinformation", accepted by AI Magazine 2024
AMITT (Adversarial Misinformation and Influence Tactics and Techniques) framework for describing disinformation incidents. Includes TTPs and countermeasures.
Vanir is a source code-based static analysis tool that automatically identifies the list of missing security patches in the target system. By default, Vanir pulls up-to-date CVEs from Open Source V…
[NeurIPS 2024] Official implementation for "AgentPoison: Red-teaming LLM Agents via Memory or Knowledge Base Backdoor Poisoning"
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
An adversarial example library for constructing attacks, building defenses, and benchmarking both
A challenge to explore adversarial robustness of neural networks on MNIST.
🔓 🔓 Find secrets and passwords in container images and file systems 🔓 🔓
Make your GenAI Apps Safe & Secure 🚀 Test & harden your system prompt
DEEPSEC: A Uniform Platform for Security Analysis of Deep Learning Model
💡 Adversarial attacks on explanations and how to defend them
A curated list of awesome resources for adversarial examples in deep learning
Adversarial attacks and defenses on Graph Neural Networks.
Implementation of Papers on Adversarial Examples
🗣️ Tool to generate adversarial text examples and test machine learning models against them
Raising the Cost of Malicious AI-Powered Image Editing
A pytorch adversarial library for attack and defense methods on images and graphs
A Toolbox for Adversarial Robustness Research
Advbox is a toolbox to generate adversarial examples that fool neural networks in PaddlePaddle、PyTorch、Caffe2、MxNet、Keras、TensorFlow and Advbox can benchmark the robustness of machine learning mode…
A Python toolbox to create adversarial examples that fool neural networks in PyTorch, TensorFlow, and JAX
Master the fundamentals of machine learning, deep learning, and mathematical optimization by building key concepts and models from scratch using Python.