[ Have a look at the presentation slides: slides-OFFZONE.pdf / slides-ODS.pdf ]
[ Related demonstration (Jupyter notebook): demo.ipynb ]
Overview |
Attacks |
Tools |
More on the topic
An overview of black-box attacks on AI and tools that might be useful during security testing of machine learning models.
demo.ipynb:
A demonstration of use of multifunctional tools during security testing of machine learning models digits_blackbox & digits_keras trained on the MNIST dataset and provided in Counterfit as example targets.
Slides:
โโโMachine Learning in products
โโโThreats to Machine Learning models
โโโExample model overview
โโโEvasion attacks
โโโModel inversion attacks
โโโModel extraction attacks
โโโDefences
โโโAdversarial Robustness Toolbox
โโโCounterfit
- Model inversion attack:
MIFaceโ code / docs / ๐DOI:10.1145/2810103.2813677 - Model extraction attack:
Copycat CNNโ code / docs / ๐arXiv:1806.05476 - Evasion attack:
Fast Gradient Method (FGM)โ code / docs / ๐arXiv:1412.6572 - + Evasion attack:
HopSkipJumpโ code / docs / ๐arXiv:1904.02144
โโโ[ Trusted AI, IBM ] Adversarial Robustness Toolbox (ART): Trusted-AI/adversarial-robustness-toolbox
โโโ[ Microsoft Azure ] Counterfit: Azure/counterfit
-
adversarial examplesevasion attacks
How MIT researchers made Google's AI think tabby cat is guacamole:โoverview / ๐arXiv:1707.07397 + arXiv:1804.08598 -
model inversion attacks
Apple's take on model inversion:โoverview / ๐arXiv:2111.03702 -
model inversion attacks
Google's demonstration of extraction of training data that the GPT-2 model has memorized:โoverview / ๐arXiv:2012.07805 -
attacks on AIadversarial attackspoisoning attacksmodel inference attacks
โ Posts on PortSwigger's "The Daily Swig" by Ben Dickson