Adversarial Network Attacks (PGD, pixel, FGSM) Noise on MNIST Images Dataset using Python (Pytorch)
-
Updated
Jun 29, 2022
10000
Adversarial Network Attacks (PGD, pixel, FGSM) Noise on MNIST Images Dataset using Python (Pytorch)
This github repository contains the official code for the papers, "Robustness Assessment for Adversarial Machine Learning: Problems, Solutions and a Survey of Current Neural Networks and Defenses" and "One Pixel Attack for Fooling Deep Neural Networks"
Crafting adversarial examples with one pixel attack
Add a description, image, and links to the pixel-attack topic page so that developers can more easily learn about it.
To associate your repository with the pixel-attack topic, visit your repo's landing page and select "manage topics."