Advbox is a toolbox to generate adversarial examples that fool neural networks in PaddlePaddle、PyTorch、Caffe2、MxNet、Keras、TensorFlow and Advbox can benchmark the robustness of machine learning models. Advbox give a command line tool to generate adversarial examples with Zero-Coding.
This repository contains the implementation of three adversarial example attack methods FGSM, IFGSM, MI-FGSM and one Distillation as defense against all attacks using MNIST dataset.
Implementation of adversarial training under fast-gradient sign method (FGSM), projected gradient descent (PGD) and CW using Wide-ResNet-28-10 on cifar-10. Sample code is re-usable despite changing the model or dataset.
This project tests multiple different machine learning algorithms that can detect adversarial attacks in multi-agent reinforcement learning settings. Baselines were used to compare performance of a proposed ensemble model. Then, using FGSM, we re-attacked the ensemble detection model with perturbed observations. Read more at the pdf titled FinalPaper.
Repository consists of pre-trained CNN model in pytorch, hitting 89% on Fashion MNIST dataset. Adversarial attack was implemented on a given model. Results are below.
WideResNet implementation on MNIST dataset. FGSM and PGD adversarial attacks on standard training, PGD adversarial training, and Feature Scattering adversarial training.