Awesome
Distributionally Adversarial Attack
Update: The convex folder is updated, you can evaluate DAA on provable network by the main_attack.py in the its examples folder.
Recently, many defense models against adversarial attacks existed. By an extensive evaluation, we figure out that one of the most effective defense methods is PGD adversarial training (Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, Adrian Vladu https://arxiv.org/abs/1706.06083)
Another potential provable defense method is: https://github.com/locuslab/convex_adversarial
We design a new first-order attack algorithm by generalizing PGD on the space of data distributions and learning an adversarial distribution that maximally increases the generalization risk of a model, namely Distributionally Adversarial Attack (DAA). Our DAA attack achieves outstanding attack success rates on those state-of-the-art defense models. Our paper link is https://arxiv.org/abs/1808.05537 (Paper Authors: Tianhang Zheng, Changyou Chen, Kui Ren)
There are 2 attack versions, i.e., DAA-BLOB and DAA-DGF. Our code is written based on MadryLab's code
Running the code
Since the files in models are not fully uploaded to Github, pls download madry's models using python3 fetch_model.py secret
python blob_rand.py
: DAA-BLOB attackpython dgf_rand.py
: DAA-DGF attackpython pgd_rand.py
: PGD attackpython mi_pgd_rand.py
: PGD variant of Momentum Iterative attack
Python3.5/3.6 is suggusted.
Our MNIST result shows on MadryLab's white-box leaderboard :-) => https://github.com/MadryLab/mnist_challenge
Our CIFAR10 result shows on MadryLab's white-box leaderboard :-) => https://github.com/MadryLab/cifar10_challenge
We also evaluate our attack against provable defense, and the code is in the directory "convex_adversarial-master".
If there is any question, pls contact tzheng4@buffalo.edu