Home

Awesome

Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation, NIPS 2017

Matthias Hein, Maksym Andriushchenko

Saarland University

From left to right: original image (1) and adversarial examples for a 1 hidden layer NN trained with weight decay (2), dropout (3) and Cross-Lipschitz (4) regularization.

<img src="images/mnist1-orig.png" width=200/> <img src="images/mnist1-wd.png" width=200/> <img src="images/mnist1-do.png" width=200/> <img src="images/mnist1-cl.png" width=200/> <img src="images/70-30-orig.png" width=200/> <img src="images/70-30-wd.png" width=200/> <img src="images/70-30-do.png" width=200/> <img src="images/70-30-cl.png" width=200/> <img src="images/stop-orig.png" width=200/> <img src="images/stop-wd.png" width=200/> <img src="images/stop-do.png" width=200/> <img src="images/stop-cl.png" width=200/>

The adversarial examples were generated using the proposed method with a binary search, such that they are located on the decision boundary between 2 classes. So there are no overshoots in terms of the required adversarial change.

We can see that adversarial examples generated for the NN trained with Cross-Lipschitz regularization in first 2 cases indeed change the class even for a human observer. However, there are still some dificulties with more abstract classes (e.g. stop signs), but in this case the required norm of the adversarial change is usually higher with Cross-Lipschitz regularization.

<!--![](images/70-30-orig.png) ![](images/70-30-wd.png) ![](images/70-30-do.png) ![](images/70-30-cl.png)--> <!--![](images/stop-orig.png) ![](images/stop-wd.png) ![](images/stop-do.png) ![](images/stop-cl.png)--> <!--![](images/mnist1-orig.png) ![](images/mnist1-wd.png) ![](images/mnist1-do.png) ![](images/mnist1-cl.png)-->

Example of usage

If we want to run a 32-layer ResNet on MNIST with adversarial training and with Cross-Lipschitz regularization:

python worker.py --experiment_name=advers_training --adv_train_flag --reg_type=cross_lipschitz --dataset=cifar10 --nn_type=resnet --gpu_number=0 --gpu_memory=0.6 --lr=0.2 --lmbd=0.0001 --batch_size=128 --n_epochs=200

You can get further information on different arguments with: python worker.py --help.

Cross-Lipschitz regularizer

The most interesting piece of code is the Cross-Lipschitz regularizer in regularizers.py: <img src="http://latex.codecogs.com/gif.latex?\Omega(f)%20=%20\frac{1}{nK^2}\sum_{i=1}^n%20\sum_{l,m=1}^K%20||\nabla%20f_l(x_i)%20-%20\nabla%20f_m(x_i)||_2^2" />

It contains ~10 lines of code and they are very straightforward due to tf.gradients used to calculate the required derivatives. It is suitable for any differentiable classifier.

Another contribution is the generation of targeted adversarial examples with box constraints, which is in attacks.py. Note, that the optimization problem being solved uses linear approximation of the classifier.

Additionally, the only well-defined point for an adversarial change is the point on the decision boundary. In order to find it we perform a binary search for the smallest c in the optimization problem, which is enough to change the class. The code is in ae_generation.py.

Running options

Supported neural network types:

Regularization types:

Supported settings (if no flag specified then it will run in plain setting):

Datasets (you will need to have them in the folder data/, for details please see data.py):

Note that the script worker.py saves optimization logs, different metrics and the model parameter to the corresponding folders. In order to evaluate the upper bounds of the minimum adversarial change for each example, you can use eval_robustness.py. This script should be called after all models in a particular experiment are trained.

Contact

For any questions regarding the code please contact Maksym Andriushchenko (m.my surname@gmail.com). Any suggestions are always welcome.

Citation

@incollection{NIPS2017_6821,
	title = {Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation},
	author = {Hein, Matthias and Andriushchenko, Maksym},
	booktitle = {Advances in Neural Information Processing Systems 30},
	editor = {I. Guyon and U. V. Luxburg and S. Bengio and H. Wallach and R. Fergus and S. Vishwanathan and R. Garnett},
	pages = {2263--2273},
	year = {2017},
	publisher = {Curran Associates, Inc.},
	url = {http://papers.nips.cc/paper/6821-formal-guarantees-on-the-robustness-of-a-classifier-against-adversarial-manipulation.pdf}
}