Awesome
ATOM: Robustifying Out-of-distribution Detection Using Outlier Mining
This project is for the paper: ATOM: Robustifying Out-of-distribution Detection Using Outlier Mining.
Illustration
Experimental Results
Preliminaries
It is tested under Ubuntu Linux 16.04.1 and Python 3.6 environment, and requires some packages to be installed:
Downloading In-distribution Dataset
Downloading Auxiliary OOD Datasets
- 80 Million Tiny Images: to download 80 Million Tiny Images dataset. In the root directory, run
cd datasets/unlabeled_datasets/80M_Tiny_Images
wget http://horatio.cs.nyu.edu/mit/tiny/data/tiny_images.bin
- Downsampled ImageNet Datasets: we use the ImageNet64x64, which could be downloaded from ImageNet Website. After downloading it, place it in this directory:
datasets/unlabeled_datasets/ImageNet
.
Downloading Out-of-distribution Test Datasets
We provide links and instructions to download each dataset:
- SVHN: download it and place it in the folder of
datasets/ood_datasets/svhn
. Then runpython select_svhn_data.py
to generate test subset. - Textures: download it and place it in the folder of
datasets/ood_datasets/dtd
. - Places365: download it and place it in the folder of
datasets/ood_datasets/places365/test_subset
. We randomly sample 10,000 images from the original test dataset. We provide the file names for the images that we sample indatasets/ood_datasets/places365/test_subset/places365_test_list.txt
. - LSUN-C: download it and place it in the folder of
datasets/ood_datasets/LSUN
. - LSUN-R: download it and place it in the folder of
datasets/ood_datasets/LSUN_resize
. - iSUN: download it and place it in the folder of
datasets/ood_datasets/iSUN
.
For example, run the following commands in the root directory to download LSUN-C:
cd datasets/ood_datasets
wget https://www.dropbox.com/s/fhtsw1m3qxlwj6h/LSUN.tar.gz
tar -xvzf LSUN.tar.gz
Downloading Pre-trained Models
We provide pre-trained DenseNet models for vanilla, SOFL, OE, ACET, CCU, ROWL, NTOM, and ATOM methods on SVHN, CIFAR-10, and CIFAR-100 datasets using TinyImages as auxiliary OOD dataset. We also provide pre-trained WideResNet models for vanilla, NTOM, and ATOM methods on SVHN, CIFAR-10, and CIFAR-100 datasets using TinyImages as auxiliary OOD dataset. They can be downloaded from Google Drive.
Overview of the Code
Running Experiments
select_svhn_data.py
: select SVHN test data.eval_ood_detection.py
: evaluate OOD detection performance of models.compute_metrics.py
: compute metrics of evaluation results.gen_rowl_train_data.py
: generate ROWL training data.gen_validation_data.py
: generate validation data used to select the best q.train_acet.py
: train ACET model.train_atom.py
: train ATOM model.train_ccu.py
: train CCU model.train_oe.py
: train OE model.train_rowl.py
: train ROWL model.train_sofl.py
: train SOFL model.train.py
: train vanilla model.tune_mahalanobis_hyperparams.py
: tune hyperparameters of Mahalanobis detector.tune_odin_hyperparams.py
: tune hyperparameters of ODIN detector.
Example
For CIFAR-10 experiments, you can run the following script to get the results:
./run_all_cifar10_exps.sh
To evaluate an OOD detection method, you can use the following command:
python eval_ood_detection.py --in-dataset {in-distribution dataset} --name {model name} --method {scoring function} [--adv or --corrupt or --adv-corrupt]
The argument --adv
is to evaluate L_inf OOD, --corrupt
is to evaluate Corrupted OOD, --adv-corrupt
is to evaluate Comp. OOD. You can only specify one of them. Without them, it will evaluate natural OOD.
{in-distribution dataset}
can be SVHN
, CIFAR-10
or CIFAR-100
.
{scoring function}
can be msp
, odin
, mahalanobis
, sofl
, rowl
, ntom
or atom
.
{model name}
is the name of the model that you have trained. The model type should match the corresponding scoring function. See the following table for the matchings:
Model Type | Scoring Function |
---|---|
vanilla | msp |
vanilla | odin |
vanilla | mahalanobis |
SOFL | sofl |
OE | msp |
ACET | msp |
CCU | msp |
ROWL | rowl |
NTOM | ntom |
ATOM | atom |
Acknowledgements
Part of this code is inspired by ODIN, Outlier Exposure, Deep Mahalanobis Detector, Corruption and Perturbation Robustness and Certified Certain Uncertainty.
Citation
Please cite our work if you use the codebase:
@article{chen2021atom,
title={ATOM: Robustifying Out-of-distribution Detection Using Outlier Mining},
author={Chen, Jiefeng and Li, Yixuan and Wu, Xi and Liang, Yingyu and Jha, Somesh},
journal={In Proceedings of European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD)},
year={2021}
}
License
Please refer to the LICENSE.