Home

Awesome

Self-produced Guidance for Weakly-supervised Object Localization

We train the SPG model on the ILSVRC dataset, and then apply the trained model on video sequences of DAVIS 2016. <img width="400" height="200" src="figs/bear_loc.gif"/><img width="400" height="200" src="figs/dog_spg_c.gif"/>

Overview of SPG

Train

We finetune the SPG model on the ILSVRC dataset.

cd scripts
sh train_imagenet_full_v5.sh

Test

Download the pretrined model at GoogleDrive(https://drive.google.com/open?id=1EwRuqfGASarGidutnYB8rXLSuzYpEoSM (IMAGENET),https://drive.google.com/open?id=1WfrELBlEoq5WO7gKUv-MLTQ8QHY-2wiX (CUB)).

Use the test script to generate attention maps.

cd scripts
sh val_imagenet_full.sh

Demo

Thanks to Jun Hao for providing the wonderful demos!

Please see the setup_demo.txt for more guidance of setuping up the demos.

Masks are getting better with the proposed easy-to-hard approach.

Citation

If you find this code helpful, please consider to cite this paper:

@inproceedings{zhang2018self,
  title={Self-produced Guidance for Weakly-supervised Object Localization},
  author={Zhang, Xiaolin and Wei, Yunchao and Kang, Guoliang and Yang, Yi and Huang, Thomas},
  booktitle={European Conference on Computer Vision},
  year={2018},
  organization={Springer}
}