Home

Awesome

Learning Generative Vision Transformer with Energy-Based Latent Space for Saliency Prediction (NeurIPS 2021)

Trained Models and Saliency Maps

We provide the trained models and saliency maps for RGB image based salient object detection and RGB-D image pair based salient object detection with both CNN backbone and transformer backbone.

For RGB image based models, we provide results on six testing datasets: DUTS_Test, ECSSD, DUT, HKU-IS, PASCAL-S and SOD.

Model and results of CNN based model: https://drive.google.com/drive/folders/1Sz-7j2Hl_oaMznkX3gX6v0Trm2WPc6XX?usp=sharing

Model and results of transformer based model: https://drive.google.com/drive/folders/1LQEXdrbiZv_BIbKhPPbeymCkav-jTeZH?usp=sharing

For RGB-D image pair based models, we provide results on six testing datasets: NJU2K, NLPR, LFSD, SIP, DES and STERE.

Model and results of CNN based model: https://drive.google.com/file/d/1vkatlwTjsTQpiw7Or_eV_Q4YKNI-E_LZ/view?usp=sharing

Model and results of transformer based model: https://drive.google.com/file/d/1UV5HBjtYuJnJIKj58r-laP8gjUCAzPTs/view?usp=sharing

Our Bib:

Please cite our paper if necessary:

@inproceedings{jing_ebm_sod21,
  title={Learning Generative Vision Transformer with Energy-Based Latent Space for Saliency Prediction},
  author={Zhang, Jing and Xie, Jianwen and Barnes, Nick and Li, Ping},
  booktitle={2021 Conference on Neural Information Processing Systems},
  year={2021}
}

Contact

Please drop me an email for further problems or discussion: zjnwpu@gmail.com

Acknowledgment

Thanks Yuxin Mao (maoyuxin@mail.nwpu.edu.cn) for setting up the transformer framework for salient object detection. Please refer to paper below for details:

@article{mao2021transformer,
  title={Transformer transforms salient object detection and camouflaged object detection},
  author={Mao, Yuxin and Zhang, Jing and Wan, Zhexiong and Dai, Yuchao and Li, Aixuan and Lv, Yunqiu and Tian, Xinyu and Fan, Deng-Ping and Barnes, Nick},
  journal={arXiv preprint arXiv:2104.10127},
  year={2021}
}