Home

Awesome

<!-- * @Author: Wenhu Zhang * @Date: 2021-06-07 17:40:58 * @LastEditTime: 2021-06-07 18:24:59 * @LastEditors: Wenhu Zhang * @Description: * @FilePath: /github/sp/DSA2F/README.md -->

DSA^2 F: Deep RGB-D Saliency Detection with Depth-Sensitive Attention and Automatic Multi-Modal Fusion (CVPR'2021, Oral)

This repo is the official implementation of "DSA^2 F: Deep RGB-D Saliency Detection with Depth-Sensitive Attention and Automatic Multi-Modal Fusion"

by Peng Sun, Wenhu Zhang, Huanyu Wang, Songyuan Li, and Xi Li.

Prerequisites

Training

Please see launch_train.sh and launch_pretrain.sh for imagenet pretraining and sod training, respectively.

Testing

Please see launch_test.sh for testing on the sod benchmarks.

Main Results

DatasetE<sub>r</sub>S<sub>λ</sub><sup>mean</sup>F<sub>β</sub><sup>mean</sup>M
DUT-RGBD0.9500.9210.9260.030
NJUD0.9230.9030.9010.039
NLPR0.9500.9180.8970.024
SSD0.9040.8760.8520.045
STEREO0.9330.9040.8980.036
LFSD0.9230.8820.8820.054
RGBD1350.9620.9200.8960.021

Saliency maps and Evaluation

All of the saliency maps mentioned in the paper are available on GoogleDrive or BaiduYun(code:juc2).

You can use the toolbox provided by jiwei0921 for evaluation.

Additionally, we also provide the saliency maps of the STERE-1000 and SIP dataset on BaiduYun(code:qxfw) for easy comparison.

DatasetE<sub>r</sub>S<sub>λ</sub><sup>mean</sup>F<sub>β</sub><sup>mean</sup>M
STERE-10000.9280.8970.8950.038
SIP0.9080.8610.8680.057

Citation

@inproceedings{Sun2021DeepRS,
  title={Deep RGB-D Saliency Detection with Depth-Sensitive Attention and Automatic Multi-Modal Fusion},
  author={P. Sun and Wenhu Zhang and Huanyu Wang and Songyuan Li and Xi Li},
  journal={IEEE Conf. Comput. Vis. Pattern Recog.},
  year={2021}
}

License

The code is released under MIT License (see LICENSE file for details).