Awesome
Memory-Aided Contrastive Consensus Learning for Co-salient Object Detection
Memory-Aided Contrastive Consensus Learning for Co-salient Object Detection
Peng Zheng, Jie Qin, Shuo Wang, Tian-Zhu Xiang, Huan Xiong
AAAI2023 (AAAI, arXiv)
Overview
<img src="README.assets/paper922_poster.png" alt="paper922_poster" style="zoom:45%;" />Abstract
Co-Salient Object Detection (CoSOD) aims at detecting common salient objects within a group of relevant source images. Most of the latest works employ the attention mechanism for finding common objects. To achieve accurate CoSOD results with high-quality maps and high efficiency, we propose a novel Memory-aided Contrastive Consensus Learning (MCCL) framework, which is capable of effectively detecting co-salient objects in real time (∼150 fps). To learn better group consensus, we propose the Group Consensus Aggregation Module (GCAM) to abstract the common features of each image group; meanwhile, to make the consensus representation more discriminative, we introduce the Memory-based Contrastive Module (MCM), which saves and updates the consensus of images from different groups in a queue of memories. Finally, to improve the quality and integrity of the predicted maps, we develop an Adversarial Integrity Learning (AIL) strategy to make the segmented regions more likely composed of complete objects with less surrounding noise. Extensive experiments on all the latest CoSOD benchmarks demonstrate that our lite MCCL outperforms 13 cutting-edge models, achieving the new state of the art (∼5.9% and ∼6.2% improvement in S-measure on CoSOD3k and CoSal2015, respectively).
Prerequisites
Python=3.8
!pip install -r requirements.txt
PyTorch==1.13.1
Datasets
Download necessary datasets:
from my google-drive: DUTS_class, COCO-9k, COCO-SEG, and CoSOD_testsets, or
from my BaiduDisk: DUTS_class, COCO-9k, COCO-SEG, and CoSOD_testsets.
The CoSOD_testsets
contains CoCA, CoSOD3k and CoSal2015.
Usage
The way to run this project is similar to our previous work GCoNet+.
- Train->test->eval: Run
go.sh
to go through training->testing->evaluation:./go.sh METHOD_NAME WHICH_CUDA
. - If you want to use the PVT_v2_b2 backbone as ours (the PVT repo), save it as it is in the
config.py
. Not necessary if you only want to do the inference. - Fast test+eval: Put the weights in
ckpt/METHOD_NAME/xx.pth
, change the params ofepoch, val_last
ingo.sh
, comment the codes for training, and run this script. For example, if you want to eval the weights of the methodtmp_method
atepoch 100
, make sure that weights are inckpt/tmp_method/ep100.pth
, and changeepochs and val_last
ingo.sh
as100, 1
, respectively.
Outputs
You can find our weights and predicted maps on both google-drive and release of this repo.
Acknowledgement
We highly recommend use the metric codes from py_sod_metrics and thanks to the codes of drawing the picture of accuracy-speed in DGNet. This repo is based on our previous project GCoNet+.
Citation
@inproceedings{zheng2022mccl,
title = {Memory-Aided Contrastive Consensus Learning for Co-salient Object Detection},
author = {Zheng Peng, Qin Jie, Wang Shuo, Xiang Tian-Zhu and Xiong Huan},
booktitle = AAAI,
year = {2023}
}
Contact
Feel free to send e-mails to me (zhengpeng0108@gmail.com).