Home

Awesome

Memory-Aided Contrastive Consensus Learning for Co-salient Object Detection

Memory-Aided Contrastive Consensus Learning for Co-salient Object Detection
Peng Zheng, Jie Qin, Shuo Wang, Tian-Zhu Xiang, Huan Xiong
AAAI2023 (AAAI, arXiv)

Overview

<img src="README.assets/paper922_poster.png" alt="paper922_poster" style="zoom:45%;" />

Abstract

Co-Salient Object Detection (CoSOD) aims at detecting common salient objects within a group of relevant source images. Most of the latest works employ the attention mechanism for finding common objects. To achieve accurate CoSOD results with high-quality maps and high efficiency, we propose a novel Memory-aided Contrastive Consensus Learning (MCCL) framework, which is capable of effectively detecting co-salient objects in real time (∼150 fps). To learn better group consensus, we propose the Group Consensus Aggregation Module (GCAM) to abstract the common features of each image group; meanwhile, to make the consensus representation more discriminative, we introduce the Memory-based Contrastive Module (MCM), which saves and updates the consensus of images from different groups in a queue of memories. Finally, to improve the quality and integrity of the predicted maps, we develop an Adversarial Integrity Learning (AIL) strategy to make the segmented regions more likely composed of complete objects with less surrounding noise. Extensive experiments on all the latest CoSOD benchmarks demonstrate that our lite MCCL outperforms 13 cutting-edge models, achieving the new state of the art (∼5.9% and ∼6.2% improvement in S-measure on CoSOD3k and CoSal2015, respectively).

Prerequisites

Python=3.8
!pip install -r requirements.txt
PyTorch==1.13.1

Datasets

Download necessary datasets:
from my google-drive: DUTS_class, COCO-9k, COCO-SEG, and CoSOD_testsets, or
from my BaiduDisk: DUTS_class, COCO-9k, COCO-SEG, and CoSOD_testsets.
The CoSOD_testsets contains CoCA, CoSOD3k and CoSal2015.

Usage

The way to run this project is similar to our previous work GCoNet+.

Outputs

You can find our weights and predicted maps on both google-drive and release of this repo.

Acknowledgement

We highly recommend use the metric codes from py_sod_metrics and thanks to the codes of drawing the picture of accuracy-speed in DGNet. This repo is based on our previous project GCoNet+.

Citation

@inproceedings{zheng2022mccl,
  title = {Memory-Aided Contrastive Consensus Learning for Co-salient Object Detection},
  author = {Zheng Peng, Qin Jie, Wang Shuo, Xiang Tian-Zhu and Xiong Huan},
  booktitle = AAAI,
  year = {2023}
}

Contact

Feel free to send e-mails to me (zhengpeng0108@gmail.com).