Home

Awesome

Advances in Deep Concealed Scene Understanding

<img align="right" src="./assets/csu-logo.png" width="350px" />

This repository contains a collection of research papers, an evaluation toolbox, and benchmarking results for the task of concealed object segmentation (COS) in images. Besides, to evaluate the generalizability of COS approaches, we re-organize a concealed defect segmentation dataset named CDS2K.

Table of Contents

CSU Background

Concealed scene understanding (CSU) is a hot computer vision topic aiming to perceive objects with camouflaged properties. The current boom in its advanced techniques and novel applications makes it timely to provide an up-to-date survey to enable researchers to understand the global picture of the CSU field, including both current achievements and major challenges.

<p align="center"> <img src="assets/dataset_sample_gallery.png" width="400"/> <br /> <em> Figure 1: Sample gallery of concealed scenarios. (a-d) show natural animals. (e) depicts a concealed human in art. (f) features a synthesized ``lion''. </em> </p>

This paper makes four contributions:

CSU Taxonomy

We introduce a taxonomy of seven popular CSU tasks. Please refer to Section 2.1 of our paper for more details.

We illustrate each task with its corresponding annotation visualization.

<p align="center"> <img src="assets/task_definition.png"/> <br /> <em> Figure 2: Illustration of representative CSU tasks. </em> </p>

CSU Survey

We recap the latest image-based research that includes 50 papers.

<p align="center"> <img src="assets/reviewed_image_methods.png"/> <br /> <em> Table 1: Essential characteristics of reviewed video-level CSU methods. </em> </p>

We also review recent nine video-based research

<p align="center"> <img src="assets/reviewed_video_methods.png"/> <br /> <em> Table 2: Essential characteristics of reviewed video-level CSU methods. </em> </p>

The following are ten datasets collected for several CSU-related tasks.

<p align="center"> <img src="assets/reviewed_datasets.png"/> <br /> <em> Table 3: Essential characteristics of reviewed video-level CSU methods. </em> </p>

CSU Benchmark

Our benchmarking is built on COS tasks since this topic is relatively well-established and offers a variety of competing approaches. WHAT DO WE PROVIDE HERE?

<p align="center"> <img src="assets/benchmark_camo.png"/> <br /> <em> Table 4: Quantitative comparison of CAMO testing set. </em> </p> <p align="center"> <img src="assets/benchmark_nc4k.png"/> <br /> <em> Table 5: Quantitative comparison on NC4K testing set. </em> </p> <p align="center"> <img src="assets/benchmark_cod10k.png"/> <br /> <em> Table 6: Quantitative comparison of COD10K testing set. </em> </p> <p align="center"> <img src="assets/cos_quali_viz.png"/> <br /> <em> Figure 3: Qualitative results of ten COS approaches. For more descriptions of visual attributes in each column refer to Section 5.6 of the paper. </em> </p>

Defect Segmentation Dataset -- CDS2K

We organize a concealed defect segmentation dataset (Google Drive, 159MB) from the five well-known defect segmentation databases. As shown in Figure 4, we present five sub-databases: (a-l) MVTecAD, (m-o) NEU, (p) CrackForest, (q) KolektorSDD, and (r) MagneticTile. The defective regions are highlighted with red rectangles. (Top-Right) Word cloud visualization of CDS2K. (Bottom) The statistic number of positive/negative samples of each category in our CDS2K.

<p align="center"> <img src="assets/cds2k.png"/> <br /> <em> Figure 4: Sample gallery of our CDS2K. </em> </p>

The average ratio of defective regions for each category is presented in Table 7, which indicates that most of the defective regions are relatively small

<p align="center"> <img src="assets/cds2k-statistics.png"/> <br /> <em> Table 7: Sample gallery of our CDS2K. </em> </p>

Next, we report the quantitative comparison on the positive samples of CDS2K. Kindly download the result map on Google Drive (116.6MB).

<p align="center"> <img src="assets/cds2k-benchmark.png"/> <br /> <em> Table 8: SQuantitative comparison on the positive samples of CDS2K. </em> </p>

Citation

Please cite our paper if you find the work useful:

@article{fan2023csu,
  title={Advances in Deep Concealed Scene Understanding},
  author={Fan, Deng-Ping and Ji, Ge-Peng and Xu, Peng and Cheng, Ming-Ming and Sakaridis, Christos and Van Gool, Luc},
  journal={Visual Intelligence (VI)},
  year={2023}
}