Home

Awesome

SDRNet

SDRNet: Camouflaged Object Detection with Independent Reconstruction of Structure and Detail, Accepted by KBS

[Paper] [Results] [Pretrained models]


Abstract: The simultaneous reconstruction of structure and detail is a prevalent strategy in camouflaged object detection. However, the reconstruction features required for structure and detail exhibit disparities, a facet overlooked in existing methods. Therefore, we present a novel methodology, termed SDRNet, which employs a dual-branch approach for the independent reconstruction of structure and detail, aiming to discern camouflaged targets and their edges. Specifically, we propose a decomposition block to segregate encoded features into distinct structure and detail components. Furthermore, structure enhancement block and detail enhancement block are proposed as feature enhancement methods to boost the capacity of structure and detail information. Subsequently, the introduced structure fusion block and detail fusion block progressively amalgamate the enhanced features. Additionally, the shared feature block is designed to serve as a bridge for the interaction between structure and detail information. Experimental results demonstrate that SDRNet outperforms existing state-of-the-art methods significantly on benchmark datasets.

<p align="center"> <img width="800" src="fr.png"> </p>

Usage

1. Training Configuration

python Train.py  --epoch 100  --lr 1e-4  --batchsize 12  --trainsize 384  --train_root YOUR_TRAININGSETPATH  --val_root  YOUR_VALIDATIONSETPATH  --save_path YOUR_CHECKPOINTPATH

2. Testing Configuration

Our well-trained model is stored in Google Drive. After downloading, please change the file path in the corresponding code.

python Test.py  --testsize YOUR_IMAGESIZE  --pth_path YOUR_CHECKPOINTPATH 

3. Evaluation

4. Results download

The prediction results of our SDRNet are stored on Google Drive please check.

5. Visualization

Change the params in cam.py as your setting, and run with

python cam.py  

the results will be saved.

model_weight_path: the trained model path; target_layers: the layer your want to visualize. eg.[model.layer1]; path: the path of dataset.

Citation

If you find our work useful in your research, please consider citing: @article{GUAN2024112051, title = {SDRNet: Camouflaged object detection with independent reconstruction of structure and detail}, journal = {Knowledge-Based Systems}, volume = {299}, pages = {112051}, year = {2024}, issn = {0950-7051}, doi = {https://doi.org/10.1016/j.knosys.2024.112051}, url = {https://www.sciencedirect.com/science/article/pii/S0950705124006853}, author = {Juwei Guan and Xiaolin Fang and Tongxin Zhu and Weiqi Qian}, keywords = {Camouflaged object detection, Structure and detail reconstruction, Feature enhancement, Feature fusion, Feature decomposition}, abstract = {The simultaneous reconstruction of structure and detail is a prevalent strategy in camouflaged object detection. However, the reconstruction features required for structure and detail exhibit disparities, a facet overlooked in existing methods. Therefore, we present a novel methodology, termed SDRNet, which employs a dual-branch approach for the independent reconstruction of structure and detail, aiming to discern camouflaged targets and their edges. Specifically, we propose a decomposition block to segregate encoded features into distinct structure and detail components. Furthermore, structure enhancement block and detail enhancement block are proposed as feature enhancement methods to boost the capacity of structure and detail information. Subsequently, the introduced structure fusion block and detail fusion block progressively amalgamate the enhanced features. Additionally, the shared feature block is designed to serve as a bridge for the interaction between structure and detail information. Experimental results demonstrate that SDRNet outperforms existing state-of-the-art methods significantly on benchmark datasets. Our code is available at https://github.com/whyandbecause/SDRNet/.} }

Concat

If you have any questions, please feel free to contact me via email at jvguan@163.com or jvguan@seu.edu.cn.