Home

Awesome

Global Semantic-Sense Aggregation Network for Salient Object Detection in Remote Sensing Images

⭐ This code has been completely released ⭐

⭐ our article ⭐

πŸ“– Introduction

<span style="font-size: 125%">Salient object detection (SOD) aims to accurately identify significant geographical objects in remote sensing images (RSI), providing reliable support and guidance for extensive geographical information analyses and decisions. However, SOD in RSI faces numerous challenges, including shadow interference, inter-class feature confusion, as well as unclear target edge contours. Therefore, we designed an effective Global Semantic-aware Aggregation Network (GSANet) to aggregate salient information in RSI. GSANet computes the information entropy of different regions, prioritizing areas with high information entropy as potential target regions, thereby achieving precise localization and semantic understanding of salient objects in remote sensing imagery. Specifically, we proposed a Semantic Detail Embedding Module (SDEM), which explores the potential connections among multi-level features, adaptively fusing shallow texture details with deep semantic features, efficiently aggregating the information entropy of salient regions, enhancing information content of salient targets. Additionally, we proposed a Semantic Perception Fusion Module (SPFM) to analyze map relationships between contextual information and local details, enhancing the perceptual capability for salient objects while suppressing irrelevant information entropy, thereby addressing the semantic dilution issue of salient objects during the up-sampling process. The experimental results on two publicly available datasets, ORSSD and EORSSD, demonstrated the outstanding performance of our method. The method achieved 93.91% SΞ±, 98.36% EΞΎ, and 89.37% FΞ² on the EORSSD dataset.</span>

<p align="center"> <img src="Images/Network.png" width=90%"></p>

If our code is helpful to you, please cite:

@Article{e26060445,
AUTHOR = {Li, Hongli and Chen, Xuhui and Yang, Wei and Huang, Jian and Sun, Kaimin and Wang, Ying and Huang, Andong and Mei, Liye},
TITLE = {Global Semantic-Sense Aggregation Network for Salient Object Detection in Remote Sensing Images},
JOURNAL = {Entropy},
VOLUME = {26},
YEAR = {2024},
NUMBER = {6},
ARTICLE-NUMBER = {445},
URL = {https://www.mdpi.com/1099-4300/26/6/445},
ISSN = {1099-4300},
ABSTRACT = {Salient object detection (SOD) aims to accurately identify significant geographical objects in remote sensing images (RSI), providing reliable support and guidance for extensive geographical information analyses and decisions. However, SOD in RSI faces numerous challenges, including shadow interference, inter-class feature confusion, as well as unclear target edge contours. Therefore, we designed an effective Global Semantic-aware Aggregation Network (GSANet) to aggregate salient information in RSI. GSANet computes the information entropy of different regions, prioritizing areas with high information entropy as potential target regions, thereby achieving precise localization and semantic understanding of salient objects in remote sensing imagery. Specifically, we proposed a Semantic Detail Embedding Module (SDEM), which explores the potential connections among multi-level features, adaptively fusing shallow texture details with deep semantic features, efficiently aggregating the information entropy of salient regions, enhancing information content of salient targets. Additionally, we proposed a Semantic Perception Fusion Module (SPFM) to analyze map relationships between contextual information and local details, enhancing the perceptual capability for salient objects while suppressing irrelevant information entropy, thereby addressing the semantic dilution issue of salient objects during the up-sampling process. The experimental results on two publicly available datasets, ORSSD and EORSSD, demonstrated the outstanding performance of our method. The method achieved 93.91% SΞ±, 98.36% EΞΎ, and 89.37% FΞ² on the EORSSD dataset.},
DOI = {10.3390/e26060445}
}

Saliency maps

We provide saliency maps of our and compared methods at here (code:o4dz) on two datasets (ORSSD and EORSSD).

DateSets

ORSSD download at here

EORSSD download at here

The structure of the dataset is as follows:

GSANet
β”œβ”€β”€ EORSSD
β”‚   β”œβ”€β”€ train
β”‚   β”‚   β”œβ”€β”€ images
β”‚   β”‚   β”‚   β”œβ”€β”€ 0001.jpg
β”‚   β”‚   β”‚   β”œβ”€β”€ 0002.jpg
β”‚   β”‚   β”‚   β”œβ”€β”€ .....
β”‚   β”‚   β”œβ”€β”€ lables
β”‚   β”‚   β”‚   β”œβ”€β”€ 0001.png
β”‚   β”‚   β”‚   β”œβ”€β”€ 0002.png
β”‚   β”‚   β”‚   β”œβ”€β”€ .....
β”‚   β”‚   
β”‚   β”œβ”€β”€ test
β”‚   β”‚   β”œβ”€β”€ images
β”‚   β”‚   β”‚   β”œβ”€β”€ 0004.jpg
β”‚   β”‚   β”‚   β”œβ”€β”€ 0005.jpg
β”‚   β”‚   β”‚   β”œβ”€β”€ .....
β”‚   β”‚   β”œβ”€β”€ lables
β”‚   β”‚   β”‚   β”œβ”€β”€ 0004.png
β”‚   β”‚   β”‚   β”œβ”€β”€ 0005.png
β”‚   β”‚   β”‚   β”œβ”€β”€ .....

Train

  1. Download the dataset.

  2. Use data_aug.m to augment the training set of the dataset.

  3. Download uniformer_base_ls_in1k.pth (code: o4dz), and put it in './pretrain/'.

  4. Modify paths of datasets, then run train_MyNet.py.

Test

  1. Download the pre-trained models of our network at here (code:o4dz)
  2. Modify paths of pre-trained models and datasets.
  3. Run test_MyNet.py.

Results

Main results on ORSSD dataset

MethodsS<sub>Ξ±</sub>MAEadp E<sub>ΞΎ</sub>mean E<sub>ΞΎ</sub>max E<sub>ΞΎ</sub>adp F<sub>Ξ²</sub>mean F<sub>Ξ²</sub>max F<sub>Ξ²</sub>
SAMNet0.87610.02170.86560.88180.94780.68430.75310.8137
HVPNet0.86100.02250.84710.87370.93200.67260.73960.7938
DAFNet0.91910.01130.93600.95390.97710.78760.85110.8928
HFANet0.93990.00920.97220.97120.97700.88190.89810.9112
MSCNet0.92270.01290.95840.96530.97540.83500.86760.8927
MJRBM0.92040.01630.93280.94150.96230.80220.85660.8842
PAFR0.89380.02110.93150.92680.94670.80250.82750.8438
CorrNet0.92010.01580.95430.94870.95750.86050.87170.8841
EMFINet0.93800.01130.96370.96570.97330.86640.88730.9019
MCCNet0.94450.00910.97330.97400.98050.89250.90450.9177
ACCoNet0.94180.00950.96940.96840.97540.86140.88470.9112
AESINet0.94270.00900.97040.97360.98170.86670.89750.9166
ERPNet0.92540.01350.95200.85660.97100.83560.87450.8974
GeleNet0.94510.00920.98160.97990.98590.90440.91230.9239
ADSTNet0.93790.00860.97850.97400.98070.89790.90420.9124
Ours0.94910.00700.98070.98150.98640.89940.90950.9253

Main results on EORSSD dataset

MethodsS<sub>Ξ±</sub>MAEadp E<sub>ΞΎ</sub>mean E<sub>ΞΎ</sub>max E<sub>ΞΎ</sub>adp F<sub>Ξ²</sub>mean F<sub>Ξ²</sub>max F<sub>Ξ²</sub>
SAMNet0.86220.01320.82840.87000.94210.61140.72140.7813
HVPNet0.87340.01100.82700.87210.94820.62020.73770.8036
DAFNet0.91660.00600.84430.92900.98590.64230.78420.8612
HFANet0.93800.00700.96440.96790.97400.83650.86810.8876
MSCNet0.90710.00900.93290.95510.96890.75530.81510.8539
MJRBM0.91970.00990.88970.93500.96460.70660.82390.8656
PAFR0.89270.01190.89590.92100.94900.71230.79610.8260
CorrNet0.91530.00970.95140.94450.95530.82590.84500.8597
EMFINet0.92840.00870.94820.95420.96650.80490.84940.8735
MCCNet0.93400.00730.96090.96760.97580.83020.86560.8884
ACCoNet0.93460.00810.95590.96220.97070.82480.86280.8846
AESINet0.93620.00720.94430.96180.97340.79080.85070.8820
ERPNet0.92100.00890.92280.94010.96030.75540.83040.8632
GeleNet0.93730.00750.97280.97400.98100.86480.87810.8910
ADSTNet0.93110.00650.96810.97090.97690.85320.87160.8804
Ours0.93910.00530.97430.97840.98360.86570.87900.8937

Visualization of results

<p align="center"> <img src="Images/Result1.png" width=95%"></p>

Evaluation Tool

You can use the evaluation tool (MATLAB version) to evaluate the above saliency maps.

ORSI-SOD Summary

Salient Object Detection in Optical Remote Sensing Images Read List at here

Acknowledgements

This code is built on PyTorch.

Contact

If you have any questions, please submit an issue on GitHub or contact me by email (cxh1638843923@gmail.com).