DKETFormer: Salient object detection in optical remote sensing images based on discriminative knowledge extraction and transfer

被引:0
作者
Sun, Yuze
Zhao, Hongwei [1 ]
Zhou, Jianhang
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Jilin, Peoples R China
关键词
Salient object detection; Optical remote sensing images; Cross-spatial knowledge extraction; Inter-layer feature transfer module; NETWORK;
D O I
10.1016/j.neucom.2025.129558
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generally, most methods for salient object detection in optical remote sensing images (ORSI-SOD) are based on convolutional neural networks (CNNs). However, CNNs, due to their architectural characteristics, can only encode local semantic information, which leads to a lack of exploration of discriminative features on a large scale. Therefore, to encode the long-term dependency within the detection image, enhance the extraction of discriminative knowledge, and transfer it at multiple scales, we introduce a Transformer architecture called DKETFormer. Specifically, DKETFormer utilizes the Transformer backbone to obtain multi-scale feature maps that have encoded long-term dependency relationships. Then, it constructs a decoder using the Cross-spatial Knowledge Extraction Module (CKEM) and the Inter-layer Feature Transfer Module (IFTM). The CKEM is capable of extracting discriminative information across receptive fields while preserving knowledge from each channel. It also utilizes global information encoding to calibrate channel weights, resulting in improved knowledge aggregation and capturing of pixel-level pairwise relationships. The IFTM utilizes encoded and extracted information from the backbone and CKEM, employing a self-attention mechanism with cosine similarity knowledge to model and propagate discriminative features. Finally, we generated the final detection map using a salient object detector. The results of comparative experiments and ablation experiments demonstrate the effectiveness of the proposed DKETFormer and its internal modules.
引用
收藏
页数:11
相关论文
共 58 条
  • [1] Alexey D, 2020, arXiv, DOI [arXiv:2010.11929, DOI 10.48550/ARXIV.2010.11929]
  • [2] Reverse Attention-Based Residual Network for Salient Object Detection
    Chen, Shuhan
    Tan, Xiuli
    Wang, Ben
    Lu, Huchuan
    Hu, Xuelong
    Fu, Yun
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 3763 - 3776
  • [3] Chen ZY, 2020, AAAI CONF ARTIF INTE, V34, P10599
  • [4] Deng ZJ, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P684
  • [5] Multiscale and Multidimensional Weighted Network for Salient Object Detection in Optical Remote Sensing Images
    Di, Lamei
    Zhang, Bin
    Wang, Yiming
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 14
  • [6] EGFNet: Edge-Aware Guidance Fusion Network for RGB-Thermal Urban Scene Parsing
    Dong, Shaohua
    Zhou, Wujie
    Xu, Caie
    Yan, Weiqing
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (01) : 657 - 669
  • [7] Boundary-Semantic Collaborative Guidance Network With Dual-Stream Feedback Mechanism for Salient Object Detection in Optical Remote Sensing Imagery
    Feng, Dejun
    Chen, Hongyu
    Liu, Suning
    Liao, Ziyang
    Shen, Xingyu
    Xie, Yakun
    Zhu, Jun
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61 : 1 - 17
  • [8] Semantic-Guided Attention Refinement Network for Salient Object Detection in Optical Remote Sensing Images
    Huang, Zhou
    Chen, Huaixin
    Liu, Biyuan
    Wang, Zhixi
    [J]. REMOTE SENSING, 2021, 13 (11)
  • [9] Segment Anything
    Kirillov, Alexander
    Mintun, Eric
    Ravi, Nikhila
    Mao, Hanzi
    Rolland, Chloe
    Gustafson, Laura
    Xiao, Tete
    Whitehead, Spencer
    Berg, Alexander C.
    Lo, Wan-Yen
    Dolla'r, Piotr
    Girshick, Ross
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 3992 - 4003
  • [10] Lee MS, 2022, AAAI CONF ARTIF INTE, P12993