Iterative Saliency Aggregation and Assignment Network for Efficient Salient Object Detection in Optical Remote Sensing Images

被引:6
作者
Yao, Zhaojian [1 ]
Gao, Wei [1 ,2 ]
机构
[1] Peking Univ, Sch Elect & Comp Engn, Shenzhen Grad Sch, Shenzhen 518055, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518066, Peoples R China
来源
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING | 2024年 / 62卷
关键词
Decoding; Accuracy; Object detection; Feature extraction; Iterative methods; Object recognition; Hierarchical interaction and dynamic integration (HIDI); iterative saliency aggregation and assignment; remote sensing; salient object detection;
D O I
10.1109/TGRS.2024.3425658
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Motivated by the pursuit of efficient salient object detection in remote sensing, researchers have devoted considerable efforts to devising lightweight models due to low running efficiency of the cumbersome models. Although the existing lightweight models have made impressive progress in improving efficiency, there is often a notable sacrifice in inference accuracy, making it challenging to attain both high-quality output and high efficiency. In this article, we propose an iterative saliency aggregation and assignment network (ISAANet) to reconcile the dilemma of balancing accuracy and efficiency. ISAANet adopts a recurrent architecture with bidirectional communication between the encoder and the decoder to boost detection performance. Specifically, a saliency aggregation mechanism is used to generate complementary information by integrating multilevel saliency cues, corresponding to the feedforward information flow from the encoder to the decoder. As for the feedback flows from the decoder to the encoder, a saliency assignment approach is proposed to inject the region and edge prompts into the encoder for hierarchical feature updates, emphasizing the representation of salient information. The encoder is progressively reinforced by alternating iterations of saliency aggregation and assignment, which further helps the decoder to produce more reliable detection results. Moreover, the effective fusion of cross-level features is key to realizing information complementarity and enhancing aggregation quality. However, conventional fusion schemes often neglect feature importance disparities, which can lead to noise accumulation. To address this issue, we introduce a hierarchical interaction and dynamic integration (HIDI) module for feature fusion. The HIDI module groups fusion features and dynamically adjusts their weights based on their information quality, which suppresses noise interference and strengthens the model's learning of meaningful features. The experimental results demonstrate that the proposed ISAANet not only effectively compensates for the accuracy deficiency of the existing lightweight models but also achieves a faster running speed of 108 FPS, which shows a favorable tradeoff between efficiency and accuracy. The code will be available at https://github.com/YiuCK/ISAANet and https://openi.pcl.ac.cn/OpenVision/ISAANet.
引用
收藏
页数:13
相关论文
共 57 条
[11]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[12]   SpectralGPT: Spectral Remote Sensing Foundation Model [J].
Hong, Danfeng ;
Zhang, Bing ;
Li, Xuyang ;
Li, Yuxuan ;
Li, Chenyu ;
Yao, Jing ;
Yokoya, Naoto ;
Li, Hao ;
Ghamisi, Pedram ;
Jia, Xiuping ;
Plaza, Antonio ;
Gamba, Paolo ;
Benediktsson, Jon Atli ;
Chanussot, Jocelyn .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (08) :5227-5244
[13]   Searching for MobileNetV3 [J].
Howard, Andrew ;
Sandler, Mark ;
Chu, Grace ;
Chen, Liang-Chieh ;
Chen, Bo ;
Tan, Mingxing ;
Wang, Weijun ;
Zhu, Yukun ;
Pang, Ruoming ;
Vasudevan, Vijay ;
Le, Quoc V. ;
Adam, Hartwig .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1314-1324
[14]   MoADNet: Mobile Asymmetric Dual-Stream Networks for Real-Time and Lightweight RGB-D Salient Object Detection [J].
Jin, Xiao ;
Yi, Kang ;
Xu, Jing .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (11) :7632-7645
[15]   CasFormer: Cascaded transformers for fusion-aware computational hyperspectral imaging [J].
Li, Chenyu ;
Zhang, Bing ;
Hong, Danfeng ;
Zhou, Jun ;
Vivone, Gemine ;
Li, Shutao ;
Chanussot, Jocelyn .
INFORMATION FUSION, 2024, 108
[16]   LRR-Net: An Interpretable Deep Unfolding Network for Hyperspectral Anomaly Detection [J].
Li, Chenyu ;
Zhang, Bing ;
Hong, Danfeng ;
Yao, Jing ;
Chanussot, Jocelyn .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
[17]   Nested Network With Two-Stream Pyramid for Salient Object Detection in Optical Remote Sensing Images [J].
Li, Chongyi ;
Cong, Runmin ;
Hou, Junhui ;
Zhang, Sanyi ;
Qian, Yue ;
Kwong, Sam .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2019, 57 (11) :9156-9166
[18]   Salient Object Detection in Optical Remote Sensing Images Driven by Transformer [J].
Li, Gongyang ;
Bai, Zhen ;
Liu, Zhi ;
Zhang, Xinpeng ;
Ling, Haibin .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 :5257-5269
[19]   Lightweight Salient Object Detection in Optical Remote-Sensing Images via Semantic Matching and Edge Alignment [J].
Li, Gongyang ;
Liu, Zhi ;
Zhang, Xinpeng ;
Lin, Weisi .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
[20]   Lightweight Salient Object Detection in Optical Remote Sensing Images via Feature Correlation [J].
Li, Gongyang ;
Liu, Zhi ;
Bai, Zhen ;
Lin, Weisi ;
Ling, Haibin .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60