Deeply Hybrid Contrastive Learning Based on Semantic Pseudo-Label for Salient Object Detection in Optical Remote Sensing Images

被引:2
作者
Qiu, Yu [1 ,2 ]
Sun, Yuhang [1 ]
Mei, Jie [3 ]
Xu, Jing [1 ]
机构
[1] Nankai Univ, Coll Artificial Intelligence, Tianjin 300350, Peoples R China
[2] Hunan Normal Univ, Coll Informat Sci & Engineer ing, Changsha 410081, Peoples R China
[3] Hunan Univ, Natl Engn Res Ctr Robot Visual Percept & Control T, Sch Robot, Changsha 410082, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Feature extraction; Semantics; Visualization; Training; Task analysis; Remote sensing; Image edge detection; Salient object detection; remote sensing images; pseudo-label; hybrid contrast; hard edge contrast; NETWORK; FUSION;
D O I
10.1109/TMM.2024.3414669
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Salient object detection in natural scene images (NSI-SOD) has undergone remarkable advancements in recent years. However, compared to those of natural images, the properties of remote sensing images (ORSIs), such as diverse spatial resolutions, complex background structures, and varying visual attributes of objects, are more complicated. Hence, how to explore the multiscale structural perceptual information of ORSIs to accurately detect salient objects is more challenging. In this paper, inspired by the superiority of contrastive learning, we propose a novel training paradigm for ORSI-SOD, named Deeply Hybrid Contrastive Learning Based on Semantic Pseudo-Label (DHCont), to force the network to extract rich structural perceptual information and further learn the better-structured feature embedding spaces. Specifically, DHCont first splits the ORSI into several local subregions composed of color- and texture-similar pixels, which act as semantic pseudo-labels. This strategy can effectively explore the underdeveloped semantic categories in ORSI-SOD. To delve deeper into multiscale structure-aware optimization, DHCont incorporates a hybrid contrast strategy that integrates "pixel-to-pixel", "region-to-region", "pixel-to-region", and "region-to-pixel" contrasts at multiple scales. Additionally, to enhance the edge details of salient regions, we develop a hard edge contrast strategy that focuses on improving the detection accuracy of hard pixels near the object boundary. Moreover, we introduce a deep contrast algorithm that adds additional deep-level constraints to the feature spaces of multiple stages. Extensive experiments on two popular ORSI-SOD datasets demonstrate that simply integrating our DHCont into the existing ORSI-SOD models can significantly improve the performance.
引用
收藏
页码:10892 / 10907
页数:16
相关论文
共 85 条
[1]   SLIC Superpixels Compared to State-of-the-Art Superpixel Methods [J].
Achanta, Radhakrishna ;
Shaji, Appu ;
Smith, Kevin ;
Lucchi, Aurelien ;
Fua, Pascal ;
Suesstrunk, Sabine .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (11) :2274-2281
[2]  
Achanta R, 2009, PROC CVPR IEEE, P1597, DOI 10.1109/CVPRW.2009.5206596
[3]   The Lovasz-Softmax loss: A tractable surrogate for the optimization of the intersection-over-union measure in neural networks [J].
Berman, Maxim ;
Triki, Amal Rannen ;
Blaschko, Matthew B. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4413-4421
[4]   Exploring Rich and Efficient Spatial Temporal Interactions for Real-Time Video Salient Object Detection [J].
Chen, Chenglizhao ;
Wang, Guotao ;
Peng, Chong ;
Fang, Yuming ;
Zhang, Dingwen ;
Qin, Hong .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 :3995-4007
[5]   Improved Robust Video Saliency Detection Based on Long-Term Spatial-Temporal Information [J].
Chen, Chenglizhao ;
Wang, Guotao ;
Peng, Chong ;
Zhang, Xiaowei ;
Qin, Hong .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :1090-1100
[6]   Video Saliency Detection via Spatial-Temporal Fusion and Low-Rank Coherency Diffusion [J].
Chen, Chenglizhao ;
Li, Shuai ;
Wang, Yongguang ;
Qin, Hong ;
Hao, Aimin .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (07) :3156-3170
[7]   Saliency Guided Inter- and Intra-Class Relation Constraints for Weakly Supervised Semantic Segmentation [J].
Chen, Tao ;
Yao, Yazhou ;
Zhang, Lei ;
Wang, Qiong ;
Xie, Guo-Sen ;
Shen, Fumin .
IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 :1727-1737
[8]  
Chen T, 2020, PR MACH LEARN RES, V119
[9]   Temporal difference-guided network for hyperspectral image change detection [J].
Chen, Zhonghao ;
Wang, Yuyang ;
Gao, Hongmin ;
Ding, Yao ;
Zhong, Qiqiang ;
Hong, Danfeng ;
Zhang, Bing .
INTERNATIONAL JOURNAL OF REMOTE SENSING, 2023, 44 (19) :6033-6059
[10]   Grid Network: Feature Extraction in Anisotropic Perspective for Hyperspectral Image Classification [J].
Chen, Zhonghao ;
Hong, Danfeng ;
Gao, Hongmin .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2023, 20