A Gather-to-Guide Network for Remote Sensing Semantic Segmentation of RGB and Auxiliary Image

被引:25
作者
Zheng, Xianwei [1 ]
Wu, Xiujie [1 ]
Huan, Linxi [1 ]
He, Wei [2 ]
Zhang, Hongyan [1 ]
机构
[1] Wuhan Univ, State Key Lab Informat Engn Surveying Mapping & R, Wuhan 430079, Peoples R China
[2] RIKEN, RIKEN Ctr Adv Intelligence Project, Tokyo 1030027, Japan
来源
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING | 2022年 / 60卷
基金
中国国家自然科学基金;
关键词
Semantics; Image segmentation; Remote sensing; Feature extraction; Convolutional neural networks; Calibration; Task analysis; Deep learning; remote sensing; semantic segmentation; CONVOLUTIONAL NETWORKS; RESOLUTION; CLASSIFICATION;
D O I
10.1109/TGRS.2021.3103517
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Convolutional neural network (CNN)-based feature fusion of RGB and auxiliary remote sensing data is known to enable improved semantic segmentation. However, such fusion is challengeable because of the substantial variance in data characteristics and quality (e.g., data uncertainties and misalignment) between two modality data. In this article, we propose a unified gather-to-guide network (G2GNet) for remote sensing semantic segmentation of RGB and auxiliary data. The key aspect of the proposed architecture is a novel gather-to-guide module (G2GM) that consists of a feature gatherer and a feature guider. The feature gatherer generates a set of cross-modal descriptors by absorbing the complementary merits of RGB and auxiliary modality data. The feature guider calibrates the RGB feature response by using the channel-wise guide weights extracted from the cross-modal descriptors. In this way, the G2GM can perform RGB feature calibration with different modality data in a gather-to-guide fashion, thus preserving the informative features while suppressing redundant and noisy information. Extensive experiments conducted on two benchmark datasets show that the proposed G2GNet is robust to data uncertainties while also improving the semantic segmentation performance of RGB and auxiliary remote sensing data.
引用
收藏
页数:15
相关论文
共 68 条
  • [1] [Anonymous], 2017, IEEE transactions on pattern analysis and machine intelligence, DOI DOI 10.1109/TPAMI.2016.2644615
  • [2] Beyond RGB: Very high resolution urban remote sensing with multimodal deep networks
    Audebert, Nicolas
    Le Saux, Bertrand
    Lefevre, Sebastien
    [J]. ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2018, 140 : 20 - 32
  • [3] Semantic Segmentation of Earth Observation Data Using Multimodal and Multi-scale Deep Networks
    Audebert, Nicolas
    Le Saux, Bertrand
    Lefevre, Sebastien
    [J]. COMPUTER VISION - ACCV 2016, PT I, 2017, 10111 : 180 - 196
  • [4] C3Net: Cross-Modal Feature Recalibrated, Cross-Scale Semantic Aggregated and Compact Network for Semantic Segmentation of Multi-Modal High-Resolution Aerial Images
    Cao, Zhiying
    Diao, Wenhui
    Sun, Xian
    Lyu, Xiaode
    Yan, Menglong
    Fu, Kun
    [J]. REMOTE SENSING, 2021, 13 (03)
  • [5] Chen KQ, 2019, INT GEOSCI REMOTE SE, P3911, DOI [10.1109/IGARSS.2019.8899217, 10.1109/igarss.2019.8899217]
  • [6] DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs
    Chen, Liang-Chieh
    Papandreou, George
    Kokkinos, Iasonas
    Murphy, Kevin
    Yuille, Alan L.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) : 834 - 848
  • [7] CCANet: Class-Constraint Coarse-to-Fine Attentional Deep Network for Subdecimeter Aerial Image Semantic Segmentation
    Deng, Guohui
    Wu, Zhaocong
    Wang, Chengjun
    Xu, Miaozhong
    Zhong, Yanfei
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [8] Fernandez-Beltran Ruben, 2018, IEEE Geoscience and Remote Sensing Letters, V15, P1347, DOI 10.1109/LGRS.2018.2843886
  • [9] Dual Attention Network for Scene Segmentation
    Fu, Jun
    Liu, Jing
    Tian, Haijie
    Li, Yong
    Bao, Yongjun
    Fang, Zhiwei
    Lu, Hanqing
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 3141 - 3149
  • [10] Gerke M., 2014, Use of the stair vision library within the ISPRS 2D semantic labeling benchmark (Vaihingen), DOI 10.13140/2.1.5015.9683