Edge Guided Context Aggregation Network for Semantic Segmentation of Remote Sensing Imagery

被引:11
作者
Liu, Zhiqiang [1 ]
Li, Jiaojiao [1 ]
Song, Rui [1 ]
Wu, Chaoxiong [1 ]
Liu, Wei [2 ]
Li, Zan [1 ]
Li, Yunsong [1 ]
机构
[1] Xidian Univ, State Key Lab Integrated Serv Networks, Xian 710000, Peoples R China
[2] State Key Lab Geoinformat Engn, Xian 710054, Peoples R China
基金
中国博士后科学基金;
关键词
remote sensing imagery; semantic segmentation; deep learning; context aggregation; CLASSIFIER; FOREST;
D O I
10.3390/rs14061353
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Semantic segmentation of remote sensing imagery (RSI) has obtained great success with the development of deep convolutional neural networks (DCNNs). However, most of the existing algorithms focus on designing end-to-end DCNNs, but neglecting to consider the difficulty of segmentation in imbalance categories, especially for minority categories in RSI, which limits the performance of RSI semantic segmentation. In this paper, a novel edge guided context aggregation network (EGCAN) is proposed for the semantic segmentation of RSI. The Unet is employed as backbone. Meanwhile, an edge guided context aggregation branch and minority categories extraction branch are designed for a comprehensive enhancement of semantic modeling. Specifically, the edge guided context aggregation branch is proposed to promote entire semantic comprehension of RSI and further emphasize the representation of edge information, which consists of three modules: edge extraction module (EEM), dual expectation maximization attention module (DEMA), and edge guided module (EGM). EEM is created primarily for accurate edge tracking. According to that, DEMA aggregates global contextual features with different scales and the edge features along spatial and channel dimensions. Subsequently, EGM cascades the aggregated features into the decoder process to capture long-range dependencies and further emphasize the error-prone pixels in the edge region to acquire better semantic labels. Besides this, the exploited minority categories extraction branch is presented to acquire rich multi-scale contextual information through an elaborate hybrid spatial pyramid pooling module (HSPP) to distinguish categories taking a small percentage and background. On the Tianzhi Cup dataset, the proposed algorithm EGCAN achieved an overall accuracy of 84.1% and an average cross-merge ratio of 68.1%, with an accuracy improvement of 0.4% and 1.3% respectively compared to the classical Deeplabv3+ model. Extensive experimental results on the dataset released in ISPRS Vaihingen and Potsdam benchmarks also demonstrate the effectiveness of the proposed EGCAN over other state-of-the-art approaches.
引用
收藏
页数:22
相关论文
共 57 条
[1]   Multi-Object Segmentation in Complex Urban Scenes from High-Resolution Remote Sensing Data [J].
Abdollahi, Abolfazl ;
Pradhan, Biswajeet ;
Shukla, Nagesh ;
Chakraborty, Subrata ;
Alamri, Abdullah .
REMOTE SENSING, 2021, 13 (18)
[2]   Beyond RGB: Very high resolution urban remote sensing with multimodal deep networks [J].
Audebert, Nicolas ;
Le Saux, Bertrand ;
Lefevre, Sebastien .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2018, 140 :20-32
[3]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[4]  
Bahdanau D., 2016, ARXIV 14090473
[5]   3D Instance Segmentation and Object Detection Framework Based on the Fusion of Lidar Remote Sensing and Optical Image Sensing [J].
Bai, Ling ;
Li, Yinguo ;
Cen, Ming ;
Hu, Fangchao .
REMOTE SENSING, 2021, 13 (16)
[6]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851
[7]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[8]  
Chen LB, 2017, IEEE INT SYMP NANO, P1, DOI 10.1109/NANOARCH.2017.8053709
[9]  
Chen Y., 2018, P ADV NEURAL INFORM
[10]   SPGNet: Semantic Prediction Guidance for Scene Parsing [J].
Cheng, Bowen ;
Chen, Liang-Chieh ;
Wei, Yunchao ;
Zhu, Yukun ;
Huang, Zilong ;
Xiong, Jinjun ;
Huang, Thomas S. ;
Hwu, Wen-Mei ;
Shi, Honghui .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :5217-5227