An attention-fused network for semantic segmentation of very-high-resolution remote sensing imagery

被引:141
作者
Yang, Xuan [1 ,5 ]
Li, Shanshan [2 ]
Chen, Zhengchao [3 ]
Chanussot, Jocelyn [1 ]
Jia, Xiuping [4 ]
Zhang, Bing [1 ,5 ]
Li, Baipeng [3 ]
Chen, Pan [1 ,5 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Digital Earth Sci, Beijing 100094, Peoples R China
[2] Chinese Acad Sci, Aerosp Informat Res Inst, China Remote Sensing Satellite Ground Stn, Beijing 100094, Peoples R China
[3] Chinese Acad Sci, Airborne Remote Sensing Ctr, Aerosp Informat Res Inst, Beijing 100094, Peoples R China
[4] Univ New South Wales, Australian Def Force Acad, Sch Engn & Informat Technol, Canberra, ACT 2612, Australia
[5] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
关键词
Semantic segmentation; Deep learning; Very-high-resolution imagery; Attention-fused network; ISPRS; Convolutional neural network; LAND-COVER; CLASSIFICATION; CHALLENGES;
D O I
10.1016/j.isprsjprs.2021.05.004
中图分类号
P9 [自然地理学];
学科分类号
0705 ; 070501 ;
摘要
Semantic segmentation is an essential part of deep learning. In recent years, with the development of remote sensing big data, semantic segmentation has been increasingly used in remote sensing. Deep convolutional neural networks (DCNNs) face the challenge of feature fusion: very-high-resolution remote sensing image multisource data fusion can increase the network's learnable information, which is conducive to correctly classifying target objects by DCNNs; simultaneously, the fusion of high-level abstract features and low-level spatial features can improve the classification accuracy at the border between target objects. In this paper, we propose a multipath encoder structure to extract features of multipath inputs, a multipath attention-fused block module to fuse multipath features, and a refinement attention-fused block module to fuse high-level abstract features and low-level spatial features. Furthermore, we propose a novel convolutional neural network architecture, named attention-fused network (AFNet). Based on our AFNet, we achieve state-of-the-art performance with an overall accuracy of 91.7% and a mean F1 score of 90.96% on the ISPRS Vaihingen 2D dataset and an overall accuracy of 92.1% and a mean F1 score of 93.44% on the ISPRS Potsdam 2D dataset.
引用
收藏
页码:238 / 262
页数:25
相关论文
共 58 条
[1]  
[Anonymous], PROC CVPR IEEE
[2]  
[Anonymous], 2015, ACS SYM SER
[3]  
[Anonymous], 2016, P 29 IEEE C COMPUTER
[4]  
[Anonymous], 2015, AUTOMATIC SEMANTIC L
[5]   Beyond RGB: Very high resolution urban remote sensing with multimodal deep networks [J].
Audebert, Nicolas ;
Le Saux, Bertrand ;
Lefevre, Sebastien .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2018, 140 :20-32
[6]   Semantic Segmentation of Earth Observation Data Using Multimodal and Multi-scale Deep Networks [J].
Audebert, Nicolas ;
Le Saux, Bertrand ;
Lefevre, Sebastien .
COMPUTER VISION - ACCV 2016, PT I, 2017, 10111 :180-196
[7]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[8]   Comprehensive survey of deep learning in remote sensing: theories, tools, and challenges for the community [J].
Ball, John E. ;
Anderson, Derek T. ;
Chan, Chee Seng .
JOURNAL OF APPLIED REMOTE SENSING, 2017, 11
[9]   Very High-Resolution Remote Sensing: Challenges and Opportunities [J].
Benediktsson, Jon Atli ;
Chanussot, Jocelyn ;
Moon, Wooil M. .
PROCEEDINGS OF THE IEEE, 2012, 100 (06) :1907-1910
[10]   Assessment of very high spatial resolution satellite image segmentations [J].
Carleer, AP ;
Debeir, O ;
Wolff, E .
PHOTOGRAMMETRIC ENGINEERING AND REMOTE SENSING, 2005, 71 (11) :1285-1294