Building Extraction from Remote Sensing Images Based on Improved U-Net

被引:4
作者
Jin Shu [1 ]
Guan Mo [1 ]
Bian Yuchan [2 ]
Wang Shulei [1 ]
机构
[1] Shenyang Univ Technol, Sch Informat Sci & Engn, Shenyang 110870, Liaoning, Peoples R China
[2] Shenyang Univ Technol, Sch Software, Shenyang 110870, Liaoning, Peoples R China
关键词
remote sensing image; semantic segmentation; building extraction; attention mechanism; multi-scale;
D O I
10.3788/LOP213004
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Building extraction from remote sensing images is of great significance to the construction of smart cities. Aiming to improve the low accuracy of traditional methods in extracting remote sensing images with a complex background, a remote sensing image building extraction method (MA-Unet) based on U-Net is proposed. This method mainly uses an encoder and a decoder. A convolutional block attention module is introduced into the encoder, in which a channel attention module is used to screen more important features and suppress invalid features, and a spatial attention module is used to screen deeper semantic features. An atrous spatial pyramid pooling module is introduced to extract features with different scales. In the decoder, to fuse object features with different scales, feature maps in the decoder are upsampled and connected in series. This information aggregation solves the difficulty of detecting objects with different scales to some extent. The experimental results show that MA-Unet method is superior to the U-Net method in terms of accuracy, precision, and intersection over union (IoU) by 1. 7 percentage points, 2. 1 percentage points, and 1. 6 percentage points on the Massachusetts building dataset and by 1. 1 percentage points, 1. 4 percentage points, and 2. 3 percentage points on the WHU building dataset, respectively. It is a more effective and practical target extraction method.
引用
收藏
页数:7
相关论文
共 20 条
[1]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[2]  
Chen L., 2021, arXiv
[3]  
Chen LC, 2016, Arxiv, DOI arXiv:1412.7062
[4]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851
[5]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[6]  
He KM, 2014, LECT NOTES COMPUT SC, V8691, P346, DOI [arXiv:1406.4729, 10.1007/978-3-319-10578-9_23]
[7]  
[何小飞 He Xiaofei], 2016, [测绘学报, Acta Geodetica et Cartographica Sinica], V45, P1073
[8]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/TPAMI.2019.2913372, 10.1109/CVPR.2018.00745]
[9]  
[季顺平 Ji Shunping], 2019, [测绘学报, Acta Geodetica et Cartographica Sinica], V48, P448
[10]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001