PCL-PTD Net: Parallel Cross-Learning-Based Pixel Transferred Deconvolutional Network for Building Extraction in Dense Building Areas With Shadow

被引:7
作者
Boonpook, Wuttichai [1 ]
Tan, Yumin [2 ]
Torsri, Kritanai [3 ]
Kamsing, Patcharin [4 ]
Torteeka, Peerapong [5 ]
Nardkulpat, Attawut [1 ]
机构
[1] Srinakharinwirot Univ, Fac Social Sci, Dept Geog, Bangkok 10110, Thailand
[2] Beihang Univ, Sch Transportat Sci & Engn, Beijing 100191, Peoples R China
[3] Minist Higher Educ Sci Res & Innovat, Hydroinformat Inst, Bangkok 10900, Thailand
[4] Int Acad Aviat Ind, King Mongkuts Inst Technol Ladkrabang, Dept Aeronaut Engn, Air Space Control Optimizat & Management Lab, Bangkok 10520, Thailand
[5] Natl Astron Res Inst Thailand, Chiangmai 10520, Thailand
关键词
Buildings; Feature extraction; Remote sensing; Deconvolution; Convolutional neural networks; Semantic segmentation; Decoding; Building extraction; building shadow; dense building; PCL-PTD net; semantic segmentation;
D O I
10.1109/JSTARS.2022.3230149
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Urban building segmentation from remote sensed imageries is challenging because there usually exists a variety of building features. Furthermore, very high spatial resolution imagery can provide many details of the urban building, such as styles, small gaps among buildings, building shadows, etc. Hence, satisfactory accuracy in detecting and extracting urban features from highly detailed images still remains. Deep learning semantic segmentation using baseline networks works well on building extraction; however, their ability in building extraction in shadows area, unclear building feature, and narrow gaps among buildings in dense building zone is still limited. In this article, we propose parallel cross-learning-based pixel transferred deconvolutional network (PCL-PTD net), and then is used to segment urban buildings from aerial photographs. The proposed method is evaluated and intercompared with traditional baseline networks. In PCL-PTD net, it is composed of parallel network, cross-learning functions, residual unit in encoder part, and PTD in decoder part. The performance is applied to three datasets (Inria aerial dataset, international society for photogrammetry and remote sensing Potsdam dataset, and UAV building dataset), to evaluate its accuracy and robustness. As a result, we found that PCL-PTD net can improve learning capacities of the supervised learning model in differentiating buildings in dense area and extracting buildings covered by shadows. As compared to the baseline networks, we found that proposed network shows superior performance compared to all eight networks (SegNet, U-net, pyramid scene parsing network, PixelDCL, DeeplabV3+, U-Net++, context feature enhancement networ, and improved ResU-Net). The experiments on three datasets also demonstrate the ability of proposed framework and indicating its performance.
引用
收藏
页码:773 / 786
页数:14
相关论文
共 35 条
[1]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[2]   Res2-Unet, a New Deep Architecture for Building Detection From High Spatial Resolution Images [J].
Chen, Fang ;
Wang, Ning ;
Yu, Bo ;
Wang, Lei .
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2022, 15 :1494-1501
[3]   A Context Feature Enhancement Network for Building Extraction from High-Resolution Remote Sensing Imagery [J].
Chen, Jinzhi ;
Zhang, Dejun ;
Wu, Yiqi ;
Chen, Yilin ;
Yan, Xiaohu .
REMOTE SENSING, 2022, 14 (09)
[4]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[5]   ResUNet-a: A deep learning framework for semantic segmentation of remotely sensed data [J].
Diakogiannis, Foivos, I ;
Waldner, Francois ;
Caccetta, Peter ;
Wu, Chen .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2020, 162 :94-114
[6]  
Ding Y., 2020, ISPRS-Int. Arch. Photogramm. Remote. Sens. Spat. Inf. Sci, V43, P35, DOI [10.5194/isprs-archives-XLIII-B3-2020-35-2020, DOI 10.5194/ISPRS-ARCHIVES-XLIII-B3-2020-35-2020]
[7]   Dual Attention Network for Scene Segmentation [J].
Fu, Jun ;
Liu, Jing ;
Tian, Haijie ;
Li, Yong ;
Bao, Yongjun ;
Fang, Zhiwei ;
Lu, Hanqing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3141-3149
[8]  
Gao HY, 2017, Arxiv, DOI arXiv:1705.06820
[9]   Contextual Attention Refinement Network for Real-Time Semantic Segmentation [J].
Hao, Shijie ;
Zhou, Yuan ;
Zhang, Youming ;
Guo, Yanrong .
IEEE ACCESS, 2020, 8 :55230-55240
[10]  
ISPRS, 2018, 2D POTSD SEM LAB DAT