A Dual Attention Neural Network for Airborne LiDAR Point Cloud Semantic Segmentation

被引:29
作者
Zhang, Ka [1 ,2 ]
Ye, Longjie [1 ]
Xiao, Wen [3 ]
Sheng, Yehua [1 ]
Zhang, Shan [1 ]
Tao, Xia [1 ]
Zhou, Yaqin [1 ]
机构
[1] Nanjing Normal Univ, Sch Geog, Key Lab Virtual Geog Environm, Nanjing 210023, Peoples R China
[2] Minist Nat Resources, Key Lab Urban Land Resources Monitoring & Simulat, Shenzhen 518034, Peoples R China
[3] China Univ Geosci, Natl Engn Res Ctr Geog Informat Syst, Sch Geog & Informat Engn, Wuhan 430074, Peoples R China
来源
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING | 2022年 / 60卷
基金
中国国家自然科学基金;
关键词
Airborne light detection and ranging (LiDAR); dual attention neural network (DA-Net); location-homogeneity module; point cloud classification; Vaihingen dataset; DEEP FEATURES; CLASSIFICATION; EXTRACTION; NET; RECOGNITION; FUSION; MODELS;
D O I
10.1109/TGRS.2022.3201902
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
With the development of airborne light detection and ranging (LiDAR) technology, it has become a common and efficient way to collect large-scale 3-D spatial information. However, efficient and automatic semantic segmentation of LiDAR data, in the form of 3-D point clouds, remains a persistent challenge. To address this, a dual attention neural network (DA-Net) is proposed, consisting of two different blocks, namely, augmented edge representation (AER) and elevation attentive pooling (EAP). First, the AER can adaptively represent local orientation and position, thereby effectively enhancing geometric information. Second, the captured local features of centroid points are utilized to further encode discriminative features using the EAP with the learned attention scores. Finally, a location homogeneity (LH) module is devised to explore the long-range relationship in an encoderdecoder network. Benefiting from the dual attention module, geometric information hidden in unorganized point clouds can be effectively propagated. Besides, the LH forces the network to pay attention to the semantic consistency of elevated objects, which facilitates both point- and object-level point cloud semantic segmentation for scene understanding. A benchmark dataset is used to assess the proposed method, which achieves an overall accuracy of 85.98% and an average F1 score of 72.31%. In addition, comparisons with other latest deep learning methods on the 2019 Data Fusion Contest dataset further demonstrate the robustness and generalization ability of the proposed method.
引用
收藏
页数:17
相关论文
共 78 条
[1]   Addressing overfitting on point cloud classification using Atrous XCRF [J].
Arief, Hasan Asy'ari ;
Indahl, Ulf Geir ;
Strand, Geir-Harald ;
Tveite, Havard .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2019, 155 :90-101
[2]   3D Semantic Parsing of Large-Scale Indoor Spaces [J].
Armeni, Iro ;
Sener, Ozan ;
Zamir, Amir R. ;
Jiang, Helen ;
Brilakis, Ioannis ;
Fischer, Martin ;
Savarese, Silvio .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1534-1543
[3]  
Axelsson P., 2000, ISPRS-International Archives of the photogrammetry. Remote Sens. Spat. Inf. Sci., V33, P110
[4]   3DmFV: Three-Dimensional Point Cloud Classification in Real-Time Using Convolutional Neural Networks [J].
Ben-Shabat, Yizhak ;
Lindenbaum, Michael ;
Fischer, Anath .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :3145-3152
[5]  
Chen C, 2019, Arxiv, DOI arXiv:1905.08705
[6]   Building change detection in very high-resolution remote sensing image based on pseudo-orthorectification [J].
Chen, Hui ;
Zhang, Ka ;
Xiao, Wen ;
Sheng, Yehua ;
Cheng, Liang ;
Zhou, Wei ;
Wang, Pengbo ;
Su, Dong ;
Ye, Longjie ;
Zhang, Shan .
INTERNATIONAL JOURNAL OF REMOTE SENSING, 2021, 42 (07) :2686-2705
[7]  
Contreras J, 2019, INT GEOSCI REMOTE SE, P5236, DOI [10.1109/IGARSS.2019.8899303, 10.1109/igarss.2019.8899303]
[8]   Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis [J].
Dai, Angela ;
Qi, Charles Ruizhongtai ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6545-6554
[9]   ScanNet: Richly-annotated 3D Reconstructions of Indoor Scenes [J].
Dai, Angela ;
Chang, Angel X. ;
Savva, Manolis ;
Halber, Maciej ;
Funkhouser, Thomas ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2432-2443
[10]  
Fan S., 2021, P IEEECVF C COMPUTER, P14504