Directionally constrained fully convolutional neural network for airborne LiDAR point cloud classification

被引:95
作者
Wen, Congcong [1 ,2 ,3 ]
Yang, Lina [1 ,2 ]
Li, Xiang [1 ,2 ,3 ]
Peng, Ling [1 ,2 ]
Chi, Tianhe [1 ,2 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] NYU, Tandon Sch Engn, New York, NY USA
关键词
Airborne LiDAR; Point cloud classification; Directionlly constrained nearest neighbor; Fully convolution networks; ISPRS 3D labeling; CONTEXTUAL CLASSIFICATION;
D O I
10.1016/j.isprsjprs.2020.02.004
中图分类号
P9 [自然地理学];
学科分类号
0705 ; 070501 ;
摘要
Point cloud classification plays an important role in a wide range of airborne light detection and ranging (LiDAR) applications, such as topographic mapping, forest monitoring, power line detection, and road detection. However, due to the sensor noise, high redundancy, incompleteness, and complexity of airborne LiDAR systems, point cloud classification is challenging. Traditional point cloud classification methods mostly focus on the development of handcrafted point geometry features and employ machine learning-based classification models to conduct point classification. In recent years, the advances of deep learning models have caused researchers to shift their focus towards machine learning-based models, specifically deep neural networks, to classify airborne LiDAR point clouds. These learning-based methods start by transforming the unstructured 3D point sets to regular 2D representations, such as collections of feature images, and then employ a 2D CNN for point classification. Moreover, these methods usually need to calculate additional local geometry features, such as planarity, sphericity and roughness, to make use of the local structural information in the original 3D space. Nonetheless, the 3D to 2D conversion results in information loss. In this paper, we propose a directionally constrained fully convolutional neural network (D-FCN) that can take the original 3D coordinates and LiDAR intensity as input; thus, it can directly apply to unstructured 3D point clouds for semantic labeling. Specifically, we first introduce a novel directionally constrained point convolution (D-Conv) module to extract locally representative features of 3D point sets from the projected 2D receptive fields. To make full use of the orientation information of neighborhood points, the proposed D-Conv module performs convolution in an orientation-aware manner by using a directionally constrained nearest neighborhood search. Then, we design a multiscale fully convolutional neural network with downsampling and upsampling blocks to enable multiscale point feature learning. The proposed D-FCN model can therefore process input point cloud with arbitrary sizes and directly predict the semantic labels for all the input points in an end-to-end manner. Without involving additional geometry features as input, the proposed method demonstrates superior performance on the International Society for Photogrammetry and Remote Sensing (ISPRS) 3D labeling benchmark dataset. The results show that our model achieves a new stateof-the-art performance on powerline, car, and facade categories. Moreover, to demonstrate the generalization abilities of the proposed method, we conduct further experiments on the 2019 Data Fusion Contest Dataset. Our proposed method achieves superior performance than the comparing methods and accomplishes an overall accuracy of 95.6% and an average F1 score of 0.810.
引用
收藏
页码:50 / 62
页数:13
相关论文
共 56 条
[1]   Estimating forest canopy fuel parameters using LIDAR data [J].
Andersen, HE ;
McGaughey, RJ ;
Reutebuch, SE .
REMOTE SENSING OF ENVIRONMENT, 2005, 94 (04) :441-449
[2]  
[Anonymous], INT ARCH PHOT REM 3A
[3]  
[Anonymous], 2006, International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences (ISPRS)
[4]  
[Anonymous], 2009, Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci
[5]  
[Anonymous], 2015, The impact of imbalanced training data for convolutional neural networks
[6]  
[Anonymous], 2012, 22 ISPRS C TECHN COM
[7]   Addressing overfitting on point cloud classification using Atrous XCRF [J].
Arief, Hasan Asy'ari ;
Indahl, Ulf Geir ;
Strand, Geir-Harald ;
Tveite, Havard .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2019, 155 :90-101
[8]  
Axelsson P., 2000, The International Archives of the Photogrammetry and Remote Sensing, Amsterdam, The Netherlands, VXXXIII, P110
[9]   Urban 3D segmentation and modelling from street view images and LiDAR point clouds [J].
Babahajiani, Pouria ;
Fan, Lixin ;
Kamarainen, Joni-Kristian ;
Gabbouj, Moncef .
MACHINE VISION AND APPLICATIONS, 2017, 28 (07) :679-694
[10]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495