Lidar Point Cloud Segmentation Model Based on Improved PointNet++

被引:2
作者
Zhang Chi [1 ]
Wang Zhijie [1 ]
Wu Hao [1 ]
Chen Dong [1 ]
机构
[1] Nanjing Forestry Univ, Coll Civil Engn, Nanjing 210037, Jiangsu, Peoples R China
关键词
point cloud segmentation; PointNet++; Lidar; feature deviation value; attention mechanism; feature fusion; residual structure;
D O I
10.3788/LOP231106
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
It is known that PointNet++ cannot deeply explore the semantic features of Lidar point clouds during feature extraction and loses features due to the use of maximum pooling during feature aggregation. These problems result in a decrease in point cloud segmentation accuracy. To address these problems, this study proposes a point cloud segmentation model based on feature deviation values and attention mechanisms by improving the feature extraction and feature aggregation modules of PointNet++. First, different local neighborhoods are obtained using spherical sampling, and then neighborhood points are selected using the K-nearest neighbor (KNN) method to calculate the feature deviation values of different neighborhoods. This enhances the model's recognition ability for different local neighborhoods and obtains deep semantic information of the point cloud. Second, an attention-based feature aggregation module is used to replace the maximum pooling module in PointNet++ to learn the weights of different features during feature aggregation. This improves the model's ability to filter information from different structures and enhances the segmentation performance of the model. To further optimize the model architecture, a residual module is added to the fully connected layer to avoid parameter redundancy and improve the model performance through weight sharing. Experimental results are validated on the Vaihingen and S3DIS datasets provided by ISPRS and Stanford University, respectively, and are compared with the results of the experiments and mainstream models provided by ISPRS. The overall accuracy (OA) and average F1 score reach 86. 69% and 73. 97%, respectively, which are 5. 49 percentage points and 8. 30 percentage points higher than those of PointNet++, respectively. The experimental results on the S3DIS dataset are compared with those of PointNet++, RandLA-Net, and ConvPoint, and show a clear improvement over those of PointNet++. The experiments show that the improved model can fully extract the semantic features of point clouds and effectively improve the segmentation accuracy of the model as compared with the segmentation results of PointNet++.
引用
收藏
页数:8
相关论文
共 23 条
[1]  
Boulch A, 2020, Arxiv, DOI arXiv:1904.02375
[2]  
Chen B., 2018, Preprints
[3]  
[黄鹤 Huang He], 2018, [测绘通报, Bulletin of Surveying and Mapping], P18
[4]  
Jiang MY, 2018, Arxiv, DOI [arXiv:1807.00652, 10.48550/arXiv.1807.00652, DOI 10.48550/ARXIV.1807.00652]
[5]   Escape from Cells: Deep Kd-Networks for the Recognition of 3D Point Cloud Models [J].
Klokov, Roman ;
Lempitsky, Victor .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :863-872
[6]   Semantic Segmentation of LiDAR Point Cloud Based on CAFF-PointNet [J].
Lai Ming ;
Zhao Jiankang ;
Liu Chuanqi ;
Cui Chao ;
Long Haihui .
LASER & OPTOELECTRONICS PROGRESS, 2021, 58 (20)
[7]   Learning of 3D Graph Convolution Networks for Point Cloud Analysis [J].
Lin, Zhi-Hao ;
Huang, Sheng Yu ;
Wang, Yu-Chiang Frank .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (08) :4212-4224
[8]  
Maturana D, 2015, IEEE INT C INT ROBOT, P922, DOI 10.1109/IROS.2015.7353481
[9]   PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation [J].
Qi, Charles R. ;
Su, Hao ;
Mo, Kaichun ;
Guibas, Leonidas J. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :77-85
[10]   RandLA-Net: Efficient Semantic Segmentation of Large-Scale Point Clouds [J].
Hu, Qingyong ;
Yang, Bo ;
Xie, Linhai ;
Rosa, Stefano ;
Guo, Yulan ;
Wang, Zhihua ;
Trigoni, Niki ;
Markham, Andrew .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11105-11114