Point Cloud Classification Segmentation Combining Inter-Region Structure Relations and Self-Attention Edge Convolution Network

被引:0
作者
Lyu, Zhiwei [1 ]
Yang, Jiazhi [1 ,2 ]
Zhou, Guoqing [1 ,2 ]
Shen, Lu [1 ]
机构
[1] School of Information Science and Engineering, Guilin University of Technology, Guangxi, Guilin
[2] Guangxi Key Laboratory of Embedded Technology and Intelligent Systems, Guangxi, Guilin
关键词
deep learning; edge convolution; regional context; regional relationship; self-attention;
D O I
10.3778/j.issn.1002-8331.2304-0141
中图分类号
学科分类号
摘要
A new network framework, ISEC-Net (inter-region structure relations and self-attention edge convolution network), is proposed to address the problem of insufficient capture of context and relational features within a region in point cloud networks for deep learning. The network consists of two modules:IrConv (inter-region convolution) and SaConv (self-attention convolution). The SaConv module can extract finer edge features, while the IrConv can dynamically integrate local structural information into point features and adaptively capture inter-regional relationships. Extensive experiments are conducted on the ModelNet40 and ShapeNet datasets for point cloud classification and part segmentation. The results show that on the ModelNet40 dataset, the overall accuracy (OA) of the ISEC-Net model reaches 93.5%, and the average accuracy (mAcc) reaches 90.7%. On the ShapeNet dataset, the average intersection-over-union (mIoU) reaches 86.1%, and the part segmentation accuracy of guitar, headphone, cup and other parts in the single-class intersection-over-union (IoU) experiment is excellent. This demonstrates that compared with traditional dynamic graph convolutional networks, ISEC-Net can accurately capture the local features and fine structure of point clouds and enhance the aggregation of global features, thus having excellent effectiveness and generalization ability. © 2024 Journal of Computer Engineering and Applications Beijing Co., Ltd.; Science Press. All rights reserved.
引用
收藏
页码:171 / 179
页数:8
相关论文
共 26 条
[1]  
LI Y, IBANEZ-GUZMAN J., Lidar for autonomous driving: the principles, challenges, and trends for automotive lidar and perception systems, IEEE Signal Processing Magazine, 37, 4, pp. 50-61, (2020)
[2]  
AHMED S M, TAN Y Z, CHEW C M, Et al., Edge and corner detection for unorganized 3D point clouds with application to robotic welding, Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 7350-7355, (2018)
[3]  
RAHMAN M M, TAN Y, XUE J, Et al., Notice of violation of IEEE publication principles: recent advances in 3D object detection in the era of deep neural networks: a survey, IEEE Transactions on Image Processing, 29, pp. 2947-2962, (2019)
[4]  
MATURANA D, SCHERER S., VoxNet: a 3D convolutional neural network for real-time object recognition, Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 922-928, (2015)
[5]  
SU H, MAJI S, KALOGERAKIS E, Et al., Multi-view convolutional neural networks for 3D shape recognition, Proceedings of the IEEE International Conference on Computer Vision, pp. 945-953, (2015)
[6]  
QI C R, SU H, MO K, Et al., PointNet: deep learning on point sets for 3D classification and segmentation, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 652-660, (2017)
[7]  
QI C R, YI L, SU H, Et al., PointNet++: deep hierarchical feature learning on point sets in a metric space, Advances in Neural Information Processing Systems, (2017)
[8]  
WANG Y, SUN Y, LIU Z, Et al., Dynamic graph CNN for learning on point clouds, ACM Transactions on Graphics, 38, 5, pp. 1-12, (2019)
[9]  
ZHAO H, JIANG L, FU C W, Et al., Pointweb: enhancing local neighborhood features for point cloud processing, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5565-5573, (2019)
[10]  
LANDRIEU L, SIMONOVSKY M., Large-scale point cloud semantic segmentation with superpoint graphs, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4558-4567, (2018)