EGCT: enhanced graph convolutional transformer for 3D point cloud representation learning

被引:0
作者
Chen, Gang [1 ]
Wang, Wenju [1 ]
Zhou, Haoran [1 ]
Wang, Xiaolin [1 ]
机构
[1] Univ Shanghai Sci & Technol, Coll Commun & Art Design, Shanghai, Peoples R China
基金
上海市自然科学基金;
关键词
Scene understanding; Graph convolution; Transformer; 3D point cloud; CLASSIFICATION; NETWORK;
D O I
10.1007/s00371-024-03600-2
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
It is an urgent problem of high-precision 3D environment perception to carry out representation learning on point cloud data, which complete the synchronous acquisition of local and global feature information. However, current representation learning methods either only focus on how to efficiently learn local features, or capture long-distance dependencies but lose the fine-grained features. Therefore, we explore transformer on topological structures of point cloud graphs, proposing an enhanced graph convolutional transformer (EGCT) method. EGCT construct graph topology for disordered and unstructured point cloud. Then it uses the enhanced point feature representation method to further aggregate the feature information of all neighborhood points, which can compactly represent the features of this local neighborhood graph. Subsequent process, the graph convolutional transformer simultaneously performs self-attention calculations and convolution operations on the point coordinates and features of the neighborhood graph. It efficiently utilizes the spatial geometric information of point cloud objects. Therefore, EGCT learns comprehensive geometric information of point cloud objects, which can help to improve segmentation and classification accuracy. On the ShapeNetPart and ModelNet40 datasets, our EGCT method achieves a mIoU of 86.8%, OA and AA of 93.5% and 91.2%, respectively. On the large-scale indoor scene point cloud dataset (S3DIS), the OA of EGCT method is 90.1%, and the mIoU is 67.8%. Experimental results demonstrate that our EGCT method can achieve comparable point cloud segmentation and classification performance to state-of-the-art methods while maintaining low model complexity. Our source code is available at https://github.com/shepherds001/EGCT.
引用
收藏
页码:3239 / 3261
页数:23
相关论文
共 50 条
[1]  
Achlioptas P, 2018, PR MACH LEARN RES, V80
[2]   3D Semantic Parsing of Large-Scale Indoor Spaces [J].
Armeni, Iro ;
Sener, Ozan ;
Zamir, Amir R. ;
Jiang, Helen ;
Brilakis, Ioannis ;
Fischer, Martin ;
Savarese, Silvio .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1534-1543
[3]   背包式激光雷达滤除低强度点云提取林木胸径 [J].
蔡硕 ;
邢艳秋 ;
端木嘉龙 .
森林工程, 2021, 37 (05) :12-19
[4]   Deep Unsupervised Learning of 3D Point Clouds via Graph Topology Inference and Filtering [J].
Chen, Siheng ;
Duan, Chaojing ;
Yang, Yaoqing ;
Li, Duanshun ;
Feng, Chen ;
Tian, Dong .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :3183-3198
[5]   Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis [J].
Dai, Angela ;
Qi, Charles Ruizhongtai ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6545-6554
[6]  
Dosovitskiy A, 2021, Arxiv, DOI [arXiv:2010.11929, 10.48550/arXiv.2010.11929, DOI 10.48550/ARXIV.2010.11929]
[7]   Coincident beach surveys using UAS, vehicle mounted and airborne laser scanner: Point cloud inter-comparison and effects of surface type heterogeneity on elevation accuracies [J].
Elsner, Paul ;
Dornbusch, Uwe ;
Thomas, Ian ;
Amos, Dan ;
Bovington, James ;
Horn, Diane .
REMOTE SENSING OF ENVIRONMENT, 2018, 208 :15-26
[8]   PCT: Point cloud transformer [J].
Guo, Meng-Hao ;
Cai, Jun-Xiong ;
Liu, Zheng-Ning ;
Mu, Tai-Jiang ;
Martin, Ralph R. ;
Hu, Shi-Min .
COMPUTATIONAL VISUAL MEDIA, 2021, 7 (02) :187-199
[9]   Deep Learning for 3D Point Clouds: A Survey [J].
Guo, Yulan ;
Wang, Hanyun ;
Hu, Qingyong ;
Liu, Hao ;
Liu, Li ;
Bennamoun, Mohammed .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (12) :4338-4364
[10]   Dual Transformer for Point Cloud Analysis [J].
Han, Xian-Feng ;
Jin, Yi-Fei ;
Cheng, Hui-Xian ;
Xiao, Guo-Qiang .
IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 :5638-5648