PSNet: Fast Data Structuring for Hierarchical Deep Learning on Point Cloud

被引:14
作者
Li, Luyang [1 ,2 ]
He, Ligang [3 ]
Gao, Jinjin [4 ]
Han, Xie [1 ]
机构
[1] North Univ China, Sch Data Sci & Technol, Taiyuan 030051, Peoples R China
[2] Shanxi Informat Ind Technol Res Inst Co Ltd, Taiyuan 030012, Peoples R China
[3] Univ Warwick, Dept Comp, Coventry CV4 7AL, W Midlands, England
[4] Shanxi Univ Finance & Econ, Expt Ctr, Taiyuan 030006, Peoples R China
基金
中国国家自然科学基金;
关键词
Point cloud compression; Data models; Deep learning; Training; Task analysis; Convolution; Computational modeling; point cloud; data structuring; computer vision; grouping; sampling; NETWORK;
D O I
10.1109/TCSVT.2022.3171968
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In order to retain more feature information of local areas on a point cloud, local grouping and subsampling are the necessary data structuring steps in most hierarchical deep learning models. Due to the disorder nature of the points in a point cloud, the significant time cost may be consumed when grouping and subsampling the points, which consequently results in poor scalability. This paper proposes a fast data structuring method called PSNet (Point Structuring Net). PSNet transforms the spatial features of the points and matches them to the features of local areas in a point cloud. PSNet achieves grouping and sampling at the same time while the existing methods process sampling and grouping in two separate steps (such as using FPS plus kNN). PSNet performs feature transformation pointwise while the existing methods uses the spatial relationship among the points as the reference for grouping. Thanks to these features, PSNet has two important advantages: 1) the grouping and sampling results obtained by PSNet is stable and permutation invariant; and 2) PSNet can be easily parallelized. PSNet can replace the data structuring methods in the mainstream point cloud deep learning models in a plug-and-play manner. We have conducted extensive experiments. The results show that PSNet can improve the training and inference speed significantly while maintaining the model accuracy.
引用
收藏
页码:6835 / 6849
页数:15
相关论文
共 56 条
[41]   Quadratic Terms Based Point-to-Surface 3D Representation for Deep Learning of Point Cloud [J].
Sun, Tiecheng ;
Liu, Guanghui ;
Li, Ru ;
Liu, Shuaicheng ;
Zhu, Shuyuan ;
Zeng, Bing .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (05) :2705-2718
[42]  
Szegedy C, 2015, PROC CVPR IEEE, P1, DOI 10.1109/CVPR.2015.7298594
[43]  
Vaswani A, 2017, ADV NEUR IN, V30
[44]  
Wang DZ, 2015, ROBOT SCI SYST
[45]   Graph Attention Convolution for Point Cloud Semantic Segmentation [J].
Wang, Lei ;
Huang, Yuchun ;
Hou, Yaolin ;
Zhang, Shenman ;
Shan, Jie .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :10288-10297
[46]   O-CNN: Octree-based Convolutional Neural Networks for 3D Shape Analysis [J].
Wang, Peng-Shuai ;
Liu, Yang ;
Guo, Yu-Xiao ;
Sun, Chun-Yu ;
Tong, Xin .
ACM TRANSACTIONS ON GRAPHICS, 2017, 36 (04)
[47]   PointConv: Deep Convolutional Networks on 3D Point Clouds [J].
Wu, Wenxuan ;
Qi, Zhongang ;
Li Fuxin .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9613-9622
[48]  
Wu ZR, 2015, PROC CVPR IEEE, P1912, DOI 10.1109/CVPR.2015.7298801
[49]  
Chang AX, 2015, Arxiv, DOI [arXiv:1512.03012, DOI 10.48550/ARXIV.1512.03012]
[50]  
Xu Q., 2020, PROC IEEECVF C COMPU, P5661