PSNet: Fast Data Structuring for Hierarchical Deep Learning on Point Cloud

被引:14
作者
Li, Luyang [1 ,2 ]
He, Ligang [3 ]
Gao, Jinjin [4 ]
Han, Xie [1 ]
机构
[1] North Univ China, Sch Data Sci & Technol, Taiyuan 030051, Peoples R China
[2] Shanxi Informat Ind Technol Res Inst Co Ltd, Taiyuan 030012, Peoples R China
[3] Univ Warwick, Dept Comp, Coventry CV4 7AL, W Midlands, England
[4] Shanxi Univ Finance & Econ, Expt Ctr, Taiyuan 030006, Peoples R China
基金
中国国家自然科学基金;
关键词
Point cloud compression; Data models; Deep learning; Training; Task analysis; Convolution; Computational modeling; point cloud; data structuring; computer vision; grouping; sampling; NETWORK;
D O I
10.1109/TCSVT.2022.3171968
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In order to retain more feature information of local areas on a point cloud, local grouping and subsampling are the necessary data structuring steps in most hierarchical deep learning models. Due to the disorder nature of the points in a point cloud, the significant time cost may be consumed when grouping and subsampling the points, which consequently results in poor scalability. This paper proposes a fast data structuring method called PSNet (Point Structuring Net). PSNet transforms the spatial features of the points and matches them to the features of local areas in a point cloud. PSNet achieves grouping and sampling at the same time while the existing methods process sampling and grouping in two separate steps (such as using FPS plus kNN). PSNet performs feature transformation pointwise while the existing methods uses the spatial relationship among the points as the reference for grouping. Thanks to these features, PSNet has two important advantages: 1) the grouping and sampling results obtained by PSNet is stable and permutation invariant; and 2) PSNet can be easily parallelized. PSNet can replace the data structuring methods in the mainstream point cloud deep learning models in a plug-and-play manner. We have conducted extensive experiments. The results show that PSNet can improve the training and inference speed significantly while maintaining the model accuracy.
引用
收藏
页码:6835 / 6849
页数:15
相关论文
共 56 条
[51]   SpiderCNN: Deep Learning on Point Sets with Parameterized Convolutional Filters [J].
Xu, Yifan ;
Fan, Tianqi ;
Xu, Mingye ;
Zeng, Long ;
Qiao, Yu .
COMPUTER VISION - ECCV 2018, PT VIII, 2018, 11212 :90-105
[52]   Modeling Point Clouds with Self-Attention and Gumbel Subset Sampling [J].
Yang, Jiancheng ;
Zhang, Qiang ;
Ni, Bingbing ;
Li, Linguo ;
Liu, Jinxian ;
Zhou, Mengdie ;
Tian, Qi .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3318-3327
[53]  
Zhang YX, 2018, 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), P6279, DOI 10.1109/ICASSP.2018.8462291
[54]   PointWeb: Enhancing Local Neighborhood Features for Point Cloud Processing [J].
Zhao, Hengshuang ;
Jiang, Li ;
Fu, Chi-Wing ;
Jia, Jiaya .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :5550-5558
[55]   Transformer3D-Det: Improving 3D Object Detection by Vote Refinement [J].
Zhao, Lichen ;
Guo, Jinyang ;
Xu, Dong ;
Sheng, Lu .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (12) :4735-4746
[56]   VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection [J].
Zhou, Yin ;
Tuzel, Oncel .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4490-4499