Direction-induced convolution for point cloud analysis

被引:3
作者
Fang, Yuan [1 ]
Xu, Chunyan [1 ]
Zhou, Chuanwei [1 ]
Cui, Zhen [1 ]
Hu, Chunlong [2 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Jiangsu, Peoples R China
[2] Jiangsu Univ Sci & Technol, Sch Comp Sci & Engn, Zhenjiang 212003, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Point cloud; Convolution; Semantic segmentation; Classification; SEGMENTATION; NETWORKS;
D O I
10.1007/s00530-021-00770-0
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Point cloud analysis becomes a fundamental but challenging problem in the field of 3D scene understanding. To deal with unstructured and unordered point clouds in the embedded 3D space, we propose a novel direction-induced convolution (DIConv) to obtain the hierarchical representations of point clouds and then boost the performance of point cloud analysis. Specifically, we first construct a direction set as the basis of spatial direction information, where its entries can denote these latent direction components of 3D points. For each neighbor point, we can project its direction information into the constructed direction set for achieving an array of direction-dependent weights, then transform its features into the canonical ordered direction set space. After that, the standard image-like convolution can be leveraged to encode the unordered neighborhood regions of point cloud data. We further develop a residual DIConv (Res_DIConv) module and a farthest point sampling residual DIConv (FPS_Res_DIConv) module for jointly capturing the hierarchical features of input point clouds. By alternately stacking Res_DIConv modules and FPS_Res_DIConv modules, a direction-induced convolution network (DICNet) can be built to perform point cloud analysis in an end-to-end fashion. Comprehensive experiments on three benchmark datasets (including ModelNet40, ShapeNet Part, and S3DIS) demonstrate that the proposed DIConv method achieves encouraging performance on both point cloud classification and semantic segmentation tasks.
引用
收藏
页码:457 / 468
页数:12
相关论文
共 47 条
[11]  
Ioffe S., 2015, PMLR, P448, DOI DOI 10.48550/ARXIV.1502.03167
[12]   Escape from Cells: Deep Kd-Networks for the Recognition of 3D Point Cloud Models [J].
Klokov, Roman ;
Lempitsky, Victor .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :863-872
[13]   A-CNN: Annularly Convolutional Neural Networks on Point Clouds [J].
Komarichev, Artem ;
Zhong, Zichun ;
Hua, Jing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :7413-7422
[14]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[15]  
LAN S, 2019, PROC CVPR IEEE, P998
[16]   Large-scale Point Cloud Semantic Segmentation with Superpoint Graphs [J].
Landrieu, Loic ;
Simonovsky, Martin .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4558-4567
[17]   Deep Projective 3D Semantic Segmentation [J].
Lawin, Felix Jaremo ;
Danelljan, Martin ;
Tosteberg, Patrik ;
Bhat, Goutam ;
Khan, Fahad Shahbaz ;
Felsberg, Michael .
COMPUTER ANALYSIS OF IMAGES AND PATTERNS, 2017, 10424 :95-107
[18]   SO-Net: Self-Organizing Network for Point Cloud Analysis [J].
Li, Jiaxin ;
Chen, Ben M. ;
Lee, Gim Hee .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9397-9406
[19]  
Li YY, 2018, ADV NEUR IN, V31
[20]   MPAN: Multi-Part Attention Network for Point Cloud Based 3D Shape Retrieval [J].
Li, Zirui ;
Xu, Junyu ;
Zhao, Yue ;
Li, Wenhui ;
Nie, Weizhi .
IEEE ACCESS, 2020, 8 :157322-157332