Faster Dynamic Graph CNN: Faster Deep Learning on 3D Point Cloud Data

被引:8
作者
Hong, Jinseok [1 ,3 ]
Kim, Keeyoung [1 ,2 ]
Lee, Hongchul [3 ]
机构
[1] Artificial Intelligence Res Inst, Seongnam 13120, South Korea
[2] Ingenio AI, Seoul 02841, South Korea
[3] Korea Univ, Sch Ind Management Engn, Seoul 02841, South Korea
来源
IEEE ACCESS | 2020年 / 8卷
关键词
Three-dimensional displays; Solid modeling; Computational modeling; Machine learning; Two dimensional displays; Neural networks; Feature extraction; Classification; deep learning; graph CNN; point cloud; segmentation;
D O I
10.1109/ACCESS.2020.3023423
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Geometric data are commonly expressed using point clouds, with most 3D data collection devices outputting data in this form. Research on processing point cloud data for deep learning is ongoing. However, it has been difficult to apply such data as input to a convolutional neural network (CNN) or recurrent neural network (RNN) because of their unstructured and unordered features. In this study, this problem was resolved by arranging point cloud data in a canonical space through a graph CNN. The proposed graph CNN works dynamically at each layer of the network and learns the global geometric features by capturing the neighbor information of the points. In addition, by using a squeeze-and-excitation module that recalibrates the information for each layer, we achieved a good trade-off between the performance and the computation cost, and a residual-type skip connection network was designed to train the deep models efficiently. Using the proposed model, we achieved a state-of-the-art performance in terms of classification and segmentation on benchmark datasets, namely ModelNet40 and ShapeNet, while being able to train our model 2 to 2.5 times faster than other similar models.
引用
收藏
页码:190529 / 190538
页数:10
相关论文
共 66 条
[1]  
[Anonymous], 2016, ShapeNet: An information-rich 3D model reposi
[2]  
[Anonymous], ICML
[3]   Inside-Outside Net: Detecting Objects in Context with Skip Pooling and Recurrent Neural Networks [J].
Bell, Sean ;
Zitnick, C. Lawrence ;
Bala, Kavita ;
Girshick, Ross .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2874-2883
[4]  
Bluche T, 2016, ADV NEUR IN, V29
[5]  
Boscaini D, 2016, ADV NEUR IN, V29
[6]   Geometric Deep Learning Going beyond Euclidean data [J].
Bronstein, Michael M. ;
Bruna, Joan ;
LeCun, Yann ;
Szlam, Arthur ;
Vandergheynst, Pierre .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (04) :18-42
[7]  
Bruna J, 2013, 2 INT C LEARN REPR I
[8]   Look and Think Twice: Capturing Top-Down Visual Attention with Feedback Convolutional Neural Networks [J].
Cao, Chunshui ;
Liu, Xianming ;
Yang, Yi ;
Yu, Yinan ;
Wang, Jiang ;
Wang, Zilei ;
Huang, Yongzhen ;
Wang, Liang ;
Huang, Chang ;
Xu, Wei ;
Ramanan, Deva ;
Huang, Thomas S. .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2956-2964
[9]   SCA-CNN: Spatial and Channel-wise Attention in Convolutional Networks for Image Captioning [J].
Chen, Long ;
Zhang, Hanwang ;
Xiao, Jun ;
Nie, Liqiang ;
Shao, Jian ;
Liu, Wei ;
Chua, Tat-Seng .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6298-6306
[10]  
Chen Y., 2017, Advances in Neural Information Processing Systems, P4470