PCUNet: A Context-Aware Deep Network for Coarse-to-Fine Point Cloud Completion

被引:7
作者
Zhao, Meihua [1 ,2 ]
Xiong, Gang [3 ,4 ]
Zhou, Mengchu [5 ]
Shen, Zhen [1 ,6 ]
Liu, Sheng [1 ]
Han, Yunjun [1 ]
Wang, Fei-Yue [1 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[3] Chinese Acad Sci, Beijing Engn Res Ctr Intelligent Syst & Technol, Inst Automat, Beijing 100190, Peoples R China
[4] Chinese Acad Sci, Cloud Comp Ctr, Guangdong Engn Res Ctr 3D Printing & Intelligent, Dongguan 523808, Peoples R China
[5] New Jersey Inst Technol, Helen & John C Hartmann Dept Elect & Comp Engn, Newark, NJ 07102 USA
[6] Qingdao Acad Intelligent Ind, Intelligent Mfg Ctr, Qingdao 266109, Peoples R China
基金
中国国家自然科学基金;
关键词
Point cloud compression; Three-dimensional displays; Decoding; Shape; Feature extraction; Sensors; Neural networks; Point cloud completion; deep learning; context-aware; attention-enhanced skip connections;
D O I
10.1109/JSEN.2022.3181675
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Point cloud completion aims at predicting a complete 3D shape from an incomplete input. It has important applications in the fields of intelligent manufacturing, augmented reality, virtual reality, self-driving cars, and intelligent robotics. Although deep learning-based point cloud completion technology has developed rapidly in recent years, there are still unsolved problems. Previous approaches predict each point independently and ignore contextual information. And, they usually predict a complete 3D shape based on a global feature vector extracted from an incomplete input, which leads to missing of some fine-grained details. In this paper, motivated by the transposed convolution and the "UNet" structure in neural networks for image processing, we propose a context-aware deep network termed as PCUNet for coarse-to-fine point cloud completion. It adopts an encoder-decoder structure, in which the encoder follows the design of the relation-shape convolutional neural network (RS-CNN), and the decoder consists of fully-connected layers and two stacked decoder modules for predicting complete point clouds. The contributions are twofold. First, we design the decoder module as a coordinate-guided context-aware upsampling module, in which contextual information can be taken into full account by neighbor aggregation. Second, to preserve fine-grained details in the input, we propose attention-enhanced skip connections for effective information propagation from the encoder to the decoder. Experiments are conducted on the widely used PCN and KITTI datasets. The results show that our proposed approach achieves competitive performance compared to the existing state-of-the-art approaches in terms of the Chamfer distance and the computational complexity metrics.
引用
收藏
页码:15098 / 15110
页数:13
相关论文
共 53 条
[1]  
Achlioptas P, 2018, PR MACH LEARN RES, V80
[2]   Denoise and Contrast for Category Agnostic Shape Completion [J].
Alliegro, Antonio ;
Valsesia, Diego ;
Fracastoro, Giulia ;
Magli, Enrico ;
Tommasi, Tatiana .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :4627-4636
[3]   Hierarchical Surface Prediction for 3D Object Reconstruction [J].
Bane, Christian ;
Tulsiani, Shubham ;
Malik, Jitendra .
PROCEEDINGS 2017 INTERNATIONAL CONFERENCE ON 3D VISION (3DV), 2017, :412-420
[4]   Semi-supervised FusedGAN for Conditional Image Generation [J].
Bodla, Navaneeth ;
Hua, Gang ;
Chellappa, Rama .
COMPUTER VISION - ECCV 2018, PT V, 2018, 11209 :689-704
[5]  
Chen Y., 2020, IEEE Trans. Neural Netw. Learn. Syst., V33, P1
[6]   Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis [J].
Dai, Angela ;
Qi, Charles Ruizhongtai ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6545-6554
[7]  
Dumoulin V, 2018, Arxiv, DOI [arXiv:1603.07285, 10.48550/arXiv.1603.07285]
[8]   A Point Set Generation Network for 3D Object Reconstruction from a Single Image [J].
Fan, Haoqiang ;
Su, Hao ;
Guibas, Leonidas .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2463-2471
[9]   Vision meets robotics: The KITTI dataset [J].
Geiger, A. ;
Lenz, P. ;
Stiller, C. ;
Urtasun, R. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1231-1237
[10]   A Papier-Mache Approach to Learning 3D Surface Generation [J].
Groueix, Thibault ;
Fisher, Matthew ;
Kim, Vladimir G. ;
Russell, Bryan C. ;
Aubry, Mathieu .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :216-224