DcTr: Noise-robust point cloud completion by dual-channel transformer with cross-attention

被引:23
作者
Fei, Ben [1 ]
Yang, Weidong [1 ,2 ]
Ma, Lipeng [1 ]
Chen, Wen-Ming [3 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai Key Lab Data Sci, Shanghai 200433, Peoples R China
[2] Zhuhai Fudan Innovat Inst, Hengqin New Area, Zhuhai 519000, Guangdong, Peoples R China
[3] Acad Engn & Technol, Shanghai 200433, Peoples R China
基金
中国国家自然科学基金;
关键词
Point cloud; 3D Vision; Transformer; Cross; -attention; Dual -channel transformer;
D O I
10.1016/j.patcog.2022.109051
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Current point cloud completion research mainly utilizes the global shape representation and local features to recover the missing regions of 3D shape for the partial point cloud. However, these methods suffer from inefficient utilization of local features and unstructured points prediction in local patches, hardly resulting in a well-arranged structure for points. To tackle these problems, we propose to employ Dual-channel Transformer and Cross-attention (CA) for point cloud completion (DcTr). The DcTr is apt at using local features and preserving a well-structured generation process. Specifically, the dual-channel transformer leverages point-wise attention and channel-wise attention to summarize the deconvolution patterns used in the previous Dual-channel Transformer Point Deconvolution (DCTPD) stage to produce the deconvolution in the current DCTPD stage. Meanwhile, we employ cross-attention to convey the geometric information from the local regions of incomplete point clouds for the generation of complete ones at different resolutions. In this way, we can generate the locally compact and structured point cloud by capturing the structure characteristic of 3D shape in local patches. Our experimental results indicate that DcTr outperforms the state-of-the-art point cloud completion methods under several benchmarks and is robust to various kinds of noise.
引用
收藏
页数:13
相关论文
共 55 条
[1]   SAANet: Spatial adaptive alignment network for object detection in automatic driving [J].
Chen, Junying ;
Bai, Tongyao .
IMAGE AND VISION COMPUTING, 2020, 94 (94)
[2]  
Cheng R, 2020, PR MACH LEARN RES, V155, P2148
[3]   SG-NN: Sparse Generative Neural Networks for Self-Supervised Scene Completion of RGB-D Scans [J].
Dai, Angela ;
Diller, Christian ;
Niessner, Matthias .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :846-855
[4]   A Field Model for Repairing 3D Shapes [J].
Duc Thanh Nguyen ;
Hua, Binh-Son ;
Minh-Khoi Tran ;
Quang-Hieu Pham ;
Yeung, Sai-Kit .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5676-5684
[5]   Superpixel/voxel medical image segmentation algorithm based on the regional interlinked value [J].
Fang, Lingling ;
Wang, Xin ;
Wang, Mengyi .
PATTERN ANALYSIS AND APPLICATIONS, 2021, 24 (04) :1685-1698
[6]   Point attention network for semantic segmentation of 3D point clouds [J].
Feng, Mingtao ;
Zhang, Liang ;
Lin, Xuefei ;
Gilani, Syed Zulqarnain ;
Mian, Ajmal .
PATTERN RECOGNITION, 2020, 107 (107)
[7]   Weakly-Supervised 3D Shape Completion in the Wild [J].
Gu, Jiayuan ;
Ma, Wei-Chiu ;
Manivasagam, Sivabalan ;
Zeng, Wenyuan ;
Wang, Zihao ;
Xiong, Yuwen ;
Su, Hao ;
Urtasun, Raquel .
COMPUTER VISION - ECCV 2020, PT V, 2020, 12350 :283-299
[8]   PCT: Point cloud transformer [J].
Guo, Meng-Hao ;
Cai, Jun-Xiong ;
Liu, Zheng-Ning ;
Mu, Tai-Jiang ;
Martin, Ralph R. ;
Hu, Shi-Min .
COMPUTATIONAL VISUAL MEDIA, 2021, 7 (02) :187-199
[9]  
Han ZZ, 2020, PR MACH LEARN RES, V119
[10]   SeqXY2SeqZ: Structure Learning for 3D Shapes by Sequentially Predicting 1D Occupancy Segments from 2D Coordinates [J].
Han, Zhizhong ;
Qiao, Guanhui ;
Liu, Yu-Shen ;
Zwicker, Matthias .
COMPUTER VISION - ECCV 2020, PT XXIV, 2020, 12369 :607-625