Pseudo-LiDAR Point Cloud Interpolation Based on 3D Motion Representation and Spatial Supervision

被引:25
作者
Liu, Haojie [1 ,2 ]
Liao, Kang [1 ,2 ]
Lin, Chunyu [1 ,2 ]
Zhao, Yao [1 ,2 ]
Guo, Yulan [3 ,4 ]
机构
[1] Beijing Jiaotong Univ, Inst Informat Sci, Beijing 100044, Peoples R China
[2] Beijing Key Lab Adv Informat Sci & Network Techno, Beijing 100044, Peoples R China
[3] Sun Yat Sen Univ, Sch Elect & Commun Engn, Guangzhou 510006, Peoples R China
[4] Natl Univ Def Technol, Coll Elect Sci & Technol, Changsha 410073, Peoples R China
基金
中国国家自然科学基金;
关键词
Pseudo-LiDAR interpolation; 3D point cloud; depth completion; scene flow; video interpolation; convolutional neural networks; DEPTH PREDICTION;
D O I
10.1109/TITS.2021.3056048
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Pseudo-LiDAR point cloud interpolation is a novel and challenging task in autonomous driving, which aims to address the frequency mismatching problem between a camera and a LiDAR. Previous works represent the 3D spatial motion relationship with a coarse 2D optical flow, and the quality of interpolated point clouds only depends on the supervision of depth maps. As a result, the generated point clouds suffer from inferior global distributions and local appearances. To solve the above problems, we propose a Pseudo-LiDAR point cloud interpolation network to generate temporally and spatially highquality point cloud sequences. By exploiting the scene flow from point clouds, the proposed network is able to learn a more accurate representation of the 3D spatial motion relationship. For a more comprehensive perception of the distribution of a point cloud, we design a novel reconstruction loss function with the chamfer distance to supervise the generation of Pseudo-LiDAR point clouds in 3D space. In addition, we introduce a multi-modal deep aggregation module to facilitate the efficient fusion of texture and depth features. As the benefits of the improved motion representation, training loss function, and model structure, our approach gains significant improvements on the Pseudo- LiDAR point cloud interpolation task. The experimental results evaluated on KITTI dataset demonstrate the state-of-the- art quantitative and qualitative performance of the proposed network.
引用
收藏
页码:6379 / 6389
页数:11
相关论文
共 43 条
[1]  
[Anonymous], 2014, ARXIV14126618
[2]  
[Anonymous], 2017, P BRIT MACH VIS C BM
[3]   Extended Patch Prioritization for Depth Filling Within Constrained Exemplar-Based RGB-D Image Completion [J].
Atapour-Abarghouei, Amir ;
Breckon, Toby P. .
IMAGE ANALYSIS AND RECOGNITION (ICIAR 2018), 2018, 10882 :306-314
[4]  
Atapour-Abarghouei A, 2016, INT C PATT RECOG, P2813, DOI 10.1109/ICPR.2016.7900062
[5]   Depth-Aware Video Frame Interpolation [J].
Bao, Wenbo ;
Lai, Wei-Sheng ;
Ma, Chao ;
Zhang, Xiaoyun ;
Gao, Zhiyong ;
Yang, Ming-Hsuan .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3698-3707
[6]   Efficient Spatio-Temporal Hole Filling Strategy for Kinect Depth Maps [J].
Camplani, Massimo ;
Salgado, Luis .
THREE-DIMENSIONAL IMAGE PROCESSING (3DIP) AND APPLICATIONS II, 2012, 8290
[7]  
Chen Y, 2019, IEEE I CONF COMP VIS, P10022, DOI [10.1109/iccv.2019.01012, 10.1109/ICCV.2019.01012]
[8]   Adaptive Nonrigid Inpainting of Three-Dimensional Point Cloud Geometry [J].
Dinesh, Chinthaka ;
Bajic, Ivan, V ;
Cheung, Gene .
IEEE SIGNAL PROCESSING LETTERS, 2018, 25 (06) :878-882
[9]   Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture [J].
Eigen, David ;
Fergus, Rob .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2650-2658
[10]  
Eldesokey A., 2018, P BRIT MACH VIS C BM, P14