Consistent Depth Prediction for Transparent Object Reconstruction from RGB-D Camera

被引:2
作者
Cai, Yuxiang [1 ]
Zhu, Yifan [1 ]
Zhang, Haiwei [1 ]
Ren, Bo [1 ]
机构
[1] Nankai Univ, Tianjin, Peoples R China
来源
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV | 2023年
关键词
SLAM;
D O I
10.1109/ICCV51070.2023.00320
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transparent objects are commonly seen in indoor scenes but are hard to estimate. Currently, commercial depth cameras face difficulties in estimating the depth of transparent objects due to the light reflection and refraction on their surface. As a result, they tend to make a noisy and incorrect depth value for transparent objects. These incorrect depth data make the traditional RGB-D SLAM method fails in reconstructing the scenes that contain transparent objects. An exact depth value of the transparent object is required to restore in advance and it is essential that the depth value of the transparent object must keep consistent in different views, or the reconstruction result will be distorted. Previous depth prediction methods of transparent objects can restore these missing depth values but none of them can provide a good result in reconstruction due to the inconsistency prediction. In this work, we propose a real-time reconstruction method using a novel stereo-based depth prediction network to keep the consistency of depth prediction in a sequence of images. Because there is no video dataset about transparent objects currently to train our model, we construct a synthetic RGB-D video dataset with different transparent objects. Moreover, to test generalization capability, we capture video from real scenes using the RealSense D435i RGB-D camera. We compare the metrics on our dataset and SLAM reconstruction results in both synthetic scenes and real scenes with the previous methods. Experiments show our significant improvement in accuracy on depth prediction and scene reconstruction.
引用
收藏
页码:3436 / 3445
页数:10
相关论文
共 46 条
  • [1] Achlioptas P, 2018, PR MACH LEARN RES, V80
  • [2] Alt N, 2013, IEEE IMAGE PROC, P4131, DOI 10.1109/ICIP.2013.6738851
  • [3] [Anonymous], 2015, IEEE C COMPUTER VISI, DOI DOI 10.1109/CVPR.2015.7299064
  • [4] [Anonymous], 2021, ICCV, DOI DOI 10.1109/ICCV48922.2021.01596
  • [5] Unsupervised Scale-Consistent Depth Learning from Video
    Bian, Jia-Wang
    Zhan, Huangying
    Wang, Naiyan
    Li, Zhichao
    Zhang, Le
    Shen, Chunhua
    Cheng, Ming-Ming
    Reid, Ian
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (09) : 2548 - 2564
  • [6] Pointwise Convolutional Neural Networks
    Binh-Son Hua
    Minh-Khoi Tran
    Yeung, Sai-Kit
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 984 - 993
  • [7] Bozic A., 2021, ADV NEURAL INF PROCE, P1403
  • [8] Structural Design of Blue-to-Red Thermally-Activated Delayed Fluorescence Molecules by Adjusting the Strength between Donor and Acceptor
    Che, Weilong
    Xie, Yujun
    Li, Zhen
    [J]. ASIAN JOURNAL OF ORGANIC CHEMISTRY, 2020, 9 (09) : 1262 - 1276
  • [9] Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis
    Dai, Angela
    Qi, Charles Ruizhongtai
    Niessner, Matthias
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6545 - 6554
  • [10] Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture
    Eigen, David
    Fergus, Rob
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 2650 - 2658