MIMONet: Structured light 3D shape reconstruction by a multi-input multi-output network

被引:12
作者
Hieu Nguyen [1 ,2 ]
Ly, Khanh L. [3 ]
Thanh Nguyen [4 ]
Wang, Yuzheng [5 ]
Wang, Zhaoyang [1 ]
机构
[1] Catholic Univ Amer, Dept Mech Engn, Washington, DC 20064 USA
[2] NIDA, Neuroimaging Res Branch, NIH, Baltimore, MD 21224 USA
[3] Catholic Univ Amer, Dept Biomed Engn, Washington, DC 20064 USA
[4] Catholic Univ Amer, Dept Elect Engn & Comp Sci, Washington, DC 20064 USA
[5] Univ Jinan, Sch Mech Engn, Jinan 250022, Shandong, Peoples R China
关键词
REAL-TIME; FRINGE-PROJECTION; ACCURACY;
D O I
10.1364/AO.426189
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Reconstructing 3D geometric representation of objects with deep learning frameworks has recently gained a great deal of interest in numerous fields. The existing deep-learning-based 3D shape reconstruction techniques generally use a single red-green-blue (RGB) image, and the depth reconstruction accuracy is often highly limited due to a variety of reasons. We present a 3D shape reconstruction technique with an accuracy enhancement strategy by integrating the structured-light scheme with deep convolutional neural networks (CNNs). The key idea is to transform multiple (typically two) grayscale images consisting of fringe and/or speckle patterns into a 3D depth map using an end-to-end artificial neural network. Distinct from the existing autoencoder-based networks, the proposed technique reconstructs the 3D shape of target using a refinement approach that fuses multiple feature maps to obtain multiple outputs with an accuracy-enhanced final output. A few experiments have been conducted to verify the robustness and capabilities of the proposed technique. The findings suggest that the proposed network approach can be a promising 3D reconstruction technique for future academic research and industrial applications. (C) 2021 Optical Society of America
引用
收藏
页码:5134 / 5144
页数:11
相关论文
共 58 条
[1]  
[Anonymous], 2015, P INT C LEARN REPR
[2]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[3]   Multi-Garment Net: Learning to Dress 3D People from Images [J].
Bhatnagar, Bharat Lal ;
Tiwari, Garvita ;
Theobalt, Christian ;
Pons-Moll, Gerard .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :5419-5429
[4]   A Comparative Analysis between Active and Passive Techniques for Underwater 3D Reconstruction of Close-Range Objects [J].
Bianco, Gianfranco ;
Gallo, Alessandro ;
Bruno, Fabio ;
Muzzupappa, Maurizio .
SENSORS, 2013, 13 (08) :11007-11031
[5]   Review of 20 years of range sensor development [J].
Blais, F .
JOURNAL OF ELECTRONIC IMAGING, 2004, 13 (01) :231-243
[6]  
Bud Andrew, 2018, Biometric Technology Today, V2018, P5, DOI 10.1016/S0969-4765(18)30010-9
[7]   Overview of three-dimensional shape measurement using optical methods [J].
Chen, F ;
Brown, GM ;
Song, MM .
OPTICAL ENGINEERING, 2000, 39 (01) :10-22
[8]   3D-R2N2: A Unified Approach for Single and Multi-view 3D Object Reconstruction [J].
Choy, Christopher B. ;
Xu, Danfei ;
Gwak, Jun Young ;
Chen, Kevin ;
Savarese, Silvio .
COMPUTER VISION - ECCV 2016, PT VIII, 2016, 9912 :628-644
[9]  
Eigen D, 2014, ADV NEUR IN, V27
[10]  
El hazzat Soulaiman, 2014, 2014 Fifth International Conference on Next-Generation Networks and Services (NGNS), P194, DOI 10.1109/NGNS.2014.6990252