Analysis of AI-Based Single-View 3D Reconstruction Methods for an Industrial Application

被引:7
作者
Hartung, Julia [1 ,2 ]
Dold, Patricia M. [1 ,2 ,3 ]
Jahn, Andreas [1 ]
Heizmann, Michael [2 ]
机构
[1] TRUMPF Laser GmbH, Aichhalder Str 39, D-78713 Schramberg, Germany
[2] Karlsruhe Inst Technol, Inst Ind Informat Technol, Hertzstr 16, D-76187 Karlsruhe, Germany
[3] Karlsruhe Inst Technol, Inst Automat & Appl Informat, Hermann von Helmholtz Pl 1, D-76344 Eggenstein Leopoldshafen, Germany
关键词
three-dimensional reconstruction; single view; stacked autoencoder (SAE); generative adversarial network (GAN); U-Net; stacked dilated U-Net (SDU-Net); artificial intelligence; deep learning; hairpin; production;
D O I
10.3390/s22176425
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Machine learning (ML) is a key technology in smart manufacturing as it provides insights into complex processes without requiring deep domain expertise. This work deals with deep learning algorithms to determine a 3D reconstruction from a single 2D grayscale image. The potential of 3D reconstruction can be used for quality control because the height values contain relevant information that is not visible in 2D data. Instead of 3D scans, estimated depth maps based on a 2D input image can be used with the advantage of a simple setup and a short recording time. Determining a 3D reconstruction from a single input image is a difficult task for which many algorithms and methods have been proposed in the past decades. In this work, three deep learning methods, namely stacked autoencoder (SAE), generative adversarial networks (GANs) and U-Nets are investigated, evaluated and compared for 3D reconstruction from a 2D grayscale image of laser-welded components. In this work, different variants of GANs are tested, with the conclusion that Wasserstein GANs (WGANs) are the most robust approach among them. To the best of our knowledge, the present paper considers for the first time the U-Net, which achieves outstanding results in semantic segmentation, in the context of 3D reconstruction tasks. Unlike the U-Net, which uses standard convolutions, the stacked dilated U-Net (SDU-Net) applies stacked dilated convolutions. Of all the 3D reconstruction approaches considered in this work, the SDU-Net shows the best performance, not only in terms of evaluation metrics but also in terms of computation time. Due to the comparably small number of trainable parameters and the suitability of the architecture for strong data augmentation, a robust model can be generated with only a few training data.
引用
收藏
页数:17
相关论文
共 50 条
[1]  
Alom M.Z., 2018, arXiv
[2]  
[Anonymous], 2015, CVPR
[3]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[4]   Face Depth Estimation With Conditional Generative Adversarial Networks [J].
Arslan, Abdullah Taha ;
Seke, Erol .
IEEE ACCESS, 2019, 7 :23222-23231
[5]   An Intelligent System for the Classification of Lung Cancer Based on Deep Learning Strategy [J].
Arslan, Ahmet Kadir ;
Yasar, Seyma ;
Colak, Cemil .
2019 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND DATA PROCESSING (IDAP 2019), 2019,
[6]   Recovery of surface orientation from diffuse polarization [J].
Atkinson, Gary A. ;
Hancock, Edwin R. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2006, 15 (06) :1653-1664
[7]  
Baby Alba Terese, 2020, 2020 Advanced Computing and Communication Technologies for High Performance Applications (ACCTHPA), P125, DOI 10.1109/ACCTHPA49271.2020.9213233
[8]  
Bundesministerium fur Wirtschaft und Klimaschutz (BMWK), 2022, IND 4 0 DIG WIRTSCH
[9]   3D-R2N2: A Unified Approach for Single and Multi-view 3D Object Reconstruction [J].
Choy, Christopher B. ;
Xu, Danfei ;
Gwak, Jun Young ;
Chen, Kevin ;
Savarese, Silvio .
COMPUTER VISION - ECCV 2016, PT VIII, 2016, 9912 :628-644
[10]   DRUNET: a dilated-residual U-Net deep learning network to segment optic nerve head tissues in optical coherence tomography images [J].
Devalla, Sripad Krishna ;
Renukanand, Prajwal K. ;
Sreedhar, Bharathwaj K. ;
Subramanian, Giridhar ;
Zhang, Liang ;
Perera, Shamira ;
Mari, Jean-Martial ;
Chin, Khai Sing ;
Tun, Tin A. ;
Strouthidis, Nicholas G. ;
Aung, Tin ;
Thiery, Alexandre H. ;
Girard, Michael J. A. .
BIOMEDICAL OPTICS EXPRESS, 2018, 9 (07) :3244-3265