Realtime 3D 360-Degree Telepresence With Deep-Learning-Based Head-Motion Prediction

被引:10
作者
Aykut, Tamay [1 ]
Xu, Jingyi [1 ]
Steinbach, Eckehard [1 ]
机构
[1] Tech Univ Munich, Dept Elect & Comp Engn, D-80333 Munich, Germany
关键词
3D vision; deep learning; telepresence; remote reality; 3D 360-degree vision; omnistereoscopic vision; COMPENSATION; DELAY;
D O I
10.1109/JETCAS.2019.2897220
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The acceptance and the dissemination of 360 degrees telepresence systems are severely restricted by the appearance of motion sickness. Such systems consist of a client-side, where the user wears a head-mounted display, a server-side, which provides a 3D 360 degrees visual representation of the remote scene, and a communication network in between. Due to the physically unavoidable latency, there is often a noticeable lag between head motion and visual response. If the sensory information from the visual system is not consistent with the perceived ego-motion of the user, and the emergence of visual discomfort is inevitable. In this paper, we present a delay-compensating 3D 360 degrees vision system, which provides omnistereoscopic vision and a significant reduction of the perceived latency. We formally describe the underlying problem and provide an algebraic description of the amount of achievable delay compensation both for perspective and fisheye camera systems, considering all the three degrees of freedom. Furthermore, we propose a generic approach that is agnostic to the underlying camera system. In addition, a novel deep-learning-based head motion prediction algorithm is presented to further improve the compensation rate. Using the naive approach, where no compensation and prediction is applied, we obtain a mean compensation rate of 72.8% for investigated latencies from 0.1 to 1.0 s. Our proposed generic delay-compensation approach, combined with our novel deep-learning-based head-motion prediction approach, manages to achieve a mean compensation rate of 97.3%. The proposed technique also substantially outperforms prior head-motion prediction techniques, both for traditional and deep learning-based methods.
引用
收藏
页码:231 / 244
页数:14
相关论文
共 31 条
[1]   Panoramic Stereo Videos with a Single Camera [J].
Aggarwal, Rajat ;
Vohra, Amrisha ;
Namboodiri, Anoop M. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3755-3763
[2]   Tolerance of temporal delay in virtual environments [J].
Allison, RS ;
Harris, LR ;
Jenkin, M ;
Jasiobedzka, U ;
Zacher, JE .
IEEE VIRTUAL REALITY 2001, PROCEEDINGS, 2001, :247-254
[3]   Jump: Virtual Reality Video [J].
Anderson, Robert ;
Gallup, David ;
Barron, Jonathan T. ;
Kontkanen, Janne ;
Snavely, Noah ;
Hernandez, Carlos ;
Agarwal, Sameer ;
Seitz, Steven M. .
ACM TRANSACTIONS ON GRAPHICS, 2016, 35 (06)
[4]  
[Anonymous], 2000, MATH HDB SCI ENG DEF
[5]  
Aykut T., 2018, P IEEE WINT C APPL C, P2010
[6]   Delay Compensation for a Telepresence System With 3D 360 Degree Vision Based on Deep Head Motion Prediction and Dynamic FoV Adaptation [J].
Aykut, Tamay ;
Karimi, Mojtaba ;
Burgmair, Christoph ;
Finkenzeller, Andreas ;
Bachhuber, Christoph ;
Steinbach, Eckehard .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :4343-4350
[7]   A Stereoscopic Vision System with Delay Compensation for 360 Remote Reality [J].
Aykut, Tamay ;
Lochbrunner, Stefan ;
Karimi, Mojtaba ;
Cizmeci, Burak ;
Steinbach, Eckehard .
PROCEEDINGS OF THE THEMATIC WORKSHOPS OF ACM MULTIMEDIA 2017 (THEMATIC WORKSHOPS'17), 2017, :201-209
[8]  
Azuma R., 1995, Computer Graphics Proceedings. SIGGRAPH 95, P401, DOI 10.1145/218380.218496
[9]  
Carbone M, 2010, ACM SIGCOMM COMP COM, V40, P13, DOI 10.1145/1764873.1764876
[10]   Human performance issues and user interface design for teleoperated robots [J].
Chen, Jessie Y. C. ;
Haas, Ellen C. ;
Barnes, Michael J. .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2007, 37 (06) :1231-1245