Fringe projection-based single-shot 3D eye tracking using deep learning and computer graphics

被引:1
作者
Zheng, Yi [1 ]
Chao, Qing [1 ]
An, Yatong [1 ]
Hirsh, Seth [1 ]
Fix, Alexander [1 ]
机构
[1] Meta Real Lab Res, Redmond, WA 98052 USA
来源
OPTICAL ARCHITECTURES FOR DISPLAYS AND SENSING IN AUGMENTED, VIRTUAL, AND MIXED REALITY, AR, VR, MR IV | 2023年 / 12449卷
关键词
eye tracking; fringe projection profilometry; 3D sensing; physically based rendering; deep learning;
D O I
10.1117/12.2667763
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
High-accuracy and high-speed 3D sensing technology plays an essential role in VR eye tracking as it can build a bridge to connect the user with virtual worlds. In VR eye tracking, fringe projection profilometry (FPP) can avoid dependence on scene textures and provide accurate results in near-eye scenarios; however, phase-shifting based FPP faces challenges like motion artifacts and may not meet the low-latency requirements of eye tracking tasks. On the other hand, Fourier transform profilometry can achieve single-shot 3D sensing, but the system is highly impacted by the texture variations on the eye. As a solution to the challenges above, researchers have explored deep learning-based single-shot fringe projection 3D sensing techniques. However, building a training dataset is expensive, and without abundant data the model is difficult to make generalized. In this paper, we built a virtual fringe projection system along with photorealistic face and eye models to synthesize large amounts of training data. Therefore, we can reduce the cost and enhance the generalization ability of the convolutional neural network (CNN). The training data synthesizer utilizes physically based rendering (PBR) and achieves high photorealism. We demonstrate that PBR can simulate the complex double refraction of structured light due to corneas. To train the CNN, we adopted the idea of transfer learning, where the CNN is first trained by PBR-generated data, then trained with the real data. We tested the CNN on real data, and the predicted results demonstrate that the synthesized data enhances the performance of the model and achieves around 3.722 degree gaze accuracy and 0.5363 mm pupil position error on an unfamiliar participant.
引用
收藏
页数:11
相关论文
共 26 条
[1]  
[Anonymous], 2015, Med. Image Comput. Comput.-Assist. Int.
[2]   Optics of the human cornea influence the accuracy of stereo eye-tracking methods: a simulation study [J].
Barsingerhorn, A. D. ;
Boonstra, F. N. ;
Goossens, H. H. L. M. .
BIOMEDICAL OPTICS EXPRESS, 2017, 8 (02) :712-725
[3]  
Chao Q., 2021, Patent, Patent No. [10977815B1, 10977815]
[4]  
Chin T. C., 2021, Multimedia university iris database (mmu)-v1 and v2
[5]  
Choi H.-S., 2018, INT C LEARN REPR
[6]   Fringe pattern analysis using deep learning [J].
Feng, Shijie ;
Chen, Qian ;
Gu, Guohua ;
Tao, Tianyang ;
Zhang, Liang ;
Hu, Yan ;
Yin, Wei ;
Zuo, Chao .
ADVANCED PHOTONICS, 2019, 1 (02)
[7]  
Funes Mora KennethAlberto., 2012, Computer Vision and Pattern Recognition Workshops (CVPRW), 2012 IEEE Computer Society Conference on, P25, DOI [DOI 10.2307/J.CTT284V4F.5, DOI 10.1109/CVPRW.2012.6239182]
[8]   Fast two-dimensional phase-unwrapping algorithm based on sorting by reliability following a noncontinuous path [J].
Herráez, MA ;
Burton, DR ;
Lalor, MJ ;
Gdeisat, MA .
APPLIED OPTICS, 2002, 41 (35) :7437-7444
[9]   Single-Shot 3D Shape Reconstruction Using Structured Light and Deep Convolutional Neural Networks [J].
Hieu Nguyen ;
Wang, Yuzeng ;
Wang, Zhaoyang .
SENSORS, 2020, 20 (13) :1-13
[10]  
Jinshu Ji, 2018, Multi-disciplinary Trends in Artificial Intelligence. 12th International Conference, MIWAI 2018. Proceedings: Lecture Notes in Artificial Intelligence (LNAI 11248), P69, DOI 10.1007/978-3-030-03014-8_6