Appearance-Based Gaze Estimation With Deep Learning: A Review and Benchmark

被引:28
作者
Cheng, Yihua [1 ]
Wang, Haofei [2 ]
Bao, Yiwei [1 ]
Lu, Feng [1 ,2 ]
机构
[1] Beihang Univ, State Key Lab Virtual Real Technol & Syst, SCSE, Beijing 100191, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518066, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Benchmark; deep learning; eye appearance; review; gaze estimation; EYE GAZE; TRACKING TECHNIQUES; MODEL; DIFFERENCE; FEATURES;
D O I
10.1109/TPAMI.2024.3393571
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Human gaze provides valuable information on human focus and intentions, making it a crucial area of research. Recently, deep learning has revolutionized appearance-based gaze estimation. However, due to the unique features of gaze estimation research, such as the unfair comparison between 2D gaze positions and 3D gaze vectors and the different pre-processing and post-processing methods, there is a lack of a definitive guideline for developing deep learning-based gaze estimation algorithms. In this paper, we present a systematic review of the appearance-based gaze estimation methods using deep learning. First, we survey the existing gaze estimation algorithms along the typical gaze estimation pipeline: deep feature extraction, deep learning model design, personal calibration and platforms. Second, to fairly compare the performance of different approaches, we summarize the data pre-processing and post-processing methods, including face/eye detection, data rectification, 2D/3D gaze conversion and gaze origin conversion. Finally, we set up a comprehensive benchmark for deep learning-based gaze estimation. We characterize all the public datasets and provide the source code of typical gaze estimation algorithms. This paper serves not only as a reference to develop deep learning-based gaze estimation methods, but also a guideline for future gaze estimation research.
引用
收藏
页码:7509 / 7528
页数:20
相关论文
共 186 条
[1]  
[Anonymous], 2006, 2006 IEEE COMP SOC C
[2]  
Balim H., 2023, IEEE/CVF Conf. Comput. Vision Pattern Recognit. Workshops (CVPRW), V2023, P2688, DOI DOI 10.1109/CVPRW59228.2023.00269
[3]   OpenFace 2.0: Facial Behavior Analysis Toolkit [J].
Baltrusaitis, Tadas ;
Zadeh, Amir ;
Lim, Yao Chong ;
Morency, Louis-Philippe .
PROCEEDINGS 2018 13TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2018), 2018, :59-66
[4]  
Baluja Shumeet., 1994, NONINTRUSIVE GAZE TR
[5]   An Individual-Difference-Aware Model for Cross-Person Gaze Estimation [J].
Bao, Jun ;
Liu, Buyu ;
Yu, Jun .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 :3322-3333
[6]   Generalizing Gaze Estimation with Rotation Consistency [J].
Bao, Yiwei ;
Liu, Yunfei ;
Wang, Haofei ;
Lu, Feng .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :4197-4206
[7]   Adaptive Feature Fusion Network for Gaze Tracking in Mobile Tablets [J].
Bao, Yiwei ;
Cheng, Yihua ;
Liu, Yunfei ;
Lu, Feng .
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, :9936-9943
[8]  
BealMJ, 2003, Variational algorithms for approximate Bayesian inference
[9]   Deep Learning based Eye Gaze Tracking for Automotive Applications: An Auto-Keras Approach [J].
Bublea, Adrian ;
Caleanu, Catalin Daniel .
2020 14TH INTERNATIONAL SYMPOSIUM ON ELECTRONICS AND TELECOMMUNICATIONS (ISETC), 2020, :48-51
[10]  
Cai X., 2021, arXiv