Contrastive Regression for Domain Adaptation on Gaze Estimation

被引:50
作者
Wang, Yaoming [1 ,2 ]
Jiang, Yangzhou [1 ,2 ]
Li, Jin [1 ]
Ni, Bingbing [1 ]
Dai, Wenrui [1 ]
Li, Chenglin [1 ]
Xiong, Hongkai [1 ]
Li, Teng [2 ,3 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[2] Huawei Inc, Shenzhen, Peoples R China
[3] Anhui Univ, Hefei, Peoples R China
来源
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022) | 2022年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52688.2022.01877
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Appearance-based Gaze Estimation leverages deep neural networks to regress the gaze direction from monocular images and achieve impressive performance. However, its success depends on expensive and cumbersome annotation capture. When lacking precise annotation, the large domain gap hinders the performance of trained models on new domains. In this paper, we propose a novel gaze adaptation approach, namely Contrastive Regression Gaze Adaptation (CRGA), for generalizing gaze estimation on the target domain in an unsupervised manner. CRGA leverages the Contrastive Domain Generalization (CDG) module to learn the stable representation from the source domain and leverages the Contrastive Self-training Adaptation (CSA) module to learn from the pseudo labels on the target domain. The core of both CDG and CSA is the Contrastive Regression (CR) loss, a novel contrastive loss for regression by pulling features with closer gaze directions closer together while pushing features with farther gaze directions farther apart. Experimentally, we choose ETH-XGAZE and Gaze-360 as the source domain and test the domain generalization and adaptation performance on MPIIGAZE, RT-GENE, Gaze-Capture, EyeDiap respectively. The results demonstrate that our CRGA achieves remarkable performance improvement compared with the baseline models and also outperforms the state-of-the-art domain adaptation approaches on gaze adaptation tasks.
引用
收藏
页码:19354 / 19363
页数:10
相关论文
共 39 条
[1]  
[Anonymous], 2015, CVPR, DOI [10.1109/CVPR.2015.7299081, DOI 10.1109/CVPR.2015.7299081]
[2]  
[Anonymous], 2020, ARXIV200610029
[3]  
[Anonymous], 2020, ARXIV200607733
[4]  
Bachman P, 2019, ADV NEUR IN, V32
[5]   Utilizing VR and Gaze Tracking to Develop AR Solutions for Industrial Maintenance [J].
Burova, Alisa ;
Makela, John ;
Hakulinen, Jaakko ;
Keskinen, Tuuli ;
Heinonen, Hanna ;
Siltanen, Sanni ;
Turunen, Markku .
PROCEEDINGS OF THE 2020 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'20), 2020,
[6]  
Caron M, 2020, ADV NEUR IN, V33
[7]   Deep semantic gaze embedding and scanpath comparison for expertise classification during OPT viewing [J].
Castner, Nora ;
Kuebler, Thomas ;
Scheiter, Katharina ;
Richter, Juliane ;
Eder, Therese ;
Huettig, Fabian ;
Keutel, Constanze ;
Kasneci, Enkelejda .
ETRA'20 FULL PAPERS: ACM SYMPOSIUM ON EYE TRACKING RESEARCH AND APPLICATIONS, 2020,
[8]  
Cheng Yihua, 2021, arXiv
[9]  
Cheng Yihua, 2021, ARXIV210313173
[10]   RT-GENE: Real-Time Eye Gaze Estimation in Natural Environments [J].
Fischer, Tobias ;
Chang, Hyung Jin ;
Demiris, Yiannis .
COMPUTER VISION - ECCV 2018, PT X, 2018, 11214 :339-357