Smartphone-based eye tracking system using edge intelligence and model optimisation

被引:2
作者
Gunawardena, Nishan [1 ]
Lui, Gough Yumu [1 ]
Ginige, Jeewani Anupama [1 ]
Javadi, Bahman [1 ]
机构
[1] Western Sydney Univ, Locked Bag 1797, Penrith, NSW 2751, Australia
关键词
Eye tracking; Edge intelligence; Deep learning; Quantisation; Pruning; Energy consumption; Memory usage;
D O I
10.1016/j.iot.2024.101481
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A significant limitation of current smartphone-based eye-tracking algorithms is their low accuracy when applied to video-type visual stimuli, as they are typically trained on static images. Also, the increasing demand for real-time interactive applications like games, VR, and AR on smartphones requires overcoming the limitations posed by resource constraints such as limited computational power, battery life, and network bandwidth. Therefore, we developed two new smartphone eye-tracking techniques for video-type visuals by combining Convolutional Neural Networks (CNN) with two different Recurrent Neural Networks (RNN), namely Long Short Term Memory (LSTM) and Gated Recurrent Unit (GRU). Our CNN+LSTM and CNN+GRU models achieved an average Root Mean Square Error of 0.955 cm and 1.091 cm, respectively. To address the computational constraints of smartphones, we developed an edge intelligence architecture to enhance the performance of smartphone-based eye tracking. We applied various optimisation methods like quantisation and pruning to deep learning models for better energy, CPU, and memory usage on edge devices, focusing on real-time processing. Using model quantisation, the model inference time in the CNN+LSTM and CNN+GRU models was reduced by 21.72% and 19.50%, respectively, on edge devices.
引用
收藏
页数:17
相关论文
共 40 条
[21]  
Li J., 2019, Trends in virtual and augmented reality research: A review of latest eye tracking research papers and beyond
[22]   Computation Offloading Toward Edge Computing [J].
Lin, Li ;
Liao, Xiaofei ;
Jin, Hai ;
Li, Peng .
PROCEEDINGS OF THE IEEE, 2019, 107 (08) :1584-1607
[23]   Eye Tracking-Based Workload and Performance Assessment for Skill Acquisition [J].
Mark, Jesse ;
Curtin, Adrian ;
Kraft, Amanda ;
Sands, Trevor ;
Casebeer, William D. ;
Ziegler, Matthias ;
Ayaz, Hasan .
ADVANCES IN NEUROERGONOMICS AND COGNITIVE ENGINEERING, 2020, 953 :129-141
[24]   COMPARISON OF PREDICTED AND OBSERVED SECONDARY STRUCTURE OF T4 PHAGE LYSOZYME [J].
MATTHEWS, BW .
BIOCHIMICA ET BIOPHYSICA ACTA, 1975, 405 (02) :442-451
[25]   Ambient intelligence framework for real-time speech-to-sign translation [J].
Otoom, Mwaffaq ;
Alzubaidi, Mohammad A. .
ASSISTIVE TECHNOLOGY, 2018, 30 (03) :119-132
[26]  
Palmero C., 2018, P BRIT MACH VIS C BM, V1-12
[27]   SearchGazer: Webcam Eye Tracking for Remote Studies of Web Search [J].
Papoutsaki, Alexandra ;
Laskey, James ;
Huang, Jeff .
CHIIR'17: PROCEEDINGS OF THE 2017 CONFERENCE HUMAN INFORMATION INTERACTION AND RETRIEVAL, 2017, :17-26
[28]   Towards End-to-End Video-Based Eye-Tracking [J].
Park, Seonwook ;
Aksan, Emre ;
Zhang, Xucong ;
Hilliges, Otmar .
COMPUTER VISION - ECCV 2020, PT XII, 2020, 12357 :747-763
[29]   Image visualization: Dynamic and static images generate users' visual cognitive experience using eye-tracking technology [J].
Pei, Huining ;
Huang, Xueqin ;
Ding, Man .
DISPLAYS, 2022, 73
[30]   Pre-learning vocabulary before viewing captioned video: an eye-tracking study [J].
Perez, Maribel Montero .
LANGUAGE LEARNING JOURNAL, 2019, 47 (04) :460-478