Smartphone-based eye tracking system using edge intelligence and model optimisation

被引:1
作者
Gunawardena, Nishan [1 ]
Lui, Gough Yumu [1 ]
Ginige, Jeewani Anupama [1 ]
Javadi, Bahman [1 ]
机构
[1] Western Sydney Univ, Locked Bag 1797, Penrith, NSW 2751, Australia
关键词
Eye tracking; Edge intelligence; Deep learning; Quantisation; Pruning; Energy consumption; Memory usage;
D O I
10.1016/j.iot.2024.101481
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A significant limitation of current smartphone-based eye-tracking algorithms is their low accuracy when applied to video-type visual stimuli, as they are typically trained on static images. Also, the increasing demand for real-time interactive applications like games, VR, and AR on smartphones requires overcoming the limitations posed by resource constraints such as limited computational power, battery life, and network bandwidth. Therefore, we developed two new smartphone eye-tracking techniques for video-type visuals by combining Convolutional Neural Networks (CNN) with two different Recurrent Neural Networks (RNN), namely Long Short Term Memory (LSTM) and Gated Recurrent Unit (GRU). Our CNN+LSTM and CNN+GRU models achieved an average Root Mean Square Error of 0.955 cm and 1.091 cm, respectively. To address the computational constraints of smartphones, we developed an edge intelligence architecture to enhance the performance of smartphone-based eye tracking. We applied various optimisation methods like quantisation and pruning to deep learning models for better energy, CPU, and memory usage on edge devices, focusing on real-time processing. Using model quantisation, the model inference time in the CNN+LSTM and CNN+GRU models was reduced by 21.72% and 19.50%, respectively, on edge devices.
引用
收藏
页数:17
相关论文
共 40 条
  • [1] Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
  • [2] Ali A. A., 2019, Asian Journal of Research in Computer Science, V2, P1, DOI [10.9734/ajrcos/2019/v4i230108, DOI 10.9734/AJRCOS/2018/V2I430080]
  • [3] Bƒce M, 2019, Arxiv, DOI arXiv:1907.11115
  • [4] Understanding the Characteristics of Mobile Augmented Reality Applications
    Chen, Huixiang
    Dai, Yuting
    Meng, Hao
    Chen, Yilun
    Li, Tao
    [J]. 2018 IEEE INTERNATIONAL SYMPOSIUM ON PERFORMANCE ANALYSIS OF SYSTEMS AND SOFTWARE (ISPASS), 2018, : 128 - 138
  • [5] Quantization of Deep Neural Networks for Accurate Edge Computing
    Chen, Wentao
    Qiu, Hailong
    Zhuang, Jian
    Zhang, Chutong
    Hu, Yu
    Lu, Qing
    Wang, Tianchen
    Shi, Yiyu
    Huang, Meiping
    Xu, Xiaowe
    [J]. ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2021, 17 (04)
  • [6] Dao T.-C., 2014, Real-Time Eye Tracking Analysis from Large Scale Dataset using Cloud Computing
  • [7] Diao E., 2023, INT C LEARN REPR ICL
  • [8] Duchowski Andrew, 2007, Eye Tracking Techniques, P51, DOI [10.1007/978-1-84628-609-4_5, DOI 10.1007/978-1-84628-609-4_5]
  • [9] Innovative Hybrid Approach for Masked Face Recognition Using Pretrained Mask Detection and Segmentation, Robust PCA, and KNN Classifier
    Eman, Mohammed
    Mahmoud, Tarek M.
    Ibrahim, Mostafa M.
    Abd El-Hafeez, Tarek
    [J]. SENSORS, 2023, 23 (15)
  • [10] Brain activity correlates with emotional perception induced by dynamic avatars
    Goldberg, Hagar
    Christensen, Andrea
    Flash, Tamar
    Giese, Martin A.
    Malach, Rafael
    [J]. NEUROIMAGE, 2015, 122 : 306 - 317