Comparison of Eye-gaze Detection using CNN and Vision Transformer

被引:0
作者
Niikura D. [1 ]
Abe K. [1 ]
机构
[1] Graduate School of System Design and Technology, Tokyo Denki University, 5, Senjuasahicho, Adachi-ku, Tokyo
关键词
convolutional neural network; eye-gaze detection; eye-gaze input; input interface; Vision Transformer;
D O I
10.1541/ieejeiss.144.683
中图分类号
TN911 [通信理论];
学科分类号
081002 ;
摘要
We propose an eye-gaze input system that utilizes a laptop PC and its inner camera. This system can discriminate the user's eye-gaze direction by using Convolutional Neural Network (CNN) or Vision Transformer (ViT). In this paper, we present the results of a comparison of the newly created eye-gaze direction discrimination model of ViT and the past model created by a CNN. We evaluated the accuracy of discrimination models created by ViT and CNN through the experiments. As a result, the ViT model has higher accuracy than the CNN model in discriminating the center direction. © 2024 The Institute of Electrical Engineers of Japan.
引用
收藏
页码:683 / 684
页数:1
相关论文
共 50 条
  • [21] Euclidean Distance based Loss Function for Eye-Gaze Estimation
    Lee, Bu Sung
    Phattharaphon, Romphet
    Yean, Seanglidet
    Liu, Jigang
    Shakya, Manoj
    2020 IEEE SENSORS APPLICATIONS SYMPOSIUM (SAS 2020), 2020,
  • [22] Efficient deepfake detection using shallow vision transformer
    Usmani, Shaheen
    Kumar, Sunil
    Sadhya, Debanjan
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (04) : 12339 - 12362
  • [23] Optimized vision transformer encoder with cnn for automatic psoriasis disease detection
    Vishwakarma, Gagan
    Nandanwar, Amit Kumar
    Thakur, Ghanshyam Singh
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (21) : 59597 - 59616
  • [24] A Hybrid Approach for Pavement Crack Detection Using Mask R-CNN and Vision Transformer Model
    Alshawabkeh, Shorouq
    Wu, Li
    Dong, Daojun
    Cheng, Yao
    Li, Liping
    CMC-COMPUTERS MATERIALS & CONTINUA, 2025, 82 (01): : 561 - 577
  • [25] Improved Deepfake Video Detection Using Convolutional Vision Transformer
    Deressa, Deressa Wodajo
    Lambert, Peter
    Van Wallendael, Glenn
    Atnafu, Solomon
    Mareen, Hannes
    2024 IEEE GAMING, ENTERTAINMENT, AND MEDIA CONFERENCE, GEM 2024, 2024, : 492 - 497
  • [26] Interface using eye-gaze and tablet input for an avatar robot control in class participation support system
    Mu, Shenglin
    Shibata, Satoru
    Yamamoto, Tomonori
    Obayashi, Haruki
    COMPUTERS & ELECTRICAL ENGINEERING, 2023, 111
  • [27] Study on eye-gaze input interface based on deep learning using images obtained by multiple cameras
    Mu, Shenglin
    Shibata, Satoru
    Chiu, Kuo-chun
    Yamamoto, Tomonori
    Liu, Tung-kuan
    COMPUTERS & ELECTRICAL ENGINEERING, 2022, 101
  • [28] Pupil Detection Using Hybrid Vision Transformer
    Wang, Li
    Wang, Changyuan
    Zhang, Yu
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (12)
  • [29] Driver Drowsiness Detection Using Vision Transformer
    Azmi, Muhammad Muizuddin Bin Mohamad
    Zaman, Fadhlan Hafizhelmi Kamaru
    2024 IEEE 14TH SYMPOSIUM ON COMPUTER APPLICATIONS & INDUSTRIAL ELECTRONICS, ISCAIE 2024, 2024, : 329 - 336
  • [30] Fall Event Detection using Vision Transformer
    Dey, Ankita
    Rajan, Sreeraman
    Xiao, George
    Lu, Jianping
    2022 IEEE SENSORS, 2022,