EmMixformer: Mix Transformer for Eye Movement Recognition

被引:0
|
作者
Qin, Huafeng [1 ,2 ]
Zhu, Hongyu [1 ,2 ]
Jin, Xin [1 ,2 ]
Song, Qun [1 ,2 ]
El-Yacoubi, Mounim A. [3 ]
Gao, Xinbo [4 ]
机构
[1] Chongqing Technol & Business Univ, Natl Res Base Intelligent Mfg Serv, Chongqing 400067, Peoples R China
[2] Chongqing Microvein Intelligent Technol Co, Chongqing 400053, Peoples R China
[3] Inst Polytech Paris, SAMOVAR, Telecom SudParis, Palaiseau 91120, France
[4] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
关键词
Feature extraction; Transformers; Biometrics; Iris recognition; Long short term memory; Gaze tracking; Fourier transforms; Support vector machines; Data mining; Training; eye movements; Fourier transform; long short-term memory (LSTM); Transformer;
D O I
10.1109/TIM.2025.3551452
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Eye movement is a new, highly secure biometric behavioral modality that has received increasing attention in recent years. Although deep neural networks, such as convolutional neural networks (CNNs), have recently achieved promising performance (e.g., achieving the highest recognition accuracy on the GazeBase database), current solutions fail to capture local and global temporal dependencies within eye movement data. To overcome this problem, we propose a mixed Transformer termed EmMixformer to extract time- and frequency-domain information for eye movement recognition in this article. To this end, we propose a mixed block consisting of three modules: a Transformer, attention long short-term memory (LSTM), and a Fourier Transformer. We are the first to attempt leveraging Transformers to learn long temporal dependencies in eye movement. Second, we incorporate the attention mechanism into the LSTM to propose attention LSTM (attLSTM) to learn short temporal dependencies. Third, we perform self-attention in the frequency domain to learn global dependencies and understand the underlying principles of periodicity. As the three modules provide complementary feature representations regarding local and global dependencies, the proposed EmMixformer can improve recognition accuracy. The experimental results on our eye movement dataset and two public eye movement datasets show that the proposed EmMixformer outperforms the state-of-the-art (SOTA) by achieving the lowest verification error. The EMg- lasses database is available at https://github.com/HonyuZhu-s/CTBU-EMglasses-database.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] User Authentication by Eye Movement Features Employing SVM and XGBoost Classifiers
    Odya, Piotr
    Gorski, Franciszek
    Czyzewski, Andrzej
    IEEE ACCESS, 2023, 11 : 93341 - 93353
  • [32] Effects of depth of field on eye movement
    Zhang, Tingting
    Xia, Ling
    Liu, Xiaofeng
    Kong, Weijie
    Wu, Xiaoli
    JOURNAL OF ENGINEERING-JOE, 2019, 2019 (23): : 9157 - 9161
  • [33] Keep Your Eye on the Best: Contrastive Regression Transformer for Skill Assessment in Robotic Surgery
    Anastasiou, Dimitrios
    Jin, Yueming
    Stoyanov, Danail
    Mazomenos, Evangelos
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (03) : 1755 - 1762
  • [34] Emotion Recognition With Multimodal Transformer Fusion Framework Based on Acoustic and Lexical Information
    Guo, Lili
    Wang, Longbiao
    Dang, Jianwu
    Fu, Yahui
    Liu, Jiaxing
    Ding, Shifei
    IEEE MULTIMEDIA, 2022, 29 (02) : 94 - 103
  • [35] A Convolutional-Transformer-Based Approach for Dynamic Gesture Recognition of Data Gloves
    Tang, Yingzhe
    Pan, Mingzhang
    Li, Hongqi
    Cao, Xinxin
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 13
  • [36] Label Enhancement-Based Multiscale Transformer for Palm-Vein Recognition
    Qin, Huafeng
    Gong, Changqing
    Li, Yantao
    Gao, Xinbo
    El-Yacoubi, Mounim A.
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [37] A Transformer-Based Unsupervised Domain Adaptation Method for Skeleton Behavior Recognition
    Yan, Qiuyan
    Hu, Yan
    IEEE ACCESS, 2023, 11 : 51689 - 51700
  • [38] A Transformer-Based Deep Learning Network for Underwater Acoustic Target Recognition
    Feng, Sheng
    Zhu, Xiaoqian
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [39] Learning 3D Skeletal Representation From Transformer for Action Recognition
    Cha, Junuk
    Saqlain, Muhammad
    Kim, Donguk
    Lee, Seungeun
    Lee, Seongyeong
    Baek, Seungryul
    IEEE ACCESS, 2022, 10 : 67541 - 67550
  • [40] In search of salience: A response-time and eye-movement analysis of bookmark recognition
    Poole, A
    Ball, LJ
    Phillips, P
    PEOPLE AND COMPUTERS XVIII - DESIGN FOR LIFE, 2005, : 363 - 378