EmMixformer: Mix Transformer for Eye Movement Recognition

被引:0
|
作者
Qin, Huafeng [1 ,2 ]
Zhu, Hongyu [1 ,2 ]
Jin, Xin [1 ,2 ]
Song, Qun [1 ,2 ]
El-Yacoubi, Mounim A. [3 ]
Gao, Xinbo [4 ]
机构
[1] Chongqing Technol & Business Univ, Natl Res Base Intelligent Mfg Serv, Chongqing 400067, Peoples R China
[2] Chongqing Microvein Intelligent Technol Co, Chongqing 400053, Peoples R China
[3] Inst Polytech Paris, SAMOVAR, Telecom SudParis, Palaiseau 91120, France
[4] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
关键词
Feature extraction; Transformers; Biometrics; Iris recognition; Long short term memory; Gaze tracking; Fourier transforms; Support vector machines; Data mining; Training; eye movements; Fourier transform; long short-term memory (LSTM); Transformer;
D O I
10.1109/TIM.2025.3551452
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Eye movement is a new, highly secure biometric behavioral modality that has received increasing attention in recent years. Although deep neural networks, such as convolutional neural networks (CNNs), have recently achieved promising performance (e.g., achieving the highest recognition accuracy on the GazeBase database), current solutions fail to capture local and global temporal dependencies within eye movement data. To overcome this problem, we propose a mixed Transformer termed EmMixformer to extract time- and frequency-domain information for eye movement recognition in this article. To this end, we propose a mixed block consisting of three modules: a Transformer, attention long short-term memory (LSTM), and a Fourier Transformer. We are the first to attempt leveraging Transformers to learn long temporal dependencies in eye movement. Second, we incorporate the attention mechanism into the LSTM to propose attention LSTM (attLSTM) to learn short temporal dependencies. Third, we perform self-attention in the frequency domain to learn global dependencies and understand the underlying principles of periodicity. As the three modules provide complementary feature representations regarding local and global dependencies, the proposed EmMixformer can improve recognition accuracy. The experimental results on our eye movement dataset and two public eye movement datasets show that the proposed EmMixformer outperforms the state-of-the-art (SOTA) by achieving the lowest verification error. The EMg- lasses database is available at https://github.com/HonyuZhu-s/CTBU-EMglasses-database.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Research on Human-Computer Interaction Intention Recognition Based on EEG and Eye Movement
    Zhao, Minrui
    Gao, Hongni
    Wang, Wei
    Qu, Jue
    IEEE ACCESS, 2020, 8 : 145824 - 145832
  • [42] An Emotion Recognition Method Based on Eye Movement and Audiovisual Features in MOOC Learning Environment
    Bao, Jindi
    Tao, Xiaomei
    Zhou, Yinghui
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (01) : 171 - 183
  • [43] Dyslexia Analysis and Diagnosis Based on Eye Movement
    Vaitheeshwari, R.
    Chen, Chih-Hsuan
    Chung, Chia-Ru
    Yang, Hsuan-Yu
    Yeh, Shih-Ching
    Wu, Eric Hsiao-Kuang
    Kumar, Mukul
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2024, 32 : 4109 - 4119
  • [44] Using results of eye movement signal analysis in the neural network recognition of otoneurological patients
    Juhola, Martti
    Aalto, Heikki
    Hirvonen, Timo
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2007, 86 (03) : 216 - 226
  • [45] Spatial--Temporal Synchronous Transformer for Skeleton-Based Hand Gesture Recognition
    Zhao, Dongdong
    Li, Hongli
    Yan, Shi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (03) : 1403 - 1412
  • [46] A Transformer Convolutional Network With the Method of Image Segmentation for EEG-Based Emotion Recognition
    Zhang, Xinyiy
    Cheng, Xiankai
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 401 - 405
  • [47] Fine-Grained Temporal-Enhanced Transformer for Dynamic Facial Expression Recognition
    Zhang, Yaning
    Zhang, Jiahe
    Shen, Linlin
    Yu, Zitong
    Gao, Zan
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 2560 - 2564
  • [48] Automatic Modulation Recognition of Underwater Acoustic Signals Using a Two-Stream Transformer
    Li, Juan
    Jia, Qingning
    Cui, Xuerong
    Gulliver, T. Aaron
    Jiang, Bin
    Li, Shibao
    Yang, Jungang
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (10): : 18839 - 18851
  • [49] Abandon Locality: Frame-Wise Embedding Aided Transformer for Automatic Modulation Recognition
    Chen, Yantao
    Dong, Binhong
    Liu, Cuiting
    Xiong, Wenhui
    Li, Shaoqian
    IEEE COMMUNICATIONS LETTERS, 2023, 27 (01) : 327 - 331
  • [50] PIDViT: Pose-Invariant Distilled Vision Transformer for Facial Expression Recognition in the Wild
    Huang, Yin-Fu
    Tsai, Chia-Hsin
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (04) : 3281 - 3293