EmMixformer: Mix Transformer for Eye Movement Recognition

被引:0
|
作者
Qin, Huafeng [1 ,2 ]
Zhu, Hongyu [1 ,2 ]
Jin, Xin [1 ,2 ]
Song, Qun [1 ,2 ]
El-Yacoubi, Mounim A. [3 ]
Gao, Xinbo [4 ]
机构
[1] Chongqing Technol & Business Univ, Natl Res Base Intelligent Mfg Serv, Chongqing 400067, Peoples R China
[2] Chongqing Microvein Intelligent Technol Co, Chongqing 400053, Peoples R China
[3] Inst Polytech Paris, SAMOVAR, Telecom SudParis, Palaiseau 91120, France
[4] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
关键词
Feature extraction; Transformers; Biometrics; Iris recognition; Long short term memory; Gaze tracking; Fourier transforms; Support vector machines; Data mining; Training; eye movements; Fourier transform; long short-term memory (LSTM); Transformer;
D O I
10.1109/TIM.2025.3551452
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Eye movement is a new, highly secure biometric behavioral modality that has received increasing attention in recent years. Although deep neural networks, such as convolutional neural networks (CNNs), have recently achieved promising performance (e.g., achieving the highest recognition accuracy on the GazeBase database), current solutions fail to capture local and global temporal dependencies within eye movement data. To overcome this problem, we propose a mixed Transformer termed EmMixformer to extract time- and frequency-domain information for eye movement recognition in this article. To this end, we propose a mixed block consisting of three modules: a Transformer, attention long short-term memory (LSTM), and a Fourier Transformer. We are the first to attempt leveraging Transformers to learn long temporal dependencies in eye movement. Second, we incorporate the attention mechanism into the LSTM to propose attention LSTM (attLSTM) to learn short temporal dependencies. Third, we perform self-attention in the frequency domain to learn global dependencies and understand the underlying principles of periodicity. As the three modules provide complementary feature representations regarding local and global dependencies, the proposed EmMixformer can improve recognition accuracy. The experimental results on our eye movement dataset and two public eye movement datasets show that the proposed EmMixformer outperforms the state-of-the-art (SOTA) by achieving the lowest verification error. The EMg- lasses database is available at https://github.com/HonyuZhu-s/CTBU-EMglasses-database.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Optimized Feature Mapping for Eye Movement Recognition using Electrooculogram Signals
    Mulam, Harikrishna
    Mudigonda, Malini
    2017 8TH INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION AND NETWORKING TECHNOLOGIES (ICCCNT), 2017,
  • [22] Automated Eye Movement Classification Based on EMG of EOM Signals Using FBSE-EWT Technique
    Khan, Sibghatullah Inayatullah
    Pachori, Ram Bilas
    IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2023, 53 (02) : 346 - 356
  • [23] Emotion Recognition by Integrating Eye Movement Analysis and Facial Expression Model
    Thong Van Huynh
    Yang, Hyung-Jeong
    Lee, Guee-Sang
    Kim, Soo-Hyung
    Na, In-Seop
    PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND SOFT COMPUTING (ICMLSC 2019), 2019, : 166 - 169
  • [24] Two-Stream Proximity Graph Transformer for Skeletal Person-Person Interaction Recognition With Statistical Information
    Li, Meng
    Wu, Yaqi
    Sun, Qiumei
    Yang, Weifeng
    IEEE ACCESS, 2024, 12 : 193091 - 193100
  • [25] Convolutional Transformer Fusion Blocks for Multi-Modal Gesture Recognition
    Hampiholi, Basavaraj
    Jarvers, Christian
    Mader, Wolfgang
    Neumann, Heiko
    IEEE ACCESS, 2023, 11 : 34094 - 34103
  • [26] Prepended Domain Transformer: Heterogeneous Face Recognition Without Bells and Whistles
    George, Anjith
    Mohammadi, Amir
    Marcel, Sebastien
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 133 - 146
  • [27] Bi-Branch Vision Transformer Network for EEG Emotion Recognition
    Lu, Wei
    Tan, Tien-Ping
    Ma, Hua
    IEEE ACCESS, 2023, 11 : 36233 - 36243
  • [28] Hybrid Autoregressive and Non-Autoregressive Transformer Models for Speech Recognition
    Tian, Zhengkun
    Yi, Jiangyan
    Tao, Jianhua
    Zhang, Shuai
    Wen, Zhengqi
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 762 - 766
  • [29] ClST: A Convolutional Transformer Framework for Automatic Modulation Recognition by Knowledge Distillation
    Hou, Dongbin
    Li, Lixin
    Lin, Wensheng
    Liang, Junli
    Han, Zhu
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2024, 23 (07) : 8013 - 8028
  • [30] Contextual Transformer Networks for Visual Recognition
    Li, Yehao
    Yao, Ting
    Pan, Yingwei
    Mei, Tao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) : 1489 - 1500