EmMixformer: Mix Transformer for Eye Movement Recognition

被引:0
|
作者
Qin, Huafeng [1 ,2 ]
Zhu, Hongyu [1 ,2 ]
Jin, Xin [1 ,2 ]
Song, Qun [1 ,2 ]
El-Yacoubi, Mounim A. [3 ]
Gao, Xinbo [4 ]
机构
[1] Chongqing Technol & Business Univ, Natl Res Base Intelligent Mfg Serv, Chongqing 400067, Peoples R China
[2] Chongqing Microvein Intelligent Technol Co, Chongqing 400053, Peoples R China
[3] Inst Polytech Paris, SAMOVAR, Telecom SudParis, Palaiseau 91120, France
[4] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
关键词
Feature extraction; Transformers; Biometrics; Iris recognition; Long short term memory; Gaze tracking; Fourier transforms; Support vector machines; Data mining; Training; eye movements; Fourier transform; long short-term memory (LSTM); Transformer;
D O I
10.1109/TIM.2025.3551452
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Eye movement is a new, highly secure biometric behavioral modality that has received increasing attention in recent years. Although deep neural networks, such as convolutional neural networks (CNNs), have recently achieved promising performance (e.g., achieving the highest recognition accuracy on the GazeBase database), current solutions fail to capture local and global temporal dependencies within eye movement data. To overcome this problem, we propose a mixed Transformer termed EmMixformer to extract time- and frequency-domain information for eye movement recognition in this article. To this end, we propose a mixed block consisting of three modules: a Transformer, attention long short-term memory (LSTM), and a Fourier Transformer. We are the first to attempt leveraging Transformers to learn long temporal dependencies in eye movement. Second, we incorporate the attention mechanism into the LSTM to propose attention LSTM (attLSTM) to learn short temporal dependencies. Third, we perform self-attention in the frequency domain to learn global dependencies and understand the underlying principles of periodicity. As the three modules provide complementary feature representations regarding local and global dependencies, the proposed EmMixformer can improve recognition accuracy. The experimental results on our eye movement dataset and two public eye movement datasets show that the proposed EmMixformer outperforms the state-of-the-art (SOTA) by achieving the lowest verification error. The EMg- lasses database is available at https://github.com/HonyuZhu-s/CTBU-EMglasses-database.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] The Tiny Eye Movement Transformer
    Fuhl, Wolfgang
    Werner, Anne Herrmann
    Nieselt, Kay
    ACM SYMPOSIUM ON EYE TRACKING RESEARCH & APPLICATIONS, ETRA 2023, 2023,
  • [2] IrisFormer: A Dedicated Transformer Framework for Iris Recognition
    Sun, Xianyun
    Wang, Caiyong
    Wang, Yunlong
    Wei, Jianze
    Sun, Zhenan
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 431 - 435
  • [3] On Designing a SwinIris Transformer Based Iris Recognition System
    Gao, Runqing
    Bourlai, Thirimachos
    IEEE ACCESS, 2024, 12 : 30723 - 30737
  • [4] Modulation of Mood on Eye Movement and Face Recognition Performance
    An, Jeehye
    Hsiao, Janet H.
    EMOTION, 2021, 21 (03) : 617 - 630
  • [5] Recognition of power transformer winding movement and deformation using FRA
    Saied, Mohamed M.
    Al-Shaher, Meshal A.
    COMPEL-THE INTERNATIONAL JOURNAL FOR COMPUTATION AND MATHEMATICS IN ELECTRICAL AND ELECTRONIC ENGINEERING, 2007, 26 (05) : 1392 - 1410
  • [6] Combined Activity Recognition Based on Continuous-Wave Radar and Vision Transformer
    Zhou, Junhao
    Sun, Chao
    Kim, Youngok
    IEEE ACCESS, 2024, 12 : 185448 - 185459
  • [7] Adversarial Domain Generalized Transformer for Cross-Corpus Speech Emotion Recognition
    Gao, Yuan
    Wang, Longbiao
    Liu, Jiaxing
    Dang, Jianwu
    Okada, Shogo
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (02) : 697 - 708
  • [8] SIGNFORMER: DeepVision Transformer for Sign Language Recognition
    Kothadiya, Deep R.
    Bhatt, Chintan M.
    Saba, Tanzila
    Rehman, Amjad
    Bahaj, Saeed Ali
    IEEE ACCESS, 2023, 11 : 4730 - 4739
  • [9] MTT: Multi-Scale Temporal Transformer for Skeleton-Based Action Recognition
    Kong, Jun
    Bian, Yuhang
    Jiang, Min
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 528 - 532
  • [10] PARFormer: Transformer-Based Multi-Task Network for Pedestrian Attribute Recognition
    Fan, Xinwen
    Zhang, Yukang
    Lu, Yang
    Wang, Hanzi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (01) : 411 - 423