Multilevel parallel attention knowledge distillation for multimodal biometric recognition

被引:0
作者
Lu, Kaikui [1 ]
Wu, Ruizhi [1 ]
Bao, Ergude [1 ]
机构
[1] Beijing Jiaotong Univ, Sch Software Engn, Beijing 100044, Peoples R China
关键词
Multimodal biometric recognition; Knowledge distillation; Attention fusion; Activation matrices;
D O I
10.1016/j.engappai.2025.110865
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multimodal Biometric Recognition (MBR) is a biometric technology designed to meet stringent security requirements in high-stakes scenarios. During the last decade, researchers have proposed various approaches to improve MBR, such as combining redundant biometric features and designing sophisticated recognition networks. However, the increasing complexity of these networks and the integration of additional features have substantially increased computational demands and prolonged training cycles, thereby hindering deployment on resource-constrained embedded systems. To address the aforementioned computational and deployment constraints, this paper proposes a Multilevel Parallel Attention Knowledge Distillation (MPAD) framework to compress MBR models. Unlike traditional knowledge distillation frameworks that only transfer unimodal features, our approach transfers attention activation maps from a complex teacher network to a lightweight student network through hierarchical layer alignment. Specifically, instead of guiding the student network's learning of each modality with isolated teacher's features, MPAD implements multilevel parallel attention fusion by combining spatial and channel attention mechanisms. This dual-attention design enables the model to simultaneously prioritize critical spatial regions and channel-wise dependencies, thereby enhancing the representational fidelity of distilled knowledge. Furthermore, a linear attention module generates batch-level activation matrices, which enforce similarity constraints between the feature extraction processes of the student and teacher networks. Extensive experiments on two real-world multimodal biometric datasets demonstrate that student models trained with our approach outperform those trained with other advanced knowledge distillation methods.
引用
收藏
页数:16
相关论文
共 45 条
[1]   Enhanced multimodal biometric recognition systems based on deep learning and traditional methods in smart environments [J].
A. El-Rahman, Sahar ;
Alluhaidan, Ala Saleh .
PLOS ONE, 2024, 19 (02)
[2]  
Alwahaishi Saleh, 2020, 2020 IEEE International Conference on Cloud Computing in Emerging Markets (CCEM), P87, DOI 10.1109/CCEM50674.2020.00027
[3]   Densely Distilled Flow-Based Knowledge Transfer in Teacher-Student Framework for Image Classification [J].
Bae, Ji-Hoon ;
Yeo, Doyeob ;
Yim, Junho ;
Kim, Nae-Soo ;
Pyo, Cheol-Sig ;
Kim, Junmo .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :5698-5710
[4]   Artificial intelligence-Enabled deep learning model for multimodal biometric fusion [J].
Byeon, Haewon ;
Raina, Vikas ;
Sandhu, Mukta ;
Shabaz, Mohammad ;
Keshta, Ismail ;
Soni, Mukesh ;
Matrouk, Khaled ;
Singh, Pavitar Parkash ;
Lakshmi, T. R. Vijaya .
MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (33) :80105-80128
[5]  
Chen K., 2025, arXiv
[6]  
Dahea W., 2018, INT J RES ADV ENG TE, V4, P25, DOI DOI 10.22271/ENGINEERING/25-31
[7]   Knowledge Distillation: A Survey [J].
Gou, Jianping ;
Yu, Baosheng ;
Maybank, Stephen J. ;
Tao, Dacheng .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (06) :1789-1819
[8]   Face-Iris multimodal biometric recognition system based on deep learning [J].
Hattab, Abdessalam ;
Behloul, Ali .
MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (14) :43349-43376
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]  
Hinton G, 2015, Arxiv, DOI [arXiv:1503.02531, 10.48550/arXiv.1503.02531, DOI 10.48550/ARXIV.1503.02531]