One-shot lip-based biometric authentication: Extending behavioral features with authentication phrase information

被引:5
作者
Koch, Brando [1 ]
Grbic, Ratko [1 ]
机构
[1] Fac Elect Engn Comp Sci & Informat Technol Osijek, Kneza Trpimira 2B, HR-31000 Osijek, Croatia
关键词
Lip-based biometric authentication; Siamese neural network; Hard-negative mining; Presentation attack detection; One-shot learning; GRID dataset; PERSON AUTHENTICATION; FACE; IDENTIFICATION; SPEECH; MOTION;
D O I
10.1016/j.imavis.2024.104900
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Lip -based biometric authentication (LBBA) is an authentication method based on a person's lip movements during speech in the form of video data. LBBA can utilize both physical and behavioral characteristics of lip movements without requiring any additional sensory equipment apart from an RGB camera. Current approaches employ deep siamese neural networks trained with one-shot learning to generate embedding vectors from lip movement features. However, most of these approaches don't discriminate against speech content which makes them vulnerable to video replay attacks. Moreover, there is a lack of comprehensive analysis regarding the impact of distinct lip characteristics or difficult dataset phrases with significant word overlap on the performance of authentication in one-shot approaches. To address this, we introduce the GRID-CCP dataset and train a siamese neural network using 3D convolutions and recurrent neural network layers to additionally discriminate against speech content. For loss calculation, we propose a custom triplet loss function for efficient and customizable batch -wise hard -negative mining. Our experimental results, using an open -set protocol, demonstrate a False Acceptance Rate (FAR) of 3.2% and a False Rejection Rate (FRR) of 3.8% on the test set of the GRID-CCP dataset. Finally, we conduct an analysis to assess the influence and discriminative power of behavioral and physical features in LBBA.
引用
收藏
页数:12
相关论文
empty
未找到相关数据