LVID: A Multimodal Biometrics Authentication System on Smartphones

被引:52
作者
Wu, Libing [1 ,2 ]
Yang, Jingxiao [1 ,2 ]
Zhou, Man [2 ,3 ]
Chen, Yanjiao [1 ,2 ]
Wang, Qian [2 ,3 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
[2] State Key Lab Cryptol, Beijing 100878, Peoples R China
[3] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Lips; Authentication; Smart phones; Biometrics (access control); Acoustics; Sensors; Microphones; multimodal biometrics; acoustic sensing; SPEAKER IDENTIFICATION; LIP MOTION;
D O I
10.1109/TIFS.2019.2944058
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Voice authentication is becoming increasingly popular, which offers potential benefits over knowledge and possession based authentication methods. Meanwhile, the unique features of lip movements during speaking have been proved to be useful for authentication. However, the unimodal biometric authentication systems based on either voice or lip movements have certain limitations. Voice authentication systems are prone to spoofing attacks and suffer from serious performance degradation in noisy environments. Lip movements authentication systems are unstable and are sensitive to the user's physical and psychological conditions. In this paper, we propose and implement LVID, a multimodal biometrics authentication system on smartphones, which resolves the defects of the original systems by combining the advantages of lip movements and voice. LVID simultaneously captures these two biometrics with the built-in audio devices on smartphones and fuses them at the data level. The reliable and effective features are then extracted from the fused data for authentication. LVID is practical as it requires neither cumbersome operations nor additional hardwares but only a speaker and a microphone that are commonly available on smartphones. Our experimental results with 104 participants show that LVID can achieve 95 accuracy for user authentication, and 93.47 of the attacks can be detected. It is also verified that LVID works well with different smartphones and is robust to different smartphone positions.
引用
收藏
页码:1572 / 1585
页数:14
相关论文
共 39 条
[1]  
[Anonymous], [No title captured]
[2]  
[Anonymous], 2006, 9 INT C CONTR AUT RO, DOI DOI 10.1109/ICARCV.2006.345473
[3]  
[Anonymous], 2007, Introduction to Biometrics
[4]   Discriminative analysis of lip motion features for speaker identification and speech-reading [J].
Cetinguel, H. Ertan ;
Yemez, Yuecel ;
Erzin, Engin ;
Tekalp, A. Murat .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2006, 15 (10) :2879-2891
[5]   Multimodal speaker/speech recognition using lip motion, lip texture and audio [J].
Cetingul, H. E. ;
Erzin, E. ;
Yemez, Y. ;
Tekalp, A. M. .
SIGNAL PROCESSING, 2006, 86 (12) :3549-3558
[6]   You Can Hear But You Cannot Steal: Defending against Voice Impersonation Attacks on Smartphones [J].
Chen, Si ;
Ren, Kui ;
Piao, Sixu ;
Wang, Cong ;
Wang, Qian ;
Weng, Jian ;
Su, Lu ;
Mohaisen, Aziz .
2017 IEEE 37TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2017), 2017, :183-195
[7]  
Dieckmann U, 1997, LECT NOTES COMPUT SC, V1206, P301, DOI 10.1007/BFb0016009
[8]   SPEECH ENHANCEMENT USING A MINIMUM MEAN-SQUARE ERROR LOG-SPECTRAL AMPLITUDE ESTIMATOR [J].
EPHRAIM, Y ;
MALAH, D .
IEEE TRANSACTIONS ON ACOUSTICS SPEECH AND SIGNAL PROCESSING, 1985, 33 (02) :443-445
[9]   BioID: A multimodal biometric identification system [J].
Frischholz, RW ;
Dieckmann, U .
COMPUTER, 2000, 33 (02) :64-+
[10]  
Gupta S., 2012, P SIGCHI C HUM FACT, V12, P1911, DOI [DOI 10.1145/2207676.2208331, 10.1145/2207676.2208331]