SonarGuard: Ultrasonic Face Liveness Detection on Mobile Devices

被引:5
作者
Zhang, Dongheng [1 ]
Meng, Jia [2 ]
Zhang, Jian [2 ]
Deng, Xinzhe [2 ]
Ding, Shouhong [2 ]
Zhou, Man [3 ]
Wang, Qian [4 ]
Li, Qi [5 ,6 ]
Chen, Yan [1 ]
机构
[1] Univ Sci & Technol China, Sch Cyber Sci & Technol, Hefei 230026, Peoples R China
[2] Tencent YouTu Lab, Shanghai 200235, Peoples R China
[3] Huazhong Univ Sci & Technol, Hubei Engn Res Ctr Big Data Secur, Sch Cyber Sci & Engn, Hubei Key Lab Distributed Syst Secur, Wuhan 430074, Peoples R China
[4] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Peoples R China
[5] Zhongguancun Lab, Beijing 100194, Peoples R China
[6] Tsinghua Univ, Inst Network Sci & Cyberspace, Beijing 100084, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Liveness detection; ultrasound signal processing; information fusion; HALLUCINATION;
D O I
10.1109/TCSVT.2023.3236303
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Liveness detection has been widely applied in face authentication systems to combat malicious attacks. However, existing methods purely depending on visual frames become vulnerable once visual perception is not reliable. The emerging face spoof and forge techniques urge the systems to exploit the defensive potential of non-visual modalities. To tackle this challenge, we introduce SonarGuard, a system combining ultrasonic and visual information to achieve robust liveness detection on mobile devices. More specifically, SonarGuard simultaneously extracts micro-doppler signatures from ultrasound reflections and motion trajectories from video frames both corresponding to the user's lip movement. To further confirm the collected ultrasonic and visual information is not derived from malicious audio/video attacks, we consolidate the system via introducing a cross-modal matching mechanism, which demands the inherent consistency between these two modalities. Extensive experiments on a new dataset collected with existing mobile devices demonstrate that the proposed system could achieve average classification error rate of 0.91% under presentation attacks. This result indicates that SonarGuard can boost the security of face authenfication systems in real world usage without additional hardware modification.
引用
收藏
页码:4401 / 4414
页数:14
相关论文
共 55 条
[2]  
Atoum Y, 2017, 2017 IEEE INTERNATIONAL JOINT CONFERENCE ON BIOMETRICS (IJCB), P319, DOI 10.1109/BTAS.2017.8272713
[3]  
Boulkenafet Z, 2015, IEEE IMAGE PROC, P2636, DOI 10.1109/ICIP.2015.7351280
[4]  
Carion N, 2020, Arxiv, DOI [arXiv:2005.12872, DOI 10.48550/ARXIV.2005.12872]
[5]   EchoFace: Acoustic Sensor-Based Media Attack Detection for Face Authentication [J].
Chen, Huangxun ;
Wang, Wei ;
Zhang, Jin ;
Zhang, Qian .
IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (03) :2152-2159
[6]   Micro-doppler effect in radar: Phenomenon, model, and simulation study [J].
Chen, VC ;
Li, FY ;
Ho, SS ;
Wechsler, H .
IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2006, 42 (01) :2-21
[7]  
Chen Y., 2019, CISC VIS NETW IND GL
[8]  
Dale PurvesGeorge J Augustine., 2004, Neuroscience, V3rd
[9]  
Devlin J, 2019, Arxiv, DOI [arXiv:1810.04805, 10.48550/arxiv.1810.04805]
[10]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929