Deep-learning-based real-time silent speech recognition using facial electromyogram recorded around eyes for hands-free interfacing in a virtual reality environment

被引:0
作者
Ho-Seung Cha
Won-Du Chang
Chang-Hwan Im
机构
[1] Hanyang University,Department of Biomedical Engineering
[2] Pukyong National University,Department of Computer Engineering
来源
Virtual Reality | 2022年 / 26卷
关键词
Deep learning; Facial electromyography; Human–computer interface; Myoelectric control; Silent speech recognition; Virtual reality;
D O I
暂无
中图分类号
学科分类号
摘要
Speech recognition technology is a promising hands-free interfacing modality for virtual reality (VR) applications. However, it has several drawbacks, such as limited usability in a noisy environment or a public place and limited accessibility to those who cannot generate loud and clear voices. These limitations may be overcome by employing a silent speech recognition (SSR) technology utilizing facial electromyograms (fEMGs) in a VR environment. In the conventional SSR systems, however, fEMG electrodes were attached around the user’s lips and neck, thereby creating new practical issues, such as the requirement of an additional wearable system besides the VR headset, necessity of a complex and time-consuming procedure for attaching the fEMG electrodes, and discomfort and limited facial muscle movements of the user. To solve these problems, we propose an SSR system using fEMGs measured by a few electrodes attached around the eyes of a user, which can also be easily incorporated into available VR headsets. To enhance the accuracy of classifying the fEMG signals recorded from limited recording locations relatively far from the phonatory organs, a deep neural network-based classification method was developed using similar fEMG data previously collected from other individuals and then transformed by dynamic positional warping. In the experiments, the proposed SSR system could classify six different fEMG patterns generated by six silently spoken words with an accuracy of 92.53%. To further demonstrate that our SSR system can be used as a hands-free control interface in practical VR applications, an online SSR system was implemented.
引用
收藏
页码:1047 / 1057
页数:10
相关论文
共 72 条
[1]  
Caserman P(2019)Real-time body tracking in virtual reality using a Vive tracker Virtual Real 23 155-168
[2]  
Garcia-Agundez A(2020)Real-time recognition of facial expressions using facial electromyograms recorded around the eyes for social virtual reality applications IEEE Access 8 62065-62075
[3]  
Konrad R(2014)Enhanced template matching using dynamic positional warping for identification of specific patterns in electroencephalogram J Appl Math 2014 1-7
[4]  
Cha H-S(2009)Dynamic positional warping: dynamic time warping for online handwriting Int J Pattern Recognit Artif Intell 23 967-986
[5]  
Choi S-J(2016)Detection of eye blink artifacts from single prefrontal channel electroencephalogram Comput Methods Programs Biomed 124 19-30
[6]  
Im C-H(2010)Silent speech interfaces Speech Commun 52 270-287
[7]  
Chang W-D(2015)User adaptation in long-term, open-loop myoelectric training: implications for EMG pattern recognition in prosthesis control J Neural Eng 12 046005-300
[8]  
Im C-H(2007)A review of signal subspace speech enhancement and its application to noise robust speech recognition EURASIP J Adv Signal Process 52 288-152
[9]  
Chang W-D(2010)Development of a silent speech interface driven by ultrasound and optical images of the tongue and lips Speech Commun 45 139-2385
[10]  
Shin J(2005)Analysis and recognition of whispered speech Speech Commun 25 2375-50