Audio-Visual Speech and Gesture Recognition by Sensors of Mobile Devices

被引:61
作者
Ryumin, Dmitry [1 ]
Ivanko, Denis [1 ]
Ryumina, Elena [1 ]
机构
[1] Russian Acad Sci SPC RAS, St Petersburg Fed Res Ctr, St Petersburg 199178, Russia
基金
俄罗斯科学基金会;
关键词
audio-visual speech recognition; model-level fusion; lip-reading; gesture recognition; spatio-temporal features; dimensionality reduction technique; computer vision; HUMAN-COMPUTER INTERACTION; VISUAL ANALYSIS; HAND GESTURES; NETWORKS; CORPUS;
D O I
10.3390/s23042284
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Audio-visual speech recognition (AVSR) is one of the most promising solutions for reliable speech recognition, particularly when audio is corrupted by noise. Additional visual information can be used for both automatic lip-reading and gesture recognition. Hand gestures are a form of non-verbal communication and can be used as a very important part of modern human-computer interaction systems. Currently, audio and video modalities are easily accessible by sensors of mobile devices. However, there is no out-of-the-box solution for automatic audio-visual speech and gesture recognition. This study introduces two deep neural network-based model architectures: one for AVSR and one for gesture recognition. The main novelty regarding audio-visual speech recognition lies in fine-tuning strategies for both visual and acoustic features and in the proposed end-to-end model, which considers three modality fusion approaches: prediction-level, feature-level, and model-level. The main novelty in gesture recognition lies in a unique set of spatio-temporal features, including those that consider lip articulation information. As there are no available datasets for the combined task, we evaluated our methods on two different large-scale corpora-LRW and AUTSL-and outperformed existing methods on both audio-visual speech recognition and gesture recognition tasks. We achieved AVSR accuracy for the LRW dataset equal to 98.76% and gesture recognition rate for the AUTSL dataset equal to 98.56%. The results obtained demonstrate not only the high performance of the proposed methodology, but also the fundamental possibility of recognizing audio-visual speech and gestures by sensors of mobile devices.
引用
收藏
页数:29
相关论文
共 146 条
[1]   Improving the Performance of Unimodal Dynamic Hand-Gesture Recognition with Multimodal Training [J].
Abavisani, Mahdi ;
Joze, Hamid Reza Vaezi ;
Patel, Vishal M. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1165-1174
[2]   Principal component analysis [J].
Abdi, Herve ;
Williams, Lynne J. .
WILEY INTERDISCIPLINARY REVIEWS-COMPUTATIONAL STATISTICS, 2010, 2 (04) :433-459
[3]   Deep Audio-Visual Speech Recognition [J].
Afouras, Triantafyllos ;
Chung, Joon Son ;
Senior, Andrew ;
Vinyals, Oriol ;
Zisserman, Andrew .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (12) :8717-8727
[4]  
Afouras Triantafyllos, 2018, arXiv
[5]   Sign Language Recognition Method Based on Palm Definition Model and Multiple Classification [J].
Amangeldy, Nurzada ;
Kudubayeva, Saule ;
Kassymova, Akmaral ;
Karipzhanova, Ardak ;
Razakhova, Bibigul ;
Kuralov, Serikbay .
SENSORS, 2022, 22 (17)
[6]  
[Anonymous], 2015, P 14 PYTH SCI C, DOI 10.25080/Majora-7b98e3ed-003
[7]   Audio-visual speech asynchrony detection using co-inertia analysis and coupled hidden markov models [J].
Argones Rua, Enrique ;
Bredin, Herve ;
Garcia Mateo, Carmen ;
Chollet, Gerard ;
Gonzalez Jimenez, Daniel .
PATTERN ANALYSIS AND APPLICATIONS, 2009, 12 (03) :271-284
[8]  
Assael Y. M., 2016, arXiv
[9]  
Axyonov A., 2021, Int. Arch. Photogramm., Remote Sens. Spatial Inf. Sci., VXLIV-2/W1-2021, P7, DOI 10.5194/isprs-archives-XLIV-2-W1-2021-7-2021
[10]   Method for visual analysis of driver's face for automatic lip-reading in the wild [J].
Axyonov, A. A. ;
Ryumin, D. A. ;
Kashevnik, A. M. ;
Ivanko, D., V ;
Karpov, A. A. .
COMPUTER OPTICS, 2022, 46 (06) :955-+