Deep learning-based sign language recognition system using both manual and non-manual components fusion

被引:3
作者
Jebali, Maher [1 ]
Dakhli, Abdesselem [1 ]
Bakari, Wided [1 ]
机构
[1] Univ Hail, Comp Sci Dept, POB 2440, Hail 100190, Saudi Arabia
来源
AIMS MATHEMATICS | 2024年 / 9卷 / 01期
关键词
CNN; CTC; recurrent neural network; sign language recognition; head pose;
D O I
10.3934/math.2024105
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Sign language is regularly adopted by speech-impaired or deaf individuals to convey information; however, it necessitates substantial exertion to acquire either complete knowledge or skill. Sign language recognition (SLR) has the intention to close the gap between the users and the non-users of sign language by identifying signs from video speeches. This is a fundamental but arduous task as sign language is carried out with complex and often fast hand gestures and motions, facial expressions and impressionable body postures. Nevertheless, non-manual features are currently being examined since numerous signs have identical manual components but vary in non-manual components. To this end, we suggest a novel manual and non-manual SLR system (MNM-SLR) using a convolutional neural network (CNN) to get the benefits of multi-cue information towards a significant recognition rate. Specifically, we suggest a model for a deep convolutional, long short-term memory network that simultaneously exploits the non-manual features, which is summarized by utilizing the head pose, as well as a model of the embedded dynamics of manual features. Contrary to other frequent works that focused on depth cameras, multiple camera visuals and electrical gloves, we employed the use of RGB, which allows individuals to communicate with a deaf person through their personal devices. As a result, our framework achieves a high recognition rate with an accuracy of 90.12% on the SIGNUM dataset and 94.87% on RWTH-PHOENIX-Weather 2014 dataset.
引用
收藏
页码:2105 / 2122
页数:18
相关论文
共 46 条
[11]   Interactive attention and improved GCN for continuous sign language recognition [J].
Guo, Qi ;
Zhang, Shujun ;
Tan, Liwei ;
Fang, Ke ;
Du, Yinghao .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 85
[12]   Research Method of Discontinuous-Gait Image Recognition Based on Human Skeleton Keypoint Extraction [J].
Han, Kun ;
Li, Xinyu .
SENSORS, 2023, 23 (16)
[13]   Self-Mutual Distillation Learning for Continuous Sign Language Recognition [J].
Hao, Aiming ;
Min, Yuecong ;
Chen, Xilin .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :11283-11292
[14]   Attention-Based 3D-CNNs for Large-Vocabulary Sign Language Recognition [J].
Huang, Jie ;
Zhou, Wengang ;
Li, Houqiang ;
Li, Weiping .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (09) :2822-2832
[15]  
Huang J, 2018, AAAI CONF ARTIF INTE, P2257
[16]   Deep Learning-Based Sign Language Recognition System for Cognitive Development [J].
Jebali, Maher ;
Dakhli, Abdesselem ;
Bakari, Wided .
COGNITIVE COMPUTATION, 2023, 15 (06) :2189-2201
[17]   Vision-based continuous sign language recognition using multimodal sensor fusion [J].
Jebali, Maher ;
Dakhli, Abdesselem ;
Jemni, Mohammed .
EVOLVING SYSTEMS, 2021, 12 (04) :1031-1044
[18]  
Koller O., 2016, BRIT C MACH VIS
[19]   Deep Sign: Enabling Robust Statistical Continuous Sign Language Recognition via Hybrid CNN-HMMs [J].
Koller, Oscar ;
Zargaran, Sepehr ;
Ney, Hermann ;
Bowden, Richard .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2018, 126 (12) :1311-1325
[20]   Re-Sign: Re-Aligned End-to-End Sequence Modelling with Deep Recurrent CNN-HMMs [J].
Koller, Oscar ;
Zargaran, Sepehr ;
Ney, Hermann .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3416-3424