Deep learning-based sign language recognition system using both manual and non-manual components fusion

被引:1
作者
Jebali, Maher [1 ]
Dakhli, Abdesselem [1 ]
Bakari, Wided [1 ]
机构
[1] Univ Hail, Comp Sci Dept, POB 2440, Hail 100190, Saudi Arabia
来源
AIMS MATHEMATICS | 2024年 / 9卷 / 01期
关键词
CNN; CTC; recurrent neural network; sign language recognition; head pose;
D O I
10.3934/math.2024105
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Sign language is regularly adopted by speech-impaired or deaf individuals to convey information; however, it necessitates substantial exertion to acquire either complete knowledge or skill. Sign language recognition (SLR) has the intention to close the gap between the users and the non-users of sign language by identifying signs from video speeches. This is a fundamental but arduous task as sign language is carried out with complex and often fast hand gestures and motions, facial expressions and impressionable body postures. Nevertheless, non-manual features are currently being examined since numerous signs have identical manual components but vary in non-manual components. To this end, we suggest a novel manual and non-manual SLR system (MNM-SLR) using a convolutional neural network (CNN) to get the benefits of multi-cue information towards a significant recognition rate. Specifically, we suggest a model for a deep convolutional, long short-term memory network that simultaneously exploits the non-manual features, which is summarized by utilizing the head pose, as well as a model of the embedded dynamics of manual features. Contrary to other frequent works that focused on depth cameras, multiple camera visuals and electrical gloves, we employed the use of RGB, which allows individuals to communicate with a deaf person through their personal devices. As a result, our framework achieves a high recognition rate with an accuracy of 90.12% on the SIGNUM dataset and 94.87% on RWTH-PHOENIX-Weather 2014 dataset.
引用
收藏
页码:2105 / 2122
页数:18
相关论文
共 46 条
  • [11] Research Method of Discontinuous-Gait Image Recognition Based on Human Skeleton Keypoint Extraction
    Han, Kun
    Li, Xinyu
    [J]. SENSORS, 2023, 23 (16)
  • [12] Self-Mutual Distillation Learning for Continuous Sign Language Recognition
    Hao, Aiming
    Min, Yuecong
    Chen, Xilin
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 11283 - 11292
  • [13] Attention-Based 3D-CNNs for Large-Vocabulary Sign Language Recognition
    Huang, Jie
    Zhou, Wengang
    Li, Houqiang
    Li, Weiping
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (09) : 2822 - 2832
  • [14] Huang J, 2018, AAAI CONF ARTIF INTE, P2257
  • [15] Deep Learning-Based Sign Language Recognition System for Cognitive Development
    Jebali, Maher
    Dakhli, Abdesselem
    Bakari, Wided
    [J]. COGNITIVE COMPUTATION, 2023, 15 (06) : 2189 - 2201
  • [16] Vision-based continuous sign language recognition using multimodal sensor fusion
    Jebali, Maher
    Dakhli, Abdesselem
    Jemni, Mohammed
    [J]. EVOLVING SYSTEMS, 2021, 12 (04) : 1031 - 1044
  • [17] Koller O., 2016, BRIT C MACH VIS
  • [18] Deep Sign: Enabling Robust Statistical Continuous Sign Language Recognition via Hybrid CNN-HMMs
    Koller, Oscar
    Zargaran, Sepehr
    Ney, Hermann
    Bowden, Richard
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2018, 126 (12) : 1311 - 1325
  • [19] Re-Sign: Re-Aligned End-to-End Sequence Modelling with Deep Recurrent CNN-HMMs
    Koller, Oscar
    Zargaran, Sepehr
    Ney, Hermann
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3416 - 3424
  • [20] Deep Hand: How to Train a CNN on 1 Million Hand Images When Your Data Is Continuous and Weakly Labelled
    Koller, Oscar
    Ney, Hermann
    Bowden, Richard
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 3793 - 3802