Prediction of sign language recognition based on multi layered CNN

被引:3
作者
Prasath, G. Arun [1 ]
Annapurani, K. [2 ]
机构
[1] SRM Inst Sci & Technol, Sch Comp, Dept Networking & Commun, Chengalpattu 603203, Tamilnadu, India
[2] SRM Inst Sci & Technol Kattankulathur, Sch Comp, Dept Networking & Commun, Chengalpattu 603203, Tamilnadu, India
关键词
Sign language recognition; Deep learning; Multi-layer convolutional neural network; Linear and non-linear features; Higher level and lower-level features; Precision; Accuracy; GESTURE RECOGNITION; SPEECH;
D O I
10.1007/s11042-023-14548-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sign Language Recognition (SLR) helps to bridge the gap between ordinary and hearing-impaired people. But various difficulties and challenges are faced by SLR system during real-time implementation. The major complexity associated with SLR is the inability to provide a consistent recognition process and it shows lesser recognition accuracy. To handle this issue, this research concentrates on adopting the finest classification approach to provide a feasible end-to-end system using deep learning approaches. This process transforms sign language into the voice for assisting the people to hear the sign language. The input is taken from the ROBITA Indian Sign Language Gesture Database and some essential pre-processing steps are done to avoid unnecessary artefacts. The proposed model is incorporated with the encoder Multi-Layer Convolutional Neural Networks (ML-CNN) for evaluating the scalability, accuracy of the end-to-end SLR. The encoder analyses the linear and non-linear features (higher level and lower level) to improve the quality of recognition. The simulation is carried out in a MATLAB environment where the performance of the ML-CNN model outperforms the existing approaches and establishes the trade-off. Some performance metrics like accuracy, precision, F-measure, recall, Matthews Correlation Coefficient (MCC), Mean Absolute Error (MAE) are evaluated to show the significance of the model. The prediction accuracy of the proposed ML-CNN with encoder is 87.5% in the ROBITA sign gesture dataset and it's increased by 1% and 3.5% over the BLSTM and HMM respectively.
引用
收藏
页码:29649 / 29669
页数:21
相关论文
共 35 条
  • [1] Dynamic Sign Language Recognition for Smart Home Interactive Application Using Stochastic Linear Formal Grammar
    Abid, Muhammad Rizwan
    Petriu, Emil M.
    Amjadian, Ehsan
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2015, 64 (03) : 596 - 605
  • [2] Hand Gesture Recognition Using 3D-CNN Model
    Al-Hammadi, Muneer
    Muhammad, Ghulam
    Abdul, Wadood
    Alsulaiman, Mansour
    Hossain, M. Shamim
    [J]. IEEE CONSUMER ELECTRONICS MAGAZINE, 2020, 9 (01) : 95 - 101
  • [3] An Automatic Health Monitoring System for Patients Suffering From Voice Complications in Smart Cities
    Ali, Zulfiqar
    Muhammad, Ghulam
    Alhamid, Mohammed F.
    [J]. IEEE ACCESS, 2017, 5 : 3900 - 3908
  • [4] Assaleh K., 2010, Journal of Intelligent Learning Systems and Applications, V2, P19, DOI 10.4236/jilsa.2010.21003
  • [5] SubUNets: End-to-end Hand Shape and Continuous Sign Language Recognition
    Camgoz, Necati Cihan
    Hadfield, Simon
    Koller, Oscar
    Bowden, Richard
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 3075 - 3084
  • [6] Recurrent Convolutional Neural Networks for Continuous Sign Language Recognition by Staged Optimization
    Cui, Runpeng
    Liu, Hu
    Zhang, Changshui
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1610 - 1618
  • [7] Learning Spatiotemporal Features with 3D Convolutional Networks
    Du Tran
    Bourdev, Lubomir
    Fergus, Rob
    Torresani, Lorenzo
    Paluri, Manohar
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 4489 - 4497
  • [8] Autoencoder node saliency: Selecting relevant latent representations
    Fan, Ya Ju
    [J]. PATTERN RECOGNITION, 2019, 88 : 643 - 653
  • [9] Graves A., 2006, P 23 INT C MACH LEAR, P369
  • [10] Guo D, 2019, PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P744