A sensing data and deep learning-based sign language recognition approach

被引:3
作者
Hao, Wei [1 ]
Hou, Chen [2 ,3 ]
Zhang, Zhihao [1 ]
Zhai, Xueyu [1 ]
Wang, Li [1 ]
Lv, Guanghao [4 ]
机构
[1] China Agr Univ, Coll Informat & Elect Engn, Beijing 100083, Peoples R China
[2] Peking Univ, Natl Engn Lab Big Data Anal & Applicat, Beijing 100871, Peoples R China
[3] PKU Changsha Inst Comp & Digital Econ, Changsha 410205, Peoples R China
[4] China Agr Univ, Yantai Res Inst, Yantai 264670, Peoples R China
基金
中国国家自然科学基金;
关键词
Sign language recognition; Sensing data; Skip connection CNN; Multi-head attention mechanism; BiLSTM; GESTURE RECOGNITION;
D O I
10.1016/j.compeleceng.2024.109339
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Correct sign language recognition helps deaf people communicate normally with hearing people. However, existing sign language recognition techniques have low recognition accuracy due to insufficient feature extraction. In this paper, we propose an approach to improve the accuracy of sign language recognition. Firstly, we design a one-dimensional convolutional neural network (CNN) using the skip connection approach Secondly, we propose an improved multi-head attention mechanism that incorporates bi-directional long short-term memory (BiLSTM) networks within a multi-head attention mechanism. Thirdly, we position this improved multi-head attention mechanism behind the final convolutional layer of the proposed CNN architecture and denote the resultant architecture as BMCNN. Finally, we verify the performance of our approach with BMCNN architecture through ten-fold cross-validation on the sensing dataset. Our approach achieves 99.13% accuracy in sign language recognition. The results show that our proposed approach outperforms the traditional machine learning approaches and other state-of-the-art techniques in this field.
引用
收藏
页数:16
相关论文
共 29 条
[1]   Incorporating Relative Position Information in Transformer-Based Sign Language Recognition and Translation [J].
Aloysius, Neena ;
Geetha, M. ;
Nedungadi, Prema .
IEEE ACCESS, 2021, 9 :145929-145942
[2]   A pattern recognition model for static gestures in malaysian sign language based on machine learning techniques [J].
Alrubayi, Ali H. ;
Ahmed, M. A. ;
Zaidan, A. A. ;
Albahri, A. S. ;
Zaidan, B. B. ;
Albahri, O. S. ;
Alamoodi, A. H. ;
Alazab, Mamoun .
COMPUTERS & ELECTRICAL ENGINEERING, 2021, 95
[3]   CROSS-VALIDATION IN REGRESSION AND COVARIANCE STRUCTURE-ANALYSIS - AN OVERVIEW [J].
CAMSTRA, A ;
BOOMSMA, A .
SOCIOLOGICAL METHODS & RESEARCH, 1992, 21 (01) :89-115
[4]  
Cun Y.Le., 1990, ADV NEURAL INFORM PR, V2, P396
[5]   Isolated Sign Recognition from RGB Video using Pose Flow and Self-Attention [J].
De Coster, Mathieu ;
Van Herreweghe, Mieke ;
Dambre, Joni .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, :3436-3445
[6]  
De Coster M, 2020, PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), P6018
[7]   A Wearable Smart Glove and Its Application of Pose and Gesture Detection to Sign Language Classification [J].
DelPreto, Joseph ;
Hughes, Josie ;
D'Aria, Matteo ;
de Fazio, Marco ;
Rus, Daniela .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04) :10589-10596
[8]  
Graves A, 2013, INT CONF ACOUST SPEE, P6645, DOI 10.1109/ICASSP.2013.6638947
[9]   American Sign Language Translation Using Wearable Inertial and Electromyography Sensors for Tracking Hand Movements and Facial Expressions [J].
Gu, Yutong ;
Zheng, Chao ;
Todoh, Masahiro ;
Zha, Fusheng .
FRONTIERS IN NEUROSCIENCE, 2022, 16
[10]   Continuous Sign Language Recognition Based on Spatial-Temporal Graph Attention Network [J].
Guo, Qi ;
Zhang, Shujun ;
Li, Hui .
CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2023, 134 (03) :1653-1670