IoT-driven smart assistive communication system for the hearing impaired with hybrid deep learning models for sign language recognition

被引:1
作者
Maashi, Mashael [1 ]
Iskandar, Huda G. [2 ,3 ]
Rizwanullah, Mohammed [4 ]
机构
[1] King Saud Univ, Coll Comp & Informat Sci, Dept Software Engn, POB 103786, Riyadh 11543, Saudi Arabia
[2] Sanaa Univ, Fac Comp & Informat Technol, Dept Informat Syst, Sanaa, Yemen
[3] King Salman Ctr Disabil Res, Riyadh 11614, Saudi Arabia
[4] Prince Sattam bin Abdulaziz Univ, Dept Comp & Self Dev, Preparatory Year Deanship, AlKharj, Saudi Arabia
关键词
Sign Language Recognition; Hybrid deep learning; Communication systems; Hearing impaired people; MobileNetV3;
D O I
10.1038/s41598-025-89975-1
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Deaf and hard-of-hearing people utilize sign language recognition (SLR) to interconnect. Sign language (SL) is vital for hard-of-hearing and deaf individuals to communicate. SL uses varied hand gestures to speak words, sentences, or letters. It aids in linking the gap of communication between individuals with hearing loss and other persons. Also, it creates comfortable for individuals with hearing loss to convey their feelings. The Internet of Things (IoTs) can help persons with disabilities sustain their desire to attain a good quality of life and permit them to contribute to their economic and social lives. Modern machine learning (ML) and computer vision (CV) developments have allowed SL gesture detection and decipherment. This study presents a Smart Assistive Communication System for the Hearing-Impaired using Sign Language Recognition with Hybrid Deep Learning (SACHI-SLRHDL) methodology in IoT. The SACHI-SLRHDL technique aims to assist people with hearing impairments by creating an intelligent solution. At the primary stage, the SACHI-SLRHDL technique utilizes bilateral filtering (BF) for image pre-processing to increase the excellence of the captured images by reducing noise while preserving edges. Furthermore, the improved MobileNetV3 model is employed for the feature extraction process. Moreover, the convolutional neural network with a bidirectional gated recurrent unit and attention (CNN-BiGRU-A) model classifier is implemented for the SLR process. Finally, the attraction-repulsion optimization algorithm (AROA) adjusts the hyperparameter values of the CNN-BiGRU-A method optimally, resulting in more excellent classification performance. To exhibit the more significant solution of the SACHI-SLRHDL method, a comprehensive experimental analysis is performed under an Indian SL dataset. The experimental validation of the SACHI-SLRHDL method portrayed a superior accuracy value of 99.19% over existing techniques.
引用
收藏
页数:22
相关论文
共 35 条
[31]   Graph data science-driven framework to aid auditory and speech impaired individuals by accelerating sign image analysis and knowledge relegation through deep learning technique [J].
Thejaswi, R. Akhila ;
Rai, Bellipady Shamantha ;
Pakkala, Permanki Guthu Rithesh .
INTERNATIONAL JOURNAL OF SYSTEM ASSURANCE ENGINEERING AND MANAGEMENT, 2025, 16 (01) :175-198
[32]   AI enabled sign language recognition and VR space bidirectional communication using triboelectric smart glove [J].
Wen, Feng ;
Zhang, Zixuan ;
He, Tianyiyi ;
Lee, Chengkuo .
NATURE COMMUNICATIONS, 2021, 12 (01)
[33]   Artificial Intelligence-Enabled Sensing Technologies in the 5G/Internet of Things Era: From Virtual Reality/Augmented Reality to the Digital Twin [J].
Zhang, Zixuan ;
Wen, Feng ;
Sun, Zhongda ;
Guo, Xinge ;
He, Tianyiyi ;
Lee, Chengkuo .
ADVANCED INTELLIGENT SYSTEMS, 2022, 4 (07)
[34]  
Zholshiyeva L, 2024, J ROBOTICS CONTROL, V6, P191, DOI 10.18196/jrc.v6i1.23879
[35]   High-precision monitoring and prediction of mining area surface subsidence using SBAS-InSAR and CNN-BiGRU-attention model [J].
Zhu, Mingfei ;
Yu, Xuexiang ;
Tan, Hao ;
Yuan, Jiajia ;
Chen, Kai ;
Xie, Shicheng ;
Han, Yuchen ;
Long, Wenjiang .
SCIENTIFIC REPORTS, 2024, 14 (01)