Gesture recognition and response system for special education using computer vision and human-computer interaction technology

被引:0
作者
Duan, Xuanfeng [1 ,2 ]
机构
[1] Leshan Normal Univ, Sichuan Prov Key Lab Philosophy & Social Sci Langu, Leshan 617000, Sichuan, Peoples R China
[2] Jose Rizal Univ, Grad Sch, Mandaluyong, Manila Province, Philippines
关键词
Gesture recognition; special education; deep learning; machine learning; genetic algorithms; model compression; AlexNet; VGG-19; ResNet; MobileNet; real-time systems; assistive technology;
D O I
10.1080/17483107.2025.2527226
中图分类号
R49 [康复医学];
学科分类号
100215 ;
摘要
Gesture recognition has emerged as a pivotal technology for enhancing human-computer interaction (HCI), especially in the context of special education. This study presents a comprehensive gesture recognition and response system that leverages advanced deep learning architectures, including AlexNet, VGG19, ResNet and MobileNet, combined with machine learning algorithms such as support vector machines (SVM) and random forest. The proposed system achieves state-of-the-art performance, with an accuracy of 95.4%, demonstrating its effectiveness in recognising complex gestures with high precision. To address the challenges of deploying gesture recognition systems on resource-constrained devices, the study incorporates genetic algorithms (GAs) for model compression. This optimisation reduces the model size by 42%, significantly enhancing its suitability for real-time applications on mobile and embedded platforms. Additionally, inference time is reduced by 45%, enabling faster response times essential for interactive educational environments. The system was evaluated using a diverse gesture dataset, ensuring robustness across varying lighting conditions, user demographics, and physical differences. The findings highlight the potential of integrating gesture recognition systems into special education, where they can serve as assistive tools for individuals with disabilities, fostering inclusive and engaging learning experiences. This work not only advances the field of gesture recognition but also underscores the importance of model optimisation for real-world applications. Future research will focus on expanding the gesture library, integrating multimodal inputs such as speech, and enhancing system adaptability through continuous learning mechanisms.
引用
收藏
页数:18
相关论文
共 48 条
[1]  
Ahmed Z., 2020, Int J Hum-Comput Interact, V36, P1157, DOI [10.1080/10447318.2020.1754700, DOI 10.1080/10447318.2020.1754700]
[2]   Multimodal fusion for multimedia analysis: a survey [J].
Atrey, Pradeep K. ;
Hossain, M. Anwar ;
El Saddik, Abdulmotaleb ;
Kankanhalli, Mohan S. .
MULTIMEDIA SYSTEMS, 2010, 16 (06) :345-379
[3]  
Bailly G., 2010, MULTIMODAL INTERACTI, P15, DOI [10.1145/1753326.1753331, DOI 10.1145/1753326.1753331]
[4]  
Buehler P., 2009, IEEE Trans Pattern Anal Mach Intell, DOI [10.1109/TPAMI.2009.48, DOI 10.1109/TPAMI.2009.48]
[5]  
Chan KT., 2021, Health Inform J, V27, P156, DOI [10.1177/1460458220971354, DOI 10.1177/1460458220971354]
[6]  
Cheng J., 2020, IEEE Trans Mobile Comput, DOI [10.1109/TMC.2020.3014072, DOI 10.1109/TMC.2020.3014072]
[7]  
Darrell T., 1996, IEEE Robot Autom Mag, V3, P30, DOI [10.1109/100.536497, DOI 10.1109/100.536497]
[8]  
Dawes M., 2021, Assistive Technol, V33, P215, DOI [10.1080/10400435.2020.1733245, DOI 10.1080/10400435.2020.1733245]
[9]   MyPGI-a methodology to yield personalized gestural interaction [J].
de Oliveira Schultz Ascari, Rubia Eliza ;
Silva, Luciano ;
Pereira, Roberto .
UNIVERSAL ACCESS IN THE INFORMATION SOCIETY, 2024, 23 (02) :795-820
[10]  
Du Y, 2015, PROC CVPR IEEE, P1110, DOI 10.1109/CVPR.2015.7298714