Deep Learning-Based Hand Gesture Recognition System and Design of a Human–Machine Interface

被引:0
作者
Abir Sen
Tapas Kumar Mishra
Ratnakar Dash
机构
[1] National Institute of Technology,Department of Computer Science and Engineering
来源
Neural Processing Letters | 2023年 / 55卷
关键词
Deep learning; Hand gesture recognition; Segmentation; Vision transformer; Kalman filter; Human machine interface; Transfer learning; Virtual mouse;
D O I
暂无
中图分类号
学科分类号
摘要
Hand gesture recognition plays an important role in developing effective human–machine interfaces (HMIs) that enable direct communication between humans and machines. But in real-time scenarios, it is difficult to identify the correct hand gesture to control an application while moving the hands. To address this issue, in this work, a low-cost hand gesture recognition system based human-computer interface (HCI) is presented in real-time scenarios. The system consists of six stages: (1) hand detection, (2) gesture segmentation, (3) feature extraction and gesture classification using five pre-trained convolutional neural network models (CNN) and vision transformer (ViT), (4) building an interactive human–machine interface (HMI), (5) development of a gesture-controlled virtual mouse, (6) smoothing of virtual mouse pointer using of Kalman filter. In our work, five pre-trained CNN models (VGG16, VGG19, ResNet50, ResNet101, and Inception-V1) and ViT have been employed to classify hand gesture images. Two multi-class datasets (one public and one custom) have been used to validate the models. Considering the model’s performances, it is observed that Inception-V1 has significantly shown a better classification performance compared to the other four CNN models and ViT in terms of accuracy, precision, recall, and F-score values. We have also expanded this system to control some multimedia applications (such as VLC player, audio player, playing 2D Super-Mario-Bros game, etc.) with different customized gesture commands in real-time scenarios. The average speed of this system has reached 25 fps (frames per second), which meets the requirements for the real-time scenario. Performance of the proposed gesture control system obtained the average response time in milisecond for each control which makes it suitable for real-time. This model (prototype) will benefit physically disabled people interacting with desktops.
引用
收藏
页码:12569 / 12596
页数:27
相关论文
共 49 条
[1]  
Al Farid F(2022)A structured and methodological review on vision-based hand gesture recognition system J Imaging 8 153-6042
[2]  
Hashim N(2011)Gabor filter-based hand-pose angle estimation for hand gesture recognition under varying illumination Expert Syst Appl 38 6031-1141
[3]  
Abdullah J(2018)Dynamic hand gesture recognition using vision-based approach for human-computer interaction Neural Comput Appl 29 1129-3951
[4]  
Bhuiyan MR(2017)Deep learning in vision-based static hand gesture recognition Neural Comput Appl 28 3941-28237
[5]  
Shahida Mohd Isa WN(2019)Gesture recognition based on CNN and DCGAN for calculation and text output IEEE Access 7 28230-2361
[6]  
Uddin J(2020)A deep convolutional neural network approach for static hand gesture recognition Proc Comput Sci 171 2353-15248
[7]  
Haque MA(2020)An efficient method for human hand gesture detection and recognition using deep learning convolutional neural networks Soft Comput 24 15239-40066
[8]  
Husen MN(2022)A novel hand gesture detection and recognition system based on ensemble-based convolutional neural network Multimed Tools Appl 81 40043-10207
[9]  
Huang D-Y(2022)Electromyography-based, robust hand motion classification employing temporal multi-channel vision transformers IEEE Robot Autom Lett 7 10200-6007
[10]  
Hu W-C(2020)Design of hand gesture recognition system for human-computer interaction Multimed Tools Appl 79 5989-undefined