mXception and dynamic image for hand gesture recognition

被引:0
作者
Bhumika Karsh
Rabul Hussain Laskar
Ram Kumar Karsh
机构
[1] National Institute of Technology,Speech and Image Processing Laboratory, Electronics and Communication Engineering Department
来源
Neural Computing and Applications | 2024年 / 36卷
关键词
UNET; Xception; Dynamic image; Gesture classification; HCI;
D O I
暂无
中图分类号
学科分类号
摘要
Gesture detection has recently attracted a lot of attention due to its wide range of applications, notably in human–computer interaction (HCI). However, when it comes to video-based gesture recognition, elements in the background unrelated to gestures slow down the system’s classification rate. This paper presents an algorithm designed for the recognition of large-scale gestures. In the training phase, we utilize RGB-D videos, where the depth modality videos are derived from RGB modality videos using UNET and subsequently employed for testing. However, it’s worth noting that in real-time applications of the proposed dynamic hand gesture recognition (DHGR) system, only RGB modality videos are needed. The algorithm begins by creating two dynamic images: one from the estimated depth video and the other from the RGB video. Dynamic images generated from RGB video excel in capturing spatial information; while, those derived from depth video excel in encoding temporal aspects. These two dynamic images are merged to form an RGB-D dynamic image (RDDI). The RDDI is then fed into a modified Xception-based CNN model for the purpose of gesture classification and recognition. In order to evaluate the system’s performance, we conducted experiments using the EgoGesture and MSR Gesture datasets. The results are highly promising, with a reported classification accuracy of 91.64% for the EgoGesture dataset and an impressive 99.41% for the MSR Gesture dataset. The results demonstrated that the suggested system outperformed some existing techniques.
引用
收藏
页码:8281 / 8300
页数:19
相关论文
共 131 条
[1]  
Mitra S(2007)Gesture recognition: a survey IEEE Trans Syst Man Cybern Part C (Appl Rev) 37 311-324
[2]  
Acharya T(2014)RETRACTED ARTICLE: human–computer interaction using vision-based hand gesture recognition systems: a survey Neural Comput Appl 25 251-261
[3]  
Hasan H(2006)New approach for static gesture recognition J Inf Sci Eng 22 1047-1057
[4]  
Abdul-Kareem S(2020)Online dynamic hand gesture recognition including efficiency analysis IEEE Trans Biom Behav Identity Sci 2 85-97
[5]  
Chang CC(1997)Visual interpretation of hand gestures for human–computer interaction: a review IEEE Trans Pattern Anal Mach Intell 19 677-695
[6]  
Chen JJ(2021)CNN based feature extraction and classification for sign language Multimed Tools Appl 80 3051-3069
[7]  
Tai WK(2018)RGB-D-based human motion recognition with deep learning: a survey Comput Vis Image Underst 171 118-139
[8]  
Han CC(2018)A systematic literature review on vision based gesture recognition techniques Multimed Tools Appl 77 28121-28184
[9]  
Köpüklü O(2012)3D convolutional neural networks for human action recognition IEEE Trans Pattern Anal Mach Intell 35 221-231
[10]  
Gunduz A(2016)3D-based deep convolutional neural network for action recognition with depth sequences Image Vis Comput 55 93-100