Deep Learning for Hand Gesture Recognition in Virtual Museum Using Wearable Vision Sensors

被引:3
作者
Zerrouki, Nabil [1 ]
Harrou, Fouzi [2 ]
Houacine, Amrane [3 ]
Bouarroudj, Riadh [3 ]
Cherifi, Mohammed Yazid [3 ]
Zouina, Ait-Djafer Amina [1 ]
Sun, Ying [2 ]
机构
[1] Ctr Dev Adv Technol, Algiers 16000, Algeria
[2] King Abdullah Univ Sci & Technol KAUST, Comp Elect & Math Sci & Engn CEMSE Div, Thuwal 23955, Saudi Arabia
[3] Univ Sci & Technol Houari Boumedienne, Fac Elect & Comp Sci, Algiers 16000, Algeria
关键词
Bidirectional long short-term memory (Bi-LSTM) classification; ego-centric vision devices; feature extraction; hand gesture recognition; wearable vision; CLASSIFICATION;
D O I
10.1109/JSEN.2024.3354784
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Hand gestures facilitate user interaction and immersion in virtual museum applications. These gestures allow users to navigate virtual exhibitions, interact with virtual artifacts, and control virtual environments naturally and intuitively. This study introduces a deep learning-driven approach for hand gesture recognition using wearable vision sensors designed for interactive virtual museum environments. The proposed approach employs an image-based feature extraction strategy that focuses on capturing five partial occupancy areas of the hand. Notably, a deep learning strategy using the bidirectional long short-term memory (Bi-LSTM) model is adopted to construct an effective model for hand gesture identification. The bidirectionality of Bi-LSTM enables it to capture dependencies in both forward and backward directions, providing a more comprehensive understanding of temporal relationships in the data. The bidirectional nature allows the model to better capture the dynamics and complexities of hand motions, leading to improved accuracy and robustness. The performance evaluation includes experiments on publicly available datasets, considering virtual and real museum scenarios. The results highlight the Bi-LSTM-based approach's superiority by accurately distinguishing various hand gestures. The experimental findings demonstrate that combining the five area ratios and Bi-LSTM classification enables robust recognition of diverse hand gestures and effectively discriminates between similar actions, such as slide left and right classes. Additionally, it shows promising detection performance compared to conventional machine learning models and state-of-the-art (SOTA) methods. The presented approach is promising for enhancing user interaction and immersion in virtual museum experiences.
引用
收藏
页码:8857 / 8869
页数:13
相关论文
共 38 条
  • [1] Confimizer: A Novel Algorithm to Optimize Cloud Resource by Confidentiality-Cost Trade-Off Using BiLSTM Network
    Achar, Sandesh
    Faruqui, Nuruzzaman
    Bodepudi, Anusha
    Reddy, Manjunath
    [J]. IEEE ACCESS, 2023, 11 : 89205 - 89217
  • [2] [Anonymous], 2012, Technical Report MSR-TR-2012-68
  • [3] [Anonymous], 2006, Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition
  • [4] [Anonymous], 2014, Economic Impact of Travel & Tourism
  • [5] [Anonymous], How the Americans Will Travel 2015
  • [6] Analysis of the Hands in Egocentric Vision: A Survey
    Bandini, Andrea
    Zariffa, Jose
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (06) : 6846 - 6866
  • [7] Gesture Recognition Using Wearable Vision Sensors to Enhance Visitors' Museum Experiences
    Baraldi, Lorenzo
    Paci, Francesco
    Serra, Giuseppe
    Benini, Luca
    Cucchiara, Rita
    [J]. IEEE SENSORS JOURNAL, 2015, 15 (05) : 2705 - 2714
  • [8] Is that my hand? An egocentric dataset for hand disambiguation
    Cruz, Sergio
    Chan, Antoni
    [J]. IMAGE AND VISION COMPUTING, 2019, 89 : 131 - 143
  • [9] A decision-theoretic generalization of on-line learning and an application to boosting
    Freund, Y
    Schapire, RE
    [J]. JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 1997, 55 (01) : 119 - 139
  • [10] Gers FA, 1999, IEE CONF PUBL, P850, DOI [10.1162/089976600300015015, 10.1049/cp:19991218]