Seeking a Hierarchical Prototype for Multimodal Gesture Recognition

被引:0
作者
Li, Yunan [1 ,2 ]
Qi, Tianyu [3 ]
Ma, Zhuoqi [3 ]
Quan, Dou [4 ]
Miao, Qiguang [1 ,2 ]
机构
[1] Xidian Univ, Key Lab Big Data & Intelligent Vis, Key Lab Smart Human Comp Interact & Wearable Tech, Xian, Peoples R China
[2] Xidian Univ, Minist Educ, Key Lab Collaborat Intelligence Syst, Xian 710071, Peoples R China
[3] Xidian Univ, Sch Comp Sci & Technol, Xian 710071, Peoples R China
[4] Xidian Univ, Sch Artificial Intelligence, Xian 710071, Peoples R China
基金
中国国家自然科学基金;
关键词
Generative adversarial network (GAN); gesture prototype; gesture recognition; memory bank; multimodal; NETWORKS; DATASET; FUSION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Gesture recognition has drawn considerable attention from many researchers owing to its wide range of applications. Although significant progress has been made in this field, previous works always focus on how to distinguish between different gesture classes, ignoring the influence of inner-class divergence caused by gesture-irrelevant factors. Meanwhile, for multimodal gesture recognition, feature or score fusion in the final stage is a general choice to combine the information of different modalities. Consequently, the gesture-relevant features in different modalities may be redundant, whereas the complementarity of modalities is not exploited sufficiently. To handle these problems, we propose a hierarchical gesture prototype framework to highlight gesture-relevant features such as poses and motions in this article. This framework consists of a sample-level prototype and a modal-level prototype. The sample-level gesture prototype is established with the structure of a memory bank, which avoids the distraction of gesture-irrelevant factors in each sample, such as the illumination, background, and the performers' appearances. Then the modal-level prototype is obtained via a generative adversarial network (GAN)-based subnetwork, in which the modal-invariant features are extracted and pulled together. Meanwhile, the modal-specific attribute features are used to synthesize the feature of other modalities, and the circulation of modality information helps to leverage their complementarity. Extensive experiments on three widely used gesture datasets demonstrate that our method is effective to highlight gesture-relevant features and can outperform the state-of-the-art methods.
引用
收藏
页码:198 / 209
页数:12
相关论文
共 50 条
  • [41] Learning Fast and Robust Gesture Recognition
    Papaioannidis, Christos
    Makrygiannis, Dimitrios
    Mademlis, Ioannis
    Pitas, Ioannis
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 761 - 765
  • [42] FabricTouch: A Multimodal Fabric Assessment Touch Gesture Dataset to Slow Down Fast Fashion
    Olugbade, Temitayo
    Lin, Lili
    Sansoni, Alice
    Warawita, Nihara
    Gan, Yuanze
    Wei, Xijia
    Petreca, Bruna
    Boccignone, Giuseppe
    Atkinson, Douglas
    Cho, Youngjun
    Baurley, Sharon
    Bianchi-Berthouze, Nadia
    2023 11TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION, ACII, 2023,
  • [43] Appearance-based Human Gesture Recognition using Multimodal Features for Human Computer Interaction
    Luo, Dan
    Gao, Hua
    Ekenel, Hazim Kemal
    Ohya, Jun
    HUMAN VISION AND ELECTRONIC IMAGING XVI, 2011, 7865
  • [44] Improving Hand Gesture Recognition Robustness to Dynamic Posture Variations by Multimodal Deep Feature Fusion
    Li, Jiwei
    Zhang, Bi
    Chen, Wanxin
    Bu, Chunguang
    Zhao, Yiwen
    Zhao, Xingang
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2024, 32 : 3104 - 3115
  • [45] Gesture Recognition and Multi-modal Fusion on a New Hand Gesture Dataset
    Schak, Monika
    Gepperth, Alexander
    PATTERN RECOGNITION APPLICATIONS AND METHODS, ICPRAM 2021, ICPRAM 2022, 2023, 13822 : 76 - 97
  • [46] Continuous gesture recognition by using gesture spotting
    Lee, Daeha
    Yoon, Hosub
    Kim, Jaehong
    2016 16TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2016, : 1496 - 1498
  • [47] Gesture Feature Extraction for Static Gesture Recognition
    Haitham Sabah Hasan
    Sameem Binti Abdul Kareem
    Arabian Journal for Science and Engineering, 2013, 38 : 3349 - 3366
  • [48] Automatic Construction of Gesture Network for Gesture Recognition
    Mori, Akihiro
    Uchida, Seiichi
    Kurazume, Ryo
    Taniguchi, Rin-ichiro
    Hasegawa, Tsutomu
    TENCON 2010: 2010 IEEE REGION 10 CONFERENCE, 2010, : 923 - 928
  • [49] Gesture Feature Extraction for Static Gesture Recognition
    Hasan, Haitham Sabah
    Kareem, Sameem Binti Abdul
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2013, 38 (12) : 3349 - 3366
  • [50] The Gesture Recognition Toolkit
    Gillian, Nicholas
    Paradiso, Joseph A.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2014, 15 : 3483 - 3487