Accurate and Real-Time Variant Hand Pose Estimation Based on Gray Code Bounding Box Representation

被引:0
作者
Wang, Yangang [1 ]
Sun, Wenqian [1 ]
Rao, Ruting [1 ]
机构
[1] Southeast Univ, Sch Automat, Key Lab Measurement & Control Complex Syst Engn, Minist Educ, Nanjing 210096, Peoples R China
基金
中国国家自然科学基金;
关键词
Pose estimation; Training; Annotations; Three-dimensional displays; Sensors; Color; Reflective binary codes; Bounding box representation; gray code; hand pose estimation; real-time;
D O I
10.1109/JSEN.2024.3389055
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Effective hand gestures are crucial for human-machine interactions, and recent advancements in neural networks offer promising avenues for efficient hand pose estimation. However, existing methods still face challenges in detecting hand poses of different scales within a single RGB image sensor. This article introduces a novel approach, drawing inspiration from modulus conversion, to enhance the efficiency of hand pose estimation from a single RGB image sensor. The method involves transforming the floating-point values of hand regions into binary codes, ensuring continuous numerical space without a significant computational overhead. This approach significantly improves accuracy for hands of varying sizes in both detection and pose estimation. Furthermore, this article addresses the challenge of datasets lacking hand keypoints annotations by introducing a novel loss computation for labeled keypoints during network training. To assess the effectiveness of the proposed method, a new benchmark for variant hand scales is presented, facilitating evaluation across different hand sizes. The proposed approach undergoes testing on diverse datasets, with experimental results demonstrating comparable performance to state-of-the-art methods, thereby validating its efficacy. Additionally, the study conducts several ablation studies, exploring aspects such as the choice of Gray code, code length, effectiveness across different hand scales, and training with labeled keypoints to affirm the efficiency and effectiveness of the proposed method.
引用
收藏
页码:18043 / 18053
页数:11
相关论文
共 49 条
  • [1] Brahmbhatt S., 2020, European Conference on Computer Vision, P361
  • [2] Contributors M., 2023, Openmmlab pose estimation toolbox andBenchmark
  • [3] domedb, 2018, CMU Panoptic Studio Hand Dataset
  • [4] Vision-based hand pose estimation: A review
    Erol, Ali
    Bebis, George
    Nicolescu, Mircea
    Boyle, Richard D.
    Twombly, Xander
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2007, 108 (1-2) : 52 - 73
  • [5] ARCTIC: A Dataset for Dexterous Bimanual Hand-Object Manipulation
    Fan, Zicong
    Taheri, Omid
    Tzionas, Dimitrios
    Kocabas, Muhammed
    Kaufmann, Manuel
    Black, Michael J.
    Hilliges, Otmar
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 12943 - 12954
  • [6] Dynamic Hand Gesture Recognition Based on 3D Hand Pose Estimation for Human-Robot Interaction
    Gao, Qing
    Chen, Yongquan
    Ju, Zhaojie
    Liang, Yi
    [J]. IEEE SENSORS JOURNAL, 2022, 22 (18) : 17421 - 17430
  • [7] First-Person Hand Action Benchmark with RGB-D Videos and 3D Hand Pose Annotations
    Garcia-Hernando, Guillermo
    Yuan, Shanxin
    Baek, Seungryul
    Kim, Tae-Kyun
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 409 - 419
  • [8] 3D Hand Shape and Pose Estimation from a Single RGB Image
    Ge, Liuhao
    Ren, Zhou
    Li, Yuncheng
    Xue, Zehao
    Wang, Yingying
    Cai, Jianfei
    Yuan, Junsong
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 10825 - 10834
  • [9] Robust 3D Hand Pose Estimation in Single Depth Images: from Single-View CNN to Multi-View CNNs
    Ge, Liuhao
    Liang, Hui
    Yuan, Junsong
    Thalmann, Daniel
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 3593 - 3601
  • [10] Google, 2023, On-Device, Real-Time Hand Tracking With Mediapipe