Accurate and Real-Time Variant Hand Pose Estimation Based on Gray Code Bounding Box Representation

被引:0
作者
Wang, Yangang [1 ]
Sun, Wenqian [1 ]
Rao, Ruting [1 ]
机构
[1] Southeast Univ, Sch Automat, Key Lab Measurement & Control Complex Syst Engn, Minist Educ, Nanjing 210096, Peoples R China
基金
中国国家自然科学基金;
关键词
Pose estimation; Training; Annotations; Three-dimensional displays; Sensors; Color; Reflective binary codes; Bounding box representation; gray code; hand pose estimation; real-time;
D O I
10.1109/JSEN.2024.3389055
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Effective hand gestures are crucial for human-machine interactions, and recent advancements in neural networks offer promising avenues for efficient hand pose estimation. However, existing methods still face challenges in detecting hand poses of different scales within a single RGB image sensor. This article introduces a novel approach, drawing inspiration from modulus conversion, to enhance the efficiency of hand pose estimation from a single RGB image sensor. The method involves transforming the floating-point values of hand regions into binary codes, ensuring continuous numerical space without a significant computational overhead. This approach significantly improves accuracy for hands of varying sizes in both detection and pose estimation. Furthermore, this article addresses the challenge of datasets lacking hand keypoints annotations by introducing a novel loss computation for labeled keypoints during network training. To assess the effectiveness of the proposed method, a new benchmark for variant hand scales is presented, facilitating evaluation across different hand sizes. The proposed approach undergoes testing on diverse datasets, with experimental results demonstrating comparable performance to state-of-the-art methods, thereby validating its efficacy. Additionally, the study conducts several ablation studies, exploring aspects such as the choice of Gray code, code length, effectiveness across different hand scales, and training with labeled keypoints to affirm the efficiency and effectiveness of the proposed method.
引用
收藏
页码:18043 / 18053
页数:11
相关论文
共 49 条
[21]  
Long J, 2015, PROC CVPR IEEE, P3431, DOI 10.1109/CVPR.2015.7298965
[22]  
Moon G., 2020, COMPUTER VISION ECCV, P548
[23]   GANerated Hands for Real-Time 3D Hand Tracking from Monocular RGB [J].
Mueller, Franziska ;
Bernard, Florian ;
Sotnychenko, Oleksandr ;
Mehta, Dushyant ;
Sridhar, Srinath ;
Casas, Dan ;
Theobalt, Christian .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :49-59
[24]   Real-time Hand Tracking under Occlusion from an Egocentric RGB-D Sensor [J].
Mueller, Franziska ;
Mehta, Dushyant ;
Sotnychenko, Oleksandr ;
Sridhar, Srinath ;
Casas, Dan ;
Theobalt, Christian .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :1163-1172
[25]  
Nogales Ruben, 2021, Advances and Applications in Computer Science, Electronics and Industrial Engineering. Proceedings of CSEI 2020. Advances in Intelligent Systems and Computing (AISC 1307), P185, DOI 10.1007/978-981-33-4565-2_12
[26]   Efficient Annotation and Learning for 3D Hand Pose Estimation: A Survey [J].
Ohkawa, Takehiko ;
Furuta, Ryosuke ;
Sato, Yoichi .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2023, 131 (12) :3193-3206
[27]   HandOccNet: Occlusion-Robust 3D Hand Mesh Estimation Network [J].
Park, JoonKyu ;
Oh, Yeonguk ;
Moon, Gyeongsik ;
Choi, Hongsuk ;
Lee, Kyoung Mu .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :1486-1495
[28]   Learning to Refine Object Segments [J].
Pinheiro, Pedro O. ;
Lin, Tsung-Yi ;
Collobert, Ronan ;
Dollar, Piotr .
COMPUTER VISION - ECCV 2016, PT I, 2016, 9905 :75-91
[29]   Embodied Hands: Modeling and Capturing Hands and Bodies Together [J].
Romero, Javier ;
Tzionas, Dimitrios ;
Black, Michael J. .
ACM TRANSACTIONS ON GRAPHICS, 2017, 36 (06)
[30]   Deep learning [J].
Rusk, Nicole .
NATURE METHODS, 2016, 13 (01) :35-35