Transformer-based 3D Human pose estimation and action achievement evaluation

被引:0
|
作者
Yang, Aolei [1 ,2 ]
Zhou, Yinghong [1 ]
Yang, Banghua [1 ]
Xu, Yulin [1 ]
机构
[1] School of Mechatronic Engineering and Automation, Shanghai University, Shanghai
[2] Shanghai Key Laboratory of Power Station Automation Technology, Shanghai
来源
Yi Qi Yi Biao Xue Bao/Chinese Journal of Scientific Instrument | 2024年 / 45卷 / 04期
关键词
3D human pose estimation; action evaluation; deep learning; dynamic time wrapping;
D O I
10.19650/j.cnki.cjsi.J2312272
中图分类号
学科分类号
摘要
According to the challenges of human pose analysis and assessment in domains such as human-computer interaction and medical rehabilitation, this paper introduces a Transformer-based methodology for 3D human pose estimation and the evaluation of action achievement. Firstly, key points of human pose and their joint angles were defined, and based on the deep pose estimation network (DPEN), a Transformer-based 3D human pose estimation model (TPEM) is proposed and constructed, the incorporation of Transformer facilitates better enhanced extraction of long-term sequential features of human pose. Secondly, the TPEM model′s outcomes in 3D human pose estimation are utilized to formulate a dynamic time warping algorithm, which focuses on weighted 3D joint angles. This algorithm temporally aligns pose keyframes for different individuals performing the same action and subsequently introduces an assessment method for action accomplishment to provide scores for the degree of action fulfillment. Finally, through experimental validation across various datasets, TPEM achieves an average joint point error of 37. 3 mm on the Human3. 6 M dataset, while the dynamic time warping algorithm based on weighted 3D joint angles yields an average error of 5. 08 frames on the Fit3D dataset. These results demonstrate the feasibility and effectiveness of the proposed approach for 3D human pose estimation and action accomplishment assessment. © 2024 Science Press. All rights reserved.
引用
收藏
页码:136 / 144
页数:8
相关论文
共 19 条
  • [1] YANG AO L, CHEN Y L, XU Y L., Humanoid motion planning of robotic arm based on reinforcement learning, Chinese Journal of Scientific Instrument, 42, 12, pp. 136-145, (2021)
  • [2] TANG X Y, SONG AI G., Human pose estimation and its implementation in scenario interaction system of rehabilitation training [ J ], Chinese Journal of Scientific Instrument, 39, 11, pp. 195-203, (2018)
  • [3] WEI W L, LIN J C, LIU T L, Et al., Capturing humans in motion: Temporal-attentive 3D human pose and shape estimation from monocular video[C], Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 13201-13210, (2022)
  • [4] LIU J Y, LIU Y K, JIA X H, Et al., Research on human pose visual recognition algorithm based on model constraints, Chinese Journal of Scientific Instrument, 41, 4, pp. 208-217, (2020)
  • [5] ZHAO W, WANG W Q, TIAN Y J., GraFormer: Graph-oriented transformer for 3D pose estimation [ C ], Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 20406-20415, (2022)
  • [6] ZHENG C, ZHU S, MENDIETA M, Et al., 3D Human pose estimation with spatial and temporal transformers[C], Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 11636-11645, (2021)
  • [7] YANG A L, LIU G C, NAEEM W, Et al., A monocular 3D human pose estimation approach for virtual character skeleton retargeting[J], Journal of Ambient Intelligence and Humanized Computing, 4, pp. 9563-9574, (2023)
  • [8] ZHANG Y T, WANG Q A, TU F Y, Et al., Automatic moving pose grading for golf swing in sports, IEEE International Conference on Image Processing, pp. 41-45, (2022)
  • [9] XU W, XIANG D H, WANG G T, Et al., Multiview video-based 3-D pose estimation of patients in computer-assisted rehabilitation environment [ J ], IEEE Transactions on Human-Machine Systems, 52, 2, pp. 196-206, (2022)
  • [10] CAO Z, HIDALGO G, SIMON T, Et al., OpenPose: Realtime multi-person 2D pose estimation using part affinity fields[J], IEEE Transactions on Pattern Analysis and Machine Intelligence, 43, 1, pp. 172-186, (2022)