SCALE-Pose: Skeletal Correction and Language Knowledge-assisted for 3D Human Pose Estimation

被引:0
作者
Ma, Xinnan [1 ]
Li, Yaochen [1 ]
Zhao, Limeng [1 ]
Zhou, ChenXu [1 ]
Xu, Yuncheng [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Software Engn, Xian 710049, Peoples R China
来源
PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XI | 2025年 / 15041卷
关键词
3D human pose estimation; Transformer; Priori knowledge; Skeletal correction; Large language model;
D O I
10.1007/978-981-97-8795-1_39
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer-based 3D human pose estimation methods typically use 2D joint sequences as inputs, leveraging spatial and temporal transformer encoders to model the 3D human pose. However, these methods often neglect to incorporate skeletal constraints to limit joint motion, and few consider integrating prior category knowledge to enhance potential joint representations. To solve these problems, we propose a new method named SCALE-Pose. Firstly, this method incorporates the spatial and temporal skeleton correction blocks to improve the ability of modeling the long-range dependency of the spatiotemporal motion of specific skeletons. Next, a four-stream radian loss based on skeleton angle error is introduced to constrain the motion space of joints. Finally, an auxiliary method employs global-local prompts from a large language model to generate prior category knowledge, improving the ability to generalize prior category knowledge. Experimental results on Human3.6M and MPI-INF-3DHP datasets demonstrate that our method outperforms existing approaches.
引用
收藏
页码:578 / 592
页数:15
相关论文
共 50 条
  • [41] GTIGNet: Global Topology Interaction Graphormer Network for 3D hand pose estimation
    Liu, Yanjun
    Fan, Wanshu
    Wang, Cong
    Wen, Shixi
    Yang, Xin
    Zhang, Qiang
    Wei, Xiaopeng
    Zhou, Dongsheng
    NEURAL NETWORKS, 2025, 185
  • [42] 3D Hand Pose Estimation From Monocular RGB With Feature Interaction Module
    Guo, Shaoxiang
    Rigall, Eric
    Ju, Yakun
    Dong, Junyu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5293 - 5306
  • [43] A Study on 3D Human Pose Estimation Using Through-Wall IR-UWB Radar and Transformer
    Kim, Gon Woo
    Lee, Sang Won
    Son, Ha Young
    Choi, Kae Won
    IEEE ACCESS, 2023, 11 : 15082 - 15095
  • [44] Diffusion-Based Hypotheses Generation and Joint-Level Hypotheses Aggregation for 3D Human Pose Estimation
    Shan, Wenkang
    Zhang, Yuhuai
    Zhang, Xinfeng
    Wang, Shanshe
    Zhou, Xilong
    Ma, Siwei
    Gao, Wen
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (11) : 10678 - 10691
  • [45] Multi-Person 3D Pose and Shape Estimation via Inverse Kinematics and Refinement
    Cha, Junuk
    Saqlain, Muhammad
    Kim, GeonU
    Shin, Mingyu
    Baek, Seungryul
    COMPUTER VISION - ECCV 2022, PT V, 2022, 13665 : 660 - 677
  • [46] A Study on 3D Human Pose Estimation with a Hybrid Algorithm of Spatio-temporal Semantic Graph Attention and Deep Learning
    Lin, Shengqing
    INFORMATION TECHNOLOGY AND CONTROL, 2024, 53 (04): : 1042 - 1059
  • [47] ESMformer: Error-aware self-supervised transformer for multi-view 3D human pose estimation
    Zhang, Lijun
    Zhou, Kangkang
    Lu, Feng
    Li, Zhenghao
    Shao, Xiaohu
    Zhou, Xiang-Dong
    Shi, Yu
    PATTERN RECOGNITION, 2025, 158
  • [48] 3D hand pose and mesh estimation via a generic Topology-aware Transformer model
    Yu, Shaoqi
    Wang, Yintong
    Chen, Lili
    Zhang, Xiaolin
    Li, Jiamao
    FRONTIERS IN NEUROROBOTICS, 2024, 18
  • [49] CLIP-Hand3D: Exploiting 3D Hand Pose Estimation via Context-Aware Prompting
    Guo, Shaoxiang
    Cai, Qing
    Qi, Lin
    Dong, Junyu
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4896 - 4907
  • [50] DDBMHT: A Diffusion-Based Double-Branch Multi-Hypothesis Transformer for 3D Human Pose Estimation in Video
    Bao, Weijie
    Xiang, Xuezhi
    2024 9TH INTERNATIONAL CONFERENCE ON ELECTRONIC TECHNOLOGY AND INFORMATION SCIENCE, ICETIS 2024, 2024, : 35 - 39