Learning to Augment Poses for 3D Human Pose Estimation in Images and Videos

被引:8
作者
Zhang, Jianfeng [1 ]
Gong, Kehong [1 ]
Wang, Xinchao [1 ]
Feng, Jiashi [2 ]
机构
[1] Natl Univ Singapore, Singapore 119077, Singapore
[2] ByteDance, Beijing 100086, Peoples R China
基金
新加坡国家研究基金会;
关键词
Three-dimensional displays; Training; Data models; Pose estimation; Training data; Solid modeling; Videos; Artificial intelligence; computational and artificial intelligence; deep learning; learning systems; machine learning; supervised learning;
D O I
10.1109/TPAMI.2023.3243400
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing 3D human pose estimation methods often suffer inferior generalization performance to new datasets, largely due to the limited diversity of 2D-3D pose pairs in the training data. To address this problem, we present PoseAug, a novel auto-augmentation framework that learns to augment the available training poses towards greater diversity and thus enhances the generalization power of the trained 2D-to-3D pose estimator. Specifically, PoseAug introduces a novel pose augmentor that learns to adjust various geometry factors of a pose through differentiable operations. With such differentiable capacity, the augmentor can be jointly optimized with the 3D pose estimator and take the estimation error as feedback to generate more diverse and harder poses in an online manner. PoseAug is generic and handy to be applied to various 3D pose estimation models. It is also extendable to aid pose estimation from video frames. To demonstrate this, we introduce PoseAug-V, a simple yet effective method that decomposes video pose augmentation into end pose augmentation and conditioned intermediate pose generation. Extensive experiments demonstrate that PoseAug and its extension PoseAug-V bring clear improvements for frame-based and video-based 3D pose estimation on several out-of-domain 3D human pose benchmarks.
引用
收藏
页码:10012 / 10026
页数:15
相关论文
共 61 条
  • [1] Ailing Zeng, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12359), P507, DOI 10.1007/978-3-030-58568-6_30
  • [2] Akhter I, 2015, PROC CVPR IEEE, P1446, DOI 10.1109/CVPR.2015.7298751
  • [3] 2D Human Pose Estimation: New Benchmark and State of the Art Analysis
    Andriluka, Mykhaylo
    Pishchulin, Leonid
    Gehler, Peter
    Schiele, Bernt
    [J]. 2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, : 3686 - 3693
  • [4] Exploiting Spatial-temporal Relationships for 3D Pose Estimation via Graph Convolutional Networks
    Cai, Yujun
    Ge, Liuhao
    Liu, Jun
    Cai, Jianfei
    Cham, Tat-Jen
    Yuan, Junsong
    Thalmann, Nadia Magnenat
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 2272 - 2281
  • [5] Unsupervised 3D Pose Estimation with Geometric Self-Supervision
    Chen, Ching-Hang
    Tyagi, Ambrish
    Agrawal, Amit
    Drover, Dylan
    Rohith, M., V
    Stojanov, Stefan
    Rehg, James M.
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 5707 - 5717
  • [6] Chen MF, 2022, LECT NOTES COMPUT SC, V13683, P222
  • [7] Synthesizing Training Images for Boosting Human 3D Pose Estimation
    Chen, Wenzheng
    Wang, Huan
    Li, Yangyan
    Su, Hao
    Wang, Zhenhua
    Tu, Changhe
    Lischinski, Dani
    Cohen-Or, Daniel
    Chen, Baoquan
    [J]. PROCEEDINGS OF 2016 FOURTH INTERNATIONAL CONFERENCE ON 3D VISION (3DV), 2016, : 479 - 488
  • [8] Weakly-Supervised Discovery of Geometry-Aware Representation for 3D Human Pose Estimation
    Chen, Xipeng
    Lin, Kwan-Yee
    Liu, Wentao
    Qian, Chen
    Lin, Liang
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 10887 - 10896
  • [9] Cascaded Pyramid Network for Multi-Person Pose Estimation
    Chen, Yilun
    Wang, Zhicheng
    Peng, Yuxiang
    Zhang, Zhiqiang
    Yu, Gang
    Sun, Jian
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7103 - 7112
  • [10] Cheng Y, 2020, AAAI CONF ARTIF INTE, V34, P10631