Quantification of Occlusion Handling Capability of a 3D Human Pose Estimation Framework

被引:11
作者
Ghafoor, Mehwish [1 ]
Mahmood, Arif [1 ]
机构
[1] Informat Technol Univ, Dept Comp Sci, Lahore 54600, Pakistan
关键词
Action classification; human pose estimation; occlusion aware networks; occlusion handling quantification; temporal dilated CNN;
D O I
10.1109/TMM.2022.3158068
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
3D human pose estimation using monocular images is an important yet challenging task. Existing 3D pose detection methods exhibit excellent performance under normal conditions however their performance may degrade due to occlusion. Recently some occlusion aware methods have also been proposed, however, the occlusion handling capability of these networks has not yet been thoroughly investigated. In the current work, we propose an occlusion-guided 3D human pose estimation framework and quantify its occlusion handling capability by using different protocols. The proposed method estimates more accurate 3D human poses using 2D skeletons with missing joints as input. Missing joints are handled by introducing occlusion guidance that provides extra information about the absence or presence of a joint. Temporal information has also been exploited to better estimate the missing joints. A large number of experiments are performed for the quantification of occlusion handling capability of the proposed method on three publicly available datasets in various settings including random missing joints, fixed body parts missing, and complete frames missing, using mean per joint position error criterion. In addition to that, the quality of the predicted 3D poses is also evaluated using action classification performance as a criterion. 3D poses estimated by the proposed method achieved significantly improved action recognition performance in the presence of missing joints. Our experiments demonstrate the effectiveness of the proposed framework for handling the missing joints as well as quantification of the occlusion handling capability of the deep neural networks.
引用
收藏
页码:3311 / 3318
页数:8
相关论文
共 33 条
  • [11] Exploring Severe Occlusion: Multi-Person 3D Pose Estimation with Gated Convolution
    Gu, Renshu
    Wang, Gaoang
    Hwang, Jenq-Neng
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 8243 - 8250
  • [12] Exploiting Temporal Information for 3D Human Pose Estimation
    Hossain, Mir Rayat Imtiaz
    Little, James J.
    [J]. COMPUTER VISION - ECCV 2018, PT X, 2018, 11214 : 69 - 86
  • [13] Hu JF, 2015, PROC CVPR IEEE, P5344, DOI 10.1109/CVPR.2015.7299172
  • [14] Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments
    Ionescu, Catalin
    Papava, Dragos
    Olaru, Vlad
    Sminchisescu, Cristian
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2014, 36 (07) : 1325 - 1339
  • [15] Propagating LSTM: 3D Pose Estimation Based on Joint Interdependency
    Lee, Kyoungoh
    Lee, Inwoong
    Lee, Sanghoon
    [J]. COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 : 123 - 141
  • [16] Multi-Person Pose Estimation Using Bounding Box Constraint and LSTM
    Li, Miaopeng
    Zhou, Zimeng
    Liu, Xinguo
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (10) : 2653 - 2663
  • [17] Attention Mechanism Exploits Temporal Contexts: Real-time 3D Human Pose Reconstruction
    Liu, Ruixu
    Shen, Ju
    Wang, He
    Chen, Chen
    Cheung, Sen-ching
    Asari, Vijayan
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 5063 - 5072
  • [18] Graph Embedded Pose Clustering for Anomaly Detection
    Markovitz, Amir
    Sharir, Gilad
    Friedman, Itamar
    Zelnik-Manor, Lihi
    Avidan, Shai
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, : 10536 - 10544
  • [19] A simple yet effective baseline for 3d human pose estimation
    Martinez, Julieta
    Hossain, Rayat
    Romero, Javier
    Little, James J.
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2659 - 2668
  • [20] Misra D, 2020, Arxiv, DOI arXiv:1908.08681