A Multiscale Coarse-to-Fine Human Pose Estimation Network With Hard Keypoint Mining

被引:2
作者
Jiang, Xiaoyan [1 ]
Tao, Hangyu [1 ]
Hwang, Jenq-Neng [2 ]
Fang, Zhijun [3 ]
机构
[1] Shanghai Univ Engn Sci, Sch Elect & Elect Engn, Shanghai 201620, Peoples R China
[2] Univ Washington, Dept Elect & Comp Engn, Seattle, WA 98195 USA
[3] Donghua Univ, Sch Comp Sci & Technol, Shanghai 200051, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2024年 / 54卷 / 03期
基金
中国国家自然科学基金;
关键词
Pose estimation; Standards; Convolution; Training; Task analysis; Heating systems; Detectors; Hard sample mining; human pose estimation; multiscale;
D O I
10.1109/TSMC.2023.3328876
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Current convolution neural network (CNN)-based multiperson pose estimators have achieved great progress, however, they pay no or less attention to "hard" samples, such as occluded keypoints, small and nearly invisible keypoints, and ambiguous keypoints. In this article, we explicitly deal with these "hard" samples by proposing a novel multiscale coarse-to-fine human pose estimation network ((HMPN)-P-2), which includes two sequential subnetworks: CoarseNet and FineNet. CoarseNet conducts a coarse prediction to locate "simple" keypoints like hands and ankles with a multiscale fusion module, which is integrated with bottleneck, resulting in a novel module called multiscale bottleneck. The new module improves the multiscale representation ability of the network in a fine-grained level, while marginally reducing the computation cost because of group convolution. FineNet further infers "hard" keypoints and refines "simple" keypoints simultaneously with a hard keypoint mining loss. Distinct from the previous works, the proposed loss deals with "hard" keypoints differentially and prevents "simple" keypoints from dominating the computed gradients during training. Experiments on the COCO keypoint benchmark show that our approach achieves superior pose estimation performance compared with other state-of-the-art methods. Source code is available for further research: https://github.com/sues-vision/C2F-HumanPoseEstimation.
引用
收藏
页码:1730 / 1741
页数:12
相关论文
共 50 条
  • [31] Attention Refined Network for Human Pose Estimation
    Wang, Xiangyang
    Tong, Jiangwei
    Wang, Rui
    NEURAL PROCESSING LETTERS, 2021, 53 (04) : 2853 - 2872
  • [32] Hierarchical Adversarial Network for Human Pose Estimation
    Radwan, Ibrahim
    Moustafa, Nour
    Keating, Byron
    Choo, Kim-Kmang Raymond
    Goecke, Roland
    IEEE ACCESS, 2019, 7 : 103619 - 103628
  • [33] Multistage attention network for human pose estimation
    Zhou, Jingyang
    Wen, Guangzhao
    Zhang, Yu
    Geng, Xin
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [34] Adversarial attack on human pose estimation network
    Zhang, Zhaoxin
    Huang, Shize
    Liu, Xiaowen
    Fan, Qianhui
    Dong, Decun
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (01)
  • [35] Modelling Sign Language with Encoder-Only Transformers and Human Pose Estimation Keypoint Data
    Woods, Luke T.
    Rana, Zeeshan A.
    MATHEMATICS, 2023, 11 (09)
  • [36] Spatial-Temporal Multiscale Constrained Learning for mmWave-Based Human Pose Estimation
    Chen, Lin
    Guo, Xuemei
    Wang, Guoli
    Li, Hongyi
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (03) : 1108 - 1120
  • [37] Coarse-to-Fine Lung Nodule Segmentation in CT Images With Image Enhancement and Dual-Branch Network
    Wu, Zhitong
    Zhou, Qianjun
    Wang, Feng
    IEEE ACCESS, 2021, 9 (09): : 7255 - 7262
  • [38] A Feature Aggregation Hourglass Network for Human Pose Estimation
    Lu, Fan
    Yang, Yan
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [39] Adaptive data augmentation network for human pose estimation
    Wang, Dong
    Xie, Wenjun
    Cai, Youcheng
    Liu, Xiaoping
    DIGITAL SIGNAL PROCESSING, 2022, 129
  • [40] Hierarchical Graph Neural Network for Human Pose Estimation
    Zheng, Guanghua
    Zhao, Zhongqiu
    Zhang, Zhao
    Yang, Yi
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 2663 - 2668