AN END-TO-END FRAMEWORK FOR POSE ESTIMATION OF OCCLUDED PEDESTRIANS

被引:0
作者
Das, Sudip [1 ]
Kishore, Perla Sai Raj [2 ]
Bhattacharya, Ujjwal [1 ]
机构
[1] Indian Stat Inst, Kolkata, India
[2] Inst Engn & Management, Kolkata, India
来源
2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP) | 2020年
关键词
Pose Estimation; Unsupervised Domain Adaptation; Multi-task Learning; Adversarial Learning;
D O I
暂无
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
Pose estimation in the wild is a challenging problem, particularly in situations of (i) occlusions of varying degrees, and (ii) crowded outdoor scenes. Most of the existing studies of pose estimation did not report the performance in similar situations. Moreover, pose annotations for occluded parts of the human figures have not been provided in any of the relevant standard datasets, which in turn creates further difficulties to the required studies for pose estimation of the entire figure for occluded humans. Well known pedestrian detection datasets such as CityPersons contains samples of outdoor scenes but it does not include pose annotations. Here we propose a novel multi-task framework for end-to-end training towards the entire pose estimation of pedestrians including in situations of any kind of occlusion. To tackle this problem, we make use of a pose estimation dataset, MS-COCO, and employ unsupervised adversarial instance-level domain adaptation for estimating the entire pose of occluded pedestrians. The experimental studies show that the proposed framework outperforms the SOTA results for pose estimation, instance segmentation and pedestrian detection in cases of heavy occlusions (HO) and reasonable + heavy occlusions (R + HO) on the two benchmark datasets.
引用
收藏
页码:1446 / 1450
页数:5
相关论文
共 26 条
[1]   2D Human Pose Estimation: New Benchmark and State of the Art Analysis [J].
Andriluka, Mykhaylo ;
Pishchulin, Leonid ;
Gehler, Peter ;
Schiele, Bernt .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :3686-3693
[2]  
[Anonymous], 2020, IEEE T PATTERN ANAL, DOI DOI 10.1109/TPAMI.2018.2844175
[3]   Learning Deep Architectures for AI [J].
Bengio, Yoshua .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2009, 2 (01) :1-127
[4]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[5]   A Coarse-Fine Network for Keypoint Localization [J].
Huang, Shaoli ;
Gong, Mingming ;
Tao, Dacheng .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :3047-3056
[6]   DeeperCut: A Deeper, Stronger, and Faster Multi-person Pose Estimation Model [J].
Insafutdinov, Eldar ;
Pishchulin, Leonid ;
Andres, Bjoern ;
Andriluka, Mykhaylo ;
Schiele, Bernt .
COMPUTER VISION - ECCV 2016, PT VI, 2016, 9910 :34-50
[7]  
Kishore P. S. R., 2019, BMVC
[8]   MultiPoseNet: Fast Multi-Person Pose Estimation Using Pose Residual Network [J].
Kocabas, Muhammed ;
Karagoz, Salih ;
Akbas, Emre .
COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 :437-453
[9]   PifPaf: Composite Fields for Human Pose Estimation [J].
Kreiss, Sven ;
Bertoni, Lorenzo ;
Alahi, Alexandre .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11969-11978
[10]   CrowdPose: Efficient Crowded Scenes Pose Estimation and A New Benchmark [J].
Li, Jiefeng ;
Wang, Can ;
Zhu, Hao ;
Mao, Yihuan ;
Fang, Hao-Shu ;
Lu, Cewu .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :10855-10864