OAF-Net: An Occlusion-Aware Anchor-Free Network for Pedestrian Detection in a Crowd

被引:19
作者
Li, Qiming [1 ,2 ]
Su, Yijing [1 ,2 ]
Gao, Yin [1 ,2 ]
Xie, Feng [3 ]
Li, Jun [1 ,2 ]
机构
[1] Chinese Acad Sci, Quanzhou Inst Equipment Mfg, Haixi Inst, Lab Robot & Intelligent Syst, Quanzhou 362216, Fujian, Peoples R China
[2] Fujian Sci & Technol, Innovat Lab Optoelect Informat China, Fuzhou 350108, Fujian, Peoples R China
[3] Inst Automat & Commun, Dept Traff & Assistance, D-39106 Magdeburg, Germany
基金
中国国家自然科学基金;
关键词
Detectors; Training; Feature extraction; Proposals; Avalanche photodiodes; Head; Benchmark testing; Pedestrian detection; occlusion-aware; anchor-free; crowd scenes; VEHICLE;
D O I
10.1109/TITS.2022.3171250
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Although pedestrian detection has achieved promising performance with the development of deep learning techniques, it remains a great challenge to detect heavily occluded pedestrians in crowd scenes. Therefore, to make the anchor-free network pay more attention to learning the hard examples of occluded pedestrians, we propose a simple but effective Occlusion-aware Anchor-Free Network (namely OAF-Net) for pedestrian detection in crowd scenes. Specifically, we first design a novel occlusion-aware detection head, which includes three separate center prediction branches combining with the scale and offset prediction branches. In the detection head of OAF-Net, occluded pedestrian instances are assigned to the most suitable center prediction branch according to the occlusion level of human body. To optimize the center prediction, we accordingly propose a novel weighted Focal Loss where pedestrian instances are assigned with different weights according to their visibility ratios, so that the occluded pedestrians are up-weighted during the training process. Our OAF-Net is able to model different occlusion levels of pedestrian instances effectively, and can be optimized towards catching a high-level understanding of the hard training samples of occluded pedestrians. Experiments on the challenging CityPersons, Caltech, and CrowdHuman benchmarks sufficiently validate the efficacy of our OAF-Net for pedestrian detection in crowd scenes.
引用
收藏
页码:21291 / 21300
页数:10
相关论文
共 77 条
[31]  
Law H., 2019, CornerNet-Lite: Efficient keypoint based object detection
[32]   CornerNet: Detecting Objects as Paired Keypoints [J].
Law, Hei ;
Deng, Jia .
COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 :765-781
[33]  
Li J., 2020, P 28 ACM INT C MULT, P1615
[34]   Scale-Aware Fast R-CNN for Pedestrian Detection [J].
Li, Jianan ;
Liang, Xiaodan ;
Shen, Shengmei ;
Xu, Tingfa ;
Feng, Jiashi ;
Yan, Shuicheng .
IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (04) :985-996
[35]   Attentive Contexts for Object Detection [J].
Li, Jianan ;
Wei, Yunchao ;
Liang, Xiaodan ;
Dong, Jian ;
Xu, Tingfa ;
Feng, Jiashi ;
Yan, Shuicheng .
IEEE TRANSACTIONS ON MULTIMEDIA, 2017, 19 (05) :944-954
[36]   Conditional random fields as message passing mechanism in anchor-free network for multi-scale pedestrian detection [J].
Li, Qiming ;
Qiang, Hua ;
Li, Jun .
INFORMATION SCIENCES, 2021, 550 :1-12
[37]   Feature Pyramid Networks for Object Detection [J].
Lin, Tsung-Yi ;
Dollar, Piotr ;
Girshick, Ross ;
He, Kaiming ;
Hariharan, Bharath ;
Belongie, Serge .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :936-944
[38]   Adaptive NMS: Refining Pedestrian Detection in a Crowd [J].
Liu, Songtao ;
Huang, Di ;
Wang, Yunhong .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :6452-6461
[39]   Receptive Field Block Net for Accurate and Fast Object Detection [J].
Liu, Songtao ;
Huang, Di ;
Wang, Yunhong .
COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 :404-419
[40]   Learning Efficient Single-Stage Pedestrian Detectors by Asymptotic Localization Fitting [J].
Liu, Wei ;
Liao, Shengcai ;
Hu, Weidong ;
Liang, Xuezhi ;
Chen, Xiao .
COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 :643-659