Focal Loss for Dense Object Detection

被引:14925
|
作者
Lin, Tsung-Yi [1 ]
Goyal, Priya [1 ]
Girshick, Ross [1 ]
He, Kaiming [1 ]
Dollar, Piotr [1 ]
机构
[1] Facebook AI Res FAIR, Menlo Pk, CA 94025 USA
来源
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV) | 2017年
关键词
D O I
10.1109/ICCV.2017.324
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The highest accuracy object detectors to date are based on a two-stage approach popularized by R-CNN, where a classifier is applied to a sparse set of candidate object locations. In contrast, one-stage detectors that are applied over a regular, dense sampling of possible object locations have the potential to be faster and simpler, but have trailed the accuracy of two-stage detectors thus far. In this paper, we investigate why this is the case. We discover that the extreme foreground-background class imbalance encountered during training of dense detectors is the central cause. We propose to address this class imbalance by reshaping the standard cross entropy loss such that it down-weights the loss assigned to well-classified examples. Our novel Focal Loss focuses training on a sparse set of hard examples and prevents the vast number of easy negatives from overwhelming the detector during training. To evaluate the effectiveness of our loss, we design and train a simple dense detector we call RetinaNet. Our results show that when trained with the focal loss, RetinaNet is able to match the speed of previous one-stage detectors while surpassing the accuracy of all existing state-of-the-art two-stage detectors.
引用
收藏
页码:2999 / 3007
页数:9
相关论文
共 50 条
  • [1] Focal Loss for Dense Object Detection
    Lin, Tsung-Yi
    Goyal, Priya
    Girshick, Ross
    He, Kaiming
    Dollar, Piotr
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (02) : 318 - 327
  • [2] Equalized Focal Loss for Dense Long-Tailed Object Detection
    Li, Bo
    Yao, Yongqiang
    Tan, Jingru
    Zhang, Gang
    Yu, Fengwei
    Lu, Jianwei
    Luo, Ye
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 6980 - 6989
  • [3] Soft focal loss: Evaluating sample quality for dense object detection
    Wang, Zhenyuan
    Xie, Xuemei
    Yang, Jianxiu
    Shi, Guangming
    NEUROCOMPUTING, 2022, 480 : 271 - 280
  • [4] Generalized Focal Loss: Towards Efficient Representation Learning for Dense Object Detection
    Li, Xiang
    Lv, Chengqi
    Wang, Wenhai
    Li, Gang
    Yang, Lingfeng
    Yang, Jian
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3139 - 3153
  • [5] Generalized Focal Loss: Learning Qualified and Distributed Bounding Boxes for Dense Object Detection
    Li, Xiang
    Wang, Wenhai
    Wu, Lijun
    Chen, Shuo
    Hu, Xiaolin
    Li, Jun
    Tang, Jinhui
    Yang, Jian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] Mask focal loss: a unifying framework for dense crowd counting with canonical object detection networks
    Zhong, Xiaopin
    Wang, Guankun
    Liu, Weixiang
    Wu, Zongze
    Deng, Yuanlong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (27) : 70571 - 70593
  • [7] Automated Focal Loss for Image based Object Detection
    Weber, Michael
    Fuerst, Michael
    Zoellner, J. Marius
    2020 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2020, : 1423 - 1429
  • [8] Focal Loss in 3D Object Detection
    Yun, Peng
    Tai, Lei
    Wang, Yuan
    Liu, Chengju
    Liu, Ming
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (02) : 1263 - 1270
  • [9] Generalized Focal Loss V2: Learning Reliable Localization Quality Estimation for Dense Object Detection
    Li, Xiang
    Wang, Wenhai
    Hu, Xiaolin
    Li, Jun
    Tang, Jinhui
    Yang, Jian
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 11627 - 11636
  • [10] Sample Weighting with Hierarchical Equalization Loss for Dense Object Detection
    Ma, Jia-Wei
    Liang, Min
    Chen, Lei
    Tian, Shu
    Chen, Song-Lu
    Qin, Jingyan
    Yin, Xu-Cheng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 5846 - 5859