Object detection using YOLO: challenges, architectural successors, datasets and applications

被引:386
作者
Diwan, Tausif [1 ]
Anirudh, G. [2 ]
Tembhurne, Jitendra, V [1 ]
机构
[1] Indian Inst Informat Technol, Dept Comp Sci & Engn, Nagpur, Maharashtra, India
[2] Cent Univ Rajasthan, Dept Data Sci & Analyt, Jaipur, Rajasthan, India
关键词
Object detection; Convolutional neural networks; YOLO; Deep learning; Computer vision; RECURRENT NEURAL-NETWORK; FASTER;
D O I
10.1007/s11042-022-13644-y
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection is one of the predominant and challenging problems in computer vision. Over the decade, with the expeditious evolution of deep learning, researchers have extensively experimented and contributed in the performance enhancement of object detection and related tasks such as object classification, localization, and segmentation using underlying deep models. Broadly, object detectors are classified into two categories viz. two stage and single stage object detectors. Two stage detectors mainly focus on selective region proposals strategy via complex architecture; however, single stage detectors focus on all the spatial region proposals for the possible detection of objects via relatively simpler architecture in one shot. Performance of any object detector is evaluated through detection accuracy and inference time. Generally, the detection accuracy of two stage detectors outperforms single stage object detectors. However, the inference time of single stage detectors is better compared to its counterparts. Moreover, with the advent of YOLO (You Only Look Once) and its architectural successors, the detection accuracy is improving significantly and sometime it is better than two stage detectors. YOLOs are adopted in various applications majorly due to their faster inferences rather than considering detection accuracy. As an example, detection accuracies are 63.4 and 70 for YOLO and Fast-RCNN respectively, however, inference time is around 300 times faster in case of YOLO. In this paper, we present a comprehensive review of single stage object detectors specially YOLOs, regression formulation, their architecture advancements, and performance statistics. Moreover, we summarize the comparative illustration between two stage and single stage object detectors, among different versions of YOLOs, applications based on two stage detectors, and different versions of YOLOs along with the future research directions.
引用
收藏
页码:9243 / 9275
页数:33
相关论文
共 84 条
  • [61] Szegedy C, 2015, PROC CVPR IEEE, P1, DOI 10.1109/CVPR.2015.7298594
  • [62] Thai L. H., 2012, INT J INFORM TECHNOL, V4, P32, DOI [DOI 10.5815/IJITCS.2012.05.05, https://doi.org/10.5815/ijitcs.2012.05.05]
  • [63] Tsang S-H, 2018, REV INCEPTION V4 EVO
  • [64] Focal Loss for Dense Object Detection
    Lin, Tsung-Yi
    Goyal, Priya
    Girshick, Ross
    He, Kaiming
    Dollar, Piotr
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2999 - 3007
  • [65] Ujjwalkarn, 2016, INT EXPL CONV NEUR N
  • [66] IMCFN: Image-based malware classification using fine-tuned convolutional neural network architecture
    Vasan, Danish
    Alazab, Mamoun
    Wassan, Sobia
    Naeem, Hamad
    Safaei, Babak
    Zheng, Qin
    [J]. COMPUTER NETWORKS, 2020, 171 (171)
  • [67] Robust real-time face detection
    Viola, P
    Jones, MJ
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2004, 57 (02) : 137 - 154
  • [68] Deep Learning for Computer Vision: A Brief Review
    Voulodimos, Athanasios
    Doulamis, Nikolaos
    Doulamis, Anastasios
    Protopapadakis, Eftychios
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2018, 2018
  • [69] CSPNet: A New Backbone that can Enhance Learning Capability of CNN
    Wang, Chien-Yao
    Liao, Hong-Yuan Mark
    Wu, Yueh-Hua
    Chen, Ping-Yang
    Hsieh, Jun-Wei
    Yeh, I-Hau
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 1571 - 1580
  • [70] Wang XG, 2018, 2018 INTERNATIONAL CONFERENCE ON SENSING, DIAGNOSTICS, PROGNOSTICS, AND CONTROL (SDPC), P676, DOI [10.1109/SDPC.2018.00130, 10.1109/SDPC.2018.8664773]