YOLOv3-A: a traffic sign detection network based on attention mechanism

被引:0
|
作者
Guo F. [1 ]
Zhang Y. [1 ]
Tang J. [1 ]
Li W. [1 ]
机构
[1] School of Automation, Central South University, Changsha
来源
Tongxin Xuebao/Journal on Communications | 2021年 / 42卷 / 01期
基金
中国国家自然科学基金;
关键词
Attention mechanism; Semantic segmentation; Small target detection; Traffic sign detection;
D O I
10.11959/j.issn.1000-436x.2021031
中图分类号
学科分类号
摘要
To solve the problem that the existing YOLOv3 algorithm had more false detections and missed detections for traffic sign detection task with small target problems and complex background, based on the YOLOv3, a channel attention method for target detection and a spatial attention method based on semantic segmentation guidance were proposed to form the YOLOv3-A (attention) algorithm. The detection features in the channel and spatial dimensions were recalibrated, allowing the network to focus and enhance the effective features, and suppress interference features, which greatly improved the detection performance. Experiments on the TT100K traffic sign data set show that the algorithm improves the detection performance of small targets, and the accuracy and recall rate of the YOLOv3 are improved by 1.9% and 2.8% respectively. © 2021, Editorial Board of Journal on Communications. All right reserved.
引用
收藏
页码:87 / 99
页数:12
相关论文
共 28 条
  • [1] DALAL N, TRIGGS B., Histograms of oriented gradients for human detection, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 886-893, (2005)
  • [2] LEE T S., Image representation using 2D Gabor wavelets, IEEE Transactions on Pattern Analysis & Machine Intelligence, 18, 10, pp. 959-971, (1996)
  • [3] VIOLA P A, JONES M J., Rapid object detection using a boosted cascade of simple features, IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 511-518, (2001)
  • [4] REN S, HE K, GIRSHICK R, Et al., Faster R-CNN: towards real-time object detection with region proposal networks, IEEE Transactions on Pattern Analysis and Machine Intelligence, 39, 6, pp. 1137-1149, (2017)
  • [5] REDMON J, DIVVALA S, GIRSHICK R, Et al., You only look once: Unified, real-time object detection, 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 779-788, (2016)
  • [6] LIU W, ANGUELOV D, ERHAN D, Et al., SSD: single shot multibox detector, European Conference on Computer Vision, pp. 21-37, (2016)
  • [7] RAJENDRAN S P, SHINE L, PRADEEP R, Et al., Fast and accurate traffic sign recognition for self driving cars using RetinaNet based detector, 2019 International Conference on Communication and Electronics Systems, pp. 784-790, (2019)
  • [8] LIN T Y, GOYAL P, GIRSHICK R, Et al., Focal loss for dense object detection, 2017 IEEE International Conference on Computer Vision, pp. 2980-2988, (2017)
  • [9] HE K, ZHANG X, REN S, Et al., Deep residual learning for image recognition, 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, (2016)
  • [10] HOUBEN S, STALLKAMP J, SALMEN J, Et al., Detection of traffic signs in real-world images: the German traffic sign detection benchmark, The 2013 International Joint Conference on Neural Networks, pp. 1-8, (2013)