Real-time traffic sign detection based on multiscale attention and spatial information aggregator

被引:32
作者
Zhang, Jianming [1 ,2 ]
Ye, Zi [1 ,2 ]
Jin, Xiaokang [3 ]
Wang, Jin [2 ]
Zhang, Jin [2 ]
机构
[1] Changsha Univ Sci & Technol, Key Lab Safety Control Bridge Engn, Minist Educ, Changsha 410114, Peoples R China
[2] Changsha Univ Sci & Technol, Sch Comp & Commun Engn, Changsha 410114, Peoples R China
[3] Jinhua Adv Res Inst, Jinhua 321013, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; Traffic sign detection; YOLO; Small objects; Multiscale attention; RECOGNITION; NETWORK;
D O I
10.1007/s11554-022-01252-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traffic sign detection, as an important part of intelligent driving, can effectively guide drivers to regulate driving and reduce the occurrence of traffic accidents. Currently, the deep learning-based detection methods have achieved very good performance. However, existing network models do not adequately consider the importance of lower-layer features for traffic sign detection. The lack of information on the lower-layer features is a major obstacle to the accurate detection of traffic signs. To solve the above problems, we propose a novel and efficient traffic sign detection method. First, we remove a prediction branch of the YOLOv3 network model to reduce the redundancy of the network model parameters and improve the real-time performance of detection. After that, we propose a multiscale attention feature module. This module fuses the feature information from different layers and refines the features to enhance the Feature Pyramid Network. In addition, we introduce a spatial information aggregator. This enables the spatial information of the lower-layer feature maps to be fused into the higher-layer feature maps. The robustness of our proposed method is further demonstrated by experiments on GTSDB, CCTSDB2021 and TT100k datasets. Specifically, the average execution time on CCTSDB2021 demonstrates the excellent real-time performance of our method. The experimental results show that the method has better accuracy than the original YOLOv3 and YOLOv5 network models.
引用
收藏
页码:1155 / 1167
页数:13
相关论文
共 39 条
[1]   Evaluation of deep neural networks for traffic sign detection systems [J].
Arcos-Garcia, Alvaro ;
Alvarez-Garcia, Juan A. ;
Soria-Morillo, Luis M. .
NEUROCOMPUTING, 2018, 316 :332-344
[2]  
Cai Z., 2012, P INT C IM PROC COMP, P1
[3]   Attentional Feature Fusion [J].
Dai, Yimian ;
Gieseke, Fabian ;
Oehmcke, Stefan ;
Wu, Yiquan ;
Barnard, Kobus .
2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, :3559-3568
[4]   Real Time Visual Trafric Lights Recognition Based on Spot Light Detection and Adaptive Traffic Lights Templates [J].
de Charette, Raoul ;
Nashashibi, Fawzi .
2009 IEEE INTELLIGENT VEHICLES SYMPOSIUM, VOLS 1 AND 2, 2009, :358-363
[5]   CRACT: Cascaded Regression-Align-Classification for Robust Tracking [J].
Fan, Heng ;
Ling, Haibin .
2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, :7013-7020
[6]   Siamese Cascaded Region Proposal Networks for Real-Time Visual Tracking [J].
Fan, Heng ;
Ling, Haibin .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :7944-7953
[7]   Dual Attention Network for Scene Segmentation [J].
Fu, Jun ;
Liu, Jing ;
Tian, Haijie ;
Li, Yong ;
Bao, Yongjun ;
Fang, Zhiwei ;
Lu, Hanqing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3141-3149
[8]  
github, CODE ADDR
[9]  
Houben S, 2013, IEEE INT C INTELL TR, P7, DOI 10.1109/ITSC.2013.6728595
[10]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/TPAMI.2019.2913372, 10.1109/CVPR.2018.00745]