An Improved Real-Time Detection Transformer Model for the Intelligent Survey of Traffic Safety Facilities

被引:5
作者
Wan, Yan [1 ]
Wang, Hui [2 ]
Lu, Lingxin [2 ,3 ]
Lan, Xin [2 ,3 ]
Xu, Feifei [1 ]
Li, Shenglin [3 ]
机构
[1] Ningbo Univ Technol, Sch Civil & Transportat Engn, Ningbo 315211, Peoples R China
[2] Chongqing Univ, Sch Civil Engn, Key Lab New Technol Construct Cities Mt Area, Minist Educ, Chongqing 400045, Peoples R China
[3] Southwest Univ, Coll Artificial Intelligence, Chongqing 400715, Peoples R China
关键词
traffic safety facility asset survey; traffic safety facility recognition; YOLO; real-time detection transformer; reparameterized generalized feature pyramid network; SIGN RECOGNITION; OPTIMIZATION; NETWORK; VISION;
D O I
10.3390/su162310172
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
The undertaking of traffic safety facility (TSF) surveys represents a significant labor-intensive endeavor, which is not sustainable in the long term. The subject of traffic safety facility recognition (TSFR) is beset with numerous challenges, including those associated with background misclassification, the diminutive dimensions of the targets, the spatial overlap of detection targets, and the failure to identify specific targets. In this study, transformer-based and YOLO (You Only Look Once) series target detection algorithms were employed to construct TSFR models to ensure both recognition accuracy and efficiency. The TSF image dataset, comprising six categories of TSFs in urban areas of three cities, was utilized for this research. The dimensions and intricacies of the Detection Transformer (DETR) family of models are considerably more substantial than those of the YOLO family. YOLO-World and Real-Time Detection Transformer (RT-DETR) models were optimal and comparable for the TSFR task, with the former exhibiting a higher detection efficiency and the latter a higher detection accuracy. The RT-DETR model exhibited a notable reduction in model complexity by 57% in comparison to the DINO (DETR with improved denoising anchor boxes for end-to-end object detection) model while also demonstrating a slight enhancement in recognition accuracy. The incorporation of the RepGFPN (Reparameterized Generalized Feature Pyramid Network) module has markedly enhanced the multi-target detection accuracy of RT-DETR, with a mean average precision (mAP) of 82.3%. The introduction of RepGFPN significantly enhanced the detection rate of traffic rods, traffic sign boards, and water surround barriers and somewhat ameliorated the problem of duplicate detection.
引用
收藏
页数:22
相关论文
共 61 条
[11]   The Cityscapes Dataset for Semantic Urban Scene Understanding [J].
Cordts, Marius ;
Omran, Mohamed ;
Ramos, Sebastian ;
Rehfeld, Timo ;
Enzweiler, Markus ;
Benenson, Rodrigo ;
Franke, Uwe ;
Roth, Stefan ;
Schiele, Bernt .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3213-3223
[12]   Real Time Visual Trafric Lights Recognition Based on Spot Light Detection and Adaptive Traffic Lights Templates [J].
de Charette, Raoul ;
Nashashibi, Fawzi .
2009 IEEE INTELLIGENT VEHICLES SYMPOSIUM, VOLS 1 AND 2, 2009, :358-363
[13]   RepVGG: Making VGG-style ConvNets Great Again [J].
Ding, Xiaohan ;
Zhang, Xiangyu ;
Ma, Ningning ;
Han, Jungong ;
Ding, Guiguang ;
Sun, Jian .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :13728-13737
[14]   Automatic Extraction of Roadside Traffic Facilities From Mobile Laser Scanning Point Clouds Based on Deep Belief Network [J].
Fang, Lina ;
Shen, Guixi ;
Luo, Haifeng ;
Chen, Chongcheng ;
Zhao, Zhiyuan .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (04) :1964-1980
[15]   Vision meets robotics: The KITTI dataset [J].
Geiger, A. ;
Lenz, P. ;
Stiller, C. ;
Urtasun, R. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1231-1237
[16]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[17]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[18]   The ApolloScape Open Dataset for Autonomous Driving and Its Application [J].
Huang, Xinyu ;
Wang, Peng ;
Cheng, Xinjing ;
Zhou, Dingfu ;
Geng, Qichuan ;
Yang, Ruigang .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (10) :2702-2719
[19]   Lightweight pruning model for road distress detection using unmanned aerial vehicles [J].
Jiang, Shengchuan ;
Wang, Hui ;
Ning, Zhipeng ;
Li, Shenglin .
AUTOMATION IN CONSTRUCTION, 2024, 168
[20]   A Model for Infrastructure Detection along Highways Based on Remote Sensing Images from UAVs [J].
Jiang, Xian ;
Cui, Qing ;
Wang, Chongguo ;
Wang, Fan ;
Zhao, Yingxiang ;
Hou, Yongjie ;
Zhuang, Rujun ;
Mei, Yunfei ;
Shi, Gang .
SENSORS, 2023, 23 (08)