A high-speed YOLO detection model for steel surface defects with the channel residual convolution and fusion-distribution

被引:10
作者
Huang, Jianhang [1 ]
Zhang, Xinliang [1 ]
Jia, Lijie [1 ]
Zhou, Yitian [2 ]
机构
[1] Henan Polytech Univ, Sch Elect Engn & Automat, Henan Int Joint Lab Direct Driveand Control Intell, Jiaozuo 454003, Peoples R China
[2] Zhoushan Yangwangnaxin Technol Co Ltd, Zhoushan 3161041, Zhejiang, Peoples R China
关键词
YOLO; channel residual; channel residual convolution module (CRCM); channel residual cross stage partial (CRCSP); fusion-distribution (FD); CRFD-YOLO;
D O I
10.1088/1361-6501/ad6281
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Accurately and efficiently detecting steel surface defects is a critical step in steel manufacturing. However, the compromise between the detection speed and accuracy remains a major challenge, especially for steel surface defects with large variations in the scale. To address the issue, an improved you only look once (YOLO) based detection model is proposed through the reinforcement of its backbone and neck. Firstly, for the reduction of the redundant parameters and also the improvement of the characterization ability of the model, an effective channel residual structure is adopted to construct a channel residual convolution module and channel residual cross stage partial module as components of the backbone network, respectively. They realize the extraction of both the shallow feature and multi-scale feature simultaneously under a small number of convolutional parameters. Secondly, in the neck of YOLO, a fusion-distribution strategy is employed, which extracts and fuses multi-scale feature maps from the backbone network to provide global information, and then distributes global information into local features of different branches through an inject attention mechanism, thus enhancing the feature gap between different branches. Then, a model called CRFD-YOLO is derived for the steel surface defect detection and localization for the situations where both speed and accuracy are demanding. Finally, extensive experimental validations are conducted to evaluate the performance of CRFD-YOLO. The validation results indicate that CRFD-YOLO achieves a satisfactory detection performance with a mean average precision of 81.3% on the NEU-DET and 71.1% on the GC10-DET. Additionally, CRFD-YOLO achieves a speed of 161 frames per second, giving a great potential in real-time detection and localization tasks.
引用
收藏
页数:14
相关论文
共 37 条
[1]   Run, Don't Walk: Chasing Higher FLOPS for Faster Neural Networks [J].
Chen, Jierun ;
Kao, Shiu-Hong ;
He, Hao ;
Zhuo, Weipeng ;
Wen, Song ;
Lee, Chul-Ho ;
Chan, S. -H. Gary .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, :12021-12031
[2]  
Chen XX, 2024, IEEE SENS J, V24, P20103, DOI 10.1109/JSEN.2024.3392674
[3]   LF-YOLOv4: a lightweight detection model for enhancing the fusion of image features of surface defects in lithium batteries [J].
Chen, Xiaoxin ;
Jiang, Zhansi ;
Cheng, Hao ;
Zheng, Hongxin ;
Du, Yixian .
MEASUREMENT SCIENCE AND TECHNOLOGY, 2024, 35 (02)
[4]   Fast vehicle detection algorithm in traffic scene based on improved SSD [J].
Chen, Zhichao ;
Guo, Haoqi ;
Yang, Jie ;
Jiao, Haining ;
Feng, Zhicheng ;
Chen, Lifang ;
Gao, Tao .
MEASUREMENT, 2022, 201
[5]   A feature enhanced RetinaNet-based for instance-level ship recognition [J].
Cheng, Jing ;
Wang, Rongjie ;
Lin, Anhui ;
Jiang, Desong ;
Wang, Yichun .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 126
[6]   Histograms of oriented gradients for human detection [J].
Dalal, N ;
Triggs, B .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :886-893
[7]   YOLOv4 Object Detection Algorithm with Efficient Channel Attention Mechanism [J].
Gao, Cui ;
Cai, Qiang ;
Ming, Shaofeng .
2020 5TH INTERNATIONAL CONFERENCE ON MECHANICAL, CONTROL AND COMPUTER ENGINEERING (ICMCCE 2020), 2020, :1764-1770
[8]  
Ge Z, 2021, Arxiv, DOI arXiv:2107.08430
[9]  
Girshick R., 2015, ARXIV, DOI [DOI 10.1109/ICCV.2015.169, 10.1109/ICCV.2015.169]
[10]  
Girshick R, 2014, PROC CVPR IEEE, P580, DOI DOI 10.1109/CVPR.2014.81