Improved YOLOV4-CSP Algorithm for Detection of Bamboo Surface Sliver Defects With Extreme Aspect Ratio

被引:18
作者
Guo, Yijing [1 ]
Zeng, Yixin [2 ]
Gao, Fengqiang [3 ]
Qiu, Yi [1 ]
Zhou, Xuqiang [2 ]
Zhong, Linwei [2 ]
Zhan, Choujun [1 ]
机构
[1] Xiamen Univ, Tan Kah Kee Coll, Sch Informat Sci & Technol, Zhangzhou 363105, Peoples R China
[2] Xiamen Univ, Sch Informat, Xiamen 361005, Peoples R China
[3] Xiamen Univ, Sch Aerosp Engn, Xiamen 361005, Peoples R China
来源
IEEE ACCESS | 2022年 / 10卷
基金
美国国家科学基金会;
关键词
Bamboo; Feature extraction; Detectors; Strips; Inspection; Deep learning; Surface treatment; Bamboo defect; defect inspection; object detection; asymmetric convolution; attention mechanism;
D O I
10.1109/ACCESS.2022.3152552
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Bamboo surface defect detection provides quality assurance for bamboo product manufacture in industrial scenarios, an integral part of the overall manufacturing process. Currently, bamboo defect inspection predominantly relies on manual operation, but manual inspection is very time-consuming as well as labor-intensive, and the quality of inspection is not guaranteed. A few visual inspection systems based on traditional image processing have been deployed in some factories in recent years. However, traditional machine vision algorithms extract features in tedious steps and have poor performance along with poor adaptability in the face of complex defects. Accordingly, many scholars are committed to seeking deep learning methods to accomplish surface defect detection. However, existing deep learning object detectors struggle with specific industrial defects when directly applied to industrial defect detection, such as sliver defects, especially for ones with extreme aspect ratios. To this end, this paper proposes an improved algorithm based on the advanced object detector YOLOV4-CSP, which introduces asymmetric convolution and attention mechanism. The introduction of asymmetric convolution enhances the feature extraction in the horizontal direction of the bamboo strip surface, improving the performance in detecting sliver defects. In addition, convolutional block attention module(CBAM), a hybrid attention module, which combines channel attention with spatial attention, is utilized to promote the representation ability of the model by increasing the weights of crucial channels and regions. The proposed model achieves outstanding performance in the general categories and excels in the hard-to-detect categories. Some enterprise's bamboo strip dataset experiments verify that the model can reach 96.74% mAP for the typical six surface defects. Meanwhile, we also observe significant improvements when extending our model to aluminum datasets with similar characteristics.
引用
收藏
页码:29810 / 29820
页数:11
相关论文
共 42 条
  • [1] Arthur D, 2007, PROCEEDINGS OF THE EIGHTEENTH ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, P1027
  • [2] Bochkovskiy A., 2020, ARXIV 200410934
  • [3] Chen L., 2017, J GUANGXI U, V42, P1069
  • [4] Control of goal-directed and stimulus-driven attention in the brain
    Corbetta, M
    Shulman, GL
    [J]. NATURE REVIEWS NEUROSCIENCE, 2002, 3 (03) : 201 - 215
  • [5] Deformable Convolutional Networks
    Dai, Jifeng
    Qi, Haozhi
    Xiong, Yuwen
    Li, Yi
    Zhang, Guodong
    Hu, Han
    Wei, Yichen
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 764 - 773
  • [6] Denton E, 2014, ADV NEUR IN, V27
  • [7] ACNet: Strengthening the Kernel Skeletons for Powerful CNN via Asymmetric Convolution Blocks
    Ding, Xiaohan
    Guo, Yuchen
    Ding, Guiguang
    Han, Jungong
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 1911 - 1920
  • [8] CenterNet: Keypoint Triplets for Object Detection
    Duan, Kaiwen
    Bai, Song
    Xie, Lingxi
    Qi, Honggang
    Huang, Qingming
    Tian, Qi
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6568 - 6577
  • [9] Gao Q., 2020, COMPUT APPL, V41, P1
  • [10] Fast R-CNN
    Girshick, Ross
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1440 - 1448