Bilateral guidance network for one-shot metal defect segmentation

被引:3
作者
Shan, Dexing [1 ]
Zhang, Yunzhou [1 ]
Liu, Xiaozheng [1 ]
Zhao, Jiaqi [1 ]
Coleman, Sonya [2 ]
Kerr, Dermot [2 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang, Liaoning, Peoples R China
[2] Univ Ulster, Intelligent Syst Res Ctr, Londonderry, England
关键词
Neural network application; One-shot metal defect segmentation; Affinity learning; Bilateral guidance mechanism; Cross-attention;
D O I
10.1016/j.engappai.2023.107802
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Metal defect inspection is critical for maintaining product quality and ensuring production safety. However, the vast majority of existing defect segmentation methods rely heavily on large-scale datasets that only cater to specific defects, making them unsuitable for the industrial sector, where training samples are often limited. To address these challenges, we propose a bilateral guidance network for one-shot metal defect segmentation that leverages the perceptual consistency of background regions within industrial images to distinguish foreground and background regions. Our model uses an interactive feature reweighting scheme that models the inter- and self -dependence of foreground and background feature maps, enabling us to build robust pixel -level correspondences. Our proposed method demonstrates good domain adaptability and accurately segments defects in multiple materials, such as steel, leather, and carpet, among others. Additionally, we have incorporated a multi -scale receptive field encoder to enhance the model's ability to perceive objects of varying scales, providing a comprehensive solution for industrial defect segmentation. Experimental results indicate that our proposed method has the potential to be effective in a variety of real -world applications where defects may not be immediately visible or where large amounts of labeled data are not readily available. With only one shot, our method achieves the state-of-the-art performance of 41.62% mIoU and 70.30% MPA on the Defect-3' dataset.
引用
收藏
页数:11
相关论文
共 41 条
[1]   Trainable guided attention based robust leather defect detection [J].
Aslam, Masood ;
Naqvi, Syed Saud ;
Khan, Tariq Mahmood ;
Holmes, Geoff ;
Naffa, Rafea .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 124
[2]   Triplet-Graph Reasoning Network for Few-Shot Metal Generic Surface Defect Segmentation [J].
Bao, Yanqi ;
Song, Kechen ;
Liu, Jie ;
Wang, Yanyan ;
Yan, Yunhui ;
Yu, Han ;
Li, Xingjie .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2021, 70
[3]   MVTec AD - A Comprehensive Real-World Dataset for Unsupervised Anomaly Detection [J].
Bergmann, Paul ;
Fauser, Michael ;
Sattlegger, David ;
Steger, Carsten .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9584-9592
[4]   A Pixel-Level Segmentation Convolutional Neural Network Based on Deep Feature Fusion for Surface Defect Detection [J].
Cao, Jingang ;
Yang, Guotian ;
Yang, Xiyun .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2021, 70
[5]  
Chen LC, 2017, Arxiv, DOI arXiv:1706.05587
[6]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851
[7]   Cross-Domain Few-Shot Classification based on Lightweight Res2Net and Flexible GNN [J].
Chen, Yu ;
Zheng, Yunan ;
Xu, Zhenyu ;
Tang, Tianhang ;
Tang, Zixin ;
Chen, Jie ;
Liu, Yiguang .
KNOWLEDGE-BASED SYSTEMS, 2022, 247
[8]   PGA-Net: Pyramid Feature Fusion and Global Context Attention Network for Automated Surface Defect Detection [J].
Dong, Hongwen ;
Song, Kechen ;
He, Yu ;
Xu, Jing ;
Yan, Yunhui ;
Meng, Qinggang .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2020, 16 (12) :7448-7458
[9]  
Dosovitskiy A, 2021, Arxiv, DOI [arXiv:2010.11929, 10.48550/arXiv.2010.11929, DOI 10.48550/ARXIV.2010.11929]
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778