A lightweight transformer with linear self-attention for defect recognition

被引:1
|
作者
Zhai, Yuwen [1 ]
Li, Xinyu [1 ]
Gao, Liang [1 ]
Gao, Yiping [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Mech Sci & Engn, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
automatic optical inspection; convolutional neural nets; image recognition; object detection;
D O I
10.1049/ell2.13292
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Visual defect recognition techniques based on deep learning models are crucial for modern industrial quality inspection. The backbone, serving as the primary feature extraction component of the defect recognition model, has not been thoroughly exploited. High-performance vision transformer (ViT) is less adopted due to high computational complexity and limitations of computational resources and storage hardware in industrial scenarios. This paper presents LSA-Former, a lightweight transformer architectural backbone that integrates the benefits of convolution and ViT. LSA-Former proposes a novel self-attention with linear computational complexity, enabling it to capture local and global semantic features with fewer parameters. LSA-Former is pre-trained on ImageNet-1K and surpasses state-of-the-art methods. LSA-Former is employed as the backbone for various detectors, evaluated specifically on the PCB defect detection task. The proposed method reduces at least 18M parameters and exceeds the baseline by more than 2.2 mAP. This paper presents LSA-Former, a lightweight transformer architectural backbone that integrates the benefits of convolution and ViT. LSA-Former proposes a novel self-attention with linear computational complexity. LSA- Former is pre-trained on large-scale data and its practicality is validated in defect detection tasks. image
引用
收藏
页数:4
相关论文
共 50 条
  • [1] Densely Connected Transformer With Linear Self-Attention for Lightweight Image Super-Resolution
    Zeng, Kun
    Lin, Hanjiang
    Yan, Zhiqiang
    Fang, Jinsheng
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [2] A lightweight and rapidly converging transformer based on separable linear self-attention for fault diagnosis
    Yin, Kexin
    Chen, Chunjun
    Shen, Qi
    Deng, Ji
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2025, 36 (01)
  • [3] Lightweight Vision Transformer with Spatial and Channel Enhanced Self-Attention
    Zheng, Jiahao
    Yang, Longqi
    Li, Yiying
    Yang, Ke
    Wang, Zhiyuan
    Zhou, Jun
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1484 - 1488
  • [4] Lightweight Smoke Recognition Based on Deep Convolution and Self-Attention
    Zhao, Yang
    Wang, Yigang
    Jung, Hoi-Kyung
    Jin, Yongqiang
    Hua, Dan
    Xu, Sen
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2022, 2022
  • [5] MULTI-VIEW SELF-ATTENTION BASED TRANSFORMER FOR SPEAKER RECOGNITION
    Wang, Rui
    Ao, Junyi
    Zhou, Long
    Liu, Shujie
    Wei, Zhihua
    Ko, Tom
    Li, Qing
    Zhang, Yu
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6732 - 6736
  • [6] Relative molecule self-attention transformer
    Łukasz Maziarka
    Dawid Majchrowski
    Tomasz Danel
    Piotr Gaiński
    Jacek Tabor
    Igor Podolak
    Paweł Morkisz
    Stanisław Jastrzębski
    Journal of Cheminformatics, 16
  • [7] Relative molecule self-attention transformer
    Maziarka, Lukasz
    Majchrowski, Dawid
    Danel, Tomasz
    Gainski, Piotr
    Tabor, Jacek
    Podolak, Igor
    Morkisz, Pawel
    Jastrzebski, Stanislaw
    JOURNAL OF CHEMINFORMATICS, 2024, 16 (01)
  • [8] Lightweight Causal Transformer with Local Self-Attention for Real-Time Speech Enhancement
    Oostermeijer, Koen
    Wang, Qing
    Du, Jun
    INTERSPEECH 2021, 2021, : 2831 - 2835
  • [9] SALSA: Swift Adaptive Lightweight Self-Attention for Enhanced LiDAR Place Recognition
    Goswami, Raktim Gautam
    Patel, Naman
    Krishnamurthy, Prashanth
    Khorrami, Farshad
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (10): : 8242 - 8249
  • [10] Attention to Emotions: Body Emotion Recognition In-the-Wild Using Self-attention Transformer Network
    Paiva, Pedro V. V.
    Ramos, Josue J. G.
    Gavrilova, Marina
    Carvalho, Marco A. G.
    COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS, VISIGRAPP 2023, 2024, 2103 : 206 - 228