Laser Stripe Segmentation of Weld Seam Based on CNN-Transformer Hybrid Networks

被引:0
|
作者
Wang, Ying [1 ]
Gao, Sheng [2 ]
Dai, Zhe [1 ]
机构
[1] Northeast Petr Univ, Sch Comp & Informat Technol, Daqing 163318, Heilongjiang, Peoples R China
[2] Northeast Petr Univ, Sch Mech Sci & Engn, Daqing 163318, Heilongjiang, Peoples R China
来源
关键词
laser stripe segmentation; semantic segmentation; Transformer; Mobile ViT block;
D O I
10.3788/CJL240710
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Objective The challenging conditions at welding construction sites-such as uneven weldment surfaces, complex bevel shapes due to the front weld channel, loss of centerline information, smoke, spatter, intense arc light, and overlapping reflections-hinder real- time and accurate tracking and control during the welding process. Projecting a laser onto the weldment surface, using a vision sensor to capture the laser streak image at the bevel, and then using the identified key point of the laser streak as the basis for weld positioning has become the most widely applied method for tracking complex weld seams. Therefore, accurately segmenting multi- layer multi- pass weld laser stripes against a complex background is a key problem in intelligent welding processes. This study proposes a lightweight weld laser stripe segmentation method based on a convolutional neural network (CNN)- Transformer hybrid network to improve the segmentation accuracy and real-time performance by acquiring fine-grained features and recognizing subtle differences, thereby enabling the tracking of complex multi- layer multi- pass welds in high- noise environments. Methods This study develops a hybrid CNN- Transformer model for weld laser streak segmentation. The encoder part of the model uses the Mobile ViT module, which has a smaller number of parameters and demands less computation, for feature extraction. It also embeds a dual non- local block (DNB) module to capture the long-distance correlation relationship on the spatial and channel domains of the weld image, which ensures feature extraction capability and improves the segmentation efficiency simultaneously. The decoder part of the model uses efficient sub- pixel convolutional neural network (ESPCN) to obtain semantic segmentation results, which reduces the feature loss in the information reconstruction process and improves the model performance in extracting laser lines from weld seams. To address the imbalance between laser- streak and background pixels in the weld image, a loss function that dynamically adjusts the weighting coefficients of laser streaks is proposed. Results and Discussions Ablation test results show that the introduction of the DNB module for feature extraction enriches the semantic information in weld laser streak images, and the adoption of the ESPCN implementation reduces the loss of weld laser streak information (Table 2). The results of loss function tests and related comparisons show that the dynamically generated weighted coefficient loss function proposed in this study can well solve the problem of pixel imbalance in weld laser streak images (Table 3). Tests and comparisons of the loss function demonstrate that the dynamically generated weighted coefficient loss function effectively addresses the pixel imbalance in weld laser streak images (Table 3). Testing and comparing different segmentation models reveal that the proposed CNN- Transformer hybrid network model is advantageous in accuracy and speed, achieving the highest pixel accuracy (PA), mean pixel accuracy (mPA), and mean intersection over union (mIoU) while retaining its lightweight computation (Table 4). Training results of the 20th round for different segmentation models indicate that the laser stripe line contour obtained by this model is clearer and closer to the labeled image (Fig. 11). Conclusions Addressing issues of incomplete, low- precision, and slow weld laser stripe segmentation caused by various factors at welding construction sites-such as harsh conditions, uneven weldment surfaces, complex bevel shapes due to the front weld channel, loss of centerline information, and numerous noises-a weld laser stripe segmentation model based on a CNN- Transformer hybrid network is established. Using the same dataset, experimental setup, and loss function, the proposed model outperforms commonly used lightweight semantic segmentation networks such as Unet, Deeplabv3+ , SegNet, PSPNet, RefineNet, and FCN- 32s in both accuracy and processing speed. As a segmentation network, the model employs different loss functions for experiments, with the improved loss function effectively addressing the imbalance between laser- stripe and background pixels, achieving the highest recognition accuracy and fastest convergence speed. The small size and low computational complexity of the proposed model, with a single image inference time of 40 ms and a pixel accuracy of 98%, meet the requirements for lightweight, high- precision, and low- latency vision tasks on resource- constrained mobile devices.
引用
收藏
页数:12
相关论文
共 29 条
  • [1] Weld Structured Light Image Segmentation Based on Lightweight DeepLab v3+Network
    Bing, Chen
    Sheng, He
    Jian, Liu
    Chen Shengfeng
    Lu Enhui
    [J]. CHINESE JOURNAL OF LASERS-ZHONGGUO JIGUANG, 2023, 50 (08):
  • [2] [陈海永 Chen Haiyong], 2018, [焊接学报, Transactions of the China Welding Institution], V39, P19
  • [3] Cheng Jiaming, 2023, Journal of Southeast University (Natural Science Edition), P86, DOI 10.3969/j.issn.1001-0505.2023.01.011
  • [4] Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
  • [5] Weld Path Identification Based on Kernel Correlation/Kalman Filters
    Du Jianzhun
    Zhang Yanxi
    Wang Jingjing
    Gao Xiangdong
    [J]. CHINESE JOURNAL OF LASERS-ZHONGGUO JIGUANG, 2022, 49 (02):
  • [6] Discontinuous Fold Line Welding Seam Recognition and Mobile Robot Tracking System in Narrow Space
    Guo L.
    Zhang H.
    [J]. Jixie Gongcheng Xuebao/Journal of Mechanical Engineering, 2019, 55 (17): : 8 - 13
  • [7] Laser Fringe Segmentation and Feature Points Location Method of Weld Image Based on Multi-Task Learning
    Huang, Yigeng
    Wang, Daqing
    Jiang, Man
    Yin, Haoyu
    Gao, Lifu
    [J]. CHINESE JOURNAL OF LASERS-ZHONGGUO JIGUANG, 2023, 50 (16):
  • [8] Rethinking the Self-Attention in Vision Transformers
    Kim, Kyungmin
    Wu, Bichen
    Dai, Xiaoliang
    Zhang, Peizhao
    Yan, Zhicheng
    Vajda, Peter
    Kim, Seon
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3065 - 3069
  • [9] A lightweight vision transformer with symmetric modules for vision tasks
    Liang, Shengjun
    Yu, Mingxin
    Lu, Wenshuai
    Ji, Xinglong
    Tang, Xiongxin
    Liu, Xiaolin
    You, Rui
    [J]. INTELLIGENT DATA ANALYSIS, 2023, 27 (06) : 1741 - 1757
  • [10] Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
    Liu, Ze
    Lin, Yutong
    Cao, Yue
    Hu, Han
    Wei, Yixuan
    Zhang, Zheng
    Lin, Stephen
    Guo, Baining
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 9992 - 10002