Laser Stripe Segmentation of Weld Seam Based on CNN-Transformer Hybrid Networks

被引:0
|
作者
Wang, Ying [1 ]
Gao, Sheng [2 ]
Dai, Zhe [1 ]
机构
[1] Northeast Petr Univ, Sch Comp & Informat Technol, Daqing 163318, Heilongjiang, Peoples R China
[2] Northeast Petr Univ, Sch Mech Sci & Engn, Daqing 163318, Heilongjiang, Peoples R China
来源
关键词
laser stripe segmentation; semantic segmentation; Transformer; Mobile ViT block;
D O I
10.3788/CJL240710
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Objective The challenging conditions at welding construction sites-such as uneven weldment surfaces, complex bevel shapes due to the front weld channel, loss of centerline information, smoke, spatter, intense arc light, and overlapping reflections-hinder real- time and accurate tracking and control during the welding process. Projecting a laser onto the weldment surface, using a vision sensor to capture the laser streak image at the bevel, and then using the identified key point of the laser streak as the basis for weld positioning has become the most widely applied method for tracking complex weld seams. Therefore, accurately segmenting multi- layer multi- pass weld laser stripes against a complex background is a key problem in intelligent welding processes. This study proposes a lightweight weld laser stripe segmentation method based on a convolutional neural network (CNN)- Transformer hybrid network to improve the segmentation accuracy and real-time performance by acquiring fine-grained features and recognizing subtle differences, thereby enabling the tracking of complex multi- layer multi- pass welds in high- noise environments. Methods This study develops a hybrid CNN- Transformer model for weld laser streak segmentation. The encoder part of the model uses the Mobile ViT module, which has a smaller number of parameters and demands less computation, for feature extraction. It also embeds a dual non- local block (DNB) module to capture the long-distance correlation relationship on the spatial and channel domains of the weld image, which ensures feature extraction capability and improves the segmentation efficiency simultaneously. The decoder part of the model uses efficient sub- pixel convolutional neural network (ESPCN) to obtain semantic segmentation results, which reduces the feature loss in the information reconstruction process and improves the model performance in extracting laser lines from weld seams. To address the imbalance between laser- streak and background pixels in the weld image, a loss function that dynamically adjusts the weighting coefficients of laser streaks is proposed. Results and Discussions Ablation test results show that the introduction of the DNB module for feature extraction enriches the semantic information in weld laser streak images, and the adoption of the ESPCN implementation reduces the loss of weld laser streak information (Table 2). The results of loss function tests and related comparisons show that the dynamically generated weighted coefficient loss function proposed in this study can well solve the problem of pixel imbalance in weld laser streak images (Table 3). Tests and comparisons of the loss function demonstrate that the dynamically generated weighted coefficient loss function effectively addresses the pixel imbalance in weld laser streak images (Table 3). Testing and comparing different segmentation models reveal that the proposed CNN- Transformer hybrid network model is advantageous in accuracy and speed, achieving the highest pixel accuracy (PA), mean pixel accuracy (mPA), and mean intersection over union (mIoU) while retaining its lightweight computation (Table 4). Training results of the 20th round for different segmentation models indicate that the laser stripe line contour obtained by this model is clearer and closer to the labeled image (Fig. 11). Conclusions Addressing issues of incomplete, low- precision, and slow weld laser stripe segmentation caused by various factors at welding construction sites-such as harsh conditions, uneven weldment surfaces, complex bevel shapes due to the front weld channel, loss of centerline information, and numerous noises-a weld laser stripe segmentation model based on a CNN- Transformer hybrid network is established. Using the same dataset, experimental setup, and loss function, the proposed model outperforms commonly used lightweight semantic segmentation networks such as Unet, Deeplabv3+ , SegNet, PSPNet, RefineNet, and FCN- 32s in both accuracy and processing speed. As a segmentation network, the model employs different loss functions for experiments, with the improved loss function effectively addressing the imbalance between laser- stripe and background pixels, achieving the highest recognition accuracy and fastest convergence speed. The small size and low computational complexity of the proposed model, with a single image inference time of 40 ms and a pixel accuracy of 98%, meet the requirements for lightweight, high- precision, and low- latency vision tasks on resource- constrained mobile devices.
引用
收藏
页数:12
相关论文
共 29 条
  • [11] Plate additive, seam-tracking technology based on feature segmentation
    Lu, Jun
    Zhang, Jun
    Luo, Jun
    Yang, Aodong
    Han, Jing
    Zhao, Zhuang
    [J]. OPTICS AND LASER TECHNOLOGY, 2024, 168
  • [12] An Efficient and Robust Complex Weld Seam Feature Point Extraction Method for Seam Tracking and Posture Adjustment
    Ma, Yunkai
    Fan, Junfeng
    Yang, Huizhen
    Wang, Hongliang
    Xing, Shiyu
    Jing, Fengshui
    Tan, Min
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (11) : 10704 - 10715
  • [13] Mehta S, 2022, Arxiv, DOI [arXiv:2110.02178, 10.48550/arXiv.2110.02178, DOI 10.48550/ARXIV.2110.02178]
  • [14] Development of a vision system integrated with industrial robots for online weld seam tracking
    Nguyen, Quoc-Chi
    Hua, Hoang Quoc Bao
    Pham, Phuong-Tung
    [J]. JOURNAL OF MANUFACTURING PROCESSES, 2024, 119 : 414 - 424
  • [15] Weld Seam Image Processing Based on Improved U-net Neural Network
    Shen J.
    Liu C.
    Hu S.
    Yang H.
    Zhang C.
    Wu D.
    [J]. Tianjin Daxue Xuebao (Ziran Kexue yu Gongcheng Jishu Ban)/Journal of Tianjin University Science and Technology, 2023, 56 (04): : 436 - 442
  • [16] Three Things Everyone Should Know About Vision Transformers
    Touvron, Hugo
    Cord, Matthieu
    El-Nouby, Alaaeldin
    Verbeek, Jakob
    Jegou, Herve
    [J]. COMPUTER VISION, ECCV 2022, PT XXIV, 2022, 13684 : 497 - 515
  • [17] Robust weld seam tracking method based on detection and tracking of laser stripe
    Wang, Nianfeng
    Yang, Jialin
    Zhang, Xianmin
    Gong, Tao
    Zhong, Kaifan
    [J]. INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2024, 130 (7-8): : 3481 - 3493
  • [18] Autonomous seam recognition and feature extraction for multi-pass welding based on laser stripe edge guidance network
    Wu, Kaixuan
    Wang, Tianqi
    He, Junjie
    Liu, Yang
    Jia, Zhenwei
    [J]. INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2020, 111 (9-10): : 2719 - 2731
  • [19] LSFP-Tracker: An Autonomous Laser Stripe Feature Point Extraction Algorithm Based on Siamese Network for Robotic Welding Seam Tracking
    Xiao, Runquan
    Xu, Yanling
    Xu, Fengjing
    Hou, Zhen
    Zhang, Huajun
    Chen, Shanben
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (01) : 1037 - 1048
  • [20] A High-Speed Seam Extraction Method Based on the Novel Structured-Light Sensor for Arc Welding Robot: A Review
    Yang, Lei
    Li, En
    Long, Teng
    Fan, Junfeng
    Liang, Zize
    [J]. IEEE SENSORS JOURNAL, 2018, 18 (21) : 8631 - 8641