Fusion multi-scale Transformer skin lesion segmentation algorithm

被引:0
作者
Liang L.-M. [1 ]
Zhou L.-S. [1 ]
Yin J. [1 ]
Sheng X.-Q. [2 ]
机构
[1] School of Electrical Engineering and Automation, Jiangxi University of Science and Technology, Ganzhou
[2] School of Computer Science and Engineering, South China University of Technology, Guangzhou
来源
Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition) | 2024年 / 54卷 / 04期
关键词
channel attention module; computer application technology; image segmentation; multi-scale fusion module; skin lesions; transformer;
D O I
10.13229/j.cnki.jdxbgxb.20220692
中图分类号
学科分类号
摘要
To address the problem of lack of multi-scale feature extraction in existing skin lesion image segmentation,which leads to lack of detailed information and incorrect segmentation of skin lesion regions,this paper proposes a fusion multi-scale Transformer encoder-decoder network skin lesion segmentation algorithm. First,a hierarchical encoder is constructed using Transformer Block,which analyses the skin lesion region from the perspective of global feature variation at multiple scales. Then,the multi-scale fusion module,channel attention module and concat layer are used to construct the fusion decoder. The multi-scale fusion module fuses shallow network information and deep network information in the hierarchical encoder to enhance the dependency between spatial and semantic information,and the channel attention module can effectively identify channels containing rich feature information and improve the segmentation accuracy of the algorithm. Finally,an expansion module is introduced to recover the image size to meet the practical requirements. The proposed algorithm was experimentally tested on three public datasets,ISBI2016,ISBI2017 and ISIC2018. The pixel accuracies were 96.70%,94.50% and 95.39%,respectively,and the mean intersection over union were 91.69%,85.74% and 89.29%,respectively,with the overall performance of the tested algorithms outperforming existing algorithms.Simulation experiments show that the multi-scale Transformer encoder-decoder network can effectively segment skin lesion images,providing a new window for the diagnosis of modern skin diseases. © 2024 Editorial Board of Jilin University. All rights reserved.
引用
收藏
页码:1086 / 1098
页数:12
相关论文
共 27 条
  • [11] Dosovitskiy A, Beyer L, Kolesnikov A, Et al., An image is worth 16×16 words: Transformers for image recognition at scale
  • [12] Cao H, Wang Y, Chen J, Et al., Swin-Unet: Unet-like pure Transformer for medical image segmentation
  • [13] Wang W, Xie E, Li X, Et al., Pyramid vision Transformer: a Versatile backbone for dense prediction without convolutions, Proceedings of the IEEE/ CVF International Conference on Computer Vision, pp. 568-578, (2021)
  • [14] Xie E, Wang W, Yu Z, Et al., SegFormer: simple and efficient design for semantic segmentation with transformers
  • [15] Petit O, Thome N, Rambour C, Et al., U-net Transformer: self and cross attention for medical image segmentation[C], International Workshop on Machine Learning in Medical Imaging, pp. 267-276, (2021)
  • [16] Islam M A, Jia S, Bruce N D B., How much position information do convolutional neural networks encode?[J/ OL]
  • [17] Chu X, Tian Z, Zhang B, Et al., Conditional positional encodings for vision transformers
  • [18] Fu J, Liu J, Tian H, Et al., Dual attention network for scene segmentation, Proceedings of the IEEE/ CVF Conference on Computer Vision and Pattern Recognition, pp. 3146-3154, (2019)
  • [19] Gutman D, Codella N C F, Celebi E, Et al., Skin lesion analysis toward melanoma detection: a challenge at the international symposium on biomedical imaging (ISBI) 2016, hosted by the international skin imaging collaboration (ISIC) [J/OL]
  • [20] Codella N C F, Gutman D, Celebi M E, Et al., Skin lesion analysis toward melanoma detection: a challenge at the 2017 international symposium on biomedical imaging (ISBI), hosted by the international skin imaging collaboration (isic) [C], IEEE 15th International Symposium on Biomedical Imaging(ISBI 2018), pp. 168-172, (2018)