ST-Unet: Swin Transformer boosted U-Net with Cross-Layer Feature Enhancement for medical image segmentation

被引:53
作者
Zhang, Jing [1 ]
Qin, Qiuge [1 ]
Ye, Qi [1 ]
Ruan, Tong [1 ]
机构
[1] East China Univ Sci & Technol, Dept Comp Sci & Engn, Shanghai 200237, Peoples R China
关键词
Medical image segmentation; Swin Transformer; ST-Unet; Cross-layer feature enhancement; BOUNDARY; NODULES;
D O I
10.1016/j.compbiomed.2022.106516
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Medical image segmentation is an essential task in clinical diagnosis and case analysis. Most of the existing methods are based on U-shaped convolutional neural networks (CNNs), and one of disadvantages is that the long-term dependencies and global contextual connections cannot be effectively established, which results in inaccuracy segmentation. For fully using low-level features to enhance global features and reduce the semantic gap between encoding and decoding stages, we propose a novel Swin Transformer boosted U-Net (ST-Unet) for medical image processing in this paper, in which Swin Transformer and CNNs are used as encoder and decoder respectively. Then a novel Cross-Layer Feature Enhancement (CLFE) module is proposed to realize cross-layer feature learning, and a Spatial and Channel Squeeze & Excitation module is adopted to highlight the saliency of specific regions. Finally, we learn the features fused by the CLFE module through CNNs to recover low-level features and localize local features for realizing more accurate semantic segmentation. Experiments on widely used public datasets Synapse and ISIC 2018 prove that our proposed ST-Unet can achieve 78.86 of dice and 0.9243 of recall performance, outperforming most current medical image segmentation methods.
引用
收藏
页数:11
相关论文
共 61 条
  • [1] Integrating semantic edges and segmentation information for building extraction from aerial images using UNet
    Abdollahi, Abolfazl
    Pradhan, Biswajeet
    [J]. MACHINE LEARNING WITH APPLICATIONS, 2021, 6
  • [2] Recurrent residual U-Net for medical image segmentation
    Alom, Md Zahangir
    Yakopcic, Chris
    Hasan, Mahmudul
    Taha, Tarek M.
    Asari, Vijayan K.
    [J]. JOURNAL OF MEDICAL IMAGING, 2019, 6 (01)
  • [3] Bi-Directional ConvLSTM U-Net with Densley Connected Convolutions
    Azad, Reza
    Asadi-Aghbolaghi, Maryam
    Fathy, Mahmood
    Escalera, Sergio
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 406 - 415
  • [4] CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification
    Chen, Chun-Fu
    Fan, Quanfu
    Panda, Rameswar
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 347 - 356
  • [5] Chen J, 2021, arXiv
  • [6] DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs
    Chen, Liang-Chieh
    Papandreou, George
    Kokkinos, Iasonas
    Murphy, Kevin
    Yuille, Alan L.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) : 834 - 848
  • [7] Codella N, 2019, Arxiv, DOI arXiv:1902.03368
  • [8] Segmentation of pulmonary nodules in thoracic CT scans: A region growing approach
    Dehmeshi, Jamshid
    Amin, Hamdan
    Valdivieso, Manlio
    Ye, Xujiong
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2008, 27 (04) : 467 - 480
  • [9] Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
  • [10] Medical image analysis: Progress over two decades and the challenges ahead
    Duncan, JS
    Ayache, N
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2000, 22 (01) : 85 - 106