Medical image segmentation with an emphasis on prior convolution and channel multi-branch attention

被引:0
作者
Wang, Yuenan [1 ]
Wang, Hua [2 ]
Zhang, Fan [3 ,4 ]
机构
[1] Shandong Technol & Business Univ, Sch Informat & Elect Engn, Yantai 264005, Peoples R China
[2] Ludong Univ, Sch Informat & Elect Engn, Yantai 264025, Peoples R China
[3] Shandong Technol & Business Univ, Sch Comp Sci & Technol, Yantai 264005, Peoples R China
[4] Shandong Future Intelligent Financial Engn Lab, Yantai 264005, Peoples R China
基金
中国国家自然科学基金;
关键词
Transformer; Medical image segmentation; Attention mechanism; NETWORKS;
D O I
10.1016/j.dsp.2025.105175
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Transformer model has received extensive attention in recent years. Its powerful ability to handle contextual relationships makes it outstanding in the accurate segmentation of medical structures such as organs and lesions. However, as the Transformer model becomes more complex, its computational overhead has also increased significantly, becoming one of the key factors limiting the performance improvement of the model. In addition, some existing methods use channel dimensionality reduction to model cross-channel relationships. Although this strategy effectively reduces the amount of computation, it may lead to information loss or poor performance in segmentation tasks on medical images with rich details. To address the above problems, we propose an innovative medical image segmentation model, PCMA Former. This model combines convolution with focused weight reparameterization and a channel multi-branch attention mechanism, aiming to effectively improve model performance while maintaining low computational overhead. Through experimental verification on multiple medical image datasets (such as Synapse, ISIC2017, and ISIC2018), PCMA Former has achieved better results than traditional convolutional neural networks and existing Transformer models.
引用
收藏
页数:11
相关论文
共 64 条
  • [1] Alexey D, 2020, arXiv, DOI [arXiv:2010.11929, DOI 10.48550/ARXIV.2010.11929]
  • [2] Asadi-Aghbolaghi M, 2020, Arxiv, DOI arXiv:2003.05056
  • [3] Azad R, 2024, IEEE WINT CONF APPL, P1276, DOI 10.1109/WACV57701.2024.00132
  • [4] DAE-Former: Dual Attention-Guided Efficient Transformer for Medical Image Segmentation
    Azad, Reza
    Arimond, Rene
    Aghdam, Ehsan Khodapanah
    Kazerouni, Amirhossein
    Merhof, Dorit
    [J]. PREDICTIVE INTELLIGENCE IN MEDICINE, PRIME 2023, 2023, 14277 : 83 - 95
  • [5] Contextual Attention Network: Transformer Meets U-Net
    Azad, Reza
    Heidari, Moein
    Wu, Yuli
    Merhof, Dorit
    [J]. MACHINE LEARNING IN MEDICAL IMAGING, MLMI 2022, 2022, 13583 : 377 - 386
  • [6] TransDeepLab: Convolution-Free Transformer-Based DeepLab v3+for Medical Image Segmentation
    Azad, Reza
    Heidari, Moein
    Shariatnia, Moein
    Aghdam, Ehsan Khodapanah
    Karimijafarbigloo, Sanaz
    Adeli, Ehsan
    Merhof, Dorit
    [J]. PREDICTIVE INTELLIGENCE IN MEDICINE (PRIME 2022), 2022, 13564 : 91 - 102
  • [7] Bi-Directional ConvLSTM U-Net with Densley Connected Convolutions
    Azad, Reza
    Asadi-Aghbolaghi, Maryam
    Fathy, Mahmood
    Escalera, Sergio
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 406 - 415
  • [8] Cao Hu, 2023, Computer Vision - ECCV 2022 Workshops: Proceedings. Lecture Notes in Computer Science (13803), P205, DOI 10.1007/978-3-031-25066-8_9
  • [9] Chen J, 2021, arXiv, DOI DOI 10.48550/ARXIV.2102.04306
  • [10] SCA-CNN: Spatial and Channel-wise Attention in Convolutional Networks for Image Captioning
    Chen, Long
    Zhang, Hanwang
    Xiao, Jun
    Nie, Liqiang
    Shao, Jian
    Liu, Wei
    Chua, Tat-Seng
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6298 - 6306