TransUMobileNet: Integrating multi-channel attention fusion with hybrid CNN-Transformer architecture for medical image segmentation

被引:0
作者
Cai, Sijing [1 ]
Jiang, Yukun [2 ]
Xiao, Yuwei [2 ]
Zeng, Jian [2 ]
Zhou, Guangming [2 ]
机构
[1] Fujian Univ Technol, Sch Transportat, Fuzhou 350108, Peoples R China
[2] Fujian Univ Technol, Sch Elect Elect Engn & Phys, Fuzhou 350108, Peoples R China
基金
中国国家自然科学基金;
关键词
Transformer; Unet; Medical image segmentation; Attention Mechanism; PLUS PLUS; NET; NETWORK;
D O I
10.1016/j.bspc.2025.107850
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
To address the common issue of strong similarity and blurred boundaries between lesion and normal tissues in medical images, we propose the TransUMobileNet model, which employs a symmetrical encoder-decoder structure. First, the feature encoder uses a hybrid CNN-Transformer architecture, where the Transformer encodes tokenized image patches from convolutional neural network (CNN) feature maps as input sequences for global context extraction. The Transformer's sequence prediction attention mechanism enhances the encoding of long-range dependencies and expressive learning, strengthening global information representation. Second, the feature decoder uses a fully symmetrical encoding form. Through symmetrical skip connections, the loss of positional information in the Transformer decoding path is mitigated, improving the depiction of target boundaries. The feature decoder utilizes cascaded upsampling to restore local spatial information and enhance finer details. Additionally, a Multi-Channel Attention Fusion (MCAF) module is incorporated into the decoding section. This module, characterized by a structure with small channels at both ends and a large one in the middle, along with an attention mechanism, enriches feature information and automatically adjusts weights for key regions, enhancing focus on target areas. TransUMobileNet was evaluated on three different public medical image segmentation datasets and a custom thyroid nodule segmentation dataset. The results show that TransUMobileNet achieves a recall rate of 82.23% and a mean average precision of 95.62%, outperforming current mainstream methods for medical image segmentation.
引用
收藏
页数:12
相关论文
共 42 条
  • [1] Alexey D, 2020, arXiv, DOI [arXiv:2010.11929, DOI 10.48550/ARXIV.2010.11929]
  • [2] Chen J, 2021, arXiv, DOI DOI 10.48550/ARXIV.2102.04306
  • [3] Chen LC, 2017, Arxiv, DOI arXiv:1706.05587
  • [4] DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs
    Chen, Liang-Chieh
    Papandreou, George
    Kokkinos, Iasonas
    Murphy, Kevin
    Yuille, Alan L.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) : 834 - 848
  • [5] Lightweight Frequency Recalibration Network for Diabetic Retinopathy Multi-Lesion Segmentation
    Fu, Yinghua
    Liu, Mangmang
    Zhang, Ge
    Peng, Jiansheng
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (16):
  • [6] TSCA-Net: Transformer based spatial-channel attention segmentation network for medical images
    Fu, Yinghua
    Liu, Junfeng
    Shi, Jun
    [J]. COMPUTERS IN BIOLOGY AND MEDICINE, 2024, 170
  • [7] RMCA U-net: Hard exudates segmentation for retinal fundus images
    Fu, Yinghua
    Zhang, Ge
    Lu, Xin
    Wu, Honghan
    Zhang, Dawei
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2023, 234
  • [8] Optic disc segmentation by U-net and probability bubble in abnormal fundus images
    Fu, Yinghua
    Chen, Jie
    Li, Jiang
    Pan, Dongyan
    Yue, Xuezheng
    Zhu, Yiming
    [J]. PATTERN RECOGNITION, 2021, 117
  • [9] ACPA-Net: Atrous Channel Pyramid Attention Network for Segmentation of Leakage in Rail Tunnel Linings
    Geng, Peng
    Tan, Ziye
    Luo, Jun
    Wang, Tongming
    Li, Feng
    Bei, Jianghui
    [J]. ELECTRONICS, 2023, 12 (02)
  • [10] CE-Net: Context Encoder Network for 2D Medical Image Segmentation
    Gu, Zaiwang
    Cheng, Jun
    Fu, Huazhu
    Zhou, Kang
    Hao, Huaying
    Zhao, Yitian
    Zhang, Tianyang
    Gao, Shenghua
    Liu, Jiang
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2019, 38 (10) : 2281 - 2292