ACU-TransNet: Attention and convolution-augmented UNet-transformer network for polyp segmentation

被引:1
作者
Huang, Lei [1 ,2 ]
Wu, Yun [1 ,2 ]
机构
[1] Guizhou Univ, State Key Lab Publ Big Data, Guiyang, Peoples R China
[2] Guizhou Univ, Coll Comp Sci & Technol, Guiyang, Peoples R China
基金
中国国家自然科学基金;
关键词
Polyp segmentation; UNet; transformer; deformable convolution; convolutional attention; ARCHITECTURE;
D O I
10.3233/XST-240076
中图分类号
TH7 [仪器、仪表];
学科分类号
0804 ; 080401 ; 081102 ;
摘要
BACKGROUND: UNet has achieved great success in medical image segmentation. However, due to the inherent locality of convolution operations, UNet is deficient in capturing global features and long-range dependencies of polyps, resulting in less accurate polyp recognition for complex morphologies and backgrounds. Transformers, with their sequential operations, are better at perceiving global features but lack low-level details, leading to limited localization ability. If the advantages of both architectures can be effectively combined, the accuracy of polyp segmentation can be further improved. METHODS: In this paper, we propose an attention and convolution-augmented UNet-Transformer Network (ACU-TransNet) for polyp segmentation. This network is composed of the comprehensive attention UNet and the Transformer head, sequentially connected by the bridge layer. On the one hand, the comprehensive attention UNet enhances specific feature extraction through deformable convolution and channel attention in the first layer of the encoder and achieves more accurate shape extraction through spatial attention and channel attention in the decoder. On the other hand, the Transformer head supplements fine-grained information through convolutional attention and acquires hierarchical global characteristics from the feature maps. RESULTS: mcU-TransNet could comprehensively learn dataset features and enhance colonoscopy interpretability for polyp detection. CONCLUSION: Experimental results on the CVC-ClinicDB and Kvasir-SEG datasets demonstrate that mcU-TransNet outperforms existing state-of-the-art methods, showcasing its robustness.
引用
收藏
页码:1449 / 1464
页数:16
相关论文
共 25 条
[1]  
Azad R, 2023, Arxiv, DOI [arXiv:2309.00121, 10.48550/arXiv.2309.00121]
[2]   WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians [J].
Bernal, Jorge ;
Javier Sanchez, F. ;
Fernandez-Esparrach, Gloria ;
Gil, Debora ;
Rodriguez, Cristina ;
Vilarino, Fernando .
COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2015, 43 :99-111
[3]  
Deng-Ping Fan, 2020, Medical Image Computing and Computer Assisted Intervention - MICCAI 2020. 23rd International Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12266), P263, DOI 10.1007/978-3-030-59725-2_26
[4]  
Gao H, 2020, Arxiv, DOI arXiv:1910.02940
[5]   Coordinate Attention for Efficient Mobile Network Design [J].
Hou, Qibin ;
Zhou, Daquan ;
Feng, Jiashi .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :13708-13717
[6]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
[7]   MultiResUNet : Rethinking the U-Net architecture for multimodal biomedical image segmentation [J].
Ibtehaz, Nabil ;
Rahman, M. Sohel .
NEURAL NETWORKS, 2020, 121 :74-87
[8]   CoInNet: A Convolution-Involution Network With a Novel Statistical Attention for Automatic Polyp Segmentation [J].
Jain, Samir ;
Atale, Rohan ;
Gupta, Anubhav ;
Mishra, Utkarsh ;
Seal, Ayan ;
Ojha, Aparajita ;
Jaworek-Korjakowska, Joanna ;
Krejcar, Ondrej .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2023, 42 (12) :3987-4000
[9]   Kvasir-SEG: A Segmented Polyp Dataset [J].
Jha, Debesh ;
Smedsrud, Pia H. ;
Riegler, Michael A. ;
Halvorsen, Pal ;
de Lange, Thomas ;
Johansen, Dag ;
Johansen, Havard D. .
MULTIMEDIA MODELING (MMM 2020), PT II, 2020, 11962 :451-462
[10]   ResUNet plus plus : An Advanced Architecture for Medical Image Segmentation [J].
Jha, Debesh ;
Smedsrud, Pia H. ;
Riegler, Michael A. ;
Johansen, Dag ;
de Lange, Thomas ;
Halvorsen, Pal ;
Johansen, Havard D. .
2019 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM 2019), 2019, :225-230