FMCA-Net: A feature secondary multiplexing and dilated convolutional attention polyp segmentation network based on pyramid vision transformer

被引:4
作者
Li, Weisheng [1 ,2 ,3 ]
Nie, Xiaolong [1 ,2 ]
Li, Feiyan [1 ,2 ]
Huang, Zhaopeng [1 ,2 ]
Zeng, Guofeng [1 ,2 ]
机构
[1] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Sch Comp Sci & Technol, Chongqing 400065, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Key Lab Cyberspace Big Data Intelligent Secur, Minist Educ, Chongqing 400065, Peoples R China
关键词
Polyp segmentation; Feature secondary reuse; Branch attention; Feature sharing; NEURAL-NETWORKS;
D O I
10.1016/j.eswa.2024.125419
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Polyp segmentation is of great significance in diagnosing and treating related symptoms. Existing polyp segmentation methods have performed well in solving the problems of intra-polyp inconsistency and inter- polyp distinguishability. However, three shortcomings still exist: (1) The decoder does not fully use the initially extracted polyp features. (2) The segmentation edges are fuzzy, and the boundaries are unclear. (3) The network structure is becoming increasingly complex and needs to be clarified. We propose a feature secondary reuse and hole convolutional attention network (FMCA-Net) based on a Pyramid Vision Transformer to solve these problems. Firstly, we propose a feature secondary reuse module (D-BFRM) to process the polyp features of different scales initially extracted in the encoder. After two stages of reuse processing, they are used as references for the remaining branches. This way, feature information such as polyp size, shape, and number can be fully obtained, ensuring the model's fitting ability. Secondly, we also propose a dilated convolutional attention module group (DCBA&DCGA), in which DCBA is used to process each branch's features further. In contrast, DCGA processes the final global features to distinguish the boundaries between polyps and backgrounds further and improve the model's overall generalization ability. Finally, we use the idea of modularization in the model to make the structure more concise and clear. We objectively evaluate the proposed method on five public polyp segmentation datasets. The experimental results show that FMCANet has excellent learning and fitting ability and strong generalization ability. At the same time, the idea of modularization also has obvious advantages in the simplicity and clarity of the model structure.
引用
收藏
页数:14
相关论文
共 46 条
[1]   Understanding Robustness of Transformers for Image Classification [J].
Bhojanapalli, Srinadh ;
Chakrabarti, Ayan ;
Glasner, Daniel ;
Li, Daliang ;
Unterthiner, Thomas ;
Veit, Andreas .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :10211-10221
[2]  
Cao Hu, 2023, Computer Vision - ECCV 2022 Workshops: Proceedings. Lecture Notes in Computer Science (13803), P205, DOI 10.1007/978-3-031-25066-8_9
[3]   GCNet: Non-local Networks Meet Squeeze-Excitation Networks and Beyond [J].
Cao, Yue ;
Xu, Jiarui ;
Lin, Stephen ;
Wei, Fangyun ;
Hu, Han .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, :1971-1980
[4]   HarDNet: A Low Memory Traffic Network [J].
Chao, Ping ;
Kao, Chao-Yang ;
Ruan, Yu-Shan ;
Huang, Chien-Hsiang ;
Lin, Youn-Long .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3551-3560
[5]  
Chen J., 2021, arXiv, DOI DOI 10.48550/ARXIV.2102.04306
[6]  
Chen LC, 2017, Arxiv, DOI [arXiv:1706.05587, DOI 10.48550/ARXIV.1706.05587, 10.48550/arXiv.1706.05587]
[7]  
Deng-Ping Fan, 2020, Medical Image Computing and Computer Assisted Intervention - MICCAI 2020. 23rd International Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12266), P263, DOI 10.1007/978-3-030-59725-2_26
[8]  
Dong B, 2024, Arxiv, DOI [arXiv:2108.06932, 10.26599/AIR.2023.9150015, DOI 10.48550/ARXIV.2108.06932]
[9]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[10]  
ELKarazle K, 2023, Arxiv, DOI arXiv:2310.05446