Polyp-LVT: Polyp segmentation with lightweight vision transformers

被引:7
作者
Lin, Long [1 ]
Lv, Guangzu [1 ]
Wang, Bin [2 ]
Xu, Cunlu [1 ]
Liu, Jun [3 ]
机构
[1] Lanzhou Univ, Sch Informat Sci & Engn, Lanzhou, Gansu, Peoples R China
[2] Nanjing Univ Finance & Econ, Sch Informat Engn, Nanjing, Peoples R China
[3] Ulster Univ, Sch Comp, Belfast, North Ireland
关键词
Polyp segmentation; Lightweight vision transformer; Pooling layer; Colorectal cancer;
D O I
10.1016/j.knosys.2024.112181
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automatic segmentation of polyps in endoscopic images is crucial for early diagnosis and surgical planning of colorectal cancer. However, polyps closely resemble surrounding mucosal tissue in both texture and indistinct borders and vary in size, appearance, and location which possess great challenge to polyp segmentation. Although some recent attempts have been made to apply Vision Transformer (ViT) to polyp segmentation and achieved promising performance, their application in clinical scenarios is still limited by high computational complexity, large model size, redundant dependencies, and significant training costs. To address these limitations, we propose a novel ViT-based approach named Polyp-LVT, strategically replacing the attention layer in the encoder with a global max pooling layer, which significantly reduces the model's parameter count and computational cost while keeping the performance undegraded. Furthermore, we introduce a network block, named Inter-block Feature Fusion Module (IFFM), into the decoder, aiming to offer a streamlined yet highly efficient feature extraction. We conduct extensive experiments on three public polyp image benchmarks to evaluate our method. The experimental results show that compared with the baseline models, our PolypLVT network achieves a nearly 44% reduction in model parameters while gaining comparable segmentation performance.
引用
收藏
页数:10
相关论文
共 50 条
[31]   MMFIL-Net: Multi-level and multi-source feature interactive lightweight network for polyp segmentation [J].
Muhammad, Zaka-Ud-Din ;
Muhammad, Usman ;
Huang, Zhangjin ;
Gu, Naijie .
DISPLAYS, 2024, 81
[32]   Adaptive Context Exploration Network for Polyp Segmentation in Colonoscopy Images [J].
Yue, Guanghui ;
Li, Siying ;
Zhou, Tianwei ;
Wang, Miaohui ;
Du, Jingfeng ;
Jiang, Qiuping ;
Gao, Wei ;
Wang, Tianfu ;
Lv, Jun .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2023, 7 (02) :487-499
[33]   Highlighted Diffusion Model as Plug-In Priors for Polyp Segmentation [J].
Du, Yuhao ;
Jiang, Yuncheng ;
Tan, Shuangyi ;
Liu, Si-Qi ;
Li, Zhen ;
Li, Guanbin ;
Wan, Xiang .
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2025, 29 (02) :1209-1220
[34]   Polyp segmentation in medical imaging: challenges, approaches and future directions [J].
Qayoom, Abdul ;
Xie, Juanying ;
Ali, Haider .
ARTIFICIAL INTELLIGENCE REVIEW, 2025, 58 (06)
[35]   Automatic Polyp Segmentation with Multiple Kernel Dilated Convolution Network [J].
Tomar, Nikhil Kumar ;
Srivastava, Abhishek ;
Bagci, Ulas ;
Jha, Debesh .
2022 IEEE 35TH INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS (CBMS), 2022, :317-322
[36]   DCANet: deep context attention network for automatic polyp segmentation [J].
Zaka-Ud-Din Muhammad ;
Zhangjin Huang ;
Naijie Gu ;
Usman Muhammad .
The Visual Computer, 2023, 39 :5513-5525
[37]   Edge-aware Feature Aggregation Network for Polyp Segmentation [J].
Zhou, Tao ;
Zhang, Yizhe ;
Chen, Geng ;
Zhou, Yi ;
Wu, Ye ;
Fan, Deng-Ping .
MACHINE INTELLIGENCE RESEARCH, 2025, 22 (01) :101-116
[38]   DCANet: deep context attention network for automatic polyp segmentation [J].
Muhammad, Zaka-Ud-Din ;
Huang, Zhangjin ;
Gu, Naijie ;
Muhammad, Usman .
VISUAL COMPUTER, 2023, 39 (11) :5513-5525
[39]   PRAPNet: A Parallel Residual Atrous Pyramid Network for Polyp Segmentation [J].
Han, Jubao ;
Xu, Chao ;
An, Ziheng ;
Qian, Kai ;
Tan, Wei ;
Wang, Dou ;
Fang, Qianqian .
SENSORS, 2022, 22 (13)
[40]   Polyp Segmentation Network Combined With Multi-Attention Mechanism [J].
Jia L. ;
Hu Y. ;
Jin Y. ;
Xue Z. ;
Jiang Z. ;
Zheng Q. .
Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2023, 35 (03) :463-473