Know your orientation: A viewpoint-aware framework for polyp segmentation

被引:4
作者
Cai, Linghan [1 ,2 ]
Chen, Lijiang [2 ]
Huang, Jianhao [1 ]
Wang, Yifeng [3 ]
Zhang, Yongbing [1 ]
机构
[1] Harbin Inst Technol Shenzhen, Sch Comp Sci & Technol, Shenzhen 518055, Peoples R China
[2] Beihang Univ, Dept Elect Informat Engn, Beijing 100191, Peoples R China
[3] Harbin Inst Technol Shenzhen, Sch Sci, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Colonoscopy image; Polyp segmentation; Viewpoint classification; Viewpoint-aware transformer; Boundary-aware transformer; DIAGNOSIS; ATTENTION; IMAGES; MAPS;
D O I
10.1016/j.media.2024.103288
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automatic polyp segmentation in endoscopic images is critical for the early diagnosis of colorectal cancer. Despite the availability of powerful segmentation models, two challenges still impede the accuracy of polyp segmentation algorithms. Firstly, during a colonoscopy, physicians frequently adjust the orientation of the colonoscope tip to capture underlying lesions, resulting in viewpoint changes in the colonoscopy images. These variations increase the diversity of polyp visual appearance, posing a challenge for learning robust polyp features. Secondly, polyps often exhibit properties similar to the surrounding tissues, leading to indistinct polyp boundaries. To address these problems, we propose a viewpoint-aware framework named VANet for precise polyp segmentation. In VANet, polyps are emphasized as a discriminative feature and thus can be localized by class activation maps in a viewpoint classification process. With these polyp locations, we design a viewpoint-aware Transformer (VAFormer) to alleviate the erosion of attention by the surrounding tissues, thereby inducing better polyp representations. Additionally, to enhance the polyp boundary perception of the network, we develop a boundary-aware Transformer (BAFormer) to encourage self-attention towards uncertain regions. As a consequence, the combination of the two modules is capable of calibrating predictions and significantly improving polyp segmentation performance. Extensive experiments on seven public datasets across six metrics demonstrate the state-of-the-art results of our method, and VANet can handle colonoscopy images in real-world scenarios effectively. The source code is available at https://github.com/1024803482/ViewpointAware-Network.
引用
收藏
页数:18
相关论文
共 79 条
[1]   A multi-centre polyp detection and segmentation dataset for generalisability assessment [J].
Ali, Sharib ;
Jha, Debesh ;
Ghatwary, Noha ;
Realdon, Stefano ;
Cannizzaro, Renato ;
Salem, Osama E. ;
Lamarque, Dominique ;
Daul, Christian ;
Riegler, Michael A. ;
Anonsen, Kim V. ;
Petlund, Andreas ;
Halvorsen, Pal ;
Rittscher, Jens ;
de Lange, Thomas ;
East, James E. .
SCIENTIFIC DATA, 2023, 10 (01)
[2]   Towards Real-Time Polyp Detection in Colonoscopy Videos: Adapting Still Frame-Based Methodologies for Video Sequences Analysis [J].
Angermann, Quentin ;
Bernal, Jorge ;
Sanchez-Montes, Cristina ;
Hammami, Maroua ;
Fernandez-Esparrach, Gloria ;
Dray, Xavier ;
Romain, Olivier ;
Javier Sanchez, F. ;
Histace, Aymeric .
COMPUTER ASSISTED AND ROBOTIC ENDOSCOPY AND CLINICAL IMAGE-BASED PROCEDURES, 2017, 10550 :29-41
[3]  
Ba J, 2014, ACS SYM SER
[4]  
Ba J.L., 2016, arXiv preprint arXiv:1607.06450, DOI DOI 10.48550/ARXIV.1607.06450
[5]   Towards automatic polyp detection with a polyp appearance model [J].
Bernal, J. ;
Sanchez, J. ;
Vilarino, F. .
PATTERN RECOGNITION, 2012, 45 (09) :3166-3182
[6]  
Bernal J., 2018, P 32 CARS C
[7]  
Bernal J., 2012, Database
[8]   WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians [J].
Bernal, Jorge ;
Javier Sanchez, F. ;
Fernandez-Esparrach, Gloria ;
Gil, Debora ;
Rodriguez, Cristina ;
Vilarino, Fernando .
COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2015, 43 :99-111
[9]  
Bond JH, 2000, AM J GASTROENTEROL, V95, P3053
[10]   Using Guided Self-Attention with Local Information for Polyp Segmentation [J].
Cai, Linghan ;
Wu, Meijing ;
Chen, Lijiang ;
Bai, Wenpei ;
Yang, Min ;
Lyu, Shuchang ;
Zhao, Qi .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT IV, 2022, 13434 :629-638