RT-Net: Region-Enhanced Attention Transformer Network for Polyp Segmentation

被引:3
|
作者
Qin, Yilin [1 ]
Xia, Haiying [1 ]
Song, Shuxiang [1 ]
机构
[1] Guangxi Normal Univ, Coll Elect Engn, Guilin 541004, Guangxi, Peoples R China
关键词
Transformer; Polyp segmentation; Region-enhanced attention; Medical image segmentation;
D O I
10.1007/s11063-023-11405-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Colonic polyps are highly correlated with colorectal cancer. Prevention of colorectal cancer is the detection and removal of polyps in the early stages of the disease. But the detection process relies on the physician's experience and is prone to missed diagnoses. The drawbacks motivate us to design an algorithm to automatically assist physicians in detection to reduce the rate of missed polyps. However, polyp segmentation encounters challenges due to the variable appearance and blurred borders with the surrounding mucosa. And it is difficult for existing CNN-based polyp segmentation algorithms to learn long-range dependencies. Therefore, we propose a region-enhanced attention transformer network (RT-Net) for polyp segmentation. Unlike existing CNN-based approaches, it employs a pyramid Transformer encoder to promote the learning ability and robustness of the network. In addition, we introduce three modules, including the residual multiscale (RMS) module, the region-enhanced attention (REA) module and the feature aggregation (FA) module. Specifically, the RMS module learns multiscale information from the features of the encoder. The REA module adopts the prediction maps of each decoder layer to guide the network in building target regions and boundary cues to compensate for the missing local fields of view in the encoder. The role of the FA module is to efficiently aggregate the features from REA with those from the decoding layer to achieve better segmentation performance. RT-Net is evaluated on five benchmark polyp datasets. Extensive experiments demonstrate that our proposed RT-Net exhibits excellent performance compared to other state-of-the-art methods.
引用
收藏
页码:11975 / 11991
页数:17
相关论文
共 50 条
  • [1] RT-Net: Region-Enhanced Attention Transformer Network for Polyp Segmentation
    Yilin Qin
    Haiying Xia
    Shuxiang Song
    Neural Processing Letters, 2023, 55 (9) : 11975 - 11991
  • [2] REU-Net: Region-enhanced nuclei segmentation network
    He, Yongjun
    Qin, Jian
    Zhou, Yang
    Zhao, Jing
    Ding, Bo
    COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 146
  • [3] Shared Hybrid Attention Transformer network for colon polyp segmentation
    Ji, Zexuan
    Qian, Hao
    Ma, Xiao
    NEUROCOMPUTING, 2025, 616
  • [4] RSAFormer: A method of polyp segmentation with region self-attention transformer
    Yin X.
    Zeng J.
    Hou T.
    Tang C.
    Gan C.
    Jain D.K.
    García S.
    Computers in Biology and Medicine, 2024, 172
  • [5] Few-Shot Medical Image Segmentation via a Region-Enhanced Prototypical Transformer
    Zhu, Yazhou
    Wang, Shidong
    Xin, Tong
    Zhang, Haofeng
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT IV, 2023, 14223 : 271 - 280
  • [6] Region-Enhanced Feature Learning for Scene Semantic Segmentation
    Kang, Xin
    Wang, Chaoqun
    Chen, Xuejin
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 954 - 964
  • [7] REET: Region-Enhanced Transformer for Person Re-Identification
    Lee, Kyoungoh
    Jang, In-su
    Kim, Kwang-Ju
    Kim, Pyong-Kun
    2022 18TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE (AVSS 2022), 2022,
  • [8] TACT: Text attention based CNN-Transformer network for polyp segmentation
    Zhao, Yiyang
    Li, Jinjiang
    Hua, Zhen
    INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2024, 34 (02)
  • [9] DHAFormer: Dual-channel hybrid attention network with transformer for polyp segmentation
    Huang, Xuejie
    Wang, Liejun
    Jiang, Shaochen
    Xu, Lianghui
    PLOS ONE, 2024, 19 (07):
  • [10] Enhanced U-Net: A Feature Enhancement Network for Polyp Segmentation
    Patel, Krushi
    Bur, Andres M.
    Wang, Guanghui
    2021 18TH CONFERENCE ON ROBOTS AND VISION (CRV 2021), 2021, : 181 - 188