Medical image segmentation using boundary-enhanced guided packet rotation dual attention decoder network

被引:2
作者
Lu, Hongchun [1 ,2 ]
Tian, Shengwei [1 ]
Yu, Long [3 ]
Xing, Yan [4 ]
Cheng, Junlong [2 ,5 ]
Liu, Lu [6 ]
机构
[1] Xinjiang Univ, Sch Software, Urumqi 830046, Xinjiang, Peoples R China
[2] Xinjiang Univ, Key Lab Software Engn Technol, Urumqi, Xinjiang, Peoples R China
[3] Xinjiang Univ, Network Ctr, Urumqi, Xinjiang, Peoples R China
[4] First Affiliated Hosp Xinjiang Med Univ, Urumqi, Xinjiang, Peoples R China
[5] Xinjiang Univ, Sch Informat Sci & Engn, Urumqi, Xinjiang, Peoples R China
[6] Xinjiang Normal Univ, Sch Educ Sci, Urumqi, Xinjiang, Peoples R China
关键词
Medical image segmentation; packet rotation convolution; dual attention mechanism; boundary enhancement; convolutional neural network;
D O I
10.3233/THC-202789
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
BACKGROUND: The automatic segmentation of medical images is an important task in clinical applications. However, due to the complexity of the background of the organs, the unclear boundary, and the variable size of different organs, some of the features are lost during network learning, and the segmentation accuracy is low. OBJECTIVE: To address these issues, this prompted us to study whether it is possible to better preserve the deep feature information of the image and solve the problem of low segmentation caused by unclear image boundaries. METHODS: In this study, we (1) build a reliable deep learning network framework, named BGRANet,to improve the segmentation performance for medical images; (2) propose a packet rotation convolutional fusion encoder network to extract features; (3) build a boundary enhanced guided packet rotation dual attention decoder network, which is used to enhance the boundary of the segmentation map and effectively fuse more prior information; and (4) propose a multi-resolution fusion module to generate high-resolution feature maps. We demonstrate the effectiveness of the proposed method on two publicly available datasets. RESULTS: BGRANet has been trained and tested on the prepared dataset and the experimental results show that our proposed model has better segmentation performance. For 4 class classification (CHAOS dataset), the average dice similarity coefficient reached 91.73%. For 2 class classification (Herlev dataset), the prediction, sensitivity, specificity, accuracy, and Dice reached 93.75%, 94.30%, 98.19%, 97.43%, and 98.08% respectively. The experimental results show that BGRANet can improve the segmentation effect for medical images. CONCLUSION: We propose a boundary-enhanced guided packet rotation dual attention decoder network. It achieved high segmentation accuracy with a reduced parameter number.
引用
收藏
页码:129 / 143
页数:15
相关论文
共 27 条
[1]   Recurrent residual U-Net for medical image segmentation [J].
Alom, Md Zahangir ;
Yakopcic, Chris ;
Hasan, Mahmudul ;
Taha, Tarek M. ;
Asari, Vijayan K. .
JOURNAL OF MEDICAL IMAGING, 2019, 6 (01)
[2]  
[Anonymous], 2018, ARXIV180703973
[3]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[4]  
Brox T., 2016, INT C MED IM COMP CO, P424
[5]   A tutorial on the cross-entropy method [J].
De Boer, PT ;
Kroese, DP ;
Mannor, S ;
Rubinstein, RY .
ANNALS OF OPERATIONS RESEARCH, 2005, 134 (01) :19-67
[6]   Res2Net: A New Multi-Scale Backbone Architecture [J].
Gao, Shang-Hua ;
Cheng, Ming-Ming ;
Zhao, Kai ;
Zhang, Xin-Yu ;
Yang, Ming-Hsuan ;
Torr, Philip .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (02) :652-662
[7]   MMJM: A High-Precision Motion Model for Human Arm Reaching Motion [J].
Gong, Shiqiu ;
Zhao, Jing .
2020 IEEE 3RD INTERNATIONAL CONFERENCE ON MECHATRONICS, ROBOTICS AND AUTOMATION (ICMRA 2020), 2020, :59-63
[8]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[9]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
[10]  
Jin-Yu Z, 2009, 2009 INT C IM AN SIG