ADF-Net: A novel adaptive dual-stream encoding and focal attention decoding network for skin lesion segmentation

被引:6
作者
Huang, Zhengwei [1 ]
Deng, Hongmin [1 ]
Yin, Shuangcai [1 ]
Zhang, Ting [2 ]
Tang, Wentang [1 ]
Wang, Qionghua [3 ]
机构
[1] Sichuan Univ, Sch Elect & Informat Engn, Chengdu 610065, Peoples R China
[2] Sichuan Univ, West China Hosp, Sch Clin Med, Dept Dermatol, Chengdu 610041, Peoples R China
[3] Beihang Univ, Sch Instrumentat & Optoelect Engn, Beijing 100191, Peoples R China
基金
中国国家自然科学基金;
关键词
Skin lesion segmentation; Dual-stream; Focal attention decoder; Transformer; IMAGES;
D O I
10.1016/j.bspc.2023.105895
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Automatic segmentation of lesion areas in dermoscopic images is a crucial step in computer -aided skin lesion examination and diagnosis systems. Efficient and accurate skin lesion segmentation benefits the quantitative analysis of diseases, such as melanoma, dermatofibroma, and seborrheic keratosis and so on. However, in practical clinical diagnosis, some lesion areas exhibit large-scale changes, fuzzy and irregular boundaries, and low contrast between the lesion and the background, leading to potential segmentation errors. To overcome this difficulty, we propose a novel network called ADF-Net, composed of a multi -stage dual -stream hybrid framework (MDHF) based on Transformer and a convolutional neural network is well designed to achieve comprehensive integration of coarse -grained and fine-grained feature representations. This framework, in combination with the adaptive feature fusion (AFF) module, enables efficient and adaptive integration of both global and local feature information. Additionally, a focal attention decoder (FAD) is proposed to suppress background noise, focusing on the target area, and promoting the fusion of encoder features and high-level features. Finally, we conduct extensive experiments on four public datasets, including ISIC 2018, ISIC 2017, ISIC 2016, and PH2. The test sets of them consist of 379, 600, 518, and 100 images, respectively. The results demonstrate that our ADF-Net outperforms other state-of-the-art methods in all four commonly used evaluation metrics, with Jaccard Index values of 84.52%, 78.92%, 87.44%, and 93.26% respectively. Furthermore, the computational complexity of the model is only 8.29G GFLOPs, and the inference time is only 20.6 ms, which is of significant importance for exploring clinical applications.
引用
收藏
页数:13
相关论文
共 42 条
[1]  
Azad R, 2022, Arxiv, DOI [arXiv:2212.13504, DOI 10.48550/ARXIV.2212.13504]
[2]  
Cao Hu, 2023, Computer Vision - ECCV 2022 Workshops: Proceedings. Lecture Notes in Computer Science (13803), P205, DOI 10.1007/978-3-031-25066-8_9
[3]   An Improved Res-UNet Model for Tree Species Classification Using Airborne High-Resolution Images [J].
Cao, Kaili ;
Zhang, Xiaoli .
REMOTE SENSING, 2020, 12 (07)
[4]   A methodological approach to the classification of dermoscopy images [J].
Celebi, M. Emre ;
Kingravi, Hassan A. ;
Uddin, Bakhtiyar ;
Lyatornid, Hitoshi ;
Aslandogan, Y. Alp ;
Stoecker, William V. ;
Moss, Randy H. .
COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2007, 31 (06) :362-373
[5]  
Chen J., 2021, arXiv, DOI DOI 10.48550/ARXIV.2102.04306
[6]  
Codella NCF, 2018, I S BIOMED IMAGING, P168, DOI 10.1109/ISBI.2018.8363547
[7]   CSWin Transformer: A General Vision Transformer Backbone with Cross-Shaped Windows [J].
Dong, Xiaoyi ;
Bao, Jianmin ;
Chen, Dongdong ;
Zhang, Weiming ;
Yu, Nenghai ;
Yuan, Lu ;
Chen, Dong ;
Guo, Baining .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :12114-12124
[8]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[9]   SLT-Net: A codec network for skin lesion segmentation [J].
Feng, Kaili ;
Ren, Lili ;
Wang, Guanglei ;
Wang, Hongrui ;
Li, Yan .
COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 148
[10]   CPFNet: Context Pyramid Fusion Network for Medical Image Segmentation [J].
Feng, Shuanglang ;
Zhao, Heming ;
Shi, Fei ;
Cheng, Xuena ;
Wang, Meng ;
Ma, Yuhui ;
Xiang, Dehui ;
Zhu, Weifang ;
Chen, Xinjian .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2020, 39 (10) :3008-3018