A multi-scale pyramid residual weight network for medical image fusion

被引:0
作者
Liu, Yiwei [1 ]
Zhang, Shaoze [1 ]
Tang, Yao [2 ]
Zhao, Xihai [1 ]
He, Zuo-Xiang [3 ,4 ]
机构
[1] Tsinghua Univ, Ctr Biomed Imaging Res, Sch Biomed Engn, 1 Shuangqing Rd, Beijing 100084, Peoples R China
[2] Tsinghua Univ, Dept Biomed Engn, Beijing, Peoples R China
[3] Tsinghua Univ, Sch Clin Med, Beijing, Peoples R China
[4] Beijing Tsinghua Changgung Hosp, Dept Nucl Med, 168 Litang Rd, Beijing 102218, Peoples R China
关键词
Medical image fusion; unsupervised learning; residual; attention; multiscale; TRANSFORM;
D O I
10.21037/qims-24-851
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Background: Due to the inherent limitations of imaging sensors, acquiring medical images that simultaneously provide functional metabolic information and detailed structural organization remains a significant challenge. Multi-modal image fusion has emerged as a critical technology for clinical diagnosis and surgical navigation, as it enables the integration of complementary information from different imaging modalities. However, existing deep learning (DL)-based fusion methods often face difficulties in effectively combining high-frequency detail information with low-frequency contextual information, which frequently leads to the degradation of high-frequency details. Therefore, there is a pressing need for a method that addresses these challenges, preserving both high-and low-frequency information while maintaining clear structural contours. In response to this issue, a novel convolutional neural network (CNN), named the multi-scale pyramid residual weight network (LYWNet), is proposed. The objective of this approach is to improve the fusion process by effectively integrating high-and low-frequency information, thereby enhancing the quality and accuracy of multimodal image fusion. This method aims to overcome the limitations of current fusion techniques and ensure the preservation of both functional and structural details, ultimately contributing to more precise clinical diagnoses and better surgical navigation outcomes. Methods: We propose a novel CNN, LYWNet, designed to address these challenges. LYWNet is composed of three modules: (I) data preprocessing module: utilizes three convolutional layers to extract both deep and shallow features from the input images. (II) Feature extraction module: incorporates three identical multi-scale pyramid residual weight (LYW) blocks in series, each featuring three interactive branches to preserve high-frequency detail information effectively. (III) Image reconstruction module: utilizes a fusion algorithm based on feature distillation to ensure the effective integration of functional and anatomical information. The proposed image fusion algorithm enhances the interaction of contextual cues and retains the metabolic details from functional images while preserving texture details from anatomical images. Results: The proposed LYWNet demonstrated its ability to retain high-frequency details during feature extraction, effectively combining them with low-frequency contextual information. The fusion results exhibited reduced differences between the fused image and the original images. The structural similarity (SSIM) and peak signal-to-noise ratio (PSNR) were 0.5592 +/- 0.0536 and 17.3594 +/- 1.0211, respectively, for single-photon emission computed tomography-magnetic resonance imaging (SPECT-MRI), 0.5195 +/- 0.0730 and 14.5324 +/- 1.7365 for PET-MRI; 0.5376 +/- 0.0442 and 13.9202 +/- 0.7265 for magnetic resonance imaging-computed tomography. Conclusions: LYWNet excels at integrating high-frequency detail information and low-frequency contextual information, addressing the deficiencies of existing DL-based image fusion methods. This approach provides superior fused images that retain the functional metabolic information and anatomical texture, making it a valuable tool for clinical diagnosis and surgical navigation.
引用
收藏
页码:1793 / 1821
页数:29
相关论文
共 54 条
[1]   Three dimensional computed tomography texture analysis of pulmonary lesions: Does radiomics allow differentiation between carcinoma, neuroendocrine tumor and organizing pneumonia? [J].
Adelsmayr, Gabriel ;
Janisch, Michael ;
Mueller, Heimo ;
Holzinger, Andreas ;
Talakic, Emina ;
Janek, Elmar ;
Streit, Simon ;
Fuchsjaeger, Michael ;
Schoellnast, Helmut .
EUROPEAN JOURNAL OF RADIOLOGY, 2023, 165
[2]   A Novel Method of Multimodal Medical Image Fusion Based on Hybrid Approach of NSCT and DTCWT [J].
Alseelawi, Nawar ;
Hazim, Hussein Tuama ;
ALRikabi, Haider Th Salim .
INTERNATIONAL JOURNAL OF ONLINE AND BIOMEDICAL ENGINEERING, 2022, 18 (03) :114-133
[3]   Infrared and visible image fusion based on target-enhanced multiscale transform decomposition [J].
Chen, Jun ;
Li, Xuejiao ;
Luo, Linbo ;
Mei, Xiaoguang ;
Ma, Jiayi .
INFORMATION SCIENCES, 2020, 508 :64-78
[4]   CT-texture analysis of subsolid nodules for differentiating invasive from in-situ and minimally invasive lung adenocarcinoma subtypes [J].
Cohen, J. G. ;
Reymond, E. ;
Medici, M. ;
Lederlin, M. ;
Lantuejoul, S. ;
Laurent, F. ;
Toffart, A. C. ;
Moreau-Gaudry, A. ;
Jankowski, A. ;
Ferretti, G. R. .
DIAGNOSTIC AND INTERVENTIONAL IMAGING, 2018, 99 (05) :291-299
[5]  
Deshmukh Manjusha, 2010, Int. J. Image Process. (IJIP), V4, P484
[6]   Multi-modal medical image fusion framework using co-occurrence filter and local extrema in NSST domain [J].
Diwakar, Manoj ;
Singh, Prabhishek ;
Shankar, Achyut .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2021, 68
[7]   Union Laplacian pyramid with multiple features for medical image fusion [J].
Du, Jiao ;
Li, Weisheng ;
Xiao, Bin ;
Nawaz, Qamar .
NEUROCOMPUTING, 2016, 194 :326-339
[8]   A METHOD FOR COMPARING 2 HIERARCHICAL CLUSTERINGS [J].
FOWLKES, EB ;
MALLOWS, CL .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1983, 78 (383) :553-569
[9]   DSAGAN: A generative adversarial network based on dual-stream attention mechanism for anatomical and functional image fusion [J].
Fu, Jun ;
Li, Weisheng ;
Du, Jiao ;
Xu, Liming .
INFORMATION SCIENCES, 2021, 576 :484-506
[10]   A multiscale residual pyramid attention network for medical image fusion [J].
Fu, Jun ;
Li, Weisheng ;
Du, Jiao ;
Huang, Yuping .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2021, 66