DFENet: A dual-branch feature enhanced network integrating transformers and convolutional feature learning for multimodal medical image fusion

被引:57
作者
Li, Weisheng [1 ]
Zhang, Yin [1 ]
Wang, Guofen [1 ]
Huang, Yuping [1 ]
Li, Ruyue [1 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Comp Sci & Technol, Chongqing 400065, Peoples R China
基金
中国国家自然科学基金;
关键词
Multimodal medical image fusion; Convolutional neural network; Vision transformer; Feature fuser; Local energy and gradient; QUALITY ASSESSMENT; PERFORMANCE; INFORMATION; ALGORITHM;
D O I
10.1016/j.bspc.2022.104402
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
In recent times, several medical image fusion techniques based on the convolutional neural network (CNN) have been proposed for various medical imaging fusion tasks. However, these methods cannot model the long-range dependencies between the fused image and the source images. To address this limitation, we propose DFENet, a multimodal medical image fusion framework that integrates CNN feature learning and vision transformer feature learning using self-supervised learning. DFENet is based on an encoder-decoder network, which can be trained on large-scale natural image dataset without the need for carefully collated ground truth fusion images. The pro-posed network consists of an encoder, a feature fuser, and a decoder. The encoder is composed of a CNN module and a transformer module, which is used to extract local and global features of images. In order to avoid the use of simple up-sampling and concatenate processing, a new global semantic information aggregation module is proposed to efficiently aggregate the multi-scale features obtained by the transformer module, which enhances the quality of the reconstructed images. The decoder is composed of six convolution layers with two skip con-nections, which are used for the reconstruction from fused features. We also propose a fusion strategy combining local energy and gradient information for the feature fusion process of magnetic resonance imaging and func-tional medical images. Compared to conventional fusion rules, our fusion strategy is more robust to noisy images. And compared with the existing competitive methods, our method retains more texture details of the original images and outputs a more natural and realistic fused image.
引用
收藏
页数:19
相关论文
共 53 条
[1]   Image fusion based on nonsubsampled contourlet transform for infrared and visible light image [J].
Adu, Jianhua ;
Gan, Jianhong ;
Wang, Yan ;
Huang, Jian .
INFRARED PHYSICS & TECHNOLOGY, 2013, 61 :94-100
[2]   Wavelet based image fusion techniques - An introduction, review and comparison [J].
Amolins, Krista ;
Zhang, Yun ;
Dare, Peter .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2007, 62 (04) :249-263
[3]   Directive Contrast Based Multimodal Medical Image Fusion in NSCT Domain [J].
Bhatnagar, Gaurav ;
Wu, Q. M. Jonathan ;
Liu, Zheng .
IEEE TRANSACTIONS ON MULTIMEDIA, 2013, 15 (05) :1014-1024
[4]   THE LAPLACIAN PYRAMID AS A COMPACT IMAGE CODE [J].
BURT, PJ ;
ADELSON, EH .
IEEE TRANSACTIONS ON COMMUNICATIONS, 1983, 31 (04) :532-540
[5]   A human perception inspired quality metric for image fusion based on regional information [J].
Chen, Hao ;
Varshney, Pramod K. .
INFORMATION FUSION, 2007, 8 (02) :193-207
[6]   A new automated quality assessment algorithm for image fusion [J].
Chen, Yin ;
Blum, Rick S. .
IMAGE AND VISION COMPUTING, 2009, 27 (10) :1421-1432
[7]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[8]   Anatomical-Functional Image Fusion by Information of Interest in Local Laplacian Filtering Domain [J].
Du, Jiao ;
Li, Weisheng ;
Xiao, Bin .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (12) :5855-5866
[9]   DSAGAN: A generative adversarial network based on dual-stream attention mechanism for anatomical and functional image fusion [J].
Fu, Jun ;
Li, Weisheng ;
Du, Jiao ;
Xu, Liming .
INFORMATION SCIENCES, 2021, 576 :484-506
[10]   Res2Net: A New Multi-Scale Backbone Architecture [J].
Gao, Shang-Hua ;
Cheng, Ming-Ming ;
Zhao, Kai ;
Zhang, Xin-Yu ;
Yang, Ming-Hsuan ;
Torr, Philip .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (02) :652-662