DSAGAN: A generative adversarial network based on dual-stream attention mechanism for anatomical and functional image fusion

被引:63
作者
Fu, Jun [1 ,2 ]
Li, Weisheng [1 ]
Du, Jiao [3 ]
Xu, Liming [1 ]
机构
[1] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
[2] Zunyi Normal Univ, Sch Informat Engn, Zunyi 563006, Guizhou, Peoples R China
[3] Guangzhou Univ, Sch Comp Sci & Educ Software, Guangzhou 510006, Peoples R China
基金
中国国家自然科学基金;
关键词
Multimodal; Image fusion; Generative adversarial network; Dual-stream attention; MULTI-FOCUS IMAGE; QUALITY ASSESSMENT; TRANSFORM; DIVERGENCE;
D O I
10.1016/j.ins.2021.06.083
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, extensive multimodal medical image fusion algorithms have been pro-posed. However, existing methods are primarily based on specific transformation theories. There are many problems with existing algorithms, such as poor adaptability, low effi-ciency and blurry details. To address these problems, this paper proposes a generative adversarial network based on dual-stream attention mechanism (DSAGAN) for anatomical and functional image fusion. The dual-stream architecture and multiscale convolutions are utilized to extract deep features. In addition, the attention mechanism is utilized to further enhance the fused features. Then, the fusion images and multimodal input images are put into the discriminator. In the update stage of the discriminator, we expect to judge the multimodal images as real, and to judge the fusion images as fake. Furthermore, the fusion images are expected to be judged as real in the update stage of the generator, forcing the generator to improve the fusion quality. The training process continues until the generator and discriminator reach a Nash equilibrium. After training, the fusion images can be obtained directly after inputting anatomical and functional images. Compared with the ref-erence algorithms, DSAGAN consumes less fusion time and achieves better objective met-rics in terms of QAG, QEN and QNIQE. (c) 2021 Elsevier Inc. All rights reserved.
引用
收藏
页码:484 / 506
页数:23
相关论文
共 48 条
[1]   Image fusion via sparse regularization with non-convex penalties [J].
Anantrasirichai, Nantheera ;
Zheng, Rencheng ;
Selesnick, Ivan ;
Achim, Alin .
PATTERN RECOGNITION LETTERS, 2020, 131 :355-360
[2]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[3]   Directive Contrast Based Multimodal Medical Image Fusion in NSCT Domain [J].
Bhatnagar, Gaurav ;
Wu, Q. M. Jonathan ;
Liu, Zheng .
IEEE TRANSACTIONS ON MULTIMEDIA, 2013, 15 (05) :1014-1024
[4]   Infrared and visible image fusion based on target-enhanced multiscale transform decomposition [J].
Chen, Jun ;
Li, Xuejiao ;
Luo, Linbo ;
Mei, Xiaoguang ;
Ma, Jiayi .
INFORMATION SCIENCES, 2020, 508 (508) :64-78
[5]   Fusion of anatomical and functional images using parallel saliency features [J].
Du, Jiao ;
Li, Weisheng ;
Xiao, Bin .
INFORMATION SCIENCES, 2018, 430 :567-576
[6]   Anatomical-Functional Image Fusion by Information of Interest in Local Laplacian Filtering Domain [J].
Du, Jiao ;
Li, Weisheng ;
Xiao, Bin .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (12) :5855-5866
[7]   An overview of multi-modal medical image fusion [J].
Du, Jiao ;
Li, Weisheng ;
Lu, Ke ;
Xiao, Bin .
NEUROCOMPUTING, 2016, 215 :3-20
[8]   Union Laplacian pyramid with multiple features for medical image fusion [J].
Du, Jiao ;
Li, Weisheng ;
Xiao, Bin ;
Nawaz, Qamar .
NEUROCOMPUTING, 2016, 194 :326-339
[9]   Multi-focus image fusion based on non-subsampled shearlet transform [J].
Gao Guorong ;
Xu Luping ;
Feng Dongzhu .
IET IMAGE PROCESSING, 2013, 7 (06) :633-639
[10]  
Goodfellow I., 2020, ADV NEUR IN, V63, P139, DOI [DOI 10.1145/3422622, 10.1145/3422622]