CsAGP: Detecting Alzheimer's disease from multimodal images via dual-transformer with cross-attention and graph pooling

被引:21
作者
Tang, Chaosheng [1 ]
Wei, Mingyang [1 ]
Sun, Junding [1 ]
Wang, Shuihua [1 ,2 ,3 ]
Zhang, Yudong [1 ,2 ,3 ]
机构
[1] Henan Polytech Univ, Sch Comp Sci & Technol, Jiaozuo 454000, Henan, Peoples R China
[2] Univ Leicester, Sch Comp & Math Sci, Leicester LE1 7RH, England
[3] King Abdulaziz Univ, Fac Comp & Informat Technol, Dept Informat Syst, Jeddah 21589, Saudi Arabia
关键词
Alzheimer's disease; Vision transformer; Multimodal image fusion; Deep learning; FUSION; MODEL;
D O I
10.1016/j.jksuci.2023.101618
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Alzheimer's disease (AD) is a terrible and degenerative disease commonly occurring in the elderly. Early detection can prevent patients from further damage, which is crucial in treating AD. Over the past few decades, it has been demonstrated that neuroimaging can be a critical diagnostic tool for AD, and the feature fusion of different neuroimaging modalities can enhance diagnostic performance. Most previous studies in multimodal feature fusion have only concatenated the high-level features extracted by neural networks from various neuroimaging images simply. However, a major problem of these studies is over-looking the low-level feature interactions between modalities in the feature extraction stage, resulting in suboptimal performance in AD diagnosis. In this paper, we develop a dual-branch vision transformer with cross-attention and graph pooling, namely CsAGP, which enables multi-level feature interactions between the inputs to learn a shared feature representation. Specifically, we first construct a brand-new cross-attention fusion module (CAFM), which processes MRI and PET images by two independent branches of differing computational complexity. These features are fused merely by the cross-attention mechanism to enhance each other. After that, a concise graph pooling algorithm-based Reshape-Pooling-Reshape (RPR) framework is developed for token selection to reduce token redundancy in the proposed model. Extensive experiments on the Alzheimer's Disease Neuroimaging Initiative (ADNI) data-base demonstrated that the suggested method obtains 99.04%, 97.43%, 98.57%, and 98.72% accuracy for the classification of AD vs. CN, AD vs. MCI, CN vs. MCI, and AD vs. CN vs. MCI, respectively.(c) 2023 The Author(s). Published by Elsevier B.V. on behalf of King Saud University. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
引用
收藏
页数:13
相关论文
共 42 条
[1]   CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification [J].
Chen, Chun-Fu ;
Fan, Quanfu ;
Panda, Rameswar .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :347-356
[2]   Generative Adversarial Networks An overview [J].
Creswell, Antonia ;
White, Tom ;
Dumoulin, Vincent ;
Arulkumaran, Kai ;
Sengupta, Biswa ;
Bharath, Anil A. .
IEEE SIGNAL PROCESSING MAGAZINE, 2018, 35 (01) :53-65
[3]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[4]   Ensemble of deep convolutional neural networks based multi-modality images for Alzheimer's disease diagnosis [J].
Fang, Xusheng ;
Liu, Zhenbing ;
Xu, Mingchang .
IET IMAGE PROCESSING, 2020, 14 (02) :318-326
[5]   Task-Induced Pyramid and Attention GAN for Multimodal Brain Image Imputation and Classification in Alzheimer's Disease [J].
Gao, Xingyu ;
Shi, Feng ;
Shen, Dinggang ;
Liu, Manhua .
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (01) :36-43
[6]   2022 Alzheimer's disease facts and figures [J].
Gaugler, Joseph ;
James, Bryan ;
Johnson, Tricia ;
Reimer, Jessica ;
Solis, Michele ;
Weuve, Jennifer ;
Buckley, Rachel F. ;
Hohman, Timothy J. .
ALZHEIMERS & DEMENTIA, 2022, 18 (04) :700-789
[7]  
Golovanevsky M, 2022, Arxiv, DOI arXiv:2206.08826
[8]  
Han Kun, 2020, Journal of Tsinghua University (Science and Technology), V60, P664, DOI 10.16511/j.cnki.qhdxxb.2020.25.003
[9]  
International A.D., 2018, World Alzheimer Report 2018 - The state of the art of dementia research: New frontiers', DOI DOI 10.1111/J.0033-0124.1950.24_14.X
[10]  
Jang J., 2022, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, P20718