A Unified CNN-ViT Network with a Feature Distribution Strategy for Multi-modal Missing MRI Sequences Imputation

被引:0
|
作者
Wang, Yulin [1 ]
Liu, Qian [1 ]
机构
[1] Hainan Univ, Sch Biomed Engn, Key Lab Biomed Engn Hainan Prov, Haikou, Peoples R China
来源
12TH ASIAN-PACIFIC CONFERENCE ON MEDICAL AND BIOLOGICAL ENGINEERING, VOL 1, APCMBE 2023 | 2024年 / 103卷
关键词
Image synthesis; Multi-modal MRI; Visual transformer; Convolutional neural network; Deep learning;
D O I
10.1007/978-3-031-51455-5_26
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modalmagnetic resonance imaging (MRI) is of great clinical use for disease assessment and diagnosis, as it provides comprehensive and complementary information through multiple contrasts. However, due to potential obstacles of the scanning process and the patient's physical condition, the available scans of each subject may vary. Here, we propose a unified adversarial network based on the convolution neural network (CNN) and vision transformer (ViT) for missing MRI image synthesis in any input-output image configurations. The purpose of our network design is to develop a robust and efficient way to integrate the local information capturing ability of the convolutional operation and global contextual sensitivity of the multi-head self-attention (MSA) mechanism. Specifically, we employ a u-shape network as our generator and mix the convolutional path and the MSA path in a parallel manner at each network stage, the mixer is in place of the original MSA block of the canonical transformer, this design enables each layer to learn both global and local information simultaneously and takes advantage of the general ViT architecture. Furthermore, since these two branches have different frequency information preference, possessing all channels in each branch inevitably renders feature redundancy and introduces extra artifacts, accordingly, we adopt a channel splitting strategy that splits input features channel-wise and feed to each branch separately, meanwhile, considering each network stage has different desire of the high- and low-frequency information, we also explore various channel distribution ratios at each network stage. Our proposed method demonstrates reliable synthesis for healthy tissues and heterogeneous enhancement on BraTS2021 dataset. Furthermore, compare with four state-of-the-art methods including convolution-based, MSA-based and hybrid models in the MRI sequence synthesis field, our approach outperforms these methods quantitatively and qualitatively. Therefore, our method has the potential to be used as a candidate for medical image synthesis.
引用
收藏
页码:238 / 244
页数:7
相关论文
共 6 条
  • [1] Unified Multi-Modal Image Synthesis for Missing Modality Imputation
    Zhang, Yue
    Peng, Chengtao
    Wang, Qiuli
    Song, Dan
    Li, Kaiyan
    Zhou, S. Kevin
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2025, 44 (01) : 4 - 18
  • [2] Missing MRI Pulse Sequence Synthesis Using Multi-Modal Generative Adversarial Network
    Sharma, Anmol
    Hamarneh, Ghassan
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2020, 39 (04) : 1170 - 1183
  • [3] Multi-Modal Modality-Masked Diffusion Network for Brain MRI Synthesis With Random Modality Missing
    Meng, Xiangxi
    Sun, Kaicong
    Xu, Jun
    He, Xuming
    Shen, Dinggang
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2024, 43 (07) : 2587 - 2598
  • [4] Learning Unified Hyper-Network for Multi-Modal MR Image Synthesis and Tumor Segmentation With Missing Modalities
    Yang, Heran
    Sun, Jian
    Xu, Zongben
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2023, 42 (12) : 3678 - 3689
  • [5] M2GCNet: Multi-Modal Graph Convolution Network for Precise Brain Tumor Segmentation Across Multiple MRI Sequences
    Zhou, Tongxue
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4896 - 4910
  • [6] RAE-Net: a multi-modal neural network based on feature fusion and evidential deep learning algorithm in predicting breast cancer subtypes on DCE-MRI
    Tang, Xiaowen
    Zhu, Yinsu
    BIOMEDICAL PHYSICS & ENGINEERING EXPRESS, 2025, 11 (02):