DLFA-NET: A HYBRID NETWORK FOR MR-TO-CT SYNTHESIS VIA GLOBAL AND LOCAL FEATURE AGGREGATION

被引:2
作者
Chen, Zeli [1 ,2 ]
Li, Chuanpu [1 ,2 ]
Zheng, Kaiyi [1 ,2 ]
Zhang, Yiwen [1 ,2 ]
Wu, Yuankui [3 ]
Feng, Qianjin [1 ,2 ]
Zhong, Liming
Yang, Wei
机构
[1] Southern Med Univ, Sch Biomed Engn, Guangzhou, Peoples R China
[2] Southern Med Univ, Guangdong Prov Key Lab Med Image Proc, Guangzhou, Peoples R China
[3] Southern Med Univ, Nanfang Hosp, Dept Med Imaging Ctr, Guangzhou 510515, Peoples R China
来源
2023 IEEE 20TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING, ISBI | 2023年
基金
中国国家自然科学基金;
关键词
MR-to-CT synthesis; MR-only treatment planning; Transformer; Convolutional neural network;
D O I
10.1109/ISBI53787.2023.10230486
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Synthesis of Computed Tomography (CT) images from Magnetic Resonance (MR) images is clinical significance for MR-only treatment planning to eliminate the co-registration errors between MR and CT images. Existing convolutional neural network-based methods, suffering from the inherent local inductive biases, struggle to distinguish bone and air which show low signals in conventional MR images. ViT-based methods can learn long-range contextual information by a global self-attention mechanism but are limited by the quadratic complexity and generating local detailed structures. Combining the merits of these two architectures, we propose a hybrid network for MR-to-CT synthesis via global and local feature aggregation from Transformer and CNN, named GLFA-NET. Specifically, we add a global patch embedding branch to supplement the patch-based global representative features directly from the image and design a residual dilated swin transformer block aggregating the local detailed features and global features to improve the synthesis performance of bone and air and reduce computational overhead. Furthermore, we adopt a wavelet PatchGAN discriminator to enhance the high-frequency detailed information of the synthetic CT. Our GLFA-NET was implemented on a dataset with 154 pairs of 3D MR-CT head and neck images. Experiments show that our GLFA-NET achieves impressive performance with MAE of 71.12 +/- 10.87, SSIM of 0.771 +/- 0.028, and PSNR of 28.91 +/- 1.33. The visual synthetic CT results also show that the proposed GLFA-NET method achieves better discrimination of bone and air and higher structural similarity than other state-of-the-art methods.
引用
收藏
页数:5
相关论文
共 17 条
[11]   U-Net: Convolutional Networks for Biomedical Image Segmentation [J].
Ronneberger, Olaf ;
Fischer, Philipp ;
Brox, Thomas .
MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION, PT III, 2015, 9351 :234-241
[12]   Radiotherapy planning using MRI [J].
Schmidt, Maria A. ;
Payne, Geoffrey S. .
PHYSICS IN MEDICINE AND BIOLOGY, 2015, 60 (22) :R323-R361
[13]   Objective quantification of intervertebral disc volume properties using MRI in idiopathic scoliosis surgery [J].
Violas, Philippe ;
Estivalezes, Erik ;
Briot, Jerome ;
de Gauzy, Jerome Sales ;
Swider, Pascal .
MAGNETIC RESONANCE IMAGING, 2007, 25 (03) :386-391
[14]   Deep embedding convolutional neural network for synthesizing CT image from T1-Weighted MR image [J].
Xiang, Lei ;
Wang, Qian ;
Nie, Dong ;
Zhang, Lichi ;
Jin, Xiyao ;
Qiao, Yu ;
Shen, Dinggang .
MEDICAL IMAGE ANALYSIS, 2018, 47 :31-44
[15]   Predicting CT Image From MRI Data Through Feature Matching With Learned Nonlinear Local Descriptors [J].
Yang, Wei ;
Zhong, Liming ;
Chen, Yang ;
Lin, Liyan ;
Lu, Zhentai ;
Liu, Shupeng ;
Wu, Yao ;
Feng, Qianjin ;
Chen, Wufan .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2018, 37 (04) :977-987
[16]  
Zhang Bowen, 2022, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, P11304
[17]   Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks [J].
Zhu, Jun-Yan ;
Park, Taesung ;
Isola, Phillip ;
Efros, Alexei A. .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2242-2251