HCformer: Hybrid CNN-Transformer for LDCT Image Denoising

被引:28
作者
Yuan, Jinli [1 ]
Zhou, Feng [1 ]
Guo, Zhitao [1 ]
Li, Xiaozeng [1 ]
Yu, Hengyong [2 ]
机构
[1] Hebei Univ Technol, Sch Elect & Informat Engn, Tianjin 300401, Peoples R China
[2] Univ Massachusetts Lowell, Dept Elect & Comp Engn, Lowell, MA 01854 USA
关键词
Low-dose CT; Deep learning; CT image denoising; LOW-DOSE CT; NETWORK;
D O I
10.1007/s10278-023-00842-9
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Low-dose computed tomography (LDCT) is an effective way to reduce radiation exposure for patients. However, it will increase the noise of reconstructed CT images and affect the precision of clinical diagnosis. The majority of the current deep learning-based denoising methods are built on convolutional neural networks (CNNs), which concentrate on local information and have little capacity for multiple structures modeling. Transformer structures are capable of computing each pixel's response on a global scale, but their extensive computation requirements prevent them from being widely used in medical image processing. To reduce the impact of LDCT scans on patients, this paper aims to develop an image post-processing method by combining CNN and Transformer structures. This method can obtain a high-quality images from LDCT. A hybrid CNN-Transformer (HCformer) codec network model is proposed for LDCT image denoising. A neighborhood feature enhancement (NEF) module is designed to introduce the local information into the Transformer's operation, and the representation of adjacent pixel information in the LDCT image denoising task is increased. The shifting window method is utilized to lower the computational complexity of the network model and overcome the problems that come with computing the MSA (Multi-head self-attention) process in a fixed window. Meanwhile, W/SW-MSA (Windows/Shifted window Multi-head self-attention) is alternately used in two layers of the Transformer to gain the information interaction between various Transformer layers. This approach can successfully decrease the Transformer's overall computational cost. The AAPM 2016 LDCT grand challenge dataset is employed for ablation and comparison experiments to demonstrate the viability of the proposed LDCT denoising method. Per the experimental findings, HCformer can increase the image quality metrics SSIM, HuRMSE and FSIM from 0.8017, 34.1898, and 0.6885 to 0.8507, 17.7213, and 0.7247, respectively. Additionally, the proposed HCformer algorithm will preserves image details while it reduces noise. In this paper, an HCformer structure is proposed based on deep learning and evaluated by using the AAPM LDCT dataset. Both the qualitative and quantitative comparison results confirm that the proposed HCformer outperforms other methods. The contribution of each component of the HCformer is also confirmed by the ablation experiments. HCformer can combine the advantages of CNN and Transformer, and it has great potential for LDCT image denoising and other tasks.
引用
收藏
页码:2290 / 2305
页数:16
相关论文
共 40 条
[11]   A Low-Dose CT Image Denoising Method Combining Multistage Network and Edge Protection [J].
Guo, Zhitao ;
Zhou, Feng ;
Chen, Yuqing ;
Yuan, Jinli .
TEHNICKI VJESNIK-TECHNICAL GAZETTE, 2022, 29 (03) :1059-1067
[12]   Neighbor2Neighbor: Self-Supervised Denoising from Single Noisy Images [J].
Huang, Tao ;
Li, Songjiang ;
Jia, Xu ;
Lu, Huchuan ;
Liu, Jianzhuang .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :14776-14785
[13]   CT-based radiomic analysis of stereotactic body radiation therapy patients with lung cancer [J].
Huynh, Elizabeth ;
Coroller, Thibaud P. ;
Narayan, Vivek ;
Agrawal, Vishesh ;
Hou, Ying ;
Romano, John ;
Franco, Idalid ;
Mak, Raymond H. ;
Aerts, Hugo J. W. L. .
RADIOTHERAPY AND ONCOLOGY, 2016, 120 (02) :258-266
[14]   SwinIR: Image Restoration Using Swin Transformer [J].
Liang, Jingyun ;
Cao, Jiezhang ;
Sun, Guolei ;
Zhang, Kai ;
Van Gool, Luc ;
Timofte, Radu .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, :1833-1844
[15]  
Liang T, ARXIV
[16]   Deep Learning for Generic Object Detection: A Survey [J].
Liu, Li ;
Ouyang, Wanli ;
Wang, Xiaogang ;
Fieguth, Paul ;
Chen, Jie ;
Liu, Xinwang ;
Pietikainen, Matti .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (02) :261-318
[17]   Swin Transformer: Hierarchical Vision Transformer using Shifted Windows [J].
Liu, Ze ;
Lin, Yutong ;
Cao, Yue ;
Hu, Han ;
Wei, Yixuan ;
Zhang, Zheng ;
Lin, Stephen ;
Guo, Baining .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :9992-10002
[18]  
Luo L., 2015, J DATA ACQUISITION P, V30, P224
[19]  
Luthra A., 2021, ARXIV
[20]  
Qiao YY., 2022, CT THEORY APPL, V31, P697, DOI [10.15953/j.ctta.2022.114, DOI 10.15953/J.CTTA.2022.114]