Information sparsity guided transformer for multi-modal medical image super-resolution

被引:1
作者
Lu, Haotian [1 ]
Mei, Jie [2 ]
Qiu, Yu [3 ]
Li, Yumeng [1 ]
Hao, Fangwei [1 ]
Xu, Jing [1 ]
Tang, Lin [4 ]
机构
[1] Nankai Univ, Coll Artificial Intelligence, Natl Key Lab Intelligent Tracking & Forecasting In, Tianjin, Peoples R China
[2] Hunan Univ, Natl Engn Res Ctr Robot Visual Percept & Control T, Sch Robot, Changsha, Peoples R China
[3] Hunan Normal Univ, Coll Informat Sci & Engn, Changsha, Peoples R China
[4] New York Inst Technol, Dept Biol & Chem Sci, New York, NY USA
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Super-resolution; Sparse transformer; Medical image; Multi-modal; NETWORK; ATTENTION;
D O I
10.1016/j.eswa.2024.125428
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modal medical image super-resolution (SR) plays a vital role in enhancing the resolution of medical images, providing more detailed visuals that aid in accurate clinical diagnosis. Recently, Transformer-based super-resolution methods have significantly promoted performance improvement in this field due to their capacity to capture global dependencies. They usually process all non-overlapping patches as tokens and densely sample these tokens without screening for calculating attention mechanisms. However, this strategy ignores the spatial sparsity of medical images, resulting in redundant or even detrimental computations on less informative regions. Hence, this paper proposes a novel sparsity-guided medical image SR network, namely SGSRNet, by exploiting the spatial sparsity characteristics of the medical images. SG-SRNet mainly consists of two components: a sparsity mask (SM) generator for image sparsity estimation, and a sparsity-guided Transformer (SGTrans) for high-resolution image reconstruction. Specially, the SM generator generates a sparsity mask by minimizing our cross-sparsity loss which can respond to the informative positions. SGTrans first screens out the informative patches according to the sparsity mask. Then, SGTrans utilizes the designed cluster-based attention to only calculate attention between information-related tokens. We perform comprehensive experiments on three datasets to show that SG-SRNet brings significant performance enhancements with low computational complexity.
引用
收藏
页数:13
相关论文
共 72 条
[21]  
Hu J, 2018, ADV NEUR IN, V31
[22]   Fast and Accurate Single Image Super-Resolution via Information Distillation Network [J].
Hui, Zheng ;
Wang, Xiumei ;
Gao, Xinbo .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :723-731
[23]   Learning multiple attention transformer super-resolution method for grape disease recognition [J].
Jin, Haibin ;
Chu, Xiaoquan ;
Qi, Jianfang ;
Feng, Jianying ;
Mu, Weisong .
EXPERT SYSTEMS WITH APPLICATIONS, 2024, 241
[24]  
Kim J, 2016, PROC CVPR IEEE, P1637, DOI [10.1109/CVPR.2016.181, 10.1109/CVPR.2016.182]
[25]  
Kitaev N., 2020, 8 INT C LEARNING REP
[26]   Unsupervised MRI Reconstruction via Zero-Shot Learned Adversarial Transformers [J].
Korkmaz, Yilmaz ;
Dar, Salman U. H. ;
Yurt, Mahmut ;
Ozbey, Muzaffer ;
Cukur, Tolga .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2022, 41 (07) :1747-1763
[27]   Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network [J].
Ledig, Christian ;
Theis, Lucas ;
Huszar, Ferenc ;
Caballero, Jose ;
Cunningham, Andrew ;
Acosta, Alejandro ;
Aitken, Andrew ;
Tejani, Alykhan ;
Totz, Johannes ;
Wang, Zehan ;
Shi, Wenzhe .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :105-114
[28]  
Lee J, 2020, I S BIOMED IMAGING, P917, DOI [10.1109/isbi45749.2020.9098416, 10.1109/ISBI45749.2020.9098416]
[29]   Transformer-empowered Multi-scale Contextual Matching and Aggregation for Multi-contrast MRI Super-resolution [J].
Li, Guangyuan ;
Lv, Jun ;
Tian, Yapeng ;
Dou, Qi ;
Wang, Chengyan ;
Xu, Chenliang ;
Qin, Jing .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :20604-20613
[30]   High-Resolution Pelvic MRI Reconstruction Using a Generative Adversarial Network With Attention and Cyclic Loss [J].
Li, Guangyuan ;
Lv, Jun ;
Tong, Xiangrong ;
Wang, Chengyan ;
Yang, Guang .
IEEE ACCESS, 2021, 9 :105951-105964