Spatial relaxation transformer for image super-resolution

被引:1
作者
Li, Yinghua [1 ]
Zhang, Ying [1 ]
Zeng, Hao [3 ]
He, Jinglu [1 ]
Guo, Jie [2 ]
机构
[1] Xian Univ Posts & Telecommun, Xian Key Lab Image Proc Technol & Applicat Publ Se, Changan West St, Xian 710121, Shaanxi, Peoples R China
[2] Xidian Univ, State Key Lab Integrated Serv Networks, 2 Southern Tai Bai Rd, Xian 710071, Shaanxi, Peoples R China
[3] Chinese Acad Sci, Inst Software, Beijing, Peoples R China
关键词
Super-resolution; Vision transformer; Feature aggregation; Image enhancement; Swin transformer;
D O I
10.1016/j.jksuci.2024.102150
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Transformer-based approaches have demonstrated remarkable performance in image processing tasks due to their ability to model long-range dependencies. Current mainstream Transformer-based methods typically confine self-attention computation within windows to reduce computational burden. However, this constraint may lead to grid artifacts in the reconstructed images due to insufficient cross-window information exchange, particularly in image super-resolution tasks. To address this issue, we propose the Multi-Scale Texture Complementation Block based on Spatial Relaxation Transformer (MSRT), which leverages features at multiple scales and augments information exchange through cross windows attention computation. In addition, we introduce a loss function based on the prior of texture smoothness transformation, which utilizes the continuity of textures between patches to constrain the generation of more coherent texture information in the reconstructed images. Specifically, we employ learnable compressive sensing technology to extract shallow features from images, preserving image features while reducing feature dimensions and improving computational efficiency. Extensive experiments conducted on multiple benchmark datasets demonstrate that our method outperforms previous state-of-the-art approaches in both qualitative and quantitative evaluations.
引用
收藏
页数:10
相关论文
共 51 条
[1]   Fast, Accurate, and Lightweight Super-Resolution with Cascading Residual Network [J].
Ahn, Namhyuk ;
Kang, Byungkon ;
Sohn, Kyung-Ah .
COMPUTER VISION - ECCV 2018, PT X, 2018, 11214 :256-272
[2]   Low-Complexity Single-Image Super-Resolution based on Nonnegative Neighbor Embedding [J].
Bevilacqua, Marco ;
Roumy, Aline ;
Guillemot, Christine ;
Morel, Marie-Line Alberi .
PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2012, 2012,
[3]   Dual Aggregation Transformer for Image Super-Resolution [J].
Chen, Zheng ;
Zhang, Yulun ;
Gu, Jinjin ;
Kong, Linghe ;
Yang, Xiaokang ;
Yu, Fisher .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, :12278-12287
[4]  
Chen Zheng, 2022, Advances in Neural Information Processing Systems
[5]   Second-order Attention Network for Single Image Super-Resolution [J].
Dai, Tao ;
Cai, Jianrui ;
Zhang, Yongbing ;
Xia, Shu-Tao ;
Zhang, Lei .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11057-11066
[6]   Swin transformer based vehicle detection in undisciplined traffic environment [J].
Deshmukh, Prashant ;
Satyanarayana, G. S. R. ;
Majhi, Sudhan ;
Sahoo, Upendra Kumar ;
Das, Santos Kumar .
EXPERT SYSTEMS WITH APPLICATIONS, 2023, 213
[7]   Accelerating the Super-Resolution Convolutional Neural Network [J].
Dong, Chao ;
Loy, Chen Change ;
Tang, Xiaoou .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :391-407
[8]   Learning a Deep Convolutional Network for Image Super-Resolution [J].
Dong, Chao ;
Loy, Chen Change ;
He, Kaiming ;
Tang, Xiaoou .
COMPUTER VISION - ECCV 2014, PT IV, 2014, 8692 :184-199
[9]  
Dosovitskiy A, 2021, Arxiv, DOI [arXiv:2010.11929, DOI 10.48550/ARXIV.2010.11929]
[10]   A Hybrid Network of CNN and Transformer for Lightweight Image Super-Resolution [J].
Fang, Jinsheng ;
Lin, Hanjiang ;
Chen, Xinyu ;
Zeng, Kun .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, :1102-1111