A multi-scale enhanced large-kernel attention transformer network for lightweight image super-resolution

被引:0
作者
Chang, Kairong [1 ]
Jun, Sun [1 ]
Biao, Yang [1 ]
Hu, Mingzhi [1 ]
Yang, Junlong [1 ]
机构
[1] Kunming Univ Sci & Technol, Sch Informat Engn & Automat, Kunming 650500, Peoples R China
基金
中国国家自然科学基金;
关键词
Image super-resolution; Lightweight; Multi-scale; Attention mechanism; Transformer;
D O I
10.1007/s11760-024-03790-1
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To address the issues of detail loss, natural texture distortion, small receptive field, overly smooth reconstructed images, and the limitations of single-scale convolution kernels in existing super-resolution models, we have designed a lightweight super-resolution network based on a multi-scale enhanced large-kernel attention transformer (MELTN). First, we propose a multi-scale shallow feature extraction module to expand the feature space, using blueprint separable convolutions instead of standard convolutions to reduce computational and memory requirements. Second, we construct the MELKAFormer module, based on multi-scale enhanced large-kernel attention and an efficient feedforward network, as the core component for efficient feature extraction. Third, we use edge-oriented convolution block and introduce a residual structure at the network output to effectively capture image edge details and texture features. Experimental results show that the proposed model maintains high image quality with significantly fewer parameters and FLOPs compared to other deep neural network models across various benchmark datasets, demonstrating the effectiveness of the model in terms of image quality and computational efficiency.
引用
收藏
页数:9
相关论文
共 28 条
[1]   HIPA: Hierarchical Patch Transformer for Single Image Super Resolution [J].
Cai, Qing ;
Qian, Yiming ;
Li, Jinxing ;
Lyu, Jun ;
Yang, Yee-Hong ;
Wu, Feng ;
Zhang, David .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 :3226-3237
[2]  
Cao H., 2023, Highlights Sci. Eng. Technol, V39, P1301, DOI [10.54097/hset.v39i.6759, DOI 10.54097/HSET.V39I.6759]
[3]   N-Gram in Swin Transformers for Efficient Lightweight Image Super-Resolution [J].
Choi, Haram ;
Lee, Jeongmin ;
Yang, Jihoon .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, :2071-2081
[4]   CFGN: A Lightweight Context Feature Guided Network for Image Super-Resolution [J].
Dai, Tao ;
Ya, Mengxi ;
Li, Jinmin ;
Zhang, Xinyi ;
Xia, Shu-Tao ;
Zhu, Zexuan .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (01) :855-865
[5]   LKASR: Large kernel attention for lightweight image super-resolution [J].
Feng, Hao ;
Wang, Liejun ;
Li, Yongming ;
Du, Anyu .
KNOWLEDGE-BASED SYSTEMS, 2022, 252
[6]  
Gao G., 2022, arXiv, DOI 10.48550/arXiv.2204.13286
[7]   Video Super-Resolution Based on Inter-Frame Information Utilization for Intelligent Transportation [J].
Guo, Kehua ;
Zhang, Zhuo ;
Guo, Haifu ;
Ren, Sheng ;
Wang, Lei ;
Zhou, Xiaokang ;
Liu, Chao .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (11) :13409-13421
[8]   Visual attention network [J].
Guo, Meng-Hao ;
Lu, Cheng-Ze ;
Liu, Zheng-Ning ;
Cheng, Ming-Ming ;
Hu, Shi-Min .
COMPUTATIONAL VISUAL MEDIA, 2023, 9 (04) :733-752
[9]   Rethinking Depthwise Separable Convolutions: How Intra-Kernel Correlations Lead to Improved MobileNets [J].
Haase, Daniel ;
Amthor, Manuel .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :14588-14597
[10]  
Hao FW, 2024, Arxiv, DOI arXiv:2405.09353