High-Similarity-Pass Attention for Single Image Super-Resolution

被引:15
作者
Su, Jian-Nan [1 ,2 ,3 ]
Gan, Min [1 ,4 ]
Chen, Guang-Yong [1 ,2 ,3 ]
Guo, Wenzhong [1 ,2 ,3 ]
Chen, C. L. Philip [1 ,5 ]
机构
[1] Fuzhou Univ, Coll Comp & Data Sci, Fuzhou 350108, Peoples R China
[2] Minist Educ, Key Lab Intelligent Metro Univ Fujian, Fujian Key Lab Network Comp & Intelligent Informa, Fuzhou, Peoples R China
[3] Minist Educ, Engn Res Ctr Big Data Intelligence, Fuzhou 350108, Peoples R China
[4] Qingdao Univ, Coll Comp Sci & Technol, Qingdao 266071, Peoples R China
[5] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510641, Peoples R China
基金
中国国家自然科学基金;
关键词
High-similarity-pass attention; softmax transformation; single image super-resolution; deep learning; SPARSE; REPRESENTATIONS;
D O I
10.1109/TIP.2023.3348293
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent developments in the field of non-local attention (NLA) have led to a renewed interest in self-similarity-based single image super-resolution (SISR). Researchers usually use the NLA to explore non-local self-similarity (NSS) in SISR and achieve satisfactory reconstruction results. However, a surprising phenomenon that the reconstruction performance of the standard NLA is similar to that of the NLA with randomly selected regions prompted us to revisit NLA. In this paper, we first analyzed the attention map of the standard NLA from different perspectives and discovered that the resulting probability distribution always has full support for every local feature, which implies a statistical waste of assigning values to irrelevant non-local features, especially for SISR which needs to model long-range dependence with a large number of redundant non-local features. Based on these findings, we introduced a concise yet effective soft thresholding operation to obtain high-similarity-pass attention (HSPA), which is beneficial for generating a more compact and interpretable distribution. Furthermore, we derived some key properties of the soft thresholding operation that enable training our HSPA in an end-to-end manner. The HSPA can be integrated into existing deep SISR models as an efficient general building block. In addition, to demonstrate the effectiveness of the HSPA, we constructed a deep high-similarity-pass attention network (HSPAN) by integrating a few HSPAs in a simple backbone. Extensive experimental results demonstrate that HSPAN outperforms state-of-the-art approaches on both quantitative and qualitative evaluations. Our code and a pre-trained model were uploaded to GitHub (https://github.com/laoyangui/HSPAN) for validation.
引用
收藏
页码:610 / 624
页数:15
相关论文
共 67 条
[1]   Densely Residual Laplacian Super-Resolution [J].
Anwar, Saeed ;
Barnes, Nick .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (03) :1192-1204
[2]   Low-Complexity Single-Image Super-Resolution based on Nonnegative Neighbor Embedding [J].
Bevilacqua, Marco ;
Roumy, Aline ;
Guillemot, Christine ;
Morel, Marie-Line Alberi .
PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2012, 2012,
[3]  
Candes E., 2006, Proc. Int. Congr. Math., V3, P1433, DOI DOI 10.4171/022-3/69
[4]   Deep MR Brain Image Super-Resolution Using Spatio-Structural Priors [J].
Cherukuri, Venkateswararao ;
Guo, Tiantong ;
Schiff, Steven J. ;
Monga, Vishal .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :1368-1383
[5]   Second-order Attention Network for Single Image Super-Resolution [J].
Dai, Tao ;
Cai, Jianrui ;
Zhang, Yongbing ;
Xia, Shu-Tao ;
Zhang, Lei .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11057-11066
[6]   Accelerating the Super-Resolution Convolutional Neural Network [J].
Dong, Chao ;
Loy, Chen Change ;
Tang, Xiaoou .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :391-407
[7]   Image Super-Resolution Using Deep Convolutional Networks [J].
Dong, Chao ;
Loy, Chen Change ;
He, Kaiming ;
Tang, Xiaoou .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (02) :295-307
[8]   Compressed sensing [J].
Donoho, DL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2006, 52 (04) :1289-1306
[9]  
Duchi J., 2008, P 25 INT C MACH LEAR, P272, DOI [DOI 10.1145/1390156.1390191, 10.1145/1390156.1390191]
[10]  
Ebrahimi M, 2007, LECT NOTES COMPUT SC, V4633, P117