Hierarchical accumulation network with grid attention for image super-resolution

被引:10
|
作者
Yang, Yue [1 ]
Qi, Yong [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Comp Sci & Technol, Xian, Shaanxi, Peoples R China
关键词
Image super-resolution; Grouping; Attention mechanism; Accumulation network;
D O I
10.1016/j.knosys.2021.107520
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep convolutional neural networks (CNNs) have recently shown promising results in single image super-resolution (SISR) due to their powerful representation ability. However, existing CNN-based SR methods mainly focus on deeper architecture design to obtain high-level semantic information, neglecting the features of intermediate layers containing fine-grained texture information and thus limiting the capacity for producing precise high-resolution images. To tackle this issue, we propose a hierarchical accumulation network (HAN) with grid attention in this paper. Specifically, a hierarchical feature accumulation (HFA) structure is proposed to accumulate outputs of intermediate layers in a grouping manner for exploiting the features of different semantic levels. Moreover, we introduce a multi-scale grid attention module (MGAM) to refine features of the same level. The MGAM employs a pyramid sampling with self-attention mechanism to efficiently model the non-local dependencies between pixel features and produces refined representations. By this means, the universal features in connection with spatial similarity and semantic levels are produced for image SR. Experimental results on five benchmark datasets with different degradation models demonstrate the superiority of our HAN in terms of quantitative metrics and visual quality. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Single-image super-resolution with multilevel residual attention network
    Ding Qin
    Xiaodong Gu
    Neural Computing and Applications, 2020, 32 : 15615 - 15628
  • [22] Residual Triplet Attention Network for Single-Image Super-Resolution
    Huang, Feng
    Wang, Zhifeng
    Wu, Jing
    Shen, Ying
    Chen, Liqiong
    ELECTRONICS, 2021, 10 (17)
  • [23] Lightweight image super-resolution with multiscale residual attention network
    Xiao, Cunjun
    Dong, Hui
    Li, Haibin
    Li, Yaqian
    Zhang, Wenming
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (04)
  • [24] Lightweight image super-resolution with sliding Proxy Attention Network
    Hu, Zhenyu
    Sun, Wanjie
    Chen, Zhenzhong
    SIGNAL PROCESSING, 2025, 227
  • [25] PYRAMID FUSION ATTENTION NETWORK FOR SINGLE IMAGE SUPER-RESOLUTION
    He, Hao
    Du, Zongcai
    Li, Wenfeng
    Tang, Jie
    Wu, Gangshan
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2165 - 2169
  • [26] Multi-scale attention network for image super-resolution
    Wang, Li
    Shen, Jie
    Tang, E.
    Zheng, Shengnan
    Xu, Lizhong
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2021, 80
  • [27] Efficient residual attention network for single image super-resolution
    Fangwei Hao
    Taiping Zhang
    Linchang Zhao
    Yuanyan Tang
    Applied Intelligence, 2022, 52 : 652 - 661
  • [28] A Novel Attention Enhanced Dense Network for Image Super-Resolution
    Niu, Zhong-Han
    Zhou, Yang-Hao
    Yang, Yu-Bin
    Fan, Jian-Cong
    MULTIMEDIA MODELING (MMM 2020), PT I, 2020, 11961 : 568 - 580
  • [29] Dynamic dual attention iterative network for image super-resolution
    Feng, Hao
    Wang, Liejun
    Cheng, Shuli
    Du, Anyu
    Li, Yongming
    APPLIED INTELLIGENCE, 2022, 52 (07) : 8189 - 8208
  • [30] Structured Fusion Attention Network for Image Super-Resolution Reconstruction
    Dai, Yaonan
    Yu, Jiuyang
    Hu, Tianhao
    Lu, Yang
    Zheng, Xiaotao
    IEEE ACCESS, 2022, 10 : 31896 - 31906