Rethinking Image Deblurring via CNN-Transformer Multiscale Hybrid Architecture

被引:0
|
作者
Zhao, Qian [1 ]
Yang, Hao [1 ]
Zhou, Dongming [1 ]
Cao, Jinde [2 ,3 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650091, Peoples R China
[2] Southeast Univ, Sch Math, Nanjing 210096, Peoples R China
[3] Yonsei Univ, Yonsei Frontier Lab, Seoul 03722, South Korea
基金
中国国家自然科学基金;
关键词
Image deblurring; motion blur; multiscale strategy; neural networks; vision transformer (ViT);
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Image deblurring is a representative low-level vision task that aims to estimate latent sharp images from blurred images. Recently, convolutional neural network (CNN)-based methods have dominated image deblurring. However, traditional CNN-based deblurring methods suffer from two essential issues: first, existing multiscale deblurring methods process blurred images at different scales through sub-networks with the same composition, which limits the model performance. Second, the convolutional layers fail to adapt to the input content and cannot effectively capture long-range dependencies. To alleviate the above issues, we rethink the multiscale architecture that follows a coarse-to-fine strategy and propose a novel hybrid architecture that combines CNN and transformer (CTMS). CTMS has three distinct features. First, the finer-scale sub-networks in CTMS are designed as architectures with larger receptive fields to obtain the pixel values around the blur, which can be used to efficiently handle large-area blur. Then, we propose a feature modulation network to alleviate the disadvantages of CNN sub-networks that lack input content adaptation. Finally, we design an efficient transformer block, which significantly reduces the computational burden and requires no pre-training. Our proposed deblurring model is extensively evaluated on several benchmark datasets, and achieves superior performance compared to state-of-the-art deblurring methods. Especially, the peak signal to noise ratio (PSNR) and structural similarity (SSIM) values are 32.73 dB and 0.959, respectively, on the popular dataset GoPro. In addition, we conduct joint evaluation experiments on the proposed method deblurring performance, object detection, and image segmentation to demonstrate the effectiveness of CTMS for subsequent high-level computer vision tasks.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Rethinking Image Deblurring via CNN-Transformer Multiscale Hybrid Architecture
    Zhao, Qian
    Yang, Hao
    Zhou, Dongming
    Cao, Jinde
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [2] Rethinking Image Deblurring via CNN-Transformer Multiscale Hybrid Architecture
    Zhao, Qian
    Yang, Hao
    Zhou, Dongming
    Cao, Jinde
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [3] CNN-Transformer Hybrid Architecture for Underwater Sonar Image Segmentation
    Lei, Juan
    Wang, Huigang
    Lei, Zelin
    Li, Jiayuan
    Rong, Shaowei
    REMOTE SENSING, 2025, 17 (04)
  • [4] Image Deblurring Based on an Improved CNN-Transformer Combination Network
    Chen, Xiaolin
    Wan, Yuanyuan
    Wang, Donghe
    Wang, Yuqing
    APPLIED SCIENCES-BASEL, 2023, 13 (01):
  • [5] CNN-Transformer Hybrid Architecture for Early Fire Detection
    Yang, Chenyue
    Pan, Yixuan
    Cao, Yichao
    Lu, Xiaobo
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT IV, 2022, 13532 : 570 - 581
  • [6] Image harmonization with Simple Hybrid CNN-Transformer Network
    Li, Guanlin
    Zhao, Bin
    Li, Xuelong
    NEURAL NETWORKS, 2024, 180
  • [7] HCformer: Hybrid CNN-Transformer for LDCT Image Denoising
    Yuan, Jinli
    Zhou, Feng
    Guo, Zhitao
    Li, Xiaozeng
    Yu, Hengyong
    JOURNAL OF DIGITAL IMAGING, 2023, 36 (05) : 2290 - 2305
  • [8] HCformer: Hybrid CNN-Transformer for LDCT Image Denoising
    Jinli Yuan
    Feng Zhou
    Zhitao Guo
    Xiaozeng Li
    Hengyong Yu
    Journal of Digital Imaging, 2023, 36 (5) : 2290 - 2305
  • [9] TFCNs: A CNN-Transformer Hybrid Network for Medical Image Segmentation
    Li, Zihan
    Li, Dihan
    Xu, Cangbai
    Wang, Weice
    Hong, Qingqi
    Li, Qingde
    Tian, Jie
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT IV, 2022, 13532 : 781 - 792
  • [10] Hybrid CNN-Transformer Feature Fusion for Single Image Deraining
    Chen, Xiang
    Pan, Jinshan
    Lu, Jiyang
    Fan, Zhentao
    Li, Hao
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 378 - 386