Heterogeneous Window Transformer for Image Denoising

被引:22
作者
Tian, Chunwei [1 ,2 ]
Zheng, Menghua [1 ]
Lin, Chia-Wen [3 ,4 ]
Li, Zhiwu [5 ]
Zhang, David [6 ,7 ]
机构
[1] Northwestern Polytech Univ, Sch Software, Xian 710072, Peoples R China
[2] Natl Engn Lab Integrated Aerospace Ground Ocean B, Xian 710129, Peoples R China
[3] Natl Tsinghua Univ, Dept Elect Engn, Hsinchu 300, Taiwan
[4] Natl Tsinghua Univ, Inst Commun Engn, Hsinchu 300, Taiwan
[5] Xidian Univ, Sch Electormech Engn, Xian 710071, Peoples R China
[6] Chinese Univ Hong Kong Shenzhen, Sch Data Sci, Shenzhen 518172, Peoples R China
[7] Shenzhen Inst Artificial Intelligence & Robot Soc, Shenzhen 518172, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2024年 / 54卷 / 11期
基金
中国国家自然科学基金;
关键词
Noise reduction; Image denoising; Transformers; Noise measurement; Data mining; Training; Tensors; Convolutional neural network (CNN); image denoising; image watermark removal; self-supervised learning; task decomposition; NETWORK; CNN;
D O I
10.1109/TSMC.2024.3429345
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep networks can usually depend on extracting more structural information to improve denoising results. However, they may ignore correlation between pixels from an image to pursue better-denoising performance. Window Transformer can use long- and short-distance modeling to interact pixels to address mentioned problem. To make a tradeoff between distance modeling and denoising time, we propose a heterogeneous window Transformer (HWformer) for image denoising. HWformer first designs heterogeneous global windows to capture global context information for improving denoising effects. To build a bridge between long and short-distance modeling, global windows are horizontally and vertically shifted to facilitate diversified information without increasing denoising time. To prevent the information loss phenomenon of independent patches, sparse idea is guided a feed-forward network to extract local information of neighboring patches. The proposed HWformer only takes 30% of popular restoration Transformer in terms of denoising time. Its codes can be obtained at <uri>https://github.com/hellloxiaotian/HWformer</uri>.
引用
收藏
页码:6621 / 6632
页数:12
相关论文
共 64 条
[1]   A High-Quality Denoising Dataset for Smartphone Cameras [J].
Abdelhamed, Abdelrahman ;
Lin, Stephen ;
Brown, Michael S. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :1692-1700
[2]   NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study [J].
Agustsson, Eirikur ;
Timofte, Radu .
2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, :1122-1131
[3]   MEAN SQUARE ERROR OF PREDICTION AS A CRITERION FOR SELECTING VARIABLES [J].
ALLEN, DM .
TECHNOMETRICS, 1971, 13 (03) :469-&
[4]   Real Image Denoising with Feature Attention [J].
Anwar, Saeed ;
Barnes, Nick .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3155-3164
[5]   Contour Detection and Hierarchical Image Segmentation [J].
Arbelaez, Pablo ;
Maire, Michael ;
Fowlkes, Charless ;
Malik, Jitendra .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (05) :898-916
[6]  
Ba J L., LAYER NORMALIZATION
[7]   Fast Gradient-Based Algorithms for Constrained Total Variation Image Denoising and Deblurring Problems [J].
Beck, Amir ;
Teboulle, Marc .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2009, 18 (11) :2419-2434
[8]   Spatial-Adaptive Network for Single Image Denoising [J].
Chang, Meng ;
Li, Qi ;
Feng, Huajun ;
Xu, Zhihai .
COMPUTER VISION - ECCV 2020, PT XXX, 2020, 12375 :171-187
[9]   Pre-Trained Image Processing Transformer [J].
Chen, Hanting ;
Wang, Yunhe ;
Guo, Tianyu ;
Xu, Chang ;
Deng, Yiping ;
Liu, Zhenhua ;
Ma, Siwei ;
Xu, Chunjing ;
Xu, Chao ;
Gao, Wen .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :12294-12305
[10]   Trainable Nonlinear Reaction Diffusion: A Flexible Framework for Fast and Effective Image Restoration [J].
Chen, Yunjin ;
Pock, Thomas .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (06) :1256-1272