Multi-modal deep convolutional dictionary learning for image denoising

被引:6
作者
Sun, Zhonggui [1 ,2 ]
Zhang, Mingzhu [1 ]
Sun, Huichao [1 ]
Li, Jie [2 ]
Liu, Tingting [3 ]
Gao, Xinbo [3 ]
机构
[1] Liaocheng Univ, Sch Math Sci, Liaocheng 252000, Peoples R China
[2] Xidian Univ, Sch Elect Engn, Video & Image Proc Syst Lab, Xian 710071, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep convolutional dictionary learning; Multi-modal; Channel attention; Image denoising; SPARSE; REMOVAL;
D O I
10.1016/j.neucom.2023.126918
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Leveraging the capabilities of traditional dictionary learning (DicL) and drawing upon the success of deep neural networks (DNNs), the recently proposed framework of deep convolutional dictionary learning (DCDicL) has exhibited remarkable behaviours in image denoising. Note that, the application of the DCDicL method is confined to single modality scenarios, whereas the images in practice often originate from diverse modalities. In this paper, to broaden the application scope of the DCDicL method, we design a multi-modal version of it, dubbed MMDCDicL. Specifically, within the mathematical model of MMDCDicL, we adopt an analytical approach to tackle the sub-problem linked to the guidance modality, harnessing its inherent reliability. Meanwhile, like in DCDicL, we utilize a network-based learning approach for the noisy modality to extract trustworthy information from the data. Based on the solution, we establish an interpretable network structure for MMDCDicL. Additionally, wherein, we design a multi-kernel channel attention block (MKCAB) in the structure to efficiently integrate the information from diverse modalities. Experimental results suggest that MMDCDicL can reconstruct higher-quality outcomes both quantitatively and perceptually. Code is available at http://www.diplab.net/lunwen/mmdcdicl.htm.
引用
收藏
页数:11
相关论文
共 47 条
[1]   A Dataset of Flash and Ambient Illumination Pairs from the Crowd [J].
Aksoy, Yagiz ;
Kim, Changil ;
Kellnhofer, Petr ;
Paris, Sylvain ;
Elgharib, Mohamed ;
Pollefeys, Marc ;
Matusik, Wojciech .
COMPUTER VISION - ECCV 2018, PT IX, 2018, 11213 :644-660
[2]   Multimodal Task-Driven Dictionary Learning for Image Classification [J].
Bahrampour, Soheil ;
Nasrabadi, Nasser M. ;
Ray, Asok ;
Jenkins, William Kenneth .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (01) :24-38
[3]   On the Global-Local Dichotomy in Sparsity Modeling [J].
Batenkov, Dmitry ;
Romano, Yaniv ;
Elad, Michael .
COMPRESSED SENSING AND ITS APPLICATIONS, 2017, :1-53
[4]  
Brown M, 2011, PROC CVPR IEEE, P177, DOI 10.1109/CVPR.2011.5995637
[5]   Siamese Dense Network for Reflection Removal with Flash and No-Flash Image Pairs [J].
Chang, Yakun ;
Jung, Cheolkon ;
Sun, Jun ;
Wang, Fengqiao .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (06) :1673-1698
[6]  
CHEN SB, 1994, CONF REC ASILOMAR C, P41, DOI 10.1109/ACSSC.1994.471413
[7]   Sparse embedded dictionary learning on face recognition [J].
Chen, Yefei ;
Su, Jianbo .
PATTERN RECOGNITION, 2017, 64 :51-59
[8]   Image restoration by sparse 3D transform-domain collaborative filtering [J].
Dabov, Kostadin ;
Foi, Alessandro ;
Katkovnik, Vladimir ;
Egiazarian, Karen .
IMAGE PROCESSING: ALGORITHMS AND SYSTEMS VI, 2008, 6812
[9]   An iterative thresholding algorithm for linear inverse problems with a sparsity constraint [J].
Daubechies, I ;
Defrise, M ;
De Mol, C .
COMMUNICATIONS ON PURE AND APPLIED MATHEMATICS, 2004, 57 (11) :1413-1457
[10]   Deep Convolutional Neural Network for Multi-Modal Image Restoration and Fusion [J].
Deng, Xin ;
Dragotti, Pier Luigi .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (10) :3333-3348