Multi-modal degradation feature learning for unified image restoration based on contrastive learning

被引:0
|
作者
Chen, Lei [1 ]
Xiong, Qingbo [1 ]
Zhang, Wei [1 ,2 ]
Liang, Xiaoli [1 ]
Gan, Zhihua [1 ]
Li, Liqiang [3 ]
He, Xin [1 ]
机构
[1] Henan Univ, Sch Software, Jinming Rd, Kaifeng 475004, Peoples R China
[2] China Univ Labor Relat, Sch Appl Technol, Zengguang Rd, Beijing 100048, Peoples R China
[3] Shangqiu Normal Univ, Sch Phys, Shangqiu 476000, Peoples R China
基金
美国国家科学基金会;
关键词
Unified image restoration; Multi-modal features; Contrastive learning; Deep learning;
D O I
10.1016/j.neucom.2024.128955
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we address the unified image restoration challenge by reframing it as a contrastive learning- based classification problem. Despite the significant strides made by deep learning methods in enhancing image restoration quality, their limited capacity to generalize across diverse degradation types and intensities necessitates the training of separate models for each specific degradation scenario. We proposes an all- encompassing approach that can restore images from various unknown corruption types and levels. We devise a method that learns representations of the latent sharp image's degradation and accompanying textual features (such as dataset categories and image content descriptions), converting these into prompts which are then embedded within a reconstruction network model to enhance cross-database restoration performance. This culminates in a unified image reconstruction framework. The study involves two stages: In the first stage, we design a MultiContentNet that learns multi-modal features (MMFs) of the latent sharp image. This network encodes the visual degradation expressions and contextual text features into latent variables, thereby exerting a guided classification effect. Specifically, MultiContentNet is trained as an auxiliary controller capable of taking the degraded input image and, through contrastive learning, extracts MMFs of the latent target image. This effectively generates natural classifiers tailored for different degradation types. The second phase integrates the learned MMFs into an image restoration network via cross-attention mechanisms. This guides the restoration model to learn high-fidelity image recovery. Experiments conducted on six blind image restoration tasks demonstrate that the proposed method achieves state-of-the-art performance, highlighting the potential significance of large-scale pretrained vision-language models' MMFs in advancing high-quality unified image reconstruction.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Overcoming the challenges of multi-modal medical image sharing: A novel data distillation strategy via contrastive learning
    Du, Taoli
    Li, Wenhui
    Wang, Zeyu
    Yang, Feiyang
    Teng, Peihong
    Yi, Xingcheng
    Chen, Hongyu
    Wang, Zixuan
    Zhang, Ping
    Zhang, Tianyang
    NEUROCOMPUTING, 2025, 617
  • [22] PolSAR Image Classification Based on Multi-Modal Contrastive Fully Convolutional Network
    Hua, Wenqiang
    Wang, Yi
    Yang, Sijia
    Jin, Xiaomin
    REMOTE SENSING, 2024, 16 (02)
  • [23] Multi-feature contrastive learning for unpaired image-to-image translation
    Gou, Yao
    Li, Min
    Song, Yu
    He, Yujie
    Wang, Litao
    COMPLEX & INTELLIGENT SYSTEMS, 2023, 9 (04) : 4111 - 4122
  • [24] Multi-feature contrastive learning for unpaired image-to-image translation
    Yao Gou
    Min Li
    Yu Song
    Yujie He
    Litao Wang
    Complex & Intelligent Systems, 2023, 9 : 4111 - 4122
  • [25] ETSCL: An Evidence Theory-Based Supervised Contrastive Learning Framework for Multi-modal Glaucoma Grading
    Yang, Zhiyuan
    Zhang, Bo
    Shi, Yufei
    Zhong, Ningze
    Loh, Johnathan
    Fang, Huihui
    Xu, Yanwu
    Yeo, Si Yong
    OPHTHALMIC MEDICAL IMAGE ANALYSIS, OMIA 2024, 2025, 15188 : 11 - 21
  • [26] Multi-modal Robustness Fake News Detection with Cross-Modal and Propagation Network Contrastive Learning
    Chen, Han
    Wang, Hairong
    Liu, Zhipeng
    Li, Yuhua
    Hu, Yifan
    Zhang, Yujing
    Shu, Kai
    Li, Ruixuan
    Yu, Philip S.
    KNOWLEDGE-BASED SYSTEMS, 2025, 309
  • [27] A Multi-modal Framework with Contrastive Learning and Sequential Encoding for Enhanced Sleep Stage Detection
    Wang, Zehui
    Zhang, Zhihan
    Wang, Hongtao
    PATTERN RECOGNITION AND COMPUTER VISION, PT V, PRCV 2024, 2025, 15035 : 3 - 17
  • [28] FMCS: Improving Code Search by Multi-Modal Representation Fusion and Momentum Contrastive Learning
    Liu, Wenjie
    Chen, Gong
    Xie, Xiaoyuan
    2024 IEEE 24TH INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY AND SECURITY, QRS, 2024, : 632 - 638
  • [29] Multi-modal brain tumor segmentation via disentangled representation learning and region-aware contrastive learning
    Zhou, Tongxue
    PATTERN RECOGNITION, 2024, 149 (149)
  • [30] Multi-Modal Feature Fusion-Based Machine Learning to Detect Abnormal Mechanical Ventilation
    Zhang, Huaqing
    Wang, Lizhu
    Xu, Jianfeng
    Xiang, Yan
    Zhang, Zhaocai
    IEEE ACCESS, 2023, 11 : 118291 - 118300