Exploring the Application of Large-Scale Pre-Trained Models on Adverse Weather Removal

被引:2
|
作者
Tan, Zhentao [1 ]
Wu, Yue [1 ]
Liu, Qiankun [2 ,3 ]
Chu, Qi [2 ,3 ]
Lu, Le [1 ]
Ye, Jieping [1 ]
Yu, Nenghai [2 ,3 ]
机构
[1] Alibaba Grp, Hangzhou 310052, Peoples R China
[2] Univ Sci & Technol China USTC, Sch Cyber Sci & Technol, CAS Key Lab Electromagnet Space Informat, Hefei 230026, Peoples R China
[3] Univ Sci & Technol China USTC, Anhui Prov Key Lab Digital Secur, Hefei 230026, Peoples R China
关键词
Meteorology; Task analysis; Training; Semantics; Image restoration; Rain; Feature extraction; Adverse weather removal; image restoration; multi-modal pre-trained model; NETWORK;
D O I
10.1109/TIP.2024.3368961
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image restoration under adverse weather conditions (e.g., rain, snow, and haze) is a fundamental computer vision problem that has important implications for various downstream applications. Distinct from early methods that are specially designed for specific types of weather, recent works tend to simultaneously remove various adverse weather effects based on either spatial feature representation learning or semantic information embedding. Inspired by various successful applications incorporating large-scale pre-trained models (e.g., CLIP), in this paper, we explore their potential benefits for leveraging large-scale pre-trained models in this task based on both spatial feature representation learning and semantic information embedding aspects: 1) spatial feature representation learning, we design a Spatially Adaptive Residual (SAR) encoder to adaptively extract degraded areas. To facilitate training of this model, we propose a Soft Residual Distillation (CLIP-SRD) strategy to transfer spatial knowledge from CLIP between clean and adverse weather images; 2) semantic information embedding, we propose a CLIP Weather Prior (CWP) embedding module to enable the network to adaptively respond to different weather conditions. This module integrates the sample-specific weather priors extracted by the CLIP image encoder with the distribution-specific information (as learned by a set of parameters) and embeds these elements using a cross-attention mechanism. Extensive experiments demonstrate that our proposed method can achieve state-of-the-art performance under various and severe adverse weather conditions. The code will be made available.
引用
收藏
页码:1683 / 1698
页数:16
相关论文
共 50 条
  • [1] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey
    Xiao Wang
    Guangyao Chen
    Guangwu Qian
    Pengcheng Gao
    Xiao-Yong Wei
    Yaowei Wang
    Yonghong Tian
    Wen Gao
    Machine Intelligence Research, 2023, 20 : 447 - 482
  • [2] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey
    Wang, Xiao
    Chen, Guangyao
    Qian, Guangwu
    Gao, Pengcheng
    Wei, Xiao-Yong
    Wang, Yaowei
    Tian, Yonghong
    Gao, Wen
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (04) : 447 - 482
  • [3] FASTERMOE: Modeling and Optimizing Training of Large-Scale Dynamic Pre-Trained Models
    He, Jiaao
    Zhai, Jidong
    Antunes, Tiago
    Wang, Haojie
    Luo, Fuwen
    Shi, Shangfeng
    Li, Qin
    PPOPP'22: PROCEEDINGS OF THE 27TH ACM SIGPLAN SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING, 2022, : 120 - 134
  • [4] Alternating Recurrent Dialog Model with Large-scale Pre-trained Language Models
    Wu, Qingyang
    Zhang, Yichi
    Li, Yu
    Yu, Zhou
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1292 - 1301
  • [5] EXPLORING LARGE SCALE PRE-TRAINED MODELS FOR ROBUST MACHINE ANOMALOUS SOUND DETECTION
    Han, Bing
    Lv, Zhiqiang
    Jiang, Anbai
    Huang, Wen
    Chen, Zhengyang
    Deng, Yufeng
    Ding, Jiawei
    Lu, Cheng
    Zhang, Wei-Qiang
    Fan, Pingyi
    Liu, Jia
    Qian, Yanmin
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 1326 - 1330
  • [6] CPM-2: Large-scale cost-effective pre-trained language models
    Zhang, Zhengyan
    Gu, Yuxian
    Han, Xu
    Chen, Shengqi
    Xiao, Chaojun
    Sun, Zhenbo
    Yao, Yuan
    Qi, Fanchao
    Guan, Jian
    Ke, Pei
    Cai, Yanzheng
    Zeng, Guoyang
    Tan, Zhixing
    Liu, Zhiyuan
    Huang, Minlie
    Han, Wentao
    Liu, Yang
    Zhu, Xiaoyan
    Sun, Maosong
    AI OPEN, 2021, 2 : 216 - 224
  • [7] 3D Semantic Novelty Detection via Large-Scale Pre-Trained Models
    Rabino, Paolo
    Alliegro, Antonio
    Tommasi, Tatiana
    IEEE ACCESS, 2024, 12 : 135352 - 135361
  • [8] Training-Free Deepfake Voice Recognition by Leveraging Large-Scale Pre-Trained Models
    Pianese, Alessandro
    Poggi, Giovanni
    Cozzolino, Davide
    Verdoliva, Luisa
    PROCEEDINGS OF THE 2024 ACM WORKSHOP ON INFORMATION HIDING AND MULTIMEDIA SECURITY, IH&MMSEC 2024, 2024, : 289 - 294
  • [9] Parameter-efficient fine-tuning of large-scale pre-trained language models
    Ning Ding
    Yujia Qin
    Guang Yang
    Fuchao Wei
    Zonghan Yang
    Yusheng Su
    Shengding Hu
    Yulin Chen
    Chi-Min Chan
    Weize Chen
    Jing Yi
    Weilin Zhao
    Xiaozhi Wang
    Zhiyuan Liu
    Hai-Tao Zheng
    Jianfei Chen
    Yang Liu
    Jie Tang
    Juanzi Li
    Maosong Sun
    Nature Machine Intelligence, 2023, 5 : 220 - 235
  • [10] Adapting Large-Scale Pre-trained Models for Uni ed Dialect Speech Recognition Model
    Toyama, T.
    Kai, A.
    Kamiya, Y.
    Takahashi, N.
    Acta Physica Polonica A, 2024, 146 (04) : 413 - 418