USCFormer: Unified Transformer With Semantically Contrastive Learning for Image Dehazing

被引:19
作者
Wang, Yongzhen [1 ]
Xiong, Jiamei [1 ]
Yan, Xuefeng [1 ,2 ]
Wei, Mingqiang [3 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Sch Comp Sci & Technol, Nanjing 210016, Peoples R China
[2] Collaborat Innovat Ctr Novel Software Technol & In, Nanjing 210093, Peoples R China
[3] Nanjing Univ Aeronaut & Astronaut, Shenzhen Res Inst, Shenzhen 518038, Peoples R China
基金
中国国家自然科学基金;
关键词
USCFormer; image dehazing; unified transformer; semantically guided; contrastive learning; ENHANCEMENT; VISIBILITY; ALGORITHM;
D O I
10.1109/TITS.2023.3277709
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Haze severely degrades the visibility of scene objects and deteriorates the performance of autonomous driving, traffic monitoring, and other vision-based intelligent transportation systems. As a potential remedy, we propose a novel unified Transformer with semantically contrastive learning for image dehazing, dubbed USCFormer. USCFormer has three key contributions. First, USCFormer absorbs the respective strengths of CNN and Transformer by incorporating them into a unified Transformer format. Thus, it allows the simultaneous capture of global-local dependency features for better image dehazing. Second, by casting clean/hazy images as the positive/negative samples, the contrastive constraint encourages the restored image to be closer to the ground-truth images (positives) and away from the hazy ones (negatives). Third, we regard the semantic information as important prior knowledge to help USCFormer mitigate the effects of haze on the scene and preserve image details and colors by leveraging intra-object semantic correlation. Experiments on synthetic datasets and real-world hazy photos fully validate the superiority of USCFormer in both perceptual quality assessment and subjective evaluation. Code is available at https://github.com/yz-wang/USCFormer.
引用
收藏
页码:11321 / 11333
页数:13
相关论文
共 69 条
  • [1] Assessment I.Q., 2004, IEEE Trans. Image Process., V13, P93
  • [2] Ba J, 2016, ARXIV160706450
  • [3] Self-Guided Image Dehazing Using Progressive Feature Fusion
    Bai, Haoran
    Pan, Jinshan
    Xiang, Xinguang
    Tang, Jinhui
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 1217 - 1229
  • [4] Non-Local Image Dehazing
    Berman, Dana
    Treibitz, Tali
    Avidan, Shai
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 1674 - 1682
  • [5] DehazeNet: An End-to-End System for Single Image Haze Removal
    Cai, Bolun
    Xu, Xiangmin
    Jia, Kui
    Qing, Chunmei
    Tao, Dacheng
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (11) : 5187 - 5198
  • [6] Rapid Detection of Blind Roads and Crosswalks by Using a Lightweight Semantic Segmentation Network
    Cao, Zhengcai
    Xu, Xiaowen
    Hu, Biao
    Zhou, MengChu
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (10) : 6188 - 6197
  • [7] Pyramid Stereo Matching Network
    Chang, Jia-Ren
    Chen, Yong-Sheng
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 5410 - 5418
  • [8] A Robust Moving Object Detection in Multi-Scenario Big Data for Video Surveillance
    Chen, Bo-Hao
    Shi, Ling-Feng
    Ke, Xiao
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (04) : 982 - 995
  • [9] Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation
    Chen, Liang-Chieh
    Zhu, Yukun
    Papandreou, George
    Schroff, Florian
    Adam, Hartwig
    [J]. COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 : 833 - 851
  • [10] Chen Ting, 2019, 25 AMERICAS C INFORM