U-Shape Transformer for Underwater Image Enhancement

被引:215
作者
Peng, Lintao [1 ,2 ]
Zhu, Chunli [1 ,2 ]
Bian, Liheng [1 ,2 ]
机构
[1] Beijing Inst Technol, MIIT Key Lab Complex Field Intelligent Sensing, Beijing 100081, Peoples R China
[2] Beijing Inst Technol Jiaxing, Yangtze Delta Reg Acad, Jiaxing 314019, Peoples R China
基金
中国国家自然科学基金;
关键词
Image color analysis; Visualization; Imaging; Circuit faults; Attenuation; Transformers; Task analysis; Underwater image enhancement; transformer; multi-color space loss function; underwater image dataset; WATER;
D O I
10.1109/TIP.2023.3276332
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The light absorption and scattering of underwater impurities lead to poor underwater imaging quality. The existing data-driven based underwater image enhancement (UIE) techniques suffer from the lack of a large-scale dataset containing various underwater scenes and high-fidelity reference images. Besides, the inconsistent attenuation in different color channels and space areas is not fully considered for boosted enhancement. In this work, we built a large scale underwater image (LSUI) dataset, which covers more abundant underwater scenes and better visual quality reference images than existing underwater datasets. The dataset contains 4279 real-world underwater image groups, in which each raw image's clear reference images, semantic segmentation map and medium transmission map are paired correspondingly. We also reported an U-shape Transformer network where the transformer model is for the first time introduced to the UIE task. The U-shape Transformer is integrated with a channel-wise multi-scale feature fusion transformer (CMSFFT) module and a spatial-wise global feature modeling transformer (SGFMT) module specially designed for UIE task, which reinforce the network's attention to the color channels and space areas with more serious attenuation. Meanwhile, in order to further improve the contrast and saturation, a novel loss function combining RGB, LAB and LCH color spaces is designed following the human vision principle. The extensive experiments on available datasets validate the state-of-the-art performance of the reported technique with more than 2dB superiority. The dataset and demo code are available at https://bianlab.github.io/.
引用
收藏
页码:3066 / 3079
页数:14
相关论文
共 68 条
  • [51] A Rapid Scene Depth Estimation Model Based on Underwater Light Attenuation Prior for Underwater Image Restoration
    Song, Wei
    Wang, Yan
    Huang, Dongmei
    Tjondronegoro, Dian
    [J]. ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT I, 2018, 11164 : 678 - 688
  • [52] MAXIM: Multi-Axis MLP for Image Processing
    Tu, Zhengzhong
    Talebi, Hossein
    Zhang, Han
    Yang, Feng
    Milanfar, Peyman
    Bovik, Alan
    Li, Yinxiao
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 5759 - 5770
  • [53] Ulyanov D, 2017, Arxiv, DOI [arXiv:1607.08022, DOI 10.48550/ARXIV.1607.08022]
  • [54] Uplavikar Pritish M., 2019, CVPR WORKSHOPS, P1
  • [55] Vaswani A, 2017, ADV NEUR IN, V30
  • [56] Wang HN, 2022, AAAI CONF ARTIF INTE, P2441
  • [57] Single Underwater Image Restoration Using Adaptive Attenuation-Curve Prior
    Wang, Yi
    Liu, Hui
    Chau, Lap-Pui
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (03) : 992 - 1002
  • [58] Uformer: A General U-Shaped Transformer for Image Restoration
    Wang, Zhendong
    Cun, Xiaodong
    Bao, Jianmin
    Zhou, Wengang
    Liu, Jianzhuang
    Li, Houqiang
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 17662 - 17672
  • [59] Underwater image enhancement based on conditional generative adversarial network
    Yang, Miao
    Hu, Ke
    Du, Yixiang
    Wei, Zhiqiang
    Sheng, Zhibin
    Hu, Jintong
    [J]. SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 81 (81)
  • [60] An In-Depth Survey of Underwater Image Enhancement and Restoration
    Yang, Miao
    Hu, Jintong
    Li, Chongyi
    Rohde, Gustavo
    Du, Yixiang
    Hu, Ke
    [J]. IEEE ACCESS, 2019, 7 : 123638 - 123657