A vision transformer based CNN for underwater image enhancement ViTClarityNet

被引:0
作者
Fathy, Mohamed E. [1 ]
Mohamed, Samer A. [1 ,3 ]
Awad, Mohammed I. [1 ]
Abd El Munim, Hossam E. [2 ]
机构
[1] Ain Shams Univ, Fac Engn, Mechatron Engn Dept, Cairo 11535, Egypt
[2] Ain Shams Univ, Fac Engn, Comp & Syst Engn Dept, Cairo 11535, Egypt
[3] Univ Bath, Fac Engn & Design, Dept Elect & Elect Engn, Bath BA2 7AY, England
关键词
Convolutional neural networks; Generative adversarial; Underwater image enhancement; Vision transformer; Synthetic dataset; SYSTEM;
D O I
10.1038/s41598-025-91212-8
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Underwater computer vision faces significant challenges from light scattering, absorption, and poor illumination, which severely impact underwater vision tasks. To address these issues, ViT-Clarity, an underwater image enhancement module, is introduced, which integrates vision transformers with a convolutional neural network for superior performance. For comparison, ClarityNet, a transformer-free variant of the architecture, is presented to highlight the transformer's impact. Given the limited availability of paired underwater image datasets (clear and degraded), BlueStyleGAN is proposed as a generative model to create synthetic underwater images from clear in-air images by simulating realistic attenuation effects. BlueStyleGAN is evaluated against existing state-of-the-art synthetic dataset generators in terms of training stability and realism. Vit-ClarityNet is rigorously tested on five datasets representing diverse underwater conditions and compared with recent state-of-the-art methods as well as ClarityNet. Evaluations include qualitative and quantitative metrics such as UCIQM, UCIQE, and the deep learning-based URanker. Additionally, the impact of enhanced images on object detection and SIFT feature matching is assessed, demonstrating the practical benefits of image enhancement for underwater computer vision tasks.
引用
收藏
页数:18
相关论文
共 52 条
[1]   Color Balance and Fusion for Underwater Image Enhancement [J].
Ancuti, Codruta O. ;
Ancuti, Cosmin ;
De Vleeschouwer, Christophe ;
Bekaert, Philippe .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (01) :379-393
[2]   Color-based underwater object recognition using water light attenuation [J].
Bazeille, Stephane ;
Quidu, Isabelle ;
Jaulin, Luc .
INTELLIGENT SERVICE ROBOTICS, 2012, 5 (02) :109-118
[3]  
Berman Dana, 2021, Zenodo, DOI 10.5281/ZENODO.5744037
[4]  
Bonin F., 2011, Journal of Maritime Research, V8, P65
[5]   Achieving domain generalization for underwater object detection by domain mixup and contrastive learning [J].
Chen, Yang ;
Song, Pinhao ;
Liu, Hong ;
Dai, Linhui ;
Zhang, Xiaochuan ;
Ding, Runwei ;
Li, Shengquan .
NEUROCOMPUTING, 2023, 528 :20-34
[6]   PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN With Dual-Discriminators [J].
Cong, Runmin ;
Yang, Wenyu ;
Zhang, Wei ;
Li, Chongyi ;
Guo, Chun-Le ;
Huang, Qingming ;
Kwong, Sam .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 :4472-4485
[7]   Underwater Depth Estimation for Spherical Images [J].
Cui, Jiadi ;
Jin, Lei ;
Kuang, Haofei ;
Xu, Qingwen ;
Schwertfeger, Soren .
JOURNAL OF ROBOTICS, 2021, 2021
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Dosovitskiy Alexey., 2021, PROC INT C LEARN REP, P2021
[10]  
Fabbri C, 2018, IEEE INT CONF ROBOT, P7159