LiDAR-assisted image restoration for extreme low-light conditions

被引:0
作者
Wang, Zhen [1 ]
Wu, Yaozu [2 ]
Li, Dongyuan [2 ]
Li, Guang [3 ]
Zhu, Peide [4 ]
Zhang, Ziqing [5 ]
Jiang, Renhe [2 ]
机构
[1] Tokyo Inst Technol, Dept Informat & Commun Engn, Tokyo, Japan
[2] Univ Tokyo, Ctr Spatial Informat Sci, Tokyo, Japan
[3] Hokkaido Univ, Educ & Res Ctr Math & Data Sci, Sapporo, Japan
[4] Fujian Normal Univ, Sch Math & Stat, Fujian, Peoples R China
[5] Shenzhen Ink Blue Technol, Shenzhen, Peoples R China
关键词
Low-light image enhancement; Multimodal; RGBD; LiDAR camera; ENHANCEMENT; NETWORK;
D O I
10.1016/j.knosys.2025.113382
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Research into enhancing images captured under low-light conditions has been a focus for several years. However, most existing image restoration techniques primarily address RGB-only images, overlooking the potential benefits of integrating additional modalities. With advancements in handheld technology, capturing images alongside depth data has become easily accessible through devices like smartphones. This presents a significant opportunity to explore the role of depth data in improving image quality in low-light environments. In this study, we introduce a novel dataset named LED (Low-light Image Enhanced with Depth Map), consisting of 1365 entries. Each entry includes a low-light image, its corresponding normal-light counterpart, and an associated depth map. To the best of our knowledge, LED is the first low-light image enhancement dataset that incorporates depth information. Building on this dataset, we propose model named DERNet with three innovative technologies to integrate depth data into the low-light image enhancement process. First, we utilize depth maps to extract precise edge details, improving object boundary clarity in low-light images. Second, we introduce a multi-scale multimodal fusion module, which combines features from both the image and depth map to enhance image quality. Third, we develop a multi-task refinement network that leverages an estimated illumination map to further refine the enhancement process. Experimental results demonstrate that our method achieves over a 0.52 improvement in PSNR compared to unimodal approaches and exceeds multimodal methods by nearly 0.3 with much less total parameters, highlighting the effectiveness of our approach. Moreover, our comprehensive ablation studies show that our method can be flexibly combined with previous approaches to enhance their accuracy, and our model demonstrates greater efficiency compared to previous methods.
引用
收藏
页数:13
相关论文
共 74 条
[1]  
[Anonymous], 1995, The handbook of brain theory and neural networks
[2]   3D Semantic Parsing of Large-Scale Indoor Spaces [J].
Armeni, Iro ;
Sener, Ozan ;
Zamir, Amir R. ;
Jiang, Helen ;
Brilakis, Ioannis ;
Fischer, Martin ;
Savarese, Silvio .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1534-1543
[3]  
Bychkovsky V, 2011, PROC CVPR IEEE, P97
[4]   nuScenes: A multimodal dataset for autonomous driving [J].
Caesar, Holger ;
Bankiti, Varun ;
Lang, Alex H. ;
Vora, Sourabh ;
Liong, Venice Erin ;
Xu, Qiang ;
Krishnan, Anush ;
Pan, Yu ;
Baldan, Giancarlo ;
Beijbom, Oscar .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11618-11628
[5]  
Cai YH, 2023, Arxiv, DOI arXiv:2303.06705
[6]   RGB-D datasets using microsoft kinect or similar sensors: a survey [J].
Cai, Ziyun ;
Han, Jungong ;
Liu, Li ;
Shao, Ling .
MULTIMEDIA TOOLS AND APPLICATIONS, 2017, 76 (03) :4313-4355
[7]   Seeing Motion in the Dark [J].
Chen, Chen ;
Chen, Qifeng ;
Do, Minh N. ;
Koltun, Vladlen .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3184-3193
[8]   Learning to See in the Dark [J].
Chen, Chen ;
Chen, Qifeng ;
Xu, Jia ;
Koltun, Vladlen .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :3291-3300
[9]   Gated Context Aggregation Network for Image Dehazing and Deraining [J].
Chen, Dongdong ;
He, Mingming ;
Fan, Qingnan ;
Liao, Jing ;
Zhang, Liheng ;
Hou, Dongdong ;
Yuan, Lu ;
Hua, Gang .
2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, :1375-1383
[10]  
Chi L., 2020, P 34 INT C NEUR INF, V33, P4479, DOI DOI 10.5555/3495724.3496100