Dehazing & Reasoning YOLO: Prior knowledge-guided network for object detection in foggy weather

被引:3
作者
Zhong, Fujin [1 ,2 ,3 ]
Shen, Wenxin [1 ,2 ,3 ]
Yu, Hong [1 ,2 ,3 ]
Wang, Guoyin [1 ,3 ]
Hu, Jun [1 ,2 ,3 ]
机构
[1] Chongqing Univ Telecommun & Posts, Chongqing Key Lab Computat Intelligence, Chongqing, Peoples R China
[2] Chongqing Univ Telecommun & Posts, Sch Comp Sci & Technol, Chongqing, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Key Lab Big Data Intelligent Comp, Chongqing, Peoples R China
基金
中国国家自然科学基金;
关键词
Object detection; Foggy weather; Prior knowledge; End-to-end network;
D O I
10.1016/j.patcog.2024.110756
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Fast and accurate object detection in foggy weather is crucial for visual tasks such as autonomous driving and video surveillance. Existing methods typically preprocess images with enhancement techniques before the object detector, so that the real-time performance of object detection decreases to some extent. Meanwhile, many popular object detection models rely solely on visual features for localization and classification. When fog is present, visual features would be so adversely impacted that the detection accuracy sharply decreases. Therefore, we propose an end-to-end prior knowledge-guided network called DR-YOLO for object detection in foggy weather. DR-YOLO integrates the atmospheric scattering model and the co-occurrence relation graph as prior knowledge into the entire training process of the detector. Firstly, Restoration Subnet Module (RSM) is designed to employ the atmospheric scattering model to guide the learning direction of the detector for dehazing features. Specifically, it is only adopted during the training process and does not increase the time cost of detection process. Secondly, for guiding the detector to pay more attention to potential cooccurring objects in the same scene, we introduce Relation Reasoning Attention Module (RRAM) that utilizes the co-occurrence relation graph to supplement deficient visual features in foggy weather. In addition, DRYOLO employs Adaptive Feature Fusion Module (AFFM) to effectively merge the key features from the backbone and neck for the needs of RRAM and RSM. Finally, we conduct experiments on clear, synthetic and real-world foggy datasets to demonstrate the effectiveness of DR-YOLO. The source code is available at https://github.com/wenxinss/DR-YOLO.
引用
收藏
页数:12
相关论文
共 46 条
[1]  
[Anonymous], 2014, arXiv
[2]   Non-Local Image Dehazing [J].
Berman, Dana ;
Treibitz, Tali ;
Avidan, Shai .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1674-1682
[3]   DehazeNet: An End-to-End System for Single Image Haze Removal [J].
Cai, Bolun ;
Xu, Xiangmin ;
Jia, Kui ;
Qing, Chunmei ;
Tao, Dacheng .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (11) :5187-5198
[4]   Grad-CAM plus plus : Generalized Gradient-based Visual Explanations for Deep Convolutional Networks [J].
Chattopadhay, Aditya ;
Sarkar, Anirban ;
Howlader, Prantik ;
Balasubramanian, Vineeth N. .
2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, :839-847
[5]  
Chen Z., 2023, ARXIV
[6]   PSD: Principled Synthetic-to-Real Dehazing Guided by Physical Priors [J].
Chen, Zeyuan ;
Wang, Yangchao ;
Yang, Yang ;
Liu, Dong .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :7176-7185
[7]   Multi-Label Image Recognition with Graph Convolutional Networks [J].
Chen, Zhao-Min ;
Wei, Xiu-Shen ;
Wang, Peng ;
Guo, Yanwen .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :5172-5181
[8]   Multi-Scale Boosted Dehazing Network with Dense Feature Fusion [J].
Dong, Hang ;
Pan, Jinshan ;
Xiang, Lei ;
Hu, Zhe ;
Zhang, Xinyi ;
Wang, Fei ;
Yang, Ming-Hsuan .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :2154-2164
[9]   The Pascal Visual Object Classes (VOC) Challenge [J].
Everingham, Mark ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) :303-338
[10]  
Fang Y, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1661