Radar Fusion Monocular Depth Estimation Based on Dual Attention

被引:1
作者
Long, JianYu [1 ]
Huang, JinGui [1 ]
Wang, ShengChun [1 ]
机构
[1] Hunan Normal Univ, Changsha 410006, Peoples R China
来源
ARTIFICIAL INTELLIGENCE AND SECURITY, ICAIS 2022, PT I | 2022年 / 13338卷
关键词
Monocular depth estimation; Radar; Attention; nuScenes;
D O I
10.1007/978-3-031-06794-5_14
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this article, we explore the integration of multimodal data into monocular depth estimation. Monocular depth estimation is performed by fusing RGB data with sparse radar data. Since the existing fusion method does not take into account the correlation between the two types of data in the channel and in space, it lacks the representation of the global information relationship on the channel and in space. Therefore, we propose a feature fusion module (DAF) based on the dual attention mechanism. The dual attention fusion module improves the global information representation capability of the model by modeling the dynamic and non-linear relationship of the two kinds of data in the channel and space, adaptively recalibrates the response to each feature, and maximizes the use of radar data. At the same time, DAF can reduce noise interference in radar data by weighting features, avoiding the loss of secondary details caused by filtering operations, and alleviating the problem of excessive noise in radar data. Finally, due to the influence of the complex weather environment and the model itself, it is difficult for the model to obtain an effective feature representation in the complex weather environment. Therefore, we introduced a batch loss function to enable the model to focus on feature extraction in a complex environment, so as to obtain a more accurate representation of feature information. It reduces model errors and speeds up the convergence of the model. The experiment was conducted on the recently released nuScenes dataset, which provides data records of the entire sensor suite of autonomous vehicles. Experiments prove that our method is superior to other fusion methods.
引用
收藏
页码:166 / 179
页数:14
相关论文
共 31 条
[1]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[2]  
Caesar H, 2020, PROC CVPR IEEE, P11618, DOI 10.1109/CVPR42600.2020.01164
[3]  
Dosovitskiy A, 2021, Arxiv, DOI [arXiv:2010.11929, DOI 10.48550/ARXIV.2010.11929]
[4]  
Eigen D, 2014, ADV NEUR IN, V27
[5]   Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture [J].
Eigen, David ;
Fergus, Rob .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2650-2658
[6]   Deep Ordinal Regression Network for Monocular Depth Estimation [J].
Fu, Huan ;
Gong, Mingming ;
Wang, Chaohui ;
Batmanghelich, Kayhan ;
Tao, Dacheng .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :2002-2011
[7]   Unsupervised Monocular Depth Estimation with Left-Right Consistency [J].
Godard, Clement ;
Mac Aodha, Oisin ;
Brostow, Gabriel J. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6602-6611
[8]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[9]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/TPAMI.2019.2913372, 10.1109/CVPR.2018.00745]
[10]   Sparse and Dense Data with CNNs: Depth Completion and Semantic Segmentation [J].
Jaritz, Maximilian ;
de Charette, Raoul ;
Wirbel, Emilie ;
Perrotton, Xavier ;
Nashashibi, Fawzi .
2018 INTERNATIONAL CONFERENCE ON 3D VISION (3DV), 2018, :52-60