Deep multi-modal data analysis and fusion for robust scene understanding in CAVs

被引:1
|
作者
Papandreou, Andreas [1 ]
Kloukiniotis, Andreas [1 ]
Lalos, Aris [2 ]
Moustakas, Konstantinos [1 ]
机构
[1] Univ Patras, Dept Elect & Comp Engn, Univ Campus, Rion 26504, Greece
[2] ISI Ind Syst Inst, Patras Sci Pk Bldg, Patras, Greece
关键词
autonomous vehicles; multi-modal scene analysis; adversarial attacks;
D O I
10.1109/MMSP53017.2021.9733604
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep learning (DL) tends to be the integral part of Autonomous Vehicles (AVs). Therefore the development of scene analysis modules that are robust to various vulnerabilities such as adversarial inputs or cyber-attacks is becoming an imperative need for the future AV perception systems. In this paper, we deal with this issue by exploring the recent progress in Artificial Intelligence (AI) and Machine Learning (ML) to provide holistic situational awareness and eliminate the effect of the previous attacks on the scene analysis modules. We propose novel multi-modal approaches against which achieve robustness to adversarial attacks, by appropriately modifying the analysis Neural networks and by utilizing late fusion methods. More specifically, we propose a holistic approach by adding new layers to a 2D segmentation DL model enhancing its robustness to adversarial noise. Then, a novel late fusion technique has been applied, by extracting direct features from the 3D space and project them into the 2D segmented space for identifying inconsistencies. Extensive evaluation studies using the KITTI odometry dataset provide promising performance results under various types of noise.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] InstaIndoor and multi-modal deep learning for indoor scene recognition
    Glavan, Andreea
    Talavera, Estefania
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (09): : 6861 - 6877
  • [22] InstaIndoor and multi-modal deep learning for indoor scene recognition
    Andreea Glavan
    Estefanía Talavera
    Neural Computing and Applications, 2022, 34 : 6861 - 6877
  • [23] Multi-Modal Deep Analysis for Multimedia
    Zhu, Wenwu
    Wang, Xin
    Li, Hongzhi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (10) : 3740 - 3764
  • [24] Multi-modal deep fusion for bridge condition assessment
    Momtaz M.
    Li T.
    Harris D.K.
    Lattanzi D.
    Journal of Infrastructure Intelligence and Resilience, 2023, 2 (04):
  • [25] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [26] Memory based fusion for multi-modal deep learning
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    INFORMATION FUSION, 2021, 67 : 136 - 146
  • [27] A hybrid multi-modal visual data cross fusion network for indoor and outdoor scene segmentation
    Hu, Sijie
    Bonardi, Fabien
    Bouchafa, Samia
    Sidibe, Desire
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2539 - 2545
  • [28] Co-inference for Multi-modal Scene Analysis
    Munoz, Daniel
    Bagnell, James Andrew
    Hebert, Martial
    COMPUTER VISION - ECCV 2012, PT VI, 2012, 7577 : 668 - 681
  • [29] Deep Object Tracking with Multi-modal Data
    Zhang, Xuezhi
    Yuan, Yuan
    Lu, Xiaoqiang
    2016 INTERNATIONAL CONFERENCE ON COMPUTER, INFORMATION AND TELECOMMUNICATION SYSTEMS (CITS), 2016, : 161 - 165
  • [30] Discriminative Multi-modal Feature Fusion for RGBD Indoor Scene Recognition
    Zhu, Hongyuan
    Weibel, Jean-Baptiste
    Lu, Shijian
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2969 - 2976