Deep multi-modal data analysis and fusion for robust scene understanding in CAVs

被引:1
|
作者
Papandreou, Andreas [1 ]
Kloukiniotis, Andreas [1 ]
Lalos, Aris [2 ]
Moustakas, Konstantinos [1 ]
机构
[1] Univ Patras, Dept Elect & Comp Engn, Univ Campus, Rion 26504, Greece
[2] ISI Ind Syst Inst, Patras Sci Pk Bldg, Patras, Greece
关键词
autonomous vehicles; multi-modal scene analysis; adversarial attacks;
D O I
10.1109/MMSP53017.2021.9733604
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep learning (DL) tends to be the integral part of Autonomous Vehicles (AVs). Therefore the development of scene analysis modules that are robust to various vulnerabilities such as adversarial inputs or cyber-attacks is becoming an imperative need for the future AV perception systems. In this paper, we deal with this issue by exploring the recent progress in Artificial Intelligence (AI) and Machine Learning (ML) to provide holistic situational awareness and eliminate the effect of the previous attacks on the scene analysis modules. We propose novel multi-modal approaches against which achieve robustness to adversarial attacks, by appropriately modifying the analysis Neural networks and by utilizing late fusion methods. More specifically, we propose a holistic approach by adding new layers to a 2D segmentation DL model enhancing its robustness to adversarial noise. Then, a novel late fusion technique has been applied, by extracting direct features from the 3D space and project them into the 2D segmented space for identifying inconsistencies. Extensive evaluation studies using the KITTI odometry dataset provide promising performance results under various types of noise.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Part-Whole Relational Fusion Towards Multi-Modal Scene Understanding
    Liu, Yi
    Li, Chengxin
    Xu, Shoukun
    Han, Jungong
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025,
  • [2] Multi-modal fusion for video understanding
    Hoogs, A
    Mundy, J
    Cross, G
    30TH APPLIED IMAGERY PATTERN RECOGNITION WORKSHOP, PROCEEDINGS: ANALYSIS AND UNDERSTANDING OF TIME VARYING IMAGERY, 2001, : 103 - 108
  • [3] Robust multi-modal fusion architecture for medical data with knowledge distillation
    Wang, Muyu
    Fan, Shiyu
    Li, Yichen
    Gao, Binyu
    Xie, Zhongrang
    Chen, Hui
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2025, 260
  • [4] Robust Dynamic Multi-Modal Data Fusion: A Model Uncertainty Perspective
    Liu, Bin
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 2107 - 2111
  • [5] Analysis of Deep Fusion Strategies for Multi-modal Gesture Recognition
    Roitberg, Alina
    Pollert, Tim
    Haurilet, Monica
    Martin, Manuel
    Stiefelhagen, Rainer
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 198 - 206
  • [6] Soft multi-modal data fusion
    Coppock, S
    Mazack, L
    PROCEEDINGS OF THE 12TH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1 AND 2, 2003, : 636 - 641
  • [7] Multi-modal data fusion: A description
    Coppock, S
    Mazlack, LJ
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 2, PROCEEDINGS, 2004, 3214 : 1136 - 1142
  • [8] Accelerating Deep Neural Networks for Efficient Scene Understanding in Multi-Modal Automotive Applications
    Nousias, Stavros
    Pikoulis, Erion-Vasilis
    Mavrokefalidis, Christos
    Lalos, Aris S.
    IEEE ACCESS, 2023, 11 : 28208 - 28221
  • [9] Robust Deep Multi-modal Learning Based on Gated Information Fusion Network
    Kim, Jaekyum
    Koh, Junho
    Kim, Yecheol
    Choi, Jaehyung
    Hwang, Youngbae
    Choi, Jun Won
    COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 90 - 106
  • [10] Survey on Deep Multi-modal Data Analytics: Collaboration, Rivalry, and Fusion
    Wang, Yang
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2021, 17 (01)