On Robustness of Multi-Modal Fusion-Robotics Perspective

被引:21
|
作者
Bednarek, Michal [1 ]
Kicki, Piotr [1 ]
Walas, Krzysztof [1 ]
机构
[1] Poznan Univ Tech, Inst Robot & Machine Intelligence, PL-60965 Poznan, Poland
关键词
multi-modal fusion; machine learning; robotics; ARCHITECTURE;
D O I
10.3390/electronics9071152
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The efficient multi-modal fusion of data streams from different sensors is a crucial ability that a robotic perception system should exhibit to ensure robustness against disturbances. However, as the volume and dimensionality of sensory-feedback increase it might be difficult to manually design a multimodal-data fusion system that can handle heterogeneous data. Nowadays, multi-modal machine learning is an emerging field with research focused mainly on analyzing vision and audio information. Although, from the robotics perspective, haptic sensations experienced from interaction with an environment are essential to successfully execute useful tasks. In our work, we compared four learning-based multi-modal fusion methods on three publicly available datasets containing haptic signals, images, and robots' poses. During tests, we considered three tasks involving such data, namely grasp outcome classification, texture recognition, and-most challenging-multi-label classification of haptic adjectives based on haptic and visual data. Conducted experiments were focused not only on the verification of the performance of each method but mainly on their robustness against data degradation. We focused on this aspect of multi-modal fusion, as it was rarely considered in the research papers, and such degradation of sensory feedback might occur during robot interaction with its environment. Additionally, we verified the usefulness of data augmentation to increase the robustness of the aforementioned data fusion methods.
引用
收藏
页码:1 / 17
页数:17
相关论文
共 50 条
  • [1] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [2] Multi-Modal Interaction for Robotics Mules
    Taylor, Glenn
    Quist, Michael
    Lanting, Matthew
    Dunham, Cory
    Muench, Paul
    UNMANNED SYSTEMS TECHNOLOGY XIX, 2017, 10195
  • [3] Is Multi-Modal Necessarily Better? Robustness Evaluation of Multi-Modal Fake News Detection
    Chen, Jinyin
    Jia, Chengyu
    Zheng, Haibin
    Chen, Ruoxi
    Fu, Chenbo
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (06): : 3144 - 3158
  • [4] Robust Dynamic Multi-Modal Data Fusion: A Model Uncertainty Perspective
    Liu, Bin
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 2107 - 2111
  • [5] On the Adversarial Robustness of Multi-Modal Foundation Models
    Schlarmann, Christian
    Hein, Matthias
    arXiv, 2023,
  • [6] On the Adversarial Robustness of Multi-Modal Foundation Models
    Schlarmann, Christian
    Hein, Matthias
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 3679 - 3687
  • [7] Soft multi-modal data fusion
    Coppock, S
    Mazack, L
    PROCEEDINGS OF THE 12TH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1 AND 2, 2003, : 636 - 641
  • [8] Multi-modal fusion for video understanding
    Hoogs, A
    Mundy, J
    Cross, G
    30TH APPLIED IMAGERY PATTERN RECOGNITION WORKSHOP, PROCEEDINGS: ANALYSIS AND UNDERSTANDING OF TIME VARYING IMAGERY, 2001, : 103 - 108
  • [9] Multi-modal data fusion: A description
    Coppock, S
    Mazlack, LJ
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 2, PROCEEDINGS, 2004, 3214 : 1136 - 1142
  • [10] MEM: Multi-Modal Elevation Mapping for Robotics and Learning
    Erni, Gian
    Frey, Jonas
    Miki, Takahiro
    Mattamala, Matias
    Hutter, Marco
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 11011 - 11018