On Robustness of Multi-Modal Fusion-Robotics Perspective

被引:21
作者
Bednarek, Michal [1 ]
Kicki, Piotr [1 ]
Walas, Krzysztof [1 ]
机构
[1] Poznan Univ Tech, Inst Robot & Machine Intelligence, PL-60965 Poznan, Poland
关键词
multi-modal fusion; machine learning; robotics; ARCHITECTURE;
D O I
10.3390/electronics9071152
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The efficient multi-modal fusion of data streams from different sensors is a crucial ability that a robotic perception system should exhibit to ensure robustness against disturbances. However, as the volume and dimensionality of sensory-feedback increase it might be difficult to manually design a multimodal-data fusion system that can handle heterogeneous data. Nowadays, multi-modal machine learning is an emerging field with research focused mainly on analyzing vision and audio information. Although, from the robotics perspective, haptic sensations experienced from interaction with an environment are essential to successfully execute useful tasks. In our work, we compared four learning-based multi-modal fusion methods on three publicly available datasets containing haptic signals, images, and robots' poses. During tests, we considered three tasks involving such data, namely grasp outcome classification, texture recognition, and-most challenging-multi-label classification of haptic adjectives based on haptic and visual data. Conducted experiments were focused not only on the verification of the performance of each method but mainly on their robustness against data degradation. We focused on this aspect of multi-modal fusion, as it was rarely considered in the research papers, and such degradation of sensory feedback might occur during robot interaction with its environment. Additionally, we verified the usefulness of data augmentation to increase the robustness of the aforementioned data fusion methods.
引用
收藏
页码:1 / 17
页数:17
相关论文
共 42 条
  • [1] [Anonymous], 2018, IEEE C COMP VIS PATT
  • [2] Multimodal Machine Learning: A Survey and Taxonomy
    Baltrusaitis, Tadas
    Ahuja, Chaitanya
    Morency, Louis-Philippe
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) : 423 - 443
  • [3] THE STRUCTURE AND FUNCTION OF PACINIAN CORPUSCLES - A REVIEW
    BELL, J
    BOLANOWSKI, S
    HOLMES, MH
    [J]. PROGRESS IN NEUROBIOLOGY, 1994, 42 (01) : 79 - 128
  • [4] Bijelic M, 2018, IEEE INT C INTELL TR, P2177, DOI 10.1109/ITSC.2018.8569911
  • [5] A Review of Data Fusion Techniques
    Castanedo, Federico
    [J]. SCIENTIFIC WORLD JOURNAL, 2013,
  • [6] Chebotar Y., 2016, ICRA 2016 WORKSH GRA
  • [7] EmbraceNet: A robust deep learning architecture for multimodal classification
    Choi, Jun-Ho
    Lee, Jong-Seok
    [J]. INFORMATION FUSION, 2019, 51 : 259 - 270
  • [8] Robotic learning of haptic adjectives through physical interaction
    Chu, Vivian
    McMahon, Ian
    Riano, Lorenzo
    McDonald, Craig G.
    He, Qin
    Perez-Tejada, Jorge Martinez
    Arrigo, Michael
    Darrell, Trevor
    Kuchenbecker, Katherine J.
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2015, 63 : 279 - 292
  • [9] Culbertson H., PENN HAPTIC TEXTURE
  • [10] Eitel A, 2015, IEEE INT C INT ROBOT, P681, DOI 10.1109/IROS.2015.7353446