Tell Me, What Do You See?-Interpretable Classification of Wiring Harness Branches with Deep Neural Networks

被引:11
作者
Kicki, Piotr [1 ]
Bednarek, Michal [1 ]
Lembicz, Pawel [2 ]
Mierzwiak, Grzegorz [2 ]
Szymko, Amadeusz [1 ]
Kraft, Marek [1 ]
Walas, Krzysztof [1 ]
机构
[1] Poznan Univ Tech, Inst Robot & Machine Intelligence, Piotrowo 3A, PL-60965 Poznan, Poland
[2] Volkswagen Poznan Sp Zoo, Ul Warszawska 349, PL-61060 Poznan, Poland
基金
欧盟地平线“2020”;
关键词
machine vision; deformable linear objects; neural networks; robot learning; computer vision for manufacturing; WIRE;
D O I
10.3390/s21134327
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
In the context of the robotisation of industrial operations related to manipulating deformable linear objects, there is a need for sophisticated machine vision systems, which could classify the wiring harness branches and provide information on where to put them in the assembly process. However, industrial applications require the interpretability of the machine learning system predictions, as the user wants to know the underlying reason for the decision made by the system. We propose several different neural network architectures that are tested on our novel dataset to address this issue. We conducted various experiments to assess the influence of modality, data fusion type, and the impact of data augmentation and pretraining. The outcome of the network is evaluated in terms of the performance and is also equipped with saliency maps, which allow the user to gain in-depth insight into the classifier's operation, including a way of explaining the responses of the deep neural network and making system predictions interpretable by humans.
引用
收藏
页数:15
相关论文
共 34 条
  • [1] [Anonymous], 2015, TENSORFLOW LARGE SCA
  • [2] On Robustness of Multi-Modal Fusion-Robotics Perspective
    Bednarek, Michal
    Kicki, Piotr
    Walas, Krzysztof
    [J]. ELECTRONICS, 2020, 9 (07) : 1 - 17
  • [3] The WIRES experiment: tools and strategies for robotized switchgear cabling
    Busi, M.
    Cirillo, A.
    De Gregorio, D.
    Indovini, M.
    De Maria, G.
    Melchiorri, C.
    Natale, C.
    Palli, G.
    Pirozzi, S.
    [J]. 27TH INTERNATIONAL CONFERENCE ON FLEXIBLE AUTOMATION AND INTELLIGENT MANUFACTURING, FAIM2017, 2017, 11 : 355 - 363
  • [4] Cirillo A, 2017, IEEE ASME INT C ADV, P490, DOI 10.1109/AIM.2017.8014065
  • [5] The Cityscapes Dataset for Semantic Urban Scene Understanding
    Cordts, Marius
    Omran, Mohamed
    Ramos, Sebastian
    Rehfeld, Timo
    Enzweiler, Markus
    Benenson, Rodrigo
    Franke, Uwe
    Roth, Stefan
    Schiele, Bernt
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 3213 - 3223
  • [6] Integration of Robotic Vision and Tactile Sensing for Wire-Terminal Insertion Tasks
    De Gregorio, Daniele
    Zanella, Riccardo
    Palli, Gianluca
    Pirozzi, Salvatore
    Melchiorri, Claudio
    [J]. IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2019, 16 (02) : 585 - 598
  • [7] Eitel A, 2015, IEEE INT C INT ROBOT, P681, DOI 10.1109/IROS.2015.7353446
  • [8] Gumingshi, 2011, 2011 INTERNATIONAL CONFERENCE ON ELECTRONICS, COMMUNICATIONS AND CONTROL (ICECC), P2397, DOI 10.1109/ICECC.2011.6066493
  • [9] Deep learning for visual understanding: A review
    Guo, Yanming
    Liu, Yu
    Oerlemans, Ard
    Lao, Songyang
    Wu, Song
    Lew, Michael S.
    [J]. NEUROCOMPUTING, 2016, 187 : 27 - 48
  • [10] Learning Rich Features from RGB-D Images for Object Detection and Segmentation
    Gupta, Saurabh
    Girshick, Ross
    Arbelaez, Pablo
    Malik, Jitendra
    [J]. COMPUTER VISION - ECCV 2014, PT VII, 2014, 8695 : 345 - 360