Towards trustworthy multi-modal motion prediction: Holistic evaluation and interpretability of outputs

被引:4
|
作者
Limeros, Sandra Carrasco [1 ,2 ,3 ]
Majchrowska, Sylwia [2 ,3 ]
Johnander, Joakim [3 ,4 ]
Petersson, Christoffer [3 ,5 ]
Sotelo, Miguel Angel [1 ]
Llorca, David Fernandez [1 ,6 ]
机构
[1] Univ Alcala, Polytech Sch, Comp Engn Dept, Madrid, Spain
[2] AI Sweden, Gothenburg, Sweden
[3] Zenseact AB, Gothenburg, Sweden
[4] Linkoping Univ, Dept Elect Engn, Linkoping, Sweden
[5] Chalmers Univ Technol, Gothenburg, Sweden
[6] European Commiss, Joint Res Ctr, Seville, Spain
关键词
autonomous vehicles; evaluation; interpretability; multi-modal motion prediction; robustness; trustworthy AI;
D O I
10.1049/cit2.12244
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Predicting the motion of other road agents enables autonomous vehicles to perform safe and efficient path planning. This task is very complex, as the behaviour of road agents depends on many factors and the number of possible future trajectories can be considerable (multi-modal). Most prior approaches proposed to address multi-modal motion prediction are based on complex machine learning systems that have limited interpretability. Moreover, the metrics used in current benchmarks do not evaluate all aspects of the problem, such as the diversity and admissibility of the output. The authors aim to advance towards the design of trustworthy motion prediction systems, based on some of the requirements for the design of Trustworthy Artificial Intelligence. The focus is on evaluation criteria, robustness, and interpretability of outputs. First, the evaluation metrics are comprehensively analysed, the main gaps of current benchmarks are identified, and a new holistic evaluation framework is proposed. Then, a method for the assessment of spatial and temporal robustness is introduced by simulating noise in the perception system. To enhance the interpretability of the outputs and generate more balanced results in the proposed evaluation framework, an intent prediction layer that can be attached to multi-modal motion prediction models is proposed. The effectiveness of this approach is assessed through a survey that explores different elements in the visualisation of the multi-modal trajectories and intentions. The proposed approach and findings make a significant contribution to the development of trustworthy motion prediction systems for autonomous vehicles, advancing the field towards greater safety and reliability.
引用
收藏
页码:557 / 572
页数:16
相关论文
共 19 条
  • [1] Probabilistic Traffic Motion Labeling for Multi-Modal Vehicle Route Prediction
    Flores Fernandez, Alberto
    Wurst, Jonas
    Sanchez Morales, Eduardo
    Botsch, Michael
    Facchi, Christian
    Garcia Higuera, Andres
    SENSORS, 2022, 22 (12)
  • [2] Multi-Modal Interaction-Aware Motion Prediction at Unsignalized Intersections
    Trentin, Vinicius
    Artunedo, Antonio
    Godoy, Jorge
    Villagra, Jorge
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (05): : 3349 - 3365
  • [3] Is Multi-Modal Necessarily Better? Robustness Evaluation of Multi-Modal Fake News Detection
    Chen, Jinyin
    Jia, Chengyu
    Zheng, Haibin
    Chen, Ruoxi
    Fu, Chenbo
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (06): : 3144 - 3158
  • [4] LiCaNet: Further Enhancement of Joint Perception and Motion Prediction Based on Multi-Modal Fusion
    Khalil, Yasser H.
    Mouftah, Hussein T.
    IEEE OPEN JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 3 : 222 - 235
  • [5] Multi-modal Motion Prediction using Temporal Ensembling with Learning-based Aggregation
    Hong, Kai-Yin
    Wang, Chieh-Chih
    Lin, Wen-Chieh
    2024 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2024), 2024, : 9691 - 9697
  • [6] Towards a Multi-Modal Methodology for User-Centred Evaluation of Adaptive Systems
    Nussbaumer, Alexander
    Steiner, Christina M.
    Conlan, Owen
    ADJUNCT PUBLICATION OF THE 27TH CONFERENCE ON USER MODELING, ADAPTATION AND PERSONALIZATION (ACM UMAP '19 ADJUNCT), 2019, : 219 - 220
  • [7] Pesti-DGI-Net: A multi-modal deep learning architecture based on dual interpretability for pesticide-likeness prediction
    Yang, Ruoqi
    Yan, Yaochao
    Wei, Zhiheng
    Wang, Fan
    Yang, Guangfu
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2024, 217
  • [8] Inherent Fusion: Towards Scalable Multi-Modal Similarity Search
    Budikova, Petra
    Batko, Michal
    Novak, David
    Zezula, Pavel
    JOURNAL OF DATABASE MANAGEMENT, 2016, 27 (04) : 1 - 23
  • [9] Usability evaluation of multi-modal biometric verification systems
    Toledano, Doroteo T.
    Fernandez Pozo, Ruben
    Hernandez Trapote, Alvaro
    Hernandez Gomez, Luis
    INTERACTING WITH COMPUTERS, 2006, 18 (05) : 1101 - 1122
  • [10] Multi-modal vehicle trajectory prediction based on mutual information
    Fei, Cong
    He, Xiangkun
    Ji, Xuewu
    IET INTELLIGENT TRANSPORT SYSTEMS, 2020, 14 (03) : 148 - 153