Explainable AI and Robustness-Based Test and Evaluation of Reinforcement Learning

被引:0
|
作者
Raz, Ali K. [1 ]
Mall, Kshitij [2 ]
Nolan, Sean Matthew [2 ]
Levin, Winston [2 ]
Mockus, Linas [2 ]
Ezra, Kris [3 ]
Mia, Ahmad [1 ]
Williams, Kyle
Parish, Julie [4 ]
机构
[1] George Mason Univ, Fairfax, VA 22030 USA
[2] Purdue Univ, W Lafayette, IN 47907 USA
[3] Crowd Strike Inc, W Lafayette, IN 47906 USA
[4] Sandia Natl Labs, Albuquerque, NM 94551 USA
关键词
Reinforcement learning; Training; Robustness; Decision making; Explainable AI; Additives; Sensitivity; high-speed aerospace systems; reinforcement learning (RL); robustness testing (RT); Shapley additive explanations (SHAP); DEEP; GUIDANCE;
D O I
10.1109/TAES.2024.3403078
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Reinforcement learning is a powerful and proven approach to generating near-optimal decision policies across domains, although characterizing performance boundaries, explaining decisions, and quantifying output uncertainties are major barriers to widespread adoption of reinforcement learning for real-time use. This is particularly true for high-risk and safety-critical aerospace systems where the cost of failure is high and performance envelopes for systems of interest may be small. To address these issues, this article presents a three-part test and evaluation framework for reinforcement learning, which is purpose-built from a systems engineering perspective on artificial intelligence. This framework employs explainable AI techniques-namely, Shapley additive explanations-to examine opaque decision-making, introduces robustness testing to characterize performance bounds and sensitivities, and incorporates output validation against accepted solutions. In this article, we consider an example problem of a high-speed aerospace vehicle emergency descent problem where a reinforcement learning agent is trained to control vehicle angle of attack (AoA). Shapley additive explanations expose the most significant features that impact the selection of AoA command while robustness testing characterizes the acceptable range of disturbances in flight parameters the trained vehicle can accommodate. Finally, the outputs from the reinforcement learning agent are compared with a baseline optimal trajectory as an acceptance criterion of RL solutions.
引用
收藏
页码:6110 / 6123
页数:14
相关论文
共 50 条
  • [21] A Robustness-Based Confidence Measure for Hybrid System Falsification
    Takisaka, Toru
    Zhang, Zhenya
    Arcaini, Paolo
    Hasuo, Ichiro
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (05) : 1718 - 1731
  • [22] A new approach to robustness-based optimization using uncertainty set constructed through machine learning
    Shahbab, R. M.
    Zaman, Kais
    STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION, 2024, 67 (07)
  • [23] A Robustness-based Approach to Determining Preventive and Emergency Control Zones
    Castellanos B, R.
    Messina, A. R.
    IEEE POWER AND ENERGY SOCIETY GENERAL MEETING 2010, 2010,
  • [24] A robustness-based approach to systems-oriented drug design
    Hiroaki Kitano
    Nature Reviews Drug Discovery, 2007, 6 : 202 - 210
  • [25] Design and Evaluation of Reinforcement Learning Based AI Agent: A Case Study in Gaming
    Jayashree, P.
    Ramakrishnan, K.
    PROCEEDINGS OF THE EIGHTH INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND PATTERN RECOGNITION (SOCPAR 2016), 2018, 614 : 330 - 339
  • [26] Robustness-based design optimization under data uncertainty
    Zaman, Kais
    McDonald, Mark
    Mahadevan, Sankaran
    Green, Lawrence
    STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION, 2011, 44 (02) : 183 - 197
  • [27] Robustness-Based Approach for Slack Time Optimization in Tram Timetables
    Zhou, Weixia
    Teng, Jing
    Chen, Enhui
    URBAN RAIL TRANSIT, 2025, 11 (01) : 90 - 107
  • [28] Sample-Based Rule Extraction for Explainable Reinforcement Learning
    Engelhardt, Raphael C.
    Lange, Moritz
    Wiskott, Laurenz
    Konen, Wolfgang
    MACHINE LEARNING, OPTIMIZATION, AND DATA SCIENCE, LOD 2022, PT I, 2023, 13810 : 330 - 345
  • [29] Container Caching Optimization based on Explainable Deep Reinforcement Learning
    Jayaram, Divyashree
    Jeelani, Saad
    Ishigaki, Genya
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 7127 - 7132
  • [30] Reinforcement Learning Based Path Exploration for Sequential Explainable Recommendation
    Li, Yicong
    Chen, Hongxu
    Li, Yile
    Li, Lin
    Yu, Philip S.
    Xu, Guandong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (11) : 11801 - 11814