Explainable AI and Robustness-Based Test and Evaluation of Reinforcement Learning

被引:0
|
作者
Raz, Ali K. [1 ]
Mall, Kshitij [2 ]
Nolan, Sean Matthew [2 ]
Levin, Winston [2 ]
Mockus, Linas [2 ]
Ezra, Kris [3 ]
Mia, Ahmad [1 ]
Williams, Kyle
Parish, Julie [4 ]
机构
[1] George Mason Univ, Fairfax, VA 22030 USA
[2] Purdue Univ, W Lafayette, IN 47907 USA
[3] Crowd Strike Inc, W Lafayette, IN 47906 USA
[4] Sandia Natl Labs, Albuquerque, NM 94551 USA
关键词
Reinforcement learning; Training; Robustness; Decision making; Explainable AI; Additives; Sensitivity; high-speed aerospace systems; reinforcement learning (RL); robustness testing (RT); Shapley additive explanations (SHAP); DEEP; GUIDANCE;
D O I
10.1109/TAES.2024.3403078
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Reinforcement learning is a powerful and proven approach to generating near-optimal decision policies across domains, although characterizing performance boundaries, explaining decisions, and quantifying output uncertainties are major barriers to widespread adoption of reinforcement learning for real-time use. This is particularly true for high-risk and safety-critical aerospace systems where the cost of failure is high and performance envelopes for systems of interest may be small. To address these issues, this article presents a three-part test and evaluation framework for reinforcement learning, which is purpose-built from a systems engineering perspective on artificial intelligence. This framework employs explainable AI techniques-namely, Shapley additive explanations-to examine opaque decision-making, introduces robustness testing to characterize performance bounds and sensitivities, and incorporates output validation against accepted solutions. In this article, we consider an example problem of a high-speed aerospace vehicle emergency descent problem where a reinforcement learning agent is trained to control vehicle angle of attack (AoA). Shapley additive explanations expose the most significant features that impact the selection of AoA command while robustness testing characterizes the acceptable range of disturbances in flight parameters the trained vehicle can accommodate. Finally, the outputs from the reinforcement learning agent are compared with a baseline optimal trajectory as an acceptance criterion of RL solutions.
引用
收藏
页码:6110 / 6123
页数:14
相关论文
共 50 条
  • [1] Explainable AI-based Federated Deep Reinforcement Learning for Trusted Autonomous Driving
    Rjoub, Gaith
    Bentahar, Jamal
    Wahab, Omar Abdel
    2022 INTERNATIONAL WIRELESS COMMUNICATIONS AND MOBILE COMPUTING, IWCMC, 2022, : 318 - 323
  • [2] From Explainable AI to Explainable Simulation: Using Machine Learning and XAI to understand System Robustness
    Feldkamp, Niclas
    Strassburger, Steffen
    PROCEEDINGS OF THE 2023 ACM SIGSIM INTERNATIONAL CONFERENCE ON PRINCIPLES OF ADVANCED DISCRETE SIMULATION, ACMSIGSIM-PADS 2023, 2023, : 96 - 106
  • [3] Quantifying Attribution-based Explainable AI for Robustness Evaluations
    Leo Wilms
    Arndt von Twickel
    Matthias Neu
    Christian Berghoff
    Datenschutz und Datensicherheit - DuD, 2023, 47 (8) : 492 - 496
  • [4] Robustness-based evaluation of hydropower infrastructure design under climate change
    Taner, Mehmet Umit
    Ray, Patrick
    Brown, Casey
    CLIMATE RISK MANAGEMENT, 2017, 18 : 34 - 50
  • [5] Robustness-Based Evaluation of GHG Emissions and Energy Use at Neighborhood Level
    Moschetti, Roberta
    Homaei, Shabnam
    ENERGIES, 2024, 17 (23)
  • [6] Robustness-Based Design of Water Distribution Systems
    Jung, Donghwi
    Kang, Doosun
    Kim, Joong Hoon
    Lansey, Kevin
    JOURNAL OF WATER RESOURCES PLANNING AND MANAGEMENT, 2014, 140 (11)
  • [7] Post-Hoc Categorization Based on Explainable AI and Reinforcement Learning for Improved Intrusion Detection
    Larriva-Novo, Xavier
    Miguel, Luis Perez
    Villagra, Victor A.
    alvarez-Campana, Manuel
    Sanchez-Zas, Carmen
    Jover, Oscar
    APPLIED SCIENCES-BASEL, 2024, 14 (24):
  • [8] Explainable AI methods on a deep reinforcement learning agent for automatic docking
    Lover, Jakob
    Gjaerum, Vilde B.
    Lekkas, Anastasios M.
    IFAC PAPERSONLINE, 2021, 54 (16): : 146 - 152
  • [9] Memory-Based Explainable Reinforcement Learning
    Cruz, Francisco
    Dazeley, Richard
    Vamplew, Peter
    AI 2019: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, 11919 : 66 - 77
  • [10] Robustness-Based Condition Evaluation Framework for Through Tied-Arch Bridge
    Fan, Bing-hui
    Wang, Su-guo
    Chen, Bao-chun
    Chao, Peng-fei
    Sun, Qi
    JOURNAL OF PERFORMANCE OF CONSTRUCTED FACILITIES, 2023, 37 (02)