Explainable reinforcement learning (XRL): a systematic literature review and taxonomy

被引:6
作者
Bekkemoen, Yanzhe [1 ]
机构
[1] Norwegian Univ Sci & Technol, Dept Comp Sci, NO-7491 Trondheim, Norway
关键词
Reinforcement learning; Explainable artificial intelligence; Interpretability; Explainability; Explanation; INTERPRETABLE POLICIES; DECISION-MAKING; VISUAL ANALYSIS; EXPLANATIONS; ATARI; GO; CLASSIFICATION; APPROXIMATION; FEATURES; MODELS;
D O I
10.1007/s10994-023-06479-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, reinforcement learning (RL) systems have shown impressive performance and remarkable achievements. Many achievements can be attributed to combining RL with deep learning. However, those systems lack explainability, which refers to our understanding of the system's decision-making process. In response to this challenge, the new explainable RL (XRL) field has emerged and grown rapidly to help us understand RL systems. This systematic literature review aims to give a unified view of the field by reviewing ten existing XRL literature reviews and 189 XRL studies from the past five years. Furthermore, we seek to organize these studies into a new taxonomy, discuss each area in detail, and draw connections between methods and stakeholder questions (e.g., "how can I get the agent to do _?"). Finally, we look at the research trends in XRL, recommend XRL methods, and present some exciting research directions for future research. We hope stakeholders, such as RL researchers and practitioners, will utilize this literature review as a comprehensive resource to overview existing state-of-the-art XRL methods. Additionally, we strive to help find research gaps and quickly identify methods that answer stakeholder questions.
引用
收藏
页码:355 / 441
页数:87
相关论文
共 345 条
[1]  
Abbeel P., 2004, P 21 INT C MACH LEAR, P1, DOI 10.1145/1015330.1015430
[2]  
Acharya A., 2020, NEURIPS WORKSH CHALL, DOI [10.48550/ARXIV.2011.09004, DOI 10.48550/ARXIV.2011.09004]
[3]  
Achiam Joshua, 2018, OpenAI
[4]  
Adebayo J, 2018, ADV NEUR IN, V31
[5]  
Adebayo Julius, 2022, 10 INT C LEARN REPR
[6]  
Agrawal Akash, 2022, 2022 IEEE Workshop on Design Automation for CPS and IoT (DESTION)., P64, DOI 10.1109/DESTION56136.2022.00017
[7]   Continuous Action Reinforcement Learning From a Mixture of Interpretable Experts [J].
Akrour, Riad ;
Tateo, Davide ;
Peters, Jan .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (10) :6795-6806
[8]   Reinforcement Learning Interpretation Methods: A Survey [J].
Alharin, Alnour ;
Doan, Thanh-Nam ;
Sartipi, Mina .
IEEE ACCESS, 2020, 8 :171058-171077
[9]  
Amir D, 2018, PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), P1168
[10]   Summarizing agent strategies [J].
Amir, Ofra ;
Doshi-Velez, Finale ;
Sarne, David .
AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2019, 33 (05) :628-644