Explainability in deep reinforcement learning

被引:176
作者
Heuillet, Alexandre [1 ]
Couthouis, Fabien [2 ]
Diaz-Rodriguez, Natalia [3 ]
机构
[1] Bordeaux INP, ENSEIRB MATMECA, 1 Ave Docteur Albert Schweitzer, F-33400 Talence, France
[2] Bordeaux INP, ENSC, 109 Ave Roul, F-33400 Talence, France
[3] Inst Polytech Paris, Inria Flowers Team, ENSTA Paris, 828 Blvd Marechaux, F-91762 Palaiseau, France
关键词
Reinforcement Learning; Explainable artificial intelligence; Machine Learning; Deep Learning; Responsible artificial intelligence; Representation learning;
D O I
10.1016/j.knosys.2020.106685
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A large set of the explainable Artificial Intelligence (XAI) literature is emerging on feature relevance techniques to explain a deep neural network (DNN) output or explaining models that ingest image source data. However, assessing how XAI techniques can help understand models beyond classification tasks, e.g. for reinforcement learning (RL), has not been extensively studied. We review recent works in the direction to attain Explainable Reinforcement Learning (XRL), a relatively new subfield of Explainable Artificial Intelligence, intended to be used in general public applications, with diverse audiences, requiring ethical, responsible and trustable algorithms. In critical situations where it is essential to justify and explain the agent's behaviour, better explainability and interpretability of RL models could help gain scientific insight on the inner workings of what is still considered a black box. We evaluate mainly studies directly linking explainability to RL, and split these into two categories according to the way the explanations are generated: transparent algorithms and post-hoc explainability. We also review the most prominent XAI works from the lenses of how they could potentially enlighten the further deployment of the latest advances in RL, in the demanding present and future of everyday problems. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 116 条
  • [1] Abbeel P., 2004, P 21 INT C MACH LEAR, P1, DOI [10.1145/1015330.1015430, DOI 10.1145/1015330.1015430]
  • [2] Achille A., 2017, ARXIV171103321
  • [3] Achille A, 2018, ADV NEUR IN, V31
  • [4] Adebayo J, 2018, ADV NEUR IN, V31
  • [5] Symbolic-Based Recognition of Contact States for Learning Assembly Skills
    Al-Yacoub, Ali
    Zhao, Yuchen
    Lohse, Niels
    Goh, Mey
    Kinnell, Peter
    Ferreira, Pedro
    Hubbard, Ella-Mae
    [J]. FRONTIERS IN ROBOTICS AND AI, 2019, 6
  • [6] Alvernaz Samuel, 2017, 2017 IEEE Conference on Computational Intelligence and Games (CIG), P1, DOI 10.1109/CIG.2017.8080408
  • [7] Andrychowicz M., 2017, ARXIV170701495
  • [8] Learning dexterous in-hand manipulation
    Andrychowicz, Marcin
    Baker, Bowen
    Chociej, Maciek
    Jozefowicz, Rafal
    McGrew, Bob
    Pachocki, Jakub
    Petron, Arthur
    Plappert, Matthias
    Powell, Glenn
    Ray, Alex
    Schneider, Jonas
    Sidor, Szymon
    Tobin, Josh
    Welinder, Peter
    Weng, Lilian
    Zaremba, Wojciech
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2020, 39 (01) : 3 - 20
  • [9] [Anonymous], 2019, ARXIV190512941
  • [10] [Anonymous], 2017, ARXIV171100867