A novel policy-graph approach with natural language and counterfactual abstractions for explaining reinforcement learning agents

被引:0
|
作者
Liu, Tongtong [1 ]
McCalmon, Joe [1 ]
Le, Thai [2 ]
Rahman, Md Asifur [1 ]
Lee, Dongwon [3 ]
Alqahtani, Sarra [1 ]
机构
[1] Wake Forest Univ, Comp Sci Dept, 1834 Wake Forest Rd, Winston Salem, NC 27109 USA
[2] Univ Mississippi, Comp & Informat Sci Dept, 1848 Univ, Oxford, MS 38677 USA
[3] Penn State Univ, Coll Informat Sci & Technol, University Pk, PA 16802 USA
关键词
Reinforcement learning; Explainable AI; XRL; Autonomous;
D O I
10.1007/s10458-023-09615-8
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As reinforcement learning (RL) continues to improve and be applied in situations along-side humans, the need to explain the learned behaviors of RL agents to end-users becomes more important. Strategies for explaining the reasoning behind an agent's policy, called policy-level explanations, can lead to important insights about both the task and the agent's behaviors. Following this line of research, in this work, we propose a novel approach, named as CAPS, that summarizes an agent's policy in the form of a directed graph with natural language descriptions. A decision tree based clustering method is utilized to abstract the state space of the task into fewer, condensed states which makes the policy graphs more digestible to end-users. We then use the user-defined predicates to enrich the abstract states with semantic meaning. To introduce counterfactual state explanations to the policy graph, we first identify the critical states in the graph then develop a novel coun-terfactual explanation method based on action perturbation in those critical states. We gen-erate explanation graphs using CAPS on 5 RL tasks, using both deterministic and stochas-tic policies. We also evaluate the effectiveness of CAPS on human participants who are not RL experts in two user studies. When provided with our explanation graph, end-users are able to accurately interpret policies of trained RL agents 80% of the time, compared to 10% when provided with the next best baseline and 68.2% of users demonstrated an increase in their confidence in understanding an agent's behavior after provided with the counterfac-tual explanations.
引用
收藏
页数:37
相关论文
共 12 条
  • [1] A novel policy-graph approach with natural language and counterfactual abstractions for explaining reinforcement learning agents
    Tongtong Liu
    Joe McCalmon
    Thai Le
    Md Asifur Rahman
    Dongwon Lee
    Sarra Alqahtani
    Autonomous Agents and Multi-Agent Systems, 2023, 37
  • [2] Explaining Reinforcement Learning Agents through Counterfactual Action Outcomes
    Amitai, Yotam
    Septon, Yael
    Amir, Ofra
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 9, 2024, : 10003 - 10011
  • [3] Explaining Black Box Reinforcement Learning Agents Through Counterfactual Policies
    Movin, Maria
    Dinis Junior, Guilherme
    Hollmen, Jaakko
    Papapetrou, Panagiotis
    ADVANCES IN INTELLIGENT DATA ANALYSIS XXI, IDA 2023, 2023, 13876 : 314 - 326
  • [4] Explaining the Behaviour of Reinforcement Learning Agents in a Multi-Agent Cooperative Environment Using Policy Graphs
    Vila, Marc
    Gnatyshak, Dmitry
    Tormos, Adrian
    Gimenez-Abalos, Victor
    Alvarez-Napagao, Sergio
    ELECTRONICS, 2024, 13 (03)
  • [5] GIS and Intelligent Agents for Multiobjective Natural Resource Allocation: A Reinforcement Learning Approach
    Bone, Christopher
    Dragicevic, Suzana
    TRANSACTIONS IN GIS, 2009, 13 (03) : 253 - 272
  • [6] Toward Collaborative Reinforcement Learning Agents that Communicate Through Text-Based Natural Language
    Eloff, Kevin M.
    Engelbrecht, Herman A.
    2021 SOUTHERN AFRICAN UNIVERSITIES POWER ENGINEERING CONFERENCE/ROBOTICS AND MECHATRONICS/PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA (SAUPEC/ROBMECH/PRASA), 2021,
  • [7] Modeling a Novel Approach for Emotion Recognition Using Learning and Natural Language Processing
    Lalitha, Lakshmi, V
    Anguraj, Dinesh Kumar
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (03)
  • [8] A Novel Graph Reinforcement Learning Approach for Stochastic Dynamic Economic Dispatch under High Penetration of Renewable Energy
    Li, Peng
    Huang, Wenqi
    Dai, Zhen
    Hou, Jiaxuan
    Cao, Shang
    Zhang, Jiayu
    Chen, Junbin
    2022 4TH ASIA ENERGY AND ELECTRICAL ENGINEERING SYMPOSIUM (AEEES 2022), 2022, : 498 - 503
  • [9] Diverse policy generation for the flexible job-shop scheduling problem via deep reinforcement learning with a novel graph representation
    Echeverria, Imanol
    Murua, Maialen
    Santana, Roberto
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 139
  • [10] A Novel Graph Reinforcement Learning-Based Approach for Dynamic Reconfiguration of Active Distribution Networks with Integrated Renewable Energy
    Zhan, Hua
    Jiang, Changxu
    Lin, Zhen
    ENERGIES, 2024, 17 (24)