Learning explainable task-relevant state representation for model-free deep reinforcement learning

被引:0
|
作者
Zhao, Tingting [1 ,2 ]
Li, Guixi [1 ]
Zhao, Tuo [1 ]
Chen, Yarui [1 ]
Xie, Ning [4 ]
Niu, Gang [2 ]
Sugiyama, Masashi [2 ,3 ]
机构
[1] Tianjin Univ Sci & Technol, Coll Artificial Intelligence, Tianjin, Peoples R China
[2] RIKEN Ctr Adv Intelligence Project AIP, Tokyo, Japan
[3] Univ Tokyo, Grad Sch Frontier Sci, Tokyo, Japan
[4] Univ Elect Sci & Technol China, Ctr Future Media, Chengdu, Peoples R China
关键词
Deep reinforcement learning; Model-free; State representation learning; Explainability; Auto-encoder;
D O I
10.1016/j.neunet.2024.106741
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State representations considerably accelerate learning speed and improve data efficiency for deep reinforcement learning (DRL), especially for visual tasks. Task-relevant state representations could focus on features relevant to the task, filter out irrelevant elements, and thus further improve performance. However, task-relevant representations are typically obtained through model-based DRL methods, which involves the challenging task of learning a transition function. Moreover, inaccuracies in the learned transition function can potentially lead to performance degradation and negatively impact the learning of the policy. In this paper, to address the above issue, we propose a novel method of explainable task-relevant state representation (ETrSR) for model-free DRL that is direct, robust, and without any requirement of learning of a transition model. More specifically, the proposed ETrSR first disentangles the features from the states based on the beta variational autoencoder (beta-VAE). Then, a reward prediction model is employed to bootstrap these features to be relevant to the task, and the explainable states can be obtained by decoding the task-related features. Finally, we validate our proposed method on the CarRacing environment and various tasks in the DeepMind control suite (DMC), which demonstrates the explainability for better understanding of the decision-making process and the outstanding performance of the proposed method even in environments with strong distractions.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Learning to Discover Task-Relevant Features for Interpretable Reinforcement Learning
    Zhang, Qiyuan
    Ma, Xiaoteng
    Yang, Yiqin
    Li, Chenghao
    Yang, Jun
    Liu, Yu
    Liang, Bin
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (04) : 6601 - 6607
  • [2] Learning task-relevant representations via rewards and real actions for reinforcement learning
    Yuan, Linghui
    Lu, Xiaowei
    Liu, Yunlong
    KNOWLEDGE-BASED SYSTEMS, 2024, 294
  • [3] Efficient Reinforcement Learning in Block MDPs: A Model-free Representation Learning Approach
    Zhang, Xuezhou
    Song, Yuda
    Uehara, Masatoshi
    Wang, Mengdi
    Agarwal, Alekh
    Sun, Wen
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [4] Model-Free Deep Inverse Reinforcement Learning by Logistic Regression
    Eiji Uchibe
    Neural Processing Letters, 2018, 47 : 891 - 905
  • [5] Model-free Deep Reinforcement Learning for Urban Autonomous Driving
    Chen, Jianyu
    Yuan, Bodi
    Tomizuka, Masayoshi
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 2765 - 2771
  • [7] Task-Relevant Chunking in Sequence Learning
    Perlman, Amotz
    Pothos, Emmanuel M.
    Edwards, Darren J.
    Tzelgov, Joseph
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-HUMAN PERCEPTION AND PERFORMANCE, 2010, 36 (03) : 649 - 661
  • [8] Learning Representations in Model-Free Hierarchical Reinforcement Learning
    Rafati, Jacob
    Noelle, David C.
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 10009 - 10010
  • [9] Unsupervised Model-Free Representation Learning
    Ryabko, Daniil
    ALGORITHMIC LEARNING THEORY (ALT 2013), 2013, 8139 : 354 - 366
  • [10] STATE REPRESENTATION LEARNING FOR EFFECTIVE DEEP REINFORCEMENT LEARNING
    Zhao, Jian
    Zhou, Wengang
    Zhao, Tianyu
    Zhou, Yun
    Li, Houqiang
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,