Topological Visualization Method for Understanding the Landscape of Value Functions and Structure of the State Space in Reinforcement Learning

被引:1
|
作者
Nakamura, Yuki [1 ]
Shibuya, Takeshi [2 ]
机构
[1] Univ Tsukuba, Grad Sch Syst & Informat Engn, 1-1-1 Tennodai, Tsukuba, Ibaraki, Japan
[2] Univ Tsukuba, Fac Engn Informat & Syst, 1-1-1 Tennodai, Tsukuba, Ibaraki, Japan
来源
ICAART: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2 | 2020年
关键词
Reinforcement Learning; Topological Data Analysis; TDA Mapper; Visualization;
D O I
10.5220/0008913303700377
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is a learning framework applied in various fields in which agents autonomously acquire control rules. Using this method, the designer constructs a state space and reward function and sets various parameters to obtain ideal performance. The actual performance of the agent depends on the design. Accordingly, a poor design causes poor performance. In that case, the designer needs to examine the cause of the poor performance; to do so, it is important for the designer to understand the current agent control rules. In the case where the state space is less than or equal to two dimensions, visualizing the landscape of the value function and the structure of the state space is the most powerful method to understand these rules. However, in other cases, there is no method for such a visualization. In this paper, we propose a method to visualize the landscape of the value function and the structure of the state space even when the state space has a high number of dimensions. Concretely, we employ topological data analysis for the visualization. We confirm the effectiveness of the proposed method via several numerical experiments.
引用
收藏
页码:370 / 377
页数:8
相关论文
共 50 条
  • [1] Structure in the space of value functions
    Foster, D
    Dayan, P
    MACHINE LEARNING, 2002, 49 (2-3) : 325 - 346
  • [2] Structure in the Space of Value Functions
    David Foster
    Peter Dayan
    Machine Learning, 2002, 49 : 325 - 346
  • [3] Reinforcement Learning Method for Continuous State Space Based on Dynamic Neural Network
    Sun, Wei
    Wang, Xuesong
    Cheng, Yuhu
    2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 750 - 754
  • [4] A reinforcement learning accelerated by state space reduction
    Senda, K
    Mano, S
    Fujii, S
    SICE 2003 ANNUAL CONFERENCE, VOLS 1-3, 2003, : 1992 - 1997
  • [5] Adaptive state space partitioning for reinforcement learning
    Lee, ISK
    Lau, HYK
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2004, 17 (06) : 577 - 588
  • [6] Adaptive state space formation in reinforcement learning
    Samejima, K
    Omori, T
    ICONIP'98: THE FIFTH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING JOINTLY WITH JNNS'98: THE 1998 ANNUAL CONFERENCE OF THE JAPANESE NEURAL NETWORK SOCIETY - PROCEEDINGS, VOLS 1-3, 1998, : 251 - 255
  • [7] A state space filter for reinforcement learning in POMDPs - Application to a continuous state space -
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 3098 - +
  • [8] A Novel State Space Exploration Method for the Sparse-Reward Reinforcement Learning Environment
    Liu, Xi
    Ma, Long
    Chen, Zhen
    Zheng, Changgang
    Chen, Ren
    Liao, Yong
    Yang, Shufan
    ARTIFICIAL INTELLIGENCE XL, AI 2023, 2023, 14381 : 216 - 221
  • [9] Reduction of state space in reinforcement learning by sensor selection
    Kishima, Yasutaka
    Kurashige, Kentarou
    ARTIFICIAL LIFE AND ROBOTICS, 2013, 18 (1-2) : 7 - 14
  • [10] Constructivist Approach to State Space Adaptation in Reinforcement Learning
    Guerian, Maxime
    Cardozo, Nicolas
    Dusparic, Ivana
    2019 IEEE 13TH INTERNATIONAL CONFERENCE ON SELF-ADAPTIVE AND SELF-ORGANIZING SYSTEMS (SASO), 2019, : 52 - 61