Topological Visualization Method for Understanding the Landscape of Value Functions and Structure of the State Space in Reinforcement Learning

被引:1
|
作者
Nakamura, Yuki [1 ]
Shibuya, Takeshi [2 ]
机构
[1] Univ Tsukuba, Grad Sch Syst & Informat Engn, 1-1-1 Tennodai, Tsukuba, Ibaraki, Japan
[2] Univ Tsukuba, Fac Engn Informat & Syst, 1-1-1 Tennodai, Tsukuba, Ibaraki, Japan
来源
ICAART: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2 | 2020年
关键词
Reinforcement Learning; Topological Data Analysis; TDA Mapper; Visualization;
D O I
10.5220/0008913303700377
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is a learning framework applied in various fields in which agents autonomously acquire control rules. Using this method, the designer constructs a state space and reward function and sets various parameters to obtain ideal performance. The actual performance of the agent depends on the design. Accordingly, a poor design causes poor performance. In that case, the designer needs to examine the cause of the poor performance; to do so, it is important for the designer to understand the current agent control rules. In the case where the state space is less than or equal to two dimensions, visualizing the landscape of the value function and the structure of the state space is the most powerful method to understand these rules. However, in other cases, there is no method for such a visualization. In this paper, we propose a method to visualize the landscape of the value function and the structure of the state space even when the state space has a high number of dimensions. Concretely, we employ topological data analysis for the visualization. We confirm the effectiveness of the proposed method via several numerical experiments.
引用
收藏
页码:370 / 377
页数:8
相关论文
共 50 条
  • [21] State-Space Closure: Revisiting Endless Online Level Generation via Reinforcement Learning
    Wang, Ziqi
    Shu, Tianye
    Liu, Jialin
    IEEE TRANSACTIONS ON GAMES, 2024, 16 (02) : 489 - 492
  • [22] Anomaly detection using state-space models and reinforcement learning
    Khazaeli, Shervin
    Nguyen, Luong Ha
    Goulet, James A.
    STRUCTURAL CONTROL & HEALTH MONITORING, 2021, 28 (06)
  • [23] Multiagent reinforcement learning with the partly high-dimensional state space
    Department of Electrical and Computer Engineering, Nagoya Institute of Technology, Nagoya, 466-8555, Japan
    Syst Comput Jpn, 2006, 9 (22-31): : 22 - 31
  • [24] A Multi-agent Reinforcement Learning Method for Role Differentiation Using State Space Filters with Fluctuation Parameters
    Nagayoshi, Masato
    Elderton, Simon J. H.
    Tamaki, Hisashi
    JOURNAL OF ROBOTICS NETWORKING AND ARTIFICIAL LIFE, 2021, 8 (01): : 6 - 9
  • [25] Hybrid Reinforcement Learning and Uneven Generalization of Learning Space Method for Robot Obstacle Avoidance
    Li, Jianghao
    Bi, Weihong
    Li, Mingda
    PROCEEDINGS OF 2013 CHINESE INTELLIGENT AUTOMATION CONFERENCE: INTELLIGENT AUTOMATION & INTELLIGENT TECHNOLOGY AND SYSTEMS, 2013, 255 : 175 - 182
  • [26] Reinforcement Learning to Create Value and Policy Functions Using Minimax Tree Search in Hex
    Takada, Kei
    Iizuka, Hiroyuki
    Yamamoto, Masahito
    IEEE TRANSACTIONS ON GAMES, 2020, 12 (01) : 63 - 73
  • [27] A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning
    Sun, Wei-Fang
    Lee, Cheng-Kuang
    See, Simon
    Lee, Chun-Yi
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [28] Distributed Multiagent Reinforcement Learning Based on Graph-Induced Local Value Functions
    Jing, Gangshan
    Bai, He
    George, Jemin
    Chakrabortty, Aranya
    Sharma, Piyush K.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (10) : 6636 - 6651
  • [29] An Experience Replay Method Based on Tree Structure for Reinforcement Learning
    Jiang, Wei-Cheng
    Hwang, Kao-Shing
    Lin, Jin-Ling
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2021, 9 (02) : 972 - 982
  • [30] Metric State Space Reinforcement Learning for a Vision-Capable Mobile Robot
    Zhumatiy, Viktor
    Gomez, Faustino
    Hutter, Marcus
    Schmidhuber, Juergen
    INTELLIGENT AUTONOMOUS SYSTEMS 9, 2006, : 272 - +