A temporal-difference learning method using gaussian state representation for continuous state space problems

被引:0
|
作者
机构
[1] Graduate School of Engineering, Osaka City University
来源
| 1600年 / Japanese Society for Artificial Intelligence卷 / 29期
关键词
Continuous state spaces; Gaussian state representation; Reinforcement learning; TD learning;
D O I
10.1527/tjsai.29.157
中图分类号
学科分类号
摘要
In this paper, we tackle the problem of reinforcement learning (RL) in a continuous state space. An appropriate discretization of the space can make many learning tasks tractable. A method using Gaussian state representation and the Rational Policy Making algorithm (RPM) has been proposed for this problem. This method discretizes the space by constructing a chain of states which represents a path to the goal of the agent exploiting past experiences of reaching it. This method exploits successful experiences strongly. Therefore, it can find a rational solution quickly in an environment with few noises. In a noisy environment, it makes many unnecessary and distractive states and does the task poorly. For learning in such an environment, we have introduced the concept of the value of a state to the above method and developed a new method. This method uses a temporal-difference (TD) learning algorithm for learning the values of states. The value of a state is used to determine the size of the state. Thus, our developed method can trim and eliminate unnecessary and distractive states quickly and learn the task well even in a noisy environment. We show the effectiveness of our method by computer simulations of a path finding task and a cart-pole swing-up task. © The Japanese Society for Artificial Intelligence 2014.
引用
收藏
页码:157 / 167
页数:10
相关论文
共 50 条
  • [21] Cooperative strategy learning in multi-agent environment with continuous state space
    Tao, Jun-Yuan
    Li, De-Sheng
    PROCEEDINGS OF 2006 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2006, : 2107 - +
  • [22] Reinforcement Learning for POMDP Environments Using State Representation with Reservoir Computing
    Yamashita, Kodai
    Hamagami, Tomoki
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2022, 26 (04) : 562 - 569
  • [23] A double layered state space construction method for reinfrocement learning agents
    Handa, H
    SICE 2004 ANNUAL CONFERENCE, VOLS 1-3, 2004, : 2698 - 2703
  • [24] A reinforcement learning algorithm for continuous state spaces using multiple Fuzzy-ART networks
    Tateyama, Takeshi
    Kawata, Seiichi
    Shimomura, Yoshiki
    2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 88 - +
  • [25] Anomaly detection using state-space models and reinforcement learning
    Khazaeli, Shervin
    Nguyen, Luong Ha
    Goulet, James A.
    STRUCTURAL CONTROL & HEALTH MONITORING, 2021, 28 (06)
  • [26] A novel state space representation for the solution of 2D-HP protein folding problem using reinforcement learning methods
    Dogan, Berat
    Olmez, Tamer
    APPLIED SOFT COMPUTING, 2015, 26 : 213 - 223
  • [27] Online Gaussian Process State-space Model: Learning and Planning for Partially Observable Dynamical Systems
    Park, Soon-Seo
    Park, Young-Jin
    Min, Youngjae
    Choi, Han-Lim
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2022, 20 (02) : 601 - 617
  • [28] Online Gaussian Process State-space Model: Learning and Planning for Partially Observable Dynamical Systems
    Soon-Seo Park
    Young-Jin Park
    Youngjae Min
    Han-Lim Choi
    International Journal of Control, Automation and Systems, 2022, 20 : 601 - 617
  • [29] A Multi-agent Reinforcement Learning Method for Role Differentiation Using State Space Filters with Fluctuation Parameters
    Nagayoshi, Masato
    Elderton, Simon J. H.
    Tamaki, Hisashi
    JOURNAL OF ROBOTICS NETWORKING AND ARTIFICIAL LIFE, 2021, 8 (01): : 6 - 9
  • [30] A Neuro-fuzzy Learning System for Adaptive Swarm Behaviors Dealing with Continuous State Space
    Kuremoto, Takashi
    Obayashi, Masanao
    Kobayashi, Kunikazu
    Adachi, Hirotaka
    Yoneda, Kentaro
    ADVANCED INTELLIGENT COMPUTING THEORIES AND APPLICATIONS, PROCEEDINGS: WITH ASPECTS OF ARTIFICIAL INTELLIGENCE, 2008, 5227 : 675 - +