A New Approach for Value Function Approximation Based on Automatic State Partition

被引:0
|
作者
Zeng, Jiaan [1 ]
Han, Yinghua [1 ]
机构
[1] S China Univ Technol, Sch Engn & Comp Sci, Guangzhou 510640, Guangdong, Peoples R China
来源
IMECS 2009: INTERNATIONAL MULTI-CONFERENCE OF ENGINEERS AND COMPUTER SCIENTISTS, VOLS I AND II | 2009年
关键词
reinforcement learning; fuzzy CMAC; automatic state partition;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Value function is usually used to deal with the reinforcement learning problems. In large or even continuous states, function approximation must be used to represent value function. Much of the current work carried out, however, has to design the structure of function approximation in advanced which cannot be adjusted during learning. In this paper, we propose a novel function approximation called Puzzy CMAC (FCMAC) with automatic state partition (ASP-FCMAC) to automate the structure design for FCMAC. Based on CMAC (also known as tile coding), ASP-FCMAC employs fuzzy membership function to avoid the setting of parameter in CMAC, and makes use of Bellman error to partition the state automatically so as to generate the structure of FCMAC. Empirical results in both mountain car and RoboCup Keepaway domains demonstrate that ASP-FCMAC can automatically generate the structure of FCMAC and agent using it can learn efficiently.
引用
收藏
页码:208 / 213
页数:6
相关论文
共 50 条
  • [31] Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
    Tamosiunaite, Minija
    Asfour, Tamim
    Woergoetter, Florentin
    BIOLOGICAL CYBERNETICS, 2009, 100 (03) : 249 - 260
  • [32] Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
    Minija Tamosiunaite
    Tamim Asfour
    Florentin Wörgötter
    Biological Cybernetics, 2009, 100 : 249 - 260
  • [33] Efficient exploration through active learning for value function approximation in reinforcement learning
    Akiyama, Takayuki
    Hachiya, Hirotaka
    Sugiyama, Masashi
    NEURAL NETWORKS, 2010, 23 (05) : 639 - 648
  • [34] Optimal feedback control of dynamical systems via value-function approximation
    Kunisch, Karl
    Walter, Daniel
    COMPTES RENDUS MECANIQUE, 2023, 351
  • [35] A review of stochastic algorithms with continuous value function approximation and some new approximate policy iteration algorithms for multidimensional continuous applications
    Warren B.POWELL
    JournalofControlTheoryandApplications, 2011, 9 (03) : 336 - 352
  • [36] A review of stochastic algorithms with continuous value function approximation and some new approximate policy iteration algorithms for multidimensional continuous applications
    Powell W.B.
    Ma J.
    Journal of Control Theory and Applications, 2011, 9 (3): : 336 - 352
  • [37] Tensor and Matrix Low-Rank Value-Function Approximation in Reinforcement Learning
    Rozada, Sergio
    Paternain, Santiago
    Marques, Antonio
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 1634 - 1649
  • [38] Restricted gradient-descent algorithm for value-function approximation in reinforcement learning
    Salles Barreto, Andre da Motta
    Anderson, Charles W.
    ARTIFICIAL INTELLIGENCE, 2008, 172 (4-5) : 454 - 482
  • [39] Using Reinforcement Learning to Control Traffic Signals in a Real-World Scenario: An Approach Based on Linear Function Approximation
    Alegre, Lucas N.
    Ziemke, Theresa
    Bazzan, Ana L. C.
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (07) : 9126 - 9135
  • [40] A fuzzy-based function approximation technique for reinforcement learning
    Wu, Cheng
    Song, Huichun
    Yan, Changsheng
    Wang, Yiming
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2017, 32 (06) : 3909 - 3920