A New Approach for Value Function Approximation Based on Automatic State Partition

被引:0
|
作者
Zeng, Jiaan [1 ]
Han, Yinghua [1 ]
机构
[1] S China Univ Technol, Sch Engn & Comp Sci, Guangzhou 510640, Guangdong, Peoples R China
来源
IMECS 2009: INTERNATIONAL MULTI-CONFERENCE OF ENGINEERS AND COMPUTER SCIENTISTS, VOLS I AND II | 2009年
关键词
reinforcement learning; fuzzy CMAC; automatic state partition;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Value function is usually used to deal with the reinforcement learning problems. In large or even continuous states, function approximation must be used to represent value function. Much of the current work carried out, however, has to design the structure of function approximation in advanced which cannot be adjusted during learning. In this paper, we propose a novel function approximation called Puzzy CMAC (FCMAC) with automatic state partition (ASP-FCMAC) to automate the structure design for FCMAC. Based on CMAC (also known as tile coding), ASP-FCMAC employs fuzzy membership function to avoid the setting of parameter in CMAC, and makes use of Bellman error to partition the state automatically so as to generate the structure of FCMAC. Empirical results in both mountain car and RoboCup Keepaway domains demonstrate that ASP-FCMAC can automatically generate the structure of FCMAC and agent using it can learn efficiently.
引用
收藏
页码:208 / 213
页数:6
相关论文
共 50 条
  • [1] A grey approximation approach to state value function in reinforcement learning
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Lee, Guar-Yuan
    2007 IEEE INTERNATIONAL CONFERENCE ON INTEGRATION TECHNOLOGY, PROCEEDINGS, 2007, : 379 - +
  • [2] Fuzzy CMAC with Automatic State Partition for Reinforcement Learning
    Min, Huaqing
    Zeng, Jiaan
    Luo, Ronghua
    WORLD SUMMIT ON GENETIC AND EVOLUTIONARY COMPUTATION (GEC 09), 2009, : 421 - 428
  • [3] Pseudorehearsal in Value Function Approximation
    Marochko, Vladimir
    Johard, Leonard
    Mazzara, Manuel
    AGENT AND MULTI-AGENT SYSTEMS: TECHNOLOGY AND APPLICATIONS, 2018, 74 : 178 - 189
  • [4] Continuous Control With Swarm Intelligence Based Value Function Approximation
    Wang, Bi
    Li, Xuelian
    Chen, Yang
    Wu, Jianqing
    Zeng, Bowen
    Chen, Junfu
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, 21 (01) : 976 - 988
  • [5] Low-rank State-action Value-function Approximation
    Rozada, Sergio
    Tenorio, Victor
    Marques, Antonio G.
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 1471 - 1475
  • [6] Least Absolute Policy Iteration-A Robust Approach to Value Function Approximation
    Sugiyama, Masashi
    Hachiya, Hirotaka
    Kashima, Hisashi
    Morimura, Tetsuro
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2010, E93D (09) : 2555 - 2565
  • [7] Geodesic Gaussian kernels for value function approximation
    Sugiyama, Masashi
    Hachiya, Hirotaka
    Towell, Christopher
    Vijayakumar, Sethu
    AUTONOMOUS ROBOTS, 2008, 25 (03) : 287 - 304
  • [8] Geodesic Gaussian kernels for value function approximation
    Masashi Sugiyama
    Hirotaka Hachiya
    Christopher Towell
    Sethu Vijayakumar
    Autonomous Robots, 2008, 25 : 287 - 304
  • [9] Differential TD Learning for Value Function Approximation
    Devraj, Adithya M.
    Meyn, Sean P.
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 6347 - 6354
  • [10] Dynamic Spectrum Anti-Jamming With Reinforcement Learning Based on Value Function Approximation
    Zhu, Xinyu
    Huang, Yang
    Wang, Shaoyu
    Wu, Qihui
    Ge, Xiaohu
    Liu, Yuan
    Gao, Zhen
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2023, 12 (02) : 386 - 390