A New Approach for Value Function Approximation Based on Automatic State Partition

被引:0
|
作者
Zeng, Jiaan [1 ]
Han, Yinghua [1 ]
机构
[1] S China Univ Technol, Sch Engn & Comp Sci, Guangzhou 510640, Guangdong, Peoples R China
来源
IMECS 2009: INTERNATIONAL MULTI-CONFERENCE OF ENGINEERS AND COMPUTER SCIENTISTS, VOLS I AND II | 2009年
关键词
reinforcement learning; fuzzy CMAC; automatic state partition;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Value function is usually used to deal with the reinforcement learning problems. In large or even continuous states, function approximation must be used to represent value function. Much of the current work carried out, however, has to design the structure of function approximation in advanced which cannot be adjusted during learning. In this paper, we propose a novel function approximation called Puzzy CMAC (FCMAC) with automatic state partition (ASP-FCMAC) to automate the structure design for FCMAC. Based on CMAC (also known as tile coding), ASP-FCMAC employs fuzzy membership function to avoid the setting of parameter in CMAC, and makes use of Bellman error to partition the state automatically so as to generate the structure of FCMAC. Empirical results in both mountain car and RoboCup Keepaway domains demonstrate that ASP-FCMAC can automatically generate the structure of FCMAC and agent using it can learn efficiently.
引用
收藏
页码:208 / 213
页数:6
相关论文
共 50 条
  • [41] Online Reinforcement Learning Control of Nonlinear Dynamic Systems: A State-action Value Function Based Solution
    Asl, Hamed Jabbari
    Uchibe, Eiji
    NEUROCOMPUTING, 2023, 544
  • [42] Local and soft feature selection for value function approximation in batch reinforcement learning for robot navigation
    Fathinezhad, Fatemeh
    Adibi, Peyman
    Shoushtarian, Bijan
    Chanussot, Jocelyn
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (08) : 10720 - 10745
  • [43] Integrating Symmetry of Environment by Designing Special Basis functions for Value Function Approximation in Reinforcement Learning
    Wang, Guo-fang
    Fang, Zhou
    Li, Bo
    Li, Ping
    2016 14TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2016,
  • [44] Local and soft feature selection for value function approximation in batch reinforcement learning for robot navigation
    Fatemeh Fathinezhad
    Peyman Adibi
    Bijan Shoushtarian
    Jocelyn Chanussot
    The Journal of Supercomputing, 2024, 80 : 10720 - 10745
  • [45] AUTOMATIC STATE SPACE AGGREGATION USING A DENSITY BASED TECHNIQUE
    Loscalzo, Steven
    Wright, Robert
    ICAART 2011: PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 1, 2011, : 249 - 256
  • [46] Gaussian Based Non-linear Function Approximation for Reinforcement Learning
    Haider A.
    Hawe G.
    Wang H.
    Scotney B.
    SN Computer Science, 2021, 2 (3)
  • [47] A Learning-based Adaptive Signal Control System with Function Approximation
    Jin, Junchen
    Ma, Xiaoliang
    IFAC PAPERSONLINE, 2016, 49 (03): : 5 - 10
  • [48] Rough Sets-based Prototype Optimization in Kanerva-based Function Approximation
    Wu, Cheng
    Li, Wei
    Meleis, Waleed
    2015 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY (WI-IAT), VOL 2, 2015, : 283 - 291
  • [49] Value-approximation-based online policy for vehicle routing problem with stochastic demand
    Zhang X.-N.
    Zhang J.-X.
    Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2022, 39 (02): : 241 - 254
  • [50] A Reinforcement Learning Approach to Understanding Procrastination: Does Inaccurate Value Approximation Cause Irrational Postponing of a Task?
    Feng, Zheyu
    Nagase, Asako Mitsuto
    Morita, Kenji
    FRONTIERS IN NEUROSCIENCE, 2021, 15