A New Approach for Value Function Approximation Based on Automatic State Partition

被引:0
|
作者
Zeng, Jiaan [1 ]
Han, Yinghua [1 ]
机构
[1] S China Univ Technol, Sch Engn & Comp Sci, Guangzhou 510640, Guangdong, Peoples R China
来源
IMECS 2009: INTERNATIONAL MULTI-CONFERENCE OF ENGINEERS AND COMPUTER SCIENTISTS, VOLS I AND II | 2009年
关键词
reinforcement learning; fuzzy CMAC; automatic state partition;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Value function is usually used to deal with the reinforcement learning problems. In large or even continuous states, function approximation must be used to represent value function. Much of the current work carried out, however, has to design the structure of function approximation in advanced which cannot be adjusted during learning. In this paper, we propose a novel function approximation called Puzzy CMAC (FCMAC) with automatic state partition (ASP-FCMAC) to automate the structure design for FCMAC. Based on CMAC (also known as tile coding), ASP-FCMAC employs fuzzy membership function to avoid the setting of parameter in CMAC, and makes use of Bellman error to partition the state automatically so as to generate the structure of FCMAC. Empirical results in both mountain car and RoboCup Keepaway domains demonstrate that ASP-FCMAC can automatically generate the structure of FCMAC and agent using it can learn efficiently.
引用
收藏
页码:208 / 213
页数:6
相关论文
共 50 条
  • [21] Offline Reinforcement Learning: Fundamental Barriers for Value Function Approximation
    Foster, Dylan J.
    Krishnamurthy, Akshay
    Simchi-Levi, David
    Xu, Yunzong
    CONFERENCE ON LEARNING THEORY, VOL 178, 2022, 178
  • [22] Distributed Value Function Approximation for Collaborative Multiagent Reinforcement Learning
    Stankovic, Milos S.
    Beko, Marko
    Stankovic, Srdjan S.
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2021, 8 (03): : 1270 - 1280
  • [23] A new approach for supervised learning based influence value reinforcement learning
    Valdivia, Andre
    Herrera Quispe, Jose
    Barrios-Aranibar, Dennis
    2ND INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND SOFT COMPUTING (ICMLSC 2018), 2015, : 24 - 28
  • [24] ON CONVERGENCE RATE OF ADAPTIVE MULTISCALE VALUE FUNCTION APPROXIMATION FOR REINFORCEMENT LEARNING
    Li, Tao
    Zhu, Quanyan
    2019 IEEE 29TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2019,
  • [25] Convergence Rates of Online Critic Value Function Approximation in Native Spaces
    Niu, Shengyuan
    Bouland, Ali
    Wang, Haoran
    Fotiadis, Filippos
    Kurdila, Andrew
    L'Afflitto, Andrea
    Paruchuri, Sai Tej
    Vamvoudakis, Kyriakos G.
    IEEE CONTROL SYSTEMS LETTERS, 2024, 8 : 2145 - 2150
  • [26] Improving Gaussian Process Value Function Approximation in Policy Gradient Algorithms
    Jakab, Hunor
    Csato, Lehel
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2011, PT II, 2011, 6792 : 221 - +
  • [27] The Divergence of Reinforcement Learning Algorithms with Value-Iteration and Function Approximation
    Fairbank, Michael
    Alonso, Eduardo
    2012 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2012,
  • [28] A UNIFIED FRAMEWORK FOR LINEAR FUNCTION APPROXIMATION OF VALUE FUNCTIONS IN STOCHASTIC CONTROL
    Sanchez-Fernandez, Matilde
    Valcarcel, Sergio
    Zazo, Santiago
    2013 PROCEEDINGS OF THE 21ST EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2013,
  • [29] From free energy to expected energy: Improving energy-based value function approximation in reinforcement learning
    Elfwing, Stefan
    Uchibe, Eiji
    Doya, Kenji
    NEURAL NETWORKS, 2016, 84 : 17 - 27
  • [30] Multi-Agent Reinforcement Learning Approach Based on Reduced Value Function Approximations
    Abouheaf, Mohammed
    Gueaieb, Wail
    2017 IEEE 5TH INTERNATIONAL SYMPOSIUM ON ROBOTICS AND INTELLIGENT SENSORS (IRIS), 2017, : 111 - 116