Fuzzy Q-learning in continuous state and action space

被引:4
作者
Xu M.-L. [1 ,2 ]
Xu W.-B. [2 ]
机构
[1] Department of Electronic Information Engineering, Wuxi City College of Vocational Technology
[2] School of Information Technology, Jiangnan University
来源
Journal of China Universities of Posts and Telecommunications | 2010年 / 17卷 / 04期
基金
中国国家自然科学基金;
关键词
adaptation; continuous; FIS; Q-learning;
D O I
10.1016/S1005-8885(09)60495-7
中图分类号
学科分类号
摘要
An adaptive fuzzy Q-learning (AFQL) based on fuzzy inference systems (FIS) is proposed. The FIS realized by a normalized radial basis function (NRBF) neural network is used to approach Q-value function, whose input is composed of state and action. The rules of FIS are created incrementally according to the novelty of each element of the pair of state-action. Moreover the premise part and consequent part of the FIS are updated using extended Kalman filter (EKF). The action that impacts on environment is the one with maximum output of FIS in the current state and generated through optimization method. Simulation results in the wall-following task of mobile robots and the inverted pendulum balancing problem demonstrate that the superiority and applicability of the proposed AFQL method. © 2010 The Journal of China Universities of Posts and Telecommunications.
引用
收藏
页码:100 / 109
页数:9
相关论文
共 50 条
[31]   Mutual Q-learning [J].
Reid, Cameron ;
Mukhopadhyay, Snehasis .
2020 3RD INTERNATIONAL CONFERENCE ON CONTROL AND ROBOTS (ICCR 2020), 2020, :128-133
[32]   Periodic Q-Learning [J].
Lee, Donghwan ;
He, Niao .
LEARNING FOR DYNAMICS AND CONTROL, VOL 120, 2020, 120 :582-598
[33]   An improved fuzzy ARTMAP and Q-learning agent model for pattern classification [J].
Pourpanah, Farhad ;
Wang, Ran ;
Lim, Chee Peng ;
Wang, Xizhao ;
Seera, Manjeevan ;
Tan, Choo Jun .
NEUROCOMPUTING, 2019, 359 :139-152
[34]   Reactive fuzzy controller design by Q-learning for mobile robot navigation [J].
张文志 ;
吕恬生 .
JournalofHarbinInstituteofTechnology, 2005, (03) :319-324
[35]   Fuzzy Q-Learning Flow Control for High-Speed Networks [J].
Li, Xin ;
Zhao, Xin ;
Jing, Yuanwei ;
Zhang, Nannan .
2008 CHINESE CONTROL AND DECISION CONFERENCE, VOLS 1-11, 2008, :383-+
[36]   Intelligent Traffic Signal Synchronization Using Fuzzy Logic and Q-Learning [J].
Iyer, Vignesh ;
Jadhav, Rashmi ;
Mavchi, Unnati ;
Abraham, Jibi .
2016 INTERNATIONAL CONFERENCE ON COMPUTING, ANALYTICS AND SECURITY TRENDS (CAST), 2016, :156-161
[37]   Genetic Based Fuzzy Q-Learning Energy Management for Smart Grid [J].
Li Xin ;
Zang Chuanzhi ;
Zeng Peng ;
Yu Haibin .
PROCEEDINGS OF THE 31ST CHINESE CONTROL CONFERENCE, 2012, :6924-6927
[38]   Virtual markets:: Q-learning sellers with simple state representation [J].
Akchurina, Natalia ;
Buening, Hans Kleine .
AUTONOMOUS INTELLIGENT SYSTEMS: AGENTS AND DATA MINING, PROCEEDINGS, 2007, 4476 :192-+
[39]   Hamilton-Jacobi-Bellman Equations for Q-Learning in Continuous Time [J].
Kim, Jeongho ;
Yang, Insoon .
LEARNING FOR DYNAMICS AND CONTROL, VOL 120, 2020, 120 :739-748
[40]   Q-Learning Lagrange Policies for Multi-Action Restless Bandits [J].
Killian, Jackson A. ;
Biswas, Arpita ;
Shah, Sanket ;
Tambe, Milind .
KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, :871-881