Neural Q-Learning Based on Residual Gradient for Nonlinear Control Systems

被引:0
|
作者
Si, Yanna [1 ]
Pu, Jiexin [1 ]
Zang, Shaofei [1 ]
机构
[1] Henan Univ Sci & Technol, Sch Informat Engn, Luoyang, Peoples R China
来源
ICCAIS 2019: THE 8TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES | 2019年
关键词
Q-learning; feedforward neural network; value function approximation; residual gradient method; nonlinear control systems;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To solve the control problem of nonlinear system under continuous state space, this paper puts forward a neural Q-learning algorithm based on residual gradient method. Firstly, the multi-layer feedforward neural network is utilized to approximate the Q-value function, overcoming the "dimensional disaster" in the classical reinforcement learning. Then based on the residual gradient method, a mini-batch gradient descent is implemented by the experience replay to update the neural network parameters, which can effectively reduce the iterations number and increase the learning speed. Moreover, the momentum optimization method is introduced to ensure the stability of the training process further and improve the convergence. In order to balance exploration and utilization better, epsilon-decreasing strategy replaces epsilon-greedy for action selection. The simulation results of CartPole control task show the correctness and effectiveness of the proposed algorithm.
引用
收藏
页数:5
相关论文
共 50 条
  • [41] Reinforcement Q-Learning for PDF Tracking Control of Stochastic Systems with Unknown Dynamics
    Yang, Weiqing
    Zhou, Yuyang
    Zhang, Yong
    Ren, Yan
    MATHEMATICS, 2024, 12 (16)
  • [42] A Combined Policy Gradient and Q-learning Method for Data-driven Optimal Control Problems
    Lin, Mingduo
    Liu, Derong
    Zhao, Bo
    Dai, Qionghai
    Dong, Yi
    2019 9TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND TECHNOLOGY (ICIST2019), 2019, : 6 - 10
  • [43] Dynamic Q-Learning for Intersection Traffic Flow Control Based on Agents
    Vista, Felipe P.
    Zhou, Xuan
    Ryu, Ji Hyoung
    Chong, Kil To
    ADVANCED SCIENCE LETTERS, 2014, 20 (01) : 120 - 123
  • [44] Feature Extraction in Q-Learning using Neural Networks
    Zhu, Henghui
    Paschalidis, Ioannis Ch.
    Hasselmo, Michael E.
    2017 IEEE 56TH ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2017,
  • [45] Reinforcement Learning-based control using Q-learning and gravitational search algorithm with experimental validation on a nonlinear servo system
    Zamfirache, Iuliu Alexandru
    Precup, Radu-Emil
    Roman, Raul-Cristian
    Petriu, Emil M.
    INFORMATION SCIENCES, 2022, 583 : 99 - 120
  • [46] Cooperative strategy based on adaptive Q-learning for robot soccer systems
    Hwang, KS
    Tan, SW
    Chen, CC
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2004, 12 (04) : 569 - 576
  • [47] Q-Learning Algorithm and CMAC Approximation Based Robust Optimal Control for Renewable Energy Management Systems
    Vy Huynh Tuyet
    Luy Nguyen Tan
    CONTROL ENGINEERING AND APPLIED INFORMATICS, 2022, 24 (01): : 15 - 25
  • [48] Adaptive Learning-Rate on Integrated Stochastic Gradient Decreasing Q-Learning
    Jin H.-D.
    Liu Q.
    Chen D.-H.
    Jisuanji Xuebao/Chinese Journal of Computers, 2019, 42 (10): : 2203 - 2215
  • [49] Deep Policy-Gradient Based Path Planning and Reinforcement Cooperative Q-Learning Behavior of Multi-Vehicle Systems
    Afifi, Ahmed M.
    Alhosainy, Omar H.
    Elias, Catherine M.
    Shehata, Omar M.
    Morgan, Elsayed I.
    2019 IEEE INTERNATIONAL CONFERENCE OF VEHICULAR ELECTRONICS AND SAFETY (ICVES 19), 2019,
  • [50] Adaptive Optimal Control via Q-Learning for Ito Fuzzy Stochastic Nonlinear Continuous-Time Systems With Stackelberg Game
    Ming, Zhongyang
    Zhang, Huaguang
    Yan, Ying
    Yang, Liu
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2024, 32 (04) : 2029 - 2038