Neural Q-Learning Based on Residual Gradient for Nonlinear Control Systems

被引:0
|
作者
Si, Yanna [1 ]
Pu, Jiexin [1 ]
Zang, Shaofei [1 ]
机构
[1] Henan Univ Sci & Technol, Sch Informat Engn, Luoyang, Peoples R China
来源
ICCAIS 2019: THE 8TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES | 2019年
关键词
Q-learning; feedforward neural network; value function approximation; residual gradient method; nonlinear control systems;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To solve the control problem of nonlinear system under continuous state space, this paper puts forward a neural Q-learning algorithm based on residual gradient method. Firstly, the multi-layer feedforward neural network is utilized to approximate the Q-value function, overcoming the "dimensional disaster" in the classical reinforcement learning. Then based on the residual gradient method, a mini-batch gradient descent is implemented by the experience replay to update the neural network parameters, which can effectively reduce the iterations number and increase the learning speed. Moreover, the momentum optimization method is introduced to ensure the stability of the training process further and improve the convergence. In order to balance exploration and utilization better, epsilon-decreasing strategy replaces epsilon-greedy for action selection. The simulation results of CartPole control task show the correctness and effectiveness of the proposed algorithm.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Elevator group control algorithm based on residual gradient and Q-learning
    Zong, ZL
    Wang, XG
    Tang, Z
    Zeng, GZ
    SICE 2004 ANNUAL CONFERENCE, VOLS 1-3, 2004, : 329 - 331
  • [2] Q-Learning Based Parameter Tuning for Model-free Adaptive Control of Nonlinear Systems
    Xu, Liuyong
    Hao, Shoulin
    Liu, Tao
    Zhu, Yong
    Wang, Haixia
    Zhang, Jiyan
    2024 14TH ASIAN CONTROL CONFERENCE, ASCC 2024, 2024, : 2078 - 2083
  • [3] Neural Q-Learning Based Mobile Robot Navigation
    Yun, Soh Chin
    Parasuraman, S.
    Ganapathy, V.
    Joe, Halim Kusuma
    MATERIALS SCIENCE AND INFORMATION TECHNOLOGY, PTS 1-8, 2012, 433-440 : 721 - +
  • [4] Hyperparameter optimization of neural networks based on Q-learning
    Xin Qi
    Bing Xu
    Signal, Image and Video Processing, 2023, 17 : 1669 - 1676
  • [5] Hyperparameter optimization of neural networks based on Q-learning
    Qi, Xin
    Xu, Bing
    SIGNAL IMAGE AND VIDEO PROCESSING, 2023, 17 (04) : 1669 - 1676
  • [6] Event-Triggered Data-Driven Control of Nonlinear Systems via Q-Learning
    Shen, Mouquan
    Wang, Xianming
    Zhu, Song
    Huang, Tingwen
    Wang, Qing-Guo
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2025, 55 (02): : 1069 - 1077
  • [7] Combining Q-learning and Deterministic Policy Gradient for Learning-based MPC
    Seel, Katrine
    Gros, Ebastien
    Gravdahl, Jan Tommy
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 610 - 617
  • [8] Q-learning based fault estimation and fault tolerant iterative learning control for MIMO systems
    Wang, Rui
    Zhuang, Zhihe
    Tao, Hongfeng
    Paszke, Wojciech
    Stojanovic, Vladimir
    ISA TRANSACTIONS, 2023, 142 : 123 - 135
  • [9] Neural Q-learning for solving PDEs
    Cohen, Samuel N.
    Jiang, Deqing
    Sirignano, Justin
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [10] Mobile robot navigation using neural Q-learning
    Yang, GS
    Chen, EK
    An, CW
    PROCEEDINGS OF THE 2004 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2004, : 48 - 52