Safe online optimization of motor speed synchronization control with incremental Q-learning

被引:0
|
作者
Huang, Jianfeng [1 ]
Lu, Guoqiang [1 ]
Yao, Xudong [1 ]
机构
[1] Shantou Univ, Coll Engn, Shantou 515063, Peoples R China
关键词
Online controller tuning; Safe reinforcement learning; Q; -learning; Motor speed synchronization; PARTICLE SWARM OPTIMIZATION; ENERGY MANAGEMENT; REINFORCEMENT; DESIGN; STRATEGY;
D O I
10.1016/j.eswa.2024.124622
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning (RL) is promising for online controller optimization. However, its practical application has been hindered by safety issues. This paper proposes an algorithm named Incremental Q-learning (IQ) and applies it to the online optimization of motor speed synchronization control. IQ ensures safe learning by adopting so-called incremental action variables which represent incremental change rather than absolute magnitude, and dividing the one-round learning process in the classic Q-learning (in this paper referred to as Absolute Qlearning, AQ) into multiple consecutive ones with the Q table getting reset at the beginning of each round. Since the permitted interval of change is restricted to be very small, the agent can learn its way safely, steadily, and robustly towards the optimal policy. Simulation results show that IQ is advantageous to AQ in optimality, safety, and adaptability. IQ converges to better final performances with significantly smaller performance variance along the whole learning process, smaller torque trajectory deviation between consecutive episodes and adapts to unknown disturbances faster. It is of great potential for online controller optimization/tuning in practical engineering projects. Source code and demos are provided.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Parallel Online Temporal Difference Learning for Motor Control
    Caarls, Wouter
    Schuitema, Erik
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (07) : 1457 - 1468
  • [42] Reinforcement Learning-based control using Q-learning and gravitational search algorithm with experimental validation on a nonlinear servo system
    Zamfirache, Iuliu Alexandru
    Precup, Radu-Emil
    Roman, Raul-Cristian
    Petriu, Emil M.
    INFORMATION SCIENCES, 2022, 583 : 99 - 120
  • [43] Q-learning approach to coordinated optimization of passenger inflow control with train skip-stopping on a urban rail transit line
    Jiang, Zhibin
    Gu, Jinjing
    Fan, Wei
    Liu, Wei
    Zhu, Bingqin
    COMPUTERS & INDUSTRIAL ENGINEERING, 2019, 127 : 1131 - 1142
  • [44] Cooperative Deep Q-Learning With Q-Value Transfer for Multi-Intersection Signal Control
    Ge, Hongwei
    Song, Yumei
    Wu, Chunguo
    Ren, Jiankang
    Tan, Guozhen
    IEEE ACCESS, 2019, 7 : 40797 - 40809
  • [45] Autonomous Vehicle Motion Control and Energy Optimization Based on Q-Learning for a 4-Wheel Independently Driven Electric Vehicle
    Hou, Shengyan
    Chen, Hong
    Liu, Jinfa
    Wang, Yilin
    Liu, Xuan
    Lin, Runzi
    Gao, Jinwu
    UNMANNED SYSTEMS, 2025,
  • [46] Interfering sensed input classification model using assimilated whale optimization and deep Q-learning for remote patient monitoring
    Johar, Sayyed
    Manjula, G. R.
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 93
  • [47] Online Markov Chain-based energy management for a hybrid tracked vehicle with speedy Q-learning
    Liu, Teng
    Wang, Bo
    Yang, Chenglang
    ENERGY, 2018, 160 : 544 - 555
  • [48] Horizontal trajectory control of stratospheric airships in wind field using Q-learning algorithm
    Yang, Xiaowei
    Yang, Xixiang
    Deng, Xiaolong
    AEROSPACE SCIENCE AND TECHNOLOGY, 2020, 106
  • [49] Intelligent Traffic Light Control by Exploring Strategies in an Optimised Space of Deep Q-Learning
    Liu, Junxiu
    Qin, Sheng
    Luo, Yuling
    Wang, Yanhu
    Yang, Su
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (06) : 5960 - 5970
  • [50] Q-Learning based Maximum Power Point Tracking Control for Microbial Fuel Cell
    Fan, Li-ping
    Feng, Xiang
    INTERNATIONAL JOURNAL OF ELECTROCHEMICAL SCIENCE, 2020, 15 (10): : 9917 - 9932