Safe online optimization of motor speed synchronization control with incremental Q-learning

被引:0
|
作者
Huang, Jianfeng [1 ]
Lu, Guoqiang [1 ]
Yao, Xudong [1 ]
机构
[1] Shantou Univ, Coll Engn, Shantou 515063, Peoples R China
关键词
Online controller tuning; Safe reinforcement learning; Q; -learning; Motor speed synchronization; PARTICLE SWARM OPTIMIZATION; ENERGY MANAGEMENT; REINFORCEMENT; DESIGN; STRATEGY;
D O I
10.1016/j.eswa.2024.124622
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning (RL) is promising for online controller optimization. However, its practical application has been hindered by safety issues. This paper proposes an algorithm named Incremental Q-learning (IQ) and applies it to the online optimization of motor speed synchronization control. IQ ensures safe learning by adopting so-called incremental action variables which represent incremental change rather than absolute magnitude, and dividing the one-round learning process in the classic Q-learning (in this paper referred to as Absolute Qlearning, AQ) into multiple consecutive ones with the Q table getting reset at the beginning of each round. Since the permitted interval of change is restricted to be very small, the agent can learn its way safely, steadily, and robustly towards the optimal policy. Simulation results show that IQ is advantageous to AQ in optimality, safety, and adaptability. IQ converges to better final performances with significantly smaller performance variance along the whole learning process, smaller torque trajectory deviation between consecutive episodes and adapts to unknown disturbances faster. It is of great potential for online controller optimization/tuning in practical engineering projects. Source code and demos are provided.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] Online Adaptive Optimal Control of Discrete-time Linear Systems via Synchronous Q-learning
    Li, Xinxing
    Wang, Xueyuan
    Zha, Wenzhong
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 2024 - 2029
  • [22] Co-evolution of synchronization and cooperation with multi-agent Q-learning
    Zhu, Peican
    Cao, Zhaoheng
    Liu, Chen
    Chu, Chen
    Wang, Zhen
    CHAOS, 2023, 33 (03)
  • [23] A resilience control method for mitigating the sudden change in online group opinion based on Q-Learning and PSO
    Wu, Di
    Hu, Bin
    Ma, Xiaomeng
    Wang, Zhichao
    APPLIED SOFT COMPUTING, 2024, 163
  • [24] On-line set-point optimization for intelligent supervisory control and improvement of Q-learning convergence
    Kim, Song Ho
    Song, Kwang Rim
    Kang, Il Yong
    Hyon, Chung Il
    CONTROL ENGINEERING PRACTICE, 2021, 114
  • [25] Hierarchical Q-learning network for online simultaneous optimization of energy efficiency and battery life of the battery/ultracapacitor electric vehicle
    Xu, Bin
    Zhou, Quan
    Shi, Junzhe
    Li, Sixu
    JOURNAL OF ENERGY STORAGE, 2022, 46
  • [26] A coordination of some basic heuristics using Q-learning for global optimization
    Mandal, Sudip Kumar
    Halim, Kazi Abdul
    Maiti, Manas Kumar
    OPSEARCH, 2024,
  • [27] Online Monitoring of Heterogeneous Partially Observable Data Streams Based on Q-Learning
    Li, Haoqian
    Ye, Honghan
    Cheng, Jing-Ru C.
    Liu, Kaibo
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, : 4802 - 4817
  • [28] MODEL PREDICTIVE CONTROL AND MODEL PREDICTIVE Q-LEARNING FOR STRUCTURAL VIBRATION CONTROL
    Phan, Minh Q.
    Azad, Seyed Mandi B.
    ASTRODYNAMICS 2017, PTS I-IV, 2018, 162 : 17 - 36
  • [29] A Q-learning based optimization method of energy management for peak load control of residential areas with CCHP systems
    Chen, Lingmin
    Wu, Jiekang
    Tang, Huiling
    Feng, Jin
    Wang, Yanan
    ELECTRIC POWER SYSTEMS RESEARCH, 2023, 214
  • [30] Q-learning Based Dynamic Optimal Relax Automatic Generation Control
    Yu, Tao
    Yuan, Ye
    Liang, Haihua
    POWER AND ENERGY ENGINEERING CONFERENCE 2010, 2010, : 797 - 800