Safe online optimization of motor speed synchronization control with incremental Q-learning

被引:0
|
作者
Huang, Jianfeng [1 ]
Lu, Guoqiang [1 ]
Yao, Xudong [1 ]
机构
[1] Shantou Univ, Coll Engn, Shantou 515063, Peoples R China
关键词
Online controller tuning; Safe reinforcement learning; Q; -learning; Motor speed synchronization; PARTICLE SWARM OPTIMIZATION; ENERGY MANAGEMENT; REINFORCEMENT; DESIGN; STRATEGY;
D O I
10.1016/j.eswa.2024.124622
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning (RL) is promising for online controller optimization. However, its practical application has been hindered by safety issues. This paper proposes an algorithm named Incremental Q-learning (IQ) and applies it to the online optimization of motor speed synchronization control. IQ ensures safe learning by adopting so-called incremental action variables which represent incremental change rather than absolute magnitude, and dividing the one-round learning process in the classic Q-learning (in this paper referred to as Absolute Qlearning, AQ) into multiple consecutive ones with the Q table getting reset at the beginning of each round. Since the permitted interval of change is restricted to be very small, the agent can learn its way safely, steadily, and robustly towards the optimal policy. Simulation results show that IQ is advantageous to AQ in optimality, safety, and adaptability. IQ converges to better final performances with significantly smaller performance variance along the whole learning process, smaller torque trajectory deviation between consecutive episodes and adapts to unknown disturbances faster. It is of great potential for online controller optimization/tuning in practical engineering projects. Source code and demos are provided.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] SEM: Safe exploration mask for q-learning
    Xuan, Chengbin
    Zhang, Feng
    Lam, Hak-Keung
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 111
  • [2] Incremental Q-learning strategy for adaptive PID control of mobile robots
    Carlucho, Ignacio
    De Paula, Mariano
    Villar, Sebastian A.
    Acosta, Gerardo G.
    EXPERT SYSTEMS WITH APPLICATIONS, 2017, 80 : 183 - 199
  • [3] An Online Home Energy Management System using Q-Learning and Deep Q-Learning
    Izmitligil, Hasan
    Karamancioglu, Abdurrahman
    SUSTAINABLE COMPUTING-INFORMATICS & SYSTEMS, 2024, 43
  • [4] An Online Learning Control Strategy for Hybrid Electric Vehicle Based on Fuzzy Q-Learning
    Hu, Yue
    Li, Weimin
    Xu, Hui
    Xu, Guoqing
    ENERGIES, 2015, 8 (10): : 11167 - 11186
  • [5] Adaptive Speed Control of Electric Vehicles Based on Multi-Agent Fuzzy Q-Learning
    Gheisarnejad, Meysam
    Mirzavand, Ghazal
    Ardeshiri, Reza Rouhi
    Andresen, Bjorn
    Khooban, Mohammad Hassan
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2023, 7 (01): : 102 - 110
  • [6] Safe Q-Learning Method Based on Constrained Markov Decision Processes
    Ge, Yangyang
    Zhu, Fei
    Lin, Xinghong
    Liu, Quan
    IEEE ACCESS, 2019, 7 : 165007 - 165017
  • [7] Q-Learning Scheduling for Tracking Current Control of Switched Reluctance Motor Drives
    Alharkan, Hamad
    Shamsi, Pourya
    Saadatmand, Sepehr
    Ferdowsi, Mehdi
    2020 IEEE POWER AND ENERGY CONFERENCE AT ILLINOIS (PECI), 2020,
  • [8] Deep Q-Learning and Particle Swarm Optimization for Bot Detection in Online Social Networks
    Lingam, Greeshma
    Rout, Rashmi Ranjan
    Somayajulu, D. V. L. N.
    2019 10TH INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION AND NETWORKING TECHNOLOGIES (ICCCNT), 2019,
  • [9] Solving Reward-Collecting Problems with UAVs: A Comparison of Online Optimization and Q-Learning
    Yixuan Liu
    Chrysafis Vogiatzis
    Ruriko Yoshida
    Erich Morman
    Journal of Intelligent & Robotic Systems, 2022, 104
  • [10] A Fuzzy Q-Learning Algorithm for Storage Optimization in Islanding Microgrid
    Yu, Yunjun
    Qin, Yang
    Gong, Hancheng
    JOURNAL OF ELECTRICAL ENGINEERING & TECHNOLOGY, 2021, 16 (05) : 2343 - 2353