Design of a reinforcement learning PID controller

被引:20
|
作者
Guan, Zhe [1 ]
Yamamoto, Toru [2 ]
机构
[1] Hiroshima Univ, Dream Driven CoCreat Res Ctr, KOBELCO Construct Machinery, 1-4-1 Kagamiyama, Higashihiroshima 7398527, Japan
[2] Hiroshima Univ, Acad Sci & Technol, 1-4-1 Kagamiyama, Higashihiroshima 7398527, Japan
关键词
reinforcement learning; PID control; Actor-Critic learning; RBF network; nonlinear system; NETWORKS;
D O I
10.1002/tee.23430
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper addresses a design scheme of a proportional-integral-derivative (PID) controller with a new adaptive updating rule based on reinforcement learning (RL) approach for nonlinear systems. A new design scheme that RL can be used to complement the conventional PID control technology is presented. In the proposed scheme, a single radial basis function (RBF) network is considered to calculate the control policy function of Actor and the value function of Critic simultaneously. Regarding the PID controller structure, the inputs of RBF network are system errors, the difference of output as well as the second-order difference of output, and they are composed of system states. The temporal difference (TD) error in the proposed scheme involves the reinforcement signal, the current and the previous stored value of the value function. The gradient descent method is adopted based on the TD error performance index, then, the updating rules can be yielded. Therefore, the network weights and the kernel function can be calculated in an adaptive way. Finally, the numerical simulations are conducted in nonlinear systems to illustrate the efficiency and robustness of the proposed scheme. (c) 2021 Institute of Electrical Engineers of Japan. Published by Wiley Periodicals LLC.
引用
收藏
页码:1354 / 1360
页数:7
相关论文
共 50 条
  • [21] Adaptive MPC path-tracking controller based on reinforcement learning and preview-based PID controller
    Feng, Kun
    Li, Xu
    Li, Wenli
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2024,
  • [22] Fuzzy PID controller: Design, tuning and comparison with conventional PID controller
    Khan, Asim Ali
    Rapal, Nishkam
    2006 IEEE INTERNATIONAL CONFERENCE ON ENGINEERING OF INTELLIGENT SYSTEMS, 2006, : 471 - 476
  • [23] Design of Equivalent Fuzzy PID Controller From The Conventional PID Controller
    Manikandan, R.
    Arulprakash, A.
    Arulmozhiyal, R.
    2015 INTERNATIONAL CONFERENCE ON CONTROL, INSTRUMENTATION, COMMUNICATION AND COMPUTATIONAL TECHNOLOGIES (ICCICCT), 2015, : 356 - 362
  • [24] AN ONLINE REINFORCEMENT LEARNING CONTROLLER DESIGN FOR MARS ASCENT VEHICLE
    Bae, Han Woong
    PROCEEDINGS OF THE 44TH ANNUAL AMERICAN ASTRONAUTICAL SOCIETY GUIDANCE, NAVIGATION, AND CONTROL CONFERENCE, AAS 2022, 2024, : 305 - 317
  • [25] Design of ATM congestion controller based on reinforcement learning method
    Li Xin
    Jing Yuan-Wei
    Proceedings of the 2007 Chinese Control and Decision Conference, 2007, : 673 - 676
  • [26] Design of robot fuzzy logic controller based on reinforcement learning
    Duan, Yong
    Liu, Xin-Gang
    Xu, Xin-He
    Xitong Fangzhen Xuebao / Journal of System Simulation, 2006, 18 (06): : 1597 - 1600
  • [27] Online reinforcement learning neural network controller design for nanomanipulation
    Yang, Qinmin
    Jagannathan, S.
    2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 225 - +
  • [28] Demystify PID controller design
    Schene, C
    de la Torre, V
    Waters, M
    CONTROL ENGINEERING, 2004, 51 (07) : IP10 - +
  • [29] Predictive PID controller design
    Uduehi, D
    Ordys, A
    Grimble, M
    PROCEEDINGS OF THE 2001 IEEE INTERNATIONAL CONFERENCE ON CONTROL APPLICATIONS (CCA'01), 2001, : 612 - 617
  • [30] Reinforcement Learning Intersection Controller
    Tolebi, Gulnur
    Dairbekov, Nurlan S.
    Kurmankhojayev, Daniyar
    Mussabayev, Ravil
    2018 14TH INTERNATIONAL CONFERENCE ON ELECTRONICS COMPUTER AND COMPUTATION (ICECCO), 2018,