Event-triggered-based online integral reinforcement learning for optimal control of unknown constrained nonlinear systems

被引:2
作者
Han, Xiumei [1 ]
Zhao, Xudong [1 ]
Wang, Ding [2 ]
Wang, Bohui [3 ,4 ]
机构
[1] Dalian Univ Technol, Key Lab Intelligent Control & Optimizat Ind Equip, Dalian, Peoples R China
[2] Beijing Univ Technol, Fac Informat, Beijing, Peoples R China
[3] Shanghai Jiao Tong Univ, Minist Educ China, Dept Automat, Automat, Shanghai, Peoples R China
[4] Shanghai Jiao Tong Univ, Minist Educ China, Key Lab Syst Control & Informat Proc, Shanghai, Peoples R China
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
Optimal event-triggered control; event-triggered-based integral reinforcement learning; unknown nonlinear systems; constrained control input; OPTIMAL TRACKING CONTROL; DISCRETE-TIME-SYSTEMS; APPROXIMATE OPTIMAL-CONTROL; ADAPTIVE OPTIMAL-CONTROL; SLIDING-MODE CONTROL; POLICY ITERATION; CONTROL DESIGN; ALGORITHM;
D O I
10.1080/00207179.2022.2137852
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
For unknown nonlinear systems with actuator saturation, an online policy iteration-based algorithm is employed to solve the optimal event-triggered control problem. To learn the system dynamics, a novel identifier is proposed to make the estimation error converge quickly and the experience replay technique is employed to release the persistence of the excitation condition. To approximate the cost function and the event-triggered control law, we present event-triggered-based critic and actor networks, whose weights are updated only at triggered instants. During the policy iteration process, an event-triggered-based integral reinforcement learning method is proposed to solve the Hamilton-Jacobi-Bellman equation. By utilising the integral reinforcement learning, the network resource is saved and learning efficiency is improved. Based on the Lyapunov method, stability for the closed-loop system and estimation errors for the three networks are analysed. At last, simulation results of two numerical examples are used to show the effectiveness of the proposed method.
引用
收藏
页码:213 / 225
页数:13
相关论文
共 39 条
  • [1] Discrete-time nonlinear HJB solution using approximate dynamic programming: Convergence proof
    Al-Tamimi, Asma
    Lewis, Frank
    [J]. 2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 38 - +
  • [2] Semi-global adaptive backstepping control for parametric strict-feedback systems with non-triangular structural uncertainties?
    Cai, Jianping
    Mei, Congli
    Yan, Qiuzhen
    [J]. ISA TRANSACTIONS, 2022, 126 : 180 - 189
  • [3] Switched-observer-based adaptive output-feedback control design with unknown gain for pure-feedback switched nonlinear systems via average dwell time
    Chang, Yi
    Zhou, Peng
    Niu, Ben
    Wang, Huanqing
    Xu, Ning
    Alassafi, M. O.
    Ahmad, A. M.
    [J]. INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2021, 52 (09) : 1731 - 1745
  • [4] Adaptive Actor-Critic Design-Based Integral Sliding-Mode Control for Partially Unknown Nonlinear Systems With Input Disturbances
    Fan, Quan-Yong
    Yang, Guang-Hong
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (01) : 165 - 177
  • [5] Distributed optimal control for continuous-time nonaffine nonlinear interconnected systems
    Farzanegan, Behzad
    Suratgar, Amir Abolfazl
    Menhaj, Mohammad Bagher
    Zamani, Mohsen
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 2022, 95 (12) : 3462 - 3476
  • [6] Zero-sum differential games-based fast adaptive robust optimal sliding mode control design for uncertain missile autopilot with constrained input
    Gao, Yuxin
    Liu, Chunsheng
    Jiang, Sen
    Zhang, Shaojie
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 2022, 95 (07) : 1789 - 1801
  • [7] Integral Reinforcement Learning for Continuous-Time Input-Affine Nonlinear Systems With Simultaneous Invariant Explorations
    Lee, Jae Young
    Park, Jin Bae
    Choi, Yoon Ho
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (05) : 916 - 932
  • [8] Reinforcement Learning and Feedback Control USING NATURAL DECISION METHODS TO DESIGN OPTIMAL ADAPTIVE CONTROLLERS
    Lewis, Frank L.
    Vrabie, Draguna
    Vamvoudakis, Kyriakos G.
    [J]. IEEE CONTROL SYSTEMS MAGAZINE, 2012, 32 (06): : 76 - 105
  • [9] Policy Iteration Adaptive Dynamic Programming Algorithm for Discrete-Time Nonlinear Systems
    Liu, Derong
    Wei, Qinglai
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (03) : 621 - 634
  • [10] Neural-network-observer-based optimal control for unknown nonlinear systems using adaptive dynamic programming
    Liu, Derong
    Huang, Yuzhu
    Wang, Ding
    Wei, Qinglai
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 2013, 86 (09) : 1554 - 1566