Robust control for affine nonlinear systems under the reinforcement learning framework

被引:2
作者
Guo, Wenxin [1 ]
Qin, Weiwei [1 ]
Lan, Xuguang [2 ]
Liu, Jieyu [1 ]
Zhang, Zhaoxiang [1 ]
机构
[1] Xian Res Inst High Tech, Xian 710025, Peoples R China
[2] Xi An Jiao Tong Univ, Xian 710049, Peoples R China
基金
中国国家自然科学基金;
关键词
Robust control; Adaptive dynamic programming; Uncertainty estimation; Utility function; TRACKING CONTROL; STABILIZATION; ALGORITHM; DESIGN;
D O I
10.1016/j.neucom.2024.127631
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article investigates the robust control problem of affine nonlinear systems with both additive and multiplicative uncertainty. Different from existing actor -critic (AC) algorithms for adaptive dynamic programming (ADP), we introduce an uncertainty estimator and propose an actor -critic -estimator (ACE) algorithm. The proposed algorithm alternates between the value evaluation, uncertainty estimation, and policy update to generate the adaptive robust control law without knowing the system dynamics. Especially, during the step of uncertainty estimation, we approximate the uncertainty by a radial basis function neural network (RBFNN) and design the appropriate utility function accordingly instead of using the supremum of the uncertainty as in existing studies. The Lyapunov stability theorem provides theoretical demonstrations of the stability and convergence. We further demonstrate that the affine nonlinear systems with uncertainty is uniformly ultimately bounded (UUB) stable when the learned adaptive robust control law is adopted. The performance of the proposed algorithm is demonstrated through a torsion pendulum system and an inverted pendulum system.
引用
收藏
页数:8
相关论文
共 30 条
[11]   H∞ Tracking Control of Completely Unknown Continuous-Time Systems via Off-Policy Reinforcement Learning [J].
Modares, Hamidreza ;
Lewis, Frank L. ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (10) :2550-2562
[12]   Optimal tracking control of nonlinear partially-unknown constrained-input systems using integral reinforcement learning [J].
Modares, Hamidreza ;
Lewis, Frank L. .
AUTOMATICA, 2014, 50 (07) :1780-1792
[13]   Integral reinforcement learning and experience replay for adaptive optimal control of partially-unknown constrained-input continuous-time systems [J].
Modares, Hamidreza ;
Lewis, Frank L. ;
Naghibi-Sistani, Mohammad-Bagher .
AUTOMATICA, 2014, 50 (01) :193-202
[14]   A policy iteration approach to online optimal control of continuous-time constrained-input systems [J].
Modares, Hamidreza ;
Sistani, Mohammad-Bagher Naghibi ;
Lewis, Frank L. .
ISA TRANSACTIONS, 2013, 52 (05) :611-621
[15]   A STABILIZATION ALGORITHM FOR A CLASS OF UNCERTAIN LINEAR-SYSTEMS [J].
PETERSEN, IR .
SYSTEMS & CONTROL LETTERS, 1987, 8 (04) :351-357
[16]   Neural Network Observers and Sensorless Robust Optimal Control for Partially Unknown PMSM With Disturbances and Saturating Voltages [J].
Tan, Luy Nguyen ;
Cong, Thanh Pham ;
Cong, Duy Pham .
IEEE TRANSACTIONS ON POWER ELECTRONICS, 2021, 36 (10) :12045-12056
[17]   Online actor-critic algorithm to solve the continuous-time infinite horizon optimal control problem [J].
Vamvoudakis, Kyriakos G. ;
Lewis, Frank L. .
AUTOMATICA, 2010, 46 (05) :878-888
[18]   Self-Learning Robust Control Synthesis and Trajectory Tracking of Uncertain Dynamics [J].
Wang, Ding ;
Cheng, Long ;
Yan, Jun .
IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (01) :278-286
[19]   Self-Learning Optimal Regulation for Discrete-Time Nonlinear Systems Under Event-Driven Formulation [J].
Wang, Ding ;
Ha, Mingming ;
Qiao, Junfei .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2020, 65 (03) :1272-1279
[20]   Neural Network Learning and Robust Stabilization of Nonlinear Systems With Dynamic Uncertainties [J].
Wang, Ding ;
Liu, Derong ;
Mu, Chaoxu ;
Zhang, Yun .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (04) :1342-1351