Robust control for affine nonlinear systems under the reinforcement learning framework

被引:2
作者
Guo, Wenxin [1 ]
Qin, Weiwei [1 ]
Lan, Xuguang [2 ]
Liu, Jieyu [1 ]
Zhang, Zhaoxiang [1 ]
机构
[1] Xian Res Inst High Tech, Xian 710025, Peoples R China
[2] Xi An Jiao Tong Univ, Xian 710049, Peoples R China
基金
中国国家自然科学基金;
关键词
Robust control; Adaptive dynamic programming; Uncertainty estimation; Utility function; TRACKING CONTROL; STABILIZATION; ALGORITHM; DESIGN;
D O I
10.1016/j.neucom.2024.127631
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article investigates the robust control problem of affine nonlinear systems with both additive and multiplicative uncertainty. Different from existing actor -critic (AC) algorithms for adaptive dynamic programming (ADP), we introduce an uncertainty estimator and propose an actor -critic -estimator (ACE) algorithm. The proposed algorithm alternates between the value evaluation, uncertainty estimation, and policy update to generate the adaptive robust control law without knowing the system dynamics. Especially, during the step of uncertainty estimation, we approximate the uncertainty by a radial basis function neural network (RBFNN) and design the appropriate utility function accordingly instead of using the supremum of the uncertainty as in existing studies. The Lyapunov stability theorem provides theoretical demonstrations of the stability and convergence. We further demonstrate that the affine nonlinear systems with uncertainty is uniformly ultimately bounded (UUB) stable when the learned adaptive robust control law is adopted. The performance of the proposed algorithm is demonstrated through a torsion pendulum system and an inverted pendulum system.
引用
收藏
页数:8
相关论文
共 30 条
[1]   Discrete-time nonlinear HJB solution using approximate dynamic programming: Convergence proof [J].
Al-Tamimi, Asma ;
Lewis, Frank L. ;
Abu-Khalaf, Murad .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2008, 38 (04) :943-949
[2]   Robust Optimal Control for the Vehicle Suspension System With Uncertainties [J].
Bai, Rui ;
Wang, He-Bin .
IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (09) :9263-9273
[3]   A Novel Reinforcement Learning-Based Robust Control Strategy for a Quadrotor [J].
Hua, Hean ;
Fang, Yongchun .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2023, 70 (03) :2812-2821
[4]   Robust Precision Position Tracking of Planar Motors Using Min-Max Model Predictive Control [J].
Huang, Su-Dan ;
Peng, Kai-Yu ;
Cao, Guang-Zhong ;
Wu, Chao ;
Xu, Junqi ;
He, Jiangbiao .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2022, 69 (12) :13265-13276
[5]   Reinforcement Learning and Adaptive Dynamic Programming for Feedback Control [J].
Lewis, Frank L. ;
Vrabie, Draguna .
IEEE CIRCUITS AND SYSTEMS MAGAZINE, 2009, 9 (03) :32-50
[6]   Adaptive Interleaved Reinforcement Learning: Robust Stability of Affine Nonlinear Systems With Unknown Uncertainty [J].
Li, Jinna ;
Ding, Jinliang ;
Chai, Tianyou ;
Lewis, Frank L. ;
Jagannathan, Sarangapani .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (01) :270-280
[7]   Off-Policy Interleaved Q-Learning: Optimal Control for Affine Nonlinear Discrete-Time Systems [J].
Li, Jinna ;
Chai, Tianyou ;
Lewis, Frank L. ;
Ding, Zhengtao ;
Jiang, Yi .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) :1308-1320
[8]   Off-Policy Reinforcement Learning for Synchronization in Multiagent Graphical Games [J].
Li, Jinna ;
Modares, Hamidreza ;
Chai, Tianyou ;
Lewis, Frank L. ;
Xie, Lihua .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (10) :2434-2445
[9]   An optimal control approach to robust control design [J].
Lin, F .
INTERNATIONAL JOURNAL OF CONTROL, 2000, 73 (03) :177-186
[10]   Reinforcement-Learning-Based Robust Controller Design for Continuous-Time Uncertain Nonlinear Systems Subject to Input Constraints [J].
Liu, Derong ;
Yang, Xiong ;
Wang, Ding ;
Wei, Qinglai .
IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (07) :1372-1385