Optimal tracking control for completely unknown nonlinear discrete-time Markov jump systems using data-based reinforcement learning method

被引:32
作者
Jiang, He [1 ]
Zhang, Huaguang [1 ]
Luo, Yanhong [1 ]
Wang, Junyi [1 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Box 134, Shenyang 110819, Peoples R China
基金
中国国家自然科学基金;
关键词
Optimal tracking control; Markov jump systems; Data-based; Reinforcement learning; Adaptive dynamic programming; Neural networks; SYNCHRONIZATION CONTROL; GRAPHICAL GAMES; CONTROL SCHEME; STABILITY; ALGORITHM;
D O I
10.1016/j.neucom.2016.02.029
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we develop a novel optimal tracking control scheme for a class of nonlinear discrete-time Markov jump systems (MJSs) by utilizing a data-based reinforcement learning method. It is not practical to obtain accurate system models of the real-world MJSs due to the existence of abrupt variations in their system structures. Consequently, most traditional model-based methods for MJSs are invalid for the practical engineering applications. In order to overcome the difficulties without any identification scheme which would cause estimation errors, a model-free adaptive dynamic programming (ADP) algorithm will be designed by using system data rather than accurate system functions. Firstly, we combine the tracking error dynamics and reference system dynamics to form an augmented system. Then, based on the augmented system, a new performance index function with discount factor is formulated for the optimal tracking control problem via Markov chain and weighted sum technique. Neural networks are employed to implement the on-line ADP learning algorithm. Finally, a simulation example is given to demonstrate the effectiveness of our proposed approach. (C) 2016 Elsevier B.V. All rights reserved.
引用
收藏
页码:176 / 182
页数:7
相关论文
共 50 条
[41]   Value Iteration Adaptive Dynamic Programming for Optimal Control of Discrete-Time Nonlinear Systems [J].
Wei, Qinglai ;
Liu, Derong ;
Lin, Hanquan .
IEEE TRANSACTIONS ON CYBERNETICS, 2016, 46 (03) :840-853
[42]   H∞ tracking control for linear discrete-time systems via reinforcement learning [J].
Liu, Ying-Ying ;
Wang, Zhan-Shan ;
Shi, Zhan .
INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2020, 30 (01) :282-301
[43]   Actor-Critic-Based Optimal Tracking for Partially Unknown Nonlinear Discrete-Time Systems [J].
Kiumarsi, Bahare ;
Lewis, Frank L. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (01) :140-151
[44]   Reinforcement learning-based neural control for discrete-time nonlinear systems via deterministic learning [J].
Zhang, Fukai ;
Zhu, Zejian ;
Chen, Tianrui ;
Wu, Weiming ;
Wang, Cong .
NONLINEAR DYNAMICS, 2024, :9981-10003
[45]   Discrete-Time Nonlinear Optimal Control Using Multi-Step Reinforcement Learning [J].
An, Ningbo ;
Wang, Qishao ;
Zhao, Xiaochuan ;
Wang, Qingyun .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (04) :2279-2283
[46]   Data-Based Self-Learning Optimal Control for Continuous-Time Unknown Nonlinear Systems With Disturbance [J].
Wei, Qinglai ;
Liu, Derong ;
Song, Ruizhuo ;
Yan, Pengfei .
PROCEEDINGS OF THE 28TH CHINESE CONTROL AND DECISION CONFERENCE (2016 CCDC), 2016, :6633-6638
[47]   Adaptive optimal control of unknown discrete-time linear systems with guaranteed prescribed degree of stability using reinforcement learning [J].
Seyed Ehsan Razavi ;
Mohammad Amin Moradi ;
Saeed Shamaghdari ;
Mohammad Bagher Menhaj .
International Journal of Dynamics and Control, 2022, 10 :870-878
[48]   Near-Optimal Tracking Control of Partially Unknown Discrete-Time Nonlinear Systems Based on Radial Basis Function Neural Network [J].
Huang, Jiashun ;
Xu, Dengguo ;
Li, Yahui ;
Ma, Yan .
MATHEMATICS, 2024, 12 (08)
[49]   Neural-Network-Based Robust Optimal Tracking Control for MIMO Discrete-Time Systems With Unknown Uncertainty Using Adaptive Critic Design [J].
Liu, Lei ;
Wang, Zhanshan ;
Zhang, Huaguang .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (04) :1239-1251
[50]   Adaptive Reinforcement Learning Control Based on Neural Approximation for Nonlinear Discrete-Time Systems With Unknown Nonaffine Dead-Zone Input [J].
Liu, Yan-Jun ;
Li, Shu ;
Tong, Shaocheng ;
Chen, C. L. Philip .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (01) :295-305