Reinforcement learning and adaptive optimization of a class of Markov jump systems with completely unknown dynamic information

被引:0
|
作者
Shuping He
Maoguang Zhang
Haiyang Fang
Fei Liu
Xiaoli Luan
Zhengtao Ding
机构
[1] Anhui University,School of Electrical Engineering and Automation
[2] Anhui University,Institute of Physical Science and Information Technology
[3] Jiangnan University,Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), Institute of Automation
[4] The University of Manchester,School of Electrical and Electronic Engineering
来源
Neural Computing and Applications | 2020年 / 32卷
关键词
Markov jump linear systems (MJLSs); Adaptive optimal control; Online; Reinforcement learning (RL); Coupled algebraic Riccati equations (AREs);
D O I
暂无
中图分类号
学科分类号
摘要
In this paper, an online adaptive optimal control problem of a class of continuous-time Markov jump linear systems (MJLSs) is investigated by using a parallel reinforcement learning (RL) algorithm with completely unknown dynamics. Before collecting and learning the subsystems information of states and inputs, the exploration noise is firstly added to describe the actual control input. Then, a novel parallel RL algorithm is used to parallelly compute the corresponding N coupled algebraic Riccati equations by online learning. By this algorithm, we will not need to know the dynamic information of the MJLSs. The convergence of the proposed algorithm is also proved. Finally, the effectiveness and applicability of this novel algorithm is illustrated by two simulation examples.
引用
收藏
页码:14311 / 14320
页数:9
相关论文
共 50 条
  • [21] Optimized Backstepping Consensus Control Using Reinforcement Learning for a Class of Nonlinear Strict-Feedback-Dynamic Multi-Agent Systems
    Wen, Guoxing
    Chen, C. L. Philip
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (03) : 1524 - 1536
  • [22] Dynamic Event-Triggered Reinforcement Learning Control of Stochastic Nonlinear Systems
    Zhu, Hao-Yang
    Li, Yuan-Xin
    Tong, Shaocheng
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2023, 31 (09) : 2917 - 2928
  • [23] Reinforcement Learning-Based Optimal Stabilization for Unknown Nonlinear Systems Subject to Inputs With Uncertain Constraints
    Zhao, Bo
    Liu, Derong
    Luo, Chaomin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (10) : 4330 - 4340
  • [24] Optimal Output Regulation of Linear Discrete-Time Systems With Unknown Dynamics Using Reinforcement Learning
    Jiang, Yi
    Kiumarsi, Bahare
    Fan, Jialu
    Chai, Tianyou
    Li, Jinna
    Lewis, Frank L.
    IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (07) : 3147 - 3156
  • [25] Reinforcement Learning of Structured Stabilizing Control for Linear Systems With Unknown State Matrix
    Mukherjee, Sayak
    Vu, Thanh Long
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (03) : 1746 - 1752
  • [26] Multiagent Meta-Reinforcement Learning for Adaptive Multipath Routing Optimization
    Chen, Long
    Hu, Bin
    Guan, Zhi-Hong
    Zhao, Lian
    Shen, Xuemin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (10) : 5374 - 5386
  • [27] A Novel Adaptive Control Design for a Class of Nonstrict-Feedback Discrete-Time Systems via Reinforcement Learning
    Bai, Weiwei
    Li, Tieshan
    Long, Yue
    Chen, C. L. Philip
    Xiao, Yang
    Li, Wenjiang
    Li, Ronghui
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (02): : 1250 - 1262
  • [28] How an adaptive learning rate benefits neuro-fuzzy reinforcement learning systems
    Kuremoto, Takashi (wu@yamaguchi-u.ac.jp), 1600, Springer Verlag (8794): : 324 - 331
  • [29] How an Adaptive Learning Rate Benefits Neuro-Fuzzy Reinforcement Learning Systems
    Kuremoto, Takashi
    Obayashi, Masanao
    Kobayashi, Kunikazu
    Mabu, Shingo
    ADVANCES IN SWARM INTELLIGENCE, PT1, 2014, 8794 : 324 - 331
  • [30] Adaptive Observation-Based Efficient Reinforcement Learning for Uncertain Systems
    Ran, Maopeng
    Xie, Lihua
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (10) : 5492 - 5503