共 18 条
Reinforcement Learning-Based Near Optimization for Continuous-Time Markov Jump Singularly Perturbed Systems
被引:12
作者:
Wang, Jing
[1
,2
]
Peng, Chuanjun
[1
,2
]
Park, Ju H.
[3
]
Shen, Hao
[1
,2
]
Shi, Kaibo
[4
]
机构:
[1] Anhui Univ Technol, Anhui Prov Key Lab Special Heavy Load Robot, Maanshan 243032, Peoples R China
[2] Anhui Univ Technol, Sch Elect & Informat Engn, Maanshan 243032, Peoples R China
[3] Yeungnam Univ, Dept Elect Engn, Gyongsan 38541, South Korea
[4] Chengdu Univ, Sch Informat Sci & Engn, Chengdu 610106, Peoples R China
基金:
新加坡国家研究基金会;
中国国家自然科学基金;
关键词:
Markov processes;
Switches;
Performance analysis;
Optimal control;
Integrated circuit modeling;
Riccati equations;
Reinforcement learning;
Fast and slow decomposition technique;
markov jump systems;
reinforcement learning;
singularly perturbed systems;
ADAPTIVE OPTIMAL-CONTROL;
LINEAR-SYSTEMS;
STABILIZATION;
D O I:
10.1109/TCSII.2022.3233790
中图分类号:
TM [电工技术];
TN [电子技术、通信技术];
学科分类号:
0808 ;
0809 ;
摘要:
The design of a suboptimal controller for continuous-time Markov jump singularly perturbed systems with partially unknown dynamics is studied in this brief. With fast and slow decomposition technique, the original Markov jump singularly perturbed systems are decomposed into fast and slow subsystems as a new attempt. On this basis, an offline parallel Kleinman algorithm and an online parallel integral reinforcement learning algorithm are presented to cope with the different subsystems, respectively. Meanwhile, the controllers obtained by the above two algorithms are used to design the suboptimal controllers for original systems. Furthermore, the suboptimality of the proposed controllers is also discussed. Finally, an example of the electric circuit model is shown to illustrate the applicability of the proposed method.
引用
收藏
页码:2026 / 2030
页数:5
相关论文