Synchronous Reinforcement Learning-Based Control for Cognitive Autonomy

被引:29
作者
Vamvoudakis, Kyriakos G. [1 ]
Kokolakis, Nick-Marios T. [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
来源
FOUNDATIONS AND TRENDS IN SYSTEMS AND CONTROL | 2020年 / 8卷 / 1-2期
基金
美国国家科学基金会;
关键词
ADAPTIVE OPTIMAL-CONTROL; OPTIMAL TRACKING CONTROL; EVENT-TRIGGERED CONTROL; TIME LINEAR-SYSTEMS; H-INFINITY CONTROL; ZERO-SUM GAMES; NONLINEAR-SYSTEMS; STACKELBERG STRATEGY; MULTIAGENT SYSTEMS; CONSENSUS PROBLEMS;
D O I
10.1561/2600000022
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This monograph provides an exposition of recently developed reinforcement learning-based techniques for decision and control in human-engineered cognitive systems. The developed methods learn the solution to optimal control, zero-sum, non zero-sum, and graphical game problems completely online by using measured data along the system trajectories and have proved stability, optimality, and robustness. It is true that games have been shown to be important in robust control for disturbance rejection, and in coordinating activities among multiple agents in networked teams. We also consider cases with intermittent (an analogous to triggered control) instead of continuous learning and apply those techniques for optimal regulation and optimal tracking. We also introduce a bounded rational model to quantify the cognitive skills of a reinforcement learning agent. In order to do that, we leverage ideas from behavioral psychology to formulate differential games where the interacting learning agents have different intelligence skills, and we introduce an iterative method of optimal responses that determine the policy of an agent in adversarial environments. Finally, we present applications of reinforcement learning to motion planning and collaborative target tracking of bounded rational unmanned aerial vehicles.
引用
收藏
页码:1 / 175
页数:175
相关论文
共 50 条
[31]   Quantifying the impact of non-stationarity in reinforcement learning-based traffic signal control [J].
Alegre, Lucas N. ;
Bazzan, Ana L. C. ;
da Silva, Bruno C. .
PEERJ COMPUTER SCIENCE, 2021,
[32]   Safe Reinforcement Learning-Based Robust Approximate Optimal Control for Hypersonic Flight Vehicles [J].
Shi, Lei ;
Wang, Xuesong ;
Cheng, Yuhu .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (09) :11401-11414
[33]   A Reinforcement Learning-Based Control Approach for Unknown Nonlinear Systems with Persistent Adversarial Inputs [J].
Zhong, Xiangnan ;
He, Haibo .
2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
[34]   Reinforcement Learning-Based Adaptive Optimal Control for Partially Unknown Systems Using Differentiator [J].
Guo, Xinxin ;
Yan, Weisheng ;
Cui, Rongxin .
2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, :1039-1044
[35]   Quantifying the Impact of Non-Stationarity in Reinforcement Learning-Based Traffic Signal Control [J].
Alegre L.N. ;
Bazzan A.L.C. ;
da Silva B.C. .
PeerJ Computer Science, 2021, 7 :1-20
[36]   Reinforcement learning-based optimal control for Markov jump systems with completely unknown dynamics [J].
Shi, Xiongtao ;
Li, Yanjie ;
Du, Chenglong ;
Chen, Chaoyang ;
Zong, Guangdeng ;
Gui, Weihua .
AUTOMATICA, 2025, 171
[37]   Reinforcement Learning-Based Event-Triggered Optimal Control of Power Systems With Control Input Saturation [J].
Gu, Zhou ;
Cao, Ruiyan ;
Tian, Engang .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2025, 21 (02) :1528-1536
[38]   Model-free Nearly Optimal Control of Constrained-Input Nonlinear Systems Based on Synchronous Reinforcement Learning [J].
Zhao, Han ;
Guo, Lei .
2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, :2162-2167
[39]   A reinforcement learning-based routing for delay tolerant networks [J].
Rolla, Vitor G. ;
Curado, Marilia .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2013, 26 (10) :2243-2250
[40]   Reinforcement Learning-based Optimization of EBike Charging infrastructure [J].
Sharif, Muddsair ;
Seker, Huseyin .
2024 IEEE/ACM INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING, APPLICATIONS AND TECHNOLOGIES, BDCAT, 2024, :18-25