Learning event-triggered control from data through joint optimization

被引:14
作者
Funk, Niklas [1 ]
Baumann, Dominik [1 ,3 ]
Berenz, Vincent [2 ]
Trimpe, Sebastian [1 ,3 ]
机构
[1] Max Planck Inst Intelligent Syst, Intelligent Control Syst Grp, Stuttgart, Germany
[2] Max Planck Inst Intelligent Syst, Empir Inference Dept, Tubingen, Germany
[3] Rhein Westfal TH Aachen, Inst Data Sci Mech Engn, Aachen, Germany
关键词
Event-triggered control; Reinforcement learning; Stability verification; Neural networks; ALGORITHM;
D O I
10.1016/j.ifacsc.2021.100144
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a framework for model-free learning of event-triggered control strategies. Event-triggered methods aim to achieve high control performance while only closing the feedback loop when needed. This enables resource savings, e.g., network bandwidth if control commands are sent via communication networks, as in networked control systems. Event-triggered controllers consist of a communication policy, determining when to communicate, and a control policy, deciding what to communicate. It is essential to jointly optimize the two policies since individual optimization does not necessarily yield the overall optimal solution. To address this need for joint optimization, we propose a novel algorithm based on hierarchical reinforcement learning. The resulting algorithm is shown to accomplish high-performance control in line with resource savings and scales seamlessly to nonlinear and high-dimensional systems. The method's applicability to real-world scenarios is demonstrated through experiments on a six degrees of freedom real-time controlled manipulator. Further, we propose an approach towards evaluating the stability of the learned neural network policies. (C) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页数:14
相关论文
共 45 条
[1]  
[Anonymous], 2003, Linear programming: Theory and extensions
[2]  
[Anonymous], ARXIV170706347
[3]   System Architectures, Protocols and Algorithms for Aperiodic Wireless Control Systems [J].
Araujo, Jose ;
Mazo, Manuel, Jr. ;
Anta, Adolfo ;
Tabuada, Paulo ;
Johansson, Karl H. .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2014, 10 (01) :175-184
[4]  
Bacon PL, 2017, AAAI CONF ARTIF INTE, P1726
[5]  
Baumann D, 2018, IEEE DECIS CONTR P, P943, DOI 10.1109/CDC.2018.8619335
[6]  
Bonassi Fabio, 2020, LEARNING DYNAMICS CO
[7]  
Brockman Greg, 2016, arXiv
[8]  
Carpentier J, 2019, IEEE/SICE I S SYS IN, P614, DOI 10.1109/SII.2019.8700380
[9]   DEEPCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling [J].
Demirel, Burak ;
Ramaswamy, Arunselvan ;
Quevedo, Daniel E. ;
Karl, Holger .
IEEE CONTROL SYSTEMS LETTERS, 2018, 2 (04) :737-742
[10]   Event-Triggered Control for String-Stable Vehicle Platooning [J].
Dolk, Victor S. ;
Ploeg, Jeroen ;
Heemels, W. P. Maurice H. .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2017, 18 (12) :3486-3500