Reduced-dimensional reinforcement learning control using singular perturbation approximations

被引:34
|
作者
Mukherjee, Sayak [1 ]
Bai, He [2 ]
Chakrabortty, Aranya [1 ]
机构
[1] North Carolina State Univ, Dept Elect & Comp Engn, Raleigh, NC 27695 USA
[2] Oklahoma State Univ, Mech & Aerosp Engn Dept, Stillwater, OK 74078 USA
基金
美国国家科学基金会;
关键词
Reinforcement learning; Linear quadratic regulator; Singular perturbation; Model-free control; Model reduction; TIME LINEAR-SYSTEMS; ALGORITHM;
D O I
10.1016/j.automatica.2020.109451
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a set of model-free, reduced-dimensional reinforcement learning (RL) based optimal control designs for linear time-invariant singularly perturbed (SP) systems. We first present a state feedback and an output feedback based RL control design for a generic SP system with unknown state and input matrices. We take advantage of the underlying time-scale separation property of the plant to learn a linear quadratic regulator (LQR) for only its slow dynamics, thereby saving significant amount of learning time compared to the conventional full-dimensional RL controller. We analyze the sub-optimality of the designs using SP approximation theorems, and provide sufficient conditions for closed-loop stability. Thereafter, we extend both designs to clustered multi-agent consensus networks, where the SP property reflects through clustering. We develop both centralized and cluster-wise block-decentralized RL controllers for such networks, in reduced dimensions. We demonstrate the details of the implementation of these controllers using simulations of relevant numerical examples, and compare them with conventional RL designs to show the computational benefits of our approach. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Traffic Signal Control Using Reinforcement Learning
    Jadhao, Namrata S.
    Jadhao, Ashish S.
    2014 FOURTH INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS AND NETWORK TECHNOLOGIES (CSNT), 2014, : 1130 - 1135
  • [42] Robot control optimization using reinforcement learning
    Song, KT
    Sun, WY
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 1998, 21 (03) : 221 - 238
  • [43] Formation Control using GQ(λ) Reinforcement Learning
    Knopp, Martin
    Aykin, Can
    Feldmaier, Johannes
    Shen, Hao
    2017 26TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2017, : 1043 - 1048
  • [44] Robot control optimization using reinforcement learning
    Natl Chiao Tung Univ, Hsinchu, Taiwan
    J Intell Rob Syst Theor Appl, 3 (221-238):
  • [45] Multivariable anesthesia control using reinforcement learning
    Sadati, N.
    Aflaki, A.
    Jahed, M.
    2006 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS, VOLS 1-6, PROCEEDINGS, 2006, : 4563 - +
  • [46] Robot Control Optimization Using Reinforcement Learning
    Kai-Tai Song
    Wen-Yu Sun
    Journal of Intelligent and Robotic Systems, 1998, 21 : 221 - 238
  • [47] Control of a Point Absorber Using Reinforcement Learning
    Anderlini, Enrico
    Forehand, David I. M.
    Stansell, Paul
    Xiao, Qing
    Abusara, Mohammad
    IEEE TRANSACTIONS ON SUSTAINABLE ENERGY, 2016, 7 (04) : 1681 - 1690
  • [48] Control of Qubit Dynamics Using Reinforcement Learning
    Koutromanos, Dimitris
    Stefanatos, Dionisis
    Paspalakis, Emmanuel
    INFORMATION, 2024, 15 (05)
  • [49] Derivative Feedback Control Using Reinforcement Learning
    Zaheer, Muhammad Hamad
    Yoon, Se Young
    Rizvi, Syed Ali Asad
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 7350 - 7355
  • [50] Control of A Polyol Process Using Reinforcement Learning
    Zhu, Wenbo
    Rendall, Ricardo
    Castillo, Ivan
    Wang, Zhenyu
    Chiang, Leo H.
    Hayot, Philippe
    Romagnoli, Jose A.
    IFAC PAPERSONLINE, 2021, 54 (03): : 498 - 503