Reduced-dimensional reinforcement learning control using singular perturbation approximations

被引:34
|
作者
Mukherjee, Sayak [1 ]
Bai, He [2 ]
Chakrabortty, Aranya [1 ]
机构
[1] North Carolina State Univ, Dept Elect & Comp Engn, Raleigh, NC 27695 USA
[2] Oklahoma State Univ, Mech & Aerosp Engn Dept, Stillwater, OK 74078 USA
基金
美国国家科学基金会;
关键词
Reinforcement learning; Linear quadratic regulator; Singular perturbation; Model-free control; Model reduction; TIME LINEAR-SYSTEMS; ALGORITHM;
D O I
10.1016/j.automatica.2020.109451
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a set of model-free, reduced-dimensional reinforcement learning (RL) based optimal control designs for linear time-invariant singularly perturbed (SP) systems. We first present a state feedback and an output feedback based RL control design for a generic SP system with unknown state and input matrices. We take advantage of the underlying time-scale separation property of the plant to learn a linear quadratic regulator (LQR) for only its slow dynamics, thereby saving significant amount of learning time compared to the conventional full-dimensional RL controller. We analyze the sub-optimality of the designs using SP approximation theorems, and provide sufficient conditions for closed-loop stability. Thereafter, we extend both designs to clustered multi-agent consensus networks, where the SP property reflects through clustering. We develop both centralized and cluster-wise block-decentralized RL controllers for such networks, in reduced dimensions. We demonstrate the details of the implementation of these controllers using simulations of relevant numerical examples, and compare them with conventional RL designs to show the computational benefits of our approach. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Using reinforcement learning for engine control
    Schoknecht, R
    Riedmiller, M
    NINTH INTERNATIONAL CONFERENCE ON ARTIFICIAL NEURAL NETWORKS (ICANN99), VOLS 1 AND 2, 1999, (470): : 329 - 334
  • [32] REINFORCEMENT LEARNING CONTROL USING INTERCONNECTED LEARNING AUTOMATA
    WU, QH
    INTERNATIONAL JOURNAL OF CONTROL, 1995, 62 (01) : 1 - 16
  • [33] Controlling quantum wavepacket motion in reduced-dimensional spaces: reaction path analysis in optimal control of HCN isomerization
    Shah, SP
    Rice, SA
    FARADAY DISCUSSIONS, 1999, 113 : 319 - 331
  • [34] Reinforcement learning control using interconnected learning automata
    Loughborough Univ of Technology, Leicestershire, United Kingdom
    Int J Control, 1 (1-16):
  • [35] SUBOPTIMAL CONTROL OF POINT REACTORS BY USING SINGULAR PERTURBATION-THEORY
    ASATANI, K
    SHIOTANI, M
    HATTORI, Y
    JOURNAL OF NUCLEAR SCIENCE AND TECHNOLOGY, 1976, 13 (01) : 1 - 8
  • [36] Nonlinear PI control of chaotic systems using singular perturbation theory
    Jiang, W
    Jing, W
    Li, HY
    CHAOS SOLITONS & FRACTALS, 2005, 25 (05) : 1057 - 1068
  • [37] AN ITERATIVE APPROACH TO REDUCED-ORDER MODELING OF SYNCHRONOUS MACHINES USING SINGULAR PERTURBATION
    SINGH, NP
    SINGH, YP
    AHSON, SI
    PROCEEDINGS OF THE IEEE, 1986, 74 (06) : 892 - 893
  • [38] Representations and Exploration for Deep Reinforcement Learning using Singular Value Decomposition
    Chandak, Yash
    Thakoor, Shantanu
    Guo, Zhaohan Daniel
    Tang, Yunhao
    Munos, Remi
    Dabney, Will
    Borsa, Diana
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [39] Reduced-dimensional design optimization of stay vane and casing of reaction hydro turbines using global sensitivity analysis
    Shrestha, Ujjwal
    Choi, Young-Do
    Park, Jungwan
    Cho, Hyunkyoo
    JOURNAL OF MECHANICAL SCIENCE AND TECHNOLOGY, 2021, 35 (04) : 1487 - 1499
  • [40] Flight Control of a Multicopter using Reinforcement Learning
    D'Apolito, Francesco
    Sulzbachner, Christoph
    IFAC PAPERSONLINE, 2021, 54 (13): : 251 - 255