On Model-free Reinforcement Learning for Switched Linear Systems: A Subspace Clustering Approach

被引:0
|
作者
Li, Hao [1 ]
Chen, Hua [1 ]
Zhang, Wei [1 ]
机构
[1] Ohio State Univ, Dept Elect & Comp Engn, Columbus, OH 43210 USA
来源
2018 56TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON) | 2018年
基金
美国国家科学基金会;
关键词
DISCRETE-TIME; TRACKING CONTROL; ALGORITHM;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we study optimal control of switched linear systems using reinforcement learning. Instead of directly applying existing model-free reinforcement learning algorithms, we propose a Q-learning-based algorithm designed specifically for discrete time switched linear systems. Inspired by the analytical results from optimal control literature, the Q function in our algorithm is approximated by a point-wise minimum form of a finite number of quadratic functions. An associated update scheme based on subspace clustering for such an approximation is also developed which preserves the desired structure during the training process. Numerical examples for both low-dimensional and high-dimensional switched linear systems are provided to demonstrate the performance of our algorithm.
引用
收藏
页码:123 / 130
页数:8
相关论文
共 50 条
  • [1] Nearly data-based optimal control for linear discrete model-free systems with delays via reinforcement learning
    Zhang, Jilie
    Zhang, Huaguang
    Wang, Binrui
    Cai, Tiaoyang
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2016, 47 (07) : 1563 - 1573
  • [2] Model-free aperiodic tracking for discrete-time systems using hierarchical reinforcement learning
    Tian, Yingqiang
    Wan, Haiying
    Karimi, Hamid Reza
    Luan, Xiaoli
    Liu, Fei
    NEUROCOMPUTING, 2024, 609
  • [3] Optimal model-free output synchronization of heterogeneous systems using off-policy reinforcement learning
    Modares, Hamidreza
    Nageshrao, Subramanya P.
    Lopes, Gabriel A. Delgado
    Babuska, Robert
    Lewis, Frank L.
    AUTOMATICA, 2016, 71 : 334 - 341
  • [4] Model-free attitude synchronization for multiple heterogeneous quadrotors via reinforcement learning
    Zhao, Wanbing
    Liu, Hao
    Wang, Bohui
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2021, 36 (06) : 2528 - 2547
  • [5] A Distributed Model-Free Ride-Sharing Approach for Joint Matching, Pricing, and Dispatching Using Deep Reinforcement Learning
    Haliem, Marina
    Mani, Ganapathy
    Aggarwal, Vaneet
    Bhargava, Bharat
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (12) : 7931 - 7942
  • [6] Model-Free Reinforcement Learning for Nonlinear Zero-Sum Games with Simultaneous Explorations
    Zhang, Qichao
    Zhao, Donghin
    Zhu, Yuanheng
    Chen, Xi
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 4533 - 4538
  • [7] Online Model-Free Reinforcement Learning for Output Feedback Tracking Control of a Class of Discrete-Time Systems With Input Saturation
    Al-Mahasneh, Ahmad Jobran
    Anavatti, Sreenatha G.
    Garratt, Matthew A.
    IEEE ACCESS, 2022, 10 : 104966 - 104979
  • [8] Model-Free Event-Triggered Optimal Consensus Control of Multiple Euler-Lagrange Systems via Reinforcement Learning
    Wang, Saiwei
    Jin, Xin
    Mao, Shuai
    Vasilakos, Athanasios V.
    Tang, Yang
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2021, 8 (01): : 246 - 258
  • [9] A novel optimal bipartite consensus control scheme for unknown multi-agent systems via model-free reinforcement learning
    Peng, Zhinan
    Hu, Jiangping
    Shi, Kaibo
    Luo, Rui
    Huang, Rui
    Ghosh, Bijoy Kumar
    Huang, Jiuke
    APPLIED MATHEMATICS AND COMPUTATION, 2020, 369
  • [10] Model-free fault-tolerant control approach for uncertain state-constrained linear systems with actuator faults
    Xie, Chun-Hua
    Yang, Guang-Hong
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2017, 31 (02) : 223 - 239