Input-Decoupled Q-Learning for Optimal Control

被引:0
|
作者
Minh Q. Phan
Seyed Mahdi B. Azad
机构
[1] Dartmouth College,Thayer School of Engineering
来源
The Journal of the Astronautical Sciences | 2020年 / 67卷
关键词
Optimal control; Reinforcement learning; Q-learning; Input-decoupled;
D O I
暂无
中图分类号
学科分类号
摘要
A design of optimal controllers based on a reinforcement learning method called Q-Learning is presented. Central to Q-Learning is the Q-function which is a function of the state and all input variables. This paper shows that decoupled-in-the-inputs Q-functions exist, and can be used to find the optimal controllers for each input individually. The method thus converts a multiple-variable optimization problem into much simpler single-variable optimization problems while achieving optimality. An explicit model of the system is not required to learn these decoupled Q-functions, but rather the method relies on the ability to probe the system and observe its state transition. Derived within the framework of modern control theory, the method is applicable to both linear and non-linear systems.
引用
收藏
页码:630 / 656
页数:26
相关论文
共 50 条
  • [1] Input-Decoupled Q-Learning for Optimal Control
    Phan, Minh Q.
    Azad, Seyed Mahdi B.
    JOURNAL OF THE ASTRONAUTICAL SCIENCES, 2020, 67 (02): : 630 - 656
  • [2] Discrete-Time Optimal Control Scheme Based on Q-Learning Algorithm
    Wei, Qinglai
    Liu, Derong
    Song, Ruizhuo
    2016 SEVENTH INTERNATIONAL CONFERENCE ON INTELLIGENT CONTROL AND INFORMATION PROCESSING (ICICIP), 2016, : 125 - 130
  • [3] Decoupled Visual Servoing With Fuzzy Q-Learning
    Shi, Haobin
    Li, Xuesi
    Hwang, Kao-Shing
    Pan, Wei
    Xu, Genjiu
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2018, 14 (01) : 241 - 252
  • [4] On-policy Q-learning for Adaptive Optimal Control
    Jha, Sumit Kumar
    Bhasin, Shubhendu
    2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning (ADPRL), 2014, : 301 - 306
  • [5] A Combined Policy Gradient and Q-learning Method for Data-driven Optimal Control Problems
    Lin, Mingduo
    Liu, Derong
    Zhao, Bo
    Dai, Qionghai
    Dong, Yi
    2019 9TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND TECHNOLOGY (ICIST2019), 2019, : 6 - 10
  • [6] Quantized measurements in Q-learning based model-free optimal control
    Tiistola, Sini
    Ritala, Risto
    Vilkko, Matti
    IFAC PAPERSONLINE, 2020, 53 (02): : 1640 - 1645
  • [7] Optimal control using adaptive resonance theory and Q-learning
    Kiumarsi, Bahare
    AlQaudi, Bakur
    Modares, Hamidreza
    Lewis, Frank L.
    Levine, Daniel S.
    NEUROCOMPUTING, 2019, 361 : 119 - 125
  • [8] Optimal Trajectory Output Tracking Control with a Q-learning Algorithm
    Vamvoudakis, Kyriakos G.
    2016 AMERICAN CONTROL CONFERENCE (ACC), 2016, : 5752 - 5757
  • [9] Inverse Q-Learning Using Input-Output Data
    Lian, Bosen
    Xue, Wenqian
    Lewis, Frank L.
    Davoudi, Ali
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (02) : 728 - 738
  • [10] Q-LEARNING
    WATKINS, CJCH
    DAYAN, P
    MACHINE LEARNING, 1992, 8 (3-4) : 279 - 292