Cooperative Model-Based Reinforcement Learning for Approximate Optimal Tracking

被引:0
|
作者
Greene, Max L. [1 ]
Bell, Zachary, I [2 ]
Nivison, Scott A. [2 ]
How, Jonathan P. [3 ]
Dixon, Warren E. [1 ]
机构
[1] Univ Florida, Dept Mech & Aerosp Engn, Gainesville, FL 32611 USA
[2] Air Force Res Lab, Munit Directorate, Eglin AFB, FL USA
[3] MIT, Dept Aeronaut & Astronaut, Cambridge, MA 02139 USA
来源
2021 AMERICAN CONTROL CONFERENCE (ACC) | 2021年
关键词
SYSTEMS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper provides an approximate online adaptive solution to the infinite-horizon optimal tracking problem for a set of agents with homogeneous dynamics and common tracking objectives. Model-based reinforcement learning is implemented by simultaneously evaluating the Bellman error (BE) at the state of each agent and on nearby off-trajectory points, as needed, throughout the state space. Each agent will calculate and share their respective on and off-trajectory BE information with a centralized estimator, which computes updates for the approximate solution to the infinite-horizon optimal tracking problem and shares the estimate with the agents. In doing so, the computational burden associated with BE extrapolation is shared between the agents and a centralized updating resource. Edge computing is leveraged to share the computational load between the agents and a centralized resource. Uniformly ultimately bounded tracking of each agent's state to the desired state and convergence of the control policy to the neighborhood of the optimal policy is proven via a Lyapunov-like stability analysis.
引用
收藏
页码:1973 / 1978
页数:6
相关论文
共 50 条
  • [21] A machine learning-based surrogate model to approximate optimal building retrofit solutions
    Thrampoulidis, Emmanouil
    Mavromatidis, Georgios
    Lucchi, Aurelien
    Orehounig, Kristina
    APPLIED ENERGY, 2021, 281
  • [22] Reinforcement Learning Tracking Control for Robotic Manipulator With Kernel-Based Dynamic Model
    Hu, Yazhou
    Wang, Wenxue
    Liu, Hao
    Liu, Lianqing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (09) : 3570 - 3578
  • [23] Distributed adaptive cooperative optimal output regulation via integral reinforcement learning
    Lin, Liquan
    Huang, Jie
    AUTOMATICA, 2024, 170
  • [24] Model-based eServices for supporting cooperative practices in AEC
    Kubicki, S.
    Guerriero, A.
    Halin, G.
    EWORK AND EBUSINESS IN ARCHITECTURE, ENGINEERING AND CONSTRUCTIO N, 2009, : 171 - 178
  • [25] Impact of provoked stress on model-free and model-based reinforcement learning in individuals with alcohol use disorder
    Wyckmans, Florent
    Chatard, Armand
    Kornreich, Charles
    Gruson, Damien
    Jaafari, Nemat
    Noel, Xavier
    ADDICTIVE BEHAVIORS REPORTS, 2024, 20
  • [26] Impairment of arbitration between model-based and model-free reinforcement learning in obsessive-compulsive disorder
    Ruan, Zhongqiang
    Seger, Carol A.
    Yang, Qiong
    Kim, Dongjae
    Lee, Sang Wan
    Chen, Qi
    Peng, Ziwen
    FRONTIERS IN PSYCHIATRY, 2023, 14
  • [27] Optimal Control of Two-Dimensional Roesser Model: Solution Based on Reinforcement Learning
    Ye, Linwei
    Zhao, Zhonggai
    Liu, Fei
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (08) : 5424 - 5430
  • [28] Output Feedback Optimal Tracking Control Using Reinforcement Q-Learning
    Rizvi, Syed Ali Asad
    Lin, Zongli
    2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, : 3423 - 3428
  • [29] A Model-Based GNN for Learning Precoding
    Guo, Jia
    Yang, Chenyang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2024, 23 (07) : 6983 - 6999
  • [30] Prosocial learning: Model-based or model-free?
    Navidi, Parisa
    Saeedpour, Sepehr
    Ershadmanesh, Sara
    Hossein, Mostafa Miandari
    Bahrami, Bahador
    PLOS ONE, 2023, 18 (06):