Optimal Synchronization Control of Multiagent Systems With Input Saturation via Off-Policy Reinforcement Learning

被引:104
|
作者
Qin, Jiahu [1 ]
Li, Man [1 ]
Shi, Yang [2 ]
Ma, Qichao [1 ]
Zheng, Wei Xing [3 ]
机构
[1] Univ Sci & Technol China, Dept Automat, Hefei 230027, Anhui, Peoples R China
[2] Univ Victoria, Dept Mech Engn, Victoria, BC V8W 2Y2, Canada
[3] Western Sydney Univ, Sch Comp Engn & Math, Sydney, NSW 2751, Australia
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
Input saturation; multiagent systems; neural networks (NNs); off-policy reinforcement learning (RL); optimal synchronization control; LINEAR-SYSTEMS; NONLINEAR-SYSTEMS; NETWORKS; GAMES;
D O I
10.1109/TNNLS.2018.2832025
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we aim to investigate the optimal synchronization problem for a group of generic linear systems with input saturation. To seek the optimal controller, Hamilton Jacobi-Bellman (HJB) equations involving nonquadratic input energy terms in coupled forms are established. The solutions to these coupled HJB equations are further proven to be optimal and the induced controllers constitute interactive Nash equilibrium. Due to the difficulty to analytically solve HJB equations, especially in coupled forms, and the possible lack of model information of the systems, we apply the data-based off-policy reinforcement learning algorithm to learn the optimal control policies. A byproduct of this off-policy algorithm is shown that it is insensitive to probing noise that is exerted to the system to maintain persistence of excitation condition. In order to implement this off-policy algorithm, we employ actor and critic neural networks to approximate the controllers and the cost functions. Furthermore, the estimated control policies obtained by this presented implementation are proven to converge to the optimal ones under certain conditions. Finally, an illustrative example is provided to verify the effectiveness of the proposed algorithm.
引用
收藏
页码:85 / 96
页数:12
相关论文
共 50 条
  • [21] Moving Target Surrounding Control of Linear Multiagent Systems With Input Saturation
    Xu, Bowen
    Zhang, Hai-Tao
    Meng, Haofei
    Hu, Binbin
    Chen, Duxin
    Chen, Guanrong
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (03): : 1705 - 1715
  • [22] Data-Driven Robust Control of Discrete-Time Uncertain Linear Systems via Off-Policy Reinforcement Learning
    Yang, Yongliang
    Guo, Zhishan
    Xiong, Haoyi
    Ding, Da-Wei
    Yin, Yixin
    Wunsch, Donald C.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (12) : 3735 - 3747
  • [23] Data-Based Optimal Synchronization of Heterogeneous Multiagent Systems in Graphical Games via Reinforcement Learning
    Xiong, Chunping
    Ma, Qian
    Guo, Jian
    Lewis, Frank L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 15984 - 15992
  • [24] Off-Policy Interleaved Q-Learning: Optimal Control for Affine Nonlinear Discrete-Time Systems
    Li, Jinna
    Chai, Tianyou
    Lewis, Frank L.
    Ding, Zhengtao
    Jiang, Yi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) : 1308 - 1320
  • [25] Input-constrained optimal output synchronization of heterogeneous multiagent systems via observer-based model-free reinforcement learning
    Zhang, Tengfei
    Jia, Yingmin
    ASIAN JOURNAL OF CONTROL, 2024, 26 (01) : 98 - 113
  • [26] Robust optimal tracking control for multiplayer systems by off-policy Q-learning approach
    Li, Jinna
    Xiao, Zhenfei
    Li, Ping
    Cao, Jiangtao
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2021, 31 (01) : 87 - 106
  • [27] Off-Policy: Model-Free Optimal Synchronization Control for Complex Dynamical Networks
    Wang, Jianfeng
    Wang, Yan
    Ji, Zhicheng
    NEURAL PROCESSING LETTERS, 2022, 54 (04) : 2941 - 2958
  • [28] Off-Policy: Model-Free Optimal Synchronization Control for Complex Dynamical Networks
    Jianfeng Wang
    Yan Wang
    Zhicheng Ji
    Neural Processing Letters, 2022, 54 : 2941 - 2958
  • [29] Optimal robust online tracking control for space manipulator in task space using off-policy reinforcement learning
    Zhuang, Hongji
    Zhou, Hang
    Shen, Qiang
    Wu, Shufan
    Razoumny, Vladimir Yu.
    Razoumny, Yury N.
    AEROSPACE SCIENCE AND TECHNOLOGY, 2024, 153
  • [30] Reinforcement learning based optimal synchronization control for multi-agent systems with input constraints using vanishing viscosity method
    Zhang, Dianfeng
    Yao, Ying
    Wu, Zhaojing
    INFORMATION SCIENCES, 2023, 637