Asynchronous iterative Q-learning based tracking control for nonlinear discrete-time multi-agent systems

被引:0
作者
Shen, Ziwen [1 ]
Dong, Tao [1 ]
Huang, Tingwen [2 ]
机构
[1] Southwest Univ, Coll Elect & Informat Engn, Chongqing 400715, Peoples R China
[2] Shenzhen Univ Adv Technol, Fac Comp Sci & Control Engn, Shenzhen 518055, Peoples R China
关键词
Multi-agent; Discrete-time; Asynchronous iterative Q-learning; Tracking control; OPTIMAL CONSENSUS CONTROL;
D O I
10.1016/j.neunet.2024.106667
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper addresses the tracking control problem of nonlinear discrete-time multi-agent systems (MASs). First, a local neighborhood error system (LNES) is constructed. Then, a novel tracking algorithm based on asynchronous iterative Q-learning (AIQL) is developed, which can transform the tracking problem into the optimal regulation of LNES. The AIQL-based algorithm has two Q values Q(i)(A) and Q(i)(B) for each agent i , where Q(i)(A) is used for improving the control policy and Q(i)(B) is used for evaluating the value of the control policy. Moreover, the convergence of LNES is given. It is shown that the LNES converges to 0 and the tracking problem is solved. A neural network-based actor-critic framework is used to implement AIQL. The critic network of AIQL is composed of two neural networks, which are used for approximating Q(i)(A) and Q(i)(B) respectively. Finally, simulation results are given to verify the performance of the developed algorithm. It is shown that the AIQLbased tracking algorithm has a lower cost value and faster convergence speed than the IQL-based tracking algorithm.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Adaptive Iterative Learning Control of Switched Nonlinear Discrete-Time Systems With Unmodeled Dynamics
    Geng, Yan
    Ruan, Xiaoe
    Xu, Jinhu
    IEEE ACCESS, 2019, 7 : 118370 - 118380
  • [32] On Stability of Consensus Control of Discrete-Time Multi-Agent Systems by Multiple Pinning Agents
    Xu, Dongwu
    Ushio, Toshimitsu
    IEEE CONTROL SYSTEMS LETTERS, 2019, 3 (04): : 1038 - 1043
  • [33] Study on Statistics Based Q-learning Algorithm for Multi-Agent System
    Xie Ya
    Huang Zhonghua
    2013 FOURTH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS DESIGN AND ENGINEERING APPLICATIONS, 2013, : 595 - 600
  • [34] Adaptive iterative learning control for consensus of multi-agent systems
    Li, Jinsha
    Li, Junmin
    IET CONTROL THEORY AND APPLICATIONS, 2013, 7 (01) : 136 - 142
  • [35] Consensusability of discrete-time linear multi-agent systems with multiple inputs
    Feng, Tao
    Zhang, Jilie
    Zhang, Huaguang
    NEUROCOMPUTING, 2020, 383 : 183 - 193
  • [36] Robust Discrete-Time Iterative Learning Control for Nonlinear Systems With Varying Initial State Shifts
    Meng, Deyuan
    Jia, Yingmin
    Du, Junping
    Yuan, Shiying
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2009, 54 (11) : 2626 - 2631
  • [37] Distributed Output Optimization for Discrete-time Linear Multi-agent Systems
    Tang, Yutao
    Zhu, Hao
    Lv, Xiaoyong
    PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 5665 - 5669
  • [38] Robust consensus for uncertain multi-agent systems with discrete-time dynamics
    Han, Dongkun
    Chesi, Graziano
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2014, 24 (13) : 1858 - 1872
  • [39] Finite-Time Tracking Consensus Control for A Class of Nonlinear Multi-Agent Systems
    Li, Zhenxing
    Chen, Xiangyong
    Wen, Yumei
    Qiu, Jianlong
    IECON 2017 - 43RD ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2017, : 5803 - 5808
  • [40] Multi-agent Q-Learning control of spacecraft formation flying reconfiguration trajectories
    Kankashvar, Mohammadrasoul
    Bolandi, Hossein
    Mozayani, Nasser
    ADVANCES IN SPACE RESEARCH, 2023, 71 (03) : 1627 - 1643