Adaptive Learning Based Output-Feedback Optimal Control of CT Two-Player Zero-Sum Games

被引:19
|
作者
Zhao, Jun [1 ]
Lv, Yongfeng [2 ]
Zhao, Ziliang [3 ]
机构
[1] Shandong Univ Sci & Technol, Coll Mech & Elect Engn, Qingdao 266590, Peoples R China
[2] Taiyuan Univ Technol, Coll Elect & Power Engn, Taiyuan 030024, Peoples R China
[3] Shandong Univ Sci & Technol, Coll Transportat, Qingdao 266590, Peoples R China
基金
中国国家自然科学基金;
关键词
Games; Optimal control; Adaptive learning; Game theory; Cost function; Observers; Estimation error; Output-feedback optimal control; adaptive learning; zero-sum games; SYSTEMS;
D O I
10.1109/TCSII.2021.3112050
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Although optimal control with full state-feedback has been well studied, online solving output-feedback optimal control problem is difficult, in particular for learning online Nash equilibrium solution of the continuous-time (CT) two-player zero-sum differential games. For this purpose, we propose an adaptive learning algorithm to address this trick problem. A modified game algebraic Riccati equation (MGARE) is derived by tailoring its state-feedback control counterpart. An adaptive online learning method is proposed to approximate the solution to the MGARE through online data, where two operations (i.e., vectorization and Kronecker's product) can be adopted to reconstruct the MGARE. Only system output information is needed to implement developed learning algorithm. Simulation results are carried out to exemplify the proposed control and learning method.
引用
收藏
页码:1437 / 1441
页数:5
相关论文
共 50 条
  • [1] Policy Similarity Measure for Two-Player Zero-Sum Games
    Tang, Hongsong
    Xiang, Liuyu
    He, Zhaofeng
    APPLIED SCIENCES-BASEL, 2025, 15 (05):
  • [2] Sufficient Conditions for Optimality and Asymptotic Stability in Two-Player Zero-Sum Hybrid Games
    Leudo, Santiago J.
    Sanfelice, Ricardo G.
    HSCC 2022: PROCEEDINGS OF THE 25TH ACM INTERNATIONAL CONFERENCE ON HYBRID SYSTEMS: COMPUTATION AND CONTROL (PART OF CPS-IOT WEEK 2022), 2022,
  • [3] A Generalized Minimax Q-Learning Algorithm for Two-Player Zero-Sum Stochastic Games
    Diddigi, Raghuram Bharadwaj
    Kamanchi, Chandramouli
    Bhatnagar, Shalabh
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (09) : 4816 - 4823
  • [4] Adaptive Optimal Output-Feedback Consensus Tracking Control of Nonlinear Multiagent Systems Using Two-Player Stackelberg Game
    Yan, Lei
    Liu, Junhe
    Lai, Guanyu
    Chen, C. L. Philip
    Wu, Zongze
    Liu, Zhi
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (09): : 5377 - 5387
  • [5] Online Minimax Q Network Learning for Two-Player Zero-Sum Markov Games
    Zhu, Yuanheng
    Zhao, Dongbin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (03) : 1228 - 1241
  • [6] Pure strategy equilibria in symmetric two-player zero-sum games
    Peter Duersch
    Jörg Oechssler
    Burkhard C. Schipper
    International Journal of Game Theory, 2012, 41 : 553 - 564
  • [7] Pure strategy equilibria in symmetric two-player zero-sum games
    Duersch, Peter
    Oechssler, Joerg
    Schipper, Burkhard C.
    INTERNATIONAL JOURNAL OF GAME THEORY, 2012, 41 (03) : 553 - 564
  • [8] Finite Horizon Stochastic Optimal Control of Nonlinear Two-Player Zero-Sum Games under Communication Constraint
    Xu, Hao
    Jagannathan, S.
    PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2014, : 239 - 244
  • [9] Model-free finite-horizon optimal control of discrete-time two-player zero-sum games
    Wang, Wei
    Chen, Xin
    Du, Jianhua
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2023, 54 (01) : 167 - 179
  • [10] Equilibrium payoffs in repeated two-player zero-sum games of finite automata
    Baskov, O. V.
    INTERNATIONAL JOURNAL OF GAME THEORY, 2019, 48 (02) : 423 - 431