Model-free finite-horizon optimal control of discrete-time two-player zero-sum games

被引:3
|
作者
Wang, Wei [1 ,2 ,3 ]
Chen, Xin [2 ,3 ,4 ]
Du, Jianhua [1 ]
机构
[1] Zhongnan Univ Econ & Law, Sch Informat & Safety Engn, Wuhan, Peoples R China
[2] Hubei Key Lab Adv Control & Intelligent Automat C, Wuhan, Peoples R China
[3] Minist Educ, Engn Res Ctr Intelligent Technol Geoexplorat, Wuhan, Peoples R China
[4] China Univ Geosci, Sch Automat, Wuhan, Peoples R China
关键词
Q-function; finite horizon; optimal control; zero-sum game; H-INFINITY-CONTROL; DIFFERENTIAL-GAMES; SYSTEMS;
D O I
10.1080/00207721.2022.2111236
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Conventionally, as the system dynamics is known, the finite-horizon optimal control of zero-sum games relies on solving the time-varying Riccati equations. In this paper, with unknown system dynamics being considered, a Q-function-based finite-horizon control method is introduced to approximate the solutions of the time-varying Riccati equations. First, a time-varying Q-function explicitly dependent on the time-varying control and disturbance is defined. Then the defined time-varying Q-function is utilised to represent the time-varying control and disturbance which are equivalent to the solutions of the time-varying Riccati equations by relaxing the system dynamics. Finally, a model-free method is introduced to approximate the defined time-varying Q-function. Simulation studies are conducted to demonstrate the validity of the developed method.
引用
收藏
页码:167 / 179
页数:13
相关论文
共 50 条
  • [31] Adaptive Learning Based Output-Feedback Optimal Control of CT Two-Player Zero-Sum Games
    Zhao, Jun
    Lv, Yongfeng
    Zhao, Ziliang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (03) : 1437 - 1441
  • [32] Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control
    Al-Tamimi, Asma
    Lewis, Frank L.
    Abu-Khalaf, Murad
    AUTOMATICA, 2007, 43 (03) : 473 - 481
  • [33] Event-triggered optimal control scheme for discrete-time nonlinear zero-sum games
    Zhang X.
    Bo Y.-C.
    Cui L.-L.
    Zhang, Xin (zhangxin@upc.edu.cn), 2018, South China University of Technology (35): : 619 - 626
  • [34] Large Scale Learning of Agent Rationality in Two-Player Zero-Sum Games
    Ling, Chun Kai
    Fang, Fei
    Kolterl, J. Zico
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 6104 - 6111
  • [35] An LP Approach for Solving Two-Player Zero-Sum Repeated Bayesian Games
    Li, Lichun
    Langbort, Cedric
    Shamma, Jeff
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2019, 64 (09) : 3716 - 3731
  • [36] A METHOD TO SOLVE TWO-PLAYER ZERO-SUM MATRIX GAMES IN CHAOTIC ENVIRONMENT
    Khalifa, Hamiden Abd El-Wahed
    Kumar, Pavan
    INDEPENDENT JOURNAL OF MANAGEMENT & PRODUCTION, 2021, 12 (01): : 115 - 126
  • [37] Provably Efficient Policy Optimization for Two-Player Zero-Sum Markov Games
    Zhao, Yulai
    Tian, Yuandong
    Lee, Jason D.
    Du, Simon S.
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [38] Upper bounds and Cost Evaluation in Dynamic Two-player Zero-sum Games
    Leudo, Santiago J.
    Ferrante, Francesco
    Sanfelice, Ricardo G.
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 424 - 429
  • [39] Corruption-Robust Offline Two-Player Zero-Sum Markov Games
    Nika, Andi
    Mandal, Debmalya
    Singla, Adish
    Radanovic, Goran
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [40] Regularized Gradient Descent Ascent for Two-Player Zero-Sum Markov Games
    Zeng, Sihan
    Doan, Thinh
    Romberg, Justin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,