Zero-sum Differential Games Guidance Law Accounting for Impact-Angle-Constrained Using Adaptive Dynamic Programming

被引:0
作者
Zhang, Xue [1 ]
Wang, Qi [2 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Aeronaut & Astronaut, Shanghai 200240, Peoples R China
[2] China Airborne Missile Acad, Luoyang 471009, Peoples R China
关键词
Computational intelligence guidance law; Impact angle constrained; Two-player zero-sum differential games; Policy iteration; SLIDING-MODE GUIDANCE; INTERCEPTION; ALGORITHM; TARGETS; SYSTEMS;
D O I
10.1007/s10846-024-02217-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To intercept a maneuvering target with a predetermined impact angle, a computational intelligence guidance law was proposed in this paper. Based on the theory of two-player zero-sum differential games, this problem is resolved efficiently by solving the Hamilton-Jacobi-Isaacs (HJI) equation. The Nash equilibrium solution of HJI equation can be solved with a policy iteration (PI) algorithm. Instead of using the offline PI algorithm, an online PI algorithm is introduced, in which the disturbance and control policies can be updated simultaneously. It can be proved that the online PI algorithm is a replacement for Newton's iterative algorithm, the convergence of which is ensured by Kantorovich's theorem. In the scenario of missiles intercepting targets, an adaptive critic structure based on a neural network (NN) is proposed to implement the online PI algorithm. Only one critic NN approximator is used in the PI algorithm to calculate a value function and the approximate Nash equilibrium solution. It is not necessary to acquire the exact internal dynamics information of nonlinear systems on the basis of online data sampling. The effectiveness of the computational intelligence guidance law is proven by simulation results.
引用
收藏
页数:12
相关论文
共 54 条
[1]   Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach [J].
Abu-Khalaf, M ;
Lewis, FL .
AUTOMATICA, 2005, 41 (05) :779-791
[2]   Neurodynamic programming and zero-sum games for constrained control systems [J].
Abu-Khalaf, Murad ;
Lewis, Frank L. ;
Huang, Jie .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (07) :1243-1252
[3]  
Arita S, 2013, AIAA GUIDANCE NAVIGA, P2013, DOI [10.2514/6.2013-4785, DOI 10.2514/6.2013-4785]
[4]  
Basar T, 1995, Systems & Control: Foundations and Applications Series, DOI [10.1109/TAC.1996.536519, DOI 10.1109/TAC.1996.536519]
[5]  
Bertsekas D P., 2005, Dynamic programming and optimal control, DOI [10.1057/jors.1996.103, DOI 10.1057/JORS.1996.103]
[6]  
Bertsekas D P., 1996, Neuro-dynamic programming, DOI [10.1007/0-306-48332-7333, DOI 10.1007/0-306-48332-7333]
[7]  
Bertsekas D. P., 2019, Reinforcement learning and optimal control
[8]   Optimal Impact Angle Control Guidance Law Based on Linearization About Collision Triangle [J].
Cho, Hangju ;
Ryoo, Chang-Kyung ;
Tsourdos, Antonios ;
White, Brian .
JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2014, 37 (03) :958-964
[9]   Indirect Impact-Angle-Control Against Stationary Targets Using Biased Pure Proportional Navigation [J].
Erer, Koray S. ;
Merttopcuoglu, Osman .
JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2012, 35 (02) :700-704
[10]   A game theoretic algorithm to compute local stabilizing solutions to HJBI equations in nonlinear H∞ control [J].
Feng, Yantao ;
Anderson, Brian D. O. ;
Rotkowitz, Michael .
AUTOMATICA, 2009, 45 (04) :881-888