All-aspect attack guidance law for agile missiles based on deep reinforcement learning

被引:18
作者
Gong, Xiaopeng [1 ]
Chen, Wanchun [1 ]
Chen, Zhongyuan [1 ]
机构
[1] Beihang Univ, Sch Astronaut, Beijing 100191, Peoples R China
基金
中国博士后科学基金;
关键词
Deep reinforcement learning; Agile turn; Angle-of-attack guidance law; Hierarchical structure; All-aspect attack; High angle-of-attack; AUTOPILOT DESIGN;
D O I
10.1016/j.ast.2022.107677
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
This paper presents an all-aspect attack guidance law for agile missiles based on deep reinforcement learning (DRL), which can effectively cope with the aerodynamic uncertainty and strong nonlinearity in the high angle-of-attack (AOA) flight phase. First, to make the training environment more authentic, the full flight envelope of the missile is modeled and highly accurate aerodynamic data is obtained through Computational Fluid Dynamics (CFD) technique. Subsequently, the DRL algorithm is applied to generate an AOA guidance law for the agile turn phase. A hierarchical scheme that consists of a meta-controller for real-time decision making according to combat scenario and a sub-controller for generating guidance command is designed, which enables the guidance law to cover the whole process of the engagement and ensures the convergence of the training in the agile turn phase. Considering the current limitations of missile maneuverability, two agile turn guidance laws are developed to accommodate both limited and unlimited AOA scenarios. Moreover, the proposed guidance law has excellent generalization capability and ensures the implementation of static training and dynamic execution, which means that the missile can adapt to the realistic combat scenarios that have not been encountered during the training. Simulation results indicate that the DRL-based guidance law is nearly optimal and robust to disturbances. In addition, the proposed guidance law enables the missile to track time-varying desired turn angles to lock the maneuvering target in the rear hemisphere during the agile turn phase, providing advantageous initial conditions for the terminal guidance. Furthermore, the computational efficiency is high enough to satisfy the requirement on onboard application. (C) 2022 Elsevier Masson SAS. All rights reserved.
引用
收藏
页数:18
相关论文
共 54 条
  • [1] [Anonymous], 2016, MATEC WEB C
  • [2] Deep Reinforcement Learning A brief survey
    Arulkumaran, Kai
    Deisenroth, Marc Peter
    Brundage, Miles
    Bharath, Anil Anthony
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) : 26 - 38
  • [3] Review of advanced guidance and control algorithms for space/ aerospace vehicles
    Chai, Runqi
    Tsourdos, Antonios
    Al Savvaris
    Chai, Senchun
    Xia, Yuanqing
    Chen, C. L. Philip
    [J]. PROGRESS IN AEROSPACE SCIENCES, 2021, 122
  • [4] Chellappan V., 2012, IFAC P, V45, P145
  • [5] Choi YS, 2003, SICE 2003 ANNUAL CONFERENCE, VOLS 1-3, P1476
  • [6] Aerodynamic optimization of high-lift devices using a 2D-to-3D optimization method based on deep reinforcement learning and transfer learning
    Dai, Jiahua
    Liu, Peiqing
    Qu, Qiulin
    Li, Ling
    Niu, Tongzhi
    [J]. AEROSPACE SCIENCE AND TECHNOLOGY, 2022, 121
  • [7] Self-learned suppression of roll oscillations based on model-free reinforcement learning
    Dong, Yizhang
    Shi, Zhiwei
    Chen, Kun
    Yao, Zhangyi
    [J]. AEROSPACE SCIENCE AND TECHNOLOGY, 2021, 116
  • [8] Fujimoto S, 2018, PR MACH LEARN RES, V80
  • [9] Adaptive generalized ZEM-ZEV feedback guidance for planetary landing via a deep reinforcement learning approach
    Furfaro, Roberto
    Scorsoglio, Andrea
    Linares, Richard
    Massari, Mauro
    [J]. ACTA ASTRONAUTICA, 2020, 171 : 156 - 171
  • [10] Gaudet B, 2022, Arxiv, DOI arXiv:2109.03880