Deep reinforcement learning based trajectory real-time planning for hypersonic gliding vehicles

被引:0
作者
Li, Jianfeng [1 ]
Song, Shenmin [1 ]
Shi, Xiaoping [2 ]
机构
[1] Harbin Inst Technol, Ctr Control Theory & Guidance Technol, 92 West Dazhi St, Harbin 150001, Peoples R China
[2] Harbin Inst Technol, Control & Simulat Ctr, Harbin, Peoples R China
关键词
Deep reinforcement learning; hypersonic gliding vehicle; trajectory planning; TD3; miss distance; PARTICLE SWARM OPTIMIZATION; ENTRY GUIDANCE; UAV;
D O I
10.1177/09544100241278023
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
To overcome the shortcomings of traditional NLP methods for trajectory planning problems, an intelligent trajectory real-time planning method is designed for hypersonic gliding vehicles (HGVs), which is composed of two stages: the agent training stage and the real-time trajectory generation stage. During the training stage, the HGV model is considered as an agent, and an environment containing flight information and relative information is constructed. Given the trajectory planning problem possessing continuous state-action space, the twin delayed deep deterministic policy gradient (TD3) is employed, based on which the HGV agent is trained in the environment. To match the real flight environment for HGVs, the process and terminal constraints are taken into consideration, such as the limit of dynamic pressure, overload, and the terminal miss distance, etc. The reward shaping technique is adopted to tackle the multiple constraints. The second stage is the real-time trajectory generation stage, during which a trajectory satisfying the multiple constraints is generated online by the TD3-based method. The simulation results verify the effectiveness of the proposed method.
引用
收藏
页码:1665 / 1682
页数:18
相关论文
共 33 条
[1]   Fairness-Based Energy-Efficient 3-D Path Planning of a Portable Access Point: A Deep Reinforcement Learning Approach [J].
Babu, Nithin ;
Donevski, Igor ;
Valcarce, Alvaro ;
Popovski, Petar ;
Nielsen, Jimmy Jessen ;
Papadias, Constantinos B. .
IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2022, 3 :1487-1500
[2]  
Bao C., 2022, PROC I MECH ENG G J, V237, P1855
[3]   Real-Time Reentry Trajectory Planning of Hypersonic Vehicles: A Two-Step Strategy Incorporating Fuzzy Multiobjective Transcription and Deep Neural Network [J].
Chai, Runqi ;
Tsourdos, Antonios ;
Savvaris, Al ;
Xia, Yuanqing ;
Chai, Senchun .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2020, 67 (08) :6904-6915
[4]   Multiconstrained Real-Time Entry Guidance Using Deep Neural Networks [J].
Cheng, Lin ;
Jiang, Fanghua ;
Wang, Zhenbo ;
Li, Junfeng .
IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2021, 57 (01) :325-340
[5]   High-Level Path Planning for an Autonomous Sailboat Robot Using Q-Learning [J].
da Silva Junior, Andouglas Goncalves ;
dos Santos, Davi Henrique ;
Fernandes de Negreiros, Alvaro Pinto ;
Boas de Souza Silva, Joao Moreno Vilas ;
Garcia Goncalves, Luiz Marcos .
SENSORS, 2020, 20 (06)
[6]  
Feher A., 2019, 16 INT C INF CONTR A
[7]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[8]   Autonomous gliding entry guidance with geographic constraints [J].
Guo Jie ;
Wu Xuzhong ;
Tang Shengjing .
CHINESE JOURNAL OF AERONAUTICS, 2015, 28 (05) :1343-1354
[9]   Aero-engine acceleration control using deep reinforcement learning with phase-based reward function [J].
Hu, Qian-Kun ;
Zhao, Yong-Ping .
PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART G-JOURNAL OF AEROSPACE ENGINEERING, 2022, 236 (09) :1878-1894
[10]   Optimal and Autonomous Control Using Reinforcement Learning: A Survey [J].
Kiumarsi, Bahare ;
Vamvoudakis, Kyriakos G. ;
Modares, Hamidreza ;
Lewis, Frank L. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (06) :2042-2062