A deep reinforcement learning control approach for high-performance aircraft

被引:12
作者
De Marco, Agostino [1 ]
D'Onza, Paolo Maria [1 ]
Manfredi, Sabato [2 ]
机构
[1] Univ Napoli Federico II, Dept Ind Engn DII, Via Claudio 21, I-80125 Naples, Italy
[2] Univ Napoli Federico II, Dept Elect Engn & Informat Technol DIETI, Via Claudio 21, I-80125 Naples, Italy
关键词
Deep reinforcement learning; Flight dynamics; UCAV; Aeroplane controllability; Nonlinear control; NONLINEAR-SYSTEMS; FLIGHT;
D O I
10.1007/s11071-023-08725-y
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
This research introduces a flight controller for a high-performance aircraft, able to follow randomly generated sequences of waypoints, at varying altitudes, in various types of scenarios. The study assumes a publicly available six-degree-of-freedom (6-DoF) rigid aeroplane flight dynamics model of a military fighter jet. Consolidated results in artificial intelligence and deep reinforcement learning (DRL) research are used to demonstrate the capability to make certain manoeuvres AI-based fully automatic for a high-fidelity nonlinear model of a fixed-wing aircraft. This work investigates the use of a deep deterministic policy gradient (DDPG) controller agent, based on the successful applications of the same approach to other domains. In the particular application to flight control presented here, the effort has been focused on the design of a suitable reward function used to train the agent to achieve some given navigation tasks. The trained controller is successful on highly coupled manoeuvres, including rapid sequences of turns, at both low and high flight Mach numbers, in simulations reproducing a prey-chaser dogfight scenario. Robustness to sensor noise, atmospheric disturbances, different initial flight conditions and varying reference signal shapes is also demonstrated.
引用
收藏
页码:17037 / 17077
页数:41
相关论文
共 39 条
[11]   Improving the Precision and Speed of Euler Angles Computation from Low-Cost Rotation Sensor Data [J].
Janota, Ales ;
Simak, Vojtech ;
Nemec, Dusan ;
Hrbcek, Jozef .
SENSORS, 2015, 15 (03) :7016-7039
[12]   Reinforcement Learning for UAV Attitude Control [J].
Koch, William ;
Mancuso, Renato ;
West, Richard ;
Bestavros, Azer .
ACM TRANSACTIONS ON CYBER-PHYSICAL SYSTEMS, 2019, 3 (02)
[13]   Formation flight of unmanned aerial vehicles using track guidance [J].
Lee, Dongwoo ;
Kim, Seungkeun ;
Suk, Jinyoung .
AEROSPACE SCIENCE AND TECHNOLOGY, 2018, 76 :412-420
[14]   Autonomous maneuver decision-making for a UCAV in short-range aerial combat based on an MS-DDQN algorithm [J].
Li, Yong-feng ;
Shi, Jing-ping ;
Jiang, Wei ;
Zhang, Wei-guo ;
Lyu, Yong-xi .
DEFENCE TECHNOLOGY, 2022, 18 (09) :1697-1714
[15]  
Lillicrap T., 2015, CONTINUOUS CONTROL WITH DEEP REINFORCEMENT LEARNING
[16]   A Multi-UCAV Cooperative Decision-Making Method Based on an MAPPO Algorithm for Beyond-Visual-Range Air Combat [J].
Liu, Xiaoxiong ;
Yin, Yi ;
Su, Yuzhan ;
Ming, Ruichen .
AEROSPACE, 2022, 9 (10)
[17]   A new adaptive neural control scheme for hypersonic vehicle with actuators multiple constraints [J].
Luo, Changxin ;
Lei, Humin ;
Li, Jiong ;
Zhou, Chijun .
NONLINEAR DYNAMICS, 2020, 100 (04) :3529-3553
[18]   Air-Combat Strategy Using Approximate Dynamic Programming [J].
McGrew, James S. ;
How, Jonathan P. ;
Williams, Brian ;
Roy, Nicholas .
JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2010, 33 (05) :1641-1654
[19]   Variable gain gradient descent-based reinforcement learning for robust optimal tracking control of uncertain nonlinear system with input constraints [J].
Mishra, Amardeep ;
Ghosh, Satadal .
NONLINEAR DYNAMICS, 2022, 107 (03) :2195-2214
[20]  
Mnih V., 2013, ARXIV, DOI DOI 10.48550/ARXIV.1312.5602