A deep reinforcement learning control approach for high-performance aircraft

被引:12
作者
De Marco, Agostino [1 ]
D'Onza, Paolo Maria [1 ]
Manfredi, Sabato [2 ]
机构
[1] Univ Napoli Federico II, Dept Ind Engn DII, Via Claudio 21, I-80125 Naples, Italy
[2] Univ Napoli Federico II, Dept Elect Engn & Informat Technol DIETI, Via Claudio 21, I-80125 Naples, Italy
关键词
Deep reinforcement learning; Flight dynamics; UCAV; Aeroplane controllability; Nonlinear control; NONLINEAR-SYSTEMS; FLIGHT;
D O I
10.1007/s11071-023-08725-y
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
This research introduces a flight controller for a high-performance aircraft, able to follow randomly generated sequences of waypoints, at varying altitudes, in various types of scenarios. The study assumes a publicly available six-degree-of-freedom (6-DoF) rigid aeroplane flight dynamics model of a military fighter jet. Consolidated results in artificial intelligence and deep reinforcement learning (DRL) research are used to demonstrate the capability to make certain manoeuvres AI-based fully automatic for a high-fidelity nonlinear model of a fixed-wing aircraft. This work investigates the use of a deep deterministic policy gradient (DDPG) controller agent, based on the successful applications of the same approach to other domains. In the particular application to flight control presented here, the effort has been focused on the design of a suitable reward function used to train the agent to achieve some given navigation tasks. The trained controller is successful on highly coupled manoeuvres, including rapid sequences of turns, at both low and high flight Mach numbers, in simulations reproducing a prey-chaser dogfight scenario. Robustness to sensor noise, atmospheric disturbances, different initial flight conditions and varying reference signal shapes is also demonstrated.
引用
收藏
页码:17037 / 17077
页数:41
相关论文
共 39 条
[1]  
[Anonymous], 2003, Aircraft control and simulation
[2]  
[Anonymous], 1957, The American Mathematical Monthly, DOI [10.2307/2309088, DOI 10.2307/2309088]
[3]  
Berndt J., 2009, AIAA MOD SIM TECHN C, P5699
[4]  
Bohn E, 2019, INT CONF UNMAN AIRCR, P523, DOI [10.1109/icuas.2019.8798254, 10.1109/ICUAS.2019.8798254]
[5]  
Brunton SL., 2022, Data-Driven Science and Engineering, DOI [10.1017/9781009089517, DOI 10.1017/9781009089517]
[6]  
Cambone S.A., 2005, UNMANNED AIRCRAFT SY, V8, P4
[7]   Soft Actor-Critic Deep Reinforcement Learning for Fault-Tolerant Flight Control [J].
Dally, K. ;
van Kampen, E. .
AIAA SCITECH 2022 FORUM, 2022,
[8]   Reinforcement learning in feedback control Challenges and benchmarks from technical process control [J].
Hafner, Roland ;
Riedmiller, Martin .
MACHINE LEARNING, 2011, 84 (1-2) :137-169
[9]   Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat [J].
Hu, Dongyuan ;
Yang, Rennong ;
Zuo, Jialiang ;
Zhang, Ze ;
Wu, Jun ;
Wang, Ying .
IEEE ACCESS, 2021, 9 :32282-32297
[10]   Observer-based adaptive fuzzy tracking control of MIMO switched nonlinear systems preceded by unknown backlash-like hysteresis [J].
Huo, Xin ;
Ma, Li ;
Zhao, Xudong ;
Niu, Ben ;
Zong, Guangdeng .
INFORMATION SCIENCES, 2019, 490 :369-386