Hierarchical Reinforcement Learning Framework in Geographic Coordination for Air Combat Tactical Pursuit

被引:6
作者
Chen, Ruihai [1 ]
Li, Hao [2 ]
Yan, Guanwei [2 ]
Peng, Haojie [1 ]
Zhang, Qian [3 ]
机构
[1] Northwestern Polytech Univ, Sch Aeronaut, Xian 710072, Peoples R China
[2] Chengdu Aircraft Design & Res Inst, Chengdu 610041, Peoples R China
[3] Northwestern Polytech Univ, Sch Aerosp, Xian 710072, Peoples R China
基金
中国博士后科学基金;
关键词
hierarchical reinforcement learning; meta-learning; reward design; decision;
D O I
10.3390/e25101409
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
This paper proposes an air combat training framework based on hierarchical reinforcement learning to address the problem of non-convergence in training due to the curse of dimensionality caused by the large state space during air combat tactical pursuit. Using hierarchical reinforcement learning, three-dimensional problems can be transformed into two-dimensional problems, improving training performance compared to other baselines. To further improve the overall learning performance, a meta-learning-based algorithm is established, and the corresponding reward function is designed to further improve the performance of the agent in the air combat tactical chase scenario. The results show that the proposed framework can achieve better performance than the baseline approach.
引用
收藏
页数:21
相关论文
共 49 条
[1]  
Barto AG, 2003, DISCRETE EVENT DYN S, V13, P41, DOI [10.1023/A:1022140919877, 10.1023/A:1025696116075]
[2]  
Byrnes M.W., 2014, AIR SPACE POWER J, V28, P48
[3]   Autonomous Maneuver Decision of UCAV Air Combat Based on Double Deep Q Network Algorithm and Stochastic Game Theory [J].
Cao, Yuan ;
Kou, Ying-Xin ;
Li, Zhan-Wu ;
Xu, An .
INTERNATIONAL JOURNAL OF AEROSPACE ENGINEERING, 2023, 2023
[4]  
Chandak Y, 2019, PR MACH LEARN RES, V97
[5]   Proximal policy optimization guidance algorithm for intercepting near-space maneuvering targets [J].
Chen, Wenxue ;
Gao, Changsheng ;
Jing, Wuxing .
AEROSPACE SCIENCE AND TECHNOLOGY, 2023, 132
[6]  
Chen YY, 2020, I C CONT AUTOMAT ROB, P817, DOI [10.1109/ICARCV50220.2020.9305467, 10.1109/icarcv50220.2020.9305467]
[7]  
Comanici G., 2010, AAMAS, V1, P709
[8]  
Duan Y, 2016, PR MACH LEARN RES, V48
[9]   Intelligent problem-solving as integrated hierarchical reinforcement learning [J].
Eppe, Manfred ;
Gumbsch, Christian ;
Kerzel, Matthias ;
Nguyen, Phuong D. H. ;
Butz, Martin, V ;
Wermter, Stefan .
NATURE MACHINE INTELLIGENCE, 2022, 4 (01) :11-20
[10]  
Ernest N., 2016, J Def Manage, V6, P2167, DOI [10.4172/2167-0374.1000144, DOI 10.4172/2167-0374.1000144]