Enhancing BVR Air Combat Agent Development With Attention-Driven Reinforcement Learning

被引:0
作者
Kuroswiski, Andre R. [1 ]
Wu, Annie S. [2 ]
Passaro, Angelo [3 ]
机构
[1] Inst Tecnol Aeronaut, BR-12228900 Sao Jose Dos Campos, SP, Brazil
[2] Univ Cent Florida, Dept Comp Sci, Orlando, FL 32816 USA
[3] Inst Estudos Avancados, BR-12228001 Sao Jose Dos Campos, SP, Brazil
关键词
Decision making; Autonomous agents; Visualization; Training; Reinforcement learning; Atmospheric modeling; Missiles; Uncertainty; Robustness; Aircraft; Adversarial learning; artificial intelligence; autonomous agents; beyond visual range air combat; multi-head attention; reinforcement learning; UCAV;
D O I
10.1109/ACCESS.2025.3561250
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This study explores the use of Reinforcement Learning (RL) to develop autonomous agents for Beyond Visual Range (BVR) air combat, addressing the challenges of dynamic and uncertain adversarial scenarios. We propose a novel approach that introduces a task-based layer, leveraging domain expertise to optimize decision-making and training efficiency. By integrating multi-head attention mechanisms into the policy model and employing an improved DQN algorithm, agents dynamically select context-aware tasks, enabling the learning of efficient emergent behaviors for variable engagement conditions. Evaluations in single- and multi-agent BVR scenarios against adversaries with diverse tactical characteristics demonstrate superior training efficiency and enhanced agent capabilities compared to leading RL algorithms commonly applied in similar domains, including PPO, DDPG, and SAC. A robustness study underscores the critical role of diverse enemy selection in the RL process, showing that adversaries with variable tactical behaviors are essential for developing robust agents. This work advances RL methodologies for autonomous BVR air combat and provides insights applicable to other problems with challenging adversarial scenarios.
引用
收藏
页码:70446 / 70463
页数:18
相关论文
共 48 条
[1]  
Armon T., 2020, Ph.D. dissertation
[2]  
Berner C., 2019, Dota 2 with Large Scale Deep Reinforcement Learning
[3]  
Bettini M, 2024, J MACH LEARN RES, V25
[4]   A comprehensive survey of multiagent reinforcement learning [J].
Busoniu, Lucian ;
Babuska, Robert ;
De Schutter, Bart .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (02) :156-172
[5]   Using agent-based modeling and a designed experiment to simulate and analyze a new air-to-air missile [J].
Connors, Casey D. ;
Miller, J. O. ;
Lunday, Brian J. .
JOURNAL OF DEFENSE MODELING AND SIMULATION-APPLICATIONS METHODOLOGY TECHNOLOGY-JDMS, 2016, 13 (03) :321-330
[6]  
Dantas J. P. A., 2021, Proc. Intell. Syst., V13074, P193
[7]  
DARPA, 2024, ACE Program Achieves World First for AI in Aerospace
[8]   Optimization of Unmanned Air Vehicle Tactical Formation in War Games [J].
De Lima Filho, Geraldo Mulato ;
Kuroswiski, Andre Rossi ;
Lobo Medeiros, Felipe Leonardo ;
Voskuijl, Mark ;
Monsuur, Herman ;
Passaro, Angelo .
IEEE ACCESS, 2022, 10 :21727-21741
[9]   A survey on multi-agent deep reinforcement learning: from the perspective of challenges and applications [J].
Du, Wei ;
Ding, Shifei .
ARTIFICIAL INTELLIGENCE REVIEW, 2021, 54 (05) :3215-3238
[10]  
Efron B., 1993, An introduction to the bootstrap