Deep Reinforcement Learning for Trajectory Generation and Optimisation of UAVs

被引:1
作者
Akhtar, Mishma [1 ]
Maqsood, Adnan [1 ]
Verbeke, Mathias [2 ]
机构
[1] Natl Univ Sci & Technol, Sch Interdisciplinary Engn & Sci, Islamabad, Pakistan
[2] Katholieke Univ Leuven, Dept Comp Sci, M Grp, Flanders Make KU Leuven, Brugge, Belgium
来源
2023 10TH INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN AIR AND SPACE TECHNOLOGIES, RAST | 2023年
关键词
Reinforcement learning; Deep Deterministic Policy Gradient; Quadcopter; Control; Continual learning; ALGORITHM;
D O I
10.1109/RAST57548.2023.10197856
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
In recent years, the rapid advancements in Machine Learning have led to substantial research in control systems for autonomous aerial vehicles. Particularly, Reinforcement Learning (RL) has attracted a lot of attention for the design and development of such control algorithms. This paper examines the control issues of autonomous flight and how these are addressed using RL approaches. The objective is to investigate how RL algorithms like Deep Deterministic Policy Gradient may be applied particularly for control actions in an unmanned aerial vehicle (UAV). This learning paradigm acts as a mechanism that continuously generates policies for tasks such as attitude and position control, which converges into an optimized trajectory. As an outlook, the application of Continual Reinforcement Learning is proposed. This is a novel RL methodology that holds the potential to advance the control system of a UAV operating in dynamic, unknown environments with the ability to reapply learnt behavior and flexibly adapt to new situations.
引用
收藏
页数:6
相关论文
共 19 条
[1]  
[Anonymous], 2013, Foundations and Trends in Robotics, DOI DOI 10.1561/2300000021
[2]  
Bengio E, 2020, PR MACH LEARN RES, V119
[3]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[4]   Context-Dependent Pre-Trained Deep Neural Networks for Large-Vocabulary Speech Recognition [J].
Dahl, George E. ;
Yu, Dong ;
Deng, Li ;
Acero, Alex .
IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2012, 20 (01) :30-42
[5]  
Din A. F., 2020, AIAA SCITECH 2020 FO, P1849
[6]   Modified model free dynamic programming :an augmented approach for unmanned aerial vehicle [J].
Din, Adnan Fayyaz Ud ;
Akhtar, Suhail ;
Maqsood, Adnan ;
Habib, Muzaffar ;
Mir, Imran .
APPLIED INTELLIGENCE, 2023, 53 (03) :3048-3068
[7]   Path planning with modified A star algorithm for a mobile robot [J].
Duchon, Frantisek ;
Babinec, Andrej ;
Kajan, Martin ;
Beno, Peter ;
Florek, Martin ;
Fico, Tomas ;
Jurisica, Ladislav .
MODELLING OF MECHANICAL AND MECHATRONIC SYSTEMS, 2014, 96 :59-69
[8]  
Hausknecht M., 2015, AAAI FALL S SEQUENTI
[9]  
Hu YR, 2004, IEEE INT CONF ROBOT, P4350
[10]   Control of a Quadrotor With Reinforcement Learning [J].
Hwangbo, Jemin ;
Sa, Inkyu ;
Siegwart, Roland ;
Hutter, Marco .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2017, 2 (04) :2096-2103