Reinforcement Learning based Neuro-control Systems for an Unmanned Helicopter

被引:0
作者
Lee, Dong Jin [1 ]
Bang, Hyochoong [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Div Aerosp Engn, Sch Mech Aerosp & Syst Engn, Taejon 305701, South Korea
来源
INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2010) | 2010年
关键词
Reinforcement Learning; Actor-Critic Learning; RUAV; Unmanned Helicopter;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper concerns with the autonomous flight control system of an unmanned helicopter, which is combined with reinforcement learning based neuro-controller. We assume that PID (proportional-integral-derivative) type, linear feedback controller is predesigned and it can stabilize the system with limited performance. The conservative control behavior is improved by the synthesis of the poor feedback controller and the neuro-controller. Actor-critic learning architecture is adopted as a learning agent. Actor network consists of feed-forward neural network and critic network is approximated with a tabular function approximator. The Q-value based critic network is trained via SARSA algorithm which is a variant of reinforcement learning. Several demonstrations are performed with a simple first-order system. Furthermore, the proposed neuro-control system is applied to an unmanned helicopter known as a highly nonlinear and complex system and the simulation results are presented.
引用
收藏
页码:2537 / 2540
页数:4
相关论文
共 9 条
[1]  
Albus J. S., 1975, Journal of Dynamic Systems, Measurement and Control
[2]  
Doyle JohnC., 1991, FEEDBACK CONTROL THE
[3]  
Lee D. J., 2008, INT C CONTR AUT SYST
[4]  
Lee D. J., 2008, GUID NAV CONTR C EXH
[5]  
LEE DJ, 2006, FLOW DYN C TOH U JOI
[6]  
LEE DJ, 2007, INT FORUM ROTORCRAFT
[7]  
LEE IJ, 2007, INFOTECH
[8]  
Sutton R.S., 2017, Introduction to reinforcement learning
[9]  
TU J, 2001, THESIS COLORADO STAT