Horizontal trajectory control of stratospheric airships in wind field using Q-learning algorithm

被引:35
作者
Yang, Xiaowei [1 ]
Yang, Xixiang [1 ]
Deng, Xiaolong [1 ]
机构
[1] Natl Univ Def Technol, Coll Aerosp Sci & Engn, Changsha 410073, Peoples R China
关键词
Stratospheric airships; Horizontal trajectory control; Q-learning algorithm; Wind field; CMAC neural network; NEURAL-NETWORK; REINFORCEMENT; ROBOT;
D O I
10.1016/j.ast.2020.106100
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
This paper proposes an adaptive horizontal trajectory control method for stratospheric airships in uncertain wind field using Q-learning algorithm. Firstly, horizontal trajectory control of the airships is decomposed into the target tracking, and the observation model of airships is constructed. Then, the Markov decision process (MDP) model of airships is established, in which the action strategy is determined by the wind direction, and a cerebellar model articulation controller (CMAC) neural network is designed to optimize the action strategy for each state. Finally, numerical simulations demonstrate that the proposed control method performs well stability and intelligent decision-making ability in the process of horizontal trajectory control for stratospheric airships. (C) 2020 Elsevier Masson SAS. All rights reserved.
引用
收藏
页数:8
相关论文
共 30 条
[1]   An Indoor and Outdoor Positioning Using a Hybrid of Support Vector Machine and Deep Neural Network Algorithms [J].
Adege, Abebe Belay ;
Lin, Hsin-Piao ;
Tarekegn, Getaneh Berie ;
Munaye, Yirga Yayeh ;
Yen, Lei .
JOURNAL OF SENSORS, 2018, 2018
[2]  
[Anonymous], 2007, AIAA BALL SYST C
[3]   Adaptive path following control of a stratospheric airship with full-state constraint and actuator saturation [J].
Chen, Tian ;
Zhu, Ming ;
Zheng, Zewei .
AEROSPACE SCIENCE AND TECHNOLOGY, 2019, 95
[4]   Gaussian process dynamic programming [J].
Deisenroth, Marc Peter ;
Rasmussen, Carl Edward ;
Peters, Jan .
NEUROCOMPUTING, 2009, 72 (7-9) :1508-1524
[5]  
Duraisamy G., 2019, AIAA SCIT 2019 FOR, P1564
[6]   Autolanding control system design with deep learning based fault estimation [J].
Eroglu, Batuhan ;
Sahin, M. Cagatay ;
Ure, Nazim Kemal .
AEROSPACE SCIENCE AND TECHNOLOGY, 2020, 102
[7]   Receding horizon guidance of a small unmanned aerial vehicle for planar reference path following [J].
Hamada, Yoshiro ;
Tsukamoto, Taro ;
Ishimoto, Shinji .
AEROSPACE SCIENCE AND TECHNOLOGY, 2018, 77 :129-137
[8]   Control of a Quadrotor With Reinforcement Learning [J].
Hwangbo, Jemin ;
Sa, Inkyu ;
Siegwart, Roland ;
Hutter, Marco .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2017, 2 (04) :2096-2103
[9]   Gaussian processes and reinforcement learning for identification and control of an autonomous blimp [J].
Ko, Jonathan ;
Klein, Daniel J. ;
Fox, Dieter ;
Haehnel, Dirk .
PROCEEDINGS OF THE 2007 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-10, 2007, :742-747
[10]   Use of Q-learning approaches for practical medium access control in wireless sensor networks [J].
Kosunalp, Selahattin ;
Chu, Yi ;
Mitchell, Paul D. ;
Grace, David ;
Clarke, Tim .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2016, 55 :146-154