A Novel Reinforcement Learning-Based Robust Control Strategy for a Quadrotor

被引:23
作者
Hua, Hean [1 ,2 ]
Fang, Yongchun [1 ,2 ]
机构
[1] Nankai Univ, Inst Robot & Automat Informat Syst, Coll Artificial Intelligence, Tianjin 300353, Peoples R China
[2] Nankai Univ, Tianjin Key Lab Intelligent Robot, Tianjin 300353, Peoples R China
基金
中国国家自然科学基金;
关键词
Quadrotors; reinforcement learning (RL) control; robust integral of the signum of the error (RISE); RISE-guided learning; real-world applications; TRAJECTORY TRACKING CONTROL; ATTITUDE-CONTROL; LEVEL CONTROL; AERIAL; SAFE;
D O I
10.1109/TIE.2022.3165288
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this article, a novel reinforcement learning (RL)-based robust control approach is proposed for quadrotors, which guarantees efficient learning and satisfactory tracking performance by simultaneously evaluating the RL and the baseline method in training. Different from existing works, the key novelty is to design a practice-reliable RL control framework for quadrotors in a two-part cooperative manner. In the first part, based on the hierarchical property, a new robust integral of the signum of the error (RISE) design is proposed to ensure asymptotic convergence, which includes the nonlinear and the disturbance rejection terms. In the second part, a one-actor-dual-critic (OADC) learning framework is proposed, where the designed switching logic in the first part works as a benchmark to guide the learning. Specifically, the two critics independently evaluate the RL policy and the switching logic simultaneously, which are utilized for policy update, only when both are positive, corresponding to the remarkable actor-better exploration actions. The asymptotic RISE controller, together with the two critics in OADC learning framework, guarantees accurate judgment on every exploration. On this basis, the satisfactory performance of the RL policy is guaranteed by the actor-better exploration based learning while the chattering problem arisen from the switching logic is addressed completely. Plenty of comparative experimental tests are presented to illustrate the superior performance of the proposed RL controller in terms of tracking accuracy and robustness.
引用
收藏
页码:2812 / 2821
页数:10
相关论文
共 47 条
[21]   Adaptive Trajectory Tracking Control using Reinforcement Learning for Quadrotor [J].
Lou, Wenjie ;
Guo, Xiao .
INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2016, 13
[22]   Uncertainty and Disturbance Estimator-Based Global Trajectory Tracking Control for a Quadrotor [J].
Lu, Qi ;
Ren, Beibei ;
Parameswaran, Siva .
IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2020, 25 (03) :1519-1530
[23]   Multirotor Aerial Vehicles Modeling, Estimation, and Control of Quadrotor [J].
Mahony, Robert ;
Kumar, Vijay ;
Corke, Peter .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 2012, 19 (03) :20-32
[24]   Safe Exploration Algorithms for Reinforcement Learning Controllers [J].
Mannucci, Tommaso ;
van Kampen, Erik-Jan ;
de Visser, Cornelis ;
Chu, Qiping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (04) :1069-1081
[25]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[26]   Low-level autonomous control and tracking of quadrotor using reinforcement learning [J].
Pi, Chen-Huan ;
Hu, Kai-Chun ;
Cheng, Stone ;
Wu, I-Chen .
CONTROL ENGINEERING PRACTICE, 2020, 95
[27]   Adaptive Position Tracking of VTOL UAVs [J].
Roberts, Andrew ;
Tayebi, Abdelhamid .
IEEE TRANSACTIONS ON ROBOTICS, 2011, 27 (01) :129-142
[28]   Deep reinforcement learning for quadrotor path following with adaptive velocity [J].
Rubi, Bartomeu ;
Morcego, Bernardo ;
Perez, Ramon .
AUTONOMOUS ROBOTS, 2021, 45 (01) :119-134
[29]   Reachability-Based Trajectory Safeguard (RTS): A Safe and Fast Reinforcement Learning Safety Layer for Continuous Control [J].
Shao, Yifei Simon ;
Chen, Chao ;
Kousik, Shreyas ;
Vasudevan, Ram .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) :3663-3670
[30]   A Fuzzy Adaptive Approach to Decoupled Visual Servoing for a Wheeled Mobile Robot [J].
Shi, Haobin ;
Xu, Meng ;
Hwang, Kao-Shing .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2020, 28 (12) :3229-3243