A Novel Reinforcement Learning-Based Robust Control Strategy for a Quadrotor

被引:23
作者
Hua, Hean [1 ,2 ]
Fang, Yongchun [1 ,2 ]
机构
[1] Nankai Univ, Inst Robot & Automat Informat Syst, Coll Artificial Intelligence, Tianjin 300353, Peoples R China
[2] Nankai Univ, Tianjin Key Lab Intelligent Robot, Tianjin 300353, Peoples R China
基金
中国国家自然科学基金;
关键词
Quadrotors; reinforcement learning (RL) control; robust integral of the signum of the error (RISE); RISE-guided learning; real-world applications; TRAJECTORY TRACKING CONTROL; ATTITUDE-CONTROL; LEVEL CONTROL; AERIAL; SAFE;
D O I
10.1109/TIE.2022.3165288
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this article, a novel reinforcement learning (RL)-based robust control approach is proposed for quadrotors, which guarantees efficient learning and satisfactory tracking performance by simultaneously evaluating the RL and the baseline method in training. Different from existing works, the key novelty is to design a practice-reliable RL control framework for quadrotors in a two-part cooperative manner. In the first part, based on the hierarchical property, a new robust integral of the signum of the error (RISE) design is proposed to ensure asymptotic convergence, which includes the nonlinear and the disturbance rejection terms. In the second part, a one-actor-dual-critic (OADC) learning framework is proposed, where the designed switching logic in the first part works as a benchmark to guide the learning. Specifically, the two critics independently evaluate the RL policy and the switching logic simultaneously, which are utilized for policy update, only when both are positive, corresponding to the remarkable actor-better exploration actions. The asymptotic RISE controller, together with the two critics in OADC learning framework, guarantees accurate judgment on every exploration. On this basis, the satisfactory performance of the RL policy is guaranteed by the actor-better exploration based learning while the chattering problem arisen from the switching logic is addressed completely. Plenty of comparative experimental tests are presented to illustrate the superior performance of the proposed RL controller in terms of tracking accuracy and robustness.
引用
收藏
页码:2812 / 2821
页数:10
相关论文
共 47 条
[1]  
Altan A., 2020, PROC INT S MULTIDISC, P1
[2]   Model predictive control of three-axis gimbal system mounted on UAV for real-time target tracking under external disturbances [J].
Altan, Aytac ;
Hacioglu, Rifat .
MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2020, 138
[3]  
Altan Aytac, 2018, 2018 6 INT C CONTROL
[4]  
[Anonymous], 2017, 5 INT C ADV MECH ROB
[5]   Adaptive Trajectory Tracking for Quadrotor MAVs in Presence of Parameter Uncertainties and External Disturbances [J].
Antonelli, Gianluca ;
Cataldi, Elisabetta ;
Arrichiello, Filippo ;
Giordano, Paolo Robuffo ;
Chiaverini, Stefano ;
Franchi, Antonio .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2018, 26 (01) :248-254
[6]   Gain-Adaptive Robust Backstepping Position Control of a BLDC Motor System [J].
Ba, Dang Xuan ;
Yeom, Hoyeon ;
Kim, Jihoon ;
Bae, Joonbum .
IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2018, 23 (05) :2470-2481
[7]   Thrust Control for Multirotor Aerial Vehicles [J].
Bangura, Moses ;
Mahony, Robert .
IEEE TRANSACTIONS ON ROBOTICS, 2017, 33 (02) :390-405
[8]   Asymptotic tracking by a reinforcement learning-based adaptive critic controller [J].
Bhasin S. ;
Sharma N. ;
Patre P. ;
Dixon W. .
Journal of Control Theory and Applications, 2011, 9 (3) :400-409
[9]   Tracking Flight Control of Quadrotor Based on Disturbance Observer [J].
Chen, Mou ;
Xiong, Shixun ;
Wu, Qingxian .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2021, 51 (03) :1414-1423
[10]  
Fujimoto S, 2018, PR MACH LEARN RES, V80