A Novel Reinforcement Learning-Based Robust Control Strategy for a Quadrotor

被引:23
作者
Hua, Hean [1 ,2 ]
Fang, Yongchun [1 ,2 ]
机构
[1] Nankai Univ, Inst Robot & Automat Informat Syst, Coll Artificial Intelligence, Tianjin 300353, Peoples R China
[2] Nankai Univ, Tianjin Key Lab Intelligent Robot, Tianjin 300353, Peoples R China
基金
中国国家自然科学基金;
关键词
Quadrotors; reinforcement learning (RL) control; robust integral of the signum of the error (RISE); RISE-guided learning; real-world applications; TRAJECTORY TRACKING CONTROL; ATTITUDE-CONTROL; LEVEL CONTROL; AERIAL; SAFE;
D O I
10.1109/TIE.2022.3165288
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this article, a novel reinforcement learning (RL)-based robust control approach is proposed for quadrotors, which guarantees efficient learning and satisfactory tracking performance by simultaneously evaluating the RL and the baseline method in training. Different from existing works, the key novelty is to design a practice-reliable RL control framework for quadrotors in a two-part cooperative manner. In the first part, based on the hierarchical property, a new robust integral of the signum of the error (RISE) design is proposed to ensure asymptotic convergence, which includes the nonlinear and the disturbance rejection terms. In the second part, a one-actor-dual-critic (OADC) learning framework is proposed, where the designed switching logic in the first part works as a benchmark to guide the learning. Specifically, the two critics independently evaluate the RL policy and the switching logic simultaneously, which are utilized for policy update, only when both are positive, corresponding to the remarkable actor-better exploration actions. The asymptotic RISE controller, together with the two critics in OADC learning framework, guarantees accurate judgment on every exploration. On this basis, the satisfactory performance of the RL policy is guaranteed by the actor-better exploration based learning while the chattering problem arisen from the switching logic is addressed completely. Plenty of comparative experimental tests are presented to illustrate the superior performance of the proposed RL controller in terms of tracking accuracy and robustness.
引用
收藏
页码:2812 / 2821
页数:10
相关论文
共 47 条
[11]  
Furrer F, 2016, STUD COMPUT INTELL, V625, P595, DOI 10.1007/978-3-319-26054-9_23
[12]  
Gao F, 2018, IEEE INT C INT ROBOT, P4715, DOI 10.1109/IROS.2018.8593579
[13]   A New Nonlinear Control Strategy Embedded With Reinforcement Learning for a Multirotor Transporting a Suspended Payload [J].
Hua, Hean ;
Fang, Yongchun ;
Zhang, Xuetao ;
Qian, Chen .
IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2022, 27 (02) :1174-1184
[14]   A Novel Robust Observer-Based Nonlinear Trajectory Tracking Control Strategy for Quadrotors [J].
Hua, Hean ;
Fang, Yongchun ;
Zhang, Xuetao ;
Lu, Biao .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2021, 29 (05) :1952-1963
[15]   Control of a Quadrotor With Reinforcement Learning [J].
Hwangbo, Jemin ;
Sa, Inkyu ;
Siegwart, Roland ;
Hutter, Marco .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2017, 2 (04) :2096-2103
[16]   Low-Level Control of a Quadrotor With Deep Model-Based Reinforcement Learning [J].
Lambert, Nathan O. ;
Drewe, Daniel S. ;
Yaconelli, Joseph ;
Levine, Sergey ;
Calandra, Roberto ;
Pister, Kristofer S. J. .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (04) :4224-4230
[17]   Geometric Tracking Control of a Quadrotor UAV on SE(3) [J].
Lee, Taeyoung ;
Leok, Melvin ;
McClamroch, N. Harris .
49TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2010, :5420-5425
[18]   Adaptive Attitude Control of a Quadrotor Using Fast Nonsingular Terminal Sliding Mode [J].
Lian, Shikang ;
Meng, Wei ;
Lin, Zemin ;
Shao, Ke ;
Zheng, Jinchuan ;
Li, Hongyi ;
Lu, Renquan .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2022, 69 (02) :1597-1607
[19]  
Lillicrap T.P., 2016, P ICLR
[20]   Control Design for UAV Quadrotors via Embedded Model Control [J].
Lotufo, Mauricio Alejandro ;
Colangelo, Luigi ;
Novara, Carlo .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2020, 28 (05) :1741-1756