Safe Learning in Robotics: From Learning-Based Control to Safe Reinforcement Learning

被引:288
作者
Brunke, Lukas [1 ,2 ,3 ]
Greeff, Melissa [1 ,2 ,3 ]
Hall, Adam W. [1 ,2 ,3 ]
Yuan, Zhaocong [1 ,2 ,3 ]
Zhou, Siqi [1 ,2 ,3 ]
Panerati, Jacopo [1 ,2 ,3 ]
Schoellig, Angela P. [1 ,2 ,3 ]
机构
[1] Univ Toronto, Inst Aerosp Studies, Toronto, ON, Canada
[2] Univ Toronto, Robot Inst, Toronto, ON, Canada
[3] Vector Inst Artificial Intelligence, Toronto, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
safe learning; robotics; robot learning; learning-based control; safe reinforcement learning; adaptive control; robust control; model predictive control; machine learning; benchmarks; MODEL-PREDICTIVE CONTROL; BARRIER FUNCTIONS; TRACKING CONTROL; OPTIMIZATION; EXPLORATION; ROBUSTNESS; ALGORITHMS; FRAMEWORK; SYSTEMS;
D O I
10.1146/annurev-control-042920-020211
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The last half decade has seen a steep rise in the number of contributions on safe learning methods for real-world robotic deployments from both the control and reinforcement learning communities. This article provides a concise but holistic review of the recent advances made in using machine learning to achieve safe decision-making under uncertainties, with a focus on unifying the language and frameworks used in control theory and reinforcement learning research. It includes learning-based control approaches that safely improve performance by learning the uncertain dynamics, reinforcement learning approaches that encourage safety or robustness, and methods that can formally certify the safety of a learned control policy. As data- and learning-based robot control methods continue to gain traction, researchers must understand when and how to best leverage them in real-world scenarios where safety is imperative, such as when operating in close proximity to humans. We highlight some of the open challenges that will drive the field of robot learning in the coming years, and emphasize the need for realistic physics-based benchmarks to facilitate fair comparisons between control and reinforcement learning approaches.
引用
收藏
页码:411 / 444
页数:34
相关论文
共 153 条
  • [1] Achiam J, 2017, PR MACH LEARN RES, V70
  • [2] Iterative learning control: Brief survey and categorization
    Ahn, Hyo-Sung
    Chen, YangQuan
    Moore, Kevin L.
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2007, 37 (06): : 1099 - 1121
  • [3] Alshiekh M, 2018, AAAI CONF ARTIF INTE, P2669
  • [4] Altman E, 1999, STOCH MODEL SER, V1st
  • [5] Ames AD, 2019, 2019 18TH EUROPEAN CONTROL CONFERENCE (ECC), P3420, DOI [10.23919/ECC.2019.8796030, 10.23919/ecc.2019.8796030]
  • [6] [Anonymous], DYN SYST LAB 2021 SA
  • [7] Deep Reinforcement Learning A brief survey
    Arulkumaran, Kai
    Deisenroth, Marc Peter
    Brundage, Miles
    Bharath, Anil Anthony
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) : 26 - 38
  • [8] Provably safe and robust learning-based model predictive control
    Aswani, Anil
    Gonzalez, Humberto
    Sastry, S. Shankar
    Tomlin, Claire
    [J]. AUTOMATICA, 2013, 49 (05) : 1216 - 1226
  • [9] Bajcsy A, 2019, IEEE DECIS CONTR P, P1758, DOI 10.1109/CDC40024.2019.9030133
  • [10] Baumann D., 2021, ARXIV210513281 CSRO