Legged Robots that Keep on Learning: Fine-Tuning Locomotion Policies in the Real World

被引:31
作者
Smith, Laura [1 ]
Kew, J. Chase [2 ]
Peng, Xue Bin [1 ]
Ha, Sehoon [2 ,3 ]
Tang, Jie [2 ]
Levine, Sergey [1 ,2 ]
机构
[1] Univ Calif Berkeley, Berkeley AI Res, Berkeley, CA 90095 USA
[2] Google Res, New York, NY USA
[3] Gcorgia Inst Technol, Atlanta, GA USA
来源
2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022) | 2022年
关键词
D O I
10.1109/ICRA46639.2022.9812166
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Legged robots are physically capable of traversing a wide range of challenging environments, but designing controllers that are sufficiently robust to handle this diversity has been a long-standing challenge in robotics. Reinforcement learning presents an appealing approach for automating the controller design process and has been able to produce remarkably robust controllers when trained in a suitable range of environments. However, it is difficult to predict all likely conditions the robot will encounter during deployment and enumerate them at training-time. What if instead of training controllers that are robust enough to handle any eventuality, we enable the robot to continually learn in any setting it finds itself in? This kind of real-world reinforcement learning poses a number of challenges, including efficiency, safety, and autonomy. To address these challenges, we propose a practical robot reinforcement learning system for fine-tuning locomotion policies in the real world. We demonstrate that a modest amount of real-world training can substantially improve performance during deployment, and this enables a real A1 quadrupedal robot to autonomously fine-tune multiple locomotion skills in a range of environments, including an outdoor lawn and a variety of indoor terrains. (Videos and code1)
引用
收藏
页码:1593 / 1599
页数:7
相关论文
共 50 条
  • [31] Knee Implant Identification by Fine-Tuning Deep Learning Models
    Sukkrit Sharma
    Vineet Batta
    Malathy Chidambaranathan
    Prabhakaran Mathialagan
    Gayathri Mani
    M. Kiruthika
    Barun Datta
    Srinath Kamineni
    Guruva Reddy
    Suhas Masilamani
    Sandeep Vijayan
    Derek F. Amanatullah
    Indian Journal of Orthopaedics, 2021, 55 : 1295 - 1305
  • [32] A statistical learning based approach for parameter fine-tuning of metaheuristics
    Calvet, Laura
    Juan, Angel A.
    Serrat, Caries
    Ries, Jana
    SORT-STATISTICS AND OPERATIONS RESEARCH TRANSACTIONS, 2016, 40 (01) : 201 - 223
  • [33] Transfer Learning Gaussian Anomaly Detection by Fine-tuning Representations
    Rippel, Oliver
    Chavan, Arnav
    Lei, Chucai
    Merhof, Dorit
    IMPROVE: PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND VISION ENGINEERING, 2022, : 45 - 56
  • [34] Adaptive fine-tuning strategy for few-shot learning
    Zhuang, Xinkai
    Shao, Mingwen
    Gao, Wei
    Yang, Jianxin
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [35] Scalable Online Planning via Reinforcement Learning Fine-Tuning
    Fickinger, Arnaud
    Hu, Hengyuan
    Amos, Brandon
    Russell, Stuart
    Brown, Noam
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [36] FedFTHA: A Fine-Tuning and Head Aggregation Method in Federated Learning
    Wang, Yansong
    Xu, Hui
    Ali, Waqar
    Li, Miaobo
    Zhou, Xiangmin
    Shao, Jie
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (14) : 12749 - 12762
  • [37] Robust adversarial uncertainty quantification for deep learning fine-tuning
    Ahmed, Usman
    Lin, Jerry Chun-Wei
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (10) : 11355 - 11386
  • [38] Fine-Tuning a Personalized OpenBioLLM Using Offline Reinforcement Learning
    Shi, Jinsheng
    Yuan, Yuyu
    Wang, Ao
    Nie, Meng
    APPLIED SCIENCES-BASEL, 2025, 15 (05):
  • [39] On the Effectiveness of Fine-tuning Versus Meta-reinforcement Learning
    Mandi, Zhao
    Abbeel, Pieter
    James, Stephen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [40] Fine-Tuning Deep Neural Networks in Continuous Learning Scenarios
    Kaeding, Christoph
    Rodner, Erik
    Freytag, Alexander
    Denzler, Joachim
    COMPUTER VISION - ACCV 2016 WORKSHOPS, PT III, 2017, 10118 : 588 - 605