Reinforcement Learning with Experience Replay for Model-Free Humanoid Walking Optimization

被引:8
作者
Wawrzynski, Pawel [1 ]
机构
[1] Warsaw Univ Technol, Inst Control & Computat Engn, PL-00665 Warsaw, Poland
关键词
Reinforcement learning; learning in robots; humanoids; bipedal walking; CONVERGENCE; ROBOTS;
D O I
10.1142/S0219843614500248
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In this paper, a control system for humanoid robot walking is approximately optimized by means of reinforcement learning. Given is a 18 DOF humanoid whose gait is based on replaying a simple trajectory. This trajectory is translated into a reactive policy. A neural network whose input represents the robot state learns to produce appropriate output that additively modifies the initial control. The learning algorithm applied is actor critic with experience replay. In 50 min of learning, the slow initial gait changes to a dexterous and fast walking. No model of the robot dynamics is engaged. The methodology in use is generic and can be applied to optimize control systems for diverse robots of comparable complexity.
引用
收藏
页数:21
相关论文
共 50 条
  • [31] Model-free LQ Control for Unmanned Helicopters using Reinforcement Learning
    Lee, Dong Jin
    Bang, Hyochoong
    [J]. 2011 11TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2011, : 117 - 120
  • [32] Model-Free Reinforcement Learning of Minimal-Cost Variance Control
    Jing, Gangshan
    Bai, He
    George, Jemin
    Chakrabortty, Aranya
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2020, 4 (04): : 916 - 921
  • [33] Hierarchical Dynamic Power Management Using Model-Free Reinforcement Learning
    Wang, Yanzhi
    Triki, Maryam
    Lin, Xue
    Ammari, Ahmed C.
    Pedram, Massoud
    [J]. PROCEEDINGS OF THE FOURTEENTH INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN (ISQED 2013), 2013, : 170 - 177
  • [34] Model-Free Decentralized Reinforcement Learning Control of Distributed Energy Resources
    Mukherjee, Sayak
    Bai, He
    Chakrabortty, Aranya
    [J]. 2020 IEEE POWER & ENERGY SOCIETY GENERAL MEETING (PESGM), 2020,
  • [35] Model-free Reinforcement Learning of Semantic Communication by Stochastic Policy Gradient
    Beck, Edgar
    Bockelmann, Carsten
    Dekorsy, Armin
    [J]. 2024 IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING FOR COMMUNICATION AND NETWORKING, ICMLCN 2024, 2024, : 367 - 373
  • [36] Budget Constrained Bidding by Model-free Reinforcement Learning in Display Advertising
    Wu, Di
    Chen, Xiujun
    Yang, Xun
    Wang, Hao
    Tan, Qing
    Zhang, Xiaoxun
    Xu, Jian
    Gai, Kun
    [J]. CIKM'18: PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2018, : 1443 - 1451
  • [37] Can model-free reinforcement learning explain deontological moral judgments?
    Ayars, Alisabeth
    [J]. COGNITION, 2016, 150 : 232 - 242
  • [38] A Model-Free Deep Reinforcement Learning Approach to Piano Fingering Generation
    Phan, Ananda
    Ahn, Chang Wook
    [J]. 2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 31 - 37
  • [39] Deep Reinforcement Learning for Autonomous Model-Free Navigation with Partial Observability
    Tapia, Daniel
    Parras, Juan
    Zazo, Santiago
    [J]. 2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,
  • [40] MODEL-FREE PREDICTIVE CONTROL OF NONLINEAR PROCESSES BASED ON REINFORCEMENT LEARNING
    Shah, Hitesh
    Gopal, M.
    [J]. IFAC PAPERSONLINE, 2016, 49 (01): : 89 - 94