Understanding the stability of deep control policies for biped locomotion

被引:3
|
作者
Park, Hwangpil [1 ,3 ]
Yu, Ri [1 ]
Lee, Yoonsang [4 ]
Lee, Kyungho [5 ]
Lee, Jehee [2 ]
机构
[1] Seoul Natl Univ, Seoul, South Korea
[2] Seoul Natl Univ, Dept Comp Sci & Engn, Seoul, South Korea
[3] Samsung Elect, Suwon, South Korea
[4] Hanyang Univ, Comp Sci, Seoul, South Korea
[5] NC Soft, Sungnam, South Korea
来源
VISUAL COMPUTER | 2023年 / 39卷 / 01期
关键词
Biped locomotion; Deep reinforcement learning; Gait analysis; Physically based simulation; Push-recovery stability; RECOVERY;
D O I
10.1007/s00371-021-02342-9
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Achieving stability and robustness is the primary goal of biped locomotion control. Recently, deep reinforcement learning (DRL) has attracted great attention as a general methodology for constructing biped control policies and demonstrated significant improvements over the previous state-of-the-art control methods. Although deep control policies are more advantageous compared with previous controller design approaches, many questions remain: Are deep control policies as robust as human walking? Does simulated walking involve strategies similar to human walking for maintaining balance? Does a particular gait pattern affect human and simulated walking similarly? What do deep policies learn to achieve improved gait stability? The goal of this study is to address these questions by evaluating the push-recovery stability of deep policies compared with those of human subjects and a previous feedback controller. Furthermore, we conducted experiments to evaluate the effectiveness of variants of DRL algorithms.
引用
收藏
页码:473 / 487
页数:15
相关论文
共 50 条
  • [21] Deep reinforcement learning method for biped robot gait control
    Feng C.
    Zhang Y.
    Huang C.
    Jiang W.
    Wu Z.
    1600, CIMS (27): : 2341 - 2349
  • [22] Simultaneous Locomotion of Biped Robot with the Transmission of Human Motion
    Prasanga, D. Kasun
    Ohnishi, Kouhei
    PROCEEDINGS OF THE IECON 2016 - 42ND ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2016, : 797 - 802
  • [23] Experimental studies of a neural oscillator for biped locomotion with QRIO
    Endo, G
    Nakanishi, J
    Morimoto, J
    Cheng, G
    2005 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), VOLS 1-4, 2005, : 596 - 602
  • [24] A framework for learning biped locomotion with dynamical movement primitives
    Nakanishi, J
    Morimoto, J
    Endo, G
    Cheng, G
    Schaal, S
    Kawato, M
    2004 4TH IEEE/RAS INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS, VOLS 1 AND 2, PROCEEDINGS, 2004, : 925 - 940
  • [25] Passive/active unified dynamic walking for biped locomotion
    Huang, Qingjiu
    Hase, Takamasa
    Ono, Kyosuke
    2007 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS, VOLS 1-5, 2007, : 964 - 971
  • [26] DIEES biped robot: A bio-inspired pneumatic platform for human locomotion analysis and stiffness control
    Spampinato, Giacomo
    Muscato, Giovanni
    2006 6TH IEEE-RAS INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS, VOLS 1 AND 2, 2006, : 478 - +
  • [27] Measurement of robustness for biped locomotion using a linearized Poincare map
    Cheng, MY
    Lin, CS
    ROBOTICA, 1996, 14 : 253 - 259
  • [28] Energy Efficient MPC for Biped Semi-passive Locomotion
    Neves, C.
    Ventura, R.
    ROBOT 2015: SECOND IBERIAN ROBOTICS CONFERENCE: ADVANCES IN ROBOTICS, VOL 2, 2016, 418 : 145 - 156
  • [29] Learning sensory feedback to CPG with policy gradient for biped locomotion
    Matsubara, T
    Morimoto, J
    Nakanishi, J
    Sato, MA
    Doya, K
    2005 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), VOLS 1-4, 2005, : 4164 - 4169
  • [30] A Disturbance Rejection Control Method Based on Deep Reinforcement Learning for a Biped Robot
    Liu, Chuzhao
    Gao, Junyao
    Tian, Dingkui
    Zhang, Xuefeng
    Liu, Huaxin
    Meng, Libo
    APPLIED SCIENCES-BASEL, 2021, 11 (04): : 1 - 17