Learning Variable Impedance Control via Inverse Reinforcement Learning for Force-Related Tasks

被引:64
作者
Zhang, Xiang [1 ]
Sun, Liting [1 ]
Kuang, Zhian [1 ,2 ]
Tomizuka, Masayoshi [1 ]
机构
[1] Univ Calif Berkeley, Dept Mech Engn, Berkeley, CA 94720 USA
[2] Harbin Inst Technol, Res Inst Intelligent Control & Syst, Harbin 150001, Peoples R China
关键词
Compliance and impedance control; learning from demonstration; machine learning for robot control;
D O I
10.1109/LRA.2021.3061374
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Many manipulation tasks require robots to interact with unknown environments. In such applications, the ability to adapt the impedance according to different task phases and environment constraints is crucial for safety and performance. Although many approaches based on deep reinforcement learning (RI) and learning from demonstration (LfD) have been proposed to obtain variable impedance skills on contact-rich manipulation tasks, these skills are typically task-specific and could be sensitive to changes in task settings. This letter proposes an inverse reinforcement learning (IRL) based approach to recover both the variable impedance policy and reward function from expert demonstrations. We explore different action space of the reward functions to achieve a more general representation of expert variable impedance skills. Experiments on two variable impedance tasks (Peg-in-Hole and Cup-on-Plate) were conducted in both simulations and on a real FANUC LR Mate 200iD/7 L industrial robot. The comparison results with behavior cloning and force-based IRL proved that the learned reward function in the gain action space has better transferability than in the force space. Experiment videos are available at https://msc.berkeley.edu/research/impedance-irl.html.
引用
收藏
页码:2225 / 2232
页数:8
相关论文
共 21 条
  • [1] Abu-Dakka F. J., 2018, P IEEE RAS 18 INT C, P1
  • [2] Learning variable impedance control
    Buchli, Jonas
    Stulp, Freek
    Theodorou, Evangelos
    Schaal, Stefan
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2011, 30 (07) : 820 - 833
  • [3] Finn C., 2016, A Connection between Generative Adversarial Networks, Inverse Reinforcement Learning, and Energy-Based Models
  • [4] Finn C, 2016, PR MACH LEARN RES, V48
  • [5] Fu Justin, 2017, Learning robust rewards with adversarial inverse reinforcement learning. arXiv
  • [6] HOGAN N, 1985, J DYN SYST-T ASME, V107, P8, DOI 10.1115/1.3140713
  • [7] Inoue T, 2017, IEEE INT C INT ROBOT, P819, DOI 10.1109/IROS.2017.8202244
  • [8] Koltun V., 2012, P 29 INT C MACH LEAR
  • [9] Kuang Z., 2021, ARXIV210203531
  • [10] Precise Linear-Motor Synchronization Control via Cross-Coupled Second-Order Discrete-Time Fractional-Order Sliding Mode
    Kuang, Zhian
    Gao, Huijun
    Tomizuka, Masayoshi
    [J]. IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2021, 26 (01) : 358 - 368