Sim-to-Real Model-Based and Model-Free Deep Reinforcement Learning for Tactile Pushing

被引:7
|
作者
Yang, Max [1 ]
Lin, Yijiong [1 ]
Church, Alex [1 ]
Lloyd, John [1 ]
Zhang, Dandan [1 ]
Barton, David A. W. [1 ]
Lepora, Nathan F. [1 ]
机构
[1] Univ Bristol, Bristol Robot Lab, Dept Engn Math, Bristol BS8 1UB, England
基金
英国工程与自然科学研究理事会;
关键词
Force and tactile sensing; dexterous manipulation; reinforcement learning;
D O I
10.1109/LRA.2023.3295236
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Object pushing presents a key non-prehensile manipulation problem that is illustrative of more complex robotic manipulation tasks. While deep reinforcement learning (RL) methods have demonstrated impressive learning capabilities using visual input, a lack of tactile sensing limits their capability for fine and reliable control during manipulation. Here we propose a deep RL approach to object pushing using tactile sensing without visual input, namely tactile pushing. We present a goal-conditioned formulation that allows both model-free and model-based RL to obtain accurate policies for pushing an object to a goal. To achieve real-world performance, we adopt a sim-to-real approach. Our results demonstrate that it is possible to train on a single object and a limited sample of goals to produce precise and reliable policies that can generalize to a variety of unseen objects and pushing scenarios without domain randomization. We experiment with the trained agents in harsh pushing conditions, and show that with significantly more training samples, a model-free policy can outperform a model-based planner, generating shorter and more reliable pushing trajectories despite large disturbances. The simplicity of our training environment and effective real-world performance highlights the value of rich tactile information for fine manipulation.
引用
收藏
页码:5480 / 5487
页数:8
相关论文
共 50 条
  • [41] Combining Model-Based and Model-Free Reinforcement Learning Policies for More Efficient Sepsis Treatment
    Liu, Xiangyu
    Yu, Chao
    Huang, Qikai
    Wang, Luhao
    Wu, Jianfeng
    Guan, Xiangdong
    BIOINFORMATICS RESEARCH AND APPLICATIONS, ISBRA 2021, 2021, 13064 : 105 - 117
  • [42] Ventral Striatum and Orbitofrontal Cortex Are Both Required for Model-Based, But Not Model-Free, Reinforcement Learning
    McDannald, Michael A.
    Lucantonio, Federica
    Burke, Kathryn A.
    Niv, Yael
    Schoenbaum, Geoffrey
    JOURNAL OF NEUROSCIENCE, 2011, 31 (07): : 2700 - 2705
  • [43] Grounded action transformation for sim-to-real reinforcement learning
    Hanna, Josiah P.
    Desai, Siddharth
    Karnan, Haresh
    Warnell, Garrett
    Stone, Peter
    MACHINE LEARNING, 2021, 110 (09) : 2469 - 2499
  • [44] Meta Reinforcement Learning for Sim-to-real Domain Adaptation
    Arndt, Karol
    Hazara, Murtaza
    Ghadirzadeh, Ali
    Kyrki, Ville
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 2725 - 2731
  • [45] Accelerated Sim-to-Real Deep Reinforcement Learning: Learning Collision Avoidance from Human Player
    Niu, Hanlin
    Ji, Ze
    Arvin, Farshad
    Lennox, Barry
    Yin, Hujun
    Carrasco, Joaquin
    2021 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION (SII), 2021, : 144 - 149
  • [46] Model-based learning retrospectively updates model-free values
    Max Doody
    Maaike M. H. Van Swieten
    Sanjay G. Manohar
    Scientific Reports, 12
  • [47] Model-Based and Model-Free Mechanisms of Human Motor Learning
    Haith, Adrian M.
    Krakauer, John W.
    PROGRESS IN MOTOR CONTROL: NEURAL, COMPUTATIONAL AND DYNAMIC APPROACHES, 2013, 782 : 1 - 21
  • [48] Model-based learning retrospectively updates model-free values
    Doody, Max
    Van Swieten, Maaike M. H.
    Manohar, Sanjay G.
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [49] Model-Free Deep Inverse Reinforcement Learning by Logistic Regression
    Eiji Uchibe
    Neural Processing Letters, 2018, 47 : 891 - 905
  • [50] Model-free Deep Reinforcement Learning for Urban Autonomous Driving
    Chen, Jianyu
    Yuan, Bodi
    Tomizuka, Masayoshi
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 2765 - 2771