CALAMARI: Contact-Aware and Language conditioned spatial Action MApping for contact-RIch manipulation

被引:0
|
作者
Wi, Youngsun [1 ]
Van der Merwe, Mark [1 ]
Zeng, Andy [2 ]
Florence, Pete [2 ]
Fazeli, Nima [1 ]
机构
[1] Univ Michigan, Robot Dept, Ann Arbor, MI 48109 USA
[2] Google Deepmind, London, England
来源
基金
美国国家科学基金会;
关键词
Contact-rich Manipulation; Visual-language guided policies;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Making contact with purpose is a central part of robot manipulation and remains essential for many household tasks - from sweeping dust into a dustpan, to wiping tables; from erasing whiteboards, to applying paint. In this work, we investigate learning language-conditioned, vision-based manipulation policies wherein the action representation is in fact, contact itself - predicting contact formations at which tools grasped by the robot should meet an observable surface. Our approach, Contact-Aware and Language conditioned spatial Action MApping for contact-RIch manipulation (CALAMARI), exhibits several advantages including (i) benefiting from existing visual-language models for pretrained spatial features, grounding instructions to behaviors, and for sim2real transfer; and (ii) factorizing perception and control over a natural boundary (i.e., contact) into two modules that synergize with each other, whereby action predictions can be aligned per pixel with image observations, and low-level controllers can optimize motion trajectories that maintain contact while avoiding penetration. Experiments show that CALAMARI outperforms existing state-of-the-art model architectures for a broad range of contact-rich tasks, and pushes new ground on embodimentagnostic generalization to unseen objects with varying elasticity, geometry, and colors in both simulated and real-world settings.
引用
收藏
页数:19
相关论文
共 49 条
  • [41] Efficient Sim-to-real Transfer of Contact-Rich Manipulation Skills with Online Admittance Residual Learning
    Zhang, Xiang
    Wang, Changhao
    Sun, Lingfeng
    Wu, Zheng
    Zhu, Xinghao
    Tomizuka, Masayoshi
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [42] A Realistic Surgical Simulator for Non-Rigid and Contact-Rich Manipulation in Surgeries with the da Vinci Research Kit
    Ou, Yafei
    Zargarzadeh, Sadra
    Sedighi, Paniz
    Tavakoli, Mandi
    2024 21ST INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS, UR 2024, 2024, : 64 - 70
  • [43] Variable-Impedance and Force Control for Robust Learning of Contact-rich Manipulation Tasks from User Demonstration
    Enayati, Nima
    Mariani, Stefano
    Wahrburg, Arne
    Zanchettin, Andrea M.
    IFAC PAPERSONLINE, 2020, 53 (02): : 9834 - 9840
  • [44] Unified HumanRobotEnvironment Interaction Control in Contact-Rich Collaborative Manipulation Tasks via Model-Based Reinforcement Learning
    Liu, Xing
    Liu, Yu
    Liu, Zhengxiong
    Huang, Panfeng
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2023, 70 (11) : 11474 - 11482
  • [45] Variable Impedance Control in End-Effector Space: An Action Space for Reinforcement Learning in Contact-Rich Tasks
    Martin-Martin, Roberto
    Lee, Michelle A.
    Gardner, Rachel
    Savarese, Silvio
    Bohg, Jeannette
    Garg, Animesh
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 1010 - 1017
  • [46] Proactive Action Visual Residual Reinforcement Learning for Contact-Rich Tasks Using a Torque-Controlled Robot
    Shi, Yunlei
    Chen, Zhaopeng
    Liu, Hongxu
    Riedel, Sebastian
    Gao, Chunhui
    Feng, Qian
    Deng, Jun
    Zhang, Jianwei
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 765 - 771
  • [47] General-Purpose Sim2Real Protocol for Learning Contact-Rich Manipulation With Marker-Based Visuotactile Sensors
    Chen, Weihang
    Xu, Jing
    Xiang, Fanbo
    Yuan, Xiaodi
    Su, Hao
    Chen, Rui
    IEEE TRANSACTIONS ON ROBOTICS, 2024, 40 : 1509 - 1526
  • [48] Adaptive Contact-Rich Manipulation Through Few-Shot Imitation Learning With Force-Torque Feedback and Pre-Trained Object Representations
    Tsuji, Chikaha
    Coronado, Enrique
    Osorio, Pablo
    Venture, Gentiane
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (01): : 240 - 247
  • [49] Diff-LfD: Contact-aware Model-based Learning from Visual Demonstration for Robotic Manipulation via Differentiable Physics-based Simulation and Rendering
    Zhu, Xinghao
    Ke, Jinghan
    Xu, Zhixuan
    Sun, Zhixin
    Bai, Bizhe
    Lv, Jun
    Liu, Qingtao
    Zeng, Yuwei
    Ye, Qi
    Lu, Cewu
    Tomizuka, Masayoshi
    Shao, Lin
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229