Visual Affordance Prediction for Guiding Robot Exploration

被引:2
作者
Bharadhwaj, Homanga [1 ]
Gupta, Abhinav [1 ]
Tulsiani, Shubham [1 ]
机构
[1] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
来源
2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA | 2023年
关键词
D O I
10.1109/ICRA48891.2023.10161288
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Motivated by the intuitive understanding humans have about the space of possible interactions, and the ease with which they can generalize this understanding to previously unseen scenes, we develop an approach for learning 'visual affordances'. Given an input image of a scene, we infer a distribution over plausible future states that can be achieved via interactions with it. To allow predicting diverse plausible futures, we discretize the space of continuous images with a VQ-VAE and use a Transformer-based model to learn a conditional distribution in the latent embedding space. We show that these models can be trained using large-scale and diverse passive data, and that the learned models exhibit compositional generalization to diverse objects beyond the training distribution. We evaluate the quality and diversity of the generations, and demonstrate how the trained affordance model can be used for guiding exploration during visual goal-conditioned policy learning in robotic manipulation.
引用
收藏
页码:3029 / 3036
页数:8
相关论文
共 50 条
  • [41] Agency and social affordance shape visual perception
    Le Besnerais, Alexis
    Prigent, Elise
    Grynszpan, Ouriel
    [J]. COGNITION, 2023, 233
  • [42] The influence of agency and affordance on visual perception of motion
    Oule, Pierre-Adams Ouattara
    Prigent, Elise
    Grynszpan, Ouriel
    [J]. INTERNATIONAL JOURNAL OF PSYCHOLOGY, 2024, 59 : 149 - 149
  • [43] Affordance-based human-robot interaction
    Moratz, Reinhard
    Tenbrink, Thora
    [J]. TOWARDS AFFORDANCE-BASED ROBOT CONTROL, 2008, 4760 : 63 - +
  • [44] TARS: Tactile Affordance in Robot Synesthesia for Dexterous Manipulation
    Wu, Qiwei
    Wang, Haidong
    Zhou, Jiayu
    Xiong, Xiaogang
    Lou, Yunjiang
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (01): : 327 - 334
  • [45] Does it help a robot navigate to call navigability an affordance?
    Hertzberg, Joachim
    Lingemann, Kai
    Loerken, Christopher
    Nuechter, Andreas
    Stiene, Stefan
    [J]. TOWARDS AFFORDANCE-BASED ROBOT CONTROL, 2008, 4760 : 16 - +
  • [46] Affordance Triggering For Arbitrary States Based on Robot Exploring
    Yi, Chang'an
    Zheng, Guofei
    Bi, Sheng
    Luo, Ronghua
    Yin, Pengshuai
    Xu, Xinshi
    Min, Huaqing
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE ROBIO 2017), 2017, : 1856 - 1861
  • [47] A Novel Formalization For Robot Cognition Based on Affordance Model
    Yi, Chang'an
    Min, Huaqing
    Luo, Ronghua
    Zhong, Zhipeng
    Shen, Xiaowen
    [J]. 2012 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO 2012), 2012,
  • [48] Topological Mapping for Robot Navigation using Affordance Features
    Varadarajan, Karthik Mahesh
    [J]. PROCEEDINGS OF THE 2015 6TH INTERNATIONAL CONFERENCE ON AUTOMATION, ROBOTICS AND APPLICATIONS (ICARA), 2015, : 42 - 49
  • [49] A method of initial welding position guiding for arc welding robot based on visual servo control
    郭振民
    陈善本
    邱涛
    吴林
    [J]. China Welding, 2003, (01) : 31 - 35
  • [50] An Evaluation of Affordance Templates for Human-Robot Interaction
    Frering, Laurent
    Mohr-Ziak, Peter
    Koenczoel, Clemens
    Mosbacher, Jochen A.
    Eder, Matthias
    Albert, Dietrich
    Kubicek, Bettina
    Steinbauer-Wagner, Gerald
    [J]. 2024 33RD IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, ROMAN 2024, 2024, : 1957 - 1962