Active Vision for Robot Manipulators Using the Free Energy Principle

被引:16
作者
Van de Maele, Toon [1 ]
Verbelen, Tim [1 ]
Catal, Ozan [1 ]
De Boom, Cedric [1 ]
Dhoedt, Bart [1 ]
机构
[1] Univ Ghent, IMEC, Dept Informat Technol, IDLab, Ghent, Belgium
来源
FRONTIERS IN NEUROROBOTICS | 2021年 / 15卷
关键词
active vision; active inference; deep learning; generative modeling; robotics; INFERENCE; RECONSTRUCTION; CONSTRUCTION;
D O I
10.3389/fnbot.2021.642780
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Occlusions, restricted field of view and limited resolution all constrain a robot's ability to sense its environment from a single observation. In these cases, the robot first needs to actively query multiple observations and accumulate information before it can complete a task. In this paper, we cast this problem of active vision as active inference, which states that an intelligent agent maintains a generative model of its environment and acts in order to minimize its surprise, or expected free energy according to this model. We apply this to an object-reaching task for a 7-DOF robotic manipulator with an in-hand camera to scan the workspace. A novel generative model using deep neural networks is proposed that is able to fuse multiple views into an abstract representation and is trained from data by minimizing variational free energy. We validate our approach experimentally for a reaching task in simulation in which a robotic agent starts without any knowledge about its workspace. Each step, the next view pose is chosen by evaluating the expected free energy. We find that by minimizing the expected free energy, exploratory behavior emerges when the target object to reach is not in view, and the end effector is moved to the correct reach position once the target is located. Similar to an owl scavenging for prey, the robot naturally prefers higher ground for exploring, approaching its target once located.
引用
收藏
页数:18
相关论文
共 64 条
  • [1] Alexander D, 2009, P BRIT MACH VIS C, DOI [10.5244/C.23.53, DOI 10.5244/C.23.53]
  • [2] ALOIMONOS J, 1987, INT J COMPUT VISION, V1, P333
  • [3] [Anonymous], 2015, INT C LEARN REPR ICL
  • [4] [Anonymous], 2018, EUR C COMP VIS
  • [5] Beal M. J., 2003, PhD Thesis, P544
  • [6] Trends and challenges in robot manipulation
    Billard, Aude
    Kragic, Danica
    [J]. SCIENCE, 2019, 364 (6446) : 1149 - +
  • [7] Blender Online Community, 2018, BlenderA 3D modelling and rendering package
  • [8] Burgess C.P., 2019, ARXIV PREPRINT ARXIV
  • [9] Learning Generative State Space Models for Active Inference
    Catal, Ozan
    Wauthier, Samuel
    De Boom, Cedric
    Verbelen, Tim
    Dhoedt, Bart
    [J]. FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2020, 14
  • [10] Chang A.X., 2015, SHAPENET INFORM RIC