Information gathering in POMDPs using active inference

被引:0
作者
Walraven, Erwin [1 ]
Sijs, Joris [2 ]
Burghouts, Gertjan J. [1 ]
机构
[1] Netherlands Org Appl Sci Res, The Hague, Netherlands
[2] Delft Univ Technol, Delft, Netherlands
关键词
Planning under uncertainty; POMDP; Information gathering; Active inference;
D O I
10.1007/s10458-024-09683-4
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Gathering information about the environment state is the main goal in several planning tasks for autonomous agents, such as surveillance, inspection and tracking of objects. Such planning tasks are typically modeled using a Partially Observable Markov Decision Process (POMDP), and in the literature several approaches have emerged to consider information gathering during planning and execution. Similar developments can be seen in the field of active inference, which focuses on active information collection in order to be able to reach a goal. Both fields use POMDPs to model the environment, but the underlying principles for action selection are different. In this paper we create a bridge between both research fields by discussing how they relate to each other and how they can be used for information gathering. Our contribution is a tailored approach to model information gathering tasks directly in the active inference framework. A series of experiments demonstrates that our approach enables agents to gather information about the environment state. As a result, active inference becomes an alternative to common POMDP approaches for information gathering, which opens the door towards more cross cutting research at the intersection of both fields. This is advantageous, because recent advancements in POMDP solvers may be used to accelerate active inference, and the principled active inference framework may be used to model POMDP agents that operate in a neurobiologically plausible fashion.
引用
收藏
页数:22
相关论文
共 38 条
[11]   Active inference on discrete state-spaces: A synthesis [J].
Da Costa, Lancelot ;
Parr, Thomas ;
Sajid, Noor ;
Veselic, Sebastijan ;
Neacsu, Victorita ;
Friston, Karl .
JOURNAL OF MATHEMATICAL PSYCHOLOGY, 2020, 99
[12]  
de Nijs F, 2021, J ARTIF INTELL RES, V70, P955
[13]  
Di Paola D, 2010, INT J ADV ROBOT SYST, V7, P19
[14]  
Doshi P, 2009, J ARTIF INTELL RES, V34
[15]  
Fountas Z., 2020, Advances in neural information processing systems, V33, P11662
[16]   Sophisticated Inference [J].
Friston, Karl ;
Da Costa, Lancelot ;
Hafner, Danijar ;
Hesp, Casper ;
Parr, Thomas .
NEURAL COMPUTATION, 2021, 33 (03) :713-763
[17]   The free-energy principle: a unified brain theory? [J].
Friston, Karl J. .
NATURE REVIEWS NEUROSCIENCE, 2010, 11 (02) :127-138
[18]  
Geffner H., 2013, A Concise Introduction to Models and Methods for Automated Planning
[19]  
Hou P, 2016, AAAI CONF ARTIF INTE, P3138
[20]   Planning and acting in partially observable stochastic domains [J].
Kaelbling, LP ;
Littman, ML ;
Cassandra, AR .
ARTIFICIAL INTELLIGENCE, 1998, 101 (1-2) :99-134