Simulating Active Inference Processes by Message Passing

被引:29
作者
van de laar, Thijs W. [1 ]
de Vries, Bert [1 ,2 ]
机构
[1] Eindhoven Univ Technol, Dept Elect Engn, Eindhoven, Netherlands
[2] GN Hearing Benelux BV, Eindhoven, Netherlands
关键词
active inference; free-energy principle; message passing; state-space models; Forney-style factor graphs; FREE-ENERGY PRINCIPLE; FACTOR GRAPH APPROACH;
D O I
10.3389/frobt.2019.00020
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
The free energy principle (FEP) offers a variational calculus-based description for how biological agents persevere through interactions with their environment. Active inference (Al) is a corollary of the FEP, which states that biological agents act to fulfill prior beliefs about preferred future observations (target priors). Purposeful behavior then results from variational free energy minimization with respect to a generative model of the environment with included target priors. However, manual derivations for free energy minimizing algorithms on custom dynamic models can become tedious and error-prone. While probabilistic programming (PP) techniques enable automatic derivation of inference algorithms on free-form models, full automation of Al requires specialized tools for inference on dynamic models, together with the description of an experimental protocol that governs the interaction between the agent and its simulated environment. The contributions of the present paper are two-fold. Firstly, we illustrate how Al can be automated with the use of ForneyLab, a recent PP toolbox that specializes in variational inference on flexibly definable dynamic models. More specifically, we describe Al agents in a dynamic environment as probabilistic state space models (SSM) and perform inference for perception and control in these agents by message passing on a factor graph representation of the SSM. Secondly, we propose a formal experimental protocol for simulated Al. We exemplify how this protocol leads to goal-directed behavior for flexibly definable Al agents in two classical RL examples, namely the Bayesian thermostat and the mountain car parking problems.
引用
收藏
页数:15
相关论文
共 58 条
[31]   Free Energy, Value, and Attractors [J].
Friston, Karl J. ;
Ao, Ping .
COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2012, 2012
[32]   The free-energy principle: a unified brain theory? [J].
Friston, Karl J. .
NATURE REVIEWS NEUROSCIENCE, 2010, 11 (02) :127-138
[33]   The free-energy principle: a rough guide to the brain? [J].
Friston, Karl J. .
TRENDS IN COGNITIVE SCIENCES, 2009, 13 (07) :293-301
[34]   Reinforcement Learning or Active Inference? [J].
Friston, Karl J. ;
Daunizeau, Jean ;
Kiebel, Stefan J. .
PLOS ONE, 2009, 4 (07)
[35]   Predictive coding under the free-energy principle [J].
Friston, Karl J. ;
Kiebel, Stefan .
PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY B-BIOLOGICAL SCIENCES, 2009, 364 (1521) :1211-1221
[36]   Preference-based reinforcement learning: a formal framework and a policy iteration algorithm [J].
Fuernkranz, Johannes ;
Huellermeier, Eyke ;
Cheng, Weiwei ;
Park, Sang-Hyeun .
MACHINE LEARNING, 2012, 89 (1-2) :123-156
[37]  
Goodman N., 2014, P COGN SCI SOC
[38]   Predictive coding [J].
Huang, Yanping ;
Rao, Rajesh P. N. .
WILEY INTERDISCIPLINARY REVIEWS-COGNITIVE SCIENCE, 2011, 2 (05) :580-593
[39]  
Kiebel Stefan J, 2009, Front Neuroinform, V3, P20, DOI 10.3389/neuro.11.020.2009
[40]  
Koller D., 2009, Probabilistic Graphical Models: Principles and Techniques