Simulating Active Inference Processes by Message Passing

被引:29
作者
van de laar, Thijs W. [1 ]
de Vries, Bert [1 ,2 ]
机构
[1] Eindhoven Univ Technol, Dept Elect Engn, Eindhoven, Netherlands
[2] GN Hearing Benelux BV, Eindhoven, Netherlands
关键词
active inference; free-energy principle; message passing; state-space models; Forney-style factor graphs; FREE-ENERGY PRINCIPLE; FACTOR GRAPH APPROACH;
D O I
10.3389/frobt.2019.00020
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
The free energy principle (FEP) offers a variational calculus-based description for how biological agents persevere through interactions with their environment. Active inference (Al) is a corollary of the FEP, which states that biological agents act to fulfill prior beliefs about preferred future observations (target priors). Purposeful behavior then results from variational free energy minimization with respect to a generative model of the environment with included target priors. However, manual derivations for free energy minimizing algorithms on custom dynamic models can become tedious and error-prone. While probabilistic programming (PP) techniques enable automatic derivation of inference algorithms on free-form models, full automation of Al requires specialized tools for inference on dynamic models, together with the description of an experimental protocol that governs the interaction between the agent and its simulated environment. The contributions of the present paper are two-fold. Firstly, we illustrate how Al can be automated with the use of ForneyLab, a recent PP toolbox that specializes in variational inference on flexibly definable dynamic models. More specifically, we describe Al agents in a dynamic environment as probabilistic state space models (SSM) and perform inference for perception and control in these agents by message passing on a factor graph representation of the SSM. Secondly, we propose a formal experimental protocol for simulated Al. We exemplify how this protocol leads to goal-directed behavior for flexibly definable Al agents in two classical RL examples, namely the Bayesian thermostat and the mountain car parking problems.
引用
收藏
页数:15
相关论文
共 58 条
[1]  
[Anonymous], 2012, THESIS ETH ZURICH
[2]  
Attias Hagai., 1999, NIPS, V12
[3]   Recent Advances in Hierarchical Reinforcement Learning [J].
Andrew G. Barto ;
Sridhar Mahadevan .
Discrete Event Dynamic Systems, 2003, 13 (4) :341-379
[4]   A DYNAMICAL-SYSTEMS PERSPECTIVE ON AGENT ENVIRONMENT INTERACTION [J].
BEER, RD .
ARTIFICIAL INTELLIGENCE, 1995, 72 (1-2) :173-215
[5]   Julia: A Fresh Approach to Numerical Computing [J].
Bezanson, Jeff ;
Edelman, Alan ;
Karpinski, Stefan ;
Shah, Viral B. .
SIAM REVIEW, 2017, 59 (01) :65-98
[6]  
Bishop C. M., 2006, PATTERN RECOGNITION, DOI DOI 10.1117/1.2819119
[7]   The free energy principle for action and perception: A mathematical review [J].
Buckley, Christopher L. ;
Kim, Chang Sub ;
McGregor, Simon ;
Seth, Anil K. .
JOURNAL OF MATHEMATICAL PSYCHOLOGY, 2017, 81 :55-79
[8]  
Camacho Eduardo F, 2013, Model predictive control
[9]   Stan: A Probabilistic Programming Language [J].
Carpenter, Bob ;
Gelman, Andrew ;
Hoffman, Matthew D. ;
Lee, Daniel ;
Goodrich, Ben ;
Betancourt, Michael ;
Brubaker, Marcus A. ;
Guo, Jiqiang ;
Li, Peter ;
Riddell, Allen .
JOURNAL OF STATISTICAL SOFTWARE, 2017, 76 (01) :1-29
[10]  
Conant R. C., 1970, International Journal of Systems Science, V1, P89, DOI [10.1080/00207727008920220, DOI 10.1080/00207727008920220]