A simple computational algorithm of model-based choice preference

被引:0
|
作者
Asako Toyama
Kentaro Katahira
Hideki Ohira
机构
[1] Nagoya University,Department of Psychology, Graduate School of Environmental Studies
[2] Nagoya University,Department of Psychology, Graduate School of Informatics
[3] Japan Society for the Promotion of Science,undefined
来源
Cognitive, Affective, & Behavioral Neuroscience | 2017年 / 17卷
关键词
Computational model; Model-free; Model-based; Eligibility trace; Reinforcement learning;
D O I
暂无
中图分类号
学科分类号
摘要
A broadly used computational framework posits that two learning systems operate in parallel during the learning of choice preferences—namely, the model-free and model-based reinforcement-learning systems. In this study, we examined another possibility, through which model-free learning is the basic system and model-based information is its modulator. Accordingly, we proposed several modified versions of a temporal-difference learning model to explain the choice-learning process. Using the two-stage decision task developed by Daw, Gershman, Seymour, Dayan, and Dolan (2011), we compared their original computational model, which assumes a parallel learning process, and our proposed models, which assume a sequential learning process. Choice data from 23 participants showed a better fit with the proposed models. More specifically, the proposed eligibility adjustment model, which assumes that the environmental model can weight the degree of the eligibility trace, can explain choices better under both model-free and model-based controls and has a simpler computational algorithm than the original model. In addition, the forgetting learning model and its variation, which assume changes in the values of unchosen actions, substantially improved the fits to the data. Overall, we show that a hybrid computational model best fits the data. The parameters used in this model succeed in capturing individual tendencies with respect to both model use in learning and exploration behavior. This computational model provides novel insights into learning with interacting model-free and model-based components.
引用
收藏
页码:764 / 783
页数:19
相关论文
共 50 条
  • [1] A simple computational algorithm of model-based choice preference
    Toyama, Asako
    Katahira, Kentaro
    Ohira, Hideki
    COGNITIVE AFFECTIVE & BEHAVIORAL NEUROSCIENCE, 2017, 17 (04) : 764 - 783
  • [2] Computational and behavioral markers of model-based decision making in childhood
    Smid, Claire R.
    Kool, Wouter
    Hauser, Tobias U.
    Steinbeis, Nikolaus
    DEVELOPMENTAL SCIENCE, 2023, 26 (02)
  • [3] Dopamine selectively remediates 'model-based' reward learning: a computational approach
    Sharp, Madeleine E.
    Foerde, Karin
    Daw, Nathaniel D.
    Shohamy, Daphna
    BRAIN, 2016, 139 : 355 - 364
  • [4] Model-based reinforcement learning: a computational model and an fMRI study
    Yoshida, W
    Ishii, S
    NEUROCOMPUTING, 2005, 63 : 253 - 269
  • [5] Is My Interface Beautiful?-A Computational Model-Based Approach
    Maity, Ranjan
    Bhattacharya, Samit
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2019, 6 (01): : 149 - 161
  • [6] Impact of Arrhythmia on Myocardial Perfusion: A Computational Model-Based Study
    Ge, Xinyang
    Simakov, Sergey
    Liu, Youjun
    Liang, Fuyou
    MATHEMATICS, 2021, 9 (17)
  • [7] Model-based learning protects against forming habits
    Gillan, Claire M.
    Otto, A. Ross
    Phelps, Elizabeth A.
    Daw, Nathaniel D.
    COGNITIVE AFFECTIVE & BEHAVIORAL NEUROSCIENCE, 2015, 15 (03) : 523 - 536
  • [8] Model-based learning protects against forming habits
    Claire M. Gillan
    A. Ross Otto
    Elizabeth A. Phelps
    Nathaniel D. Daw
    Cognitive, Affective, & Behavioral Neuroscience, 2015, 15 : 523 - 536
  • [9] Model-Based and Model-Free Replay Mechanisms for Reinforcement Learning in Neurorobotics
    Massi, Elisa
    Barthelemy, Jeanne
    Mailly, Juliane
    Dromnelle, Remi
    Canitrot, Julien
    Poniatowski, Esther
    Girard, Benoit
    Khamassi, Mehdi
    FRONTIERS IN NEUROROBOTICS, 2022, 16
  • [10] Comparing Model-free and Model-based Algorithms for Offline Reinforcement Learning
    Swazinna, Phillip
    Udluft, Steffen
    Hein, Daniel
    Runkler, Thomas
    IFAC PAPERSONLINE, 2022, 55 (15): : 19 - 26