A simple computational algorithm of model-based choice preference

被引:0
|
作者
Asako Toyama
Kentaro Katahira
Hideki Ohira
机构
[1] Nagoya University,Department of Psychology, Graduate School of Environmental Studies
[2] Nagoya University,Department of Psychology, Graduate School of Informatics
[3] Japan Society for the Promotion of Science,undefined
来源
Cognitive, Affective, & Behavioral Neuroscience | 2017年 / 17卷
关键词
Computational model; Model-free; Model-based; Eligibility trace; Reinforcement learning;
D O I
暂无
中图分类号
学科分类号
摘要
A broadly used computational framework posits that two learning systems operate in parallel during the learning of choice preferences—namely, the model-free and model-based reinforcement-learning systems. In this study, we examined another possibility, through which model-free learning is the basic system and model-based information is its modulator. Accordingly, we proposed several modified versions of a temporal-difference learning model to explain the choice-learning process. Using the two-stage decision task developed by Daw, Gershman, Seymour, Dayan, and Dolan (2011), we compared their original computational model, which assumes a parallel learning process, and our proposed models, which assume a sequential learning process. Choice data from 23 participants showed a better fit with the proposed models. More specifically, the proposed eligibility adjustment model, which assumes that the environmental model can weight the degree of the eligibility trace, can explain choices better under both model-free and model-based controls and has a simpler computational algorithm than the original model. In addition, the forgetting learning model and its variation, which assume changes in the values of unchosen actions, substantially improved the fits to the data. Overall, we show that a hybrid computational model best fits the data. The parameters used in this model succeed in capturing individual tendencies with respect to both model use in learning and exploration behavior. This computational model provides novel insights into learning with interacting model-free and model-based components.
引用
收藏
页码:764 / 783
页数:19
相关论文
共 50 条
  • [21] An adaptive solar photovoltaic array using model-based reconfiguration algorithm
    Nguyen, Dzung
    Lehman, Brad
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2008, 55 (07) : 2644 - 2654
  • [22] Evidence for Model-Based Action Planning in a Sequential Finger Movement Task
    Fermin, Alan
    Yoshida, Takehiko
    Ito, Makoto
    Yoshimoto, Junichiro
    Doya, Kenji
    JOURNAL OF MOTOR BEHAVIOR, 2010, 42 (06) : 371 - 379
  • [23] The modulation of acute stress on model-free and model-based reinforcement learning in gambling disorder
    Wyckmans, Florent
    Banerjee, Nilosmita
    Saeremans, Melanie
    Otto, Ross
    Kornreich, Charles
    Vanderijst, Laetitia
    Gruson, Damien
    Carbone, Vincenzo
    Bechara, Antoine
    Buchanan, Tony
    Noel, Xavier
    JOURNAL OF BEHAVIORAL ADDICTIONS, 2022, 11 (03) : 831 - 844
  • [24] Investigation of Autonomic Dysfunction in Alzheimer's Disease-A Computational Model-Based Approach
    Nair, Sajitha Somasundaran
    Govindankutty, Mini Maniyelil
    Balakrishnan, Minimol
    Prasad, Krishna
    Sathyaprabha, Talakad N.
    Udupa, Kaviraja
    BRAIN SCIENCES, 2023, 13 (09)
  • [25] Dopamine enhances model-free credit assignment through boosting of retrospective model-based inference
    Deserno, Lorenz
    Moran, Rani
    Michely, Jochen
    Lee, Ying
    Dayan, Peter
    Dolan, Raymond J.
    ELIFE, 2021, 10
  • [26] Combining Model-Based and Model-Free Reinforcement Learning Policies for More Efficient Sepsis Treatment
    Liu, Xiangyu
    Yu, Chao
    Huang, Qikai
    Wang, Luhao
    Wu, Jianfeng
    Guan, Xiangdong
    BIOINFORMATICS RESEARCH AND APPLICATIONS, ISBRA 2021, 2021, 13064 : 105 - 117
  • [27] Model-based average reward reinforcement learning
    Tadepalli, P
    Ok, D
    ARTIFICIAL INTELLIGENCE, 1998, 100 (1-2) : 177 - 224
  • [28] Asynchronous Methods for Model-Based Reinforcement Learning
    Zhang, Yunzhi
    Clavera, Ignasi
    Tsai, Boren
    Abbeel, Pieter
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [29] The involvement of model-based but not model-free learning signals during observational reward learning in the absence of choice
    Dunne, Simon
    D'Souza, Arun
    O'Doherty, John P.
    JOURNAL OF NEUROPHYSIOLOGY, 2016, 115 (06) : 3195 - 3203
  • [30] A New Semi-explicit Model-based Integration Algorithm with High Accuracy
    Fu B.
    Zhang F.
    Chen J.
    Tongji Daxue Xuebao/Journal of Tongji University, 2023, 51 (05): : 738 - 746