The drift diffusion model as the choice rule in reinforcement learning

被引:142
|
作者
Pedersen, Mads Lund [1 ,2 ]
Frank, Michael J. [3 ]
Biele, Guido [1 ,4 ]
机构
[1] Univ Oslo, Dept Psychol, Oslo, Norway
[2] Natl Hosp Norway, Intervent Ctr, Oslo Univ Hosp, Oslo, Norway
[3] Brown Univ, Brown Inst Brain Sci, Dept Cognit Linguist & Psychol Sci, Providence, RI 02912 USA
[4] Norwegian Inst Publ Hlth, Oslo, Norway
基金
美国国家科学基金会;
关键词
Decision making; Reinforcement learning; Bayesian modeling; Mathematical models; SPEED-ACCURACY TRADEOFF; DECISION FIELD-THEORY; PERCEPTUAL DECISION; COMPUTATIONAL MODELS; SUBTHALAMIC NUCLEUS; WORKING-MEMORY; DOPAMINE; BRAIN; PERFORMANCE; ATTENTION;
D O I
10.3758/s13423-016-1199-y
中图分类号
B841 [心理学研究方法];
学科分类号
040201 ;
摘要
Current reinforcement-learning models often assume simplified decision processes that do not fully reflect the dynamic complexities of choice processes. Conversely, sequential-sampling models of decision making account for both choice accuracy and response time, but assume that decisions are based on static decision values. To combine these two computational models of decision making and learning, we implemented reinforcement-learning models in which the drift diffusion model describes the choice process, thereby capturing both within- and across-trial dynamics. To exemplify the utility of this approach, we quantitatively fit data from a common reinforcement-learning paradigm using hierarchical Bayesian parameter estimation, and compared model variants to determine whether they could capture the effects of stimulant medication in adult patients with attention-deficit hyperactivity disorder (ADHD). The model with the best relative fit provided a good description of the learning process, choices, and response times. A parameter recovery experiment showed that the hierarchical Bayesian modeling approach enabled accurate estimation of the model parameters. The model approach described here, using simultaneous estimation of reinforcement-learning and drift diffusion model parameters, shows promise for revealing new insights into the cognitive and neural mechanisms of learning and decision making, as well as the alteration of such processes in clinical groups.
引用
收藏
页码:1234 / 1251
页数:18
相关论文
共 50 条
  • [1] The drift diffusion model as the choice rule in reinforcement learning
    Mads Lund Pedersen
    Michael J. Frank
    Guido Biele
    Psychonomic Bulletin & Review, 2017, 24 : 1234 - 1251
  • [2] Reinforcement learning in women remitted from anorexia nervosa: Preliminary examination with a hybrid reinforcement learning/drift diffusion model
    Wierenga, Christina E.
    Bischoff-Grethe, Amanda
    Brown, Carina S.
    Brown, Gregory G.
    JOURNAL OF THE INTERNATIONAL NEUROPSYCHOLOGICAL SOCIETY, 2025,
  • [3] The model of the reward choice basing on the theory of reinforcement learning
    Smirnitskaya, I. A.
    Frolov, A. A.
    Merzhanova, G. Kh.
    ZHURNAL VYSSHEI NERVNOI DEYATELNOSTI IMENI I P PAVLOVA, 2007, 57 (02) : 133 - 143
  • [4] A model of reward choice based on the theory of reinforcement learning
    Smirnitskaya I.A.
    Frolov A.A.
    Merzhanova G.Kh.
    Neuroscience and Behavioral Physiology, 2008, 38 (3) : 269 - 278
  • [5] An attentional drift diffusion model over binary-attribute choice
    Fisher, Geoffrey
    COGNITION, 2017, 168 : 34 - 45
  • [6] Drift diffusion model of reward and punishment learning in schizophrenia: Modeling and experimental data
    Moustafa, Ahmed A.
    Keri, Szabolcs
    Somlai, Zsuzsanna
    Balsdon, Tarryn
    Frydecka, Dorota
    Misiak, Blazej
    White, Corey
    BEHAVIOURAL BRAIN RESEARCH, 2015, 291 : 147 - 154
  • [7] A Reinforcement Learning Approach for Graph Rule Learning
    Mai, Zhenzhen
    Wang, Wenjun
    Liu, Xueli
    Feng, Xiaoyang
    Wang, Jun
    Fu, Wenzhi
    BIG DATA MINING AND ANALYTICS, 2025, 8 (01): : 31 - 44
  • [8] The drift diffusion model as the choice rule in inter-temporal and risky choice: A case study in medial orbitofrontal cortex lesion patients and controls
    Peters, Jan
    D'Esposito, Mark
    PLOS COMPUTATIONAL BIOLOGY, 2020, 16 (04)
  • [9] Learning to use working memory: a reinforcement learning gating model of rule acquisition in rats
    Lloyd, Kevin
    Becker, Nadine
    Jones, Matthew W.
    Bogacz, Rafal
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2012, 6
  • [10] Combining Choice and Response Time Data: A Drift-Diffusion Model of Mobile Advertisements
    Chiong, Khai Xiang
    Shum, Matthew
    Webb, Ryan
    Chen, Richard
    MANAGEMENT SCIENCE, 2024, 70 (02) : 1238 - 1257