Literature DB >> 30664916

Joint modeling of reaction times and choice improves parameter identifiability in reinforcement learning models.

Ian C Ballard1, Samuel M McClure2.   

Abstract

BACKGROUND: Reinforcement learning models provide excellent descriptions of learning in multiple species across a variety of tasks. Many researchers are interested in relating parameters of reinforcement learning models to neural measures, psychological variables or experimental manipulations. We demonstrate that parameter identification is difficult because a range of parameter values provide approximately equal quality fits to data. This identification problem has a large impact on power: we show that a researcher who wants to detect a medium sized correlation (r = .3) with 80% power between a variable and learning rate must collect 60% more subjects than specified by a typical power analysis in order to account for the noise introduced by model fitting. NEW
METHOD: We derive a Bayesian optimal model fitting technique that takes advantage of information contained in choices and reaction times to constrain parameter estimates.
RESULTS: We show using simulation and empirical data that this method substantially improves the ability to recover learning rates. COMPARISON WITH EXISTING
METHODS: We compare this method against the use of Bayesian priors. We show in simulations that the combined use of Bayesian priors and reaction times confers the highest parameter identifiability. However, in real data where the priors may have been misspecified, the use of Bayesian priors interferes with the ability of reaction time data to improve parameter identifiability.
CONCLUSIONS: We present a simple technique that takes advantage of readily available data to substantially improve the quality of inferences that can be drawn from parameters of reinforcement learning models.
Copyright © 2019 Elsevier B.V. All rights reserved.

Entities:  

Keywords:  Delay discounting; Intertemporal choice; Parameter estimation; Power; Q-learning; Reproducibility; Striatum

Mesh:

Year:  2019        PMID: 30664916      PMCID: PMC8930195          DOI: 10.1016/j.jneumeth.2019.01.006

Source DB:  PubMed          Journal:  J Neurosci Methods        ISSN: 0165-0270            Impact factor:   2.390


  40 in total

Review 1.  From reinforcement learning models to psychiatric and neurological disorders.

Authors:  Tiago V Maia; Michael J Frank
Journal:  Nat Neurosci       Date:  2011-02       Impact factor: 24.884

2.  Separate neural systems value immediate and delayed monetary rewards.

Authors:  Samuel M McClure; David I Laibson; George Loewenstein; Jonathan D Cohen
Journal:  Science       Date:  2004-10-15       Impact factor: 47.728

Review 3.  Prediction error in reinforcement learning: a meta-analysis of neuroimaging studies.

Authors:  Jane Garrison; Burak Erdeniz; John Done
Journal:  Neurosci Biobehav Rev       Date:  2013-04-06       Impact factor: 8.989

Review 4.  The relative merit of empirical priors in non-identifiable and sloppy models: Applications to models of learning and decision-making : Empirical priors.

Authors:  Mikhail S Spektor; David Kellen
Journal:  Psychon Bull Rev       Date:  2018-12

5.  Episodic memory encoding interferes with reward learning and decreases striatal prediction errors.

Authors:  G Elliott Wimmer; Erin Kendall Braun; Nathaniel D Daw; Daphna Shohamy
Journal:  J Neurosci       Date:  2014-11-05       Impact factor: 6.167

6.  Amygdala and Ventral Striatum Make Distinct Contributions to Reinforcement Learning.

Authors:  Vincent D Costa; Olga Dal Monte; Daniel R Lucas; Elisabeth A Murray; Bruno B Averbeck
Journal:  Neuron       Date:  2016-10-06       Impact factor: 17.173

7.  Learning the opportunity cost of time in a patch-foraging task.

Authors:  Sara M Constantino; Nathaniel D Daw
Journal:  Cogn Affect Behav Neurosci       Date:  2015-12       Impact factor: 3.282

8.  Reinforcement learning signals in the human striatum distinguish learners from nonlearners during reward-based decision making.

Authors:  Tom Schönberg; Nathaniel D Daw; Daphna Joel; John P O'Doherty
Journal:  J Neurosci       Date:  2007-11-21       Impact factor: 6.167

9.  Genetic triple dissociation reveals multiple roles for dopamine in reinforcement learning.

Authors:  Michael J Frank; Ahmed A Moustafa; Heather M Haughey; Tim Curran; Kent E Hutchison
Journal:  Proc Natl Acad Sci U S A       Date:  2007-10-03       Impact factor: 11.205

10.  Model-based choices involve prospective neural activity.

Authors:  Bradley B Doll; Katherine D Duncan; Dylan A Simon; Daphna Shohamy; Nathaniel D Daw
Journal:  Nat Neurosci       Date:  2015-03-23       Impact factor: 24.884

View more
  10 in total

Review 1.  Advances in modeling learning and decision-making in neuroscience.

Authors:  Anne G E Collins; Amitai Shenhav
Journal:  Neuropsychopharmacology       Date:  2021-08-27       Impact factor: 7.853

2.  Reinforcement learning with associative or discriminative generalization across states and actions: fMRI at 3 T and 7 T.

Authors:  Jaron T Colas; Neil M Dundon; Raphael T Gerraty; Natalie M Saragosa-Harris; Karol P Szymula; Koranis Tanwisuth; J Michael Tyszka; Camilla van Geen; Harang Ju; Arthur W Toga; Joshua I Gold; Dani S Bassett; Catherine A Hartley; Daphna Shohamy; Scott T Grafton; John P O'Doherty
Journal:  Hum Brain Mapp       Date:  2022-07-21       Impact factor: 5.399

3.  Reinforcement learning modeling reveals a reward-history-dependent strategy underlying reversal learning in squirrel monkeys.

Authors:  Bilal A Bari; Megan J Moerke; Hank P Jedema; Devin P Effinger; Jeremiah Y Cohen; Charles W Bradberry
Journal:  Behav Neurosci       Date:  2021-09-27       Impact factor: 1.912

4.  Improving the reliability of model-based decision-making estimates in the two-stage decision task with reaction-times and drift-diffusion modeling.

Authors:  Nitzan Shahar; Tobias U Hauser; Michael Moutoussis; Rani Moran; Mehdi Keramati; Raymond J Dolan
Journal:  PLoS Comput Biol       Date:  2019-02-13       Impact factor: 4.779

5.  The drift diffusion model as the choice rule in inter-temporal and risky choice: A case study in medial orbitofrontal cortex lesion patients and controls.

Authors:  Jan Peters; Mark D'Esposito
Journal:  PLoS Comput Biol       Date:  2020-04-20       Impact factor: 4.475

6.  Ten simple rules for the computational modeling of behavioral data.

Authors:  Robert C Wilson; Anne Ge Collins
Journal:  Elife       Date:  2019-11-26       Impact factor: 8.140

7.  Reliability assessment of temporal discounting measures in virtual reality environments.

Authors:  Luca R Bruder; Lisa Scharer; Jan Peters
Journal:  Sci Rep       Date:  2021-03-29       Impact factor: 4.379

8.  Simultaneous Hierarchical Bayesian Parameter Estimation for Reinforcement Learning and Drift Diffusion Models: a Tutorial and Links to Neural Data.

Authors:  Mads L Pedersen; Michael J Frank
Journal:  Comput Brain Behav       Date:  2020-05-26

Review 9.  Modeling the influence of working memory, reinforcement, and action uncertainty on reaction time and choice during instrumental learning.

Authors:  Samuel D McDougle; Anne G E Collins
Journal:  Psychon Bull Rev       Date:  2021-02

10.  Response time models separate single- and dual-process accounts of memory-based decisions.

Authors:  Peter M Kraemer; Laura Fontanesi; Mikhail S Spektor; Sebastian Gluth
Journal:  Psychon Bull Rev       Date:  2021-02
  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.