Literature DB >> 28573384

A simple computational algorithm of model-based choice preference.

Asako Toyama1,2,3, Kentaro Katahira4,5, Hideki Ohira4,5.   

Abstract

A broadly used computational framework posits that two learning systems operate in parallel during the learning of choice preferences-namely, the model-free and model-based reinforcement-learning systems. In this study, we examined another possibility, through which model-free learning is the basic system and model-based information is its modulator. Accordingly, we proposed several modified versions of a temporal-difference learning model to explain the choice-learning process. Using the two-stage decision task developed by Daw, Gershman, Seymour, Dayan, and Dolan (2011), we compared their original computational model, which assumes a parallel learning process, and our proposed models, which assume a sequential learning process. Choice data from 23 participants showed a better fit with the proposed models. More specifically, the proposed eligibility adjustment model, which assumes that the environmental model can weight the degree of the eligibility trace, can explain choices better under both model-free and model-based controls and has a simpler computational algorithm than the original model. In addition, the forgetting learning model and its variation, which assume changes in the values of unchosen actions, substantially improved the fits to the data. Overall, we show that a hybrid computational model best fits the data. The parameters used in this model succeed in capturing individual tendencies with respect to both model use in learning and exploration behavior. This computational model provides novel insights into learning with interacting model-free and model-based components.

Entities:  

Keywords:  Computational model; Eligibility trace; Model-based; Model-free; Reinforcement learning

Mesh:

Year:  2017        PMID: 28573384     DOI: 10.3758/s13415-017-0511-2

Source DB:  PubMed          Journal:  Cogn Affect Behav Neurosci        ISSN: 1530-7026            Impact factor:   3.282


  27 in total

1.  Prefrontal cortex and decision making in a mixed-strategy game.

Authors:  Dominic J Barraclough; Michelle L Conroy; Daeyeol Lee
Journal:  Nat Neurosci       Date:  2004-03-07       Impact factor: 24.884

2.  Reinforcement learning through modulation of spike-timing-dependent synaptic plasticity.

Authors:  Răzvan V Florian
Journal:  Neural Comput       Date:  2007-06       Impact factor: 2.026

3.  Model-based and model-free decisions in alcohol dependence.

Authors:  Miriam Sebold; Lorenz Deserno; Stephan Nebe; Stefan Nebe; Daniel J Schad; Maria Garbusow; Claudia Hägele; Jürgen Keller; Elisabeth Jünger; Norbert Kathmann; Michael N Smolka; Michael Smolka; Michael A Rapp; Florian Schlagenhauf; Andreas Heinz; Quentin J M Huys
Journal:  Neuropsychobiology       Date:  2014-10-30       Impact factor: 2.328

4.  Rostrolateral prefrontal cortex and individual differences in uncertainty-driven exploration.

Authors:  David Badre; Bradley B Doll; Nicole M Long; Michael J Frank
Journal:  Neuron       Date:  2012-02-09       Impact factor: 17.173

5.  Short-term memory traces for action bias in human reinforcement learning.

Authors:  Rafal Bogacz; Samuel M McClure; Jian Li; Jonathan D Cohen; P Read Montague
Journal:  Brain Res       Date:  2007-03-24       Impact factor: 3.252

Review 6.  A unified framework for addiction: vulnerabilities in the decision process.

Authors:  A David Redish; Steve Jensen; Adam Johnson
Journal:  Behav Brain Sci       Date:  2008-08       Impact factor: 21.357

7.  Model-based learning protects against forming habits.

Authors:  Claire M Gillan; A Ross Otto; Elizabeth A Phelps; Nathaniel D Daw
Journal:  Cogn Affect Behav Neurosci       Date:  2015-09       Impact factor: 3.282

8.  Disruption of dorsolateral prefrontal cortex decreases model-based in favor of model-free control in humans.

Authors:  Peter Smittenaar; Thomas H B FitzGerald; Vincenzo Romei; Nicholas D Wright; Raymond J Dolan
Journal:  Neuron       Date:  2013-10-24       Impact factor: 17.173

Review 9.  Goals and habits in the brain.

Authors:  Ray J Dolan; Peter Dayan
Journal:  Neuron       Date:  2013-10-16       Impact factor: 17.173

10.  Goal-Directed Decision Making with Spiking Neurons.

Authors:  Johannes Friedrich; Máté Lengyel
Journal:  J Neurosci       Date:  2016-02-03       Impact factor: 6.167

View more
  6 in total

1.  Reinforcement learning with associative or discriminative generalization across states and actions: fMRI at 3 T and 7 T.

Authors:  Jaron T Colas; Neil M Dundon; Raphael T Gerraty; Natalie M Saragosa-Harris; Karol P Szymula; Koranis Tanwisuth; J Michael Tyszka; Camilla van Geen; Harang Ju; Arthur W Toga; Joshua I Gold; Dani S Bassett; Catherine A Hartley; Daphna Shohamy; Scott T Grafton; John P O'Doherty
Journal:  Hum Brain Mapp       Date:  2022-07-21       Impact factor: 5.399

2.  Cardiac Cycle Affects the Asymmetric Value Updating in Instrumental Reward Learning.

Authors:  Kenta Kimura; Noriaki Kanayama; Asako Toyama; Kentaro Katahira
Journal:  Front Neurosci       Date:  2022-06-02       Impact factor: 5.152

3.  Active inference and the two-step task.

Authors:  Sam Gijsen; Miro Grundei; Felix Blankenburg
Journal:  Sci Rep       Date:  2022-10-21       Impact factor: 4.996

4.  Revisiting the importance of model fitting for model-based fMRI: It does matter in computational psychiatry.

Authors:  Kentaro Katahira; Asako Toyama
Journal:  PLoS Comput Biol       Date:  2021-02-09       Impact factor: 4.475

5.  Model-based learning retrospectively updates model-free values.

Authors:  Max Doody; Maaike M H Van Swieten; Sanjay G Manohar
Journal:  Sci Rep       Date:  2022-02-11       Impact factor: 4.996

6.  Psychiatric symptoms influence reward-seeking and loss-avoidance decision-making through common and distinct computational processes.

Authors:  Yuichi Yamashita; Kentaro Katahira; Shinsuke Suzuki
Journal:  Psychiatry Clin Neurosci       Date:  2021-07-17       Impact factor: 5.188

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.