| Literature DB >> 33839955 |
E Pomarol-Clotet1,2, J Radua1,2,3,4,5, A Santo-Angles6,7,8,9, P Fuentes-Claramonte1,2, I Argila-Plaza1, M Guardiola-Ripoll1,2, C Almodóvar-Payá1,2, J Munuera10, P J McKenna1,2.
Abstract
Reward prediction error, the difference between the expected and obtained reward, is known to act as a reinforcement learning neural signal. In the current study, we propose a model fitting approach that combines behavioral and neural data to fit computational models of reinforcement learning. Briefly, we penalized subject-specific fitted parameters that moved away too far from the group median, except when that deviation led to an improvement in the model's fit to neural responses. By means of a probabilistic monetary learning task and fMRI, we compared our approach with standard model fitting methods. Q-learning outperformed actor-critic at both behavioral and neural level, although the inclusion of neuroimaging data into model fitting improved the fit of actor-critic models. We observed both action-value and state-value prediction error signals in the striatum, while standard model fitting approaches failed to capture state-value signals. Finally, left ventral striatum correlated with reward prediction error while right ventral striatum with fictive prediction error, suggesting a functional hemispheric asymmetry regarding prediction-error driven learning.Entities:
Keywords: Counterfactual; Fictive prediction error; Model fitting; Reward prediction error; fMRI
Mesh:
Year: 2021 PMID: 33839955 DOI: 10.1007/s00429-021-02270-3
Source DB: PubMed Journal: Brain Struct Funct ISSN: 1863-2653 Impact factor: 3.270