Literature DB >> 35662490

The computational roots of positivity and confirmation biases in reinforcement learning.

Stefano Palminteri1, Maël Lebreton2.   

Abstract

Humans do not integrate new information objectively: outcomes carrying a positive affective value and evidence confirming one's own prior belief are overweighed. Until recently, theoretical and empirical accounts of the positivity and confirmation biases assumed them to be specific to 'high-level' belief updates. We present evidence against this account. Learning rates in reinforcement learning (RL) tasks, estimated across different contexts and species, generally present the same characteristic asymmetry, suggesting that belief and value updating processes share key computational principles and distortions. This bias generates over-optimistic expectations about the probability of making the right choices and, consequently, generates over-optimistic reward expectations. We discuss the normative and neurobiological roots of these RL biases and their position within the greater picture of behavioral decision-making theories.
Copyright © 2022 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  confirmation; decision; gain; learning; loss; update

Mesh:

Year:  2022        PMID: 35662490     DOI: 10.1016/j.tics.2022.04.005

Source DB:  PubMed          Journal:  Trends Cogn Sci        ISSN: 1364-6613            Impact factor:   24.482


  1 in total

1.  Choice perseverance underlies pursuing a hard-to-get target in an avatar choice task.

Authors:  Michiyo Sugawara; Kentaro Katahira
Journal:  Front Psychol       Date:  2022-09-06
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.