| Literature DB >> 22972994 |
Michael S Landy1, Julia Trommershäuser, Nathaniel D Daw.
Abstract
Humans take into account their own movement variability as well as potential consequences of different movement outcomes in planning movement trajectories. When variability increases, planned movements are altered so as to optimize expected consequences of the movement. Past research has focused on the steady-state responses to changing conditions of movement under risk. Here, we study the dynamics of such strategy adjustment in a visuomotor decision task in which subjects reach toward a display with regions that lead to rewards and penalties, under conditions of changing uncertainty. In typical reinforcement learning tasks, subjects should base subsequent strategy by computing an estimate of the mean outcome (e.g., reward) in recent trials. In contrast, in our task, strategy should be based on a dynamic estimate of recent outcome uncertainty (i.e., squared error). We find that subjects respond to increased movement uncertainty by aiming movements more conservatively with respect to penalty regions, and that the estimate of uncertainty they use is well characterized by a weighted average of recent squared errors, with higher weights given to more recent trials.Entities:
Mesh:
Year: 2012 PMID: 22972994 PMCID: PMC3477850 DOI: 10.1523/JNEUROSCI.6160-11.2012
Source DB: PubMed Journal: J Neurosci ISSN: 0270-6474 Impact factor: 6.167