Literature DB >> 32065253

Modeling awake hippocampal reactivations with model-based bidirectional search.

Mehdi Khamassi1, Benoît Girard2.   

Abstract

Hippocampal offline reactivations during reward-based learning, usually categorized as replay events, have been found to be important for performance improvement over time and for memory consolidation. Recent computational work has linked these phenomena to the need to transform reward information into state-action values for decision making and to propagate it to all relevant states of the environment. Nevertheless, it is still unclear whether an integrated reinforcement learning mechanism could account for the variety of awake hippocampal reactivations, including variety in order (forward and reverse reactivated trajectories) and variety in the location where they occur (reward site or decision-point). Here, we present a model-based bidirectional search model which accounts for a variety of hippocampal reactivations. The model combines forward trajectory sampling from current position and backward sampling through prioritized sweeping from states associated with large reward prediction errors until the two trajectories connect. This is repeated until stabilization of state-action values (convergence), which could explain why hippocampal reactivations drastically diminish when the animal's performance stabilizes. Simulations in a multiple T-maze task show that forward reactivations are prominently found at decision-points while backward reactivations are exclusively generated at reward sites. Finally, the model can generate imaginary trajectories that are not allowed to the agent during task performance. We raise some experimental predictions and implications for future studies of the role of the hippocampo-prefronto-striatal network in learning.

Keywords:  Computational neuroscience; Hippocampal replay; Navigation; Reinforcement learning

Mesh:

Year:  2020        PMID: 32065253     DOI: 10.1007/s00422-020-00817-x

Source DB:  PubMed          Journal:  Biol Cybern        ISSN: 0340-1200            Impact factor:   2.086


  3 in total

1.  Model-Based and Model-Free Replay Mechanisms for Reinforcement Learning in Neurorobotics.

Authors:  Elisa Massi; Jeanne Barthélemy; Juliane Mailly; Rémi Dromnelle; Julien Canitrot; Esther Poniatowski; Benoît Girard; Mehdi Khamassi
Journal:  Front Neurorobot       Date:  2022-06-24       Impact factor: 3.493

2.  Reward prediction errors drive declarative learning irrespective of agency.

Authors:  Kate Ergo; Luna De Vilder; Esther De Loof; Tom Verguts
Journal:  Psychon Bull Rev       Date:  2021-06-15

Review 3.  From spatial navigation via visual construction to episodic memory and imagination.

Authors:  Michael A Arbib
Journal:  Biol Cybern       Date:  2020-04-13       Impact factor: 2.086

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.