Literature DB >> 19196231

A spiking neural network model of an actor-critic learning agent.

Wiebke Potjans1, Abigail Morrison, Markus Diesmann.   

Abstract

The ability to adapt behavior to maximize reward as a result of interactions with the environment is crucial for the survival of any higher organism. In the framework of reinforcement learning, temporal-difference learning algorithms provide an effective strategy for such goal-directed adaptation, but it is unclear to what extent these algorithms are compatible with neural computation. In this article, we present a spiking neural network model that implements actor-critic temporal-difference learning by combining local plasticity rules with a global reward signal. The network is capable of solving a nontrivial gridworld task with sparse rewards. We derive a quantitative mapping of plasticity parameters and synaptic weights to the corresponding variables in the standard algorithmic formulation and demonstrate that the network learns with a similar speed to its discrete time counterpart and attains the same equilibrium performance.

Entities:  

Mesh:

Year:  2009        PMID: 19196231     DOI: 10.1162/neco.2008.08-07-593

Source DB:  PubMed          Journal:  Neural Comput        ISSN: 0899-7667            Impact factor:   2.026


  32 in total

1.  Alternative time representation in dopamine models.

Authors:  François Rivest; John F Kalaska; Yoshua Bengio
Journal:  J Comput Neurosci       Date:  2009-10-22       Impact factor: 1.621

Review 2.  Building functional networks of spiking model neurons.

Authors:  L F Abbott; Brian DePasquale; Raoul-Martin Memmesheimer
Journal:  Nat Neurosci       Date:  2016-03       Impact factor: 24.884

3.  Computational models of reinforcement learning: the role of dopamine as a reward signal.

Authors:  R D Samson; M J Frank; Jean-Marc Fellous
Journal:  Cogn Neurodyn       Date:  2010-03-21       Impact factor: 5.082

4.  Neuromodulation of STDP through short-term changes in firing causality.

Authors:  Simon M Vogt; Ulrich G Hofmann
Journal:  Cogn Neurodyn       Date:  2012-07-25       Impact factor: 5.082

5.  Reinforcement learning with Marr.

Authors:  Yael Niv; Angela Langdon
Journal:  Curr Opin Behav Sci       Date:  2016-10

6.  Spike-based decision learning of Nash equilibria in two-player games.

Authors:  Johannes Friedrich; Walter Senn
Journal:  PLoS Comput Biol       Date:  2012-09-27       Impact factor: 4.475

7.  Reinforcement learning of two-joint virtual arm reaching in a computer model of sensorimotor cortex.

Authors:  Samuel A Neymotin; George L Chadderdon; Cliff C Kerr; Joseph T Francis; William W Lytton
Journal:  Neural Comput       Date:  2013-09-18       Impact factor: 2.026

8.  Spike-based reinforcement learning in continuous state and action space: when policy gradient methods fail.

Authors:  Eleni Vasilaki; Nicolas Frémaux; Robert Urbanczik; Walter Senn; Wulfram Gerstner
Journal:  PLoS Comput Biol       Date:  2009-12-04       Impact factor: 4.475

9.  Enabling functional neural circuit simulations with distributed computing of neuromodulated plasticity.

Authors:  Wiebke Potjans; Abigail Morrison; Markus Diesmann
Journal:  Front Comput Neurosci       Date:  2010-11-23       Impact factor: 2.380

10.  Reinforcement learning using a continuous time actor-critic framework with spiking neurons.

Authors:  Nicolas Frémaux; Henning Sprekeler; Wulfram Gerstner
Journal:  PLoS Comput Biol       Date:  2013-04-11       Impact factor: 4.475

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.