Literature DB >> 15169045

Learning in neural networks by reinforcement of irregular spiking.

Xiaohui Xie1, H Sebastian Seung.   

Abstract

Artificial neural networks are often trained by using the back propagation algorithm to compute the gradient of an objective function with respect to the synaptic strengths. For a biological neural network, such a gradient computation would be difficult to implement, because of the complex dynamics of intrinsic and synaptic conductances in neurons. Here we show that irregular spiking similar to that observed in biological neurons could be used as the basis for a learning rule that calculates a stochastic approximation to the gradient. The learning rule is derived based on a special class of model networks in which neurons fire spike trains with Poisson statistics. The learning is compatible with forms of synaptic dynamics such as short-term facilitation and depression. By correlating the fluctuations in irregular spiking with a reward signal, the learning rule performs stochastic gradient ascent on the expected reward. It is applied to two examples, learning the XOR computation and learning direction selectivity using depressing synapses. We also show in simulation that the learning rule is applicable to a network of noisy integrate-and-fire neurons.

Entities:  

Mesh:

Year:  2004        PMID: 15169045     DOI: 10.1103/PhysRevE.69.041909

Source DB:  PubMed          Journal:  Phys Rev E Stat Nonlin Soft Matter Phys        ISSN: 1539-3755


  27 in total

1.  Dynamic afferent synapses to decision-making networks improve performance in tasks requiring stimulus associations and discriminations.

Authors:  Mark A Bourjaily; Paul Miller
Journal:  J Neurophysiol       Date:  2012-03-28       Impact factor: 2.714

2.  Computational models of reinforcement learning: the role of dopamine as a reward signal.

Authors:  R D Samson; M J Frank; Jean-Marc Fellous
Journal:  Cogn Neurodyn       Date:  2010-03-21       Impact factor: 5.082

3.  Cerebellar learning using perturbations.

Authors:  Guy Bouvier; Johnatan Aljadeff; Claudia Clopath; Célian Bimbard; Jonas Ranft; Antonin Blot; Jean-Pierre Nadal; Nicolas Brunel; Vincent Hakim; Boris Barbour
Journal:  Elife       Date:  2018-11-12       Impact factor: 8.140

4.  A reward-modulated hebbian learning rule can explain experimentally observed network reorganization in a brain control task.

Authors:  Robert Legenstein; Steven M Chase; Andrew B Schwartz; Wolfgang Maass
Journal:  J Neurosci       Date:  2010-06-23       Impact factor: 6.167

5.  Spike-based reinforcement learning in continuous state and action space: when policy gradient methods fail.

Authors:  Eleni Vasilaki; Nicolas Frémaux; Robert Urbanczik; Walter Senn; Wulfram Gerstner
Journal:  PLoS Comput Biol       Date:  2009-12-04       Impact factor: 4.475

6.  Synaptic theory of replicator-like melioration.

Authors:  Yonatan Loewenstein
Journal:  Front Comput Neurosci       Date:  2010-06-17       Impact factor: 2.380

7.  Enabling functional neural circuit simulations with distributed computing of neuromodulated plasticity.

Authors:  Wiebke Potjans; Abigail Morrison; Markus Diesmann
Journal:  Front Comput Neurosci       Date:  2010-11-23       Impact factor: 2.380

8.  Reinforcement learning using a continuous time actor-critic framework with spiking neurons.

Authors:  Nicolas Frémaux; Henning Sprekeler; Wulfram Gerstner
Journal:  PLoS Comput Biol       Date:  2013-04-11       Impact factor: 4.475

9.  Functional network reorganization in motor cortex can be explained by reward-modulated Hebbian learning.

Authors:  Robert Legenstein; Steven M Chase; Andrew B Schwartz; Wolfgang Maass
Journal:  Adv Neural Inf Process Syst       Date:  2009

10.  A Dynamic Connectome Supports the Emergence of Stable Computational Function of Neural Circuits through Reward-Based Learning.

Authors:  David Kappel; Robert Legenstein; Stefan Habenschuss; Michael Hsieh; Wolfgang Maass
Journal:  eNeuro       Date:  2018-04-24
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.