Literature DB >> 18249797

New results on recurrent network training: unifying the algorithms and accelerating convergence.

A F Atiya1, A G Parlos.   

Abstract

How to efficiently train recurrent networks remains a challenging and active research topic. Most of the proposed training approaches are based on computational ways to efficiently obtain the gradient of the error function, and can be generally grouped into five major groups. In this study we present a derivation that unifies these approaches. We demonstrate that the approaches are only five different ways of solving a particular matrix equation. The second goal of this paper is develop a new algorithm based on the insights gained from the novel formulation. The new algorithm, which is based on approximating the error gradient, has lower computational complexity in computing the weight update than the competing techniques for most typical problems. In addition, it reaches the error minimum in a much smaller number of iterations. A desirable characteristic of recurrent network training algorithms is to be able to update the weights in an on-line fashion. We have also developed an on-line version of the proposed algorithm, that is based on updating the error gradient approximation in a recursive manner.

Year:  2000        PMID: 18249797     DOI: 10.1109/72.846741

Source DB:  PubMed          Journal:  IEEE Trans Neural Netw        ISSN: 1045-9227


  16 in total

1.  Exploiting the Dynamics of Soft Materials for Machine Learning.

Authors:  Kohei Nakajima; Helmut Hauser; Tao Li; Rolf Pfeifer
Journal:  Soft Robot       Date:  2018-04-30       Impact factor: 8.071

2.  Spintronic reservoir computing without driving current or magnetic field.

Authors:  Tomohiro Taniguchi; Amon Ogihara; Yasuhiro Utsumi; Sumito Tsunegi
Journal:  Sci Rep       Date:  2022-06-23       Impact factor: 4.996

3.  Generating coherent patterns of activity from chaotic neural networks.

Authors:  David Sussillo; L F Abbott
Journal:  Neuron       Date:  2009-08-27       Impact factor: 17.173

4.  Information processing using a single dynamical node as complex system.

Authors:  L Appeltant; M C Soriano; G Van der Sande; J Danckaert; S Massar; J Dambre; B Schrauwen; C R Mirasso; I Fischer
Journal:  Nat Commun       Date:  2011-09-13       Impact factor: 14.919

5.  Optoelectronic reservoir computing.

Authors:  Y Paquot; F Duport; A Smerieri; J Dambre; B Schrauwen; M Haelterman; S Massar
Journal:  Sci Rep       Date:  2012-02-27       Impact factor: 4.379

6.  Optimal nonlinear information processing capacity in delay-based reservoir computers.

Authors:  Lyudmila Grigoryeva; Julie Henriques; Laurent Larger; Juan-Pablo Ortega
Journal:  Sci Rep       Date:  2015-09-11       Impact factor: 4.379

7.  Constructing optimized binary masks for reservoir computing with delay systems.

Authors:  Lennert Appeltant; Guy Van der Sande; Jan Danckaert; Ingo Fischer
Journal:  Sci Rep       Date:  2014-01-10       Impact factor: 4.379

8.  Predicting physical time series using dynamic ridge polynomial neural networks.

Authors:  Dhiya Al-Jumeily; Rozaida Ghazali; Abir Hussain
Journal:  PLoS One       Date:  2014-08-26       Impact factor: 3.240

9.  Physical reservoir computing with origami and its application to robotic crawling.

Authors:  Priyanka Bhovad; Suyi Li
Journal:  Sci Rep       Date:  2021-06-21       Impact factor: 4.379

10.  A soft body as a reservoir: case studies in a dynamic model of octopus-inspired soft robotic arm.

Authors:  Kohei Nakajima; Helmut Hauser; Rongjie Kang; Emanuele Guglielmino; Darwin G Caldwell; Rolf Pfeifer
Journal:  Front Comput Neurosci       Date:  2013-07-09       Impact factor: 2.380

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.