Literature DB >> 9377276

Long short-term memory.

S Hochreiter1, J Schmidhuber.   

Abstract

Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient-based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O(1). Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.

Entities:  

Mesh:

Year:  1997        PMID: 9377276     DOI: 10.1162/neco.1997.9.8.1735

Source DB:  PubMed          Journal:  Neural Comput        ISSN: 0899-7667            Impact factor:   2.026


  2000 in total

1.  Interactions between frontal cortex and basal ganglia in working memory: a computational model.

Authors:  M J Frank; B Loughry; R C O'Reilly
Journal:  Cogn Affect Behav Neurosci       Date:  2001-06       Impact factor: 3.282

2.  Initialization and self-organized optimization of recurrent neural network connectivity.

Authors:  Joschka Boedecker; Oliver Obst; N Michael Mayer; Minoru Asada
Journal:  HFSP J       Date:  2009-10-26

3.  Deep Learning Models of the Retinal Response to Natural Scenes.

Authors:  Lane T McIntosh; Niru Maheswaranathan; Aran Nayebi; Surya Ganguli; Stephen A Baccus
Journal:  Adv Neural Inf Process Syst       Date:  2016

4.  Knowledge-Based Biomedical Word Sense Disambiguation with Neural Concept Embeddings

Authors:  Akm Sabbir; Antonio Jimeno-Yepes; Ramakanth Kavuluru
Journal:  Proc IEEE Int Symp Bioinformatics Bioeng       Date:  2018-01-11

5.  A customizable deep learning model for nosocomial risk prediction from critical care notes with indirect supervision.

Authors:  Travis R Goodwin; Dina Demner-Fushman
Journal:  J Am Med Inform Assoc       Date:  2020-04-01       Impact factor: 4.497

6.  Identification of Patients with Sarcopenia Using Gait Parameters Based on Inertial Sensors.

Authors:  Jeong-Kyun Kim; Myung-Nam Bae; Kang Bok Lee; Sang Gi Hong
Journal:  Sensors (Basel)       Date:  2021-03-04       Impact factor: 3.576

7.  Generating Classical Chinese Poems from Vernacular Chinese.

Authors:  Zhichao Yang; Pengshan Cai; Yansong Feng; Fei Li; Weijiang Feng; ElenaSuet-Ying Chiu; Hong Yu
Journal:  Proc Conf Empir Methods Nat Lang Process       Date:  2019-11

8.  Context-Aware Emotion Recognition in the Wild Using Spatio-Temporal and Temporal-Pyramid Models.

Authors:  Nhu-Tai Do; Soo-Hyung Kim; Hyung-Jeong Yang; Guee-Sang Lee; Soonja Yeom
Journal:  Sensors (Basel)       Date:  2021-03-27       Impact factor: 3.576

Review 9.  A roadmap to integrate astrocytes into Systems Neuroscience.

Authors:  Ksenia V Kastanenka; Rubén Moreno-Bote; Maurizio De Pittà; Gertrudis Perea; Abel Eraso-Pichot; Roser Masgrau; Kira E Poskanzer; Elena Galea
Journal:  Glia       Date:  2019-05-06       Impact factor: 7.452

Review 10.  How cognitive theory guides neuroscience.

Authors:  Michael J Frank; David Badre
Journal:  Cognition       Date:  2014-12-08
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.