Literature DB >> 27732574

Hybrid computing using a neural network with dynamic external memory.

Alex Graves1, Greg Wayne1, Malcolm Reynolds1, Tim Harley1, Ivo Danihelka1, Agnieszka Grabska-Barwińska1, Sergio Gómez Colmenarejo1, Edward Grefenstette1, Tiago Ramalho1, John Agapiou1, Adrià Puigdomènech Badia1, Karl Moritz Hermann1, Yori Zwols1, Georg Ostrovski1, Adam Cain1, Helen King1, Christopher Summerfield1, Phil Blunsom1, Koray Kavukcuoglu1, Demis Hassabis1.   

Abstract

Artificial neural networks are remarkably adept at sensory processing, sequence learning and reinforcement learning, but are limited in their ability to represent variables and data structures and to store data over long timescales, owing to the lack of an external memory. Here we introduce a machine learning model called a differentiable neural computer (DNC), which consists of a neural network that can read from and write to an external memory matrix, analogous to the random-access memory in a conventional computer. Like a conventional computer, it can use its memory to represent and manipulate complex data structures, but, like a neural network, it can learn to do so from data. When trained with supervised learning, we demonstrate that a DNC can successfully answer synthetic questions designed to emulate reasoning and inference problems in natural language. We show that it can learn tasks such as finding the shortest path between specified points and inferring the missing links in randomly generated graphs, and then generalize these tasks to specific graphs such as transport networks and family trees. When trained with reinforcement learning, a DNC can complete a moving blocks puzzle in which changing goals are specified by sequences of symbols. Taken together, our results demonstrate that DNCs have the capacity to solve complex, structured tasks that are inaccessible to neural networks without external read-write memory.

Year:  2016        PMID: 27732574     DOI: 10.1038/nature20101

Source DB:  PubMed          Journal:  Nature        ISSN: 0028-0836            Impact factor:   49.962


  34 in total

Review 1.  Using computational theory to constrain statistical models of neural data.

Authors:  Scott W Linderman; Samuel J Gershman
Journal:  Curr Opin Neurobiol       Date:  2017-07-18       Impact factor: 6.627

Review 2.  Reevaluating the Role of Persistent Neural Activity in Short-Term Memory.

Authors:  Nicolas Y Masse; Matthew C Rosen; David J Freedman
Journal:  Trends Cogn Sci       Date:  2020-01-29       Impact factor: 20.229

Review 3.  Direct Fit to Nature: An Evolutionary Perspective on Biological and Artificial Neural Networks.

Authors:  Uri Hasson; Samuel A Nastase; Ariel Goldstein
Journal:  Neuron       Date:  2020-02-05       Impact factor: 17.173

Review 4.  If deep learning is the answer, what is the question?

Authors:  Andrew Saxe; Stephanie Nelli; Christopher Summerfield
Journal:  Nat Rev Neurosci       Date:  2020-11-16       Impact factor: 34.870

Review 5.  Finding numbers in the brain.

Authors:  C R Gallistel
Journal:  Philos Trans R Soc Lond B Biol Sci       Date:  2017-02-19       Impact factor: 6.237

6.  Transforming task representations to perform novel tasks.

Authors:  Andrew K Lampinen; James L McClelland
Journal:  Proc Natl Acad Sci U S A       Date:  2020-12-10       Impact factor: 11.205

7.  Meta-learning synaptic plasticity and memory addressing for continual familiarity detection.

Authors:  Danil Tyulmankov; Guangyu Robert Yang; L F Abbott
Journal:  Neuron       Date:  2021-12-02       Impact factor: 17.173

8.  Placing language in an integrated understanding system: Next steps toward human-level performance in neural language models.

Authors:  James L McClelland; Felix Hill; Maja Rudolph; Jason Baldridge; Hinrich Schütze
Journal:  Proc Natl Acad Sci U S A       Date:  2020-09-28       Impact factor: 11.205

9.  Use of deep learning to develop continuous-risk models for adverse event prediction from electronic health records.

Authors:  Christopher Nielson; Martin G Seneviratne; Joseph R Ledsam; Shakir Mohamed; Nenad Tomašev; Natalie Harris; Sebastien Baur; Anne Mottram; Xavier Glorot; Jack W Rae; Michal Zielinski; Harry Askham; Andre Saraiva; Valerio Magliulo; Clemens Meyer; Suman Ravuri; Ivan Protsyuk; Alistair Connell; Cían O Hughes; Alan Karthikesalingam; Julien Cornebise; Hugh Montgomery; Geraint Rees; Chris Laing; Clifton R Baker; Thomas F Osborne; Ruth Reeves; Demis Hassabis; Dominic King; Mustafa Suleyman; Trevor Back
Journal:  Nat Protoc       Date:  2021-05-05       Impact factor: 13.491

10.  Predictive learning as a network mechanism for extracting low-dimensional latent space representations.

Authors:  Mattia Rigotti; Eric Shea-Brown; Stefano Recanatesi; Matthew Farrell; Guillaume Lajoie; Sophie Deneve
Journal:  Nat Commun       Date:  2021-03-03       Impact factor: 14.919

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.