Literature DB >> 26073971

Learning Orthographic Structure With Sequential Generative Neural Networks.

Alberto Testolin1,2, Ivilin Stoianov2,3, Alessandro Sperduti4, Marco Zorzi2,5,6.   

Abstract

Learning the structure of event sequences is a ubiquitous problem in cognition and particularly in language. One possible solution is to learn a probabilistic generative model of sequences that allows making predictions about upcoming events. Though appealing from a neurobiological standpoint, this approach is typically not pursued in connectionist modeling. Here, we investigated a sequential version of the restricted Boltzmann machine (RBM), a stochastic recurrent neural network that extracts high-order structure from sensory data through unsupervised generative learning and can encode contextual information in the form of internal, distributed representations. We assessed whether this type of network can extract the orthographic structure of English monosyllables by learning a generative model of the letter sequences forming a word training corpus. We show that the network learned an accurate probabilistic model of English graphotactics, which can be used to make predictions about the letter following a given context as well as to autonomously generate high-quality pseudowords. The model was compared to an extended version of simple recurrent networks, augmented with a stochastic process that allows autonomous generation of sequences, and to non-connectionist probabilistic models (n-grams and hidden Markov models). We conclude that sequential RBMs and stochastic simple recurrent networks are promising candidates for modeling cognition in the temporal domain.
Copyright © 2015 Cognitive Science Society, Inc.

Keywords:  Connectionist modeling; Generative models; Orthographic structure; Probabilistic graphical models; Recurrent neural networks; Restricted Boltzmann machines; Statistical sequence learning; Unsupervised learning

Mesh:

Year:  2015        PMID: 26073971     DOI: 10.1111/cogs.12258

Source DB:  PubMed          Journal:  Cogn Sci        ISSN: 0364-0213


  5 in total

1.  Learning representation hierarchies by sharing visual features: a computational investigation of Persian character recognition with unsupervised deep learning.

Authors:  Zahra Sadeghi; Alberto Testolin
Journal:  Cogn Process       Date:  2017-02-25

2.  The Role of Architectural and Learning Constraints in Neural Network Models: A Case Study on Visual Space Coding.

Authors:  Alberto Testolin; Michele De Filippo De Grazia; Marco Zorzi
Journal:  Front Comput Neurosci       Date:  2017-03-21       Impact factor: 2.380

3.  Visual sense of number vs. sense of magnitude in humans and machines.

Authors:  Alberto Testolin; Serena Dolfi; Mathijs Rochus; Marco Zorzi
Journal:  Sci Rep       Date:  2020-06-22       Impact factor: 4.379

4.  Emergence of Network Motifs in Deep Neural Networks.

Authors:  Matteo Zambra; Amos Maritan; Alberto Testolin
Journal:  Entropy (Basel)       Date:  2020-02-11       Impact factor: 2.524

5.  Probabilistic Models and Generative Neural Networks: Towards an Unified Framework for Modeling Normal and Impaired Neurocognitive Functions.

Authors:  Alberto Testolin; Marco Zorzi
Journal:  Front Comput Neurosci       Date:  2016-07-13       Impact factor: 2.380

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.