| Literature DB >> 35401316 |
Ludovica Pannitto1, Aurelie Herbelot1,2.
Abstract
It has been shown that Recurrent Artificial Neural Networks automatically acquire some grammatical knowledge in the course of performing linguistic prediction tasks. The extent to which such networks can actually learn grammar is still an object of investigation. However, being mostly data-driven, they provide a natural testbed for usage-based theories of language acquisition. This mini-review gives an overview of the state of the field, focusing on the influence of the theoretical framework in the interpretation of results.Entities:
Keywords: construction grammar; grammar; language acquisition; recurrent neural networks; usage-based linguistics
Year: 2022 PMID: 35401316 PMCID: PMC8984258 DOI: 10.3389/fpsyg.2022.741321
Source DB: PubMed Journal: Front Psychol ISSN: 1664-1078
Figure 1LSTM networks are capable of keeping track of long-term dependencies. As recurrent neural networks (upper layer of the figure), they present a chain-like structure: at each time step t, the network's output is computed based on both the input of time t(x) and the network's state at time t−1(h). As opposed to a simple recurrent cell, an LSTM cell (lower layer of the figure) has the ability to regulate how the two kinds of information (input and previous state) are weighted towards the computation of the output. The first gate, the forget gate, evaluates C (a representation of the previous state different from h) against x and learns what information to keep from previous steps, including it in a vector f. Next, a candidate value for the current state Ĉ is computed along with the input gate vector i that weighs how much of the input will contribute to the current state. Finally, the state of the cell C is computed by weighting C with the forget gate vector f and the at Ĉ with the input vector i. h is then computed from C. A complete and easy to read guide to LSTMs can be found at https://colah.github.io/posts/2015-08-Understanding-LSTMs/.