| Literature DB >> 33286220 |
Yair Lakretz1, Stanislas Dehaene1,2, Jean-Rémi King3.
Abstract
Sentence comprehension requires inferring, from a sequence of words, the structure of syntactic relationships that bind these words into a semantic representation. Our limited ability to build some specific syntactic structures, such as nested center-embedded clauses (e.g., "The dog that the cat that the mouse bit chased ran away"), suggests a striking capacity limitation of sentence processing, and thus offers a window to understand how the human brain processes sentences. Here, we review the main hypotheses proposed in psycholinguistics to explain such capacity limitation. We then introduce an alternative approach, derived from our recent work on artificial neural networks optimized for language modeling, and predict that capacity limitation derives from the emergence of sparse and feature-specific syntactic units. Unlike psycholinguistic theories, our neural network-based framework provides precise capacity-limit predictions without making any a priori assumptions about the form of the grammar or parser. Finally, we discuss how our framework may clarify the mechanistic underpinning of language processing and its limitations in the human brain.Entities:
Keywords: artificial neural networks; double center-embeddings; language model; long-range dependencies; sentence processing
Year: 2020 PMID: 33286220 PMCID: PMC7516924 DOI: 10.3390/e22040446
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Figure 1Two sentences that carry the same meaning, with (top) and without (bottom) double center-embedding.
Figure 2An illustration of reading times in self-paced reading in object-extracted relative (blue) and subject-extracted (orange) relative clauses (adapted from Ref. [3]). While reading times are similar for the main verb (’hoped’), they vary substantially for the embedded one (’sent’). This suggests that the processing load for object-extracted relative clauses diverges most from subject-extracted relative-clauses at the embedded verb.
Figure 3A chunk representation of a syntactic structure in Adaptive Control of Thought–Rational (ACT-R) (adapted from [16]).
Figure 4The neural mechanism for long-range grammatical-number agreement. Cell activities of the syntax (green), singular (red) and plural (blue) units are presented during the processing of a sentence with a long-range subject-verb dependency across a subject-extracted relative clause. Panel (A) describes unit activity during the processing of sentences in which the main subject is singular. Continuous/dashed lines correspond to cases in which the intervening noun has the same/opposite number as the main subject (i.e., singular/plural). Similarly, Panel (B) describes conditions in which the main subject is plural. Error bars represent standard deviation across 2000 sentences.
Summary of the main elements of the cognitive models for sentence processing. The source for capacity limitation of each theory is described on the right column (see Section 3 for details).
| Cognitive Theories for Sentence Processing | ||||
|---|---|---|---|---|
|
| Grammar | Parsing Algorithm | Limiting Resource | Explanation for Capacity Limitation and processing breakdowns |
|
| Dependency grammar | Dependency parsing | Energy units | Too many long-range structural integrations take place at a given word, exceeding unit resources. |
|
| pCFG | Left-corner | Temporal activity | High similarity among memory items cause unresolvable interference. |
|
| pCFG | None | Probability mass | Frequent syntactic structures ‘consume’ most of the probability mass, leading rarer structures to generate high surprisal. |
|
| None | None | Dimensionality | Highly complex syntactic structures require higher state-space dimensionality than that available. |
|
| None | None | Specialized syntax units | The neural circuit for long-range dependencies is sparse and can therefore process only a limited number of nested, or cross, dependencies. |