| Literature DB >> 34305728 |
Abstract
Recent work on the application of neural networks to language modeling has shown that models based on certain neural architectures can capture syntactic information from utterances and sentences even when not given an explicitly syntactic objective. We examine whether a fully data-driven model of language development that uses a recurrent neural network encoder for utterances can track how child language utterances change over the course of language development in a way that is comparable to what is achieved using established language assessment metrics that use language-specific information carefully designed by experts. Given only transcripts of child language utterances from the CHILDES Database and no pre-specified information about language, our model captures not just the structural characteristics of child language utterances, but how these structures reflect language development over time. We establish an evaluation methodology with which we can examine how well our model tracks language development compared to three known approaches: Mean Length of Utterance, the Developmental Sentence Score, and the Index of Productive Syntax. We discuss the applicability of our model to data-driven assessment of child language development, including how a fully data-driven approach supports the possibility of increased research in multilingual and cross-lingual issues.Entities:
Keywords: IPSyn; child language assessment; computational linguistics; language model; natural language processing; neural network
Year: 2021 PMID: 34305728 PMCID: PMC8295984 DOI: 10.3389/fpsyg.2021.674402
Source DB: PubMed Journal: Front Psychol ISSN: 1664-1078
Figure 1A simple feedforward neural network with an input layer, an embedding layer, a hidden layer, and an output layer. This network can implement a bigram language model with units in the input and output layers representing different words in a vocabulary. When a word is active in the input layer, the unit with highest activation in the output layer is the model's prediction for the next word.
Figure 2A simple recurrent neural network. The hidden layer receives activation from the embedding layer and from the hidden layer in the previous time step.
Figure 3A simple recurrent neural network unrolled for L time steps, where L is the length of the input string.
Figure 4Our model for encoding and scoring language samples composed of utterances. Each utterance is encoded by a Bidirectional LSTM network. Utterance representations consisting of the concatenations of the first and last tokens are averaged into a vector that represents the entire language sample. From this language sample vector, a score is computed for the entire language sample.
Spearman rank correlation coefficients between age in months and four language development scores for the 16 children in our dataset.
| Braunwald: Laura | 0.732 ± 0.001 | 0.794 ± 0.001 | 0.867 ± 0.001 | 0.888 ± 0.001 |
| Brown: Adam | 0.942 ± 0.000 | 0.739 ± 0.001 | 0.906 ± 0.001 | 0.964 ± 0.000 |
| Brown: Eve | 0.976 ± 0.001 | 0.958 ± 0.000 | 1.000 ± 0.000 | 1.000 ± 0.000 |
| Brown: Sarah | 0.935 ± 0.000 | 0.953 ± 0.000 | 0.966 ± 0.000 | 0.959 ± 0.000 |
| Clark: Shem | 0.842 ± 0.002 | 0.855 ± 0.001 | 0.936 ± 0.001 | 0.889 ± 0.001 |
| Demetras: Trevor | 0.618 ± 0.003 | 0.567 ± 0.002 | 0.609 ± 0.003 | 0.727 ± 0.003 |
| Kuczaj: Abe | 0.855 ± 0.001 | 0.804 ± 0.002 | 0.943 ± 0.001 | 0.801 ± 0.001 |
| MacWhinney: Ross | 0.610 ± 0.002 | 0.588 ± 0.002 | 0.458 ± 0.002 | 0.604 ± 0.002 |
| Sachs: Naomi | 0.732 ± 0.002 | 0.869 ± 0.001 | 0.933 ± 0.001 | 0.92 ± 0.001 |
| Snow: Nathaniel | 0.190 ± 0.004 | 0.892 ± 0.001 | 0.905 ± 0.001 | 0.881 ± 0.001 |
| Suppes: Nina | 0.896 ± 0.001 | 0.896 ± 0.001 | 0.896 ± 0.001 | 0.974 ± 0.001 |
| Weist: Benjamin | 0.607 ± 0.004 | 0.927 ± 0.000 | 0.964 ± 0.001 | 1.000 ± 0.000 |
| Weist: Emily | 0.336 ± 0.003 | 0.643 ± 0.002 | 0.629 ± 0.002 | 0.432 ± 0.003 |
| Weist: Jillian | 0.321 ± 0.005 | 0.243 ± 0.005 | 0.126 ± 0.006 | 0.657 ± 0.003 |
| Weist: Matt | 0.685 ± 0.002 | 0.741 ± 0.001 | 0.622 ± 0.002 | 0.713 ± 0.001 |
| Weist: Roman | 0.311 ± 0.003 | 0.735 ± 0.003 | 0.566 ± 0.002 | 0.509 ± 0.002 |
| Average | 0.662 | 0.763 | 0.770 | 0.807 |
The four language development scores include our data-driven approach and three baselines: Mean Length of Utterance (MLU), the Developmental Sentence Score (DSS), and the Index of Productive Syntax (IPSyn).
Accuracy in detection of the 20 most common syntactic dependency types in our dataset from utterance encodings produced by our model.
| SUBJ | 82.3 |
| ROOT | 84.4 |
| JCT | 74.1 |
| DET | 72.3 |
| OBJ | 73.6 |
| AUX | 79.2 |
| POBJ | 73.9 |
| PRED | 72.7 |
| LINK | 72.1 |
| MOD | 61.3 |
| COMP | 78.9 |
| INF | 81.8 |
| NEG | 75.1 |
| QUANT | 64.1 |
| NJCT | 74.3 |
| COORD | 72.5 |
| CONJ | 75.8 |
| CJCT | 85.9 |
| CMOD | 75.1 |
| XMOD | 69.2 |
| AVERAGE | 74.9 |