Literature DB >> 34758151

Word Order Variation is Partially Constrained by Syntactic Complexity.

Yingqi Jing1,2,3, Paul Widmer1,2, Balthasar Bickel1,2.   

Abstract

Previous work suggests that when speakers linearize syntactic structures, they place longer and more complex dependents further away from the head word to which they belong than shorter and simpler dependents, and that they do so with increasing rigidity the longer expressions get, for example, longer objects tend to be placed further away from their verb, and with less variation. Current theories of sentence processing furthermore make competing predictions on whether longer expressions are preferentially placed as early or as late as possible. Here we test these predictions using hierarchical distributional regression models that allow estimates of word order and word order variation at the level of individual dependencies in corpora from 71 languages, while controlling for confounding effects from the type of dependency (e.g., subject vs. object), and the type of clause (main vs. subordinate) involved as well as from trends that are characteristic of individual languages, language families, and language contact areas. Our results show the expected correlations of length with position and variation only for two out of six dependency types (obliques and nominal modifiers) and no difference between clause types. These findings challenge received theories of across-the-board effects of complexity on word order and word order variation and call for theoretical models that relativize effects to specific kinds of syntactic structures and dependencies.
© 2021 The Authors. Cognitive Science published by Wiley Periodicals LLC on behalf of Cognitive Science Society (CSS).

Entities:  

Keywords:  Dependency length; Dependency treebanks; Distributional regression model; Heaviness serialization principle; Word order variation

Mesh:

Year:  2021        PMID: 34758151      PMCID: PMC9287024          DOI: 10.1111/cogs.13056

Source DB:  PubMed          Journal:  Cogn Sci        ISSN: 0364-0213


Introduction

Our communication channels (speech, gesture, writing) impose a strict linearization of syntactic structure in language. While linearizations differ widely across languages (e.g., English puts the verb before its object, Japanese puts it after the object) (Greenberg, 1963), there is a growing body of evidence that they are partially shaped by general principles of efficiency and economy that facilitate the processing, planning, and learning of language (Culbertson, Smolensky, & Legendre, 2012; Dryer, 1992; Fedzechkina, Chu, & Florian Jaeger, 2018; Futrell, Levy, & Gibson, 2020; Gennari, Sloman, Malt, & Fitch, 2002; Gibson, 1998; Gibson et al., 2019; Hall, Mayberry, & Ferreira, 2013; Hawkins, 1983, 2014; MacDonald, 2013). While current proposals differ in their more specific predictions, they agree in the general hypothesis that the more complex a syntactic element is, the more rigidly it is placed at the periphery of a dependency structure, that is it is more rigidly final or more rigidly initial. While syntactic complexity can be measured in various ways, a useful approximation has proven to be the sheer length of a dependent or constituent in the number of word‐like segments. To illustrate the intuition behind this idea, consider the following examples from English and Japanese, where directed arcs denote which element depends on which other element in terms of “theme‐of” and “recipient‐of” dependencies. While the two languages differ in where they place the verb (initial in English, final in Japanese), both prefer to place the longer and more complex dependent (“to the man wearing glasses”) at the periphery of the verb's dependency structure. But when both arguments are shorter and simpler, they are more freely expressed as either “give the man the book” or “give the book to the man.” English and Japanese ditransitive examples The hypothesis that longer dependents are preferentially placed in the periphery is primarily motivated by more general observations of dependency length minimization (DLM) (Ferrer i Cancho, 2004; Futrell et al., 2020; Gibson, 1998, 2000; Jing, Blasi, & Bickel, in press; Liu, 2008; Temperley, 2008). The core finding of DLM research is that speakers seek to minimize the total length of dependencies in a given structure because this reduces pressure on working memory and keeps adjacent information that belongs together (Choi, 2007; Faghiri & Samvelian, 2014, 2020; Futrell et al., 2020; Gibson et al., 2019; Hawkins, 1994; Ros, Santesteban, Fukumura, & Laka, 2015; Yamashita, 2002; Yamashita & Chang, 2001). One specific effect of this bears on the peripheral placement of long dependents: when dependents occur on the same side of their head, placing the shortest dependent closest to the verb reduces the total length of dependencies. Thus, in both English and Japanese, placing the shorter theme (“the book”) closer to the verb (“give”) than the recipient (“the man wearing glasses”) ensures that the total distance between the verb and its dependents is the shortest possible. If the recipient was placed closer to the verb (“give the man wearing glasses the book”), this would lengthen the dependency between the verb and the theme (“the book”) in both languages. These specific effects of DLM are best ensured when linearization is not entirely left to speakers' contextual needs, but when grammatical rules or constraints evolve in such a way that they make the peripheral placement of longer dependents relatively rigid and allow only relatively short dependents to vary in their positions. Such rules or constraints will generate more dependencies with shorter overall distances than dependencies with longer overall distances (Gildea & Temperley, 2010; Temperley, 2008). The associations of long dependents with rigid placement and of short dependents with variable placement have often been noted in the linguistic literature, for example, in terms of a “mobility” principle that predicts short dependents to be more flexible than long dependents (Hawkins, 1983). One effect of the mobility principle is that it can disrupt the oft‐noted harmony in head/dependent ordering across different types of dependencies (Culbertson et al., 2012; Dryer, 1992; Greenberg, 1963; Lehmann, 1973; Temperley, 2008; Vennemann, 1974). For example, the order of verbs and their object (“see the house”) correlates with the order of nouns and dependent nominal modifiers (“the owner of the house”). The order of nouns and adjectives, however, appears to escape the correlation when adjectives are short, single‐word expressions, but not when they are fully fledged adjective phrases (Dryer, 1992). For example, the English noun phrases with a single adjective do not follow the general head‐initial (HI) preference of the language (cf. head‐final [HF] “rich countries”) while noun phrases with fully fledged adjective phrases do (cf. HI “countries rich in resources”). Gulordava (2018) shows that this is a general pattern in corpus data across Romance languages. While these considerations predict a peripheral and rigid placement of more complex dependents, they do not make a prediction on whether the periphery is the first or the last position in a dependency structure. On this question, there are two competing theories. One theory goes back to Behaghel's “law of growing parts” and privileges the last position, following a general short/simple before long/complex linearization (Behaghel, 1909). Behaghel's law has been recast in various ways in the literature, such as in terms of an end‐weight preference (Quirk, 1972; Wasow, 1997), the heaviness serialization principle (Hawkins, 1983), a heavy NP shift (Arnold, Losongco, Wasow, & Ginstrom, 2000), a short‐before‐long ordering (Fedzechkina et al., 2018; Temperley, 2008), and an easy first bias (MacDonald, 2013). Behaghel's law is commonly motivated by the cognitive simplicity of short elements (Bock, 1982; Bock & Levelt, 1994; Christianson & Ferreira, 2005; Levelt, 1999; MacDonald, 2013). This simplicity effect is potentially strengthened by the fact that short elements (such as pronouns) tend to represent easy‐to‐access, known information (Arnold et al., 2000; J. M. Gundel, 1975; J. K. Gundel, 1988; Haviland & Clark, 1974; Krifka, 2008; MacDonald, 2013). For pairs of heads and dependents (where the head is usually a single word), Behaghel's law predicts that longer dependents tend to be placed after their head (as they are in the English example in 1a), while short dependents are expected to be more variably placed on either side of the head (as is the case of some English adverbs which can be ordered both as in “she quickly finished” or as in “she finished quickly”) (Gildea & Temperley, 2010; Gildea & Jaeger, 2015; Temperley, 2008). In other words, across languages, one would expect more HI and less variable structures with longer and more complex dependents (longhead‐initialless variable). The other theory privileges the first position and is based on the notion that speakers seek to convey new and rich information as early as possible. This has been motivated by a general decrease in communicative importance over the course of an utterance since short elements (e.g., pronouns) represent information that is less important and can be inferred from the previous discourse (Givón, 1988; J. K. Gundel, 1988). An alternative motivation is grounded in expectation‐based theories, which propose that the early placement of semantically rich information reduces surprisal and facilitates the processing of the upcoming head by increasing its predictability (Husain, Vasishth, & Srinivasan, 2014; Konieczny, 2000; R. Levy, 2008; R. P. Levy & Keller, 2013; Vasishth & Lewis, 2006), and they also facilitate sentence planning by keeping plans distinct from alternatives (Sauppe et al., 2021). For example, a complex object noun phrase facilitates the processing of a final verb in cases like the following: Expectation‐based facilitation in German (Konieczny 2000) Er hat das Buch, [das Lisa gestern gekauft hatte], hingelegt. he has the book that Lisa yesterday bought had laid.down Er hat das Buch hingelegt, [das Lisa gestern gekauft hatte]. he has the book laid.down that Lisa yesterday bought had ‘He has laid down the book that Lisa had bought yesterday.' In self‐paced reading, the verb is read faster in such examples when it follows the semantically rich and complex noun phrase in 2a (“the book that Lisa bought yesterday”) than when it follows the short object noun phrase “the book” in 2b, where the relative clause is extraposed (Konieczny, 2000). Since these facilitation effects matter most for long and information‐rich dependents, one would expect less variation in placement with longer dependencies (longhead‐finalless variable). In what follows, we use cross‐linguistic corpus data to assess the evidence for the overall hypothesis that longer dependents tend to be placed at the periphery of a dependency structure and to evaluate the relative evidence for the more specific hypotheses of a cross‐linguistic preference for early versus late placements of longer dependents. We do this by statistically modeling the order and variation of where dependents are placed in relation to their length in individual dependency structures, controlling for general trends in languages, families, and geographical areas. We furthermore control for differences arising from types of clauses and types of dependencies. With regard to clause types, it has been hypothesized that word order varies more in main than in dependent clauses (Ross, 1973), and that, consistently with this, main clauses are more amenable to variation over time (Givón, 1979; Hock, 1991; Lightfoot, 1982; Vennemann, 1975). This hypothesis is generally motivated by the assumption that subordinate clauses tend to contain presupposed background information (Givón, 1979; Hooper & Thompson, 1973) or that they tend to be processed as whole chunks (Bybee, 2002). With regard to dependency types, it has been noted that they differ in the variation they tend to tolerate. For instance, overall, word orders in noun modifiers tend to have less flexible order than core verbal arguments (Gulordava & Merlo, 2015; Tily, 2010), and oblique arguments are found to be the most variable (Levshina, 2019). Our focus is entirely on individual dependencies, and we seek to assess the proposed general principles that push longer dependents to early or late in a given dependency and that they show less variation than short dependents. Therefore, we do not presently explore potential interactions of a given dependency with other dependencies in the same sentence or the same language, nor do we address effects from the combination of several dependencies into complex structures. As a consequence, our study differs from DLM research, where the total lengths over all dependencies are compared to randomized baselines generated by entire grammars. A randomization‐based approach is not readily suitable for simultaneous measurements of order and variation at the level of individual dependencies instead of entire grammars.

Data and methods

Data

The corpus data in this study comes from the Universal Dependencies (UD) database, version 2.7 (Nivre et al., 2020). The UD database provides a collection of dependency‐annotated corpora of diverse languages in the general framework of Dependency Grammar (de Marneffe & Manning, 2008; de Marneffe & Nivre, 2019; Hudson, 1984; Liu, 2009; Melčuk, 1988; Petrov, Das, & McDonald, 2012; Zeman, 2008). Each pair of words in a sentence is linked via directed arcs that indicate their head directions and dependency types. Given continued controversies about whether functional categories, such as determiners or auxiliary verbs, should be considered as heads or dependents(Futrell et al., 2020; Groß & Osborne, 2015; Osborne & Gerdes, 2019; Osborne & Maxwell, 2015), we focus exclusively on dependencies between major lexical categories, such as V(erb), N(oun), Adj(ective), and Adv(erb). These have a clear head/dependent structure and are cross‐linguistically more comparable than functional categories. Furthermore, the position of pronouns and auxiliaries is often subject to confounding factors from phonology (as when auxiliaries need a host for cliticization). This gives us a set of six dependencies (verb/nominal subject (the boy ran away), verb/nominal object (read the book), verb/oblique (read the book [over the weekend]), verb/adverbial modifier (read silently), noun/nominal modifier (my friend's book), and noun/adjectival modifier (a new book)) in 71 treebanks from UD v2.7. The choice of these dependencies guarantees a sufficient number of tokens in each corpus and across main and subordinate clauses (), and a minimum coverage of 40 languages. For those languages with more than one treebank, we choose the largest one. The information of these treebanks is summarized in Table 1.
Table 1

Overview of 71 dependency treebanks from UD v2.7. We simplify the geographical distribution as Europe versus not since most languages are from Europe (see Fig. S4 for the geographic distribution of languages)

LanguageFamilyEuropeSentencesWord TokenLanguageFamilyEuropeSentencesWord token
AfrikaansIndo‐EuropeanF1,93449,276IrishIndo‐EuropeanT4,910115,969
AkkadianAfro‐AsiaticF1,80421,962ItalianIndo‐EuropeanT14,167298,343
Ancient GreekIndo‐EuropeanT17,080213,999JapaneseJapaneseF57,0281,250,875
ArabicAfro‐AsiaticF19,738738,889KoreanKoreanF27,363350,090
ArmenianIndo‐EuropeanT2,50252,630LatinIndo‐EuropeanT26,977450,515
BambaraMandeF1,02613,823LatvianIndo‐EuropeanT13,643219,955
BasqueBasqueT8,993121,443LithuanianIndo‐EuropeanT3,64270,047
BelarusianIndo‐EuropeanT23,534275,153MalteseAfro‐AsiaticF2,07444,162
BretonIndo‐EuropeanT88810,054NaijaCreoleF9,242140,729
BulgarianIndo‐EuropeanT11,138156,149North SamiUralicT3,12226,845
BuryatMongolicF92710,185NorwegianIndo‐EuropeanT20,044310,221
CantoneseSino‐TibetanF1,00413,918Old Church SlavonicIndo‐EuropeanT6,33857,563
CatalanIndo‐EuropeanT16,678531,971Old FrenchIndo‐EuropeanT17,678170,740
ChineseSino‐TibetanF4,997123,291Old RussianIndo‐EuropeanT16,944149,780
Classical ChineseSino‐TibetanF48,434233,122PersianIndo‐EuropeanF29,107501,776
CopticAfro‐AsiaticF1,87348,631PolishIndo‐EuropeanT22,152350,036
CroatianIndo‐EuropeanT9,010199,409PortugueseIndo‐EuropeanT12,078319,853
CzechIndo‐EuropeanT87,9131,506,484RomanianIndo‐EuropeanT26,225572,436
DanishIndo‐EuropeanT5,512100,733RussianIndo‐EuropeanT61,8891,106,296
DutchIndo‐EuropeanT13,578208,470SanskritIndo‐EuropeanF3,99727,117
EnglishIndo‐EuropeanT16,622254,829Scottish GaelicIndo‐EuropeanT3,17360,417
ErzyaUralicT1,69017,148SerbianIndo‐EuropeanT4,38497,673
EstonianUralicT30,972437,769SlovakIndo‐EuropeanT10,604106,043
FaroeseIndo‐EuropeanT1,62140,484SlovenianIndo‐EuropeanT8,000140,670
FinnishUralicT15,136202,194SpanishIndo‐EuropeanT17,680549,569
FrenchIndo‐EuropeanT18,535573,370SwedishIndo‐EuropeanT6,02696,819
GalicianIndo‐EuropeanT3,993138,837TamilDravidianF6009,581
GermanIndo‐EuropeanT189,9283,399,390Turkish GermanMixedF1,89131,946
GothicIndo‐EuropeanT5,40155,336TurkishTurkicF9,761122,383
GreekIndo‐EuropeanT2,52163,441UkrainianIndo‐EuropeanT7,060122,091
HebrewAfro‐AsiaticF6,216161,417Upper SorbianIndo‐EuropeanT64611,196
Hindi EnglishMixedF1,89826,909UrduIndo‐EuropeanF5,130138,077
HindiIndo‐EuropeanF16,647351,704VietnameseAustro‐AsiaticF3,00043,754
HungarianUralicT1,80042,032WelshIndo‐EuropeanT1,65732,911
IcelandicIndo‐EuropeanT44,029985,057WolofNiger‐CongoF2,10744,258
IndonesianAustronesianF5,593121,923
Overview of 71 dependency treebanks from UD v2.7. We simplify the geographical distribution as Europe versus not since most languages are from Europe (see Fig. S4 for the geographic distribution of languages) The dependencies in UD are defined by combining semantic and syntactic criteria : a nominal subject (“nsubj”) is the most agent‐like argument in an active transitive or the single argument in an intransitive clause, and an object of the verb (“obj”) is the most patient‐like argument in a transitive clause. The oblique relation (“obl”) subsumes both obliquely marked arguments (e.g., adpositional objects like “to the reader” in give it to the reader or “by”‐agents in passives) and adjuncts (e.g., time or location expressions like “in the city”). An adverbial modifier (“advmod”) is an adverb or adverbial phrase that serves to modify a predicate, and a nominal modifier (“nmod”) is used for nominal dependents of another noun or noun phrase and functionally corresponds to an attribute or genitive complement. An adjectival modifier (“amod”) is an adjectival word or phrase that modifies a noun. For illustration, Fig. 1 provides an annotated English sentence containing a dependent clause, where the target dependencies are displayed in italic. We strictly separate the word orders in main clauses, including only simple clauses, and in subordinate clauses, including adjectival (relative) clauses (“acl”), adverbial clauses (“advcl”), clausal subjects (“csubj”), and clausal complements (“ccomp”). Accordingly, the main clause part in Fig. 1 is also excluded to avoid any potential confounds from the overall the complexity of the sentence.
Fig. 1

Dependency graph of an English sentence with an adverbial clause: each terminal is annotated by its parts of speech tag (e.g., PRON for “pronoun”), and each arc for its dependency types (e.g., “nsubj” for “nominal subject”). The directed arcs go from the head to its dependents, indicating the head directions, for example, the subject “hotel” precedes the verbal head “had” forming a HF dependency, and the object “rooms” forms a HI dependency with the governor “had.”

Dependency graph of an English sentence with an adverbial clause: each terminal is annotated by its parts of speech tag (e.g., PRON for “pronoun”), and each arc for its dependency types (e.g., “nsubj” for “nominal subject”). The directed arcs go from the head to its dependents, indicating the head directions, for example, the subject “hotel” precedes the verbal head “had” forming a HF dependency, and the object “rooms” forms a HI dependency with the governor “had.” We measure dependency length (DL) as the linear distance between each pair of a head and its dependents in numbers of words (as defined by the UD word segmentation). For example, the adjacent dependency between the verb had and nominal subject hotel yields a length of , and the DL between the head verb had and object rooms is . The adjective modifier free is adjacent to its head noun rooms with a DL of .

Methods

Previous corpus‐based studies on word order variation mostly use entropy to quantify word order freedom at the language level. For example, Futrell, Mahowald, and Gibson (2015) use entropy to quantify free ordering of subjects and objects as a predictor of case marking. Gulordava (2018) takes entropy as a measure of a diachronic change towards more rigid word order in the history of Greek. Levshina (2019) uses entropy to show that function elements generally show the least degree of word order variation, followed by noun modifiers, core verbal arguments, and adjuncts/obliques. However, since we seek to estimate effects of length within single dependency structures, entropies would have to be estimated for each length, or each bin of lengths, over the corpus of a given language. Since lengths show a power‐law distribution, this severely reduces sample sizes for longer lengths, making estimates unstable, reducing statistical power, and introducing problematic associations between sample size and entropy (Chao, Wang, & Jost, 2013; DeDeo, Hawkins, Klingenstein, & Hitchcock, 2013). Moreover, measuring entropies at the corpus (language) level risks underestimating the actual variation within it, since entropies based on observed frequencies may have a poor representation of the real distributions, especially for rare dependency types or in small corpora (see Futrell et al., 2015, for bootstrap estimators of entropy using subcorpora and complete corpora). It is in fact challenging to get a reliable estimate of entropy in high‐dimensional and undersampled datasets, and some bias correction is often needed (Chao et al., 2013; DeDeo et al., 2013; Hausser & Strimmer, 2009; Nemenman, Shafee, & Bialek, 2002; Wolpert & Wolf, 1995). In response to this, we move away from entropy measures of word order flexibility. As an alternative, we adopt distributional (location/scale) regression models to simultaneously predict head directionality (location) and variation (scale) of different word orders. The distributional models also allow us to directly evaluate the role of each predictor (DL, dependency type, clause type) on both word order preferences and their variation at both population (“fixed effects”) and language (“random effects”) levels (Nalborczyk, Batailler, Lœvenbruck, Vilain, & Bürkner, 2019). The distributional models relax the assumption of a constant scale parameter, but this comes with a degree of complexity that makes these models hard to fit in a frequentist maximum‐likelihood approach. In response to this, we fit the models in a Bayesian framework. The development of Markov chain Monte Carlo (MCMC) algorithms, especially the Hamiltonian Monte Carlo and its variant No‐U‐Turn Sampler, makes it possible to efficiently sample from high‐dimensional distributions and estimate parameters in a complex distributional model. We predict the direction and variation of each word order in a model assuming a beta‐binomial likelihood function. The beta‐binomial distribution is a mixture of a binomial and a beta distribution. It generalizes the binomial distribution, and can capture overdispersion. With the beta‐binomial model, the binomial probability is randomly drawn from a beta distribution with hyperparameters and . To make it easier to interpret the results, we choose a slightly different parameterization by using the parameters of a mean () for the location and the reciprocal of the precision parameter () for the scale [cf. (Bürkner, 2018; Jørgensen, 1997; Kruschke, 2014) for similar parameterizations]. The parameter represents the expected binomial probability of each word order, and is the dispersion parameter indicating the degree of variation of word orders. This model allows us to predict the probability of each word order (H[ead]I[nitial] or H[ead]F[inal]) and its variation at the same time and given the same covariates. We build and compare distributional regression models as shown in Table 2. At the population level (“fixed effects”), we estimate coefficients ofDL, clause type (CLS), and dependency type (DEP), allowing for group‐level adjustments per language (“lang”), language family (“fam”), and area. This model structure can not only capture the language‐specific variation but also allow us to compare the effect of DL on word orders between population level and language level.
Table 2

Models for estimating head‐finality of each word order () and word order variation () using brms‐style notation. CLS: clause type; DEP: dependency type; DL: dependency length; lang: language; fam: family. The family and area groups control for phylogenetic and spatial autocorrelation, respectively

NameRegression
Intercept μ1+(1lang)+(1fam)+(1area)
ϕ1+(1lang)+(1fam)+(1area)
Clause type μ1+CLS+(1+CLSlang)+(1fam)+(1area)
ϕ1+CLS+(1+CLSlang)+(1fam)+(1area)
Dependency type μ1+DEP+(1+DEPlang)+(1fam)+(1area)
ϕ1+DEP+(1+DEPlang)+(1fam)+(1area)
DL μ1+DL+(1+DLlang)+(1fam)+(1area)
ϕ1+DL+(1+DLlang)+(1fam)+(1area)
Dependency type by clause type μ1+DEP+(1+DEPCLS)+(1+DEPlang)+(1fam)+(1area)
ϕ1+DEP+(1+DEPCLS)+(1+DEPlang)+(1fam)+(1area)
DL by clause type μ1+DL+(1+DLCLS)+(1+DLlang)+(1fam)+(1area)
ϕ1+DL+(1+DLCLS)+(1+DLlang)+(1fam)+(1area)
DL by dependency type μ1+DL+(1+DLDEP)+(1+DLlang)+(1fam)+(1area)
ϕ1+DL+(1+DLDEP)+(1+DLlang)+(1fam)+(1area)
Models for estimating head‐finality of each word order () and word order variation () using brms‐style notation. CLS: clause type; DEP: dependency type; DL: dependency length; lang: language; fam: family. The family and area groups control for phylogenetic and spatial autocorrelation, respectively We model interactions between DL and types of clauses and dependencies by allowing slopes to vary across types. We opt for this strategy (rather than models with population‐level interactions) in order to profit from the shrinkage effects that arise when information is partially pooled across types (McElreath, 2020; Nalborczyk et al., 2019). This strategy avoids overfitting at the population level, and at the same time it accommodates the unequal and sometimes small sample sizes that arise when estimating effects across particular combinations of lengths and types. In these models, the response is determined by the observed number of HF dependencies in the total number of dependencies at a given combination of levels in the terms of the model. It is linked to the expected binomial probability of head‐finality for each word order via the log odds (logit) function. The response is the dispersion parameter for word order variation. Higher values suggest more word order variation, while lower values suggest more rigid word order. We treat DL as a monotonic predictor, mo(DL), which accommodates non‐linearity and is specifically designed for integer predictors and ordered factors more generally (Bürkner & Charpentier, 2020). The parameterization of monotonic effects is realized by joint estimations of the average word order differences due to DL and the relative proportions of observed differences between two adjacent categories. If the word order differences are evenly distributed between all adjacent levels of DL, a linear relationship of DL on word orders is guaranteed. Otherwise, a non‐linear relationship is achieved. Thus, the mo(DL) term can not only evaluate the overall effect of DL on word orders but also identify the most influential region between consecutive levels of DL. We fit the models in Stan (Stan Development Team, 2020) via the brms (Bürkner, 2017) and cmdstanr (Gabry & Češnovar, 2020) interfaces to R (R Core Team, 2020). We choose weakly informative priors by setting Student‐t(20, 0, 1) for intercepts and slopes, and half‐Cauchy(0, 1) for the standard deviations. For each model, we run two independent MCMC chains for 3,000 iterations with a warm‐up of 1,000 iterations. We check for the convergence by ensuring effective sample sizes and diagnostic for each parameter. We assess the performance of models by leave‐one‐out cross‐validation in terms of the expected log pointwise predictive density (elpd). We approximate the eldp through Pareto‐smoothed importance sampling (PSIS) from the posterior log probabilities (McElreath, 2020; Yao et al., 2018). Apart from examining raw eldp estimates, we furthermore compare model performance by model stacking, that is by allocating weights to models in such a way that they jointly maximize prediction quality (Vehtari et al., 2020). To get the posterior predictive effects for each predictor, we furthermore build a full model by including all predictors (DL, dependency types, and clause types) and their mutual interactions. We report 90% credible intervals since they are computationally more stable than wider intervals when the effective sample size is below 10k (Kruschke, 2014). Code and data summaries are available in the Supporting Information; executable code and input data are accessible at https://osf.io/4s67a/.

Results

Fig. 2 visualizes the relationship between DL and word order across clause and dependency types in the raw data, using medians to capture the proportion of HF orders and boxplot quartiles to capture the variation in that proportion. The plot suggests that dependency type, but not clause type, has an appreciable impact on the relationship between DL and word order. These impressions are confirmed by the results from model comparison where dependency type turns out to be the most important predictor of word order, by itself (leveraging 46% of the total weight; Table 3) or as a source of variation for the effect of DL on word order (15%). Clause type is relevant for prediction only when modeled as a source of variation for the effect of dependency type (38%). Consistent with this, the posterior probability of population‐level differences between clause types includes 0 even in its 60% credible interval ( CI for head direction; CI for word order variation). In what follows, we focus on the conditional effects of DL across dependency types in the full model, which subsumes all models from Table 3.
Fig. 2

Boxplots for the distribution of raw HF dependencies in relation to DL across main and subordinate clauses.

Table 3

Model comparisons via approximate leave‐one‐out cross‐validation and model stacking. Models are ordered by their expected log pointwise densities (elpd) as estimated via Pareto‐smoothed importance sampling (psis). A higher elpd value indicates better model performance

Model elpdpsis Δelpdpsis SE(Δelpdpsis) Weights
Dependency type by clause type−13,029.310.000.000.38
Dependency type−13,049.78−20.478.420.46
DL by dependency type−14,059.59−1,030.2954.020.15
DL by clause type−15,412.45−2,383.1453.70.01
DL−15,416.04−2,386.7453.560.00
Clause type−15,450.81−2,421.551.940.00
Intercept−15,457.31−2,428.0051.640.00
Model comparisons via approximate leave‐one‐out cross‐validation and model stacking. Models are ordered by their expected log pointwise densities (elpd) as estimated via Pareto‐smoothed importance sampling (psis). A higher elpd value indicates better model performance Boxplots for the distribution of raw HF dependencies in relation to DL across main and subordinate clauses. We first present the direction estimates (Fig. 3) and then the variation estimates (Fig. 4). Since the effect of DL on word order does not vary much by clause type (Fig. 2) and models that allow this variation leverage almost no weight (Table 3), we collapse over main and subordinate clauses in what follows.
Fig. 3

Monotonic effect of DL on word order directions ( in log odds) across dependency types, after conditioning on the language‐specific effects (“random slopes”) and the overall variation between families and areas (“random intercepts”). Higher values indicate higher (log) odds of HF orders. The thick black lines show the estimated mean effect of DL at the population level (“fixed effects”), and the thin gray lines show the estimated mean effects in individual languages (“random effects”). The colored ribbons indicate 90% credible intervals around the population‐level mean.

Fig. 4

Monotonic effect of DL on word order variation ( on the log scale) across dependency types, after conditioning on the language‐specific effects (“random slopes”) and the overall variation between families and areas variable (“random intercepts”) by family and area. Higher values indicate more word order variation. The thick black lines show the estimated mean effects of DL at the population level (“fixed effects”), and the thin gray lines show the estimated mean effects at the language level (“random effects”). The colored ribbons indicate 90% credible intervals around the population‐level mean.

Monotonic effect of DL on word order directions ( in log odds) across dependency types, after conditioning on the language‐specific effects (“random slopes”) and the overall variation between families and areas (“random intercepts”). Higher values indicate higher (log) odds of HF orders. The thick black lines show the estimated mean effect of DL at the population level (“fixed effects”), and the thin gray lines show the estimated mean effects in individual languages (“random effects”). The colored ribbons indicate 90% credible intervals around the population‐level mean. Monotonic effect of DL on word order variation ( on the log scale) across dependency types, after conditioning on the language‐specific effects (“random slopes”) and the overall variation between families and areas variable (“random intercepts”) by family and area. Higher values indicate more word order variation. The thick black lines show the estimated mean effects of DL at the population level (“fixed effects”), and the thin gray lines show the estimated mean effects at the language level (“random effects”). The colored ribbons indicate 90% credible intervals around the population‐level mean. With regard to preferred head orderings, we observe strong differences between dependency types: subjects (“S”), adverbs (“Adv”), and adjectives (“Adj”) slightly prefer to precede their heads (i.e., have higher (log) odds for HF dependencies), while the other dependencies, especially objects and nominal modifiers, tend to follow their heads in our sample languages. However, with the exception of the initial placement of subjects, there is substantial cross‐linguistic variation (as shown by the thin gray lines in the figure). The overall effect of DL on word order direction includes 0 ( CI ; ), but there is appreciable variation across dependency types: obliques (estimated mean slope: ; CI ; ) and nominal modifiers (estimated mean slope: ; CI ; ) show negative correlations, suggesting strong tendencies for long obliques and nominal modifiers to be post‐posed (see Table 4 for summaries). The major differences in head directionality (on average ) occur between short dependencies at the lengths of 1 and 2 (cf. the simplex estimates of the monotonic parameterization of DL in Section S4.2 in the Supporting Information). By contrast, there are no strong correlations between DL and head‐finality for the other orders, as all 90% CI includes 0. With the increase of DL, adjective modifiers show a slight negative trend (estimated mean slope: ; CI ; ), whereas adverb modifiers exhibit a slight positive correlation (estimated mean slope: ; CI ; ). The placements of subjects and objects seem to be independent of DL (see Table 4 for estimated slopes of DL on word order directions). The relationship between DL and head directionality is relatively consistent between population (“fixed effects”) and language (“random effects”) levels for obliques (range of estimated slopes: ; ), and nominal modifiers (range of estimated slopes: ; ). More language‐specific relationships of DL on head ordering are observed for the other orders (cf. the thin gray lines in Fig. 3 and more detailed reports in Figs. S7– S12 and Tables S3– S8 in the Supporting Information).
Table 4

Estimated intercepts and mean slopes for DL on word order directions ( in log odds) across dependency types. The last column reports the proportion of the posterior estimates that is higher than 0 when the mean is higher than 0, and the proportion that is lower than 0 if the mean is lower than 0

Dependency TypeParameterMean90% CI p(parameter<>0)
V and SIntercept1.33[−0.04, 2.63].95
Slope−0.02[−0.37, 0.33].60
V and OIntercept−1.42[−2.77, 0.10].96
Slope−0.08[−0.41, 0.27].78
V and OblIntercept1.25[−0.13, 2.58].93
Slope−0.56[−0.90, 0.22].97
V and AdvIntercept0.92[−0.46, 2.25].87
Slope0.13[−0.21, 0.47].86
N and AdjIntercept1.93[0.46, 3.35].98
Slope−0.25[−0.59, 0.10].91
N and NmodIntercept−0.79[−2.23, 0.63].83
Slope−0.38[−0.73, −0.03].95
Estimated intercepts and mean slopes for DL on word order directions ( in log odds) across dependency types. The last column reports the proportion of the posterior estimates that is higher than 0 when the mean is higher than 0, and the proportion that is lower than 0 if the mean is lower than 0 Estimated intercepts and mean slopes for DL on word order variation ( on the log scale) across dependency types. The last column reports the proportion of the posterior estimates that is higher than 0 when the mean is higher than 0, and the proportion that is lower than 0 if the mean is lower than 0 Turning to the predictive effects of DL on word order variation in Fig. 4 and Table 5, we note similar extents of variation across all dependency types (with a mean value of about ). There is no evidence for a general effect of DL across all types at the population level ( CI ; ) and effects differ across types: there are strong effects of DL on variation in obliques (estimated mean slope = ; CI ; ) and nominal modifiers (estimated mean slope =; CI ; ), that is longer dependencies of obliques and nominal modifiers yield less word order variation. A strong negative effect of DL is observed for objects (estimated mean slope ; CI ; ) and adverbial modifiers (estimated mean slope ; CI ; ). A slight positive relationship between DL and variation is found for adjectives (estimated mean slope ; CI ; ), and no correlation is observed for subjects (estimated mean slope ; CI ; ). All the observations are remarkably consistent across individual languages. This can be seen by the consistent downward slopes of gray lines, which represent the individual languages in Fig. 4. The estimated slopes for obliques range from to , with a standard deviation of , and the estimated slopes for nominal modifiers range from to , with a standard deviation of across different languages (see Figs. S14– S19 and Tables S10– S15 in the Supporting Information). The strongest effect occurs between lengths 1 and 2, although with 52% the difference to the other levels is not as big as in the case of word order direction (see Section S4.2).
Table 5

Estimated intercepts and mean slopes for DL on word order variation ( on the log scale) across dependency types. The last column reports the proportion of the posterior estimates that is higher than 0 when the mean is higher than 0, and the proportion that is lower than 0 if the mean is lower than 0

Dependency TypeParameterEstimates90% CI p(parameter<>0)
V and SIntercept−2.21[−3.28, −0.38].97
Slope0.01[−0.39, 0.47].48
V and OIntercept−1.83[−2.89, 0.01].95
Slope−0.42[−0.79, 0.05].94
V and OblIntercept−0.77[−1.81, 1.08].84
Slope−0.75[−1.13, −0.30].97
V and AdvIntercept−2.10[−3.22, −0.23].96
Slope−0.24[−0.62, 0.22].89
N and AdjIntercept−3.40[−4.48, −1.56]1.00
Slope0.21[−0.17, 0.68].88
N and NmodIntercept−1.94[−3.07, −0.09].96
Slope−0.73[−1.13, −0.28].97

Discussion

Our results suggest that DL is not sufficient to predict the directions and variation of word orders, and no general correlations are obtained across all dependency types. Strong effects of DL on direction and variation are limited to obliques and nominal modifiers: These dependency types show strong evidence for longer dependents to be placed after the head and more rigidly so than shorter dependents. This limitation of effects challenges across‐the‐board principles like Behaghel's law, the easy first principle or heaviness serialization principle, which all favor a consistently rightward positioning of longer dependents (or constituents) (Arnold et al., 2000; Behaghel, 1909; Hawkins, 1983; MacDonald, 2013; Wasow, 1997). Our results are also difficult to reconcile with the theory of predictability maximization, which would favor a consistent placement of dependents before the head across types (Ferrer‐i‐Cancho, 2017; Konieczny, 2000; Vasishth & Lewis, 2006). While we did not formally test this, our findings are only partially consistent with the observation in DLM research that longer elements tend to be placed later in HI languages and earlier in HF languages, that is a language‐level effect of harmonic ordering across types (Dryer, 1992; Futrell et al., 2020; Greenberg, 1963; Hawkins, 1983). This expectation shows up in our results in a partial segregation of language‐specific estimates in verb/object and verb/subject dependencies, with some languages showing an increase of HF positions and others a decrease with length (also see Figs. S7 and S8 in the Supporting Information for more detailed plots). However, this segregation into HF versus HI types is weak since many languages fall in between, and slopes are inconsistent even in the more clearly HF (above 0) and more clearly HI (below 0) languages. This suggests that the expected language‐level effect has only limited scope and that our results are not an artifact of pooling HI and HF languages with opposite length effects. DL has a very weak effect on the order of adjectives and adverbs. While there is a general preference for placing them before the head (means are above 0 in Fig. 3) and for constraining their order (all values are below 0 in Fig. 4), the preferences go in opposite, albeit very weak directions: the order preference decreases and variation increases with DL in the case of adjectives; the order preference increases and variation decreases with DL in the case of adverbs. While the pattern we observed for adverbs is consistent with the notion that single‐word dependents are more variably placed at the opposite side of the head to achieve DLM (Gulordava, 2018; Temperley, 2008; Temperley & Gildea, 2017), this does not generalize to adjectives. Intriguingly, the effect of DL is not evenly distributed over lengths. The effects are mostly due to differences between adjacent and non‐adjacent dependencies: on average of the differences in direction, and of the difference in variation are located between the first two levels of DL (lengths of 1 and 2). This implies that the increase of DL has a stronger negative effect for short dependencies, but has less influence on longer dependencies. Furthermore, it suggests a major distinction between single‐word versus phrasal dependents. This is consistent with proposals that draw a categorical difference between non‐branching and branching elements (Dryer, 1992), but the effect is limited to obliques and nominal modifiers. Word order directions differ considerably across dependency types in Fig. 3, although they all show a consistent overall bias against word order variation 4. A strong and cross‐linguistically robust preference for HF order is limited to the subject‐verb (SV) dependency. This is likely to reflect an oft‐noted preference to place agents and topics first (Bornkessel‐Schlesewsky & Schlesewsky, 2009; Comrie, 1981; Gibson et al., 2013, 2019; Givón, 1983; Goldin‐Meadow, So, Özyürek, & Mylander, 2008; Hall et al., 2013; Kemmerer, 2012; Krebs, Malaia, Wilbur, & Roehm, 2018; Meir, Sandler, Padden, & Aronoff, 2010; Riesberg, Malcher, & Himmelmann, 2019; Schouwstra & de Swart, 2014; Tomlin, 1986). We also note a general bias towards HI VO order, although it varies strongly across individual languages. As noted above, with the increase of DL, we furthermore observe a somewhat divergent trend for initial or final placement of long dependents at the language level, and different languages seem to explore their own ways of ordering complex elements. Further research is needed to evaluate to what extent DL effects in an individual object/verb dependency interact with other dependencies in the same sentence or in the same language. Assuming the VO bias holds independently of such potential interactions, it combines with the SV bias in a weak overall preference for subject‐verb‐object (SVO) order that emerges when estimates are conditioned on cross‐linguistic variation in our sample. To some extent this finding might be driven by the overrepresentation of European languages in the sample although our models condition of areal and family relations. To the extent it holds, such a preference is consistent with DLM observations: when there are two simultaneous dependents (S and O) of the same head (V), the total DL is minimal when they occur on either side of the head (Ferrer‐i Cancho, 2008, 2015). However, DLM observations elsewhere furthermore suggest that long dependents should be more constrained than short dependents. Yet this is not supported by our findings (Fig. 3): the directions of S and O barely change with increasing lengths. Our results show very similar word order patterns across main and subordinate clauses. This is in line with previous findings of roughly similar amount of word order variation across contexts (Kroch, 1989, 2001), and it challenges theories that assume more variation in main than in subordinate clauses (Bybee, 2002; Givón, 1979; Hooper & Thompson, 1973; Ross, 1973) and therefore a higher innovation potential of main clauses (Bean, 1983; MacLeish, 1969; Stockwell & Minkova, 1991). Previous work has also examined the relationship between word order direction and variation (Levshina, 2019). Our results are only partly consistent with the results from this work, although a full comparison is limited by the fact that we select narrower dependency types by relying on both dependency (“nsubj,” “obj,” “obl,” “advmod,” “amod,” and “nmod”) and part‐of‐speech tags (NOUN, VERB, ADJ, and ADV), whereas Levshina (2019) collapses some types, relies on dependency tags only, and examines variability depending on overall language type (VO vs. OV) rather than individual dependencies. Fig. 5 plots our estimates of variation against word order for dependency type. Overall, the correlations are relatively weak and virtually absent in the subject (S/V) and adverb (V/Adv) dependencies. One potential correlation seems to be a slightly lower variation in pre‐nominal (log odds of HF > 0 in 5) than in postnominal (log odds < 0) adjectives (Greenberg, 1963). This is partially consistent with Levshina's findings although her analysis combines adjectival with nominal modifiers, for which we found the opposite trend, that is a slight increase of variation in more HF dependencies. A possible reason for this discrepancy is that Levshina's analysis also includes numeral modifiers in the same overall noun‐headed dependency type, and to the extent that numerals behave like adjectives, they might have strengthened the correlation despite the opposite trend among nominal modifiers.
Fig. 5

Relationship between word order direction () and variation (). The x‐axis is the estimated log odds of HF dependencies, and the y‐axis is the estimated log of variation. Each data point indicates the estimated direction and variation at a given combination of levels (DL, clause type, and dependency type), conditioned on the variation by language, language family, and area.

Relationship between word order direction () and variation (). The x‐axis is the estimated log odds of HF dependencies, and the y‐axis is the estimated log of variation. Each data point indicates the estimated direction and variation at a given combination of levels (DL, clause type, and dependency type), conditioned on the variation by language, language family, and area. For objects and obliques, we observe potentially more variation before the head (log odds of HF > 0) than after the head (Dryer, 1996; Dryer & Gensler, 2005; Hawkins, 2014). To the extent that these orders (unlike adjectives and adverbs) follow a general VO versus OV distinction at the language level (Dryer, 1992), our findings do not support Levshina's generalization of more variation among core arguments and oblique/adverbial dependencies in HI (VO) and flexible languages (i.e., log odds in our terms) than in OV languages (log odds > 0). At first sight, a possible cause for this discrepancy might be subordinate clauses, which are included in Levshina's generalization but not in our study. However, the total number of datapoints that this inclusion would add is unlikely to be so large that they would not only fully reverse the trend we see in 5 but also strengthen it at the same time. Another possibility is that Levshina includes S/V and Adv/V orders among core arguments and oblique/adverbial dependencies, respectively, but our findings on these suggest that they would at best decrease the strength of the positive trend we find, but again, not fully reverse and at the same time sharpen it. In view of this, we suspect that the most likely reason for the discrepancy across all dependencies is that entropy measures might have increased error rates in sparse regions (Chao et al., 2013; DeDeo et al., 2013) and that language‐level estimates of word order types might miss finer distinctions between individual dependencies and their ordering. More methodological research, ideally with simulation studies, is needed to fully resolve the issue.

Conclusions

Unlike previous work, we estimated word order and word order variation at the level of individual dependencies. This revealed substantially more diversity than current theories would predict. Our results support the general expectation of DL effects on word order only for two out of six types of dependencies: obliques (e.g., “over the weekend” in “read the book over the weekend”) and nominal modifiers (e.g., “of the cabin” in “the owner of the cabin”) show an increase of final and more rigid placement, with the difference being driven mostly by the distinction between single‐word and multiword dependents. While this is consistent with theories of DLM and theories that favor early placement of simple items, all these theories overgenerate, since they make the same predictions for other dependency types as well. Yet we found only very partial or no effects in the other types: Objects and adverbs show an effect for word order variation, but not head placement; adjectives show a very weak effect for head placement but an (equally weak) inverse effect for word order variation; subjects show no effect for either measure. Overall then, this suggests limitations of current theories and invites more detailed experimental and corpus‐based research on why DL effects are limited to certain dependency types. Given that the dependency types that we conditioned on rest on a combination of syntactic with semantic criteria (de Marneffe & Manning, 2008; Nivre, 2015), we propose that future research might profit from relativizing expected effects to specific kinds and degrees of dependencies (e.g. in terms of mutual information). At the same time, future models need to take into account the major difference we found between single‐word and multiword dependencies, and they might need to further control for interaction effects between dependency types at the level of individual sentences or entire languages. This invites a reconceptualization of research on how syntactic complexity interacts with order and, more specifically of research on DLM. Rather than assessing effects at the level of entire grammars, we submit that future theories need to carefully disentangle effects from individual dependencies and effects from the interaction between dependencies. In other words, progress might need a similar move from holistic grammars to specific properties and their interactions that have characterized progress in linguistic typology (Bickel, 2007). As our results suggest, holistic research might miss important differences between dependencies.

Open Research Badges

This article has earned Open Data and Open Materials badges. Data are available at https://universaldependencies.org/ and materials are available at https://osf.io/4s67a/.
  23 in total

1.  Locality and parsing complexity.

Authors:  L Konieczny
Journal:  J Psycholinguist Res       Date:  2000-11

2.  Motion events in language and cognition.

Authors:  Silvia P Gennari; Steven A Sloman; Barbara C Malt; W Tecumseh Fitch
Journal:  Cognition       Date:  2002-02

3.  Learning biases predict a word order universal.

Authors:  Jennifer Culbertson; Paul Smolensky; Géraldine Legendre
Journal:  Cognition       Date:  2011-12-28

4.  Euclidean distance between syntactically linked words.

Authors:  Ramon Ferrer i Cancho
Journal:  Phys Rev E Stat Nonlin Soft Matter Phys       Date:  2004-11-30

5.  The natural order of events: how speakers of different languages represent events nonverbally.

Authors:  Susan Goldin-Meadow; Wing Chee So; Asli Ozyürek; Carolyn Mylander
Journal:  Proc Natl Acad Sci U S A       Date:  2008-07-01       Impact factor: 11.205

Review 6.  Linguistic complexity: locality of syntactic dependencies.

Authors:  E Gibson
Journal:  Cognition       Date:  1998-08

7.  A noisy-channel account of crosslinguistic word-order variation.

Authors:  Edward Gibson; Steven T Piantadosi; Kimberly Brink; Leon Bergen; Eunice Lim; Rebecca Saxe
Journal:  Psychol Sci       Date:  2013-05-06

Review 8.  How Efficiency Shapes Human Language.

Authors:  Edward Gibson; Richard Futrell; Steven P Piantadosi; Isabelle Dautriche; Kyle Mahowald; Leon Bergen; Roger Levy
Journal:  Trends Cogn Sci       Date:  2019-04-18       Impact factor: 20.229

9.  Expectation and Locality Effects in German Verb-final Structures.

Authors:  Roger P Levy; Frank Keller
Journal:  J Mem Lang       Date:  2013-02-01       Impact factor: 3.059

10.  Strong expectations cancel locality effects: evidence from Hindi.

Authors:  Samar Husain; Shravan Vasishth; Narayanan Srinivasan
Journal:  PLoS One       Date:  2014-07-10       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.