| Literature DB >> 32161560 |
Abstract
Spoken word recognition involves a perceptual tradeoff between the reliance on the incoming acoustic signal and knowledge about likely sound categories and their co-occurrences as words. This study examined how adult second language (L2) learners navigate between acoustic-based and knowledge-based spoken word recognition when listening to highly variable, multi-talker truncated speech, and whether this perceptual tradeoff changes as L2 listeners gradually become more proficient in their L2 after multiple months of structured classroom learning. First language (L1) Mandarin Chinese listeners and L1 English-L2 Mandarin adult listeners took part in a gating experiment. The L2 listeners were tested twice - once at the start of their intermediate/advanced L2 language class and again 2 months later. L1 listeners were only tested once. Participants were asked to identify syllable-tone words that varied in syllable token frequency (high/low according to a spoken word corpus) and syllable-conditioned tonal probability (most probable/least probable in speech given the syllable). The stimuli were recorded by 16 different talkers and presented at eight gates ranging from onset-only (gate 1) through onset +40 ms increments (gates 2 through 7) to the full word (gate 8). Mixed-effects regression modeling was used to compare performance to our previous study which used single-talker stimuli (Wiener et al., 2019). The results indicated that multi-talker speech caused both L1 and L2 listeners to rely greater on knowledge-based processing of tone. L1 listeners were able to draw on distributional knowledge of syllable-tone probabilities in early gates and switch to predominantly acoustic-based processing when more of the signal was available. In contrast, L2 listeners, with their limited experience with talker range normalization, were less able to effectively transition from probability-based to acoustic-based processing. Moreover, for the L2 listeners, the reliance on such distributional information for spoken word recognition appeared to be conditioned by the nature of the acoustic signal. Single-talker speech did not result in the same pattern of probability-based tone processing, suggesting that knowledge-based processing of L2 speech may only occur under certain acoustic conditions, such as multi-talker speech.Entities:
Keywords: Mandarin Chinese; distributional learning; gating; lexical tone; second language acquisition; spoken word recognition; talker variability
Year: 2020 PMID: 32161560 PMCID: PMC7052525 DOI: 10.3389/fpsyg.2020.00214
Source DB: PubMed Journal: Front Psychol ISSN: 1664-1078
FIGURE 1Proportion of correct syllable-tone words at gate 8 (full acoustic signal). Individual participant means (points), group means (solid line), 95% confidence intervals (white box), and group density (violin) are shown.
Mixed effect logistic regression model on correct syllable-tone at Gate 8.
| (Intercept) | 3.54 | 0.34 | 10.23 | <0.001 |
| L1 Multi-talker | −0.96 | 0.43 | −2.18 | 0.029 |
| L2 Multi-talker Test 1 | −3.26 | 0.39 | −8.26 | <0.001 |
| L2 Multi-talker Test 2 | −2.99 | 0.40 | −7.50 | <0.001 |
| L2 Single-talker Test 1 | −3.26 | 0.42 | −7.67 | <0.001 |
| L2 Single-talker Test 2 | −2.54 | 0.43 | −5.96 | <0.001 |
| L1 Multi-talker – L2 Multi-talker Test 1 | 2.30 | 0.34 | 6.74 | <0.001 |
| L1 Multi-talker – L2 Multi-talker Test 2 | 2.03 | 0.34 | 5.87 | <0.001 |
| L1 Multi-talker – L2 Single-talker Test 1 | 2.29 | 0.37 | 6.11 | <0.001 |
| L1 Multi-talker – L2 Single-talker Test 2 | 1.58 | 0.37 | 4.18 | <0.001 |
| L2 Multi-talker Test 1 – L2 Multi-talker Test 2 | −0.26 | 0.11 | −2.27 | 0.207 |
| L2 Multi-talker Test 1 – L2 Single-talker Test 1 | −0.01 | 0.36 | −0.01 | 0.999 |
| L2 Multi-talker Test 2 – L2 Single-talker Test 2 | −0.45 | 0.37 | −1.25 | 0.809 |
| L2 Single-talker Test 1 – L2 Single-talker Test 2 | −0.72 | 0.39 | −1.83 | 0.443 |
FIGURE 2Proportion of correct syllable-tone words at gates 2–3, 4–5, and 6–7 by syllable token frequency and group. Error bars represent 95% confidence intervals.
Mixed effect logistic regression model on correct syllable-tone, Gates 2-3, 4-5, 6-7.
| (Intercept) | −1.42 | 0.28 | −4.99 | <0.001 |
| L1 Multi-talker | −1.48 | 0.29 | −5.03 | <0.001 |
| L2 Multi-talker Test 1 | −1.65 | 0.29 | −5.70 | <0.001 |
| L2 Multi-talker Test 2 | −1.12 | 0.29 | −3.82 | <0.001 |
| L2 Single-talker Test 1 | −2.06 | 0.32 | −6.38 | <0.001 |
| L2 Single-talker Test 2 | −1.18 | 0.31 | −3.76 | <0.001 |
| Frequency | 0.78 | 0.18 | 4.23 | <0.001 |
| L1 Multi-talker – L2 Multi-talker Test 1 | 0.18 | 0.27 | 0.66 | 0.986 |
| L1 Multi-talker – L2 Multi-talker Test 2 | −0.36 | 0.27 | −1.31 | 0.781 |
| L1 Multi-talker – L2 Single-talker Test 1 | 0.59 | 0.30 | 1.91 | 0.395 |
| L1 Multi-talker – L2 Single-talker Test 2 | −0.29 | 0.29 | −0.99 | 0.922 |
| L2 Multi-talker Test 1 – L2 Multi-talker Test 2 | −0.53 | 0.11 | −2.53 | 0.100 |
| L2 Multi-talker Test 1 – L2 Single-talker Test 1 | 0.40 | 0.30 | 1.33 | 0.763 |
| L2 Multi-talker Test 2 – L2 Single-talker Test 2 | 0.06 | 0.29 | 0.21 | 0.999 |
| L2 Single-talker Test 1 – L2 Single-talker Test 2 | −0.88 | 0.32 | −2.68 | 0.078 |
| (Intercept) | 0.15 | 0.27 | 0.54 | 585 |
| L1 Multi-talker | −1.28 | 0.29 | −4.36 | <0.001 |
| L2 Multi-talker Test 1 | −1.91 | 0.29 | −6.53 | <0.001 |
| L2 Multi-talker Test 2 | −1.61 | 0.29 | −5.49 | <0.001 |
| L2 Single-talker Test 1 | −2.11 | 0.32 | −6.59 | <0.001 |
| L2 Single-talker Test 2 | −1.41 | 0.32 | −4.46 | <0.001 |
| Frequency | 0.55 | 0.16 | 3.34 | <0.001 |
| L1 Multi-talker – L2 Multi-talker Test 1 | 0.63 | 0.267 | 2.36 | 0.170 |
| L1 Multi-talker – L2 Multi-talker Test 2 | 0.34 | 0.26 | 1.25 | 0.810 |
| L1 Multi-talker – L2 Single-talker Test 1 | 0.83 | 0.30 | 2.79 | 0.058 |
| L1 Multi-talker – L2 Single-talker Test 2 | 0.14 | 0.29 | 0.46 | 0.997 |
| L2 Multi-talker Test 1 – L2 Multi-talker Test 2 | −0.29 | 0.09 | −2.24 | 0.190 |
| L2 Multi-talker Test 1 – L2 Single-talker Test 1 | 0.20 | 0.29 | 0.68 | 0.983 |
| L2 Multi-talker Test 2 – L2 Single-talker Test 2 | −0.20 | 0.29 | −0.67 | 0.984 |
| L2 Single-talker Test 1 – L2 Single-talker Test 2 | −0.69 | 0.32 | −2.16 | 0.255 |
| (Intercept) | 1.02 | 0.26 | 3.97 | <0.001 |
| L1 Multi-talker | −0.84 | 0.27 | −3.05 | 0.002 |
| L2 Multi-talker Test 1 | −2.05 | 0.27 | −7.50 | <0.001‘ |
| L2 Multi-talker Test 2 | −1.89 | 0.28 | −6.86 | <0.001 |
| L2 Single-talker Test 1 | −2.32 | 0.30 | −7.76 | <0.001 |
| L2 Single-talker Test 2 | −1.56 | 0.30 | −5.23 | <0.001 |
| Frequency | 0.37 | 0.15 | 2.45 | 0.014 |
| L1 Multi-talker – L2 Multi-talker Test 1 | 1.21 | 0.25 | 4.90 | <0.001 |
| L1 Multi-talker – L2 Multi-talker Test 2 | 1.05 | 0.25 | 4.22 | <0.001 |
| L1 Multi-talker – L2 Single-talker Test 1 | 1.49 | 0.27 | 5.38 | <0.001 |
| L1 Multi-talker – L2 Single-talker Test 2 | 0.72 | 0.27 | 3.62 | 0.012 |
| L2 Multi-talker Test 1 – L2 Multi-talker Test 2 | −0.16 | 0.084 | −1.88 | 0.416 |
| L2 Multi-talker Test 1 – L2 Single-talker Test 1 | 0.27 | 0.27 | 0.99 | 0.922 |
| L2 Multi-talker Test 2 – L2 Single-talker Test 2 | −0.34 | 0.27 | −1.23 | 0.819 |
| L2 Single-talker Test 1 – L2 Single-talker Test 2 | −0.77 | 0.30 | −2.58 | 0.103 |
Mixed effect linear regression model on empirical log error ratio.
| (Intercept) | 0.08 | 0.04 | 1.86 | 0.065 |
| Talker | 0.31 | 0.04 | 6.92 | <0.001 |
| Window | 0.28 | 0.04 | 6.26 | <0.001 |
| Talker:Window:Frequency | 0.11 | 0.05 | 2.39 | 018 |
| Multi-talker – Single-talker | 0.62 | 0.10 | 5.87 | <0.001 |
| Early (Gates 2-3) – Late (Gates 4-7) | 0.56 | 0.09 | 5.70 | <0.001 |
| Early Multi (F−) – Early Single (F−) | 0.24 | 0.18 | 1.31 | 0.893 |
| Early Multi (F+) – Early Single (F+) | 0.93 | 0.18 | 5.17 | <0.001 |
| Late Multi (F−) – Late Single (F-) | 0.74 | 0.18 | 4.13 | 0.002 |
| Late Multi (F+) – Late Single (F+) | 0.57 | 0.18 | 3.21 | 0.036 |
| (Intercept) | 0.10 | 0.04 | 2.49 | 0.015 |
| Talker | 0.48 | 0.04 | −4.59 | <0.001 |
| Frequency | −0.14 | 0.03 | 11.35 | <0.001 |
| Multi-talker – Single-talker | 0.96 | 0.08 | 11.24 | <0.001 |
| (F−) – (F+) | 0.28 | 0.06 | 4.28 | <0.001 |
FIGURE 3Log ratio of acoustic-based and probability-based errors by group and window. Individual participant means (points), group medians (solid line), and first and third quantiles (lower and upper hinges) are shown.