Literature DB >> 34469016

Individual differences in representational similarity of first and second languages in the bilingual brain.

Emily S Nichols1,2, Yue Gao3, Sofia Fregni4, Li Liu3, Marc F Joanisse2,5.   

Abstract

Current theories of bilingualism disagree on the extent to which separate brain regions are used to maintain or process one's first and second language. The present study took a novel multivariate approach to address this question. We examined whether bilinguals maintain distinct neural representations of two languages; specifically, we tested whether brain areas that are involved in processing word meaning in either language are reliably representing each language differently, and whether language representation is influenced by individual differences in proficiency level and age of acquisition (AoA) of L2. Thirty-one English-Mandarin bilingual adults performed a picture-word matching task in both languages. We then used representational similarity analysis to examine which brain regions reliably showed different patterns of activity for each language. We found that both proficiency and AoA predicted dissimilarity between language representations in several brain areas within the language network as well as several regions of the ventral visual pathway, demonstrating that top-down language knowledge and individual language experience shapes concept representation in this processing stream. The results support the model of an integrated language system in bilinguals, along with a novel description of how representations for each language change with proficiency level and L2 AoA.
© 2021 The Authors. Human Brain Mapping published by Wiley Periodicals LLC.

Entities:  

Keywords:  bilingualism; individual differences; representational similarity analysis; word recognition

Mesh:

Year:  2021        PMID: 34469016      PMCID: PMC8519873          DOI: 10.1002/hbm.25633

Source DB:  PubMed          Journal:  Hum Brain Mapp        ISSN: 1065-9471            Impact factor:   5.038


INTRODUCTION

Current theories of bilingual language processing hold that bilingual speakers coactivate their two languages during speech, and that they maintain similar, overlapping representations for both (Goldrick, Putnam, & Schwarz, 2016; Pyers & Emmorey, 2008; Spalek, Hoshino, Wu, Damian, & Thierry, 2014; Starreveld, De Groot, Rossmark, & Van Hell, 2014). In addition, past neuroimaging research has provided much evidence that a second language (L2) is processed similarly to the speaker's first language (Cao, Tao, Liu, Perfetti, & Booth, 2013; Indefrey, 2006; Perani et al., 2003; Sun et al., 2015; Wang et al., 2011). Even when L1–L2 differences exist, such as more extensive activity in L2 (Lucas, McKhann, & Ojemann, 2004; Nichols & Joanisse, 2016), there remains extensive overlap (Indefrey, 2006). This suggests that similar underlying language networks are engaged regardless of which language is being used. The concept of language coactivation in bilinguals is widely accepted, as is that of a single, integrated lexicon (van Heuven & Dijkstra, 2010; Zeng, Branigan, & Pickering, 2020). Proponents of the BIA+/BIA‐d framework also propose that there is an L1/L2 language “node,” which determines which language will be accessed. While this model accounts for cross‐language priming and interference (Ando, Matsuki, Sheridan, & Jared, 2015; Jared & Szucs, 2002; Jouravlev, Lupker, & Jared, 2014; Midgley, Holcomb, van Heuven, & Grainger, 2008; Smits, Martensen, Dijkstra, & Sandra, 2006), the mechanism of the language node remains unclear at the neurobiological level. While neuroimaging provides much support for an integrated lexicon through activation of similar structures, the separation of patterns of activity within the shared L1/L2 brain areas may provide evidence for some degree of distinction of how the lexicon is accessed between L1 and L2. That is, these separable representational patterns may reflect the action of a regulating ‘language node’ that allows access to language‐specific semantic information within shared brain regions. Proponents of the various current models of bilingual word recognition generally agree that individual differences such as proficiency and age of acquisition (AoA) influence bilingual lexico‐semantic processing based on the strength of the associations between phonology, orthography, and semantics. For example, within the BIA+/BIA‐d framework, in unbalanced bilinguals resting activation levels (i.e., the parameter assigned to word nodes within the model, reflecting the frequency of their usage) are lower in the nondominant language (often L2) leading to slower word recognition. However, as bilinguals become balanced in proficiency, resting activation levels become similar between languages, leading to symmetrical priming effects across languages. Similarly, within the Multilink framework (Dijkstra et al., 2019), a computational model of bilingual word recognition, word forms are characterized by a frequency‐dependent resting‐level activation. As bilinguals with different L2‐proficiency have used L2 words more or less frequently, this implies differences in the resting level activation for L2 words. In addition, task demands are assumed to be distributed throughout the processing pathway; that is, different sets of elementary processing operations are activated depending on the task requirements (Dijkstra & van Heuven, 2002). It is possible then, that differences in proficiency and AoA would place different task demands on speakers, in turn leading to greater separation within the areas responsible for task demands. With more similar resting activation levels (i.e., at higher proficiencies and earlier AoAs), task demands become more similar, potentially leading to more overlap in these regions. It is important to note that both models are based on reading; however, it has been assumed that they are also valid for bilingual auditory word recognition (Dijkstra & van Heuven, 2002). Despite L1 and L2 sharing a network of structures, traditional univariate contrasts cannot tell us how languages are being represented in those areas, and while there is extensive overlap in brain areas that represent L1 and L2 (Chee, Tan, & Thiel, 1999; Indefrey, 2006; Perani et al., 2003; Wang et al., 2011), how the languages are represented may vary. That is, regions coding for language‐specific information, such as spoken codes (e.g., left superior temporal gyrus and left inferior frontal gyrus) may represent each language differently. In contrast, regions involved in executive and attentional control (e.g., dorsolateral prefrontal cortex and insula) are likely to show less differentiation in how each language is represented as the function of these regions should not differ qualitatively from one language to another. Individual differences in language ability and experience also play an important role in bilingual language processing (Newman, Tremblay, Nichols, Neville, & Ullman, 2012; Nichols & Joanisse, 2016, 2017) and may affect the integration of the neural representation of each language. Previous research indicates that low proficiency speakers and late L2 learners have greater separation of their two languages' conceptual knowledge (Van Hell & Tanner, 2012), and this separation may also be reflected in the neural representation of words and concepts within co‐activated brain areas. Representational similarity analysis (RSA) is one of several multivariate functional magnetic resonance imaging (fMRI) analysis techniques relying on reproducible spatial patterns of activity that correlate with distinct experimental conditions (Kriegeskorte, Mur, & Bandettini, 2008). RSA has been used to reveal differences between conditions within individual brain regions that were previously undetectable using standard univariate methods; it reveals cortical patterns sensitive to differences in stimuli even when the degree of activation is similar (Connolly et al., 2012; Ejaz, Hamada, & Diedrichsen, 2015; Fabbri, Stubbs, Cusack, & Culham, 2016; Joanisse & DeSouza, 2014). For example, this approach has been used to identify how regions differentiate between languages during reading (Xu, Baldauf, Chang, Desimone, & Tan, 2017), supporting the view that during visual word recognition, brain regions that are similarly active for both may nevertheless maintain representationally distinct patterns for them. Similar studies of bilingual reading have found that high proficiency readers showed more similarity in the reading network than low proficiency readers (Li et al., 2019; Qu et al., 2019). However, others have found that AoA, rather than proficiency, predicted degree of similarity within the reading network, with more dissimilarity in left inferior and middle frontal regions in earlier AoAs (Ou, Li, Yang, Wang, & Xu, 2020). While these studies have demonstrated the effect of individual differences in visual word recognition, it is much less clear how this relates to language representation within the auditory word recognition pathway, how this pertains to theories of access to a shared bilingual lexicon, and the effects of proficiency and AoA on these representational differences. Considering that much of language occurs within the spoken domain, the many differences between speech and reading (Guediche, Baart, & Samuel, 2020), and the assumption that the tenets of the BIA+ hold for spoken word recognition, it is important to investigate these questions. In the absence of a reading task, RSA may be particularly relevant to describing bilingual word processing, as it has the potential to identify differences between languages that were previously thought to not exist. RSA allows us to examine possible language‐processing differences in areas that are assumed to be engaged similarly for both languages, suggesting that they are representing L1 and L2 differently. In addition, by measuring individual differences in bilingual experience, it is possible to determine how the neural representation of each language changes with these measures.

Rationale for the present study

The present study examined whether brain areas involved in both L1 and L2 representationally distinguish the two languages during bilingual auditory word recognition. We recruited a unique sample of L1 English, L2 Mandarin adults living in Beijing, whose AoA and proficiency levels naturally varied (Oh et al., 2019). Participants performed a lexico‐semantic recognition task in each language, and we examined whether brain regions showed reliably different patterns of activity for each language within regions that significantly activated to both. We predicted that, consistent with models of an integrated bilingual lexicon, representational similarity would increase with increasing proficiency and earlier ages of L2 AoA due to increasingly similar resting activation levels. In contrast, areas involved in domain general cognitive processes, such as executive function, were not expected to show language‐selective patterns.

MATERIAL AND METHODS

Participants

Thirty‐two (13 female) neurologically healthy right‐handed native speakers of English were recruited via posters and word of mouth in Beijing, China. All participants were second‐language learners of Mandarin, aged 18–37 (M = 23.84, SD = 4.59), and had begun learning Mandarin between the ages of 0 and 28 years (M = 18.09, SD = 7.10). This study was approved by the Beijing Normal University research ethics board and all participants gave informed consent prior to participation. Demographic and language information is summarized in Table 1.
TABLE 1

Participant demographic and language information

Measure M (SD)
N 32
Sex13 female, 19 male
Age (years)23.84 (4.59)
Age of L2 acquisition18.90 (7.10)
Time spent speaking Mandarin (years)5.75 (5.55)
Country of origin (n)
United States16
United Kingdom9
Canada3
India1
Japan1
New Zealand1
Singapore1
Proficiency (%)
English88.93 (5.88)
Mandarin38.54 (18.15)
Reaction time (ms)
English1,203.60 (202.21)
Mandarin1,607.88 (231.78)
Accuracy (%)
English94.17 (4.21)
Mandarin83.07 (10.35)
Participant demographic and language information

Behavioral tests

L1 English and L2 Mandarin proficiency levels were assessed prior to scanning using a subset of 48 questions from the Test of English as a Second Language (ETS, Princeton, NJ) and 48 questions from the Hanyu Shuiping Kaoshi measure of Mandarin (HSK Centre, Beijing, China), respectively. Both tests consisted of three sections, grammar, reading comprehension, and vocabulary, which were combined to give a final score for each language, representing overall proficiency in these three domains. AoA was obtained by self‐report, defined as the age at which individuals first began learning Mandarin. To verify handedness, participants completed an abridged version of the Edinburgh Handedness Inventory (Oldfield, 1971). Behavioral measures, informed consent and task instructions were administered in English, aside from the Mandarin proficiency test, which was administered in Mandarin.

fMRI task

Participants completed a picture–word matching task during scanning, in alternating runs of English and Mandarin. This task has been used extensively in the past to study lexical knowledge (Nichols & Joanisse, 2016; Weniger, Crelier, Alkadhi, & Kollias, 2000; Breining et al., 2014; Dräger et al., 2004), and was chosen in order to examine lexical‐semantic processing. Pictures were presented via LCD projector to the center of a screen mounted at the head of the scanner bore, which was viewed through a mirror placed above the head coil. At the same time, a word was played binaurally through insert earphones (Sensimetrics Corporation, Malden, MA). Participants were required to indicate as quickly as possible with a button press whether the picture and word matched. Each picture was visible for 2.5 s. They viewed a fixation crosshair between trials as baseline. Stimulus presentation and response recording was controlled with E‐Prime software (Psychology Software Tools, Inc., Sharpsburg, PA) and a Windows laptop. The scanning session was divided into eight alternating English and Mandarin runs. Four English runs were interleaved with four Mandarin runs, with starting language counterbalanced, so that a run in the first language was always followed by a run in the other language. Four orderings were produced: one version starting with English, one version starting with Mandarin, and an additional version of each in which runs were presented in the reverse order. Each run began with an image reminding participants of which buttons to respond with, and the language in which the next run would be performed. Each run consisted of 20 trials for a total of 160 trials (80 in each language, with 40 matching and 40 mismatching). A short break was provided between each 3.5‐min scanning run. Each image appeared twice during the experiment, once in a matching pair and once in a semantically unrelated mismatching pair, with order counterbalanced so matching and unrelated trials occurred equally often in either order. Each trial was 2.5 s in duration, with intertrial interval jittered between 2.5 and 12.5 s in 2.5 s increments, to optimize the deconvolution of the blood oxygen‐level‐dependent signal. Stimulus words consisted of 40 common single‐word concepts, retrieved using Google Image Search, with the constraint that they are expressed as single two‐syllable words in both English and Mandarin, and have frequencies greater than 40 per million in both languages (English: CELEX Lexical Database (Baayen, Piepenbrock, & Gulikers, 1995) and Mandarin: SUBTLEX‐CH (Cai & Brysbaert, 2010)). In a separate pilot study involving different participants (42 native speakers of English and 33 native speakers of Mandarin), we asked groups of native speakers of English or Mandarin to rate the imageability and familiarity of the stimulus words, as well as the correspondence of the pictures to target words, on a Likert scale of 1–7. Both groups showed equally high ratings on familiarity (M Mandarin = 5.78, M English = 5.48) and picture/word correspondence (M Mandarin = 6.08, M English = 5.95).

Data acquisition and processing

Imaging was conducted on a Siemens Magnetom TIM Trio whole‐body 3 Tesla scanner with a 32‐channel head coil. T2*‐weighted functional scans were acquired in the transverse plane with 45 slices per volume (TR = 2.5 s; TE = 38 ms; flip angle = 80°; FOV = 192 × 192 mm; voxel size 3 × 3 × 3 mm3) using an iPAT parallel acquisition sequence (generalized autocalibrating partially parallel acquisition [GRAPPA]; acceleration factor = 2), providing full coverage of the cerebrum and the superior portion of the cerebellum. A total of 576 functional scans were acquired for each participant over 8 runs (3.5 min per run). After the final functional run, a whole‐head high‐resolution 3D anatomical scan was acquired in the sagittal plane, using a 3D pulse sequence weighted for T1 contrast (MPRAGE; TR = 2.3 s; TE = 2.98 ms; FOV = 256 × 256 mm; voxel size = 1 mm3; 176 slices; GRAPPA acceleration factor = 2). Raw data were converted from DICOM to BIDS format and preprocessed using FMRIPREP version 1.0.0 (Gorgolewski et al., 2018) a Nipype (Gorgolewski et al., 2011, 2018) based tool. Each T1‐weighted volume was corrected for bias field using N4BiasFieldCorrection v2.1.0 (Tustison et al., 2010) and skull‐stripped using antsBrainExtraction.sh v2.1.0 (using OASIS template). Cortical surface was estimated using FreeSurfer v6.0.0 (Dale, Fischl, & Sereno, 1999). The skull‐stripped T1w volume was coregistered to skull‐stripped ICBM 152 Nonlinear Asymmetrical template version 2009c (Fonov, Evans, McKinstry, Almli, & Collins, 2009) using nonlinear transformation implemented in ANTs v2.1.0 (Avants, Epstein, Grossman, & Gee, 2008). Functional data were slice time corrected using AFNI (Cox, 1996) and motion corrected using MCFLIRT v5.0.9 (Jenkinson, Bannister, Brady, & Smith, 2002). This was followed by co‐registration to the corresponding T1‐weighted volume using boundary‐based registration with 9 degrees of freedom—implemented in FreeSurfer v6.0.0 (Greve & Fischl, 2009). Motion‐correcting transformations, T1‐weighted transformation and MNI template warp were applied in a single step using antsApplyTransformations v2.1.0 with Lanczos interpolation. Three tissue classes were extracted from T1w images using FSL FAST v5.0.9 (Zhang, Brady, & Smith, 2001). Voxels from cerebrospinal fluid and white matter were used to create a mask which was in turn used to extract physiological noise regressors using aCompCor (Behzadi, Restom, Liau, & Liu, 2007). The mask was eroded and limited to subcortical regions to limit overlap with gray matter, and six principal components were estimated. Frame‐wise displacement (Power, Schlaggar, & Petersen, 2014) was calculated for each functional run using Nipype implementation. For more details of the pipeline, see https://fmriprep.readthedocs.io/en/latest/workflows.html.

First‐ and second‐level statistics

Single‐subject statistical maps were formed in the context of the general linear model using the AFNI 3dDeconvolve function. Linear trends in the functional data were removed, and first‐level analysis was conducted by modeling all English trials together and all Mandarin trials together. The statistical maps were formed in the context of the general linear model using AFNI 3dDeconvolve function. Additional regressors were included for the six motion parameters, an estimate of physiological noise from the preprocessing step, and trial‐wise response times. This yielded one English and one Mandarin statistical map per subject that we used to compute the second‐level (cross‐subject) univariate contrasts. One sample t tests against zero were computed for each language (AFNI 3dttest++, results shown in Figure S1 and Table S1) and a conjunction analysis (AFNI 3dcalc) was performed to identify areas that significantly activated for both English and Mandarin thresholded at 2.596 (p = 0.01 uncorrected). Note a relatively liberal threshold was used at this stage in order to include as many areas in the multivariate search space as possible. A brain mask was then created using the results of this conjunction analysis. Finally, first‐level single‐subject statistics were recomputed for English and Mandarin, this time creating separate models for even and odd runs. In addition, results of a between language t tests are shown in Figure S2 and Table S2. Only correct trials were included in both first‐level analyses, with accuracy ranging from 81.25% to 100% correct on the English task and from 61.25% to 96.25% correct on the Mandarin task.

Split‐half correlation searchlight analysis

Searchlight RSA was then performed to identify regions in which the representations of L1 and L2 were reliably different, regardless of groupwise differences in activation levels. The search space for the analysis was constrained to regions within the English–Mandarin conjunction mask, shown in Figure 1. A split‐half correlation searchlight was performed within the CoSMoMVPA Matlab toolbox (Oosterhof, Connolly, & Haxby, 2016), using a search sphere radius of three voxels (Connolly et al., 2012). Within each searchlight sphere Pearson correlations were performed for activity patterns between even and odd runs, within‐language (English–English and Mandarin–Mandarin) and between‐language (English–Mandarin), yielding a 2 × 2 similarity matrix for each individual at each point of the searchlight. Next, the degree of dissimilarity of between‐language versus within‐language patterns (on‐diagonal versus off‐diagonal) was computed using a pairwise t test based on the difference of Fisher‐transformed mean correlations (Haxby et al., 2001). Significant differences in an area within the searchlight sphere indicated this region differentially encodes L1 and L2. The center of the searchlight was then moved to the next location of the search space, and the statistical analysis was repeated, ultimately yielding a statistical map of all voxels falling within the conjunction map. Analyses were performed using subject‐wise coefficient maps in MNI space. Once single‐subject searchlight results were computed, a group statistic was created via a one‐sample t test, which identified voxels showing significantly greater representational similarity within‐language than between‐languages, across all subjects. Next, we computed random‐effect cluster statistics corrected for multiple comparison (cosmo_montecarlo_cluster_stat) with a mean of zero under the null hypothesis and 10,000 iterations, and significant clusters were converted to z scores.
FIGURE 1

Areas that significantly activated for both L1 English and L2 Mandarin at p = .01 uncorrected. Results are overlaid on a stereotaxic brain in MNI space. L = left, R = right

Areas that significantly activated for both L1 English and L2 Mandarin at p = .01 uncorrected. Results are overlaid on a stereotaxic brain in MNI space. L = left, R = right

Regression with proficiency and AoA

We then conducted a linear regression to examine whether AoA and the difference in proficiency level between L1 and L2 predicted the degree of representational dissimilarity within‐subject. Two linear models were constructed, the first with the difference in proficiency scores between L1 and L2 as a continuous regressor and adjusting for AoA, the second with AoA as a continuous regressor and adjusting for the difference in proficiency scores. The minimum cluster‐size threshold was determined in two steps. First, we estimated the smoothness of the residuals for each subject output by 3dDeconvolve using the autocorrelation function (ACF) option (AFNI 3dFWHMx), and the mean smoothness level was calculated. Next, minimum cluster size was determined using a 10,000 iteration Monte Carlo simulation (AFNI 3dClustSim) at a voxelwise alpha level of p = .01, using bi‐sided thresholding and first‐nearest neighbor clustering. Correction for multiple comparisons at p = .01 was achieved by setting a minimum cluster size of 7 voxels.

RESULTS

Behavioral

Performance on the L1 (English) proficiency test ranged from 72.92 to 100%, and performance on the L2 (Mandarin) proficiency test ranged from 12.5 to 77.08%. Analysis of the proficiency test data acquired prior to scanning indicated that L2 proficiency was significantly lower than L1 proficiency (M = 88.93%, SD = 5.88, M = 38.54%, SD = 18.15, respectively; t[31] = −15.93, p < .001, 95% CI [43.94, 56.84]). L2 proficiency did not significantly correlate with L2 AoA (r[30] = −0.21, p = .255). Participants responded faster on English trials than Mandarin trials (M = 1,203.60 ms, SD = 202.21, M = 1,607.88 ms, SD = 231.78, respectively; t[31] = −14.67, p < .001, 95% CI [−460.48, −348.09]) and were more accurate on English trials than Mandarin trials (M = 94.17%, SD = 4.21, M = 83.07%, SD = 10.35; t[31] = 6.84, p < .001, 95% CI [7.78, 14.40]).

Conjunction analysis

Results of the conjunction analysis are shown in Figure 1 and Table 2. Both L1 English and L2 Mandarin produced significant activation at a voxelwise p value of 0.01 (uncorrected) in an extensive network of bilateral brain regions including the Heschl's gyrus, superior temporal gyrus (STG), inferior frontal gyrus (IFG), fusiform and lingual gyri, and occipital and parietal cortices.
TABLE 2

Areas of common activation revealed by a conjunction of English and Mandarin

RegionMNI coordinatesVoxels
x y z
LLingual gyrus−3−6059,969
RPlanum temporale56−144671
RPremotor cortex1346536
LMiddle frontal gyrus−451035461
RPutamen271−3113
RCerebellum19−41−4530
LTemporal fusiform cortex−32−7−3520
RInsula40−21512
RParietal operculum cortex48−322511
LPlanum Polare−45−17−26
REntorhinal cortex31−8−365
LBrainstem0−17−175
LPrecentral gyrus−59−23505
LInsula−37074
RInsula38754
LParietal operculum cortex−43−37164
RBrainstem14−30−273
LPars triangularis−573333
RAnterior intraparietal sulcus39−42473
RPremotor cortex15−12813

Note: Coordinates denote the center of mass. L/R = left/right.

Areas of common activation revealed by a conjunction of English and Mandarin Note: Coordinates denote the center of mass. L/R = left/right.

Searchlight with split‐half correlation analysis

Group‐level RSA searchlight

As a group, no regions showed significantly greater representational similarity within‐language (Mandarin–Mandarin and English–English) compared to between‐language (Mandarin–English).

Regression with proficiency and AoA

In order to determine whether proficiency or AoA predicted the degree of difference in representational similarity within‐ and between‐language, subject‐wise searchlight maps were next submitted to linear regression analysis. L1–L2 proficiency score difference predicted greater within‐language representational similarity than between‐language similarity in several areas including the left fusiform, IFG, bilateral STG, and right lingual gyrus, shown in Figure 2 and Table 3. All areas showed a positive relationship, indicating that as the difference in proficiency between languages increased, so did the degree of difference in representation between English and Mandarin.
FIGURE 2

The relationship between difference in L1 − L2 proficiency scores (%) and L1 − L2 similarity. z‐Score values represent the mean across the entire region of interest (ROI). Higher z‐scores indicate greater differences between correlation values between‐language versus within‐language. Statistical maps are thresholded at p = .01, overlaid on an MNI brain atlas. L = left, R = right. Cluster locations and sizes are reported in Table 3

TABLE 3

Regions where proficiency or age of acquisition (AoA) significantly predicted z score

PredictorRegionMNI coordinatesVoxels t p
x y z
L1–L2 proficiencyRCerebellum6−81−30355.23<.001
LOccipital fusiform gyrus−33−75−18964.72<.001
LSupramarginal gyrus−66−4521134.20<.001
LPrecentral gyrus−33−966124.10<.001
LPars opercularis−39324204.07<.001
RCerebellum33−66−48254.03<.001
RMiddle occipital gyrus51−81093.97<.001
LAnterior intra‐parietal sulcus−36−484293.90.001
RPrimary visual cortex18−609263.88.001
RInferior temporal gyrus48−51−24123.76.001
LLingual gyrus−15−45−9133.71.001
RSupramarginal gyrus66−4524173.70.001
ROrbitofrontal cortex3333−3103.63.001
LPars triangularis−51331883.61.001
RVisual cortex ventral V321−78−683.51.001
RCerebellum3−57−4573.31.003
AoARVisual cortex ventral V342−93−685.09<.001
LMiddle temporal gyrus−54−24−994.34<.001
LAnterior intraparietal sulcus−39−454815−3.50.002
RInsula3921016−3.46.002
RCalcarine sulcus18−5198−3.32.002

Note: Coordinates denote the location of peak activation. L/R = Left/Right. P values have been cluster‐size corrected to p < .05.

The relationship between difference in L1 − L2 proficiency scores (%) and L1 − L2 similarity. z‐Score values represent the mean across the entire region of interest (ROI). Higher z‐scores indicate greater differences between correlation values between‐language versus within‐language. Statistical maps are thresholded at p = .01, overlaid on an MNI brain atlas. L = left, R = right. Cluster locations and sizes are reported in Table 3 Regions where proficiency or age of acquisition (AoA) significantly predicted z score Note: Coordinates denote the location of peak activation. L/R = Left/Right. P values have been cluster‐size corrected to p < .05. The relationship between AoA and representational similarity is shown in Figure 3 and Table 3. AoA positively predicted greater within‐language than between‐language representational similarity in the left middle temporal gyrus and right inferior occipital gyrus, indicating that later AoAs were associated with larger differences between L1 and L2 in these areas. In contrast, AoA showed a negative correlation with the left inferior parietal lobe and right insula and calcarine sulcus, indicating that earlier AoAs were associated with smaller L1–L2 representational differences in these areas.
FIGURE 3

The relationship between L2 age of acquisition (AoA) and L1 − L2 dissimilarity. z‐Score values represent the mean across the entire region of interest (ROI). Higher z‐scores indicate greater differences between correlation values between‐language versus within‐language. Statistical maps are thresholded at p = .01, overlaid on an MNI brain atlas. L = left, R = right. Cluster locations and sizes are reported in Table 3

The relationship between L2 age of acquisition (AoA) and L1 − L2 dissimilarity. z‐Score values represent the mean across the entire region of interest (ROI). Higher z‐scores indicate greater differences between correlation values between‐language versus within‐language. Statistical maps are thresholded at p = .01, overlaid on an MNI brain atlas. L = left, R = right. Cluster locations and sizes are reported in Table 3

DISCUSSION

The present study investigated the hypothesis that bilinguals maintain similar, overlapping lexical representations for both of their languages. Using a lexico‐semantic recognition task, we found both similarity and dissimilarity in the representation of bilinguals' two languages within the bilingual word recognition network. There were no regions that significantly differed in their representation of English and Mandarin at the group level, however both proficiency and AoA predicted the degree of representational similarity in several areas. That is, individual differences predicted differentiation in the representation of bilinguals' two languages in areas that were significantly activated during the word recognition task in both Mandarin and English. These results extend behavioral and event‐related potential findings that bilinguals have a single, integrated lexicon (Ando et al., 2015; Jouravlev et al., 2014; Midgley et al., 2008; van Heuven & Dijkstra, 2010), demonstrating how the neural representations within activated regions change with language experience. While prior meta‐analyses and reviews have argued this on the basis of relative intensity of fMRI activity (Indefrey, 2006; van Heuven & Dijkstra, 2010), degree of activation cannot tell us about how each language is being represented. It can also be difficult to interpret the meaning of greater activity: increased fMRI signal may represent a greater specialization of a region to a certain task, or it might alternatively indicate increased neural effort required due to weaker familiar with that task. Here, we reconceptualize regional activation in terms of degree of language differentiation, focusing on regions that do not show clear specialization for one language or the other. Consistent with our hypotheses, several regions of the language network showed patterns of representation that differentiated languages depending on individual differences. For example, one of these regions was the left IFG (including both the pars opercularis and the pars triangularis), an area engaged in representing and planning articulatory codes for speech and tone (Broca, 1865; Fedorenko, Fillmore, Smith, Bonilha, & Fridriksson, 2015; Gao, Guo, Liu, Mo, & Wang, 2020; Wise, Greene, Buchel, & Scott, 1999). Indeed, these features differ between English and Mandarin in that each language has phonological features that are not present in the other (e.g., tone in Mandarin, consonant clusters in English). The left IFG showed greater representational similarity between languages when the difference between L1 and L2 proficiency was smaller, suggesting that as bilinguals become more matched in proficiency across their two languages, the phonological representations become more integrated. Similarly, language similarity within the bilateral supramarginal gyrus was greater with smaller proficiency differences, an area important for auditory‐motor integration during word recognition (Catani, Jones, & Ffytche, 2005). One notable result was that of representational dissimilarity in lower L2 proficiency and later AoA speakers throughout the ventral visual stream, a cortical pathway responsible for object recognition and concept representation (Martin, Douglas, Newsome, Man, & Barense, 2018). The language differentiation in visual areas is especially interesting as participants in the present study saw the same images in each language; the manipulation here was only the language in which they heard the names of these objects. As a result, language‐dependent differences in this region suggest a top‐down modulation of high‐level visual processing by the linguistic input. Although visual processing of the same images may appear to be a domain‐general process, support for it being language‐specific comes from the label‐feedback hypothesis, which suggests that language modulates ongoing cognitive and perceptual processing (Lupyan, 2012). In line with this hypothesis, each language's verbal label for the paired image influences the perception of that image. Thus, while the image remains the same, the top‐down influence of the language is producing separable representations in high‐level visual areas, distinguishing between the visual perception of the spoken word table versus that of the spoken word 桌子 (the Mandarin word for table) (Jared, Poh, & Paivio, 2013). In addition to areas showing a positive relationship between AoA and representational similarity, three regions showed greater dissimilarity at earlier AoAs. The right insula has been implicated in inhibitory control (Cai et al., 2014), and larger differences in neural representation within this area may reflect differences in how inhibition of the other language is performed. One tenet of the BIA‐d Model is that once conceptual connections are in place, inhibitory connections between translation equivalents develop, allowing for improved processing (Grainger et al., 2010). Thus, these inhibitory connections, more likely to have had time to develop in earlier bilinguals, may be controlled though the right insula. Interestingly, the left anterior intraparietal sulcus, another area that showed a negative relationship between representational dissimilarity and AoA, is important in the planning and control of fine finger movement, and has strong connections to the bilateral insula (Uddin et al., 2010). The third region, the area prostriata of the calcarine sulcus, has been shown to have cortical connections to the cingulate motor cortex in monkeys (Morecraft, Rockland, & Van Hoesen, 2000). While the reason for these links with AoA is less clear, it is possible that the process of initiating a response, done through a button press, was more separable for participants with greater inhibitory connections between languages. Future research should aim to further elucidate the role of these areas in bilingual auditory word recognition, including their relationship with AoA. Finally, although it may seem surprising that no regions showed higher within‐language representational similarity, this result is likely due to the high degree of variability in AoA and proficiency in our sample. Based on our predictions that individual differences in these factors will predict degree of similarity, it then logically follows that when considering everyone together, no consistent pattern would emerge. Even the external properties of the stimuli that differ between languages, such as the phonology of tone that exists in Mandarin and English, have been shown to be affected by individual differences (e.g., Yu et al., 2019), perhaps explaining why we still did not see higher within‐language representational similarity within‐language than between‐language. There have been numerous studies showing activation differences between L1 and L2, showing greater activation in language areas for one language versus another (Chee, Hon, Lee, & Soon, 2001; Ding et al., 2003; Perani et al., 2003), or showing additional areas recruited for L2 processing versus L1 processing (Wang et al., 2011). These differences have largely been attributed to later acquisition of L2, differences in proficiency, or other external factors affecting how L2 was acquired (Chee et al., 2001; De Bleser et al., 2003; Indefrey, 2006). In contrast, matched bilinguals tend to show overlapping activity in language regions, with little or no differentiation between languages at the univariate level (Hernandez, Dapretto, Mazziotta, & Bookheimer, 2001; Hernandez, Martinez, & Kohnert, 2000; Rodriguez‐Fornells et al., 2005). L2 speakers in the present study showed experience‐dependent representational differences between L1 and L2 in both the language network as well as throughout the ventral visual stream, providing further evidence for integration of bilinguals' two languages but only when speakers are matched in ability across those two languages. This represents an important but underexplored perspective on how neuroplasticity is expressed in bilingualism.

CONCLUSION

We investigated first and second language representation in English‐Mandarin bilinguals. Using RSA, we identified both regions in which individual differences predicted differentiation in representation between English and Mandarin. Experience‐modulated within‐language representational similarity was present in language‐network areas (e.g., portions of the left IFG) as well as several regions of the ventral visual pathway. The results thus support a model in which the degree to which bilinguals maintain separable representations of their two languages varies by brain region, and that the strength of such effects also varies appreciably as a function of both maturational and experiential factors. A logical extension of present study is the examination of representational differences in different types of second language processing. For instance, results may differ when comparing two languages that are more similar than English and Mandarin, such as Spanish and French, or when using items that vary in similarity, such as cognates and noncognates. In addition, word processing does not involve grammatical processing, which is also an important aspect of bilingual language processing that can differ greatly between L1 and L2. Univariate approaches that contrast degree of brain activation may miss important differences in this regard. The multivariate approach used here may thus provide a way forward in our ability to fully discern how L1 and L2 are represented in the brain.

CONFLICT OF INTEREST

The authors declare no potential conflicts of interest.

ETHICS STATEMENT

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Beijing Normal University research ethics board. All participants gave written informed consent prior to participation. FIGURE S1 Areas that significantly activated for (a) L1 English versus Baseline and (b) L2 Mandarin versus Baseline, at a cluster‐size‐corrected threshold of p = .01. Results are overlaid on a stereotaxic brain in MNI space. L = Left, R = Right. FIGURE S2. Areas that showed significantly different levels of activation for L1 English versus L2 Mandarin, at a cluster‐size‐corrected threshold of p = .01. Results are overlaid on a stereotaxic brain in MNI space. L = Left, R = Right. TABLE S1. Areas of significant activation for L1 English versus Baseline and L2 Mandarin versus Baseline at a voxel‐wise alpha of p = .01 TABLE S2. Clusters of significant activation in L1 English – L2 Mandarin contrast at a voxel‐wise alpha of p = .01 Click here for additional data file.
  53 in total

1.  Localization of area prostriata and its projection to the cingulate motor cortex in the rhesus monkey.

Authors:  R J Morecraft; K S Rockland; G W Van Hoesen
Journal:  Cereb Cortex       Date:  2000-02       Impact factor: 5.357

2.  Relative language proficiency modulates BOLD signal change when bilinguals perform semantic judgments. Blood oxygen level dependent.

Authors:  M W Chee; N Hon; H L Lee; C S Soon
Journal:  Neuroimage       Date:  2001-06       Impact factor: 6.556

3.  Effects of native language experience on Mandarin lexical tone processing in proficient second language learners.

Authors:  Keke Yu; Li Li; Yuan Chen; Yacong Zhou; Ruiming Wang; Yang Zhang; Ping Li
Journal:  Psychophysiology       Date:  2019-07-29       Impact factor: 4.016

4.  The influence of language proficiency on lexical semantic processing in native and late learners of English.

Authors:  Aaron J Newman; Antoine Tremblay; Emily S Nichols; Helen J Neville; Michael T Ullman
Journal:  J Cogn Neurosci       Date:  2011-10-07       Impact factor: 3.225

5.  Functional activity and white matter microstructure reveal the independent effects of age of acquisition and proficiency on second-language learning.

Authors:  Emily S Nichols; Marc F Joanisse
Journal:  Neuroimage       Date:  2016-08-25       Impact factor: 6.556

6.  Language switching and language representation in Spanish-English bilinguals: an fMRI study.

Authors:  A E Hernandez; M Dapretto; J Mazziotta; S Bookheimer
Journal:  Neuroimage       Date:  2001-08       Impact factor: 6.556

7.  Earlier second language acquisition is associated with greater neural pattern dissimilarity between the first and second languages.

Authors:  Jian Ou; Wenlong Li; Yang Yang; Nizhuan Wang; Min Xu
Journal:  Brain Lang       Date:  2020-01-23       Impact factor: 2.381

8.  Dissociable roles of right inferior frontal cortex and anterior insula in inhibitory control: evidence from intrinsic and task-related functional parcellation, connectivity, and response profile analyses across multiple datasets.

Authors:  Weidong Cai; Srikanth Ryali; Tianwen Chen; Chiang-Shan R Li; Vinod Menon
Journal:  J Neurosci       Date:  2014-10-29       Impact factor: 6.167

9.  Sensitivity of human auditory cortex to rapid frequency modulation revealed by multivariate representational similarity analysis.

Authors:  Marc F Joanisse; Diedre D DeSouza
Journal:  Front Neurosci       Date:  2014-09-30       Impact factor: 4.677

10.  Individual differences in representational similarity of first and second languages in the bilingual brain.

Authors:  Emily S Nichols; Yue Gao; Sofia Fregni; Li Liu; Marc F Joanisse
Journal:  Hum Brain Mapp       Date:  2021-09-01       Impact factor: 5.038

View more
  1 in total

1.  Individual differences in representational similarity of first and second languages in the bilingual brain.

Authors:  Emily S Nichols; Yue Gao; Sofia Fregni; Li Liu; Marc F Joanisse
Journal:  Hum Brain Mapp       Date:  2021-09-01       Impact factor: 5.038

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.