Literature DB >> 29484971

Music Training Can Improve Music and Speech Perception in Pediatric Mandarin-Speaking Cochlear Implant Users.

Xiaoting Cheng1,2, Yangwenyi Liu1,2, Yilai Shu1,2, Duo-Duo Tao3, Bing Wang1,2, Yasheng Yuan1,2, John J Galvin4, Qian-Jie Fu5, Bing Chen1,2.   

Abstract

Due to limited spectral resolution, cochlear implants (CIs) do not convey pitch information very well. Pitch cues are important for perception of music and tonal language; it is possible that music training may improve performance in both listening tasks. In this study, we investigated music training outcomes in terms of perception of music, lexical tones, and sentences in 22 young (4.8 to 9.3 years old), prelingually deaf Mandarin-speaking CI users. Music perception was measured using a melodic contour identification (MCI) task. Speech perception was measured for lexical tones and sentences presented in quiet. Subjects received 8 weeks of MCI training using pitch ranges not used for testing. Music and speech perception were measured at 2, 4, and 8 weeks after training was begun; follow-up measures were made 4 weeks after training was stopped. Mean baseline performance was 33.2%, 76.9%, and 45.8% correct for MCI, lexical tone recognition, and sentence recognition, respectively. After 8 weeks of MCI training, mean performance significantly improved by 22.9, 14.4, and 14.5 percentage points for MCI, lexical tone recognition, and sentence recognition, respectively ( p < .05 in all cases). Four weeks after training was stopped, there was no significant change in posttraining music and speech performance. The results suggest that music training can significantly improve pediatric Mandarin-speaking CI users' music and speech perception.

Entities:  

Keywords:  cochlear implant; melodic contour identification; music training; pitch

Mesh:

Year:  2018        PMID: 29484971      PMCID: PMC5833165          DOI: 10.1177/2331216518759214

Source DB:  PubMed          Journal:  Trends Hear        ISSN: 2331-2165            Impact factor:   3.293


Introduction

Cochlear implantation is an effective treatment for severe-to-profound sensorineural hearing loss. While perception of nontonal (e.g., English) and tonal languages (e.g., Mandarin Chinese) is comparable in cochlear implant (CI) users, tonal language perception depends strongly on fundamental frequency (F0) cues, especially for speech segments (Lin, 1988). For the most part, F0 information is weakly represented by CI signal processing and is primarily conveyed in the temporal envelopes used to modulate pulse trains delivered to the relevant intracochlear electrodes. The functional spectral resolution is too limited to provide good pitch perception (Croghan, Duran, & Smith, 2017; Friesen, Shannon, Baskent, & Wang, 2001; Madsen, Whiteford, & Oxenham, 2017; Shannon, Fu, & Galvin, 2004). Indeed, CI users’ pitch perception is much poorer than that of normal-hearing (NH) listeners (Brockmeier et al., 2011; Gfeller & Lansing, 1991; Kong, Cruz, Jones, & Zeng, 2004), which limits CI users’ perception of music, prosody, vocal emotion, and tonal language (Brockmeier et al., 2011; Chatterjee & Peng, 2008; Gfeller & Lansing, 1991; Fu, Zeng, Shannon, & Soli, 1998; Galvin, Fu, & Nogaki, 2007; Kong et al., 2004; Luo & Fu, 2004; Luo, Fu, & Galvin, 2007). However, music training has been shown to improve CI users’ melodic pitch perception (Fu, Galvin, Wang, & Wu, 2015; Galvin et al., 2007; Galvin, Fu, & Shannon, 2009; Galvin, Eskridge, Oba, & Fu, 2012). Such melodic pitch training might also improve tonal language perception, as pitch cues are important for both listening tasks. Mandarin Chinese includes four distinctive tones: Tone 1 (high-flat), Tone 2 (high-rising), Tone 3 (falling-rising), and Tone 4 (high-falling). While F0 contributes strongly to lexical tone perception (Abramson, 1978), CI users’ may make use of other acoustic cues such as duration and amplitude that covary with F0 (Fu, Hsu, & Horng, 2004; Lin, 1988). Luo and Fu (2004) showed that manipulating the amplitude envelope to more closely resemble the F0 contour can improve lexical tone recognition. Exaggerated pitch contours have also been shown to improve CI users’ tone recognition (He, Deroche, Doong, Jiradejvong, & Limb, 2016). In terms of acoustic information, music and speech have much in common, such as pitch (F0), timbre (spectral envelope), and timing (rhythm; Kraus, Skoe, Parbery-Clark, & Ashley, 2009; Patel, 2003; Tzounopoulos & Kraus, 2009). Previous studies with adult Mandarin-speaking CI users have shown significant correlations between melodic pitch and lexical tone perception (Looi, Teo, & Loo, 2015; Wang et al., 2012; Wang, Zhou, & Xu, 2011), suggesting that both listening tasks may share a similar pitch processing mechanism. However, Tao et al. (2015) found no correlation between lexical tone perception and melodic pitch perception in prelingual or postlingual CI users, possibly due to highly variable performance and a wide age range across subjects. If there is a relationship between music and speech perception, music training may benefit Chinese CI users’ speech perception. Previous studies have observed advantages in speech perception for musicians or for people with extensive music training (Kraus et al., 2009; Parbery-Clark, Skoe, Lam, & Kraus, 2009; Patel, 2003; Tzounopoulos & Kraus, 2009). This “musician effect” may show some advantage when listening to spectrotemporally degraded signals (as in CIs), especially for pitch-mediated speech such as vocal emotion (Fuller, Galvin, Maat, Free, & Başkent, 2014). There is evidence that music training may be related to brain plasticity (Hyde et al., 2009; Kraus & Chandrasekaran, 2010). Music training has been shown to generate structural and functional changes in the brain that may benefit development of musical skills as well as general auditory skills (Besson, Schön, Moreno, Santos, & Magne, 2007; Hyde et al., 2009; Kraus & Chandrasekaran, 2010; Musacchia, Sams, Skoe, & Kraus, 2007). Chen et al. (2010) showed that music training improved pitch perception in prelingual CI children, and that the duration of training was significantly correlated with improvements in pitch perception. Vandali, Sly, Cowan, and van Hoesel (2014) found improved pitch and timbre perception in CI listeners using computer-based music training. Fu et al. (2015) also found improved melodic contour identification (MCI) performance in Mandarin-speaking pediatric CI users after MCI training with different stimuli, similar to outcomes with adult, English-speaking CI users (Galvin et al., 2007, 2009, 2012). The above studies show some relationship between music and speech, especially where pitch cues are important (e.g., vocal emotion, lexical tones, etc.). Musical training can improve music perception in NH and CI listeners. However, it is not clear whether music training can also improve CI users’ speech perception. Besides the importance of pitch cues for lexical tone perception, pitch cues are also important for language development (e.g., infant-directed speech; Trainor & Desjardins, 2002); as such, music training may be especially beneficial for pediatric Mandarin-speaking CI users. In this study, the effect of music training on music and speech perception was studied in young Mandarin-speaking CI users. Given the importance of pitch cues to both listening tasks, we hypothesized that the MCI training would improve both music and speech perception.

Materials and Methods

Ethics Statement

The study and its consent procedure were approved by the local ethics committee (Ethics Committee of Eye and Ear, Nose, and Throat Hospital of Fudan University, approval number: KY2012-009) and written informed consent was obtained from children’s parents before participation.

Participants

Sixteen (5 females and 11 males) Mandarin-speaking CI children were recruited from the Shanghai Eye and Ear, Nose, and Throat Hospital, Fudan University, China. The inclusion criteria were that all pediatric participants be prelingually deaf and diagnosed with severe-to-profound sensorineural hearing loss at ≤1 year old. The exclusion criteria were previous formal music training experience as well as any cognitive, visual, or intelligence disorders. The mean age at testing was 6.3 years (range: 4.5 to 9.3 years), the mean age at implantation was 3.4 years (range: 1.7 to 6.1 years), and the mean CI experience was 2.8 years (range: 0.8 to 6.0 years). Twelve subjects used hearing aids for nearly 6 months prior to cochlear implantation. After implantation, none of the subjects used a hearing aid on the contralateral ear. The mean unaided pure-tone average threshold across 500, 1000, and 2000 Hz was 105 dB HL for the nonimplanted ear. Relevant demographic information is shown in Table 1. Twenty-two (11 females and 11 males) Mandarin-speaking NH children were also tested to ensure that the listening tasks were age appropriate. The inclusion criterion was normal hearing, and exclusion criteria included previous formal music training as well as any cognitive, visual, and intelligence disorders. The mean age at testing for NH subjects was 6.2 years (range: 4.5 to 9.3 years).
Table 1.

CI Subject Demographic Information.

SubjectGenderAge at testing (years)Age at CI (years)CI exp. (years)CI deviceCI strategy
T1M5.82.83.0Cochlear N-24ACE
T2F6.86.00.8Cochlear N-24ACE
T3M5.83.32.5Cochlear N-24ACE
T4M9.36.13.2Cochlear N-24ACE
T5M6.41.64.8Cochlear N-24ACE
T6M8.82.86.0AB HiRes 90KF120
T7F7.85.82.0Cochlear N-24ACE
T8F6.32.34.0MED-EL PulsarFSP
T9M4.51.72.8MED-EL PulsarFSP
T10M5.53.02.5Cochlear N-24ACE
T11M5.63.12.5AB HiRes 90KF120
T12M5.53.52.0Cochlear N-24ACE
T13F5.52.53.0MED-EL PulsarFSP
T14M5.53.81.7Cochlear N-24ACE
T15M5.32.82.5Cochlear N-24ACE
T16F5.63.62.0MED-EL PulsarFSP
AVE6.33.42.8
(SE)(0.3)(0.4)(0.3)

Note. CI exp. = amount of CI experience; N-24 = Nucleus 24; AB = Advanced Bionics; ACE = Advanced combination encoder; F120 = Fidelity 120; FSP = Fine-structure processing.

CI Subject Demographic Information. Note. CI exp. = amount of CI experience; N-24 = Nucleus 24; AB = Advanced Bionics; ACE = Advanced combination encoder; F120 = Fidelity 120; FSP = Fine-structure processing. All subjects were trained and tested at the Xinsheng rehabilitation center in Jiangsu province, China. The Xinsheng center offers special education services for hearing impaired people, including pediatric CI users. However, the center offers little direct auditory habilitation for CI users, and speech perception is largely learned during regular class offerings (e.g., mathematics, Chinese language, etc.). During training and testing, all stimuli were presented in sound field at 65 dBA via a single loudspeaker positioned 1 m away from the subject, who directly faced the speaker. For the closed-set MCI and tone recognition tasks, children were instructed how to use the computer interface; for the open-set Mandarin Speech Perception (MSP) sentence recognition, an experimenter scored subject responses. Parents or supervisors accompanied their children for all test and training sessions.

MCI Stimuli and Testing

MCI stimuli and testing were similar to that in Galvin et al. (2007, 2009). Stimuli consisted of nine melodic contours (rising, rising-flat, rising-falling, flat-rising, flat, flat-falling, falling-rising, falling-flat, or falling), composed of five notes of equal duration (250 ms, with 50 ms of silence between each note). The lowest note in any contour was A4 (440 Hz). The spacing between successive notes in each contour was varied to be 1, 2, 3, or 5 semitones. As such, a rising contour might range from A4 (440 Hz) to D4 (587 Hz) with 1-semitone spacing, or from A4 (440 Hz) to F6 (1397 Hz) with 5-semitone spacing. The instrument used for the contour was a piano sample, as in Galvin, Fu, and Oba (2008). Thus, the stimulus set consisted of 36 stimuli (9 melodic contours × 4 semitone spacing), and all 36 stimuli were presented during each test run. MCI testing was performed using a closed-set, nine-alternative forced-choice (9AFC) procedure. Prior to formal testing, a practice session was conducted to familiarize subjects with the stimuli, task, and procedures. During testing, a contour would be randomly selected from the stimulus set and presented to the subject, who responded by clicking on one of the response boxes shown on the computer screen, labeled with a picture of the contour as well as Chinese text describing the pitch direction. Subjects were allowed to repeat the stimulus up to three times; no trial-by-trial feedback was provided. A minimum of two test runs was conducted for each subject, and performance was averaged across test runs.

Lexical Tone Stimuli and Testing

Lexical tone stimuli consisted of four tonal patterns spoken by two males and two females, taken from the Standard Chinese Database (Wang, 1993). The four tonal patterns included Tone 1 (high-level), Tone 2 (high-rising), Tone 3 (falling-rising), and Tone 4 (high-falling) produced for four monosyllables (/ba/, /bo/, /bu/, /bi/). Thus, the stimulus set consisted of 64 stimuli (4 tone × 4 monosyllables × 4 talkers), and all 64 stimuli were presented during each test run. Tone contour features such as F0, duration, and amplitude were preserved. The mean duration was 273 ms for Tone 1, 340 ms for Tone 2, 410 ms for Tone 3, and 213 ms for Tone 4. Table 2 shows the change in F0 (in semitones) for the four lexical tones produced in the four vowel contexts by the four talkers.
Table 2.

Change in F0 (in Semitones) Over Vowels for Lexical Tones Produced by Two Female (F1, F2) and Two Male (M1, M2) Talkers.

Change in F0 (semitones)
TalkerVowelTone 1Tone 2Tone 3Tone 4
F1/ba/1.07.212.310.1
/bi/0.97.36.114.5
/bo/2.07.019.916.9
/bu/0.67.19.714.9
F2/ba/2.58.36.625.9
/bi/2.78.210.720.7
/bo/2.58.98.224.3
/bu/2.29.111.326.0
M1/ba/1.36.68.115.4
/bi/0.88.98.319.4
/bo/1.48.58.317.5
/bu/2.110.18.418.7
M2/ba/2.08.27.616.6
/bi/2.89.07.518.0
/bo/3.78.010.418.7
/bu/1.88.39.317.1
AVE1.98.29.518.4
(SE)(0.2)(0.2)(0.8)(1.1)
Change in F0 (in Semitones) Over Vowels for Lexical Tones Produced by Two Female (F1, F2) and Two Male (M1, M2) Talkers. Tone recognition was measured in quiet as in Tao et al. (2015), using a four-alternative forced-choice procedure. Stimuli were presented in sound field at 65 dBA. During testing, a stimulus would be randomly selected from the stimulus set and presented to the subject, who responded by clicking on one of the four response boxes shown on the computer screen, labeled in Chinese as Tone 1—flat, Tone 2—rising, Tone 3—falling-rising, and Tone 4—falling. Subjects were allowed to repeat the stimulus up to three times; no trial-by-trial feedback was provided. A minimum of two test runs was conducted for each subject, and performance was averaged across test runs.

Sentence Stimuli and Testing

Sentence recognition in quiet was measured using sentences from the MSP test (Fu, Zhu, & Wang, 2011; Zhu, Wang, & Fu, 2012). Lists of 20 sentences developed for testing CI listeners were used (Li, Wang, Su, Galvin, & Fu, 2016; Su, Galvin, Zhang, Li, & Fu, 2016). The MSP sentences are relatively easy and can be used to test pediatric CI users. Sentence recognition was measured using an open-set paradigm. During testing, a sentence was randomly selected from the list and presented to the subject, who repeated as many words as possible. The experimenter scored the correctly identified words, and then a new sentence was presented. One MSP list was presented for each test session.

MCI Training

After baseline measures were completed, subjects began MCI training at the Xinsheng rehabilitation center in Jiangsu, using customized training software loaded onto the rehabilitation center computers. All subjects were required to train for 5 days per week, for 8 weeks. Subjects completed three to six sessions each training day; the average time for each training session was 15 min (range = 12 to 18 min). MCI training was similar to that in previous studies (Fu et al., 2015; Galvin et al., 2007, 2012). The lowest note of the training stimuli was randomly varied from trial to trial to be any note between A3 (220 Hz) and A5 (880 Hz) except A4 (440 Hz), thus avoiding direct training of the pitch range used for testing. Each training run contained 25 stimuli. Contours were presented with either 3 to 4 or 5 to 6 semitone spacing (i.e., 3 semitones between successive notes, 4 semitones between successive notes, etc.). These two spacing conditions were randomly assigned across training runs. During training, a contour was randomly selected from the stimulus set and the contour would not be presented again in the remaining training procedure. Subjects responded by clicking on one of the nine response boxes shown onscreen. If the subject responded correctly, a new contour would be presented. If not, audio and visual feedback was provided allowing subjects to compare their response to the correct response, after which a new contour was presented. MCI, tone, and sentence recognition were remeasured after 2, 4, and 8 weeks of training. Four weeks after training was stopped, MCI, tone, and sentence recognition were remeasured to observe whether any training benefits were retained.

Results

Figure 1 shows boxplots of MCI scores for each semitone spacing condition, as a function of test week; Test Week 0 = baseline, Test Weeks 2, 4, and 8 = posttraining, and Test Week 12 = follow-up measures 4 weeks after training was stopped. A two-way repeated measures analysis of variance (RM ANOVA) with semitone spacing (1, 2, 3, 5) and test week (0, 2, 4, 8, 12) as factors was performed on the data shown in Figure 1. Results showed a significant effect for test week, F(4, 168) = 17.5; p < .001, but not for semitone spacing, F(3, 168) = 2.4; p = .081; there was no significant interaction, F(12, 168) = 0.3; p = .986. Post hoc Bonferroni pairwise comparisons showed that performance was significantly better at Weeks 4, 8, and 12 than at Week 0 (p < .05 in all cases), and significantly better at Weeks 8 and 12 than at Week 2 (p < .05 in both cases); there were no significant differences among the remaining test weeks (p > .05 in all cases).
Figure 1.

Boxplots of CI users’ MCI scores for each semitone spacing as a function of test week; Week 0 = baseline, Weeks 2, 4, and 8 = posttraining, and Week 12 = follow-up performance 4 weeks after training was stopped. The solid horizontal line shows mean performance for a group of NH peers. The boxes show the 25th to 75th percentiles, the error bars show the 5th and 95th percentiles, the circles show outliers, the solid horizontal lines show median performance, and the dashed horizontal lines shows mean performance.

Boxplots of CI users’ MCI scores for each semitone spacing as a function of test week; Week 0 = baseline, Weeks 2, 4, and 8 = posttraining, and Week 12 = follow-up performance 4 weeks after training was stopped. The solid horizontal line shows mean performance for a group of NH peers. The boxes show the 25th to 75th percentiles, the error bars show the 5th and 95th percentiles, the circles show outliers, the solid horizontal lines show median performance, and the dashed horizontal lines shows mean performance. Figure 2 shows boxplots of recognition scores for each lexical tone as a function of test week. A two-way RM ANOVA, with tone (1, 2, 3, 4) and test week as factors, showed significant effects for tone, F(3, 168) = 4.4; p = .009, and test week, F(4, 168) = 13.5; p < .001; there was no significant interaction, F(12, 168) = 0.7; p = .735. Post hoc Bonferroni pairwise comparisons showed that recognition was significantly better with Tones 1 and 4 than with Tone 2 (p < .05 in both cases); there were no significant differences among the remaining tones. Post hoc Bonferroni pairwise comparisons also showed that performance was significantly better at Weeks 2, 4, 8, and 12 than at Week 0 (p < .05 in all cases); there were no significant differences among the remaining test weeks (p > .05 in all cases).
Figure 2.

Boxplots of CI users’ recognition scores for individual lexical tones, as a function of test week. The solid horizontal line shows mean performance for a group of NH peers. The boxes show the 25th to 75th percentiles, the error bars show the 5th and 95th percentiles, the circles show outliers, the solid horizontal lines show median performance, and the dashed horizontal lines shows mean performance.

Boxplots of CI users’ recognition scores for individual lexical tones, as a function of test week. The solid horizontal line shows mean performance for a group of NH peers. The boxes show the 25th to 75th percentiles, the error bars show the 5th and 95th percentiles, the circles show outliers, the solid horizontal lines show median performance, and the dashed horizontal lines shows mean performance. Figure 3 shows boxplots of MCI (averaged across all semitone spacing conditions), tone recognition (averaged across all tones), and sentence recognition scores, as a function of test week. One-way RM ANOVAs were performed on the MCI, tone, and sentence recognition data shown in Figure 3, with test week as the factor. There was a significant effect of test week for MCI, F(4, 59) = 16.1; p < .001, tone recognition, F(4, 59) = 12.3; p < .001, and sentence recognition, F(4, 59) = 13.2; p < .001. Post hoc Bonferroni pairwise comparisons showed that MCI was significantly better at Weeks 4, 8, and 12 than at Week 0 (p < .05 in all cases), and significantly better at Weeks 8 and 12 than at Week 2 (p < .05 in both cases); there were no significant differences among the remaining test weeks (p > .05 in all cases). Tone recognition was significantly better at Weeks 2, 4, 8, and 12 than at Week 0 (p < .05 in all cases); there were no significant differences among the remaining test weeks (p > .05 in all cases). Performance was significantly better at Week 8 relative to Week 0 for Tones 1, 2, and 3 (p < .05 in all cases). Sentence recognition was significantly better at Weeks 4, 8, and 12 than at Week 0 (p < .05 in all cases), and significantly better at Weeks 8 and 12 than at Week 2 (p < .05 in both cases); there were no significant differences among the remaining test weeks (p > .05 in all cases).
Figure 3.

Boxplots of CI users’ MCI (across all semitone spacings), tone recognition (across all lexical tones), and sentence recognition scores as a function of test week. The solid horizontal line shows mean performance for a group of NH peers. The boxes show the 25th to 75th percentiles, the error bars show the 5th and 95th percentiles, the circles show outliers, the solid horizontal lines show median performance, and the dashed horizontal lines shows mean performance.

Boxplots of CI users’ MCI (across all semitone spacings), tone recognition (across all lexical tones), and sentence recognition scores as a function of test week. The solid horizontal line shows mean performance for a group of NH peers. The boxes show the 25th to 75th percentiles, the error bars show the 5th and 95th percentiles, the circles show outliers, the solid horizontal lines show median performance, and the dashed horizontal lines shows mean performance. Demographic factors such as age at testing, age at implantation, and CI experience were compared with MCI, tone recognition, and sentence recognition performance at the different test intervals; results of Pearson correlations are shown in Table 3. After Bonferroni adjustment for multiple comparisons, significant correlations were observed between CI experience and MCI at Week 0 (p = .012), between age at implantation and tone recognition at Week 4 (p < .001), and between age at testing and tone recognition at Week 8 (p = .006).
Table 3.

Results of Pearson Correlations Between Demographic Factors and Music and Speech Perception for CI Users at Each Test Week, and for NH Listeners.

MCI
Tone
Sentence
Week r p r p r p
CI 0Age at test0.260.335−0.340.203−0.030.919
Age at CI−0.300.259−0.360.165−0.300.264
CI exp.0.610.012*0.050.8390.300.352
2Age at test0.300.263−0.310.2400.100.721
Age at CI−0.140.599−0.310.249−0.230.389
CI exp.0.470.0630.020.9560.360.169
4Age at test0.040.869−0.560.023−0.010.992
Age at CI−0.350.179−0.75<0.001*−0.260.330
CI exp.0.440.0840.240.3620.300.277
8Age at test0.100.709−0.700.006*0.010.984
Age at CI−0.200.457−0.570.022−0.250.349
CI exp.0.330.209−0.060.8370.290.288
12Age at test0.070.797−0.440.1050.010.964
Age at CI−0.360.163−0.540.037−0.170.547
CI exp.0.480.0690.150.5990.200.470
NHAge at test0.410.0570.050.8570.460.030

Note. The asterisks indicate significant correlations after Bonferroni correction for multiple comparisons. Age at CI = age at cochlear implantation; CI exp. = amount of CI experience; MCI = melodic contour identification.

Results of Pearson Correlations Between Demographic Factors and Music and Speech Perception for CI Users at Each Test Week, and for NH Listeners. Note. The asterisks indicate significant correlations after Bonferroni correction for multiple comparisons. Age at CI = age at cochlear implantation; CI exp. = amount of CI experience; MCI = melodic contour identification. The mean amount of time spent training over the 8-week period was 16.3 hr (range: 12.7 to 20.0). After 8 weeks of training, the mean improvement was 22.9 (range: −5.7 to 47.2), 14.5 (range: 4.7 to 32.8), and 14.5 (−1.5 to 34.3) percentage points for MCI, tone recognition, and sentence recognition, respectively. Pearson correlations showed no significant relationships between the time spent training and training benefit for MCI, tone recognition, or sentence recognition (p > .05 in all cases). There were no significant correlations observed in terms of training benefits (Week 8—Week 0) among MCI, tone, and sentence recognition (p > .005 in all cases). After Bonferroni correction for multiple comparisons, Pearson correlation showed a significant relationship between MCI and tone recognition only at Week 12 (r = .65, p = .009). There were no other significant correlations among MCI, tone, and sentence recognition at the other test intervals (p > .05 in all cases).

Discussion

Consistent with our hypothesis, 8 weeks of music training significantly improved music and speech perception in Mandarin-speaking pediatric CI users. Training benefits were largely retained 4 weeks after training was stopped. However, posttraining MCI, tone, and sentence recognition remained much poorer for CI users than for NH peers. We next discuss the results in greater detail.

Baseline Music and Speech Performance

Baseline MCI performance for CI subjects was generally poor (mean = 33.2% correct) and highly variable (range: 8.3 to 83.3% correct), similar to previous reports with Mandarin-speaking CI users (Fu et al., 2015), but poorer than observed in previous studies with adult postlingual English-speaking CI users (Galvin et al., 2007, 2009). Mean MCI recognition in this study was better than the 18.5% correct observed in Tao et al. (2015). In that study, the mean age at testing for prelingual subjects was 10.8 years (range: 6 to 16 years), compared with 6.1 years (range: 5 to 9 years) in this study. The mean age at implantation for prelingual subjects was 4.3 years (range: 2 to 12 years) in Tao et al. (2015), compared with 3.5 years (range: 1.2 to 6.5 years) in this study. Earlier implantation and shorter duration of deafness may have contributed to differences in MCI performance and the relationship between speech and music perception observed between Tao et al. (2015) and the present study. There was no significant effect of semitone spacing on baseline performance, different from previous studies with adult CI users that showed better MCI performance with increased semitone spacing (Galvin et al., 2007, 2009). However, the present results are consistent with Tao et al. (2015), who showed no significant differences among semitone spacing, except between 1 and 6 semitones. In this study, some melodic contours (rising-flat, flat-rising, flat, flat-falling, and falling-flat) were more easily identified than others (rising, rising-falling, falling-rising, and falling), which might have offset advantages associated with larger spacings. Differences in subject age, duration of deafness, and previous acoustic hearing experience may have also contributed to differences in MCI performance observed between this and previous studies. Mean baseline tone recognition in this study (77.9% correct) was similar to that of prelingual CI subjects in Tao et al. (2015; 80.9% correct). Tone 2 recognition was significantly poorer than Tone 1 or Tone 4 recognition, consistent with previous studies (Su et al., 2016). In this study, all covarying cues (F0, amplitude, and duration) were preserved. Given the weak coding of F0 cues in CI signal processing, differences in duration may have contributed to the present pattern of results, with Tone 2 having a relatively long duration and Tone 4 having a relatively short duration (Wei, Cao, & Zeng, 2004). MCI performance was significantly correlated with tone recognition at Week 12 (follow-up measure), but not at baseline. The correlation at Week 12 is consistent with previous CI studies involving tonal languages (Looi et al., 2015; Wang S et al., 2012; Wang W et al., 2011), suggesting that music and lexical tone perception may share perceptual mechanisms most likely related to pitch perception. However, the present data at Week 0 (baseline) were consistent with Tao et al. (2015), who found no correlation between MCI performance and lexical tone perception in pre- or postlingual Mandarin-speaking CI users. This suggests that music training may be necessary to observe relationships between speech and music perception. Mean baseline MSP sentence recognition was only 45.8% correct, much poorer than that of NH subjects (97.5% correct). The present performance was also much poorer than that observed for MSP sentences with pediatric CI users (84.7% correct in Su et al., 2016), observed for MSP sentences with NH adult subjects listening to four-channel acoustic CI simulations (90.1% correct in Fu et al., 2011), and poorer than observed with the similar MEST sentences for adult CI users (82% correct in Li et al., 2016). It is possible than differences in age at testing may have contributed to the discrepancies in sentence recognition across studies. In Fu et al. (2011) and Li et al. (2016), adult subjects were tested. In Su et al. (2016), the mean age at testing for pediatric CI subjects was 9.7 years, compared with 6.1 years in this study. Eisenberg, Shannon, Martinez, Wygonski, and Boothroyd (2000) found significantly better performance for older (10 to 12 years) than for younger NH children (5 to 7 years) listening to acoustic CI simulations, suggesting some developmental contribution to perception of spectrally degraded speech. The lower sentence recognition performance, compared with tone recognition, may have also been due to different testing paradigms (closed-set for tones and open-set for sentences). After Bonferroni correction for multiple comparisons, the amount of CI experience was significantly correlated with baseline MCI performance (p < .05), but not with baseline tone or sentence recognition (p > .05 in both cases). There were no significant correlations between age at testing and MCI, tone, or sentence recognition (p > .05 in all cases). Thus, CI experience appears to have contributed more strongly to baseline music perception than to speech perception. Perception of melodic contours may have depended more strongly on adaptation to the spectrotemporal resolution of the CI, rather than development of specific patterns, as would be required for speech perception. Also, prelingual pediatric CI users must develop speech patterns with the limited spectrotemporal resolution available with their device. Perception of melodic contours requires attention to pitch cues, while lexical tone perception depends on F0, duration, and amplitude cues, which may interact with development (age at testing and age at implantation). Note that CI experience was not significantly correlated with MCI performance (or with tone or sentence recognition) at any of the other test intervals (p > .05 in all cases), suggesting that the MCI training may have been more of a factor than CI experience for MCI performance.

Benefits of Music Training

Eight weeks of MCI training significantly improved MCI performance (p < .05). The range of posttraining gains for MCI was quite large (−5.7 to 47.2 percentage points). Some subjects with relatively low baseline scores (e.g., T8, T15) received relatively small training benefits (<12 percentage points), while others (e.g., T14, T16) received large benefits (>38 percentage points). Posttraining gains were retained 4 weeks after training was stopped and did not appear to be due to selective improvements for different semitone spacings, as there was no significant difference among the different spacings after 8 weeks of MCI training. The present subjects trained only with the 3 to 4 and 5 to 6 semitone spacings, which seemed to improve performance for all semitone spacing conditions. Although baseline MCI performance was poorer in this study than in previous studies with adult, postlingual English-speaking CI users (Galvin et al., 2007, 2009, 2012), the mean training benefit was comparable across studies (approximately 25 percentage points). The present training benefit was much less than reported by Fu et al. (2015) for six pediatric Mandarin-speaking CI users. Note that subjects in Fu et al. (2015) trained with a relatively simple harmonic complex, compared with the more spectrotemporally complex piano sound used in this study. The mean age of this study was also lower than that of previous studies, which may have influenced the training benefit. Significant improvements were observed for Tones 1, 2, and 3 after 8 weeks of MCI training, relative to baseline (p < .05); improvement in Tone 4 recognition may have been limited by ceiling performance effects. The greatest posttraining improvement was observed for Tone 3 (rising-falling), which has greater changes in F0 than in the other tones. The improvement in MSP sentence recognition may have been due to improved tone recognition, as lexical tone recognition has been shown to significantly contribute to sentence recognition (Chen, Wong, Chen, & Xi, 2014; Fu et al., 1998). The music training benefit for tone recognition is consistent with previous pediatric CI studies that showed a benefit for music training in terms of speech prosody perception, for which pitch cues are important (Good et al., 2017; Torppa et al., 2014). While many studies have shown a musician advantage for speech perception (Besson et al., 2007; Fuller et al., 2014; Musacchia et al., 2007; Parbery-Clark et al., 2009), others have not (Deroche, Limb, Chatterjee, & Gracco, 2017; Madsen, Whiteford, & Oxenham, 2017; Ruggles, Freyman, & Oxenham, 2014). The variability in musician effects for speech recognition may be due to the importance of pitch cues to the listening task. In this study, the improvements in speech performance were for a speech task where pitch cues are lexically meaningful (as opposed to segregation of competing talkers or perception of voiced vs. whispered speech). While the improved melodic pitch perception via MCI training appeared to generalize to improved speech perception where voice pitch cues are critical, it is worth noting that the time scale of pitch changes in the MCI task (1,500 ms) was much greater than for lexical tones (309 ms, averaged across all tones). Interestingly, the range of changes in F0 was comparable between the MCI and lexical tone stimuli (Table 2). There seemed to be a global benefit for the MCI training, as recognition significantly improved for all semitone spacings in the MCI task and for three of the four tones in the tone recognition task. It could be that subjects’ functional spectrotemporal resolution was improved as a result of the melodic pitch training. Alternatively, subjects’ attention, memory, and cognitive processing were improved merely by participating in the training, in which case, the music training did not necessarily improve specific aspects of auditory perception. However, Oba, Galvin, and Fu (2013) found significant improvements in speech understanding in noise perception with auditory training, and not with a similar visual training task, suggesting that auditory training is needed to improve auditory perception. While there was no proper control group for this study, MCI, tone, and sentence recognition were each repeatedly measured in a group of six Mandarin-speaking pediatric CI users at the same test intervals as for the above CI users who received MCI training. One-way RM ANOVAs showed no significant effect of test interval on MCI, tone, or sentence recognition performance (p > .05), suggesting that the benefit of observed in subjects who received MCI training may not have been due to procedural learning. Given the benefit of MCI training observed in this study, further study with a rigorously matched control group and experimental blinding seems warranted. Finally, note that the present MCI training is not at all similar to the instrument training experienced by musicians in studies that show musician advantages in speech performance and auditory perception (Besson et al., 2007; Fuller et al., 2014; Kraus and Chandrasekaran, 2010; Kraus et al., 2009; Musacchia et al., 2007; Parbery-Clark et al., 2009). Active participation in music and learning to play an instrument may further improve the benefits of music training for CI users. Patel (2014) presented preliminary data collected with Galvin and coworkers showing improved MCI training benefits when subjects were instructed to play the contours on a musical keyboard, rather than just listen to the contours as in previous MCI studies (Fu et al., 2015; Galvin et al., 2007, 2009, 2012). Previous studies have shown that MCI training generalized to improved familiar melody recognition, and that MCI training with one instrument or pitch range can generalize to better MCI performance with other instruments or pitch ranges (Galvin et al., 2007, 2009, 2012). While these generalized performance gains are encouraging, it is unclear whether MCI training might generalize to other music perception tasks involving more complex listening (e.g., polyphonic music perception). Still, the present data suggest that music training can significantly benefit Mandarin-speaking pediatric CI users’ melodic pitch and speech perception.

Conclusions

In this study, the benefits of music training for speech and music perception were studied in young, pediatric, Mandarin-speaking CI users. The results suggest that melodic pitch training can improve melodic pitch perception as well as lexical tone and sentence recognition. Major findings include: Baseline MCI performance was poor, while tone and sentence recognition were moderately good. After 8 weeks of MCI training, significant improvements were observed for MCI, tone, and sentence recognition. However, posttraining performance for CI users remained much poorer than that of NH peers. Posttraining gains were largely retained 4 weeks after training was stopped. Significant correlations were observed between baseline MCI performance and CI experience, between tone recognition and age at implantation after 4 weeks of training, and between age at testing and tone recognition after 8 weeks of training. Training benefits for MCI, tone, and sentence recognition were not significantly correlated with age at testing, age at implantation, or CI experience.
  47 in total

1.  Mandarin tone recognition in cochlear-implant subjects.

Authors:  Chao-Gang Wei; Keli Cao; Fan-Gang Zeng
Journal:  Hear Res       Date:  2004-11       Impact factor: 3.208

2.  Processing F0 with cochlear implants: Modulation frequency discrimination and speech intonation recognition.

Authors:  Monita Chatterjee; Shu-Chen Peng
Journal:  Hear Res       Date:  2007-11-23       Impact factor: 3.208

3.  Vocal emotion recognition by normal-hearing listeners and cochlear implant users.

Authors:  Qian-Jie Fu; John J Galvin
Journal:  Trends Amplif       Date:  2007-12

4.  Melodic, rhythmic, and timbral perception of adult cochlear implant users.

Authors:  K Gfeller; C R Lansing
Journal:  J Speech Hear Res       Date:  1991-08

5.  Validation of list equivalency for Mandarin speech materials to use with cochlear implant listeners.

Authors:  Yongxin Li; Shuncheng Wang; Qiaodang Su; John J Galvin; Qian-Jie Fu
Journal:  Int J Audiol       Date:  2016-07-14       Impact factor: 2.117

6.  Musician enhancement for speech-in-noise.

Authors:  Alexandra Parbery-Clark; Erika Skoe; Carrie Lam; Nina Kraus
Journal:  Ear Hear       Date:  2009-12       Impact factor: 3.570

Review 7.  Experience-induced malleability in neural encoding of pitch, timbre, and timing.

Authors:  Nina Kraus; Erika Skoe; Alexandra Parbery-Clark; Richard Ashley
Journal:  Ann N Y Acad Sci       Date:  2009-07       Impact factor: 5.691

8.  Minimal effects of visual memory training on auditory performance of adult cochlear implant users.

Authors:  Sandra I Oba; John J Galvin; Qian-Jie Fu
Journal:  J Rehabil Res Dev       Date:  2013

9.  The perception of prosody and associated auditory cues in early-implanted children: the role of auditory working memory and musical activities.

Authors:  Ritva Torppa; Andrew Faulkner; Minna Huotilainen; Juhani Järvikivi; Jari Lipsanen; Marja Laasonen; Martti Vainio
Journal:  Int J Audiol       Date:  2014-01-27       Impact factor: 2.117

10.  Musicians do not benefit from differences in fundamental frequency when listening to speech in competing speech backgrounds.

Authors:  Sara M K Madsen; Kelly L Whiteford; Andrew J Oxenham
Journal:  Sci Rep       Date:  2017-10-03       Impact factor: 4.379

View more
  5 in total

1.  Singing Proficiency of Members of a Choir Formed by Prelingually Deafened Children With Cochlear Implants.

Authors:  Jing Yang; Qi Liang; Haotong Chen; Yanjun Liu; Li Xu
Journal:  J Speech Lang Hear Res       Date:  2019-05-21       Impact factor: 2.297

2.  No Evidence That Music Training Benefits Speech Perception in Hearing-Impaired Listeners: A Systematic Review.

Authors:  Colette M McKay
Journal:  Trends Hear       Date:  2021 Jan-Dec       Impact factor: 3.293

3.  Effect of Sound Coding Strategies on Music Perception with a Cochlear Implant.

Authors:  Gaëlle Leterme; Caroline Guigou; Geoffrey Guenser; Emmanuel Bigand; Alexis Bozorg Grayeli
Journal:  J Clin Med       Date:  2022-07-29       Impact factor: 4.964

4.  Rapid Assessment of Non-Verbal Auditory Perception in Normal-Hearing Participants and Cochlear Implant Users.

Authors:  Agathe Pralus; Ruben Hermann; Fanny Cholvy; Pierre-Emmanuel Aguera; Annie Moulin; Pascal Barone; Nicolas Grimault; Eric Truy; Barbara Tillmann; Anne Caclin
Journal:  J Clin Med       Date:  2021-05-13       Impact factor: 4.241

5.  A Preliminary Study of the Effects of Attentive Music Listening on Cochlear Implant Users' Speech Perception, Quality of Life, and Behavioral and Objective Measures of Frequency Change Detection.

Authors:  Gabrielle M Firestone; Kelli McGuire; Chun Liang; Nanhua Zhang; Chelsea M Blankenship; Jing Xiang; Fawen Zhang
Journal:  Front Hum Neurosci       Date:  2020-03-31       Impact factor: 3.169

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.