Literature DB >> 35833941

Enhanced mismatch negativity in harmonic compared with inharmonic sounds.

David Ricardo Quiroga-Martinez1,2, Krzysztof Basiński3, Jonathan Nasielski4, Barbara Tillmann5,6, Elvira Brattico2,7, Fanny Cholvy5,6, Lesly Fornoni5,6, Peter Vuust2, Anne Caclin5,6.   

Abstract

Many natural sounds have frequency spectra composed of integer multiples of a fundamental frequency. This property, known as harmonicity, plays an important role in auditory information processing. However, the extent to which harmonicity influences the processing of sound features beyond pitch is still unclear. This is interesting because harmonic sounds have lower information entropy than inharmonic sounds. According to predictive processing accounts of perception, this property could produce more salient neural responses due to the brain's weighting of sensory signals according to their uncertainty. In the present study, we used electroencephalography to investigate brain responses to harmonic and inharmonic sounds commonly occurring in music: Piano tones and hi-hat cymbal sounds. In a multifeature oddball paradigm, we measured mismatch negativity (MMN) and P3a responses to timbre, intensity, and location deviants in listeners with and without congenital amusia-an impairment of pitch processing. As hypothesized, we observed larger amplitudes and earlier latencies (for both MMN and P3a) in harmonic compared with inharmonic sounds. These harmonicity effects were modulated by sound feature. Moreover, the difference in P3a latency between harmonic and inharmonic sounds was larger for controls than amusics. We propose an explanation of these results based on predictive coding and discuss the relationship between harmonicity, information entropy, and precision weighting of prediction errors.
© 2022 The Authors. European Journal of Neuroscience published by Federation of European Neuroscience Societies and John Wiley & Sons Ltd.

Entities:  

Keywords:  amusia; auditory perception; event-related potentials; harmonicity; pitch perception

Mesh:

Year:  2022        PMID: 35833941      PMCID: PMC9543822          DOI: 10.1111/ejn.15769

Source DB:  PubMed          Journal:  Eur J Neurosci        ISSN: 0953-816X            Impact factor:   3.698


analysis of variance congenital amusia confidence interval degrees of freedom electroencephalography event‐related potential fundamental frequency independent component analysis mean Montreal Battery of Evaluation of Amusia mismatch negativity pitch discrimination threshold standard deviation

INTRODUCTION

Many naturally occurring sounds are composed of frequencies that are integer multiples of a single, fundamental frequency (F0). These are termed harmonic sounds and include human and animal voices as well as many musical instrument sounds. Harmonic sounds have an easily detectable pitch, a subjective perceptual quality that enables them to be ordered from low to high (McDermott, 2018; Oxenham, 2012). Pitch information is used for solving many cognitive tasks. In tonal languages, pitch contours determine the meaning of words, whereas in non‐tonal languages they are used for prosody. In music, pitch is critical for the processing of melody and harmony (Huron, 2016). Pitch information is also used in auditory scene analysis, where simultaneously perceived sounds can be segregated using pitch discrepancies (Bregman, 1994; Wightman & Green, 1974). Inharmonic sounds possess frequency spectra that do not follow the harmonic series. These are often described as metallic, crackling, sizzling, or noisy. Inharmonic sounds are sparingly used in Western tonal music (mainly in percussion instruments) as they do not produce a clear pitch sensation that may be used to convey melody. Historically, research in auditory neuroscience focused on simple stimuli, mostly limited to sine waves and harmonic complex tones (McPherson et al., 2022). Not much is known about the neural mechanisms of perception of inharmonic sounds or how inharmonicity influences the perception of auditory features other than pitch. Recent studies have shown harmonicity to be an important factor in several auditory tasks not directly related to pitch. Violations of harmonicity impair speech recognition in cocktail‐party scenarios (Popham et al., 2018). Harmonic sounds are easier to discriminate than inharmonic sounds over time delays, suggesting more efficient coding of harmonic sounds in memory (McPherson et al., 2022). Harmonic signals are also more easily detected in noise than are inharmonic signals, pointing to an important role of harmonicity in auditory scene analysis (McPherson et al., 2022). Taken together, these studies suggest that harmonicity is important for processing different auditory features. It is thus likely that the ease of detection of deviances in features, such as timbre, intensity, or location, could depend on harmonicity. This view is also consistent with a predictive coding account of auditory perception, where sensory information, usually cast as ascending prediction error, is weighted by the information entropy (i.e., uncertainty or inverse precision) of the stimuli (Clark, 2015; Friston, 2005; Rao & Ballard, 1999). As inharmonic sounds have higher spectral entropy, prediction errors elicited by inharmonic deviants may be down‐weighted in comparison with harmonic deviants. In this study, we used electroencephalography (EEG) to investigate the brain responses to harmonic and inharmonic sounds in both typical listeners and participants with congenital amusia (CA)—an impairment in pitch processing that cannot be attributed to causes such as intellectual disability, lack of music exposure, or brain damage (Peretz, 2016). Aiming for ecological validity, our present study used sounds commonly occurring in popular music: Piano tone (harmonic, low spectral entropy) and hi‐hat cymbal (inharmonic, high spectral entropy). In a multifeature oddball paradigm (Näätänen et al., 2004; Paavilainen, 2013), we measured mismatch negativity (MMN) and P3a responses to timbre, intensity, and location deviants. These event‐related potentials (ERPs) are elicited without attentional control and index the brain's ability to perform automatic comparisons of incoming auditory stimuli (Garrido et al., 2009; Paavilainen, 2013). In contrast to traditional paradigms that employ a single type of deviant, multifeature paradigms record MMN responses to several sound features in a relatively short time, which make them an optimal choice to address our question (Kliuchko et al., 2016; Vuust et al., 2011). Based on previous studies highlighting the role of harmonicity in auditory scene analysis, we hypothesized that timbre, intensity, and location deviants would elicit larger MMN and P3a responses for harmonic than for inharmonic sounds. Finally, given that listeners with CA are impaired in the processing of pitch information (which relies on harmonicity), here we also explored whether amusics process harmonic and inharmonic sounds differently from their matched non‐musician controls (Cousineau et al., 2012; Marin et al., 2012).

METHODS

The code and materials employed to conduct the experiment and analyses presented here can be found at: https://doi.org/10.17605/OSF.IO/JSEU8. Due to data protection regulations, data cannot be publicly shared but can be made privately available upon reasonable request.

Participants

Thirty‐four participants took part in the experiment (same participants as in Quiroga‐Martinez et al., 2021), 17 amusics and 17 matched controls (see Table 1). All subjects were recruited in the Lyon area in France, were French speakers, and were screened for CA using the Montreal Battery of Evaluation of Amusia (MBEA) (Peretz et al., 2003). The total MBEA scores including six subtests were significantly lower for amusics than for controls, t(25.6) = −10.49, p < 0.001, as were the average combined scores for the three pitch subtests, t(23.7) = −11.36, p < 0.001. An individual was considered amusic if the total MBEA score was less than 23 (maximum score = 30) or their average score for the pitch subtests was lower than 21.7 (maximum score = 30). We measured pitch discrimination thresholds (PDTs) using an adaptive tracking staircase procedure (Tillmann et al., 2009). PDTs were significantly larger for amusics in comparison to controls. The groups did not differ significantly in age, years of education, or musical training. The study was conducted and approved by a national ethics committee. Participants gave their written informed consent prior to the experiment and received a small financial compensation for their time.
TABLE 1

Participant demographics (mean ± SD, t statistics, degrees of freedom, and p values)

AmusicsControls t df p
Sample size1717
Female89
Right‐handed1314
Education (years)15.06 (± 2.7)15.12 (± 2.34)−0.0731.40.946
Music training (years)00.24 (± 0.66)127.5 (U)0.163
Age (years)38.43 (± 15.96)37.61 (± 17.02)0.1531.90.886
MBEA21.79 (± 1.81)27.09 (± 1.04)−10.4925.6<0.001
MBEA pitch20.79 (± 2.15)27.43 (± 1.09)−11.3623.7<0.001
PDT (semitones)1.57 (± 1.53)0.31 (± 0.17)3.37516.40.004

Note: MBEA: Montreal battery of evaluation of Amusia (maximum score = 30, average of the six sub‐tests of the battery); MBEA pitch: Average of the three sub‐tests of the battery assessing pitch (maximum score = 30); PDT: Pitch discrimination threshold. For music training, Mann–Whitney U test results are reported.

Participant demographics (mean ± SD, t statistics, degrees of freedom, and p values) Note: MBEA: Montreal battery of evaluation of Amusia (maximum score = 30, average of the six sub‐tests of the battery); MBEA pitch: Average of the three sub‐tests of the battery assessing pitch (maximum score = 30); PDT: Pitch discrimination threshold. For music training, Mann–Whitney U test results are reported.

Stimuli

Two types of sound sequences were presented to participants, corresponding to a modified version of the optimal MMN paradigm (Näätänen et al., 2004). In the piano condition, participants listened to repeating piano tones with a fundamental frequency at C3 pitch (F0 = 262 Hz). In the hi‐hat condition, broadband percussive hi‐hat cymbal sounds were used (Figure 1). The sounds were samples taken from the Cubase sample library (Steinberg Media Technology, version 8). The spectral entropies of the sounds were 2.94 bits and 9.66 bits for the piano and hi‐hat tones, respectively (as calculated with AntroPy v.0.1.4 ; Vallat, 2021). During the experiment, one sound was played repeatedly, and a deviant was pseudorandomly introduced in every group of four consecutive stimuli in the sequence. No two deviants were played consecutively, and no deviant feature was presented again before a whole iteration of the five features was played. Each condition (piano or hihat) was presented in a separate block lasting approximately 13 min. Both sequences included the same number of standard and deviant tones. The piano block and the hi‐hat block, together with two other conditions involving melodies with different complexity levels (presented in Quiroga‐Martinez et al., 2021), were counterbalanced between participants and their order matched across groups. Other conditions were included in the experiment, which have been reported elsewhere (Quiroga‐Martinez et al., 2021).
FIGURE 1

(a) Waveforms for standard hi‐hat and piano sounds. (b) Example of the stimulus deviants in each condition for each feature (red rectangles), represented as a power spectrogram showing left and right audio channels for two sound conditions. Vertical axes show (log‐spaced) frequency while horizontal axes show the time in seconds. For every condition, standard sounds are intertwined with different feature deviants. Note: The order of the stimuli in this figure is for illustration purposes; see main text for stimulus ordering procedures. (c) Normalized autocorrelations of standard piano and hi‐hat sounds in comparison with white noise. Piano sounds are much more periodic than hi‐hat sounds and white noise.

(a) Waveforms for standard hi‐hat and piano sounds. (b) Example of the stimulus deviants in each condition for each feature (red rectangles), represented as a power spectrogram showing left and right audio channels for two sound conditions. Vertical axes show (log‐spaced) frequency while horizontal axes show the time in seconds. For every condition, standard sounds are intertwined with different feature deviants. Note: The order of the stimuli in this figure is for illustration purposes; see main text for stimulus ordering procedures. (c) Normalized autocorrelations of standard piano and hi‐hat sounds in comparison with white noise. Piano sounds are much more periodic than hi‐hat sounds and white noise. The sounds lasted 250 ms each, were loudness normalized, and were presented excluding any silent gaps between the sounds. Pitch, intensity, timbre, location, and rhythm deviants were introduced in the melodies. Pitch deviants appeared in the piano condition but were not part of the analysis because there are no pitch deviants in the hi‐hat condition (results for the pitch deviant in the piano block can be seen in fig. 2 in Quiroga‐Martinez et al., 2021). All deviants were created with Adobe Audition (Adobe Systems Incorporated, version 8) by modifying the standard tones as follows: Intensity: −12 dB gain; timbre: a combination of filters (low‐shelf −10 dB at 500 Hz, peak +10 dB at 2 kHz, notch filter at 6 kHz, all filters Q = 1); location: leftward bias (20‐ms time shift between channels); rhythm: −60 ms for sound onset. Note that rhythm violations implied a shortening of the preceding tone and a lengthening of the actual deviant tone by 60 ms. The rhythm deviants were excluded from the analysis due to baseline contamination issues, as reported in Quiroga‐Martinez et al. (2021). Overall, a total of 2339 standards and 153 (5%) deviants per feature were presented in each of the two conditions.

Procedure

Participants were informed about the procedures at their arrival, willingly gave their written consent, and filled out the required forms. Subsequently, EEG caps were placed on their scalp accompanied by conductive gel applied to the electrodes. A Sennheiser HD280 Pro headset was carefully placed on top of the EEG cap with foam padding to avoid pressure on electrodes. The impedances were checked again after the headphones were on. All participants had the sound volume set to an identical, comfortable level. During testing, participants were looking at a computer screen from a distance of about 1.5 m, sitting on an armchair inside a sound‐attenuated booth, electrically shielded with a Faraday cage. They were told sounds would be playing in the background and were instructed to watch a movie of their choice while ignoring the auditory stimuli. Additionally, they were asked to remain still and relaxed, knowing there would be pauses between blocks during which they could stretch and change posture. During stimulation, the blocks were presented in such a way that for nine matched pairs of participants, the counterbalanced piano and hi‐hat conditions, together with the two additional complexity conditions, came before two counterbalanced additional conditions, whereas for the remaining pairs the order was inverted. Two additional blocks were included at the end of the experiment, in which participants listened freely to entire pieces of music. Their analysis, however, is beyond the scope of this article and will be reported elsewhere. The whole recording session lasted around 1 h and a half, plus half an hour of preparation.

EEG recording and preprocessing

Scalp potentials were recorded with a 64‐channel Biosemi system with active electrodes and a sampling rate of 1024 Hz. Additional electrodes were used to track horizontal and vertical eye movements. Data analyses were conducted with MNE‐Python v.0.23.0 (Gramfort et al., 2014). EEG signals were first cleaned from eyeblink artifacts using independent component analysis (ICA) with a semiautomatic routine (fastICA algorithm). Visual inspection was used as a quality check. After removing ICA components, the raw signals were filtered with a pass band of 0.5–35 Hz and re‐referenced to the mastoids. Epochs from −100 to 400 ms from tone onsets were extracted and baseline corrected with a prestimulus baseline of 100 ms. Epochs with an amplitude exceeding 150 μV were rejected to further clean the data from remaining artifacts. For each participant, ERPs were obtained by averaging epochs for the standard tones and each of the deviant features separately, per condition. Standard tones preceded by a deviant were excluded from the averages. Both deviant‐specific MMN and P3a responses were calculated by subtracting standard from deviant ERPs for each feature and condition.

Statistical analyses

We performed analyses on mean amplitudes and latencies for both MMN and P3a responses. MMN peak latencies were extracted within a time window of 70–250 ms, whereas P3a peak latencies were extracted for 150–350 ms. Mean amplitudes were obtained from electrodes Fz, F1, F2, F3, F4, FCz, FC1, FC2, FC3, and FC4 and calculated as the average activity ±25 ms around the participant‐wise peak, for each condition (piano or hi‐hat) and feature (intensity, location, timbre). The chosen electrodes correspond to those that typically show the largest responses in MMN studies and exhibited the largest P50 amplitudes (thus making sure that they properly captured auditory evoked activity). Using R v. 4.0.5 (R Core Team, 2021) and the ezANOVA library (Lawrence, 2016), several mixed analyses of variance (ANOVAs) including within‐ and between‐subject effects were run on the mean amplitudes and latencies of MMN and P3a difference waves. Factors included group (amusics and control), condition (piano and hi‐hat), and sound feature (intensity, location, and timbre). Mauchly's test of sphericity was checked for each ANOVA, and no correction was applied. Post hoc, pairwise contrasts were also performed with the emmeans library (Lenth et al., 2021) after fitting linear mixed‐effects models using lme4 library (Bates et al., 2015) with the two main effects and their interactions as predictors.

RESULTS

Consistent with the literature (Näätänen et al., 2007; Polich & Criado, 2006), the MMN and P3a manifested themselves as a fronto‐central negativity and a fronto‐central positivity, respectively (Figures 2 and 3).
FIGURE 2

Time‐course of grand‐average measured mismatch negativity (MMN) (blue) and P3a (red) topographies for different features (location, timbre, and intensity), groups (amusics and controls), and conditions (piano and hi‐hat).

FIGURE 3

Standard, deviant, and measured mismatch negativity (MMN) difference waves, averaged according to conditions, groups, and features. Vertical axes show neural activity in microvolts and horizontal axes show the time in seconds. Grey traces represent MMN difference waves for each participant. Shaded areas indicate 95% confidence intervals (for the difference waves). The displayed activity corresponds to the channel Fz.

Time‐course of grand‐average measured mismatch negativity (MMN) (blue) and P3a (red) topographies for different features (location, timbre, and intensity), groups (amusics and controls), and conditions (piano and hi‐hat). Standard, deviant, and measured mismatch negativity (MMN) difference waves, averaged according to conditions, groups, and features. Vertical axes show neural activity in microvolts and horizontal axes show the time in seconds. Grey traces represent MMN difference waves for each participant. Shaded areas indicate 95% confidence intervals (for the difference waves). The displayed activity corresponds to the channel Fz.

MMN analysis

Mixed effects ANOVA analyses revealed significant main effects and two‐way interactions on MMN mean amplitudes and peak latencies for group, condition, and feature (Figures 4 and 5).
FIGURE 4

Effect of condition on measured mismatch negativity (MMN) and P3a responses for each group. Grey traces represent difference waves between piano and hi‐hat condition for each participant. Shaded areas indicate 95% confidence intervals (for the difference of difference waves). The displayed activity corresponds to the channel Fz.

FIGURE 5

Mean measured mismatch negativity (MMN) amplitudes (left) and peak latencies (right) as a function of conditions and features in both groups. Boxes display median and interquartile ranges. Beans depict the estimated densities. Lines connect measurements for individual participants.

Effect of condition on measured mismatch negativity (MMN) and P3a responses for each group. Grey traces represent difference waves between piano and hi‐hat condition for each participant. Shaded areas indicate 95% confidence intervals (for the difference of difference waves). The displayed activity corresponds to the channel Fz. Mean measured mismatch negativity (MMN) amplitudes (left) and peak latencies (right) as a function of conditions and features in both groups. Boxes display median and interquartile ranges. Beans depict the estimated densities. Lines connect measurements for individual participants. For mean amplitudes (see Table A1), the main effect of condition was significant F(1,32) = 101.129, η 2 = 0.338, p < 0.001, with larger MMNs for piano tones compared with hi‐hat tones. No significant main effects of group and feature were found, with F(1,32) = 0.063, p = 0.803 and F(2,64) = 1.131, p = 0.329, respectively. No significant interaction was found between group and condition F(1,32) = 0.486, p = 0.491. Furthermore, the interaction between group and feature was not significant either, with F(2,64) = 2.561, p = 0.085. Interestingly, a significant interaction was found for condition and feature with F(2,64) = 24.485, η 2 = 0.068, p < 0.001. No significant three‐way interaction between group, condition and feature was observed, F(2,64) = 0.775, p = 0.465.
TABLE A1

ANOVA table for MMN amplitude analysis: main effects, two‐way and three‐way interaction effects on group, condition, and feature

EffectDFnDFdSSnSSd F p ges
(Intercept)1321098.317278.878126.027 <0.001* 0.687
Group1320.55278.8780.0630.8030.001
Condition132256.00181.006101.129 <0.001* 0.338
Feature2643.31893.8891.1310.3290.007
Group:condition1321.23181.0060.4860.4910.002
Group:feature2647.51393.8892.5610.0850.015
Condition:feature26436.49747.69824.485 <0.001* 0.068
Group:condition:feature2641.15547.6980.7750.4650.002

Note: Significant effects are marked with an asterisk.

By looking at the post hoc pairwise comparison (Table 2), we can observe that there are significant differences in amplitude, contrasting hi‐hat, and piano conditions, for all features. We notice in Table 2 that the estimates of between‐condition MMN differences (hi‐hat − piano) for intensity and location were relatively high compared with timbre. This is reflected in the significant differences in between‐condition contrasts comparing intensity to timbre (p < 0.001) and location to timbre (p = 0.001) but not intensity to location (p = 0.504). Taken together, these differences account for the significant condition‐by‐feature interaction.
TABLE 2

Pairwise contrasts of MMN mean amplitudes between conditions for each feature separately, as well as pairwise comparisons between features for differences between conditions (contrasts of contrasts)

FeatureContrastDifference (μV)CI 2.5%CI 97.5% t d p
IntensityHi‐hat − piano3.102.533.6610.821.92 <0.001*
LocationHi‐hat − piano2.541.973.108.871.57 <0.001*
TimbreHi‐hat − piano1.090.521.653.810.67 <0.001*
(Hi‐hat − piano intensity) − (hi‐hat − piano location)0.56−0.421.541.390.350.504
(Hi‐hat − piano intensity) − (hi‐hat − piano timbre)2.011.032.994.961.24 <0.001*
(Hi‐hat − piano location) − (hi‐hat − piano timbre)1.450.472.433.580.9 0.001*

Note: Data are pooled from both groups. Significant contrasts are highlighted in bold and marked with an asterisk. Standard effect sizes (d) are calculated as the difference between conditions divided by the square root of the sum of the residual and the random effects variance. The condition by feature interaction was not further modulated by group, p = 0.465.

Pairwise contrasts of MMN mean amplitudes between conditions for each feature separately, as well as pairwise comparisons between features for differences between conditions (contrasts of contrasts) Note: Data are pooled from both groups. Significant contrasts are highlighted in bold and marked with an asterisk. Standard effect sizes (d) are calculated as the difference between conditions divided by the square root of the sum of the residual and the random effects variance. The condition by feature interaction was not further modulated by group, p = 0.465. For mean latencies (see Table A2), the main effect of feature was significant F(2,64) = 9.252, η 2 = 0.073, p < 0.001. No significant main effects were found for group and condition, with F(1,32) = 0.195, p = 0.662 and F(1,32) = 0.475, p = 0.496, respectively. No significant interaction was found for group‐by‐condition F(1,32) = 0.064, p = 0.802 or group‐by‐feature F(2,64) = 0.252, p = 0.778. Interestingly, an interaction was found for condition and feature with F(2,64) = 11.520, η 2 = 0.105, p < 0.001. No significant three‐way interaction between group, condition, and feature was observed, F(2,64) = 0.457, p = 0.635.
TABLE A2

ANOVA table for MMN latency analysis: main effects, two‐way and three‐way interaction effects on group, condition, and feature

EffectDFnDFdSSnSSd F p ges
(Intercept)1322869123.7733201.0592765.332 <0.001* 0.955
Group132202.00533201.0590.1950.6620.001
Condition132313.7721153.3330.4750.4960.002
Feature26410732.39237121.4719.252 <0.001* 0.073
Group:condition13242.39721153.3330.0640.8020.0
Group:feature264291.80437121.4710.2520.7780.002
Condition:feature26415958.74544329.66711.52 <0.001* 0.105
Group:condition:feature264632.58844329.6670.4570.6350.005

Note: Significant effects are marked with an asterisk.

By looking at the post hoc pairwise comparisons decomposing the condition‐by‐feature interaction (Table 3), we can observe that there are significant differences in latencies, contrasting hi‐hat and piano conditions, for the intensity with 23.62 ms (earlier MMN for piano tones, p < 0.001) and timbre with −19.68 ms (earlier MMN for hi‐hat tones, p = 0.002). No significant differences in latency between conditions were found for location (p = 0.570). Similarly to MMN mean amplitudes (Table 2), the condition‐by‐feature interaction could be explained with significant differences in between‐condition contrasts comparing intensity to timbre (p < 0.001) and location to timbre (p = 0.025) but not intensity to location (p = 0.065).
TABLE 3

Pairwise contrasts of MMN peak latencies between conditions for each feature separately, as well as pairwise comparisons between features for differences between conditions (contrasts of contrasts)

FeatureContrastDifference (ms)CI 2.5%CI 97.5%tdp
IntensityHi‐hat − piano23.6211.4935.753.850.89 <0.001*
LocationHi‐hat − piano3.50−8.6315.630.570.130.570
TimbreHi‐hat − piano−19.68−31.81−7.55−3.20−0.74 0.002*
(Hi‐hat − piano intensity) − (hi‐hat − piano location)20.12−0.941.132.320.760.065
(Hi‐hat − piano intensity) − (hi‐hat − piano timbre)43.2922.2864.314.981.63 <0.001*
(Hi‐hat − piano location) − (hi‐hat − piano timbre)23.182.1644.192.670.87 0.025*

Note: Data are pooled from both groups. Significant contrasts are highlighted in bold and marked with an asterisk. Standard effect sizes (d) are calculated as the difference between conditions divided by the square root of the sum of the residual and the random effects variance.

Pairwise contrasts of MMN peak latencies between conditions for each feature separately, as well as pairwise comparisons between features for differences between conditions (contrasts of contrasts) Note: Data are pooled from both groups. Significant contrasts are highlighted in bold and marked with an asterisk. Standard effect sizes (d) are calculated as the difference between conditions divided by the square root of the sum of the residual and the random effects variance.

P3a analysis

Mixed ANOVA analyses revealed significant main effects and two‐way interactions on P3a mean amplitudes and peak latencies for group, condition, and feature (Figures 4 and 6).
FIGURE 6

Mean P3a amplitudes (left) and peak latencies (right) as a function of conditions and features in both groups. Boxes display median and interquartile ranges. Beans depict the estimated densities. Lines connect measurements for individual participants.

Mean P3a amplitudes (left) and peak latencies (right) as a function of conditions and features in both groups. Boxes display median and interquartile ranges. Beans depict the estimated densities. Lines connect measurements for individual participants. For P3a mean amplitudes (see Table A3), the main effect of feature was significant with F(2,64) = 24.580, η 2 = 0.130, p < 0.001. No significant main effects of group and condition were found F(1,32) = 0.223, p = 0.640 and F(1,32) = 0.863, p = 0.360. Interestingly, a significant condition‐by‐feature interaction was found with F(2,64) = 44.171, η 2 = 0.180, p < 0.001. No significant interactions were found for group‐by‐condition with F(1,32) = 0.341, p = 0.563 and group‐by‐feature with F(2,64) = 0.518, p = 0.598. No significant three‐way interaction was found for group‐by‐condition‐by‐feature with F(2,64) = 1.493, p = 0.232.
TABLE A3

ANOVA table for P3a amplitude analysis: main effects, two‐way and three‐way interaction effects on group, condition, and feature

EffectDFnDFdSSnSSd F p ges
(Intercept)1321338.018502.81685.154 <0.001* 0.594
Group1323.505502.8160.2230.640.004
Condition1322.37888.180.8630.360.003
Feature264136.148177.24924.58 <0.001* 0.13
Group:condition1320.9488.180.3410.5630.001
Group:feature2642.87177.2490.5180.5980.003
Condition:feature264199.872144.79744.171 <0.001* 0.18
Group:condition:feature2646.755144.7971.4930.2320.007

Note: Significant effects are marked with an asterisk.

By looking at the post hoc pairwise comparisons (Table 4), we can observe that there is a significant difference in amplitude, both for location and timbre in both groups, contrasting hi‐hat and piano conditions. However, whereas for location, P3a was smaller for hi‐hat tones compared with piano tones (as already observed for the MMN), for timbre, P3a was larger for hi‐hat tones compared with piano tones. This is reflected in the significant difference between location and timbre for differences between conditions (p < 0.001). Furthermore, while intensity contrast was by itself not significant (p = 0.123), there was a significant difference between intensity and timbre for differences between conditions (p < 0.001). Taken together, these results explain the condition‐by‐feature interaction.
TABLE 4

Pairwise contrasts of P3a mean amplitudes between conditions for each feature separately, as well as pairwise comparisons between features for differences between conditions (contrasts of contrasts)

FeatureContrastDifference (μV)CI 2.5%CI 97.5% t d p
IntensityHi‐hat − piano−0.60−1.370.16−1.55−0.280.123
LocationHi‐hat − piano−1.69−2.46−0.93−4.36−0.78 <0.001*
TimbreHi‐hat − piano2.942.183.717.581.35 <0.001*
(Hi‐hat − piano intensity) − (hi‐hat − piano location)1.09−0.242.421.990.50.146
(Hi‐hat − piano intensity) − (hi‐hat − piano timbre)−3.55−4.88−2.22−6.46−1.63 <0.001*
(Hi‐hat − piano location) − (hi‐hat − piano timbre)−4.64−5.97−3.31−8.44−2.13 <0.001*

Note: Data are pooled from both groups. Significant contrasts are highlighted in bold and marked with an asterisk. Standard effect sizes (d) are calculated as the difference between conditions divided by the square root of the sum of the residual and the random effects variance.

Pairwise contrasts of P3a mean amplitudes between conditions for each feature separately, as well as pairwise comparisons between features for differences between conditions (contrasts of contrasts) Note: Data are pooled from both groups. Significant contrasts are highlighted in bold and marked with an asterisk. Standard effect sizes (d) are calculated as the difference between conditions divided by the square root of the sum of the residual and the random effects variance. For mean P3a latencies (see Table A4), the main effects of condition and feature were significant, F(1,32) = 14.474, η 2 = 0.033, p < 0.001 and F(2,64) = 45.605, η 2 = 0.289, p < 0.001, respectively. No significant main effect was found for group F(1,32) = 0.297, p = 0.590. No significant interaction was found for group‐by‐feature F(2,64) = 1.203, p = 0.307. Interestingly, an interaction was found for group‐by‐condition F(1,32) = 5.094, η 2 = 0.012, p = 0.031 and condition‐by‐feature F(2,64) = 4.892, η 2 = 0.039, p = 0.011. No significant three‐way interaction between group, condition, and feature was observed, F(2,64) = 0.242, p = 0.786.
TABLE A4

ANOVA table for P3a latency analysis: main effects, two‐way and three‐way interaction effects on group, condition, and feature

EffectDFnDFdSSnSSd F p ges
(Intercept)13211495477.824115803.5293176.546 <0.001* 0.974
Group1321073.647115803.5290.2970.590.003
Condition13210737.25523738.90214.474 0.001* 0.033
Feature264127284.35389313.41245.605 <0.001* 0.289
Group:condition1323778.84323738.9025.094 0.031* 0.012
Group:feature2643356.23589313.4121.2030.3070.011
Condition:feature26412800.27583738.514.892 0.011* 0.039
Group:condition:feature264633.21683738.510.2420.7860.002

Note: Significant effects are marked with an asterisk.

Post hoc pairwise comparisons (Table 5) revealed significant differences in latencies, contrasting hi‐hat and piano conditions, for intensity with 31.59 ms and location with 18.53 ms (earlier P3a with piano tones in both cases). For timbre, no significant differences in latency between conditions were found (p = 0.440). Furthermore, while the difference between hi‐hat and piano conditions was significant for the control group (23.12 ms; 95% CI: 9.40–36.83; t(160) = 3.33; d = 0.57; p = 0.001), this was not the case for amusics (5.90 ms; 95% CI: −7.81‐19.62; t(160) = 0.85; d = 0.15; p = 0.396).
TABLE 5

Pairwise contrasts of P3a peak latencies between conditions for each feature separately, as well as pairwise comparisons between features for differences between conditions (contrasts of contrasts)

FeatureContrastDifference (ms)CI 2.5%CI 97.5% t d p
IntensityHi‐hat − piano31.5914.7948.393.710.78 <0.001*
LocationHi‐hat − piano18.531.7335.332.180.46 0.031*
TimbreHi‐hat − piano−6.59−23.3910.21−0.77−0.160.440
(Hi‐hat − piano intensity) − (hi‐hat − piano location)13.06−16.0442.161.090.320.838
(Hi‐hat − piano intensity) − (hi‐hat − piano timbre)38.189.0767.283.170.95 0.005*
(Hi‐hat − piano location) − (hi‐hat − piano timbre)25.12−3.9954.222.090.620.115

Note: Data are pooled from both groups. Significant contrasts are highlighted in bold and marked with an asterisk. Standard effect sizes (d) are calculated as the difference between conditions divided by the square root of the sum of the residual and the random effects variance.

Pairwise contrasts of P3a peak latencies between conditions for each feature separately, as well as pairwise comparisons between features for differences between conditions (contrasts of contrasts) Note: Data are pooled from both groups. Significant contrasts are highlighted in bold and marked with an asterisk. Standard effect sizes (d) are calculated as the difference between conditions divided by the square root of the sum of the residual and the random effects variance.

DISCUSSION

In this study, we found that the harmonicity of musical sounds enhances MMN in response to intensity, timbre, and location deviants, both in control and amusic participants. This is consistent with the hypothesis that harmonicity influences the processing of auditory features other than pitch. In the following, an explanation based on precision weighting of prediction errors will be discussed.

Larger MMN for harmonic sounds

Harmonic sounds produced larger MMNs than inharmonic sounds, as evidenced by statistically significant main effects of condition (piano vs. hi‐hat) on MMN amplitudes. Piano tones had clear harmonicity, whereas hi‐hat tones had a more broadband spectrum (Figure 1). Other acoustic features were matched between the two: Duration, impulsivity (sharp attack and fast decay as both types of sounds are generated by a single percussion), intensity, location; both had ecological validity. One possible interpretation for the larger MMN amplitudes in harmonic sounds comes from a predictive coding account of perception (Clark, 2015; Friston, 2005; Rao & Ballard, 1999). According to predictive coding, the brain employs a generative model to form predictions about the incoming sensory stimuli. The discrepancies between predictions and the sensory data, in the form of prediction errors, are used to update the generative model and enable better predictions in the future. This system is suggested to be hierarchical, with higher levels predicting the future states of lower levels, and the hierarchy is thought to be reflected in the neuroanatomical structure of the nervous system (Kanai et al., 2015). In predictive coding, the generative model is only updated by prediction errors with low self‐estimated sensory uncertainty, or high precision. This precision‐weighting process enables the system to “filter out” sensory data that are noisy, uncertain, or otherwise unreliable and are analogous to attention (Kanai et al., 2015; Koelsch et al., 2019). ERP responses to deviances, such as MMN, are thought to be electrophysiological markers of ascending precision‐weighted prediction errors (Garrido et al., 2007, 2009; Koelsch et al., 2019; Lecaignard et al., 2022; Winkler & Czigler, 2012; see also Grimm & Schroger, 2007, for a related, predictive account of MMN generation). Precision of prediction errors is related to Shannon entropy, a metric that captures the amount of information content (uncertainty, expected surprise) of a given signal (Shannon, 1948). In individual sounds, entropy crucially depends on the amount of periodicity in the waveform shape and is thus directly related to harmonicity. Harmonic sounds have low spectral entropy because their waveforms can be reliably predicted using low amounts of information (as their frequency spectra consist of F0 and its integer multiples). Conversely, sounds that are inharmonic have higher spectral entropy and their waveforms are harder to predict (the frequency components do not form a harmonic series). Thus, the precision of prediction errors arising from deviances in inharmonic sounds is bound to be lower than in harmonic sounds. Consequently, the lower amplitudes of MMN signals for inharmonic sounds may result from a stronger down‐weighting of sensory signals due to the spectral entropy of the acoustic input. Similar effects of information entropy on prediction error responses measured by MMN (or its magnetic counterpart, the MMNm) were reported previously for rhythmic (Lumaca et al., 2019) and melodic sequences (Quiroga‐Martinez et al., 2019, 2020). Furthermore, Takegata et al. (2008) investigated MMN in response to duration deviants for noise and other auditory stimuli. They found smaller MMN amplitudes in the noise condition in comparison with other (harmonic) stimuli. Although consistent with findings of the present study, these results were interpreted in terms of lower ecological validity of noise. The results of the present study suggest an effect of precision weighting on prediction errors on a very short timescale during the unfolding of auditory signals, that is, related to the instantaneous processing of the spectral content of sound.

Differences between features in harmonic and non‐harmonic sounds

We found that the strength of MMN responses was dependent on sound feature in different ways for the two types of sounds (piano vs. hi‐hat), as evidenced by the significant interaction effect between condition and feature for MMN amplitudes and latencies. While the MMN was generally larger for harmonic than for inharmonic sounds, the differences between MMN amplitudes were less pronounced for timbre than for intensity and location deviants. Additionally, the MMN amplitudes for intensity and location deviants in the inharmonic condition were very small. MMN latencies were also dependent on sound feature, and this relationship was moderated by condition. For intensity deviants, harmonic sounds produced earlier peaks than inharmonic sounds, which is consistent with the idea that harmonic sounds are processed faster. Conversely, for timbre deviants, harmonic sounds produced later MMN peaks. This effect may be attributed to a large P3a response that followed the initial negativity (see below), possibly partly masking the MMN, especially for the hi‐hat sound. No differences in latency between harmonic and inharmonic stimuli were found for the location deviants. As broadband noise stimuli are easier to locate than tones (Terhune, 1985), the broadband nature of the hi‐hat sound may have counterbalanced the expected harmonicity effect. Additionally, MMN latency for location deviants was short (in comparison with latencies observed for classical MMN responses; Näätänen et al., 2007), suggesting that processing of location could be performed faster, on lower levels of the auditory hierarchy. While MMN amplitudes were smaller for timbre than for other features, P3a amplitudes were larger, and were larger for hi‐hat tones than piano tones. P3a indicates when the mental representation of a stimulus is updated (Donchin, 1981). Because of its longer latency, P3a is suggested to originate at higher levels of the auditory processing hierarchy than the MMN (Koelsch et al., 2019). The P3a is also associated with the initiation of an automatic attention‐orienting response to stimulus deviance (Polich & Criado, 2006). Stronger P3a in our study could indicate a stronger implicit shift of attention to timbre deviants in comparison to other features. This shift may happen faster for timbre deviants, as evidenced by shorter P3a latencies. Furthermore, these strong and fast P3a responses for timbre could have interfered with activity in earlier latencies, potentially resulting in a diminished MMN response. The distinct pattern of responses for timbre deviants may have to do with the filtering applied to the sounds to generate the deviants. Due to differences in the spectral characteristics of the piano and hi‐hat, the filter changed these sounds in a different way. Because the hi‐hat sound had a more uniformly distributed frequency spectrum, the filter introduced a timbral change that might have been perceived as a stronger deviance in comparison to the piano sound. It is thus possible that for the timbre condition, the different patterns of MMN and P3a might have resulted not from differences related to harmonicity, but from a more salient spectral change in the hi‐hat condition. This problem is inherently related to comparing real‐world harmonic and inharmonic sounds that (necessarily) differ in frequency spectra. One way to address this problem would be to investigate natural sounds with artificially manipulated harmonic components (as in Popham et al., 2018). Another approach could involve fully synthesized sounds with carefully controlled manipulations in the frequency spectra, yet this may negatively impact the ecological validity of the stimuli. This filter‐related restriction however applies only to timbre deviants and does not extend to other auditory features.

Preserved MMN and P3a in CA for harmonic and non‐harmonic sounds

We found significant MMN and P3a responses, both in CA and control participants and no significant main effects of group or group–condition, group–feature, or group–feature–condition interactions. This was true for mean amplitudes as well as mean latencies in both MMN and P3a analyses, with the exception of a group‐by‐condition interaction in P3a latency (discussed below). Whether or not the processing deficits in CA extend to the processing of harmonicity is debated. It was reported that amusics are insensitive to harmonicity cues, contrarily to beating cues (Cousineau et al., 2012; but see Graves et al., 2021), and they might rely on roughness cues more than harmonicity cues to evaluate the pleasantness of a sound (Marin et al., 2012). On the other hand, the identification of pitch processing areas with fMRI by contrasting harmonic sounds with noises lead to results very similar to that obtained in controls (Norman‐Haignere et al., 2016), yet with subtle differences revealed by multivariate analyses (Albouy, Caclin, et al., 2019). During active pitch short‐term memory tasks, functional impairments are observed in a fronto‐temporal network in CA (Albouy et al., 2013; Albouy, Peretz, et al., 2019). Our results suggest that individuals with CA do not differ from matched controls in terms of early neural processing of auditory features other than pitch (see Quiroga‐Martinez et al., 2021, for converging evidence). ERP studies with pitch deviants in classical oddball paradigms also report intact MMN responses in CA listeners, indicating normal pitch processing at an early level of auditory processing (Moreau et al., 2013; Zendel et al., 2015), whereas abnormalities might be observed in more complex sound sequences (Quiroga‐Martinez et al., 2021). Taken together, these results provide further evidence that deficits in CA might mainly result from abnormalities at higher levels of the perceptual hierarchy, possibly related to impaired top‐down connectivity in the fronto‐temporal network (Albouy et al., 2013; Peretz, 2016). Our results can be also interpreted as evidence that low‐level auditory processing in CA remains greatly spared, not only for pitch but also for harmonicity, even though the deficit might lead to delayed processing with impaired auditory cortex contributions (see Albouy et al., 2013). This is consistent with the hypothesis that CA individuals might not have conscious access to the results of low‐level processing and might thus be less able to use them in perceptual decision making (Peretz, 2016). In P3a latency analysis, we found a significant group‐by‐condition interaction. This was accompanied with a significant main effect of condition (piano sounds yielded generally shorter P3a latencies than hi‐hat sounds) but not a main effect of group. The strength of this effect is however very small (η 2 = 0.012), much smaller than the main effects of condition (η 2 = 0.033) or feature (η 2 = 0.289). This interaction can be interpreted as suggesting that for inharmonic sounds, the CA individuals require more processing time to consciously detect deviance. However, because of the small effect size, further research is needed to clarify this finding.

LIMITATIONS AND CONCLUSIONS

A limitation of our study is that subjects listened to the stimuli passively, without any perceptual tasks. Our goal was to focus on low‐level processing of auditory deviances and attention‐orientation as measured with the MMN and P3a, respectively. We chose a passive listening approach because it is optimal for evaluating the MMN. More active tasks (such as actively detecting deviants with button‐presses) could be used to investigate neural processing that involves higher levels of the perceptual hierarchy, with a focus on later ERPs. These types of tasks could also arguably reveal differences in processing specific to CA that were not found in the current study. Finally, it would be beneficial to investigate a more diverse set of auditory stimuli with different levels of inharmonicity to better understand the relationship between the spectral properties of sound and the MMN. Piano and hi‐hat sounds differ not only in harmonicity, but in other features as well (e.g., amplitude envelope, spectral content). Future studies may attempt to control for this issue using either complex tones with jittered frequencies, or use natural sounds with artificially manipulated harmonics. In a recent study with synthetic tones (Graves et al., in revision), we observed larger MMNs for harmonic deviants within a sequence of inharmonic standards than the reverse (inharmonic deviants within a sequence of harmonic standards), paving the way for future studies were individual sound entropies could be carefully controlled in oddball sequences. The results of the present study provide neurophysiological evidence consistent with the notion that harmonicity plays an important role in human auditory perception, starting in early stages of sound processing in the auditory networks. A detailed investigation into the relationships between early auditory ERPs and harmonicity can provide valuable information about the nature of auditory information processing in the brain.

CONFLICT OF INTEREST

The authors declare no conflict of interest.

AUTHOR CONTRIBUTION

David Ricardo Quiroga‐Martinez: Conceptualization, Methodology, Software, Formal analysis, Data curation, Writing ‐ original draft, Visualization, Investigation. Krzysztof Basiński: Software, Writing ‐ original draft, Visualization. Jonathan Nasielski: Software, Writing ‐ original draft, Visualization. Barbara Tillmann: Conceptualization, Supervision, Writing ‐ original draft. Elvira Brattico: Conceptualization, Supervision, Writing ‐ original draft. Fanny Cholvy: Investigation, Data curation. Lesly Fornoni: Investigation, Data curation. Peter Vuust: Conceptualization, Methodology, Supervision, Writing ‐ original draft. Anne Caclin: Conceptualization, Methodology, Writing ‐ original draft, Supervision.

PEER REVIEW

The peer review history for this article is available at https://publons.com/publon/10.1111/ejn.15769.
  39 in total

1.  The mismatch negativity (MMN): towards the optimal paradigm.

Authors:  Risto Näätänen; Satu Pakarinen; Teemu Rinne; Rika Takegata
Journal:  Clin Neurophysiol       Date:  2004-01       Impact factor: 3.708

Review 2.  Evidence from auditory and visual event-related potential (ERP) studies of deviance detection (MMN and vMMN) linking predictive coding theories and perceptual object representations.

Authors:  István Winkler; István Czigler
Journal:  Int J Psychophysiol       Date:  2011-10-30       Impact factor: 2.997

3.  Weighting of neural prediction error by rhythmic complexity: A predictive coding account using mismatch negativity.

Authors:  Massimo Lumaca; Niels Trusbak Haumann; Elvira Brattico; Manon Grube; Peter Vuust
Journal:  Eur J Neurosci       Date:  2019-01-20       Impact factor: 3.386

4.  Attending to pitch information inhibits processing of pitch information: the curious case of amusia.

Authors:  Benjamin Rich Zendel; Marie-Élaine Lagrois; Nicolas Robitaille; Isabelle Peretz
Journal:  J Neurosci       Date:  2015-03-04       Impact factor: 6.167

5.  Localization of pure tones and click trains by untrained humans.

Authors:  J M Terhune
Journal:  Scand Audiol       Date:  1985

6.  The basis of musical consonance as revealed by congenital amusia.

Authors:  Marion Cousineau; Josh H McDermott; Isabelle Peretz
Journal:  Proc Natl Acad Sci U S A       Date:  2012-11-12       Impact factor: 11.205

7.  Congenital amusia: a short-term memory deficit for non-verbal, but not verbal sounds.

Authors:  Barbara Tillmann; Katrin Schulze; Jessica M Foxton
Journal:  Brain Cogn       Date:  2009-09-16       Impact factor: 2.310

8.  Specialized neural dynamics for verbal and tonal memory: fMRI evidence in congenital amusia.

Authors:  Philippe Albouy; Isabelle Peretz; Patrick Bermudez; Robert J Zatorre; Barbara Tillmann; Anne Caclin
Journal:  Hum Brain Mapp       Date:  2018-11-01       Impact factor: 5.038

9.  Inharmonic speech reveals the role of harmonicity in the cocktail party problem.

Authors:  Sara Popham; Dana Boebinger; Dan P W Ellis; Hideki Kawahara; Josh H McDermott
Journal:  Nat Commun       Date:  2018-05-29       Impact factor: 14.919

10.  Decoding Task-Related Functional Brain Imaging Data to Identify Developmental Disorders: The Case of Congenital Amusia.

Authors:  Philippe Albouy; Anne Caclin; Sam V Norman-Haignere; Yohana Lévêque; Isabelle Peretz; Barbara Tillmann; Robert J Zatorre
Journal:  Front Neurosci       Date:  2019-10-30       Impact factor: 4.677

View more
  1 in total

1.  Enhanced mismatch negativity in harmonic compared with inharmonic sounds.

Authors:  David Ricardo Quiroga-Martinez; Krzysztof Basiński; Jonathan Nasielski; Barbara Tillmann; Elvira Brattico; Fanny Cholvy; Lesly Fornoni; Peter Vuust; Anne Caclin
Journal:  Eur J Neurosci       Date:  2022-07-26       Impact factor: 3.698

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.