Literature DB >> 33742708

Deficient auditory emotion processing but intact emotional multisensory integration in alexithymia.

Zhihao Wang1,2, Mai Chen3, Katharina S Goerlich2, André Aleman1,2, Pengfei Xu4,5,6, Yuejia Luo1,4,7,8,5.   

Abstract

Alexithymia has been associated with emotion recognition deficits in both auditory and visual domains. Although emotions are inherently multimodal in daily life, little is known regarding abnormalities of emotional multisensory integration (eMSI) in relation to alexithymia. Here, we employed an emotional Stroop-like audiovisual task while recording event-related potentials (ERPs) in individuals with high alexithymia levels (HA) and low alexithymia levels (LA). During the task, participants had to indicate whether a voice was spoken in a sad or angry prosody while ignoring the simultaneously presented static face which could be either emotionally congruent or incongruent to the human voice. We found that HA performed worse and showed higher P2 amplitudes than LA independent of emotion congruency. Furthermore, difficulties in identifying and describing feelings were positively correlated with the P2 component, and P2 correlated negatively with behavioral performance. Bayesian statistics showed no group differences in eMSI and classical integration-related ERP components (N1 and N2). Although individuals with alexithymia indeed showed deficits in auditory emotion recognition as indexed by decreased performance and higher P2 amplitudes, the present findings suggest an intact capacity to integrate emotional information from multiple channels in alexithymia. Our work provides valuable insights into the relationship between alexithymia and neuropsychological mechanisms of emotional multisensory integration.
© 2021 The Authors. Psychophysiology published by Wiley Periodicals LLC on behalf of Society for Psychophysiological Research.

Entities:  

Keywords:  TAS-20; alexithymia; emotion recognition deficits; emotional multisensory integration; event-related potentials (ERP)

Mesh:

Year:  2021        PMID: 33742708      PMCID: PMC9285530          DOI: 10.1111/psyp.13806

Source DB:  PubMed          Journal:  Psychophysiology        ISSN: 0048-5772            Impact factor:   4.348


INTRODUCTION

Alexithymia, a subclinical personality trait, is characterized as an impaired ability to identify, describe, and regulate one's emotions (Luminet et al., 2018). Ample evidence has demonstrated that alexithymia is not only associated with difficulties in identifying visual emotional stimuli (Van der Velde et al., 2013), especially sad and angry facial expressions (Kano et al., 2003; Lee et al., 2011), but high alexithymia levels are also related to a blunted sensitivity to the emotional qualities of speech (Goerlich‐Dobre et al., 2014). Indeed, individuals with alexithymia may need more attentional resources to process emotion‐laden contents (Franz et al., 2004). Emotions, however, are inherently multimodal in daily life (Tang et al., 2016). It has been shown that emotional multisensory integration (eMSI), the ability to bundle emotional information synergistically from different modalities, plays a key role in emotion recognition (Klasen et al., 2012). However, it is currently unclear whether individuals with alexithymia have deficits in eMSI. This is of particular importance given its ecological relevance, as humans are constantly exposed to competing, complex audiovisual emotional information in social interaction contexts. Indeed, it has been shown that a reliable interpretation of others' affective states requires the integration of multimodal stimuli into a single, coherent percept (see De Gelder & Bertelson, 2003, for a review). There have been two eMSI‐related studies in alexithymia. Delle‐Vigne et al. (2014) observed a larger N1 component in alexithymia during an emotional oddball audiovisual task, suggesting that individual with alexithymia need more attentional resources to process emotionally congruent audiovisual cues. However, the impact of emotion recognition from visual clues or from the voice on eMSI in alexithymia was not investigated. Another study showed a negative correlation of alexithymia levels with the N400 for emotionally incongruent compared with congruent audiovisual stimuli during an affective priming task, indicating reduced sensitivity to affective mismatches in audiovisual information with higher levels of alexithymia (Goerlich et al., 2011). This study did not test clinically relevant alexithymia levels (only two individuals had alexithymia scores exceeding the clinical threshold). Therefore, it remains uncertain whether the observed effects can be attributed to abnormal eMSI in clinically‐relevant alexithymia. Previous studies suggested that eMSI is impaired in patients with affective disorders (Müller et al., 2013), even in subclinical populations (Campanella et al., 2010). Given that alexithymia has long been considered as a risk factor for affective disorders (highly correlated with anxiety and depression; Hendryx et al., 1991; Li et al., 2015), it is reasonable to speculate that eMSI may be reduced in alexithymia (hypo‐integration hypothesis). However, the opposite may also be hypothesized: stronger integration as a compensation for reduced emotion perception ability (hyper‐integration hypothesis). That is, given that alexithymia seems to be associated with a reduced intensity perception of emotional visual and auditory stimuli (for a review, see Donges & Suslow, 2017), individuals with high alexithymia levels may show better eMSI than a control group with average alexithymia levels. If so, eMSI would function as a compensation for deficits in visual or auditory emotion processing. Event‐related potential (ERP) techniques, with high temporal resolution, have been used to shed light on the time course of eMSI (Klasen et al., 2012; Wang et al., 2012). It has been shown that there are two stages of eMSI processing, including multisensory facilitation and multisensory inhibition (Campanella & Belin, 2007; Wang et al., 2012). Multisensory facilitation, a sensory process during which the emotional attention triggered by one domain is amplified by another modality to increase efficiency (Klasen et al., 2012), occurs when bimodal stimuli have a close spatial and temporal coincidence (Stein & Stanford, 2008) or when the bimodal stimuli are congruent, leading to priming effects (Chen et al., 2016; Föcker & Röder, 2019). Multisensory facilitation can be reflected by a larger auditory N1 component, a negative deflection over frontal‐central regions (Luck, 2014; Luck et al., 2000). By contrast, multisensory inhibition occurs when simultaneously presenting bimodal stimuli that are incongruent with each other, for example in their affective or semantic connotation (Wang et al., 2012). This incongruency presents a conflict, causing a need to allocate more attentional resources to disambiguate and select a correct response, resulting in prolonged reaction times (Donohue et al., 2013). At the neural level, increases in the frontal‐central N2 have been repeatedly reported for emotionally incongruent audiovisual stimuli, signaling conflict monitoring and disambiguation processes in eMSI (Zinchenko et al., 2017). Taken together, abnormally high or low amplitude of the classic eMSI‐related components N1 and N2 can be regarded as indices of abnormal neuropsychological processes based on the hyper‐ or hypo‐integration hypotheses. The aim of the current study was to examine eMSI ability in relation to alexithymia by applying ERP techniques in a Stroop‐like visual‐auditory task. During the task, participants were asked to identify whether someone's voice speaking in a sad or angry prosody while ignoring the simultaneously presented static face which could be either emotionally congruent or incongruent to the human voice. According to the hypo‐ or hyper‐integration hypothesis, we hypothesized that individuals with high alexithymia levels (HA) would show abnormal integrative performance and differences in early ERP components during emotional multisensory integration compared with individuals with low alexithymia levels (LA). Specifically, regarding the hypothesis of hypo‐integration, one would expect worse integrative performance in HA versus LA. At the neural level, this would be reflected in smaller differences in N1 and N2 between congruent and incongruent conditions in HA as compared with LA. With regard to the hypothesis of hyper‐integration, better integrative performance would be predicted in HA versus LA. At the neural level, this would be reflected in larger differences in N1 and N2 between congruent and incongruent conditions in HA as compared with LA. In addition to eMSI‐relevant hypotheses, we hypothesized that HA would show deficits in auditory emotion recognition as indexed by the P2 (replicating previous research; Franz et al., 2004).

METHOD

Participants

Forty‐eight healthy adults from a pool of 543 (368 females; age: 17–38 years, mean ± SD: 20.03 ± 2.11 years) students at Shenzhen University participated in the experiment. Each participant in the pool completed the Chinese version of the 20‐item Toronto Alexithymia Scale (TAS‐20; Bagby et al., 1994a; Zhu et al., 2007). In light of the international cutoff to assess clinical‐relevant alexithymia in TAS‐20 (Taylor et al., 1988), individuals with TAS‐20 scores higher or equal to 61 (14.9% of the pool) were identified as high alexithymia (HA) whereas those with TAS‐20 scores lower or equal to 51 (54.6% of the pool) were classified as low alexithymia (LA; Table 1). The sample size of 48 participants (Table 1) was determined based on the medium effect size by G*power (version: 3.1; Faul et al., 2007). Twenty‐three participants per group were needed to detect a reliable effect (Cohen f = 0.25, α = .05, 1 − β = .9, ANOVA Repeated‐measures, within‐between interaction; Faul et al., 2007). All participants had normal or corrected‐to‐normal vision and normal hearing and reported no psychiatric illness in present or past. This study was approved by the Ethics Committee of Shenzhen University and informed written consents were obtained from all participants.
TABLE 1

Demographics and questionnaire scores

HA (25; 12 females)LA (23; 12 females) t p Cronbach's alpha
Mean (SD)[Min, max]Mean (SD)[Min, max]
Age19.96 (1.46)[18, 24]19.78 (1.81)[17, 24]0.376.709
TAS‐2066.08 (2.96)[62, 73]36.09 (5.01)[22, 43]24.992<.0010.856
DIF23.56 (2.38)[20, 27]11.09 (2.68)[7, 16]17.077<.0010.916
DDF18.52 (1.98)[14, 22]8.74 (1.91)[5, 13]17.373<.0010.534
EOT24.00 (3.01)[19, 32]16.26 (3.31)[10, 24]8.485<.0010.369
BDI12.92 (6.36)[3, 27]4.00 (4.68)[0, 17]5.492<.0010.870
BAI33.92 (6.36)[24, 48]25.00 (4.68)[21, 38]5.492<.0010.857
AQ124.44 (10.21)[100, 142]109.39 (8.42)[89, 120]5.543<.0010.673

Abbreviations: AQ, Autism Spectrum Quotient;BAI, beck anxiety inventory; BDI, beck depression inventory; DDF, difficulty describing feelings; DIF, difficulty identifying feelings; EOT, externally oriented thinking; HA, individuals with alexithymia; LA, individuals without alexithymia; TAS‐20, twenty‐item Toronto Alexithymia Scale.

Demographics and questionnaire scores Abbreviations: AQ, Autism Spectrum Quotient;BAI, beck anxiety inventory; BDI, beck depression inventory; DDF, difficulty describing feelings; DIF, difficulty identifying feelings; EOT, externally oriented thinking; HA, individuals with alexithymia; LA, individuals without alexithymia; TAS‐20, twenty‐item Toronto Alexithymia Scale.

Stimulus materials

The vocal stimuli were 64 pseudo‐sentences uttered in sad and angry prosody selected from the Database of Chinese Vocal Emotions (DoCVE; Liu & Pell, 2012). These angry (32 items) and sad (32 items) pseudo‐utterances were recorded by four native speakers (two females; eight items per speaker). In light of previous studies (Liu et al., 2015), each item was cut down to 800 ms using the Praat software (Boersma & Weenink, 2021) to avoid ceiling effects and to control the amount of expressed information across stimuli. The original reflected the whole utterance. Although the whole utterances selected from this dataset had matched (Table S1), it is uncertain how cutting to 800 ms would affect them. Therefore, we conducted a behavioral follow‐up. To check whether emotional properties of the 800 ms voice datasets were matched as those of the original categorizations, a follow‐up rating was conducted on the auditory stimuli in each group (see Supporting Information for details). Although angry and sad utterances used in the current study were not strictly matched in emotional properties of intensity and valence (Table 2), the absence of significant group differences in these parameters suggests that alexithymia‐related effects were not confounded by basic emotional characteristics (for details, see Supporting Information). The amplitude values (maximum amplitude, minimum amplitude) were normalized in relation to the average minimum amplitude value of all neutral utterances by that speaker, to correct for individual differences in a speaker's intensity by subtracting the normalized minimum amplitude values from the normalized maximum amplitude values (Liu & Pell, 2012). Thus, all vocal stimuli had approximate equal intensity. Note that pseudo‐sentences are composed of real function words, but the sentence content is meaningless to avoid influences of the linguistic‐semantic content during speech. The vocal stimuli were expressed in stereo mode. The visual stimuli of 64 grey‐scaled emotional faces were selected from the Chinese Facial Affective Picture System (CFAPS; Gong et al., 2011). These emotional faces consisted of 32 angry (16 females) and 32 sad (16 females) faces. We also balanced the emotional characteristics between angry and sad faces (Table S1). Specifically, no significant differences were found between angry and sad faces in valence, arousal, and attractiveness (Table S1), suggesting that angry and sad facial stimuli matched in these emotional dimensions. The visual stimuli of faces subtended a visual angle of 3° × 4°.
TABLE 2

Follow‐up ratings for voice stimuli with angry and sad emotions in each group

GroupCategorizationRecognition rateIntensityValence
HAAngry0.67 ± 0.194.04 ± 0.602.01 ± 0.46
Sad0.68 ± 0.183.65 ± 0.652.18 ± 0.60
MAAngry0.66 ± 0.184.10 ± 0.372.08 ± 0.48
Sad0.64 ± 0.223.71 ± 0.492.19 ± 0.38
LAAngry0.72 ± 0.224.21 ± 0.311.86 ± 0.38
Sad0.75 ± 0.193.74 ± 0.442.05 ± 0.43

Descriptive data are presented as mean (standard deviation).

Abbreviations: HA, individuals with high level of alexithymia; LA, individuals with low level of alexithymia; MA, individuals with middle levels of alexithymia.

Follow‐up ratings for voice stimuli with angry and sad emotions in each group Descriptive data are presented as mean (standard deviation). Abbreviations: HA, individuals with high level of alexithymia; LA, individuals with low level of alexithymia; MA, individuals with middle levels of alexithymia.

Self‐report questionnaires

The TAS‐20 measures three facets of alexithymia: (a) difficulty identifying feelings (DIF; 7 items); (b) difficulty describing feelings (DDF; 5 items), and (c) externally oriented thinking (EOT; 8 items). In a self‐reporting manner, each item is rated on a five‐point Likert scale ranging from 1 (“strongly disagree”) to 5 (“strongly agree”), with five items being negatively scored. For analysis, the negatively keyed items are reverted and item scores for each respective dimension are summed up. The total score is calculated as the sum of all items. High scores represent high levels of alexithymia (Bagby et al., 1994; Bagby, Taylor, et al., 1994; Taylor et al., 2003). Importantly, the Chinese version of the TAS‐20 has been established with acceptable reliability and validity (Zhu et al., 2007). To control for the potential confounding effect of depression, anxiety, and autism (Bird & Cook, 2013; Li et al., 2015; Van der Velde et al., 2013), participants also completed the Beck Anxiety Inventory (BAI; Beck et al., 1988), the Beck Depression Inventory (BDI; Beck, 1967), and the Autism Spectrum Quotient (AQ; Baron‐Cohen et al., 2001).

Task and procedure

We adopted an emotional Stroop‐like visual‐auditory task (Figure 1). In each trial, a fixation cross was first shown on the screen center with a random duration (800~1,500 ms). Subsequently, the face‐voice pairs were presented simultaneously. The voice lasted for 800 ms maximally while the face was displayed for 2,500 ms maximally. Once they had identified the emotion (angry or sad) from the voice, participants pressed button “f” if the voice was angry and “j” if the voice was sad. The assignment of the two response buttons was counterbalanced across participants. While viewing the static face all the time (2,500 ms maximally), presentation of the voice stimuli stopped once a button was pressed (800 ms maximally). Finally, a blank screen lasted for 1,000 ms, and then the next trial began.
FIGURE 1

Trial design. In each trial, the face‐voice pair expressing congruent or incongruent emotions (angry–angry, angry–sad, sad–angry, and sad–sad pairs) was presented simultaneously. Faces were presented with a duration of 2,500 ms maximally, accompanied simultaneously by the voice stimuli (pseudowords) lasting for 800 ms maximally. Participants pressed a button as soon as they had identified the emotion from the voice (self‐paced, maximally trial duration 2,500 ms). Finally, a blank screen lasted for 1,000 ms and then the next trial began. The upward and downward arrows mark the beginning of the auditory and visual stimuli. SOA, stimulus onset asynchrony

Trial design. In each trial, the face‐voice pair expressing congruent or incongruent emotions (angry–angry, angry–sad, sad–angry, and sad–sad pairs) was presented simultaneously. Faces were presented with a duration of 2,500 ms maximally, accompanied simultaneously by the voice stimuli (pseudowords) lasting for 800 ms maximally. Participants pressed a button as soon as they had identified the emotion from the voice (self‐paced, maximally trial duration 2,500 ms). Finally, a blank screen lasted for 1,000 ms and then the next trial began. The upward and downward arrows mark the beginning of the auditory and visual stimuli. SOA, stimulus onset asynchrony There were four same‐gender face‐voice pairs (sad–sad, sad–angry, angry–sad, angry–angry). After dividing four pairs into two conditions: congruent (Con; sad–sad, angry–angry) and incongruent (Incon; sad–angry, angry–sad), each condition included 128 trials. Trial presentation was randomized. Participants were instructed to identify the vocal expression (sad or angry) as accurate and quick as possible while ignoring the facial expression. All experimental procedures were presented using the E‐Prime 2.0 software (Psychology Software Tools, Pittsburgh, PA). In addition to the approximate equal intensity of all vocal stimuli (mentioned above), the vocal stimuli were presented binaurally by an insert headphone with equal sound intensity across participants (half of sound intensity of our computer). All participants reported to hear the vocal stimuli clearly. No participant reported having felt uncomfortable at any part of the experiment. The current study adopted the voice task instead of the facial task, because it has been demonstrated that the eMSI effect on Chinese participants is more sensitive to the voice task (Wang et al., 2012). Specifically, during a timed target detection task (visual target and auditory target), the significant integration effect was found from the auditory target whereas no significant integration effect from the visual target was observed (Wang et al., 2012). Participants completed several practice rounds until achieving at least 80% of accuracy before the formal experiment started. There were four blocks with 64 trials per block. Participants could take self‐paced breaks between blocks. After finishing the task, participants rated the intensity of all stimuli (face and voice) on a nine‐point Likert scale.

Electrophysiology (EEG) recording and preprocessing

We recorded EEG data from a 64‐electrode scalp cap according to the international 10–20 system (Brain Products, Munich, Germany), with the reference to the channel FCz. The electrooculogram (EOG; vertical) was recorded with electrodes placed below the right eye. Electrodes impedances of EEG and EOG were maintained <5 kΩ. All electrodes were amplified using a 0.01 online high‐pass filter and continuously sampled at 1,000 Hz per channel for offline analysis. EEG data were preprocessed with EEGLAB 14.1.2b (http://sccn.ucsd.edu/eeglab/) in Matlab 2014b (MathWorks Inc). It comprised of the following steps: (a) resampling to 250Hz; (b) low‐pass filtering of 30Hz by FIR filter; (c) segmentation from 200ms before to 800ms after stimuli onset; (d) baseline correction from −200 to 0 ms; (e) manually rejecting salient muscle epochs and bad channels if any; (f) Independent Component Analysis (ICA); (g) visually inspecting and rejecting artifact components (horizontal and vertical eye movements and muscle components); (h) interpolating bad channels if any; (i) re‐referencing offline to the average of all electrodes; (j) rejecting trials in which EEG voltages were out of range (−80, 80) μV.

Behavioral statistics

We used SPSS 17.0 to perform statistical analyses, with the significance level at p = .05. Trials with response times (RTs) out of range (mean‐3 × SD, mean + 3×SD) were removed for each participant. The impact of alexithymia on congruency was tested in 2 (Con vs. Incon) by 2 (HA vs. LA) repeated‐measures ANOVA with the corrected performance (ACC/RT; speed‐accuracy tradeoff) as the dependent variable. To change corrected performance between HA and LA into the common scales (because growing evidence suggested deficits of auditory emotional processing in alexithymia; Goerlich et al., 2011; Goerlich‐Dobre et al., 2014), we normalized the corrected performance for each participant as the index of integration effect ([Con − Incon]/[Con + Incon]). This index signals the individual point of the best balance between ACC and RT and thus indicates each participant's “sweet spot” at which sensory integration is most efficient. Importantly, this procedure eliminates the impact of the variable of noninterest (main effect of alexithymia in auditory emotion processing) on the variable of interest (integration efficiency in the HA and LA groups). Repeated‐measures ANOVA of normalized performance was conducted with alexithymia groups as the independent variable. To assess the difference between HA and LA in intensity rating of each stimulus, repeated‐measures ANOVA was conducted.

ERP statistics

Trials of ERP data were included for statistical analysis only if the behavioral response was correct. The current study focused on three ERP components, the frontal‐central N1, the central P2, and the frontal‐central N2. We also focused on the P2 to test neural correlates of alexithymia because alexithymia has been demonstrated to impact emotion recognition from both auditory and visual emotional stimuli, which can be indexed by P2 (Franz et al., 2004). Visual detection on the grand‐averaged waveform in the specified channel and the topography across all conditions confirmed the peak latency of each component. The time window of each component was in line with previous studies (Crowley & Colrain, 2004; Folstein & Van Petten, 2008; Mercado et al., 2006; Miller et al., 2011; Mittag et al., 2013; Wang et al., 2012). Specifically, the N1 was identified in a window of 100–150 ms over the frontal‐central electrode (FCz; Wang et al., 2012). The P2 was defined as the mean activity between 165 and 215 ms over the central electrode (Cz; Crowley & Colrain, 2004; Mercado et al., 2006; Miller et al., 2011). The N2 was measured in a time‐window of 250–350 ms over the frontal‐central electrode (FCz; Folstein & Van Petten, 2008; Mittag et al., 2013). The 2 (Con vs. Incon) by 2 (HA vs. LA) repeated‐measures ANOVA was conducted for each component of interest. Please note that the minimum numbers of trials in each condition is higher than 55. The number of bad channels is 0.48 ± 1.13 (mean ± SD), with ranges from 0 to 5.

Correlations among subscales of TAS‐20, P2 and corrected performance

To examine the correlations of sub‐dimensions of alexithymia with behavioral and electrophysiological differences between HA and LA (if difference is significant), we performed correlations among subscales of the TAS‐20 (DIF, DDF, and EOT), the significant ERPs, and the significant behavioral performance.

RESULTS

Note that anxiety, depression, and autism have been demonstrated to co‐occur with alexithymia (Hendryx et al., 1991; Li et al., 2015). In our sample, we indeed found significant correlations of TAS‐20 with BAI (r = .635, p < .001), BDI (r = .635, p < .001), and AQ (r = .644, p < .001). Thus, we performed additional ANCOVAs with BAI, BDI, and AQ scores as covariates to control for potential influences of these covariates on the results. Upon discovering null findings in our analyses, Bayes Factors (BFs) were computed for the respective comparisons to test whether the negative result supported the alternative or the null hypothesis. These were done using JASP (JASP Team, 2020) on both behavioral and ERP data. Descriptive data (mean ± SD) are given in Table 3.
TABLE 3

Behavioral and electrophysiological responses in each experimental condition of each group

HALA
ConInconConIncon
Correct performance1.46 ± 0.201.37 ± 0.191.65 ± 0.181.55 ± 0.17
Integrative effect0.03 ± 0.020.03 ± 0.02
Intensity rating5.85 ± 0.775.72 ± 1.42
N1−4.81 ± 1.99−4.59 ± 2.27−4.83 ± 2.12−4.56 ± 1.94
N2−3.73 ± 2.57−4.02 ± 2.66−3.65 ± 2.32−4.00 ± 2.64
P22.02 ± 2.072.04 ± 1.901.20 ± 1.721.13 ± 1.84
P2_indi2.33 ± 2.202.36 ± 2.021.58 ± 1.711.51 ± 1.84

Descriptive data are presented as mean (standard deviation).

Abbreviations: Con, congruent condition; HA, individuals with alexithymia; Incon, incongruent condition; LA, individuals without alexithymia; P2_indi, P2 amplitudes based on individual peak.

Behavioral and electrophysiological responses in each experimental condition of each group Descriptive data are presented as mean (standard deviation). Abbreviations: Con, congruent condition; HA, individuals with alexithymia; Incon, incongruent condition; LA, individuals without alexithymia; P2_indi, P2 amplitudes based on individual peak.

Behavioral results

For corrected performance, the 2 (Con vs. Incon) by 2 (HA vs. LA) repeated‐measures ANOVA revealed main effects of Congruency (F (1,46) = 90.849, p < .001,  = 0.664; Con > Incon; 95% Confidence interval [CI] of difference between Con and Incon [0.075, 0.115]) and Group (F (1,46) = 12.197, p = .001,  = 0.210; HA < LA; 95% CI of difference between HA and LA [−0.293, −0.079]). No significant interaction effect between Congruency and Group was found (F (1,46) = 0.726,  = 0.016). After normalization, the significant main effect of Group disappeared (F (1,46) = 0.115,  = 0.002; 95% CI of difference between HA and LA [−0.015, 0.011]). The independent‐sample BF t‐test provided substantial evidence for the null hypothesis of no difference between HA and LA regarding the integration effect (BF10 = 0.301; Dienes, 2014). Furthermore, ANCOVA revealed same patterns with ANOVA. The ANCOVA in corrected performance revealed main effects of Congruency (F (1,44) = 92.453, p < .001,  = 0.678; Con > Incon; 95% CI of difference between Con and Incon [0.075, 0.115]) and Group (F (1,44) = 4.325, p = .043,  = 0.089; HA < LA; 95% CI of difference between HA and LA [−0.329, −0.005]; Figure 3a). No significant interaction effect between Congruency and Group was found (F (1,44) = 0.002,  < 0.001). After normalization, the significant main effect of Group disappeared (F (1,44) = 0.206,  = 0.005; 95% CI of difference between HA and LA [−0.015, 0.023]). The independent‐sample BF t‐test provided substantial evidence for the null hypothesis of no difference between HA and LA regarding the integration effect (BF10 = 0.301; Figure 2a; Dienes, 2014).
FIGURE 3

Group differences (independent of congruency effect). (a) Corrected performance between HA and LA. (b) Time course at Cz electrode and the topographic map of the difference between HA and LA. (c) Mean amplitude values of P2 between HA and LA. (d) Mean amplitude values of P2 based on individual peak between HA and LA. (e‐g) Correlations between DIF and P2, between DDF and P2, and between P2 and corrected performance. Con, congruent condition; Incon, incongruent condition; DDF, difficulty describing feelings; DIF, difficulty identifying feelings; HA, individuals with alexithymia; LA, individuals without alexithymia. **p < .01; *p < .05

FIGURE 2

Group differences in eMSI effect. (a) Integrative performance between HA and LA. (b) Intensity rating between HA and LA. (c) Time course at FCz electrode. (d & e) Topographic maps of each condition in N1 and N2. (f & g) Mean amplitude values of N1 and N2 in Con versus Incon between HA and LA. Con, congruent condition; HA, individuals with alexithymia; Incon, incongruent condition; LA, individuals without alexithymia

Group differences in eMSI effect. (a) Integrative performance between HA and LA. (b) Intensity rating between HA and LA. (c) Time course at FCz electrode. (d & e) Topographic maps of each condition in N1 and N2. (f & g) Mean amplitude values of N1 and N2 in Con versus Incon between HA and LA. Con, congruent condition; HA, individuals with alexithymia; Incon, incongruent condition; LA, individuals without alexithymia For intensity ratings, no significant group main effect was found from ANOVA (F (1,46) = 0.164,  = 0.004; 95% CI of difference between HA and LA [−0.524, 0.788]). Similarly, for this nonsignificant effect, the BF was calculated to test whether the negative result supported the alternative or the null hypothesis. Again, substantial evidence for the null hypothesis of an absence of group differences for intensity ratings was provided by the independent‐sample BF t‐test (BF10 = 0.308; Dienes, 2014). The same pattern was replicated when adding covariates. ANCOVA in intensity ratings found no significant group main effect (F (1,44) = 0.939,  = 0.021; 95% CI of difference between HA and LA [−0.510, 1.454]). Similarly, for this nonsignificant effect, the BF was calculated to test whether the negative result supported the alternative or the null hypothesis. Again, substantial evidence for the null hypothesis of an absence of group differences for intensity ratings was provided by the independent‐sample BF t‐test (BF10 = 0.308; Figure 2b; Dienes, 2014).

ERP analyses

N1

For the N1, ANOVA showed a significant main effect of Congruency (F (1,46) = 5.244, p = .027,  = 0.102; 95% CI of difference between Con and Incon [−0.459, −0.030]). Specifically, Con elicited a more negative N1 than the Incon condition. No other significant effect was observed (main effect of Group: F (1,46) < 0.001,  < 0.001; 95% CI of difference between HA and LA [−1.206, 1.182]; interaction effect between Group and Congruency: F (1,46) = 0.053,  = 0.001). We then calculated the BF for this nonsignificant effect. The BF ANOVA test confirmed that there was no interaction effect between Group and Congruency for the N1 (BF10 = 0.306; Dienes, 2014). Importantly, we found same patterns in ANCOVA. Regarding the N1, there was a significant main effect of Congruency (F (1,44) = 4.970, p = .031,  = 0.101; 95% CI of difference between Con and Incon [−0.462, −0.023]). Specifically, Con elicited a more negative N1 than the Incon condition. No other significant effect was observed (main effect of Group: F (1,44) = 0.333,  = 0.008; 95% CI of difference between HA and LA [−1.281, 2.309]; interaction effect between Group and Congruency: F (1,44) = 0.014,  < 0.001). We then calculated the BF for this nonsignificant effect. The BF ANCOVA test confirmed that there was no interaction effect between Group and Congruency for the N1 (BF10 = 0.253; Figure 2f; Dienes, 2014).

P2

With regard to the P2, we did not observe a significant main Group effect from ANOVA (F (1,46) = 2.553, p = .117,  = 0.053; 95% CI of difference between HA and LA [−0.224, 1.947]). No other significant effect was found (main effect of Congruency: F (1,46) = 0.066,  = 0.001; 95% CI of difference between Con and Incon [−0.156, 0.202]; interaction effect between Group and Congruency: F (1,46) = 0.265,  = 0.006). We observed a significant main Group effect from ANCOVA (F (1,44) = 7.974, p = .007,  = 0.153; HA > LA; 95% CI of difference between HA and LA [0.615, 3.683]; Figure 3c). No other significant effect was found [main effect of Congruency: F (1,44) = 0.036,  = 0.001; 95% CI of difference between Con and Incon (−0.160, 0.194); interaction effect between Group and Congruency: F (1,44) = 0.617,  = 0.014]. To eliminate the impact of the previous negative‐going (N1) peak on P2, we also computed the average waveform (50 ms) around the individual peak to validate our result. A similar pattern was observed: a significant main effect of Group (F (1,44) = 5.775, p = .021,  = 0.116; HA > LA; 95% CI of difference between HA and LA [0.311, 3.540]; Figure 3d) and no other significant effect (main effect of Congruency: F (1,44) = 0.023,  = 0.001; 95% CI of difference between Con and Incon [−0.167, 0.193]; interaction effect between Group and Congruency: F (1,44) = 0.835,  = 0.019). The significant effect of alexithymia on P2 disappeared when not controlling for anxiety, depression, and autism. It indicates that the absence of the significant group effect on P2 might be contaminated by these potential variables. Group differences (independent of congruency effect). (a) Corrected performance between HA and LA. (b) Time course at Cz electrode and the topographic map of the difference between HA and LA. (c) Mean amplitude values of P2 between HA and LA. (d) Mean amplitude values of P2 based on individual peak between HA and LA. (e‐g) Correlations between DIF and P2, between DDF and P2, and between P2 and corrected performance. Con, congruent condition; Incon, incongruent condition; DDF, difficulty describing feelings; DIF, difficulty identifying feelings; HA, individuals with alexithymia; LA, individuals without alexithymia. **p < .01; *p < .05

N2

For the N2, ANOVA revealed a significant main effect of Congruency (F (1,46) = 6.715, p = .013,  = 0.127; 95% CI of difference between Con and Incon [0.071, 0.569]), whereas no significant main effect of Group or interaction effect between Congruency and Group was found (main effect of Group: F (1,46) = 0.005,  < 0.001; 95% CI of difference between HA and LA [−1.514, 1.415]; interaction effect between Congruency and Group: F (1,46) = 0.072,  = 0.002). Specifically, Con evoked a less negative N2 than Incon. Once again, we conducted Bayesian statistics for the interaction effect between Congruency and Group to test whether this negative result supported the alternative or null hypothesis. The BF ANOVA test showed that there was no interaction effect between Group and Congruency for the N2 (BF10 = 0.287; Dienes, 2014). Importantly, ANCOVA replicated the results from ANOVA. A significant main effect of Congruency was found from ANCOVA (F (1,44) = 6.613, p = .014,  = 0.131; 95% CI of difference between Con and Incon [0.070, 0.576]), whereas no significant main effect of Group or interaction effect between Congruency and Group were found (main effect of Group: F (1,44) = 0.006,  < 0.001; 95% CI of difference between HA and LA [−2.128, 2.301]; interaction effect between Congruency and Group: F (1,44) = 0.296,  = 0.007). Specifically, Con evoked a less negative N2 than Incon. Once again, we conducted Bayesian statistics for the interaction effect between Congruency and Group to test whether this negative result supported the alternative or null hypothesis. The BF ANCOVA test showed that there was no interaction effect between Group and Congruency for the N2 (BF10 = 0.274; Figure 2g; Dienes, 2014). Notably, the significant integrative effects (Con vs. Incon) at both behavioral and neural levels indicate a successful manipulation in the current study, consistent with previous studies comparing Con and Incon audiovisual conditions (Campanella & Belin, 2007; Klasen et al., 2012). Given that the significant group difference in P2 was observed only when adding covariates, we conducted partial correlations among subscales of the TAS‐20 (difficulty identifying feelings [DIF], difficulty describing feelings [DDF], and externally oriented thinking [EOT]), the P2, and corrected behavioral performance (with BAI, BDI and AQ as covariates). These post hoc analyses revealed a positive correlation between DIF and P2 (r (43) = .505, p = .004; Figure 3e), between DDF and P2 (r (43) = .404, p = .006; Figure 3f), and a negative correlation between P2 and corrected performance (r (43) = −.320, p = .032; Figure 3g). Please note that analyses without covariates were as planned initially (e.g., ANOVA), whereas analyses with covariates were conducted post hoc, including ANCOVAs, partial correlations, and Bayesian statistics.

DISCUSSION

The current study examined whether alexithymia is related to deficits in the integration of multisensory emotional information. Bayesian statistics showed no group differences in classical integration‐related ERP components (N1 and N2). These results do not support the assumption of abnormal emotional multisensory integration (eMSI; hypo‐ and hyper integration) in relation to alexithymia. Instead, HA appear to show normal behaviors and neuropsychological responses during eMSI. However, the findings of worse performance and higher P2 amplitudes in HA versus LA, independent of emotional congruency, do support the notion that HA have deficits processing auditory emotional information. In light of two stages of eMSI (Klasen et al., 2012; Wang et al., 2012), we observed that Con elicited an increased N1 and a decreased N2 compared with Incon. The larger N1 may reflect amplification of attention: that is attention triggered by stimulation of one sensory modality can be amplified by stimulation of another sensory modality (Ho et al., 2013; Klasen et al., 2012; Luck et al., 2000). The N1 effect has also been demonstrated in blind‐sight patients who cannot consciously perceive a visual stimulus (but are able to react to it, implying that stimulus processing took place outside of conscious awareness), reflecting autonomous eMSI in the early stage (de Gelder et al., 2002). Therefore, the observed larger N1 in Con versus Incon conditions may indicate that attention to auditory stimuli, unconsciously amplified by implicit visual (i.e., facial) processing, facilitates early processing of eMSI in the congruent context, resulting in better performance (i.e., multisensory facilitation). The increased N2 has been repeatedly reported for processing incongruent compared with congruent stimuli in the Stroop task (Folstein & Van Petten, 2008), reflecting conflict processing (Mittag et al., 2013; Zinchenko et al., 2017). The larger N2 evoked in the current Incon versus Con conditions may indicate inhibition of conflicting stimuli (multisensory inhibition). However, the current results failed to support the hypothesis of either hypo‐ or hyper‐integration in relation to alexithymia. Instead, our evidence supports normal behavioral and neuropsychological responses of eMSI in HA, resulting from Bayesian analysis. Bayes Factor has long been considered as a central tool to enable conclusions that could support a null hypothesis over a specific hypothesis (Gallistel, 2009). To counteract the problem of publication bias, reporting a lack of differences is as important as are claims of “significant effects” (Gallistel, 2009; Tendeiro & Kiers, 2019). Indeed, being able to provide substantial evidence for the absence of differences plays a crucial role in the development of theories in neuroscience and cognitive science (Dienes, 2014). Based on two stages of eMSI (Klasen et al., 2012; Wang et al., 2012), observed absence of N1 difference in Con versus Incon between HA and LA suggests that individuals with high alexithymia levels are able to successfully integrate an attended emotional voice with an unattended facial expression, thus implying intact emotional multisensory facilitation. It goes without saying, however, that strong conclusions regarding intact eMSI in alexithymia can only be made with more studies (using different tasks) supporting our finding. No N2 difference in Con versus Incon between HA and LA were observed, and integrative behavioral performance was also unaffected. This may reflect successful inhibition of the incongruent emotional information in individuals with high alexithymia levels. However, previous studies did report deficits in inhibiting conflict processing (Zhang et al., 2011), especially in emotional conflicting inhibition (Zhang et al., 2012). One possible explanation is the task difference. Although previous studies adopted ANT and Go/Nogo tasks with unisensory stimuli (Zhang et al., 2011, 2012), we, instead, used a Stroop‐like task with multisensory (visual‐auditory) stimuli. Consequently, the absence of alexithymia‐related differences in conflict inhibition may be task‐specific or modality‐specific. Another possible reason could be that the TAS‐20 may not be sensitive to alexithymia‐related differences in conflict inhibition. Koven and Thomas (2010) used principal components analysis on the TAS‐20 and the Mood Awareness Scale to decompose alexithymia into two facets: emotional clarity and emotional monitoring. They found that only the emotional clarity component correlated with behavioral inhibition (Koven & Thomas, 2010). Thus, it is possible that the use of the TAS‐20 as the only alexithymia measure may have prevented us from detecting conflict inhibition differences in relation to alexithymia. Future studies should employ multiple alexithymia assessment tools to identify the nature of differences in behavioral inhibition and executive control in relation to alexithymia. It should be noted that a previous study (with western participants) found that alexithymia was associated with deficits in emotional integration during an audiovisual affective priming task with a stimulus onset asynchrony (SOA) of 200 ms (the presentation of audio stimuli occurred 200 ms after visual stimuli; Goerlich et al., 2011). One potential explanation for the different results of our current and the previous study might be the asynchrony between auditory and visual stimuli in that study. Indeed, time delay has been reported to be essential for congruency effects (Föcker & Röder, 2019). It has been demonstrated that the maximum eMSI appeared when the temporal difference between auditory and visual stimuli was less than 100 ms, and integrative performance decreased as the SOA increased (Ren et al., 2017). It is conceivable that the 200 ms SOA used in the study by Goerlich and colleagues (2011) might be a limiting factor for the integration of emotions from multisensory stimulus modalities, revealing alexithymia‐related differences. It would be interesting to test the effect of asynchrony of stimulus presentation in future studies on eMSI in relation to alexithymia. Another reason might be attributed to cultural differences between western participants in the previous study (Goerlich et al., 2011) and eastern participants in our present study. It has been suggested that culture plays an important role in emotion processing in relation to alexithymia (Zhu et al., 2007). Indeed, experience and expression of emotions, defining features of alexithymia, are significantly different between eastern and western cultures (Le et al., 2002). Therefore, it is possible that cultural differences contributed to the absence of eMSI differences in relation to alexithymia in the present study. One may argue that the absence of the behavioral performance difference between Con and Incon in HA versus LA may be due to the similarity of the emotions. Indeed, the angry and sad stimuli we used are similar in that both convey negative information. However, despite sharing some characteristics, different emotion expressions are thought to serve different functions (Horstmann, 2003). For example, expressions of anger (an approach emotion) are thought to cue the perceiver to modulate his or her behavior, whereas expressions of sadness (a withdrawal emotion) are thought to serve an emotion‐sharing function (Garcia & Tully, 2020). We also found significant congruency effects (Con vs. Incon) both at behavioral and neural (N1 and N2) levels, further indicating that angry and sad emotions were perceived as different stimuli. Thus, the absence of the behavioral performance difference between Con and Incon in HA versus LA was likely not due to the similarity of the emotions. Instead, the absence of effects, in combination with Bayesian analysis, support normal behavioral and neuropsychological responses of eMSI in HA. We also investigated abnormal emotion recognition from the voice and its corresponding neuropsychological mechanism. It is well‐known that alexithymia impacts emotion recognition from both audio and visual emotional stimuli (Franz et al., 2004; Goerlich‐Dobre et al., 2014; Kano et al., 2003; Lee et al., 2011; Van der Velde et al., 2013). For example, HA exhibited a larger P2 during the processing of negative pictures than LA (Franz et al., 2004). The P2 component is linked to the early processing of categorizing different stimuli, especially negative emotions (Crowley & Colrain, 2004; Zhu et al., 2015). The P2 component has also been implicated in the involvement of attentional resources (Mercado et al., 2006; Miller et al., 2011). Indeed, it has been reported that HA need more attentional resources to process emotional information (Franz et al., 2004; Van der Velde et al., 2013). Therefore, worse performance and increased P2 in HA versus LA in the current study may reflect that individuals with high alexithymia levels called upon attentional resources to a stronger extent to categorize vocal emotions. This deficit was further corroborated by correlations between the DIF and DDF subscales of alexithymia and P2, and between P2 and corrected behavioral performance. Recent reviews on alexithymia have identified DIF and—to a somewhat lesser extent—DDF as the most important TAS‐20 contributors to deficits in perceiving and recognizing negative emotions during the early, automatic (unconscious) processing level (Donges & Suslow, 2017; Goerlich, 2018). The present finding that P2 was correlated with DIF and DDF provides further evidence for altered emotion processing at early, automatic stages in relation to alexithymia and suggests that individuals with difficulty identifying and describing their feelings need more attentional resources to recognize negative emotions from the voice. However, these conclusions should be made cautiously given poor Cronbach's alpha coefficient of DDF. Several limitations of the present study should be mentioned. First, we adopted the Stroop‐like visual‐auditory task with the judgement of auditory emotion only (rather than also including the visual judgement task). Although stronger eMSI effects were found in the voice task than in the visual task in Chinese participants (Wang et al., 2012), whether our findings can be extended to the visual task remains to be tested. Second, static facial pictures were used, whereas in real life, we recognize emotion from dynamic visual information, for example facial movements (Collignon et al., 2008; Jessen & Kotz, 2011; Kokinous et al., 2015). Future studies should aim to explore eMSI in a dynamic context for higher ecological validity. Third, we only adopted negative stimuli in the current study. It is thus not clear whether the current conclusions would extrapolate to neutral or positive emotions during eMSI. Although differences in processing negative emotions seem to be at the core of alexithymia ((Kano et al., 2003; Lee et al., 2011), both positive and negative emotions are associated with processing differences in alexithymia (Van der Velde et al., 2013). Future studies would thus benefit from investigating positive, neutral, and negative emotion integration in relation to alexithymia. Next, we did not measure the sound intensity participants received, despite the approximate equal intensity of all vocal stimuli. Participants also reported that they can hear the voice clearly without any uncomfortable state. However, it has been demonstrated that intensity affects eMSI (Pan et al., ). Future studies might thus benefit from assessing perceived sound intensities. Finally, although the voice stimuli selected from the dataset were matched in emotional properties between angry and sad stimuli, we did observe stronger intensity and more negative valence being perceived from angry than sad utterances. Importantly, however, the absence of significant group differences for these parameters suggests that the current result of intact eMSI in alexithymia was not confounded by basic stimulus characteristics. To conclude, both behavioral and electrophysiological data indicated intact emotional multisensory integration in relation to alexithymia. These results do not support alexithymia‐related differences in the integration of emotional information. However, we provide evidence (replicating previous research) for an emotion recognition deficit from the voice and show that this deficit is driven by the DIF and DDF factors of the TAS‐20. Our work provides valuable insights into the effects of alexithymia on the neuropsychological mechanisms of eMSI and the processing of emotions expressed by the voice.

CONFLICT OF INTEREST

The authors have indicated they have no potential conflicts of interest to disclose.

AUTHOR CONTRIBUTION

Zhihao Wang: Conceptualization; Data curation; Formal analysis; Investigation; Methodology; Project administration; Visualization; Writing‐original draft; Writing‐review & editing. Mai Chen: Conceptualization; Investigation; Methodology; Project administration; Visualization; Writing‐original draft. Katharina S. Goerlich: Conceptualization; Supervision; Writing‐review & editing. André Aleman: Conceptualization; Supervision; Writing‐review & editing. Pengfei Xu: Conceptualization; Funding acquisition; Supervision; Writing‐review & editing. Yuejia Luo: Conceptualization; Funding acquisition; Supervision; Writing‐review & editing. Table S1 Click here for additional data file.
  57 in total

1.  Culture and alexithymia: mean levels, correlates, and the role of parental socialization of emotions.

Authors:  Huynh-Nhu Le; Howard Berenbaum; Chitra Raghavan
Journal:  Emotion       Date:  2002-12

2.  Multisensory integration, perception and ecological validity.

Authors:  Beatrice De Gelder; Paul Bertelson
Journal:  Trends Cogn Sci       Date:  2003-10       Impact factor: 20.229

3.  Visual event-related potentials in subjects with alexithymia: modified processing of emotional aversive information?

Authors:  Matthias Franz; Ralf Schaefer; Christine Schneider; Wolfgang Sitte; Jessica Bachor
Journal:  Am J Psychiatry       Date:  2004-04       Impact factor: 18.112

Review 4.  Integrating face and voice in person perception.

Authors:  Salvatore Campanella; Pascal Belin
Journal:  Trends Cogn Sci       Date:  2007-11-09       Impact factor: 20.229

Review 5.  Influence of cognitive control and mismatch on the N2 component of the ERP: a review.

Authors:  Jonathan R Folstein; Cyma Van Petten
Journal:  Psychophysiology       Date:  2007-09-10       Impact factor: 4.016

6.  The temporal dynamics of processing emotions from vocal, facial, and bodily expressions.

Authors:  S Jessen; S A Kotz
Journal:  Neuroimage       Date:  2011-06-22       Impact factor: 6.556

7.  The integration of facial and vocal cues during emotional change perception: EEG markers.

Authors:  Xuhai Chen; Zhihui Pan; Ping Wang; Xiaohong Yang; Peng Liu; Xuqun You; Jiajin Yuan
Journal:  Soc Cogn Affect Neurosci       Date:  2015-06-30       Impact factor: 3.436

8.  Dimensions of alexithymia and their relationships to anxiety and depression.

Authors:  M S Hendryx; M G Haviland; D G Shaw
Journal:  J Pers Assess       Date:  1991-04

9.  The rapid distraction of attentional resources toward the source of incongruent stimulus input during multisensory conflict.

Authors:  Sarah E Donohue; Alexandra E Todisco; Marty G Woldorff
Journal:  J Cogn Neurosci       Date:  2012-12-18       Impact factor: 3.225

10.  The Multifaceted Nature of Alexithymia - A Neuroscientific Perspective.

Authors:  Katharina S Goerlich
Journal:  Front Psychol       Date:  2018-08-29
View more
  2 in total

1.  Social-specific impairment of negative emotion perception in alexithymia.

Authors:  Zhihao Wang; Katharina S Goerlich; Yue-Jia Luo; Pengfei Xu; André Aleman
Journal:  Soc Cogn Affect Neurosci       Date:  2022-04-01       Impact factor: 3.436

2.  Deficient auditory emotion processing but intact emotional multisensory integration in alexithymia.

Authors:  Zhihao Wang; Mai Chen; Katharina S Goerlich; André Aleman; Pengfei Xu; Yuejia Luo
Journal:  Psychophysiology       Date:  2021-03-20       Impact factor: 4.348

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.