Literature DB >> 34975623

The Borderline Bias in Explicit Emotion Interpretation.

Sylwia Hyniewska1, Joanna Dąbrowska2, Iwona Makowska3, Kamila Jankowiak-Siuda4, Krystyna Rymarczyk4.   

Abstract

Atypical emotion interpretation has been widely reported in individuals with borderline personality disorder (iBPD); however, empirical studies reported mixed results so far. We suggest that discrepancies in observations of emotion interpretation by iBPD can be explained by biases related to their fear of rejection and abandonment, i.e., the three moral emotions of anger, disgust, and contempt. In this study, we hypothesized that iBPD would show a higher tendency to correctly interpret these three displays of social rejection and attribute more negative valence. A total of 28 inpatient iBPDs and 28 healthy controls were asked to judge static and dynamic facial expressions in terms of emotions, valence, and self-reported arousal evoked by the observed faces. Our results partially confirmed our expectations. The iBPD correctly interpreted the three unambiguous moral emotions. Contempt, a complex emotion with a difficulty in recognizing facial expressions, was recognized better by iBPD than by healthy controls. All negative emotions were judged more negatively by iBPD than by controls, but no difference was observed in the neutral or positive emotion. Alexithymia and anxiety trait and state levels were controlled in all analyses.
Copyright © 2021 Hyniewska, Dąbrowska, Makowska, Jankowiak-Siuda and Rymarczyk.

Entities:  

Keywords:  borderline personality; emotion bias; emotion perception; face interpretation; nonverbal communication

Year:  2021        PMID: 34975623      PMCID: PMC8715824          DOI: 10.3389/fpsyg.2021.733742

Source DB:  PubMed          Journal:  Front Psychol        ISSN: 1664-1078


Introduction

Adaptive emotion interpretation is fundamental for healthy human interactions and the mental health of individuals. Atypical appraisal of emotional cues of others could be related to traits, such as anger, anxiety, and alexithymia (Schlegel et al., 2017; Kiliç et al., 2020), and are characteristics of various mental disorders, such as borderline personality disorder (BPD) (Domes et al., 2009, 2011; De Panfilis et al., 2015). Interestingly, research into emotion perception in individuals with borderline personality disorder (iBPD) reported heterogeneous results, with different studies suggesting deficits in emotion understanding, generalized negative biases, or, in some cases, even high sensitivity and more accurate labeling of subtle emotions. Studies highlighting sensitivity to emotional signals, i.e., low threshold for the naming of emotional stimuli, showed, for example, that iBPD were able to correctly classify emotions at a lower intensity level of facial expression compared with healthy individuals (Lynch et al., 2006). Higher emotional reactivity in iBPD compared with controls was also reported as greater amygdala activation to emotional and neutral faces (Donegan et al., 2003), as well as to aversive stimuli in general (Herpertz et al., 2001). However, in several studies, neither increased psychophysiological responses, e.g., no greater potentiation of the startle response to negative pictures (Herpertz et al., 1999, 2000), nor any increased facial mimicry to facial expressions of emotions was observed (Matzke et al., 2014) in iBPD compared with controls. The authors of the latter study observed, however, a general tendency in iBPD to react with augmented activation of the corrugator supercilii muscle, i.e., frowning, to all displays of negative expressions. The authors concluded that, rather than heightened affective empathy in iBPD, a potential negativity bias could explain the diverse emotion interpretation deficits reported in the literature (Matzke et al., 2014). Moreover, visual search tasks where iBPD had to spot schematic happy and angry faces among neutral ones did not show any higher performance to angry stimuli compared with healthy participants either (Hagenhoff et al., 2013). Although only one negatively valenced emotion was presented, the authors suggested that these visual search results most probably can be explained by a lack of bias to negative stimuli. Different studies also showed more general deficits in the interpretation of emotional displays. Deficits were observed in the naming of surprise when iBPD watched morphs from neutral to basic emotion displays and, after having reported the face to be emotional, were asked to attribute one of the six basic emotions (Domes et al., 2008). However, the authors did not provide any explanation of confusions or eventual biases observed in their study. Another study on morphs showed biases in the attribution of anger, with iBPD being more likely to respond “anger,” when anger and disgust faces were blended 50%/50% or anger and happiness faces were blended 40%/60% (Domes et al., 2008). Other teams reported deficits in the interpretation of negatively valenced emotional displays (Levine et al., 1997; Wagner and Linehan, 1999; Bland et al., 2004), which is sometimes interpreted as contradicting the existence of an increased vigilance to social threat stimuli as postulated by Linehan (1993). Some questions were raised, however, regarding the stimuli used, e.g., Bland et al. (2004) reported changes in anger, sadness, and disgust, but in each case, only one picture of the three presented for each emotion led to interpretation differences in iBPD compared with healthy participants, and the authors presented no confusion matrix. Another study showed a higher tendency to attribute fear to emotional displays and to neutral faces, which led to a higher correct attribution of fear in iBPD compared with healthy participants, as well as to a high number of false alarms when appraising neutral faces (Wagner and Linehan, 1999). Although only fear showed these results, the authors interpreted them as reflecting a negativity bias in iBPD when appraising social cues. Different teams tried to explain the inconsistencies through the prism of comorbidities at play in BPD, particularly elevated alexithymia (Domes et al., 2011; Kiliç et al., 2020), or anxiety, which is reported at high levels in this population (Domes et al., 2008). Another approach to elucidating emotion interpretation skills is to look at the characteristics of emotional stimuli encountered in naturalistic settings, i.e., before all the dynamic modality. Previous studies that have attempted to investigate the question of dynamic facial expression being easier to interpret have yielded inconsistent findings (for a review, see Kätsyri, 2006; Fiorentini and Viviani, 2011; Alves, 2013; Krumhuber et al., 2013; Rymarczyk et al., 2016). Interestingly, however, clinical and neuropsychological conditions have been shown to influence the extent to which dynamic displays lead to processing benefits (Ambadar et al., 2005; Torro-Alves et al., 2016; Bala et al., 2018; Żurowska et al., 2018). Thus, individuals with major depression have been observed to present atypical emotion interpretation patterns depending on whether they watched static or dynamic facial expressions of emotions, namely, greater accuracy in labeling static sadness and angry faces and less accuracy in labeling dynamic happiness faces (Bomfim et al., 2019). Patients with brain lesions in the mesial temporal zone have shown lower performance in interpreting social information from movement compared with healthy individuals (Bala et al., 2018). Patients during and after benzodiazepine detoxification recognized dynamic facial displays better than static displays, whereas no similar emotional recognition enhancement for the dynamic modality was observed in the healthy controls (Żurowska et al., 2018). Although numerous empirical scientists support the importance of testing both static and dynamic stimuli to improve the understanding of processes at play in emotion interpretation, whether in healthy or clinical populations, no study so far has investigated this aspect in the population with BPD. Therefore, we decided to investigate how the presentation modality of stimuli, i.e., dynamic vs. static facial displays, affects emotion interpretation in BPD while controlling for alexithymia and anxiety levels. Given that fear of rejection and abandonment might be a defining feature of iBPD (Gunderson, 2008; De Panfilis et al., 2015), presenting iBPD with stimuli relating to these specific emotions would be invaluable for the understanding of atypical emotion perception in this disorder. In fact, most of the former studies in iBPD investigated some or all of the basic emotions as well as neutral faces and, to the best of our knowledge, have never included any expressions of contempt. Although we do not have any expectations regarding how the presentation modality of stimuli might influence emotion interpretation in iBPD, we predict that facial expressions of emotions associated with social threat might exhibit a high unbiased hit rate, mostly anger, disgust, and contempt, the three being the so-called moral emotions (Hutcherson and Gross, 2011). Given the mentioned fear of rejection and abandonment, we expected higher arousal and higher negativity to be attributed to these three emotion stimuli by iBPD compared with healthy controls.

Materials and Methods

Participants

A total of 28 inpatients, meeting the diagnosis of BPD according to the DSM-5 criteria, and 28 healthy individuals participated in this study. The sample size of this study was determined regarding an a priori power analysis, for which we used the G*Power software application (version 3.1.9.2, Heinrich-Heine-Universität Düsseldorf, Düsseldorf, Germany; Faul et al., 2007). According to the calculations, 17 samples per group were required to accomplish an ANOVA with an α of 0.05, a power (1 − β) of 0.80, and an effect size f of 0.40 based on the data provided in similar designs comparing emotion recognition in patients with BPD and healthy controls (Fenske et al., 2015; Lowyck et al., 2016; Kiliç et al., 2020). These seemed to be the only similar studies to report sufficient prior information to run power analyses. Participants from both groups were of similar age (t = − 1.77, p = 0.083, d = − 0.473). All patients (24 females and 4 males), aged 19–57 years (M = 26.86; SD = 8.78, SE = 1.66), were referred to the study by psychiatrists from the 24/7 Department of Neurosis, Personality Disorders and Eating Disorders at the Institute of Psychiatry and Neurology in Warsaw, Poland. The higher ratio of females to males is a reflection of this specific patient population and in accordance with the DSM-5, which records a higher prevalence of women among those being diagnosed clinically with BPD. For the control group, individuals were involved from the general population in Warsaw (23 females and 5 males), aged 18–54 years (M = 30.96; SD = 8.60, SE = 1.62), through online advertisements and mailing groups. None from the control group had any current or past history of mental health conditions, nor any excessive consumption of alcohol or recreational drugs as verified through self-reports.

Procedure

Videos and static photographs of forward-facing actors (two women and two men) were presented in a semi-random sequence. Each actor displayed nine emotional faces (i.e., joy, sadness, anger, fear, disgust, surprise, embarrassment, contempt, and pride) and one neutral face in dynamic and static format from the Amsterdam Dynamic Facial Expression Set (ADFES; Van der Schalk et al., 2011). In the neutral ADFES dynamic condition, actors could be observed blinking, closing their eyes, or slightly changing the position of their heads. All stimuli were 576 pixels in height and 720 pixels in width, presented on a gray background. For our study, facial displays from four ADFES actors were selected (two males and two females), presenting each emotion once in a dynamic format and once in a static format. Both formats are accessible in a usable form directly from the ADFES dataset. All stimuli were unambiguous, and as per the wish of authors, the expressions were to be highly standardized: they were to be included in the dataset exclusively when following closely established atheoretical prototypes for each emotion display and have received very high recognition rates in healthy populations (see Van der Schalk et al., 2011). Each participant saw and evaluated 80 stimuli in total (10 facial displays × 4 actors × 2 modalities). The experimental session was preceded by an explanatory session with two faces to be judged in order for the participants to become acquainted with the experimental procedure. Each face stimulus was preceded by a fixation dot (5 s duration) presented in the place where the face of the stimulus actor would follow (Figure 1). Each stimulus was presented for 5 s independently of whether in a photograph or video format. Each facial stimulus was followed automatically by three evaluative questions.
FIGURE 1

Experimental procedure.

Experimental procedure. First, participants were asked to use the pictorial Self-Assessment Manikin (SAM; Bradley and Lang, 1994) to judge whether the presented emotion is more positive or more negative (valence). Second, they were asked to use SAM to judge to what degree the presented emotion triggers a reaction in them, in other words, to report their arousal level. Finally, the emotion attributions of participants were recorded through a multiple-choice task, where participants had to choose 1 label out of 11 to name the emotion of the stimuli they observed. All participants answered the self-report Toronto Alexithymia Scale (TAS-20). This questionnaire measures 20 items with a five-point Likert scale, with a focus on identifying feelings, describing feelings, and externally oriented thinking (Bagby et al., 1994a,b). The two groups differed in terms of total TAS scores (t = 6.902, p < 0.001, d = 1.864), with higher alexithymia in iBPD (M = 69.00, SD = 8.94, SE = 1.660) than in controls (M = 52.89, SD = 8.31, SE = 1.63), with 24 and 4, respectively, being classified as alexithymic given the following interpretation: ≤51, no alexithymia; 52–60, borderline alexithymia; and ≥61, alexithymia. Levels of anxiety were measured in both populations using the State-Trait Anxiety Inventory (STAI; Spielberger, 2010). STAI state was higher for iBPD (M = 55.66, SD = 10.955, SE = 2.034) than for controls (M = 34.52, SD = 9.323, SE = 1.865) and so was the STAI trait (M = 58.72, SD = 10.42, SE = 1.94 and M = 41.76, SD = 9.02, SE = 1.80, respectively).

Results

Label Attribution

To investigate how facial displays were perceived in terms of emotion label attributions, unbiased hit rates, confusion matrices, and factorial analyses were computed. The “unbiased hit rate” (H) was calculated as proposed by Wagner (1993) to account for response biases. H was calculated as the squared frequency of correct attributions for an emotion stimulus category divided by the product of the number of times the category was assessed and the overall frequency of this emotion label being attributed. Its value ranges from zero to one, one indicating that all stimuli of an emotion have been correctly identified and the respective emotion has never been falsely chosen for a different emotion. A repeated-measures ANOVA on unbiased hit rates and STAI trait, STAI state, and TAS as covariates showed a participant group and emotion interaction effect, as well as an emotion effect (see Table 1 and Figure 2). There was also a group effect. No effects of the modality of stimuli were observed (Figure 3).
TABLE 1

Repeated measures ANOVA on the number of unbiased hits.

CasesSum of squaresdfMean squareF p η2
Within subjects effects
Modality0.03510.0350.5150.4764.309e-4
Modality × group0.03110.0310.4610.5003.854e-4
Modality × fear_state0.01410.0140.2100.6491.758e-4
Modality × fear_trait0.22010.2203.2480.0770.003
Modality × TAS_total0.16910.1692.4960.1200.002
Residuals3.449510.068
Emotion1.58290.1762.2410.0190.020
Emotion × group1.53090.1702.1670.0230.019
Emotion × fear_state0.82990.0921.1740.3090.010
Emotion × fear_trait0.32490.0360.4590.9020.004
Emotion × TAS_total0.66390.0740.9390.4910.008
Residuals36.0084590.078
Modality × Emotion0.20190.0220.5110.8670.002
Modality × Emotion × group0.19490.0220.4920.8800.002
Modality × Emotion × fear_state0.21890.0240.5540.8340.003
Modality × Emotion × fear_trait0.09390.0100.2360.9890.001
Modality × Emotion × TAS0.30890.0340.7810.6340.004
Residuals20.0724590.044
Between subjects effects
Group1.15511.1554.8590.0320.014
Fear_state0.01910.0190.0780.7812.294e-4
Fear_trait0.07910.0790.3310.5679.747e-4
TAS_total1.49811.4986.3020.0150.019
Residuals12.126510.238

Type III sum of squares.

FIGURE 2

Unbiased hits for 2 groups (i.e., BPD, controls) and 10 emotions (i.e., anger, contempt, disgust, embarrassment, fear, joy, neutrality, pride, sadness, surprise).

FIGURE 3

Unbiased hit rates observed in the interpretation of dynamic (A) and static (B) stimuli by healthy controls and in patients with BPD in terms of emotional labels. Confidence interval 95%.

Repeated measures ANOVA on the number of unbiased hits. Type III sum of squares. Unbiased hits for 2 groups (i.e., BPD, controls) and 10 emotions (i.e., anger, contempt, disgust, embarrassment, fear, joy, neutrality, pride, sadness, surprise). Unbiased hit rates observed in the interpretation of dynamic (A) and static (B) stimuli by healthy controls and in patients with BPD in terms of emotional labels. Confidence interval 95%. Arousal judgments in the control and iBPD population. Confidence interval 95%. Contempt expressions were labeled as contempt more often by iBPD (H = 0.34 M = 2.11; SD = 1.26) vs. controls (Hu = 0.18, M = 2.11; SD = 1.26, p < 0.05). The misattribution profiles were slightly different (see Table 2), with the attribution of a surprise to the contempt expression significantly higher in controls (p < 0.05). Thus, controls misattributed the expression mostly to surprise (25%) and to the “none of the above” emotion category (22%). iBPD attributed the “none of the above” category most often (20%) than surprise (19%).
TABLE 2

Confusion matrix for (A) inpatients with BPD and (B) the control group.

Label attributions
AngerContemptDisgustEmbarrassmentFearJoyNeutralPrideSadnessSurpriseNone
(A) Confusion matrix for iBPD
StimuliAnger755411010725
Contempt14435004041919
Disgust18106700000122
Embarrassment031731031936
Fear1073690102133
Joy000009321002
Neutral251110730906
Pride0801018267004
Sadness224220108124
Surprise000190000862
Total10079878784112877111312752

(B) Confusion matrix for the control group
Controls
StimuliAnger7553100101022
Contempt23527003042422
Disgust2086700000022
Embarrassment011730031967
Fear2053690003152
Joy000109441000
Neutral240100780805
Pride0601019269002
Sadness133200008323
Surprise200110001930
Total10463829073114927311814744

Rounded percentages of label responses (%) attributed to each emotion stimulus category.

Confusion matrix for (A) inpatients with BPD and (B) the control group. Rounded percentages of label responses (%) attributed to each emotion stimulus category. The overattribution of surprise labels in controls went to contempt and fear: 25% of all contempt stimuli were labeled as surprise as well as 16% of all fear stimuli. In iBPD and controls, anger stimuli were most often mislabeled as sadness (7%; 9%), disgust as anger (18%; 20%), and contempt as surprise (19%; 25%). In iBPD and controls, the anger label was most often misattributed to disgust (18%; 20%), disgust was most often misattributed to fear (8%; 5%), while contempt to disgust (9%; 9%), pride (8%; 6%), and anger (5%; 5%).

Arousal

The repeated measures ANOVA (Table 3) showed differences between groups of participants in the arousal reports (Figure 4), with TAS, STAI trait, and STAI state as covariants. There was an emotion effect, an emotion × group interaction but no effect of modality nor any group × emotion × modality interaction.
TABLE 3

Repeated measures ANOVA on arousal levels and mean scores per category.

CasesSum of squaresdfMean squareF p η2
Arousal scores: repeated measures ANOVA
Within subjects effects
Dynamics0.52910.5292.8380.0982.352e-4
Dynamics × group0.39810.3982.1370.151.771e-4
Dynamics × fear_state0.1910.191.0190.3188.446e-5
Dynamics × fear_trait0.910.94.8270.0334.001e-4
Dynamics × TAS_total0.13310.1330.7130.4025.909e-5
Residuals9.506510.186
Emotion27.3849a3.043a3.312a<0.001a0.012
Emotion × group20.5429a2.282a2.484a0.009a0.009
Emotion × fear_state6.4219a0.713a0.777a0.638a0.003
Emotion × fear_trait13.6459a1.516a1.65a0.099a0.006
Emotion × TAS_total6.9519a0.772a0.841a0.579a0.003
Residuals421.694590.919
Dynamics × Emotion1.1899a0.132a0.581a0.813a5.287ea-4
Dynamics × Emotion × group1.9929a0.221a0.974a0.461a8.857e-4
Dynamics × Emotion × fear_state0.8389a0.093a0.41a0.93a3.729e-4
Dynamics × Emotion × fear_trait0.8589a0.095a0.419a0.925a3.814e-4
Dynamics × Emotion × TAS_total0.5779a0.064a0.282a0.979a2.567e-4
Residuals104.2934590.227
Between subjects effects
Cases
Group82.398182.3982.7480.1040.037
Fear_state1.86211.8620.0620.8048.281e-4
Fear_trait16.560116.5600.5520.4610.007
TAS_total0.73610.7360.0250.8763.271e-4
Residuals1529.1365129.983

Dynamics Emotion Group Mean SD N

Descriptives
DynamicNeutralBPD2.2141.26528
Control2.4291.07328
PrideBPD3.9381.18228
Control3.5121.41028
SadnessBPD3.9291.35928
Control3.6161.56928
SurpriseBPD3.1961.37728
Control2.9381.24128
AngerBPD4.0451.63428
Control3.1591.39328
ContemptBPD3.4461.43628
Control2.7591.25028
DisgustBPD4.0711.61128
Control3.0181.42228
EmbarrassmentBPD3.4201.32828
Control2.8841.14828
FearBPD3.9021.42528
Control3.3211.40928
JoyBPD4.1701.42128
Control3.9851.71828
StaticNeutralBPD2.6251.44928
Control2.4961.08828
PrideBPD3.5981.28628
Control3.3751.38928
SadnessBPD3.9381.35328
Control3.4381.43028
SurpriseBPD3.1881.43128
Control2.9821.33328
AngerBPD3.8131.59528
Control3.2021.39828
ContemptBPD3.2681.35028
Control2.8201.18128
DisgustBPD4.0891.64528
Control3.2681.45028
EmbarrassmentBPD3.4201.19828
Control2.7591.06628
FearBPD3.9291.67228
Control3.2861.42228
JoyBPD3.9461.62828
Control3.8571.83628

Type III sum of squares.

Type III sum of squares.

FIGURE 4

Arousal judgments in the control and iBPD population. Confidence interval 95%.

Repeated measures ANOVA on arousal levels and mean scores per category. Type III sum of squares. Type III sum of squares. Both populations reported the strongest arousal for joy (M = 4.05; SD = 1.67). In iBPD, this was followed by high arousal scores for disgust (M = 4.0; SD = 1.59), followed closely by anger (M = 3.90; SD = 1.9). In controls, arousal levels that followed those of joy were for pride (M = 3.71; SD = 1.75) and sadness (M = 3.29; SD = 1.56).

Valence

To check the differences between groups of participants in the emotional valence attribution, repeated-measures ANOVA was computed, using TAS, STAI trait, and STAI state as covariants. There was an emotion effect (p < 0.001) as well as a modality × emotion × group interaction (p = 0.002) (see Table 4 and Figure 5). The modality × group interaction did not reach significance (p = 0.051).
TABLE 4

Repeated measures ANOVA on valence scores and mean scores per category.

CasesSum of squaresdfMean squareF p η2
Valence scores: repeated measures ANOVA
Within subjects effects
Dynamics0.00210.0020.0160.9014.564e-6
Dynamics × group0.63110.6314.0090.0510.001
Dynamics × fear_state0.09310.0930.5920.4451.738e-4
Dynamics × fear_trait0.19910.1991.2670.2663.720e-4
Dynamics × TAS_total0.02610.0260.1640.6874.813e-5
Residuals8.024510.157
Emotion41.832a9a4.648a6.715a<0.001 a0.078
Emotion × group6.51a9a0.723a1.045a0.403a0.012
Emotion × fear_state1.729a9a0.192a0.277a0.981a0.003
Emotion × fear_trait2.21a9a0.246a0.355a0.956a0.004
Emotion × TAS_total2.379a9a0.264a0.382a0.944a0.004
Residuals317.6954590.692
Dynamics × Emotion1.366a9a0.152a1.017a0.425a0.003
Dynamics × Emotion × group2.988a9a0.332a2.225a0.02a0.006
Dynamics × Emotion × fear_state1.412a9a0.157a1.052a0.398a0.003
Dynamics × Emotion × fear_trait1.418a9a0.158a1.056a0.394a0.003
Dynamics × Emotion × TAS_total1.097a9a0.122a0.817a0.601a0.002
Residuals68.4764590.149
Between subjects effects
Cases
Group0.41210.4120.2800.5997.686e-4
Fear_state2.12612.1261.4440.2350.004
Fear_trait0.07110.0710.0480.8271.329e-4
TAS_total0.34110.3410.2320.6326.365e-4
Residuals75.060511.472

Dynamics Emotion Group Mean SD N

Descriptives
DynamicNeutralBPD3.0000.42528
Control3.0180.42528
PrideBPD4.7771.30128
Control5.0271.01928
SadnessBPD1.7680.46628
Control1.8660.56728
SurpriseBPD2.9020.48328
Control3.3130.46028
AngerBPD1.8480.56228
Control2.1960.52428
ContemptBPD2.4110.56228
Control2.7660.47928
DisgustBPD1.7140.43928
Control2.1400.56828
EmbarrassmentBPD2.2590.57928
Control2.7590.48328
FearBPD1.7140.51728
Control2.1790.53128
JoyBPD5.8570.93928
Control5.8870.97428
staticNeutralBPD2.8570.52028
Control2.9460.39928
PrideBPD4.5361.14028
Control4.9430.91728
SadnessBPD1.7680.58128
Control1.9200.50928
SurpriseBPD3.2860.63728
Control2.9550.78828
AngerBPD1.9290.60028
Control2.0740.53328
ContemptBPD2.6070.67228
Control2.8040.48328
DisgustBPD1.5630.47028
Control1.9730.54228
EmbarrassmentBPD2.5270.62928
Control2.7140.53528
FearBPD1.8480.49228
Control2.0890.48728
JoyBPD5.7950.85528
Control5.7201.15928

Type III sum of squares.

Type III sum of squares.

FIGURE 5

Valence judgments in the control and iBPD population and 10 emotions (i.e., anger, contempt, disgust, embarrassment, fear, joy, neutrality, pride, sadness, and surprise). (A) Valence: Dynamic facial display stimuli. (B) Valence: Static facial display stimuli.

Repeated measures ANOVA on valence scores and mean scores per category. Type III sum of squares. Type III sum of squares. Valence judgments in the control and iBPD population and 10 emotions (i.e., anger, contempt, disgust, embarrassment, fear, joy, neutrality, pride, sadness, and surprise). (A) Valence: Dynamic facial display stimuli. (B) Valence: Static facial display stimuli.

Discussion

When studying BPD, researchers have focused until now on the recognition of emotions, i.e., the attribution of an expected emotional label to a specific expression, and sensitivity to threat, i.e., interpretation of facial displays of anger or sometimes fear. The aim of this study was to change the focus to moral emotions (Hutcherson and Gross, 2011). These emotions, namely, anger, contempt, and disgust, are particularly relevant to iBPD given that a strong fear of rejection is a diagnostic feature of BPD (Goodman et al., 2014). In line with former studies describing comorbidities associated with BPD, we recorded and anxiety scores in all participants as their levels are known to influence emotion interpretation and define perception biases. As expected, we observed a more accurate interpretation of contempt in iBPD compared with controls. These results could be explained by these emotional displays being perceived by individuals as predictors of rejection and abandonment, therefore, having survival value for this population and necessitating correct interpretation. The high sensitivity to the displays of contempt may also be related to their vulnerability to harm from others as well as their generalized belief that other individuals are hostile (Beck and Freeman, 1990). As expected, dynamic presentations improved recognition of emotions in both populations, especially for the emotion of contempt, which is a complex emotion often poorly recognized (see former uses of the ADFES database in the general population: Van der Schalk et al., 2011; Wingenbach et al., 2016). To summarize, our results support that iBPD have good emotion recognition skills, on par with healthy individuals, except facial expressions of contempt, which are recognized more accurately by iBPD, most probably due to their survival value iBPD attributes to this social information. Confirming former studies reporting sensitivity to threat, iBPD, compared with healthy controls, attributed more negative valence to all presented emotions, more specifically, contempt, embarrassment, fear, disgust, and surprise. The iBPD reported higher arousal to observed stimuli than healthy individuals, which means a higher reactivity to facial expressions. The highest arousal was observed in dynamically presented emotions of anger and disgust. This is in line with numerous studies showing that iBPD tend to react more strongly to anger expressions and to judge strangers performing simple tasks such as entering a room and sitting down as more aggressive than healthy controls do (Barnow et al., 2009). This bias might be explained by the belief expressed by iBPD that all other individuals are malevolent (Pretzer, 1990; Arntz and Veen, 2001). Greater emotional reactivity seen in BDP controls has also been observed in their brain activity during the perception of negative social signals (van Zutphen et al., 2015), particularly greater activation of the amygdala during the perception of facial expressions of anger (Donegan et al., 2003; Minzenberg et al., 2007). The high arousal induced by disgust expressions in iBPD has previously been reported (Bland et al., 2004; Guitart-Masip et al., 2009; Jovev et al., 2011), and according to Rusch et al. (2011), not only the perceived disgust but also the experience of disgust toward the self may be a prominent emotion in BPD pathology, stronger than anxiety or anger. The general greater emotional sensitivity reflected in the high arousal levels reported by iBPD could be an expression of non-adaptive emotion regulation strategies, such as the less frequent redirection of attention from negative to more positive stimuli (Porter et al., 2016). BPD is an important psychological disorder characterized by emotional, interpersonal, and behavioral instability (American Psychiatric Association [APA], 2013). Following the biosocial theory proposed by Linehan (1993), iBPD can exhibit difficulties with the identification and correct reaction to relevant social stimuli, particularly to the facial expressions of emotions of others. This might be one of the factors contributing to difficulties in interpersonal functioning. Our results support previous studies suggesting a higher sensitivity to negative emotions (Lynch et al., 2006; Zanarini and Frankenburg, 2007) and are in line with reports of iBPD exhibiting higher emotional reactivity (Ebner-Priemer et al., 2007). This higher mood lability and emotional fluctuations in iBPD could explain some of the previously reported divergent results reported in different studies, e.g., neutral faces being colored by own emotional states of patients (e.g., Wagner and Linehan, 1999; Herr and Meier, 2020). In this study, we observed no higher attribution of negative valence to neutral faces, which is aligned with findings from several studies failing to find any significant differences in neutral facial expression recognition accuracy (Levine et al., 1997; Bland et al., 2004; Minzenberg et al., 2006a,b; Merkl et al., 2010; Mier et al., 2012; Hagenhoff et al., 2013). Given how important it is to use due to their better ecological validity and the added complexity seen in a naturalistic emotional context (e.g., Dziobek, 2012; Hyniewska et al., 2019) and the fact that previous studies on emotion perception in iBPD mostly relied on static stimuli, we introduced dynamic stimuli, and their comparison with static ones, to try to elucidate discrepancies observed in iBPD and emotion labeling of facial expressions. Our study did not show any particular advantage of modality for iBPD; however, more studies are necessary to understand the process at play and whether any conditions (e.g., for low-intensity emotions) require iBPD to rely on dynamic vs. static information in facial emotion interpretation tasks. Intensity of facial expressions is a factor that will need to be integrated in future studies, especially for complex emotions such as contempt (see Wingenbach et al., 2016). To comprehend the conflicting results regarding emotion interpretation in iBPD, the great heterogeneity of this clinical population needs to be acknowledged (Mitchell et al., 2014). Numerous factors could play a role in influencing the performance of patients, from the emotional and clinical state of the individual to comorbidities. Mitchell et al. (2014) suggested that emotional states and personal experience could be influencing emotion interpretation in iBPD. For example, following the mood-congruency hypothesis (Bower, 1981), patients with BPD who regularly experience negative states might be more skilled at processing and interpretating negative stimuli. Given data on neglect and childhood abuse often reported in iBPD (Wagner and Linehan, 1999), which are considered factors influencing the shaping of the borderline traits, further studies on emotion interpretation would need to record these characteristics for this population and healthy counterparts. This is on par with the quantifying of the degree of BPD-related dysfunctions, along with the study of BPD traits in non-clinical populations (Trull et al., 1997). The sensitivity to negativity and more generally emotional reactivity observed in patients with BPD is in line with Linehan’s biosocial model of emotion dysregulation. This dysregulation can be explained by an interplay of biological vulnerabilities and an early environment characterized by invalidation (Linehan, 1993). Particularly, sensitivity to injustice predisposes to emotional and cognitive biases and to intense reactions when expecting and perceiving potential rejection (Downey and Feldman, 1996) either as a victim, an observer, or a perpetrator (Schmitt et al., 2010). Furthermore, sensitivity to moral disgust predisposes to a stronger experience of disgust when confronted with moral norm violations (Tybur et al., 2009). These sensitivities could help explain cognitive and emotional biases observed in individuals with high BPD scores, who show a tendency to ascribe negative and hostile intent to observed social interactions and more generally ambiguous or explicit behavior of others (De Panfilis et al., 2015). Possibly, the development of atypical coping strategies, including emotional perception biases, might be functional attempts to deal with the fear of abandonment and emotional overstimulation, which in specific life circumstances might appear to be effective coping.

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding authors.

Ethics Statement

The studies involving human participants were reviewed and approved by Department of Biological and Behavioral Psychology, Behavioral Neuroscience Lab, SWPS University of Social Sciences and Humanities, Warsaw, Poland. The patients/participants provided their written informed consent to participate in this study. The animal study was reviewed and approved by Department of Biological and Behavioral Psychology, Behavioral Neuroscience Lab, SWPS University of Social Sciences and Humanities, Warsaw, Poland.

Author Contributions

SH and KR were responsible for the conceptual definition of the research. JD obtained the data. SH, JD, and KR analyzed the data. All authors wrote the manuscript.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
  52 in total

1.  Facial expression recognition ability among women with borderline personality disorder: implications for emotion regulation?

Authors:  A W Wagner; M M Linehan
Journal:  J Pers Disord       Date:  1999

2.  G*Power 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences.

Authors:  Franz Faul; Edgar Erdfelder; Albert-Georg Lang; Axel Buchner
Journal:  Behav Res Methods       Date:  2007-05

3.  Reduced sensitivity to emotional facial expressions in borderline personality disorder: effects of emotional valence and intensity.

Authors:  M Hagenhoff; N Franzen; L Gerstner; G Koppe; G Sammer; P Netter; B Gallhofer; S Lis
Journal:  J Pers Disord       Date:  2013-02

4.  Social-emotion recognition in borderline personality disorder.

Authors:  Michael J Minzenberg; John H Poole; Sophia Vinogradov
Journal:  Compr Psychiatry       Date:  2006-05-03       Impact factor: 3.735

5.  Heightened sensitivity to facial expressions of emotion in borderline personality disorder.

Authors:  Thomas R Lynch; M Zachary Rosenthal; David S Kosson; Jennifer S Cheavens; C W Lejuez; R J R Blair
Journal:  Emotion       Date:  2006-11

6.  When social inclusion is not enough: Implicit expectations of extreme inclusion in borderline personality disorder.

Authors:  Chiara De Panfilis; Paolo Riva; Emanuele Preti; Chiara Cabrino; Carlo Marchesi
Journal:  Personal Disord       Date:  2015-07-06

7.  Mood and memory.

Authors:  G H Bower
Journal:  Am Psychol       Date:  1981-02

8.  Psychophysiological ambulatory assessment of affective dysregulation in borderline personality disorder.

Authors:  Ulrich W Ebner-Priemer; Stacy S Welch; Paul Grossman; Thomas Reisch; Marsha M Linehan; Martin Bohus
Journal:  Psychiatry Res       Date:  2007-02-23       Impact factor: 3.222

9.  Implications of rejection sensitivity for intimate relationships.

Authors:  G Downey; S I Feldman
Journal:  J Pers Soc Psychol       Date:  1996-06

Review 10.  Emotion recognition in borderline personality disorder-a review of the literature.

Authors:  Gregor Domes; Lars Schulze; Sabine C Herpertz
Journal:  J Pers Disord       Date:  2009-02
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.