Literature DB >> 35258359

Infusing Context Into Emotion Perception Impacts Emotion Decoding Accuracy.

Ursula Hess1, Konstantinos Kafetsios2,3.   

Abstract

The accurate decoding of facial emotion expressions lies at the center of many research traditions in psychology. Much of this research, while paying lip service to the importance of context in emotion perception, has used stimuli that were carefully created to be deprived of contextual information. The participants' task is to associate the expression shown in the face with a correct label, essentially changing a social perception task into a cognitive task. In fact, in many cases, the task can be carried out correctly without engaging emotion recognition at all. The present article argues that infusing context in emotion perception does not only add an additional source of information but changes the way that participants approach the task by rendering it a social perception task rather than a cognitive task. Importantly, distinguishing between accuracy (perceiving the intended emotions) and bias (perceiving additional emotions to those intended) leads to a more nuanced understanding of social emotion perception. Results from several studies that use the Assessment of Contextual Emotions demonstrate the significance and social functionality of simultaneously considering emotion decoding accuracy and bias for social interaction in different cultures, their key personality and societal correlates, and their function for close relationships processes.

Entities:  

Keywords:  accuracy; bias; emotion perception; social perception

Mesh:

Year:  2021        PMID: 35258359      PMCID: PMC9446470          DOI: 10.1027/1618-3169/a000531

Source DB:  PubMed          Journal:  Exp Psychol        ISSN: 1618-3169


Most interactions – even trivial ones – are colored in emotion. Be it the salesperson trying to sound enthusiastic about a product or a loved one complaining about their problems, emotions play a central role in everyday human communication. As such, the ability to perceive and understand the emotions of others is a central social skill (Salovey & Mayer, 1989–1990). The accurate recognition of emotions (Emotion Decoding Accuracy, EDA)[1] is central for the regulation of social and personal relationships (Manstead et al., 1999) because it helps coordination with others, communication in general, and provides the necessary affective glue in dyadic interactions (Feldman et al., 1991; Niedenthal & Brauer, 2012). Emotions can be expressed via voice, body postures, gestures (e.g., Bänziger et al., 2009), and even through touch (Hertenstein et al., 2006), yet one of the most prominent sources of emotion communication is the face. In what follows, we focus on facial expressions of emotions and their accurate decoding. However, it should be noted that the basic principle of what we propose is not in any way limited to facial affect but applies to emotion communication in general. Although the central role of EDA is acknowledged in theoretical thinking, the available evidence regarding relationships between EDA and key psychological outcomes/correlates such as social relations or personality variables is limited (Elfenbein et al., 2002). In the present article, we argue that to reveal relationships between EDA and social relations outcomes such as interaction quality, conflict, negotiation tactics, or personality variables, EDA has to be conceived of as a social perception task rather than as a cognitive task and that this is achieved by measuring EDA in a social context. Specifically, we will argue that facial expressions can be decoded using two different ways, through pattern matching and through perspective taking, but that the commonly used approach to EDA only engages the cognitive route in tasks that are less likely to reflect social competencies. We propose that a task that infuses context into EDA better captures how observers perceive emotion expression. Furthermore, such an approach can distinguish between EDA (perceiving the intended emotions) and bias (perceiving additional emotions to those expressed) and thus is better suited to capture social emotion perception as it occurs in everyday life. Results from several studies that assessed EDA using the Assessment of Contextual Emotions – a model and method that infuses context in emotion perception and distinguishes between EDA and bias – demonstrate the utility of this approach for understanding the role of EDA for social interaction in different cultures, its key personality and societal correlates, and its function for close relationships processes.

EDA Research: The Tradition

Research on EDA has traditionally focused on questions related to the universality of emotion expressions (see, e.g., Hess, 2017) or on how factors such as sex (Hall et al., 2000), cultural background (Elfenbein & Ambady, 2002), or psychiatric status (Penton-Voak et al., 2017) may influence accuracy. The goal of this line of research was to assess whether some groups make more errors in decoding than others, where errors are defined as selecting the wrong label for the expression shown. In these studies, participants were therefore presented with prototypical facial expressions drawn from standardized sets of facial expressions such the Pictures of Facial Affect (Ekman & Friesen, 1976) or with (facial) recognition tests such as the Diagnostic Analysis of Nonverbal Accuracy (Nowicki & Duke, 2001), or the Reading the Mind in the Eyes Test (Baron-Cohen et al., 2001). Participants were then asked to select from a list of emotion labels the one that best describes the depicted emotional expression. That is, accuracy is defined as the ability to associate one (correct) label with a single emotion expression shown without social context. Notably, there is limited evidence for links between EDA assessed using these traditional tasks and social relations or personality outcomes that have been postulated to rely on EDA. Toner and Gates (1985) initially noted the scant success in finding evidence for a link between EDA and personality. Elfenbein et al. (2002) and Matsumoto et al. (2000) also emphasized the scarcity of findings, which often also did not replicate across studies. Even more discouraging, given that EDA is a central aspect of emotional intelligence (EI), is null or inconsistent evidence for a relationship between EDA and EI measured either as a trait (Matthews et al., 2015) or as an ability (Farrelly & Austin, 2007). Yet more concerning is an almost total lack of established EDA models that predict well-being. This is because, as outlined in the introduction, EDA is considered a key part of social interaction and quality of social interaction is a key correlate of well-being (Kawachi & Berkman, 2001). More recently, evidence for a link between emotion recognition ability and alexithymia (Ihme et al., 2014; Jongen et al., 2014), as well as narcissism (Martins et al., 2019), and attachment style (Chris Fraley et al., 2006) has emerged, especially when considering extreme groups, but often only for some emotions and not for others. Thus, overall, the evidence for a link between EDA on the one hand and personality or social outcomes on the other is considerably less strong than theory (e.g., on emotional or social intelligence) would predict. The present paper addresses this paradox. In particular, we outline two problems with the way EDA is traditionally measured and conceived of. First, the underlying definition of what constitutes accuracy in decoding emotion is limited. Second, the lack of social context in emotion decoding transforms those classic EDA tasks into a cognitive rather than a social perception task. We will then introduce a different approach, one that understands emotion perception as a contextualized social process in which accuracy can take different forms, and show evidence that EDA as measured by a social perception accuracy task such as the Assessment of Contextualized Emotion (ACE) can be meaningfully linked to both personality and social outcomes.

The Problem With Traditional EDA Models

A first question to address is the definition of accuracy. This question seems simple and straightforward at first glance, but in fact, scholars who have reflected on social perception accuracy have long recognized the difficulties surrounding an exact definition of accuracy (Funder, 1989; Kruglanski, 1989; Zaki & Ochsner, 2011a). The classic emotion recognition literature typically starts out with a judgment accuracy approach: A rating is considered accurate when the chosen label corresponds to the criterion label established by the researcher; otherwise, it is considered inaccurate. Yet, this definition of accuracy assumes that there is one and only one correct answer. That is, emotion expression is presumed to reflect a single “pure” emotion and that the decoders are accurate when they are able to label this one pure emotion. We argue that the assumption that a single emotion label adequately describes an emotion expression is problematic. The general problem with any performance-based measure is the establishment of the correct answer or ground truth (Funder, 1995). That is, to decide whether a judgment is accurate, one needs to decide on a suitable criterion. For emotion expressions, there are several options. For example, the label can be derived from the expressive parameters for a given prototypical emotion described by Ekman and Friesen (1978). Alternatively, the label could be derived from the emotion the expresser felt during the expression (Levenson et al., 1991). The typical solution to the ground truth problem is to start with either one of these, that is, to either ask actors to pose a specific configuration or to induce a specific emotional state and select expressions according to these criteria. Since this approach usually leads to a range of expressive materials, a second validation phase is added where observers are asked to rate the expressions, and only those expressions for which a majority choose the same desired label are retained. That is, the recognition test will, in fact, be scored based on a consensus scoring procedure (see also Mayer et al., 2003). This procedure implies two problematic issues. First, it is not clear that the portrayed expressions, whether posed to be prototypical or captured while a person reports feeling a specific emotion, are in fact pure representations of a given emotional state. For instance, people who experience emotions often report blends (Watson & Stanton, 2017). Second, even if one assumes that the researchers who created the test succeeded in capturing pure emotions, there is good evidence that they would not be perceived as such. In fact, observers tend to see multiple emotions even when judging emotional expressions considered to be pure (Russell et al., 1993; Russell & Fehr, 1987; Yrizarry et al., 1998). This is especially the case in naturally occurring social interactions where people are likely to show subtle expressions that are more open to different interpretations (Ekman, 2003; Motley & Camden, 1988). Thus, focusing on a single label does not provide a realistic measure of the actual perception. These key limitations lead to a further problem associated with the response-rating method. In the existing methods, participants can typically choose only one label out of several; hence, only one form of inaccuracy can be assessed: mistaking one emotion for another. Lyusin and Ovsyannikova (2016) criticize this approach and suggest the use of a multidimensional response format or scalar rating scales (see also Matsumoto, 2005) where participants are asked to indicate all the emotions they can discern in an expression. This procedure captures the actual perception process much better and thus allows a different form of inaccuracy to be revealed – the common tendency mentioned above – to perceive more than one emotion in a given expression: in other words, to see emotions as mixed rather than as pure. However, this form of inaccuracy – contrary to the mislabeling of emotions in a forced choice task – does not necessarily result in a trade-off such that more accuracy automatically entrains less inaccuracy. In fact, the tendency to inaccurately perceive such “secondary” emotions is, arguably, theoretically independent of accuracy for the target emotion. That is, the fact that someone perceives some level of sadness in an expression that is primarily considered angry does not have to impinge on the perception of anger. Yet, the fact that in this example sadness is also perceived is very relevant as there are good reasons why this tendency should show a link to individual differences as we will outline below.

Two Ways to Decode Emotion Expressions

Specifically, there are two ways to decode emotion in facial emotion expressions. First is pattern matching where specific features of the expression are associated with specific emotions (Buck, 1984). For example, upturned corners of the mouth or lowered eyebrows are recognized as smiles or frowns respectively and a perceiver can thus conclude that the individual is happy or angry. This process can be conceived of as a cognitive task that does not rely on the perceiver's wider social knowledge. The perceiver only has to match a label to a perceived constellation of features. In fact, not even that is always required as participants can often reduce the candidate emotions based on a single cue. For example, many positive emotions include a smile element (Shiota et al., 2003). On seeing the teeth, an observer may conclude that the target showed a positive emotion and then select the only emotion in the list that qualifies (Bänziger et al., 2009). As such, the classic EDA tasks listed above may assess emotion discrimination rather than emotion recognition, especially when the list of labels is short (Bänziger et al., 2009; Nelson & Russell, 2016). However, there is a second process, which is based specifically on the perceiver's social knowledge: perspective taking. Knowing about the event that elicited an emotion allows people to use their naïve emotion theories to predict the most likely emotion that would follow such an event. When the event is unknown, such as in classic emotion recognition tasks, any social category that the perceiver is aware of and for which expectations regarding emotional reactions exist can influence emotion identification (Kirouac & Hess, 1999) in that the perceiver is more likely to attribute the more expected emotion evidenced in the ambiguous expression. For example, knowing that a (male) expresser is black or of high status leads observers to more readily label those social target's expression as angry (Hugenberg & Bodenhausen, 2003; Ratcliff et al., 2012). This second process engages participants' mental state attribution system (North et al., 2010; Zaki & Ochsner, 2011b). As this process is not solely based on the expressions shown but also involves the implicit social information provided by the image, the result is more likely to vary between participants and to lead to the differential attribution of secondary emotions (emotions that are not directly communicated by the target), in line with the observers' social knowledge and personality. Importantly, however, to properly tap the effect of social knowledge on EDA, it does not suffice to simply assess secondary emotions. As noted above, secondary emotions are more likely to be perceived when participants use perspective taking in their efforts to understand others. Yet, this process depends on the availability of a social context. Although it is widely acknowledged that emotion perception in real life rarely operates devoid of context (Barrett & Kensinger, 2010; Hess & Hareli, 2016), emotion perception research has typically used context-free facial expressions as stimuli. Even more surprising is that emotion research has largely ignored the most common form of context we experience in everyday life – other people. As emotions usually occur in (real or imagined) interactions, the presence of other people is a feature that is common to many emotion eliciting contexts. Nonetheless, the presence of others has only been considered from a cultural perspective (e.g., Kafetsios & Hess, 2015; Hess, Blaison, & Kafetsios, 2016; Masuda et al., 2008), when, in fact, it is a pervasive element of everyday interaction. The facial expressions of bystanders to an event may influence how the event itself is perceived (Hess et al., 2018), and the facial response of recipients of an expression can influence the meaning attributed to the expression (Hareli & David, 2017). We argue that presenting participants with emotion expressions shown by a group of individuals provides a very relevant and important social framing for the task. Social framing fosters the use of perspective taking which, in turn, infuses the perception process with “biases” that reflect the personality and values of the perceiver. In this sense, biases are not so much errors as they constitute an expression of the perceiver's social cognition and personality. We will revisit this point below. In sum, we have argued that perceiving expressions in social context changes the EDA task from a cognitive puzzle into a social task. Notably, we do not claim that people never utilize the cognitive puzzle approach in real life – they definitely do, for example, when they point out expressive features in a picture, such as a pleasant smile, or an ironic look; we maintain, however, that the more commonly used approach involves a predominance of perspective taking. We further propose that classic EDA tasks do not measure an important aspect of emotion perception, namely the attribution of secondary emotions. Yet, secondary emotions are an integral part of what people really perceive.

The ACE Model

Based on the above outlined considerations, we have proposed the ACE (Hess et al., 2016; Kafetsios & Hess, 2013, 2015) as a new model and method of EDA. ACE unites key considerations outlined above by infusing context in the process of decoding emotion, and by doing so, it allows for a more holistic consideration of accuracy, one that simultaneously assesses accuracy and bias. ACE inserts context in EDA through the presentation of pictorial stimuli of naturalistic facial expressions of a group of three persons (the surrounding two persons express congruent or incongruent emotions to those expressed by the central person's expressions to be decoded; for example, see Figure 1). Specifically, while there are many possible contextual elements that can be imagined,[2] a typical feature of most of these is the presence of others, which can serve to prime social processing modes. Observers rate these expressions on an emotion profile by indicating the intensity of a series of emotions using dimensional scales. Hence, the method permits the distinction between accurate evaluation of the presented focal emotions (accuracy) and the simultaneous evaluation of nonpresented, secondary, emotions (bias). Accuracy and bias are hypothesized, and found, to be largely independent EDA dimensions. Information about the ACE can be found in Hess et al. (2016) and Kafetsios and Hess (in press), and the ACE stimuli are available by request from the authors.
Figure 1

Example stimulus from the ACE faces. ACE = Assessment of Contextualized Emotion.

Accuracy and Bias

As noted above, in classic EDA research, the decoder is either right or wrong. Hence, inaccuracy is simply the proportion of responses that are not accurate. Yet, dominant models of social perception (Funder, 1995; West & Kenny, 2011) strongly maintain that accuracy and inaccuracy/bias in social perception are theoretically distinct processes and bias is distinct from error. In their seminal Truth and Bias model, West and Kenny (2011) specifically argue (p. 358) that “certain psychological mechanisms lead perceivers to be both accurate and biased, other mechanisms lead to more accuracy and less bias.” Much of the literature that inspired the Truth and Bias model and that it since has been applied to focuses on the social perception of personality traits (e.g., Overall et al., 2015). Yet, the model is very applicable to the perception of physical characteristics (cues), which also points to likely bridges between the physical and the social worlds in social perception (Zaki, 2013). We return to this point at the end of the paper. The ACE embraces this view by considering both accuracy and bias in the perception of facial expression cues. Accuracy is defined as the rating of the target emotion within the profile. Thus, for an expression that is considered an exemplar of a sadness expression, the sadness ratings on an emotion profile are considered an index of accuracy. By contrast, the mean of all other emotion ratings on the profile is considered bias. Notably, bias does not mean that the rating is absolutely wrong. It may well be that a given emotion expression in context conveys more than one message. However, we assume that facial expression cues – like any message – have a main theme or information that they intend to convey and this would be captured in the accuracy rating. In sum, the ACE approach to EDA aims to situate the task as a social task and hence to elicit social processing of the material – rather than their treatment as a cognitive puzzle. Doing so was predicted to lead to a measurement of EDA that preserves its theoretically posited relevance to EI and social functioning. We thus follow a more recent thinking about the dynamic nature of accuracy by using a measure that allows the separate measurement of accuracy and bias. We predict that EDA thus measured would allow us to reveal its relevance for social functioning and its groundedness in personality. In what follows, we want to summarize findings from our research that support this notion.

ACE and Social Functioning

Recent studies have demonstrated that ACE accuracy and bias have unique, measurable, and meaningful effects for social interaction. In three studies, two conducted in Greece and one in Germany (Hess et al., 2016), participants completed the ACE task in the laboratory (ACE cartoons in study 1 and ACE faces in studies 2 and 3) and then participated in a self-report event sampling diary of all dyadic interactions during a period of 10 days that lasted 10 minutes or more. Accuracy and Bias in the ACE meaningfully predicted self-reported parameters of interaction quality, whereas a standard emotion perception task, the MSCEIT faces (Mayer, Salovey, and Caruso Emotional Intelligence Test; Mayer et al., 2003) did not. Specifically, ACE accuracy was associated with higher quality indicators in social interaction with in-group members in Greece, whereas bias in ACE cartoons and ACE faces was associated with lower social interaction quality primarily within and outside more intimate relationships. In Germany, higher ACE faces accuracy was associated with all social interaction quality indicators across levels of intimacy (Hess et al., 2016). Importantly, ACE accuracy and bias had unique effects on social interaction quality, suggesting that one can be simultaneously both accurate and inaccurate. This point is further elaborated in the context of the Truth and Bias model of social perception (West & Kenny, 2011). Moreover, bias as measured by the ACE was associated with alexithymia, the difficulty in identifying and describing emotion, and the two were found to contribute to problems in dyadic interactions and relationships (Kafetsios & Hess, 2019). Participants completed the Toronto Alexithymia Scale (TAS) and the ACEs in a laboratory session, followed by a 10-day event sampling study of the quality of their naturally occurring social interactions. The Difficulties in Identifying Feelings (DIF) subscale of the TAS negatively related to all indices of quality of social interaction and DIF was positively and moderately strongly correlated with bias in emotion perception using the ACE. Importantly, ACE bias mediated DIF effects on social interaction outcomes. It seems that bias as measured in the ACE can tap the lack of attunement in dyadic social interactions observed in people with alexithymia (e.g., Foran & O'Leary, 2012). The notion that a contextualized test of EDA is associated with social interaction parameters leads to the very likely possibility that higher EDA can also contribute to overall well-being. In fact, there has been initial correlational evidence for a positive association between well-being (measured with Diener et al.'s 2010 flourishing scale) with ACE accuracy and a negative association with ACE bias (Kafetsios & Hess, in press). Following this up, we tested whether dyadic interaction quality during the week following the ACE assessment in the laboratory predicted well-being at the end of that week (Kafetsios et al., 2021). Results from Multilevel Structural Equation model analyses suggested direct associations between ACE accuracy and bias with well-being indicators in the expected direction and also indirect associations via social interaction quality during the week preceding the well-being assessment. The implications of this finding are important and point to a new way at looking at the lack of attunement in social relations and the effects this may have for rapport and responsiveness in social and personal relationships (Reis et al., 2017). Elements of attunement were tested with a shorter version of the ACE faces of emotion regulation in 220 Greek dating couples (Papachiou et al., 2021). Both members of couples completed the ACE faces short, a 16-item version of the ACE faces, measures of intrapersonal and interpersonal emotion regulation (Gross & John, 2003; Little et al., 2012) and participated in a 10-day event sampling study of dyadic emotion regulation. ACE accuracy was negatively, and ACE bias was positively associated with dysfunctional intrapersonal and interpersonal emotion regulation strategies. Importantly, results from actor–partner interdependence models on the event sampling part of the study found both actor and partner ACE accuracy and bias effects on dyadic emotion regulation in theoretically meaningful ways.

Context and Personality in EDA

As noted above, classic EDA research strives to show expressions devoid of context. Yet, when considering real-life emotion communication, next to the expression itself, additional sources of information contribute to the contextual characteristics of the stimulus and the decoders' social schemas (e.g., Hess & Hareli, 2016). The latter are predicated on the decoder's previous social experiences, but also on their personality. To name an example, individuals with an insecure attachment style overattribute negative affect to peoples' faces, that is, they show a bias for the attribution of negative secondary emotions (Magai et al., 2000). In fact, personality has to be understood as a system that mediates how the individual selects, construes, and processes social information and generates social behaviors (Mischel & Shoda, 1995). The Cognitive-Affective Processing System model posits social situation as a critical component of personality in that what happens in social situations very much reflects individual differences in cognitive-affective processing (Zayas et al., 2002). Personality characteristics are likely to emerge in particular situations, and it is information from both individual differences in personality and those situations that can provide predictions regarding personality–behavior links (Zayas et al., 2002). Such an approach reflects the meaning of Lewin's (1951) B = f(P,E) formula (see Funder, 2009). As such, presenting information in a socially disengaged manner, such as typical of the traditional EDA tasks, reduces its pertinence for personality-mediated processes. A set of recent studies has highlighted this close link between personality and the contextually focused approach to EDA proposed by the ACE model (Kafetsios & Hess, in press). In seven studies conducted with different samples in Greece and in Germany, accuracy and inaccuracy in ACE faces were consistently associated with personality characteristics that tap the social domain (attachment orientations, emotion regulation strategies, cultural self-construal, self-reported EI, loneliness, alexithymia, and well-being). In many of those associations, there were unique effects of accuracy and inaccuracy further supporting the notion that the two EDA processes are largely independent from one another. Notably, in none of these studies did MSCEIT faces (Mayer et al., 2003) predict the personality correlates. Importantly, a central question tested in all studies was whether a traditional hit rates approach – associating one (correct) label to a single emotion expression – can provide the same information as the accuracy and bias approach of the ACE. Across all studies, the assessment of accuracy and inaccuracy in a contextualized assessment of emotion was superior to simple hit rates in revealing associations with personality traits.

On the Path of Uniting Social Cognition and Accuracy Research

Theorizing and research in social perception has highlighted the important role that context holds for emotion perception. Some theorists have gone so far as to deny that emotion expressions are intrinsically anything but ambiguous and claim that they can only be correctly perceived as part of a given context (Hassin et al., 2013), whereas others have emphasized bidirectional influences between context and expression information (Hess & Hareli, 2018). By placing facial expressions into a context, specifically, by using stimuli that involve three people who are interacting and who show congruent or incongruent facial expressions, the ACE places the EDA task in a social frame which invites the use of social schemas in perspective taking. This novel approach to EDA opens up a number of possibilities to further link social cognition with accuracy processes, as already suggested a decade ago (Zaki & Ochsner, 2011b). For example, the ACE approach can be used to assess the impact of social attribution processes on accuracy and bias. Appraisal theories of emotion (Ellsworth & Scherer, 2003) spell out the evaluation processes that underlie emotion elicitation. Conversely, observers can reconstruct these appraisals based on the expressions of the emoter (Hareli & Hess, 2010). In a situation where more than one person is shown, the expressions of witnesses to an emoter's emotional reactions to an event can influence the evaluation of both the event and the emoter (Hareli & David, 2017; Hess et al., 2018). These influences depend on the social knowledge and the engagement of the observer. As such, elements of the group context (i.e., the presence and specific emotion expressions of others and the congruence or incongruence between the emotions of the focal and the peripheral facial expressions) influence the perception of emotions in terms of both accuracy and bias, and this influence also meaningfully reflects individual differences. This could lead to advances in linking research on social cognition with accuracy research and ultimately with outcomes (Zaki & Ochsner, 2011a). For example, given bidirectional relationships between context and social appraisal of emotion (Hess et al., 2019), one could vary the social situations and context cues that can give rise to different social appraisals to assess how levels of accuracy and bias and their combination are affected. Also, more elaborate analytic frames and models can be developed to allow a more exact mapping of how characteristics of the perceiver (e.g., social, individual, personality, socioeconomic) interact with contextual elements of the ACE to shape accuracy, bias, and their interrelationship. In both cases, application of social computational models (e.g., Ong et al., 2019) would greatly enhance the understanding of process – outcome relationships. A social computational approach could also shed light on the social perception processes responsible for the level of covariation between accuracy and bias, especially the conditions under which both target/social situation characteristics and perceiver characteristics (Hehman et al., 2017) contribute to the level of covariation between accuracy and bias. As discussed above, ACE accuracy and bias are theoretically independent, yet, depending on the specific experimental context, nontrivial covariations can be observed.

Conclusion

The present article argues that infusing context in emotion perception in the form of a group of other persons does not only add an additional source of information but changes the way that participants approach the task of emotion decoding by rendering it a social perception task rather than a cognitive task. The ACE approach to EDA presented in this paper constitutes a conceptual contribution to models of social and emotional perception. We provide evidence in line with conceptual arguments that accuracy and inaccuracy in emotion perception are theoretically distinct processes (Funder, 1995) and that accuracy and bias in social perception constitute two nonexclusive dimensions (West & Kenny, 2011; Zaki & Ochsner, 2011a). Such an approach also promotes a view of accuracy in terms of its utility for social emotion perception and its adaptive value (Kruglanski, 1989), as is evident in ACE accuracy and bias predicting several social functionality correlates.

Outstanding Questions

How do appraisals of the social situation influence ACE accuracy and bias and their covariation? How can social perception processes inherent in ACE be computationally modeled to predict decoding accuracy and bias? What are the behavioral proxies to accuracy and bias in dyadic interaction? (How) Do they mediate social functioning effects? Can a contextualized view of emotion decoding revise our understanding of cultural differences in EDA?
  35 in total

1.  The faces of positive emotion: prototype displays of awe, amusement, and pride.

Authors:  Michelle N Shiota; Belinda Campos; Dacher Keltner
Journal:  Ann N Y Acad Sci       Date:  2003-12       Impact factor: 5.691

Review 2.  Social functionality of human emotion.

Authors:  Paula M Niedenthal; Markus Brauer
Journal:  Annu Rev Psychol       Date:  2011-10-10       Impact factor: 24.137

3.  A facial expression of pax: Assessing children's "recognition" of emotion from faces.

Authors:  Nicole L Nelson; James A Russell
Journal:  J Exp Child Psychol       Date:  2015-08-27

4.  Do facial expressions signal specific emotions? Judging emotion from the face in context.

Authors:  J M Carroll; J A Russell
Journal:  J Pers Soc Psychol       Date:  1996-02

5.  On the malleability of the meaning of contexts: the influence of another person's emotion expressions on situation perception.

Authors:  Ursula Hess; Shlomo Hareli
Journal:  Cogn Emot       Date:  2016-12-21

6.  On the accuracy of personality judgment: a realistic approach.

Authors:  D C Funder
Journal:  Psychol Rev       Date:  1995-10       Impact factor: 8.934

7.  Emotion, physiology, and expression in old age.

Authors:  R W Levenson; L L Carstensen; W V Friesen; P Ekman
Journal:  Psychol Aging       Date:  1991-03

Review 8.  Personality in context: an interpersonal systems perspective.

Authors:  Vivian Zayas; Yuichi Shoda; Ozlem N Ayduk
Journal:  J Pers       Date:  2002-12

9.  Alexithymic features and the labeling of brief emotional facial expressions - An fMRI study.

Authors:  Klas Ihme; Julia Sacher; Vladimir Lichev; Nicole Rosenberg; Harald Kugel; Michael Rufer; Hans-Jörgen Grabe; André Pampel; Jöran Lepsien; Anette Kersting; Arno Villringer; Richard D Lane; Thomas Suslow
Journal:  Neuropsychologia       Date:  2014-10-02       Impact factor: 3.139

10.  The "Reading the Mind in the Eyes" Test revised version: a study with normal adults, and adults with Asperger syndrome or high-functioning autism.

Authors:  S Baron-Cohen; S Wheelwright; J Hill; Y Raste; I Plumb
Journal:  J Child Psychol Psychiatry       Date:  2001-02       Impact factor: 8.982

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.