| Literature DB >> 32123191 |
Maria Gendron1, Katie Hoemann2, Alyssa N Crittenden3, Shani Msafiri Mangola4, Gregory A Ruark5, Lisa Feldman Barrett6,7.
Abstract
It has long been claimed that certain configurations of facial movements are universally recognized as emotional expressions because they evolved to signal emotional information in situations that posed fitness challenges for our hunting and gathering hominin ancestors. Experiments from the last decade have called this particular evolutionary hypothesis into doubt by studying emotion perception in a wider sample of small-scale societies with discovery-based research methods. We replicate these newer findings in the Hadza of Northern Tanzania; the Hadza are semi-nomadic hunters and gatherers who live in tight-knit social units and collect wild foods for a large portion of their diet, making them a particularly relevant population for testing evolutionary hypotheses about emotion. Across two studies, we found little evidence of universal emotion perception. Rather, our findings are consistent with the hypothesis that people infer emotional meaning in facial movements using emotion knowledge embrained by cultural learning.Entities:
Mesh:
Year: 2020 PMID: 32123191 PMCID: PMC7051983 DOI: 10.1038/s41598-020-60257-2
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Tests of the Universality Hypothesis for Facial Configurations and Vocal Cues in Small-Scale Societies. Findings summarized for anger, disgust, fear, sadness, and surprise; happiness is the only pleasant category tested in all studies except Tracy and Robins (2008), and therefore perception can be (and likely is) guided by distinguishing valence in those studies. All studies used photographs of posed facial configurations or posed vocalizations, except Crivelli, Jarillo et al. (2016), Study 2, and Crivelli et al. (2017), Study 1, which used dynamic as well as static posed configurations and static spontaneous configurations from Papua New Guinea (PNG), respectively. In Bryant and HC Barrett (2008), participants were tested in a second language (Spanish) in which they received training. A subset of choice-from-array studies did not control whether foils and target facial configurations could be distinguished by valence and/or arousal, with the exception of Gendron et al. 2014a, Study 2, which controlled for valence and arousal; Sauter et al. 2015 (2010 re-analysis) and Cordaro et al. (2015) controlled for valence only. N = sample size. Unsupported = consistency and specificity at chance, or any level of consistency above chance combined with evidence of no specificity. Weak support = consistency between 20% and 40% (weak) for at least a single emotion category (other than happiness) combined with above-chance specificity for that category, or consistency between 41% and 70% (moderate) for at least a single category (other than happiness) with unknown specificity. Moderate support = consistency between 41% and 70% (moderate) combined with any evidence of above-chance specificity those categories, or consistency above 70% (strong) for at least a single category (other than happiness) with unknown specificity. Strong support = strong evidence of consistency (above 70%) and strong evidence of specificity for at least a single emotion category (other than happiness). Superscript a: Specificity levels were not reported. Superscript a1: Specificity inferred from reported results. Superscript a2: Traditional specificity and consistency tests are inappropriate for this method, but the results are placed here based on the original author’s interpretation of multidimensional scaling and clustering results. Superscript b: The sample size, marginal means, and exact pattern of errors reported for the Sadong samples is identical in Sorenson (1975), Sample 3 and Ekman et al. (1969); Sorenson described using a free-labeling method and Ekman et al. (1969) described using a choice-from-array method in which participants were shown photographs and asked to choose a label from a small list of emotion words; Ekman (1994) indicated, however, that he did not use a free-labeling method, implying that the samples are distinct. Superscript c: Sorenson (1975), Sample 2 included three groups of Fore participants (those with little, moderate, and most other-group contact). The pattern of findings is nearly identical for the subgroup with the most contact and the data reported for the Fore in Ekman et al. (1969); again, Sorenson described using a free-labeling method and Ekman et al. (1969) described using a choice-from-array method. It is questionable whether the Sadong and the Fore subgroup should be considered a small-scale society (see Sorenson, 1975, p. 362 and 363), but we include them here to avoid falsely dichotomizing cultures as “isolated from” versus “exposed to” one another (Fridlund, 1994; Gewald, 2010). Superscript d: These are likely the same sample because the sample sizes and pattern of data are identical for all emotion categories except for the fear category, which is extremely similar, and for the disgust category, which includes responses for contempt in Ekman and Friesen (1971) but was kept separate in Sorenson (1975). Superscript e: Participants were children. Superscript f: Participants were adolescents. Superscript g: The Dani sample reported in Ekman (1972) is likely a subset of the data from Ekman, Heider, Friesen, and Heider (unpublished manuscript).
Figure 2Coded responses from Study 1. Top panel depicts verbal responses produced by Hadza (left) and US (right) samples that were coded as “mental states”. The proportion of labels produced by a given sample are plotted, with higher intensity (yellow) values indicating a higher proportion and lower intensity (blue) values indicating a lower proportion; numerical proportion is also presented in each cell. Responses are plotted by the coded label types produced (y-axis) for each facial configuration of interest (x-axis). Other mental = other mental labels offered that did not conform to otherwise coded categories. Lower panel depicts verbal responses produced by Hadza (left) and US (right) samples that were coded as consistent with a set of “functional” descriptions derived from the prior literature. Functional descriptions are clustered according to their theoretically proposed links to specific emotions. Other action = other action labels offered that did not conform to otherwise coded categories.
Free-Labeling Results: Study 1.
| Coded Response | Facial Configurations | |||||
|---|---|---|---|---|---|---|
| Scowl | Nose-Wrinkle | Wide-Eyed Gasp | Smile | Pout | Wide-Eyed | |
| Anger | 0.67a [0.51 0.79] | 0.17 [0.08 0.31] | 0.21 [0.11 0.36] | 0.02 [−0.01 0.13] | 0.26 [0.15 0.41] | 0.12 [0.05 0.25] |
| Disgust | 0 | 0b | 0.05 [0 0.17] | 0 | 0 | 0 |
| Fear | 0 | 0 | 0.07b [0.02 0.2] | 0 | 0 | 0.02 [−0.01 0.14] |
| Happy | 0.02 [−0.01 0.13] | 0.05 [0 0.16] | 0.07 [0.02 0.19] | 0.44a [0.3 0.59] | 0.07 [0.02 0.19] | 0.15 [0.07 0.29] |
| Sad | 0 | 0.07 [0.02 0.19] | 0.02 [−0.01 0.13] | 0 | 0.12b [0.05 0.25] | 0 |
| Surprised | 0 | 0 | 0.12 [0.05 0.26] | 0 | 0 | 0.15b [0.07 0.29] |
| 133.41*** | 19** | 10.65† | 88.60*** | 29.95*** | 14.67* | |
| 0.80 | 0.30 | 0.23 | 0.64 | 0.37 | 0.27 | |
| Other Mental | 0.17 [0.08 0.31] | 0.39 [0.26 0.54] | 0.24 [0.14 0.4] | 0.12 [0.05 0.25] | 0.12 [0.05 0.25] | 0.2 [0.1 0.34] |
| Lash out | 0.02 [−0.01 0.13] | 0 | 0 | 0 | 0 | 0 |
| Withdraw | 0 | 0 | 0.02 [−0.01 0.14] | 0 | 0 | 0 |
| Escape | 0 | 0 | 0 | 0 | 0 | 0 |
| Affiliate | 0 | 0 | 0 | 0 | 0.02 [−0.01 0.13] | 0.02 [−0.01 0.14] |
| Laugh/Smile | 0 | 0 | 0.15 [0.07 0.29] | 0.58 [0.43 0.72] | 0.02 [−0.01 0.13] | 0.10 [0.03 0.23] |
| Cry | 0.02 [−0.01 0.13] | 0.17 [0.08 0.32] | 0.07 [0.02 0.2] | 0 | 0.42 [0.28 0.57] | 0.02 [−0.01 0.14] |
| 4.00 | 35.00*** | 17.60* | 125.00*** | 77.80*** | 12.00† | |
| 0.15 | 0.46 | 0.33 | 0.85 | 0.67 | 0.27 | |
| Other Actions | 0.26 [0.15 0.41] | 0.22 [0.12 0.37] | 0.24 [0.14 0.4] | 0.05 [0 0.16] | 0.16 [0.08 0.3] | 0.41 [0.28 0.57] |
| Constrict Orifices | 0 | 0 | 0 | 0 | 0 | 0 |
| Olfaction | 0 | 0 | 0 | 0 | 0 | 0 |
| Vomit | 0 | 0 | 0.02 [−0.01 0.14] | 0 | 0 | 0.02 [−0.01 0.14] |
| Wide Eyes/Visual Monitoring | 0 | 0 | 0 | 0 | 0 | 0 |
| Wide Eyes/See Unexpected | 0 | 0 | 0.02 [−0.01 0.14] | 0 | 0 | 0.05 [0 0.17] |
| Vision | 0 | 0.05 [0 0.17] | 0.15 [0.07 0.29] | 0 | 0.02 [−0.01 0.13] | 0.10 [0.03 0.23] |
| — | 10.00 | 20.50*** | — | 5.00 | 12.00† | |
| — | 0.22 | 0.32 | — | 0.15 | 0.24 | |
| Threatening | 0 | 0 | 0 | 0 | 0 | 0 |
| Dominance | 0 | 0 | 0 | 0 | 0 | 0 |
| Aversive Food Warn | 0 | 0 | 0 | 0 | 0 | 0 |
| Distasteful Idea/Behavior Warn | 0 | 0.02 [−0.01 0.14] | 0 | 0 | 0 | 0 |
| Alert to Threat | 0.02 [−0.01 0.13] | 0 | 0 | 0 | 0.02 [−0.01 0.13] | 0 |
| Aggressor Appeasement | 0 | 0 | 0 | 0 | 0 | 0 |
| No Threat | 0 | 0 | 0 | 0.05 [0 0.16] | 0 | 0 |
| Appease/Extract Sympathy | 0 | 0 | 0 | 0 | 0 | 0 |
| Anger | 0.69a [0.54 0.81] | 0 | 0.02 [−0.01 0.13] | 0 | 0 | 0 |
| Disgust | 0 | 0.69a [0.54 0.81] | 0.07 [0.02 0.19] | 0 | 0 | 0 |
| Fear | 0 | 0.13 [0.06 0.27] | 0.67a [0.52 0.79] | 0 | 0 | 0.11 [0.04 0.24] |
| Happy | 0 | 0 | 0 | 0.78a [0.64 0.88] | 0.02 [−0.01 0.13] | 0.07 [0.02 0.19] |
| Sad | 0.02 [−0.01 0.13] | 0.07 [0.02 0.19] | 0 | 0 | 0.76a [0.61 0.86] | 0 |
| Surprised | 0 | 0 | 0.07 [0.02 0.19] | 0 | 0 | 0.76a [0.61 0.86] |
| 148.38*** | 110.90*** | 112.03*** | 175.00*** | 163.34*** | 128.00*** | |
| 0.81 | 0.70 | 0.71 | 0.88 | 0.85 | 0.75 | |
| Other Mental | 0.24 [0.14 0.39] | 0.09 [0.03 0.21] | 0.13 [0.06 0.27] | 0.18 [0.09 0.32] | 0.22 [0.12 0.36] | 0.04 [0 0.16] |
| Lash out | 0 | 0 | 0 | 0 | 0 | 0 |
| Withdraw | 0 | 0 | 0.04 [0 0.16] | 0 | 0 | 0 |
| Escape | 0 | 0 | 0 | 0 | 0 | 0 |
| Affiliate | 0 | 0 | 0 | 0.02 [−0.01 0.13] | 0 | 0 |
| Laugh/Smile | 0 | 0 | 0 | 0.07 [0.02 0.19] | 0 | 0 |
| Cry | 0 | 0 | 0 | 0 | 0.02 [−0.01 0.13] | 0 |
| — | — | 10.00 | 11.00† | 5.00 | — | |
| — | — | 0.24 | 0.25 | 0.17 | — | |
| Other Actions | 0.07 [0.02 0.19] | 0.04 [0 0.16] | 0.07 [0.02 0.19] | 0.02 [−0.01 0.13] | 0.04 [0 0.16] | 0.04 [0 0.16] |
| Constrict Orifices | 0 | 0.02 [−0.01 0.13] | 0 | 0 | 0 | 0 |
| Olfaction | 0 | 0.02 [−0.01 0.13] | 0 | 0 | 0 | 0 |
| Vomit | 0 | 0 | 0 | 0 | 0 | 0 |
| Wide Eyes/Visual Monitoring | 0 | 0 | 0 | 0 | 0 | 0 |
| Wide Eyes/See Unexpected | 0 | 0 | 0.02 [−0.01 0.13] | 0 | 0 | 0.02 [−0.01 0.13] |
| Vision | 0 | 0 | 0 | 0 | 0 | 0 |
| — | 4.00 | 5.00 | — | — | 10.00 | |
| — | 0.13 | 0.15 | — | — | 0.21 | |
| Threatening | 0 | 0 | 0 | 0.02 [−0.01 0.13] | 0 | 0 |
| Dominance | 0.02 [−0.01 0.13] | 0 | 0 | 0 | 0 | 0 |
| Aversive Food Warn | 0 | 0 | 0 | 0 | 0 | 0 |
| Distasteful Idea/Behavior Warn | 0.02 [−0.01 0.13] | 0.02 [−0.01 0.13] | 0 | 0 | 0 | 0 |
| Alert to Threat | 0 | 0 | 0 | 0 | 0 | 0 |
| Aggressor Appeasement | 0 | 0 | 0 | 0 | 0 | 0 |
| No Threat | 0 | 0 | 0 | 0.02 [−0.01 0.13] | 0 | 0 |
| Appease/Extract Sympathy | 0 | 0 | 0 | 0 | 0 | 0 |
Note. Proportion of coded responses provided by participants for each facial configuration in the US and Hadza samples with 95% Agresti-Coull Confidence Intervals in brackets; CI for US 0 frequency cells: [−0.02, 0.09]; CI for Hadza 0 frequency cells: [−0.02, 0.10]. Responses along the main diagonals are consistent with theoretical predictions for universal emotion or feature labeling. Cochran’s Q and McNemar pairwise comparisons were computed for the data represented along the diagonal for each cultural context separately, with superscripts indicating which responses (within culture) are statistically different from one another. χ2 goodness-of-fit tests are reported for each column, within each code type (again, within each cultural group). p-values are based on Monte-Carlo simulations with 10,000 replicates. Significant χ2 goodness-of-fit tests indicate that the distribution of a given response feature was not uniform across the codes for a given facial configuration. †p ≤ 0.10. *p ≤ 0.05. ** p ≤ 0.01. *** p ≤ 0.001. Bolded proportions indicate facial configurations for which a given coded response was characteristic for that facial configuration (defined as greater than two standardized residuals based on the χ2 goodness-of-fit test; after Crivelli et al., 2017). Dashes indicate rows for which a χ2 test could not be computed because no responses were coded consistent with the feature. No χ2 test was computed for the social communication codes (final block of the US and Hadza portions of the table) due to low frequencies of the response feature and low reliability in coding. *Note that the the Hadza dataset contains variable number of participants across target facial configurations (Nscowl = 42, Nnose-wrinkle = 41, Nwide-eyed gasp = 41, Nsmile = 43, Npout = 43, Nwide-eyed = 41) due to a subset of responses for which a reliable translation could not be achieved (see Supplementary Information for more details).
Figure 3Study 2 Task Conditions (a) and Performance for US (b,e), Hadza (c,f) and Hadza participants with minimized exposure to other cultural groups (Hadza-M; subset based on proxy variables of second language skill and formal schooling) (d,g). (a) Examples of vignettes (for all scenarios, see Supplementary Information Table 6), targets and foils for the four trial types. Facial configurations are examples because stimulus sets restrict publication of actual photographs. Arousal-controlled trials: the foil face differed from the target only in depicting positivity or negativity, or valence (e.g., a smiling facial configuration hypothesized to be the universal expression of happiness vs. a scowling facial configuration hypothesized to be the universal expression of anger). Valence is a descriptive feature of affect, along with a second feature, level of arousal. For example, some evidence suggests that perceivers may be able to distinguish scowling from pouting not because scowling is perceived as “anger” and pouting is perceived as “sadness” but because scowling is typically perceived as high arousal and pouting as low arousal. Valence-controlled trials: the foil face differed from the target only in depicting level of arousal (e.g., a scowling vs. a pouting configuration hypothesized to be the universal expressions of anger and sadness, respectively). Affect-uncontrolled trials: the foil face differed from the target in depicting both valence and level arousal (e.g., a smiling vs. a pouting configuration). Affect-controlled trials: the foil face matched the target in depicting valence and arousal (e.g., a scowling vs. a wide-eyed gasping facial configuration hypothesized to be the universal expressions of anger and fear, respectively). Performance for each of the 4 experimental conditions (x-axis) is plotted for US participants (b), Hadza participants (c) and Hadza-M participants (d). Performance within the affect-controlled condition, for each of the 3 target facial configurations (x-axis) is plotted for US participants (e), Hadza participants (f) and Hadza-M participants (g). Individual data points represent mean proportion agreement (i.e., selecting a target matching the presumed universal model) for a given participant within a given condition. Contours of violin plots represent density of data points at a given agreement level. Horizontal red bar represents chance-level performance, and significance against chance-level responding is noted at the top of each violin plot: ***p < 0.001 **p < 0.01 *p < 0.05 p < 0.10. Means combined with brackets represent conditions that do not statistically differ in χ tests (ps > 0.25). Statistically significant differences between conditions based on follow-up χ tests are notated using the same conventions, with the following exception: **(*) indicates statistical significance for individual tests ranged between p < 0.01 and p < 0.001.
Choice-From Array Results: Study 2.
| Model | Fixed Effect | ||||
|---|---|---|---|---|---|
| Hadza | Affect Uncontrolled, | 1.414a, *** | 4.112 | (3.057, 5.532) | 0.804 |
| Arousal Controlled, | 1.222a, *** | 3.395 | (2.597, 4.438) | 0.772 | |
| Valence Controlled, | 0.491b, *** | 1.634 | (1.278, 2.088) | 0.620 | |
| Affect Controlled, | 0.567b, *** | 1.762 | (1.336, 2.325) | 0.638 | |
| US | Affect Uncontrolled, | 2.219a, *** | 9.198 | (7.750, 10.916) | 0.902 |
| Arousal Controlled, | 1.936a, *** | 6.929 | (5.475, 8.768) | 0.874 | |
| Valence Controlled, | 1.600b, *** | 4.954 | (3.651, 6.724) | 0.832 | |
| Affect Controlled, | 1.922a, *** | 6.834 | (5.217, 8.952) | 0.872 |
Note. Table reports generalized linear model (HGLM) population average results with robust standard errors for Level-1 intercepts. b = regression coefficient, OR = odds ratio, CI = confidence interval, Prob = estimated probability of success. ***p ≤ 0.001. Superscripts denote whether coefficients are statistically different from one another based on χ2 hypothesis testing within a model (i.e., each society was modeled separately). Superscripts only hold for comparisons within a given society.
Influence of Other-Culture Exposure in Hadza Participants: Study 2.
| Fixed Effect | ||||
|---|---|---|---|---|
| Minimal cultural exposure, | 1.24*** | 3.44 | (2.383,4.969) | 0.775 |
| Formal schooling, | 0.01 | 1.01 | (0.921,1.111) | |
| Swahili skill, | 0.54 | 1.72 | (0.877,3.358) | |
| Minimal cultural exposure, | 1.02*** | 2.76 | (2.060,3.698) | 0.734 |
| Formal schooling, | 0.02 | 1.02 | (0.862,1.197) | |
| Swahili language skill, | 0.64 | 1.89 | (0.774,4.607) | |
| Minimal cultural exposure, | 0.42** | 1.52 | (1.124,2.068) | 0.604 |
| Formal schooling, | 0.045 | 1.05 | (0.961,1.144) | |
| Swahili language skill, | −0.04 | 0.965 | (0.530,1.752) | |
| Minimal cultural exposure, | 0.43* | 1.535 | (1.097,2.139) | 0.605 |
| Formal schooling, | 0.07 | 1.07 | (0.943,1.209) | |
| Swahili language skill, | 0.10 | 1.11 | (0.600,2.045) | |
Note. Table reports hierarchical generalized linear model (HGLM) population average results with robust standard errors for Level-1 intercepts. b = regression coefficient, OR = odds ratio, CI = confidence interval, Prob = estimated probability of success. The HGLM includes both formal schooling and Swahili language skill and as Level-2 predictors. Self-reported formal schooling was entered based on the number of years completed. Self-reported Swahili language skill was dichotomized as 0=poor, 1=good. The intercept for each condition, what we call “minimal cultural exposure”, tests performance against chance-level responding for participants who self-reported poor Swahili language skill and had no years of formal schooling. For models separately examining Schooling and Swahili as Level-2 predictors, see Supplementary Information Table 10.
Choice-From-Array Performance in Affect-Controlled Trials: Study 2.
| Model | Fixed Effect | ||||
|---|---|---|---|---|---|
| Hadza | Scowl, | 0.45a, ** | 1.57 | (1.021, 2.421) | 0.61 |
| Nose Wrinkle, | 0.34a, † | 1.41 | (0.980, 2.031) | 0.59 | |
| Wide Eyed Gasp, | 0.97b, *** | 2.63 | (1.653, 4.180) | 0.72 | |
| US | Scowl, | 2.11a, *** | 8.21 | (5.201, 12.976) | 0.89 |
| Nose Wrinkle, | 1.87a, *** | 6.51 | (4.169, 10.158) | 0.87 | |
| Wide Eyed Gasp, | 1.91a, *** | 6.77 | (4.448, 10.317) | 0.87 |
Note. Table reports generalized linear model (HGLM) population average results with robust standard errors for Level-1 intercepts. b = regression coefficient, OR = odds ratio, CI = confidence interval, Prob = estimated probability of success. †p ≤ 0.10. **p ≤ 0.01. **p ≤ 0.001. Superscripts denote whether coefficients are statistically different from one another based on χ2 hypothesis testing within a model (i.e., each society was modeled separately). Superscripts only hold for comparisons within a given society.
Influence of Other-Culture Exposure in Hadza Participants During Affect-Controlled Trials: Study 2.
| Fixed Effect | ||||
|---|---|---|---|---|
| Minimal cultural exposure, | 0.305 | 1.356 | (0.810,2.272) | 0.576 |
| Formal schooling, | 0.030 | 1.031 | (0.819,1.298) | |
| Swahili skill, | 0.306 | 1.358 | (0.379,4.858) | |
| Minimal cultural exposure, | 0.390† | 1.477 | (0.960,2.271) | 0.596 |
| Formal schooling, | −0.032 | 0.969 | (0.806,1.164) | |
| Swahili language skill, | 0.018 | 1.019 | (0.476,2.180) | |
| Minimal cultural exposure, | 0.621* | 1.861 | (1.033,3.353) | 0.650 |
| Formal schooling, | 0.271** | 1.312 | (1.072,1.605) | |
| Swahili language skill, | −0.059 | 0.942 | (0.332,2.676) | |
Note. Table reports hierarchical generalized linear model (HGLM) population average results with robust standard errors for Level-1 intercepts. SE = standard error, df = approximate degrees of freedom, OR = odds ratio, CI = confidence interval, Prob = estimated probability of success. **p ≤ 0.01. **p ≤ 0.001. The HGLM model includes both formal schooling and Swahili language skill and as Level-2 predictors. Self-reported formal schooling was entered based on the number of years completed. Self-reported Swahili language skill was dichotomized as 0=poor, 1=good. The intercept for each condition, what we call “minimal cultural exposure”, tests performance against chance-level responding for participants who self-reported poor Swahili language skill and had no years of formal schooling. For models separately examining Schooling and Swahili as Level-2 predictors, see Supplementary Information Table 11.