| Literature DB >> 26246649 |
Rebecca Jürgens1, Annika Grass2, Matthis Drolet3, Julia Fischer1.
Abstract
Both in the performative arts and in emotion research, professional actors are assumed to be capable of delivering emotions comparable to spontaneous emotional expressions. This study examines the effects of acting training on vocal emotion depiction and recognition. We predicted that professional actors express emotions in a more realistic fashion than non-professional actors. However, professional acting training may lead to a particular speech pattern; this might account for vocal expressions by actors that are less comparable to authentic samples than the ones by non-professional actors. We compared 80 emotional speech tokens from radio interviews with 80 re-enactments by professional and inexperienced actors, respectively. We analyzed recognition accuracies for emotion and authenticity ratings and compared the acoustic structure of the speech tokens. Both play-acted conditions yielded similar recognition accuracies and possessed more variable pitch contours than the spontaneous recordings. However, professional actors exhibited signs of different articulation patterns compared to non-trained speakers. Our results indicate that for emotion research, emotional expressions by professional actors are not better suited than those from non-actors.Entities:
Keywords: Acoustics; Actors; Emotion; Play-acting; Vocal expressions
Year: 2015 PMID: 26246649 PMCID: PMC4519627 DOI: 10.1007/s10919-015-0209-5
Source DB: PubMed Journal: J Nonverbal Behav ISSN: 0191-5886
Fig. 1Recognition of encoding condition (a) and emotion recognition (b) across all conditions. a Unbiased hit rates for recognition of encoding condition (responding “authentic” when the stimulus is authentic, “play-acted” when stimulus is either from actors or non-actors), given as mean values ± SEM. b Unbiased hit rates for the emotion classification (mean values ± SEM). X marks the mean chance levels per condition. Asterisks refer to the significance level. < .1; *p < .05; **p < .01, ***p < .001. Asterisks at the bottom of the bars indicate differences between unbiased hit rates and individual chance level
Post-hoc comparisons for recognition of encoding condition and emotion (unbiased hit rates)
| Type | Emotion | Encoding conditions | Estimatea |
| z value |
| |
|---|---|---|---|---|---|---|---|
| Encoding condition | Anger | Authentic | Actors | 0.00848 | 0.002049 | 4.138 | <.001*** |
| Authentic | Non-actors | 0.01520 | 0.002049 | 7.418 | <.001*** | ||
| Actors | Non-actors | 0.00672 | 0.002049 | 3.279 | .013* | ||
| Fear | Authentic | Actors | 0.01057 | 0.002049 | 5.159 | <.001*** | |
| Authentic | Non-actors | 0.02029 | 0.002049 | 9.901 | <.001*** | ||
| Actors | Non-actors | 0.00972 | 0.002049 | 4.743 | <.001*** | ||
| Joy | Authentic | Actors | 0.00545 | 0.002049 | 2.661 | .094 | |
| Authentic | Non-actors | 0.01795 | 0.002049 | 8.76 | <.001*** | ||
| actors | Non-actors | 0.01250 | 0.002049 | 6.099 | <.001*** | ||
| Sadness | Authentic | Actors | 0.00800 | 0.002049 | 3.902 | .001** | |
| Authentic | Non-actors | 0.00772 | 0.002049 | 3.769 | .002** | ||
| Actors | Non-actors | −0.00027 | 0.002049 | −0.133 | 1 | ||
| Emotion | Anger | Authentic | Actors | −0.01111 | 0.002314 | −4.8 | <.001*** |
| Authentic | Non-actors | −0.00465 | 0.002314 | −2.009 | .535 | ||
| Actors | Non-actors | 0.00646 | 0.002314 | 2.792 | .063. | ||
| Fear | Authentic | Actors | 0.00116 | 0.002314 | 0.501 | 1 | |
| Authentic | Non-actors | 0.00405 | 0.002314 | 1.749 | .964 | ||
| Actors | Non-actors | 0.00289 | 0.002314 | 1.248 | 1 | ||
| Joy | Authentic | Actors | 0.76010 | 0.002314 | 3.285 | .012* | |
| Authentic | Non-actors | 0.00464 | 0.002314 | 2.005 | .539 | ||
| Actors | Non-actors | −0.00296 | 0.002314 | −1.28 | 1 | ||
| Sadness | Authentic | Actors | 0.00776 | 0.002314 | 3.353 | .0096** | |
| Authentic | Non-actors | 0.00886 | 0.002314 | 3.829 | .0016** | ||
| Actors | Non-actors | 0.00110 | 0.002314 | 0.47 | 1 | ||
Asterisks mark the significance level at *p < .05; **p < .01; ***p < .001
aBased on arscine transformed data
bAdjusted p values (Bonferroni correction)
Fig. 2Selected acoustic parameters separated for emotion and encoding condition. Mean values are given for vowels a, e, i ± SEM
Results of the linear mixed models on the acoustic structure of vowels
| Parameter | Vowel | Emotiona | Encoding conditiona | Gendera | ||||
|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
| Estimates ± SEc | ||
| F0 | a | 15.34 | .005** | 0.42 | 1 | 43.56 | <.001*** | 70.49 ± 9.72 |
| e | 15.02 | .005** | 0.86 | 1 | 45.07 | <.001*** | 69.72 ± 9.30 | |
| i | 14.48 | .007** | 3.41 | .540 | 32.12 | <.001*** | 70.18 ± 11.46 | |
| HNR | a | 1.35 | 1 | 1.28 | 1 | 21.15 | <.001*** | 0.046 ± 0.01 |
| e | 3.32 | 1 | 0.19 | 1 | 22.99 | <.001*** | 0.045 ± 0.009 | |
| i | 11.91 | .023* | 2.59 | .822 | 41.78 | <.001*** | 0.076 ± 0.010 | |
| Amplitude ratio | a | 5.61 | .397 | 52.07 | <.001*** | 5.09 | .072 | −0.221 ± 0.099d |
| e | 1.74 | 1 | 26.15 | <.001*** | 27.61 | <.001*** | −0.651 ± 0.117d | |
| i | 11.65 | .026* | 3.79 | .450 | 32.44 | <.001*** | −1.028 ± 0.164d | |
| Peak frequency | a | 21.42 | <.001*** | 24.22 | <.001*** | 0.76 | 1 | 0.067 ± 0.0774d |
| e | 8.13 | .13 | 17.53 | <.001*** | 6.99 | .024* | 0.169 ± 0.064d | |
| i | 0.55 | 1 | 24.58 | <.001*** | 6.50 | .032* | 0.121 ± 0.048d | |
| First formant | a | 4.66 | .198 | 19.08 | <.001*** | 5.00 | .025* | 0.190 ± 0.086d |
Asterisks mark the significance level at *p < .05; **p < .01; ***p < .001
aStatistical values are obtained from the model comparison (full model to reduced model excluding the respective predictor)
b p value adjustments (Bonferroni correction) were done for the different vowels within one acoustic parameter and one predictor
cEstimates for the predictor gender were gained from the LMM, with male speakers included in the intercept. Estimates refer to the female speakers in comparison to the male speakers
dValues base on log transformed data
Results of the post-hoc analyses on the influence of encoding condition on the acoustic structure of vowels
| Parameter | Vowel | Encoding condition | Estimate |
| z value |
| |
|---|---|---|---|---|---|---|---|
| Amplitude ratio | a | Authentic | Actor | 0.552 | 0.115 | 4.949 | <.001*** |
| Authentic | Non-actor | −0.423 | 0.109 | −3.895 | <.001*** | ||
| Actors | Non-actors | −0.975 | 0.123 | −7.941 | <.001*** | ||
| e | Authentic | Actor | 0.385 | 0.121 | 3.177 | .004** | |
| Authentic | Non-actor | −0.323 | 0.125 | −2.591 | .029* | ||
| Actors | Non-actors | −0.709 | 0.133 | −5.323 | <.001*** | ||
| Peak frequency | a | Authentic | Actor | 0.196b | 0.077b | 2.550 | .0323 |
| Authentic | Non-actor | −0.231b | 0.075b | −3.093 | .006** | ||
| Actors | Non-actors | −0.428b | 0.084b | −5.077 | <.001*** | ||
| e | Authentic | Actor | 0.103b | 0.567b | 1.814 | .209 | |
| Authentic | Non-actor | −0.163b | 0.058b | −2.798 | .015* | ||
| Actors | Non-actors | −0.266b | 0.063b | −4.243 | <.001*** | ||
| i | Authentic | Actor | 0.022b | 0.050b | 0.450 | 1 | |
| Authentic | Non-actor | −0.206b | 0.045b | −4.555 | <.001*** | ||
| Actors | Non-actors | −0.229b | 0.052b | −4.381 | <.001*** | ||
| First formant | a | Authentic | Actor | −0.444b | 0.105b | −4.299 | <.001*** |
| Authentic | Non-actor | −0.052b | 0.102b | −0.511 | 1 | ||
| Actors | Non-actors | 0.392b | 0.114b | 3.442 | .002** | ||
Asterisks mark the significance level at *p < .05; **p < .01; ***p < .001
aAdjusted p values (Bonferroni correction)
bValues base on log transformed data
Fig. 3Speech tempo for the different encoding conditions and emotion categories. Mean ± SEM is given for a speech rate and b articulation rate