Literature DB >> 34294630

Weighting of Prosodic and Lexical-Semantic Cues for Emotion Identification in Spectrally Degraded Speech and With Cochlear Implants.

Margaret E Richter1, Monita Chatterjee2.   

Abstract

OBJECTIVES: Normally-hearing (NH) listeners rely more on prosodic cues than on lexical-semantic cues for emotion perception in speech. In everyday spoken communication, the ability to decipher conflicting information between prosodic and lexical-semantic cues to emotion can be important: for example, in identifying sarcasm or irony. Speech degradation in cochlear implants (CIs) can be sufficiently overcome to identify lexical-semantic cues, but the distortion of voice pitch cues makes it particularly challenging to hear prosody with CIs. The purpose of this study was to examine changes in relative reliance on prosodic and lexical-semantic cues in NH adults listening to spectrally degraded speech and adult CI users. We hypothesized that, compared with NH counterparts, CI users would show increased reliance on lexical-semantic cues and reduced reliance on prosodic cues for emotion perception. We predicted that NH listeners would show a similar pattern when listening to CI-simulated versions of emotional speech.
DESIGN: Sixteen NH adults and 8 postlingually deafened adult CI users participated in the study. Sentences were created to convey five lexical-semantic emotions (angry, happy, neutral, sad, and scared), with five sentences expressing each category of emotion. Each of these 25 sentences was then recorded with the 5 (angry, happy, neutral, sad, and scared) prosodic emotions by 2 adult female talkers. The resulting stimulus set included 125 recordings (25 Sentences × 5 Prosodic Emotions) per talker, of which 25 were congruent (consistent lexical-semantic and prosodic cues to emotion) and the remaining 100 were incongruent (conflicting lexical-semantic and prosodic cues to emotion). The recordings were processed to have 3 levels of spectral degradation: full-spectrum, CI-simulated (noise-vocoded) to have 8 channels and 16 channels of spectral information, respectively. Twenty-five recordings (one sentence per lexical-semantic emotion recorded in all five prosodies) were used for a practice run in the full-spectrum condition. The remaining 100 recordings were used as test stimuli. For each talker and condition of spectral degradation, listeners indicated the emotion associated with each recording in a single-interval, five-alternative forced-choice task. The responses were scored as proportion correct, where "correct" responses corresponded to the lexical-semantic emotion. CI users heard only the full-spectrum condition.
RESULTS: The results showed a significant interaction between hearing status (NH, CI) and congruency in identifying the lexical-semantic emotion associated with the stimuli. This interaction was as predicted, that is, CI users showed increased reliance on lexical-semantic cues in the incongruent conditions, while NH listeners showed increased reliance on the prosodic cues in the incongruent conditions. As predicted, NH listeners showed increased reliance on lexical-semantic cues to emotion when the stimuli were spectrally degraded.
CONCLUSIONS: The present study confirmed previous findings of prosodic dominance for emotion perception by NH listeners in the full-spectrum condition. Further, novel findings with CI patients and NH listeners in the CI-simulated conditions showed reduced reliance on prosodic cues and increased reliance on lexical-semantic cues to emotion. These results have implications for CI listeners' ability to perceive conflicts between prosodic and lexical-semantic cues, with repercussions for their identification of sarcasm and humor. Understanding instances of sarcasm or humor can impact a person's ability to develop relationships, follow conversation, understand vocal emotion and intended message of a speaker, following jokes, and everyday communication in general.
Copyright © 2021 Wolters Kluwer Health, Inc. All rights reserved.

Entities:  

Mesh:

Year:  2021        PMID: 34294630      PMCID: PMC8545870          DOI: 10.1097/AUD.0000000000001057

Source DB:  PubMed          Journal:  Ear Hear        ISSN: 0196-0202            Impact factor:   3.570


  42 in total

1.  Evaluation of nonverbal emotion in face and voice: some preliminary findings on a new battery of tests.

Authors:  Marc David Pell
Journal:  Brain Cogn       Date:  2002 Mar-Apr       Impact factor: 2.310

2.  Pitch contour identification with combined place and temporal cues using cochlear implants.

Authors:  Xin Luo; Monica Padilla; David M Landsberger
Journal:  J Acoust Soc Am       Date:  2012-02       Impact factor: 1.840

3.  Validation of a simple response-time measure of listening effort.

Authors:  Carina Pals; Anastasios Sarampalis; Hedderik van Rijn; Deniz Başkent
Journal:  J Acoust Soc Am       Date:  2015-09       Impact factor: 1.840

4.  Channel interaction limits melodic pitch perception in simulated cochlear implants.

Authors:  Joseph D Crew; John J Galvin; Qian-Jie Fu
Journal:  J Acoust Soc Am       Date:  2012-11       Impact factor: 1.840

5.  Interactions Between Item Set and Vocoding in Serial Recall.

Authors:  Adam K Bosen; Mary C Luckasen
Journal:  Ear Hear       Date:  2019 Nov/Dec       Impact factor: 3.570

6.  Speech Recognition in Adults With Cochlear Implants: The Effects of Working Memory, Phonological Sensitivity, and Aging.

Authors:  Aaron C Moberly; Michael S Harris; Lauren Boyce; Susan Nittrouer
Journal:  J Speech Lang Hear Res       Date:  2017-04-14       Impact factor: 2.297

7.  Recognizing spoken words: the neighborhood activation model.

Authors:  P A Luce; D B Pisoni
Journal:  Ear Hear       Date:  1998-02       Impact factor: 3.570

Review 8.  Pitch perception and auditory stream segregation: implications for hearing loss and cochlear implants.

Authors:  Andrew J Oxenham
Journal:  Trends Amplif       Date:  2008-10-30

9.  Multisensory perception of the six basic emotions is modulated by attentional instruction and unattended modality.

Authors:  Sachiko Takagi; Saori Hiramatsu; Ken-Ichi Tabei; Akihiro Tanaka
Journal:  Front Integr Neurosci       Date:  2015-02-02

10.  Multisensory emotion perception in congenitally, early, and late deaf CI users.

Authors:  Ineke Fengler; Elena Nava; Agnes K Villwock; Andreas Büchner; Thomas Lenarz; Brigitte Röder
Journal:  PLoS One       Date:  2017-10-12       Impact factor: 3.240

View more
  1 in total

1.  Perception of speaker sincerity in complex social interactions by cochlear implant users.

Authors:  Kathrin Rothermich; Susannah Dixon; Marti Weiner; Madison Capps; Lixue Dong; Sébastien Paquette; Ning Zhou
Journal:  PLoS One       Date:  2022-06-08       Impact factor: 3.752

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.