| Literature DB >> 35675356 |
Kathrin Rothermich1, Susannah Dixon1, Marti Weiner1, Madison Capps1, Lixue Dong1, Sébastien Paquette2, Ning Zhou1.
Abstract
Understanding insincere language (sarcasm and teasing) is a fundamental part of communication and crucial for maintaining social relationships. This can be a challenging task for cochlear implant (CIs) users who receive degraded suprasegmental information important for perceiving a speaker's attitude. We measured the perception of speaker sincerity (literal positive, literal negative, sarcasm, and teasing) in 16 adults with CIs using an established video inventory. Participants were presented with audio-only and audio-visual social interactions between two people with and without supporting verbal context. They were instructed to describe the content of the conversation and answer whether the speakers meant what they said. Results showed that subjects could not always identify speaker sincerity, even when the content of the conversation was perfectly understood. This deficit was greater for perceiving insincere relative to sincere utterances. Performance improved when additional visual cues or verbal context cues were provided. Subjects who were better at perceiving the content of the interactions in the audio-only condition benefited more from having additional visual cues for judging the speaker's sincerity, suggesting that the two modalities compete for cognitive recourses. Perception of content also did not correlate with perception of speaker sincerity, suggesting that what was said vs. how it was said were perceived using unrelated segmental versus suprasegmental cues. Our results further showed that subjects who had access to lower-order resolved harmonic information provided by hearing aids in the contralateral ear identified speaker sincerity better than those who used implants alone. These results suggest that measuring speech recognition alone in CI users does not fully describe the outcome. Our findings stress the importance of measuring social communication functions in people with CIs.Entities:
Mesh:
Year: 2022 PMID: 35675356 PMCID: PMC9176755 DOI: 10.1371/journal.pone.0269652
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.752
Example scenes used in the experiment.
| Scene “Wedding” (with verbal context) | Scene “Party” (no verbal context) | ||
|---|---|---|---|
| Literal positive | Lisa on the phone:… I’m looking forward already! I’ll call you later! | Literal positive | Lisa: Do you think the party was a success? |
| Paul: Are you gonna come with me to Sarah’s wedding? | Paul: Yeah, I had a great time! | ||
| Lisa: Yeah, it is gonna be fun! | |||
| Literal negative | Lisa on the phone:… and I don’t really wanna go there, anyways. I’ll call you later! | Literal negative | Lisa: Do you think the party was a success? |
| Paul: Are you gonna come with me to Sarah’s wedding? | Paul: No, no one had fun. | ||
| Lisa: No, weddings aren’t really my thing. | |||
| Sarcasm | Lisa on the phone:… and I don’t really wanna go there, anyways. I’ll call you later! | Sarcasm | Lisa: Do you think the party was a success? |
| Paul: Are you gonna come with me to Sarah’s wedding? | Paul (sarcastic): Yeah, I had a great time! | ||
| Lisa (sarcastic): Yeah, it is gonna be fun! | |||
| Teasing | Lisa on the phone:… I’m looking forward already! I’ll call you later! | Teasing | Lisa: Do you think the party was a success? |
| Paul: Are you gonna come with me to Sarah’s wedding? | Paul: No, no one had fun (laughs). | ||
| Lisa: No, weddings aren’t really my thing (laughs). | |||
Demographic information.
| Subject | Ear | Gender | Age | CI use (yrs) | Hearing Loss Duration | Implant Type | Processor Type | Speech processing strategy | Device type |
|---|---|---|---|---|---|---|---|---|---|
| DB1 | L | M | 80.35 | 17.27 | 13.06 | CI24R (CS) | CP910 | ACE | Bilateral CI |
| DB1 | R | M | 80.35 | 11.30 | 19.04 | CI24RE (CA) | CP810 | ACE | Bilateral CI |
| CJ3 | L | F | 68.99 | 12.98 | 15.35 | CI24RE (CA) | CP920 | ACE | Bilateral CI |
| CJ3 | R | F | 68.99 | 14.50 | 13.83 | CI24RE (CA) | CP920 | ACE | Bilateral CI |
| NM4 | L | F | 60.15 | 7.77 | 22.38 | CI24RE (CA) | CP810 | ACE | Unilateral CI |
| SG7 | R | F | 73.69 | 8.50 | 34.83 | CR24RE (CA) | CP810 | ACE | Bimodal |
| TH16 | L | M | 57.69 | 12.60 | 45.09 | CI24RE (CA) | CP810 | ACE | Bilateral CI |
| TH16 | R | M | 57.69 | 10.68 | 47.01 | CI24RE (CA) | CP810 | ACE | Bilateral CI |
| SP18 | L | F | 67.43 | 4.71 | 33.63 | CI422 | CP910 | ACE | Bimodal |
| FP19 | L | F | 72.64 | 12.00 | 44.33 | CI24RE (CA) | CP1000 | ACE | Unilateral CI |
| DA22 | R | F | 74.47 | 6.94 | 17.39 | CI4RE (CA) | CP920 | ACE | Unilateral CI |
| KS25 | L | F | 62.18 | 11.67 | 18.67 | CI24RE (CA) | CP900 | ACE | Bilateral CI |
| KS25 | R | F | 62.18 | 10.92 | 19.42 | CI24RE (CA) | CP900 | ACE | Bilateral CI |
| DK27 | L | M | 59.76 | 12.33 | 42.43 | CI512 | CP920 | ACE | Bilateral CI |
| DK27 | R | M | 59.76 | 13.33 | 41.43 | CI24RE (CA) | CP920 | ACE | Bilateral CI |
| KK28 | R | F | 76.64 | 12.58 | 58.06 | HR90K | Naida CI Q70 | HiRes Optima-S | Unilateral CI |
| CP30 | L | F | 60.77 | 6.91 | 53.86 | HiFocus 1J | Naida CI Q90 | HiRes Optima-S | Bilateral CI |
| CP30 | R | F | 60.77 | 12.68 | 48.09 | HiFocus 1J | Naida CI Q90 | HiRes Optima-S | Bilateral CI |
| SS31 | L | M | 69.71 | 3.82 | 26.51 | CI422 | Kanso | ACE | Unilateral CI |
| RS32 | L | M | 68.88 | 2.51 | 47.36 | HiFocus ms | Naida CI Q90 | HiRes Optima-S | Bimodal |
| PV33 | R | F | 68.93 | 2.13 | 62.20 | HiFocus 1J | Naida CI Q90 | HiRes Optima Paired | Unilateral CI |
| BB34 | L | F | 72.68 | 1.59 | 14.74 | HiFocus ms | Naida CI Q90 | HiRes Optima-S | Bimodal |
| BB35 | L | M | 77.66 | 1.60 | 21.06 | HiFocus ms | Naida CI Q90 | HiRes Optima-P | Bimodal |
Fig 1Performance for individual subjects.
Blue bars represent performance in the audio-only condition. The red bars represent the audio-visual condition. The bars represent the percentages of correctly perceived conversations in terms of the content question (dark-colored bars) and identifying speaker sincerity (light-colored bars) relative to the total number of stimuli.
Fig 2Box plot of perception of speaker’s intentions collapsed across subjects.
The top panel shows performance with context, and the bottom panel shows performance without context. Percent correct scores were calculated based on the number of correctly perceived intentions relative to the number of correctly perceived contents. Bars represent group medium. Asterisks show outliers.
Fig 3Left panel: The relationship between the advantage of visual cues in identifying speakers’ sincerity and perception of content in the audio-only condition.
Right panel: Correlation between the perception of content in the audio-only condition and duration of hearing loss.