Literature DB >> 36097603

Interpersonal synchrony across vocal and lexical modalities in interactions involving children with autism spectrum disorder.

Rimita Lahiri1, Md Nasir2, Manoj Kumar3, So Hyun Kim4, Somer Bishop5, Catherine Lord6, Shrikanth Narayanan1.   

Abstract

Quantifying behavioral synchrony can inform clinical diagnosis, long-term monitoring, and individualised interventions in neuro-developmental disorders characterized by deficit in communication and social interaction, such as autism spectrum disorder. In this work, three different objective measures of interpersonal synchrony are evaluated across vocal and linguistic communication modalities. For vocal prosodic and spectral features, dynamic time warping distance and squared cosine distance of (feature-wise) complexity are used, and for lexical features, word mover's distance is applied to capture behavioral synchrony. It is shown that these interpersonal vocal and linguistic synchrony measures capture complementary information that helps in characterizing overall behavioral patterns.
© 2022 Author(s).

Entities:  

Year:  2022        PMID: 36097603      PMCID: PMC9462442          DOI: 10.1121/10.0013421

Source DB:  PubMed          Journal:  JASA Express Lett        ISSN: 2691-1191


Introduction

Autism spectrum disorder (ASD) (Baio, 2018; Maenner ) is a developmental condition primarily characterised by differences in social communication skills and abilities along with restricted repetitive behavior, interests, and movements. Prevalence of ASD in children in the United States has steadily risen from 1 in 150 in 2002 to 1 in 44 in 2021. ASD is a spectrum disorder with wide individual heterogeneity. Human-centered technological advances offer promise for supporting new evidence-driven possibilities in support of behavioral stratification as well as diagnosis and personalized treatment (Bone ; Narayanan and Georgiou, 2013). In the recent years, computational approaches using signal processing and machine learning have been proposed for both research and clinical translation in mental health (Bone ). Computational methods have shown promise in supporting the diagnostic efforts by identifying the essential nosological modules, eliminating redundancy without compromising on accuracy (Bone ; Bone ; Kumar ). Computational techniques have also provided tools to further scientific understanding of interaction mechanisms. For instance, Bone connects objective signal-derived descriptors of vocal prosody to subjective perceptions of prosodic awkwardness and reports differences in acoustic–prosodic features between ASD and neurotypical individuals, including demonstrating interaction pattern differences in vocal prosody coordination that varied in accordance with a child's ASD symptom severity (Bone ). Synchrony (Johnson and Jacob, 2000; Nasir ) in interactions can be broadly viewed as an individual's reciprocity, coordination, or adaptation to other participants(s) in, and during, the interaction. Typically, synchrony can be exhibited across multiple communication modalities, such as through vocal patterns (Nasir ), hand and head motions (Lee ; Xiao ), and facial expressions (Guha ). Since interpersonal synchrony in dyadic conversations provides insights toward understanding behavioral dynamics, it can potentially aid in scientific and clinical studies of interactions in the domain of ASD, which is characterized by differences in social communication and interaction (Bone ; Bone ). Prior work related to behavioral synchrony in the ASD domain is somewhat limited, focusing largely on individual modalities, such as vocal prosody or facial movements. Bone investigated synchrony in vocal arousal patterns in ASD child–clinician (adult) interactions, and showed its variation based on the child's ASD severity levels. To understand multimodal synchrony patterns, it is also important to consider the coordination and interplay between the communication modalities within an individual, in addition to across individuals. For example, Guha have reported the role of localized dynamics between different facial regions and their movements, and differences therein between typically developing children and children with high functioning autism. In this paper, we investigate three distinct measures of behavioral synchrony in speech and language patterns in an interaction based on DTW distance (Guha ), cosine distance (Nasir ; Yang and Narayanan, 2016), and word mover's distance (Nasir ). The primary contribution of this work is in quantifying synchrony across different information modalities related to voice, articulation, and language through the joint consideration of prosody, acoustic spectral features, and language patterns to capture interaction synchrony. Experiments performed on data from real world clinical interactions show that the proposed measures capture coordination in dyadic interactions. Since individuals with ASD exhibit wide differences in social communication, we believe that these coordination features can offer additional objective measures for behavior characterization and further stratification. Importantly, we experimentally investigate whether coordination features across the speech and language communication channels can capture complementary information that can be used as an additional source of information in characterizing ASD individuals and distinguishing them from those that have not received an ASD diagnosis. We analyze differences in the synchrony measures across children with and without an ASD determination through post hoc classification experiments. Classification experiments carried out with the three proposed coordination features reveal their importance in differentiating ASD and non-ASD groups through improved (F1 score) performance with respect to baseline classifiers. Furthermore, we analyse the variation of the mean value of the proposed coordination measures throughout the interactions across two different subtasks for both children diagnosed with ASD and without ASD. We also examine age-dependency in these results through two-way analysis of variance of the classification F1 scores computed across three different age groups of young (2.5–7.5 y), middle-band (7.5–10 y), and older children (above 10 y), as well as male and female children.

Dataset description

The vocal and language behavioral synchrony measures are evaluated in the context of interactions between a child and a clinician. The data are drawn from two specific domains (Emotions and Social difficulties and annoyance subtasks) involving behavioral observation. The Autism Diagnostic Observation Schedule (ADOS-2) (Lord ) instrument refers to semi-structured interactions between a child and a clinician trained to score the different behaviors associated with ASD. These interactive sessions are typically 40–60 min long and broken down into a variety of subtasks (e.g., construction, joint-interactive play, creating a story, demonstration, etc.) which are likely to evoke prominent response from a child under different social circumstances. Based on the child's response, the clinician provides assessment of ASD symptoms following module-specific coding and finally, all these codes are aggregated to compute an autism severity score (Gotham ). For this study, we focus on a subset of data from the administrations of module-3 meant for verbally fluent children. Specifically, we choose to work with Emotions and Social difficulties and annoyance subtasks because of their ability to elicit spontaneous speech from the children under significant cognitive demand. The dataset consists of recordings from 165 children (86 ASD, 79 non-ASD), collected from two different clinics: the University of Michigan Autism and Communication Disorders Center (UMACC) and the Cincinnati Children's Medical Center (CCHMC). For our experiments, we have 1 recording for each of the mentioned subtasks from each participant resulting in 330 recordings. The demographic details are presented in Table 1. The average duration of each session is about 3 min (192 s). The lexical features are extracted based on manual transcriptions following SALT (Miller and Iglesias, 2012) guidelines. Since we aggregate turn-level coordination measures, sessions with fewer than 10 turns are discarded as a sufficient number of turns is required to aggregate and average out local irregularities.
Table 1.

Demographic details of ADOS dataset.

CategoryStatistics
Age (y)Range: 3.58–13.17 (mean, std): (8.61, 2.49)
Gender123 male, 42 female
Non-verbal IQRange: 47–141 (mean, std): (96.01, 18.79)
Clinical diagnosis86 ASD, 42 ADHD
14 mood/anxiety disorder
12 language disorder
10 intellectual disability, 1 no diagnosis
Age distribution clinicwiseCincinnati 84 (2.5–7.5 y: 28, 7.5–10 y : 31, 10 y: 25)
Michigan 81 (2.5–7.5 y: 24, 7.5–10 y: 30, 10 y: 27)
Age distribution ASD/Non-ASDASD 86 (2.5–7.5 y: 25, 7.5–10 y : 30, 10 y: 31)
Non-ASD 79 (2.5–7.5 y: 27, 7.5–10 y: 31, 10 y: 21)
Demographic details of ADOS dataset.

Quantification of interpersonal synchrony

In this section, we describe the different signal feature descriptors used and the proposed coordination measures in detail. First, we outline the feature descriptors and based on those features, we define the coordination measures. For this study, we consider three different sets of feature descriptors: vocal prosodic features, acoustic spectral features, and lexical features. We use DTWD and SCDC for vocal prosodic and acoustic spectral features to quantify interpersonal synchrony. For lexical features, we apply WMD to capture behavioral synchrony.

Acoustic spectral and vocal prosodic features

All the features are extracted using the OpenSMILE toolkit (Eyben ). The feature extraction is carried out using a sliding Hamming window of 25 ms duration with an interval of 10 ms. We use 15 dimensional Mel frequency cepstral coefficients (MFCC) (Yang and Narayanan, 2016) as acoustic spectral features and pitch, intensity, jitter, and shimmer as vocal prosodic features. The prosodic features are smoothed and extrapolated over the unvoiced regions.

DTW distance measure (DTWD)

We use the classic DTW (Myers and Rabiner, 1981) method to measure the similarity between the acoustic–prosodic features extracted from two consecutive speaker turns. This method computes the (dis)similarity between two time sequences, of possibly varying lengths, after aligning them to the maximum extent in terms of a warping distance. Herein, we employ the DTW method to compute the dissimilarity between vocal feature time series obtained from the child and clinician's speech turn-pairs. We introduce the average warping distance as a measure for interpersonal synchrony. For two m dimensional time series X and Y with length T and T, respectively, such that and , DTW finds the (dis)similarity between these sequences by optimally aligning them. A distance matrix is calculated where every element d(i, j) denotes the Euclidean distance between the ith vector of X and jth vector of Y. Based on the distance matrix values, an optimal warping path yielding the overall minimum cost (distance) is found. A warping path is a mapping from X to Y tracing the elements of W where every element is such that , i.e., X can be warped to the same length as Y by corresponding ith element of X to jth element of Y. An optimal path is the one associated with the minimum cost where the cost is computed as the sum of absolute distances for every matched pair of indices present in the path, Since it is a dissimilarity measure, a larger warping distance is deemed to signify lesser coordination or synchrony.

Squared cosine distance of complexity measure (SCDC)

Prior work (Nasir ) has attempted to measure coordination between speakers in a dyadic conversation from a nonlinear dynamical system approach based on the underlying model's complexity. Following a similar approach, we analyze the complexity pattern underlying the signals observed in dyadic conversations by framing them as arising from a coupled nonlinear dynamical system. However, while Nasir relies on computing coordination in different features separately, we measure the coordination between speakers as a whole based on all the audio features considered. We capture the difference between vocal characteristics of the speaker (turns) by comparing the complexity underlying their prosodic and spectral feature values. We use sample entropy as the complexity measure. It is an information-theoretic measure of complexity which signifies the amount of new information introduced across a temporal sequence. Based on our definition of complexity, we hypothesize that local changes of complexity in a well-coordinated conversation will be smaller than a less well-coordinated conversation. More specifically, the distance between complexity patterns corresponding to consecutive turns should be expected to be lower in a well-coordinated conversation as compared to a randomly generated conversation. As will be shown in Sec. 4, the experimental findings are found to support this hypothesis, and confirm that a larger value of the proposed measure corresponds to lower synchrony. To calculate the proposed measure, for a time sequence of length M such that , a length m subsequence is formed as . Let denote the Chebyshev distance between any two such vector pairs where . Now, if denotes the number of vector pairs such that where r is a predefined threshold then sample entropy S is defined as From the definition, is always less than or equal to , so S is non-negative. Smaller values of S signify greater self-similarity across the values. Here, we consider, m = 2 and the standard deviation of the time series. For any two consecutive pairs of turns, first, we compute the sample entropy values for every feature, yielding two vectors X1 and X2 consisting of feature-wise complexity values corresponding to the two turns. Next, we calculate the synchrony measure σ as We hypothesize the difference between sample entropy values corresponding to a turn-pair in a well-coordinated conversation will be smaller as compared to a less well-coordinated conversation. σ captures this information and is therefore introduced as a coordination measure.

Lexical features

A word embedding is a contextually derived numerical representation in a vector space for every word. In this study, we use 768 dimensional bidirectional encoder representations from transformer (BERT) (Devlin ) embeddings as lexical features. We choose BERT embeddings since these embeddings incorporate context which handles polysemy and nuances better. Here, we extract BERT embeddings (BERT BASE model with 12 transformer blocks, 12 attention heads, and hidden layer dimension 768) corresponding to each word to form the feature matrix corresponding to every turn, where each row of the matrix is an embedding corresponding to one word of the turn. Once the feature matrices corresponding to the speaker turns are obtained, WMD is computed between these feature matrices corresponding to consecutive turn-pairs.

Word mover's distance (WMD)

Word mover's distance (WMD) was introduced by Kusner as a similarity measure between two text documents by comparing between word2vec (Mikolov ) neural word embeddings. It is calculated as the optimum distance the embedded words in one document needs to travel to reach the embedded words in the other document. It can be considered as a special case of the popular transportation problem of Earth mover's distance (Rubner ). If b and b are the BERT embeddings corresponding to words w and w, respectively, the distance between these words can be defined as If X1 and X2 are the matrices corresponding to two turns of Speaker A and Speaker B, respectively, the WMD between the turns can be expressed as constrained on , , where m and n are the number of words in turns X1 and X2, respectively, and T is the associated weight. Since we are working with contextual embeddings, the same word in two different positions will have distinct embeddings, so the weights are chosen to be uniform for WMD calculation. Similar to the previously introduced measures, it is hypothesized that a conversation with greater synchrony is likely to have a smaller average WMD compared to a less synchronized conversation.

Empirical validation of the proposed synchrony measures

In this work, the DTWD and SCDC of prosodic and spectral features are introduced as interaction coordination measures. While these methods have been previously employed to capture (dis)similarity between different time series, they have not been used together in the context of quantifying interpersonal synchrony from speech audio. It should be noted that prior work (Nasir ) has established the potential of WMD in capturing linguistic coordination based on word embeddings in dyadic conversation setting. Since the use of WMD as a viable measure for interaction coordination has already been validated in previous work, in this section, we only consider validating the usefulness of measures based on DTWD and SCDC for characterising synchrony in dyadic interactions. We use the USC CreativeIT database (Metallinou ; the details of the dataset are described in the supplementary material1) for these analysis experiments. It is a publicly available multimodal database consisting of dyadic conversations portraying theatrical improvisations. We generate approximately 2500 regular (“real”) and random turn-pairs from these interactions. Any consecutive pair of turns from the same interaction is considered as a regular turn-pair while the random pairs are generated by arbitrarily choosing two turns from two different interaction sessions not involving same speakers. Our hypothesis is that for any feature set, the synchrony should be higher in the actual turn-pairs when compared to randomly shuffled pairs. For validating our hypothesis, we design a paired t-test to compare the coordination measures across the actual and random pairs with a null hypothesis being the sample mean of those two time series are equal. The paired sample t-test result shows that the prosodic feature synchrony is significantly higher in real turn-pairs compared to the random turn-pairs, based on DTWD (F-statistic = 4.277, p value = 0.000 01) and SCDC (F-statistic = 3.705, p value = 0.000 21). A similar trend is seen for synchrony of acoustic segmental features in terms of both DTWD (F-statistic = 2.515, p value = 0.0119) and SCDC (F-statistic = 3.705, p value = 0.0002).

Experimental results on ASD interaction datasets

In this section, we report and analyze the results of the different experiments conducted to understand the differences in the proposed synchrony measures in interactions involving children with and without an ASD diagnosis.

Classification experiment

The experiments in this subsection explore whether the proposed measures of interaction synchrony reveal differences between interactions involving children with an ASD diagnosis and those that do not have one. This is set up as a series of classification experiments aimed at how well the children diagnosed with and without ASD can be distinguished using the proposed synchrony measures. For the experiments, each interaction session is partitioned into child and adult speaker turns. Once the speaker turns are defined, the turns are collected into N segments and the coordination measures are calculated for every such segment. Hence, for each interactive session, the classifier will be input with N features to predict the ASD or non-ASD output label. The experiments consider three classifiers, all well established in the literature: (i) support vector machine (SVM) with linear kernel, (ii) SVM with radial basis function (RBF) kernel, and (iii) logistic regression. We also consider a classifier predicting the majority class as the baseline classifier. The classification experimental results with individual modalities are tabulated in Table 2, while those with fused features are reported in Fig. 1. We use early fusion for concatenating the features from individual modalities before feeding to the classifier.
Table 2.

F1-score for ASD diagnosis.

ClassifiersAudio featuresLexical features
Spectral featuresProsodic featuresWord embeddings
DTWDSCDCDTWDSCDCWMD
Majority Classifier0.3446
SVM (RBF)0.55540.53450.51500.34280.5060
SVM (Linear)0.53150.34460.54170.34480.4394
Logistic Regression0.54240.34460.48650.35240.4861
Fig. 1.

F1-score for ASD diagnosis with fused features.

F1-score for ASD diagnosis. F1-score for ASD diagnosis with fused features.

ANOVA analysis based on age group and gender

In addition to reporting the F1 scores for ASD/non-ASD classification experiments with the different coordination features, we also carry out a two-way analysis of variance of classification F1 scores across three age groups and gender. We partition the data into 3 different age groups (2.5–7.5 y, 7.5–10 y, y) to gain insights into the synchrony features across different age groups and gender within children diagnosed with ASD and children who are not. Figure 2 presents a box plot showing the median, maximum, minimum, 75% percentile, and 25% percentile values of F1 scores for each age group and gender.
Fig. 2.

Two-way analysis of variance of F1 scores across age-groups and gender.

Two-way analysis of variance of F1 scores across age-groups and gender.

Comparison of the distribution of the proposed measures across different subtasks

We also report the mean value of these coordination measures for children diagnosed with and without ASD, across the two subtasks. The comparison of the distribution of these values is shown in Fig. 3. We calculate the mean value of the coordination measures for the turns collected in five segments and plot those values across the corresponding segments.
Fig. 3.

Comparison of different coordination measures across subtasks.

Comparison of different coordination measures across subtasks.

Discussion

From Table 2, we observe that in the case of both of spectral and prosodic features, all classifiers yield better performance with DTWD based synchrony features compared to SCDC based features. Comparing the results from Table 2 and Fig. 1, it can be seen that fusing the synchrony features across modalities improves the classification performance over using individual modality features. Among the fused feature-based experiments reported in Fig. 1, prosodic and spectral features together show the best performance, which indicates that there is complementary information across these modalities helpful for this classification task. While all the classifiers considered provide similar performance levels, the variants of SVM with radial basis function kernel appears to be the most consistent across the experiments. For the age group–based analysis reported in Fig. 2, the p value = 0.000 196 suggests that the null hypothesis can be rejected implying that age difference and gender both significantly affect the classifier F1 scores for differentiating between children with and without ASD diagnosis based on the proposed coordination measures. Moreover, we can also find an improvement in ASD/non-ASD classification performance amongst female children in the oldest age group as compared to the other age groups. This finding stands consistent with the investigation presented in the prior work (Sturrock ) and motivates us to seek more insight to why females are more likely to go undiagnosed than males until an older age. Figure 3 presents variation of the mean of the proposed measures across two different subtasks and ASD status. In most cases, higher mean value of these proposed measures is reported for children with an ASD diagnosis in both social and emotion subtasks. This finding indicates that children with an ASD diagnosis exhibit less synchrony in terms of these measures as compared to the children without ASD diagnosis. It is interesting to note that for all the children, emotion subtask is shown to have less synchrony for most of the duration of the session as revealed by all the measures.

Conclusion

Previous behavioral science research has established the importance of interpersonal synchrony in understanding behavior patterns in human interaction. In this work, we propose three different measures of synchrony across different aspects of speech communication (vocal acoustics, prosody, and language use). To investigate whether these synchrony features offer insights into potential differences in interaction patterns involving children diagnosed with ASD and those that do not, we set up a classification experiment utilizing the synchrony features. Results show that the proposed synchrony features can distinguish interactions involving ASD and non-ASD children indicating the role of coordination as an element of difference in social communication patterns. Moreover, the analysis shows that the synchrony features across different information modalities of spoken interactions captured by spectral features, prosodic features, and language patterns provide complementary information distinguishing the two groups: children with ASD diagnosis and without ASD diagnosis. There are several challenging research directions to explore in future. We plan to investigate more data driven approaches to quantify synchrony instead of knowledge driven approaches. Knowing that neural networks can efficiently learn non-linear mappings between feature and coordination measures, future works can explore usage of deep neural network–based models to learn representations related to synchrony.
  12 in total

1.  The autism diagnostic observation schedule-generic: a standard measure of social and communication deficits associated with the spectrum of autism.

Authors:  C Lord; S Risi; L Lambrecht; E H Cook; B L Leventhal; P C DiLavore; A Pickles; M Rutter
Journal:  J Autism Dev Disord       Date:  2000-06

2.  Sequential interactions in the marital communication of depressed men and women.

Authors:  S L Johnson; T Jacob
Journal:  J Consult Clin Psychol       Date:  2000-02

3.  Applying machine learning to facilitate autism diagnostics: pitfalls and promises.

Authors:  Daniel Bone; Matthew S Goodwin; Matthew P Black; Chi-Chun Lee; Kartik Audhkhasi; Shrikanth Narayanan
Journal:  J Autism Dev Disord       Date:  2015-05

4.  Standardizing ADOS scores for a measure of severity in autism spectrum disorders.

Authors:  Katherine Gotham; Andrew Pickles; Catherine Lord
Journal:  J Autism Dev Disord       Date:  2008-12-12

5.  Use of machine learning to improve autism screening and diagnostic instruments: effectiveness, efficiency, and multi-instrument fusion.

Authors:  Daniel Bone; Somer L Bishop; Matthew P Black; Matthew S Goodwin; Catherine Lord; Shrikanth S Narayanan
Journal:  J Child Psychol Psychiatry       Date:  2016-04-19       Impact factor: 8.982

6.  Behavioral Signal Processing: Deriving Human Behavioral Informatics From Speech and Language: Computational techniques are presented to analyze and model expressed and perceived human behavior-variedly characterized as typical, atypical, distressed, and disordered-from speech and language cues and their applications in health, commerce, education, and beyond.

Authors:  Shrikanth Narayanan; Panayiotis G Georgiou
Journal:  Proc IEEE Inst Electr Electron Eng       Date:  2013-02-07       Impact factor: 10.961

7.  QUANTIFYING ATYPICALITY IN AFFECTIVE FACIAL EXPRESSIONS OF CHILDREN WITH AUTISM SPECTRUM DISORDERS.

Authors:  Angeliki Metallinou; Ruth B Grossman; Shrikanth Narayanan
Journal:  Proc (IEEE Int Conf Multimed Expo)       Date:  2013

8.  Prevalence of Autism Spectrum Disorder Among Children Aged 8 Years - Autism and Developmental Disabilities Monitoring Network, 11 Sites, United States, 2014.

Authors:  Jon Baio; Lisa Wiggins; Deborah L Christensen; Matthew J Maenner; Julie Daniels; Zachary Warren; Margaret Kurzius-Spencer; Walter Zahorodny; Cordelia Robinson Rosenberg; Tiffany White; Maureen S Durkin; Pamela Imm; Loizos Nikolaou; Marshalyn Yeargin-Allsopp; Li-Ching Lee; Rebecca Harrington; Maya Lopez; Robert T Fitzgerald; Amy Hewitt; Sydney Pettygrove; John N Constantino; Alison Vehorn; Josephine Shenouda; Jennifer Hall-Lande; Kim Van Naarden Braun; Nicole F Dowling
Journal:  MMWR Surveill Summ       Date:  2018-04-27

9.  Speaking the Same Language? A Preliminary Investigation, Comparing the Language and Communication Skills of Females and Males with High-Functioning Autism.

Authors:  Alexandra Sturrock; Natalie Yau; Jenny Freed; Catherine Adams
Journal:  J Autism Dev Disord       Date:  2020-05

10.  Prevalence and Characteristics of Autism Spectrum Disorder Among Children Aged 8 Years - Autism and Developmental Disabilities Monitoring Network, 11 Sites, United States, 2018.

Authors:  Matthew J Maenner; Kelly A Shaw; Amanda V Bakian; Deborah A Bilder; Maureen S Durkin; Amy Esler; Sarah M Furnier; Libby Hallas; Jennifer Hall-Lande; Allison Hudson; Michelle M Hughes; Mary Patrick; Karen Pierce; Jenny N Poynter; Angelica Salinas; Josephine Shenouda; Alison Vehorn; Zachary Warren; John N Constantino; Monica DiRienzo; Robert T Fitzgerald; Andrea Grzybowski; Margaret H Spivey; Sydney Pettygrove; Walter Zahorodny; Akilah Ali; Jennifer G Andrews; Thaer Baroud; Johanna Gutierrez; Amy Hewitt; Li-Ching Lee; Maya Lopez; Kristen Clancy Mancilla; Dedria McArthur; Yvette D Schwenk; Anita Washington; Susan Williams; Mary E Cogswell
Journal:  MMWR Surveill Summ       Date:  2021-12-03
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.