Literature DB >> 24740236

The potential of accelerating early detection of autism through content analysis of YouTube videos.

Vincent A Fusaro1, Jena Daniels2, Marlena Duda2, Todd F DeLuca1, Olivia D'Angelo1, Jenna Tamburello1, James Maniscalco1, Dennis P Wall3.   

Abstract

Autism is on the rise, with 1 in 88 children receiving a diagnosis in the United States, yet the process for diagnosis remains cumbersome and time consuming. Research has shown that home videos of children can help increase the accuracy of diagnosis. However the use of videos in the diagnostic process is uncommon. In the present study, we assessed the feasibility of applying a gold-standard diagnostic instrument to brief and unstructured home videos and tested whether video analysis can enable more rapid detection of the core features of autism outside of clinical environments. We collected 100 public videos from YouTube of children ages 1-15 with either a self-reported diagnosis of an ASD (N = 45) or not (N = 55). Four non-clinical raters independently scored all videos using one of the most widely adopted tools for behavioral diagnosis of autism, the Autism Diagnostic Observation Schedule-Generic (ADOS). The classification accuracy was 96.8%, with 94.1% sensitivity and 100% specificity, the inter-rater correlation for the behavioral domains on the ADOS was 0.88, and the diagnoses matched a trained clinician in all but 3 of 22 randomly selected video cases. Despite the diversity of videos and non-clinical raters, our results indicate that it is possible to achieve high classification accuracy, sensitivity, and specificity as well as clinically acceptable inter-rater reliability with nonclinical personnel. Our results also demonstrate the potential for video-based detection of autism in short, unstructured home videos and further suggests that at least a percentage of the effort associated with detection and monitoring of autism may be mobilized and moved outside of traditional clinical environments.

Entities:  

Mesh:

Year:  2014        PMID: 24740236      PMCID: PMC3989176          DOI: 10.1371/journal.pone.0093533

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Background

In the United States, autism spectrum disorder (ASD) is typically diagnosed in children at four years of age with an estimated 27% undiagnosed by eight years of age [1]. Yet, the majority of parents of ASD children report developmental concerns before one year of age [1]–[3]. To better understand the developmental progression of ASD and identify distinguishing behaviors between ASD and non-ASD children, a number of studies focused on retrospective home video analysis [4]–[12]. Previous home video studies showed that ASD and non-ASD children differ in frequency of responding to name [10], gaze [13], smiling [14], and stereotypic motor behaviors [15]. It is important to note that home videos generally will not reflect all aspects of the clinical evaluation and can present challenges to proper diagnosis such as not clearly capturing eye contact, capturing context-dependent behaviors that may not exemplify routine, and poorly demonstrating social behavior with unfamiliar person(s). Nevertheless, home videos are considered a more accurate representation of early events than parental recall [10], and could be of value in future efforts focused on lowering the average age of diagnosis in the Unites States and abroad. Significant prior work has focused on creating structured questions and standardized approaches to evaluate home videos by retrospectively examining early videos of children later diagnosed with ASD. This allows researchers to combine additional clinical features, such as age and IQ, with the video to develop a more complete representation of behavioral development. To standardize the home videos, most research focuses on a common event such as first or second birthday, and/or pivotal developmental milestones (such as not pointing, not responding to name calling, or not playing [stereotypically or pretend]). We sought to expand the concept and feasibility of home video analysis by applying the Autism Diagnostic Observation Schedule (ADOS) [16] questions, not the full exam, to ASD and non-ASD videos from YouTube, the largest public video-sharing Internet website. The ADOS is widely considered a gold standard and is one of the most common behavioral instruments used to aid in diagnosis of ASD [16]–[18]. Clinical application of the ADOS requires a clinician trained in administrating the exam in order to elicit specific types of responses during structured activities. The exam is divided into four modules that correspond to an individual’s language and developmental level, ensuring coverage for a wide variety of behavioral manifestations. We used module 1, which contains 10 activities and 29 questions, for assessment because this module is largely focused on the behavioral characteristics most appropriate for characterizing the younger children found in a majority of the YouTube videos collected for this study, and also because this age group serves to benefit most from early detection and early intervention. Our aims were first to test the feasibility of answering the ADOS module 1 questions when viewing short (<10 min) unstructured videos, and second to test the accuracy of non-clinical raters to distinguish videos containing children with autism from videos of children who have no signs of autism.

Methods

Human Subjects

The use of videos from YouTube was approved for exemption by the Institutional Review Board at Beth Israel Deaconess Medical Center (protocol #2012P-000307), under exemption number 4 (“research involving the collection or study of existing data, documents, records, pathological specimens, or diagnostic specimens, if these sources are publicly available or if the information is recorded by the investigator in such a manner that subjects cannot be identified, directly or through identifiers linked to the subjects”) of the Code of Federal Regulations, 45 CFR 46.101(b).

Video Selection

We searched YouTube (www.youtube.com) for videos of ASD and non-ASD children using the following keywords in the video title or description: autism, ASD, autism spectrum disorder, Asperger’s, autistic child, autistic kid, autistic behavior, PDD-NOS, son, daughter, child, birthday party. We required each video to be under 10 minutes long, have clear audible sound quality, have continuous activity, and to contain a majority of footage on a child between the ages of 1 and 15. When age was not provided it was estimated based on a consensus from two or more investigators (pertinent to 7 of the 100 videos collected). In total, we identified 100 unique subject videos, 45 ASD and 55 non-ASD. We identified non-ASD videos as controls by matching age, gender, and race as closely as possible. Videos were labeled as ASD if the video title, video description, or meta-tag included autism, ASD, Asperger’s, or hand-flapping/stimming; otherwise, the video was labeled non-ASD. The full list of videos is provided in .

Raters and ADOS Scoring

ADOS scoring instruments were purchased from Western Psychological Services (http://www.wpspublish.com/). Four non-clinical raters independently scored the 29 questions on ADOS module 1 for each video. Raters were purposely given minimal instructions to code only when the video clearly depicted a behavior and/or contained opportunities for the child to exhibit the behavior in question, and otherwise to code the behavioral item as not applicable (N/A). Two of the four raters participated in the acquisition of the home videos from YouTube, and thus were not naïve to videos titles. An ADOS-certified clinical practitioner performed a blinded evaluation of a random subset of 22 videos (13 ASD and 8 non-ASD) that we used as a clinical standard for comparison to outcomes from the four non-clinical raters. We calculated the typical ADOS score as indicated in the exam instructions for the ADOS algorithm and followed the convention of converting all scores of three to two [19]. Videos were labeled ASD if the score was ≥7 and otherwise labeled as non-ASD. Any item that was scored N/A was removed from scoring and statistical analysis.

Statistical Analysis

We computed intra-class correlations across pairs of raters for the communication and social domains and overall for the ADOS module 1. We computed the mean item-level agreement across all raters and compared the mean item-level agreement between the four non-clinical raters to the clinical rater. We calculated the accuracy, sensitivity and specificity for classifications only if the majority of raters agreed (3 out of 4). For videos where there was no majority (2 out 4) we did not make a classification and labeled it “no call.” We performed all such analyses in R [20],[21] using the “IRR” package.

Results

We identified and analyzed 100 videos on YouTube that met our criteria for ASD (45 videos) and non-ASD (55 videos). A complete summary of video characteristics is provided in . The average age of children with ASD was 4.35 and 2.89 for children without an ASD. The age of the child was available in the metadata associated with the video in all but 7 of the 100 videos. For these 7, age was estimated by 2 raters independently. We found more male videos (n = 37) than female (n = 18) for ASD, which is consistent with the higher percentage of autism found in males. The most common types of video were of a child playing inside with his/her collection of toys. Nearly all videos (91%) included interaction with an adult.
Table 1

Characteristics of ASD and Non-ASD videos on YouTube.

CharacteristicASD (n = 55)Non-ASD (n = 45)
Gendera, % (M/F)37/1823/22
Agea, mean (range)4.35 (1.5–15)2.89 (0.92–6)
Racea
White, %89.182.2
Black, %1.83.6
Other, %9.013.3
Video length (min), mean (range)3∶19 (0∶31–9∶16)2∶37 (0∶36–6∶09)
YouTube viewsb, mean (range)17,661 (17–96,469)4,107,244 (146–102,871,735)
Date posted on YouTube, range03/21/07–06/19/1206/25/06–11/04/12
Appearance of other peoplec, %
One person20.033.3
Two or more people18.217.8
Interacting with adult, %96.495.6
Interacting with peer, %16.417.8
Types of videosd, %
Exhibiting a talent5.533.3
Having a conversation60.080.0
Playing49.042.2
Inside85.577.8
Outside9.013.3
In a car0.06.7
Party or birthday1.88.9
Eating10.922.2

If not explicitly stated in metadata associated with the video value was estimated based on consensus of two or more investigators.

The number of views on YouTube accessed on 12/10/13.

Number of additional people in the video excluding the child and person recording the video.

Videos were broadly categorized, after initial identification, to illustrate the diversity of videos evaluated.

If not explicitly stated in metadata associated with the video value was estimated based on consensus of two or more investigators. The number of views on YouTube accessed on 12/10/13. Number of additional people in the video excluding the child and person recording the video. Videos were broadly categorized, after initial identification, to illustrate the diversity of videos evaluated. The high classification accuracy (96.8%) supported the feasibility of applying ADOS module 1 questions to videos on YouTube ( ). In addition, the high sensitivity (94.1%) and perfect specificity suggested that home video classification could yield both high positive and negative predictive value. While all behavioral items on the ADOS could be addressed by at least 1 rater in all 100 videos, 7 behaviors, including imagination and functional play, had 30% or higher N/A answer codes indicating that they were not adequately captured in the videos ( ). The intra-class correlation among raters for specific questions on ADOS module 1 were consistent with previous results [22] and ranged from 0.83–0.88. The mean rater agreement was high (73.3%) despite the diversity of environmental contexts represented in the videos. The lowest mean rater agreement was 58.1% for item B9 (“showing”) possibly due in part to the fact that it had the a relatively high frequency of N/A codes across the raters ( ). The highest mean rater agreement was 95.3% for item D3 (“self-injurious behavior”) as the children in the videos rarely exhibited this behavior. The distribution of raters’ scores for all videos is shown in which provides additional insight into the classification accuracy and inter-rater agreement.
Table 2

Video Scoring Performance.

CharacteristicTotal (n = 100)
Accuracy, %96.8
Sensitivity, %94.1
Specificity, %100
No call (ASD/non-ASD)a 4/1
Interrater classification agreement (ASD vs. non-ASD), % mean (range)92.2 (89–95)
Intraclass correlation across items
Communication0.84
Social0.83
Total0.88
Rater item-level mean agreementb, % mean (range)73.3 (58.1–95.3)
Rater vs. Expert item-level mean agreementc, % mean (range)71.3 (37.5–100)

We required majority rater agreement (3 out of 4) for classification.

Item-level agreement shown is the mean and range of 29 item agreements across four non-clinical raters.

Clinical evaluation compared to four non-clinical raters for item-level agreement. All 29 items across a subset of 22 videos were considered.

Table 3

Scorability of each behavior across the ASD and non-ASD video collections.

ASDNon-ASD
Behaviors (Code)Code%N/ACode%N/A
Anxiety (E3)E30D30
Facial Expressions (B3)B30.0045E20
Gaze to Initiate Interaction (B4)B40.0045E30
Repetitive Interests (D4)D40.0045A30.0056
Overactivity (E1)E10.0045D10.0056
Tantrums/Aggression (E2)E20.0045D40.0056
Spontaneous Expressive Language (A2)A20.0091E10.0056
Social Overtures (B12)B120.0091A10.01
Complex Mannerisms (D2)D20.0091A50.01
Gestures (A8)A80.0182D20.01
Self-Injurious Behavior (D3)D30.023B120.017
Eye Contact (B1)B10.023A40.02
Sensory Interest (D1)D10.023A80.028
Shared Enjoyment (B5)B50.045B10.028
Intonation (A3)A30.086B40.028
Responsive Smile (B2)B20.1A20.033
Idiosyncratic Use of Words (A5)A50.12B50.039
Pointing (A7)A70.14B30.044
Initiation of Joint Attention (B10)B100.24A60.061
Response to Name (B6)B60.24B20.14
Requesting (B7)B70.27B70.17
Showing (B9)B90.29B60.22
Spontaneous Expressive Language (A1)A10.3A70.3
Functional Play (C1)C10.35B110.34
Giving (B8)B80.38B100.35
Use of Other’s Body to Communicate (A6)A60.44B90.38
Echolalia (A4)A40.49C10.57
Imagination/Creativity (C2)C20.60B80.65
Social Overtures (B11)B110.62C20.79

The frequency of N/A (not applicable) answer codes per video over all video raters is listed in descending order for both video collections. The lowest values correspond to the most readily scored behaviors. A large majority of items were readily detectable and resulted in only a small fraction of N/As.

Figure 1

Inter-rater variability among cases previously diagnosed with autism and controls with no known diagnosis of autism.

While subjectivity shifts did occur among the four independent raters, these shifts in qualitative judgment did not significantly impact the agreement among the reviewers. Both the inter-rater classification agreement (>90%) and classification accuracy (>95%) were found to be high in this analysis.

Inter-rater variability among cases previously diagnosed with autism and controls with no known diagnosis of autism.

While subjectivity shifts did occur among the four independent raters, these shifts in qualitative judgment did not significantly impact the agreement among the reviewers. Both the inter-rater classification agreement (>90%) and classification accuracy (>95%) were found to be high in this analysis. We required majority rater agreement (3 out of 4) for classification. Item-level agreement shown is the mean and range of 29 item agreements across four non-clinical raters. Clinical evaluation compared to four non-clinical raters for item-level agreement. All 29 items across a subset of 22 videos were considered. The frequency of N/A (not applicable) answer codes per video over all video raters is listed in descending order for both video collections. The lowest values correspond to the most readily scored behaviors. A large majority of items were readily detectable and resulted in only a small fraction of N/As. Because two of the raters helped to locate videos appropriate for this study, they were not always naïve to the diagnoses. To address this possible bias, we recruited a research reliable and professionally trained ADOS clinical practitioner (blinded to the diagnosis) to score 22 randomly selected videos to ensure the non-clinical raters were not unduly influenced by their prior knowledge [23] or lack of clinical experience. The raters agreed with classification provided by the clinical expert in all but 3 of the 22 randomly selected videos. In each of these 3 cases, the classification provided by the clinician did not agree with the self-reported diagnosis of ASD. The mean agreement between each non-clinical and clinical rater was 71.3%, which was consistent with the mean among non-clinical raters (73.3%) indicating that prior knowledge did not obviously bias the non-clinical rater’s scoring.

Conclusions

The absence of any reliable molecular, neurological or physical features to characterize ASD means the best-estimate behavioral diagnosis is still the gold standard. For this reason, previous studies focused on identifying a core set of primary deficits from home videos for early age detection, to allow behavioral intervention at key stages in child development. Here, we take a different approach and apply standard diagnostic questions from ADOS module 1 to a diverse collection of videos of children in a non-clinical environment using non-clinical raters. Our results show high classification accuracy and inter-rater reliability and together demonstrate that the ADOS module 1 questions can be used by non-clinicians on unstructured videos to effectively distinguish behavioral differences among children with and without ASD. While not all questions on the ADOS were expected to be relevant to the YouTube videos, we did find that a majority of the questions could be applied. In particular, questions regarding vocalization, use of words or phrases, unusual eye contact, responsive social smile, and repetitive interests or behaviors were the most relevant to the broad assortment of videos. Adding even a limited set of criteria to meet when capturing home videos for prospective studies may increase the likelihood that the less detectable behaviors identified in this study are captured. We observed the length of time to screen a video was never more than one minute over the length of the video and that the average assessment times were ∼4 min for ASD and ∼3∶30 min for non-ASD. These results could have important implications for faster screening approaches in the future. There are important limitations to this proof-of-concept study. First, because we did not contact the individuals, we cannot confirm the diagnosis, gender, age, and race. Second, although by design, our raters were not trained clinicians, were not certified as research-reliable to perform ADOS and were not professionally qualified to confirm a diagnosis. Third, we applied ADOS module 1 to all the videos despite the typical recommendation that it be applied to children with single word vocabulary and mental age less than 3 years of age. However, the high agreement among the non-clinical raters, as well as the agreement of the raters to a research-reliable ADOS practitioner suggests that these limitations did not heavily bias the results. In conclusion, we demonstrate the feasibility of applying a standard diagnostic questions to ASD and non-ASD videos from YouTube. Our results show that despite the short length (<4 minutes), home-quality, and diversity of scenarios captured within the videos it is possible to achieve good inter-rater reliability and high classification accuracy, sensitivity, and specificity. Our results also indicate that it is possible for non-clinical raters to correctly detect the presence of autism with high inter-rater reliability and >94% accuracy. These results support the potential role for short home videos in rapidly screening for a potential autism diagnosis. Based on these results, we hypothesize that it may be possible to apply other testing modules, including highly abbreviated approaches [24] to home videos to further reduce the complexity of detecting autism outside of clinical settings, without appreciable loss of accuracy in comparison to the current standard of care. Our future efforts will focus on testing this and other related hypotheses, including the potential of combining a video-based classification with brief parent-directed classification systems [25] to increase accuracy. Our results also suggest a potential larger role for public video repositories such as YouTube in the detection of other human conditions that have behavioral symptomatology. Details on the 100 YouTube videos. Forty-five autism spectrum disorder (ASD) and 55 non-ASD samples were collected in total. The Table contains the original download URL, the self-reported diagnosis, the gender and age of the subject, length of video, and the total number of YouTube views (as of 12/10/13). (XLSX) Click here for additional data file.
  24 in total

1.  Object play in infants with autism: methodological issues in retrospective video analysis.

Authors:  Grace T Baranek; Carolyn R Barnett; Erin M Adams; Nancy A Wolcott; Linda R Watson; Elizabeth R Crais
Journal:  Am J Occup Ther       Date:  2005 Jan-Feb

2.  Combining information from multiple sources for the diagnosis of autism spectrum disorders for toddlers and young preschoolers from 12 to 47 months of age.

Authors:  So Hyun Kim; Catherine Lord
Journal:  J Child Psychol Psychiatry       Date:  2011-09-01       Impact factor: 8.982

3.  Integrating computerized primitives and annotated video patterns: a proposed model for autism diagnosis and research.

Authors:  Robert Jameson; Daniel Lorence; James Lee
Journal:  J Med Syst       Date:  2011-05-03       Impact factor: 4.460

4.  Computerized home video detection for motherese may help to study impaired interaction between infants who become autistic and their parents.

Authors:  Ammar Mahdhaoui; Mohamed Chetouani; Raquel S Cassel; Catherine Saint-Georges; Erika Parlato; Marie Christine Laznik; Fabio Apicella; Filippo Muratori; Sandra Maestro; David Cohen
Journal:  Int J Methods Psychiatr Res       Date:  2011-03       Impact factor: 4.035

5.  Onset patterns in autism: correspondence between home video and parent report.

Authors:  Sally Ozonoff; Ana-Maria Iosif; Gregory S Young; Susan Hepburn; Meagan Thompson; Costanza Colombi; Ian C Cook; Emily Werner; Stacy Goldring; Fottfam Baguio; Sally J Rogers
Journal:  J Am Acad Child Adolesc Psychiatry       Date:  2011-06-02       Impact factor: 8.829

6.  Same but different: 9-month-old infants at average and high risk for autism look at the same facial features but process them using different brain mechanisms.

Authors:  Alexandra P F Key; Wendy L Stone
Journal:  Autism Res       Date:  2012-06-01       Impact factor: 5.216

7.  Early recognition of children with autism: a study of first birthday home videotapes.

Authors:  J Osterling; G Dawson
Journal:  J Autism Dev Disord       Date:  1994-06

8.  Coordination of precision grip in 2-6 years-old children with autism spectrum disorders compared to children developing typically and children with developmental disabilities.

Authors:  Fabian J David; Grace T Baranek; Chris Wiesen; Adrienne F Miao; Deborah E Thorpe
Journal:  Front Integr Neurosci       Date:  2012-12-31

9.  Use of artificial intelligence to shorten the behavioral diagnosis of autism.

Authors:  Dennis P Wall; Rebecca Dally; Rhiannon Luyster; Jae-Yoon Jung; Todd F Deluca
Journal:  PLoS One       Date:  2012-08-27       Impact factor: 3.240

10.  Use of machine learning to shorten observation-based screening and diagnosis of autism.

Authors:  D P Wall; J Kosmicki; T F Deluca; E Harstad; V A Fusaro
Journal:  Transl Psychiatry       Date:  2012-04-10       Impact factor: 6.222

View more
  23 in total

1.  The impact of the COVID-19 pandemic on the assessment of autism spectrum disorder.

Authors:  Annio Posar; Paola Visconti; Virginia Giuberti
Journal:  Turk Arch Pediatr       Date:  2021-05-01

Review 2.  Towards a Multivariate Biomarker-Based Diagnosis of Autism Spectrum Disorder: Review and Discussion of Recent Advancements.

Authors:  Troy Vargason; Genevieve Grivas; Kathryn L Hollowood-Jones; Juergen Hahn
Journal:  Semin Pediatr Neurol       Date:  2020-03-05       Impact factor: 1.636

3.  "Frank" presentations as a novel research construct and element of diagnostic decision-making in autism spectrum disorder.

Authors:  Ashley de Marchena; Judith Miller
Journal:  Autism Res       Date:  2016-10-21       Impact factor: 5.216

4.  Social media video analysis methodology for sarin exposure.

Authors:  Sadik Toprak; Emine Yilmaz Can; Bulent Altinsoy; John Hart; Zekeriya Dogan; Mustafa Ozcetin
Journal:  Forensic Sci Res       Date:  2020-11-05

5.  Computer Decision Support Changes Physician Practice But Not Knowledge Regarding Autism Spectrum Disorders.

Authors:  N S Bauer; A E Carroll; C Saha; S M Downs
Journal:  Appl Clin Inform       Date:  2015-07-15       Impact factor: 2.342

Review 6.  Data-Driven Diagnostics and the Potential of Mobile Artificial Intelligence for Digital Therapeutic Phenotyping in Computational Psychiatry.

Authors:  Peter Washington; Natalie Park; Parishkrita Srivastava; Catalin Voss; Aaron Kline; Maya Varma; Qandeel Tariq; Haik Kalantarian; Jessey Schwartz; Ritik Patnaik; Brianna Chrisman; Nathaniel Stockham; Kelley Paskov; Nick Haber; Dennis P Wall
Journal:  Biol Psychiatry Cogn Neurosci Neuroimaging       Date:  2019-12-13

7.  Searching for a minimal set of behaviors for autism detection through feature selection-based machine learning.

Authors:  J A Kosmicki; V Sochat; M Duda; D P Wall
Journal:  Transl Psychiatry       Date:  2015-02-24       Impact factor: 6.222

8.  Crowdsourced validation of a machine-learning classification system for autism and ADHD.

Authors:  M Duda; N Haber; J Daniels; D P Wall
Journal:  Transl Psychiatry       Date:  2017-05-16       Impact factor: 6.222

9.  Sparsifying machine learning models identify stable subsets of predictive features for behavioral detection of autism.

Authors:  Sebastien Levy; Marlena Duda; Nick Haber; Dennis P Wall
Journal:  Mol Autism       Date:  2017-12-19       Impact factor: 7.509

10.  The GapMap project: a mobile surveillance system to map diagnosed autism cases and gaps in autism services globally.

Authors:  Jena Daniels; Jessey Schwartz; Nikhila Albert; Michael Du; Dennis P Wall
Journal:  Mol Autism       Date:  2017-10-23       Impact factor: 7.509

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.