| Literature DB >> 22952789 |
Dennis P Wall1, Rebecca Dally, Rhiannon Luyster, Jae-Yoon Jung, Todd F Deluca.
Abstract
The Autism Diagnostic Interview-Revised (ADI-R) is one of the most commonly used instruments for assisting in the behavioral diagnosis of autism. The exam consists of 93 questions that must be answered by a care provider within a focused session that often spans 2.5 hours. We used machine learning techniques to study the complete sets of answers to the ADI-R available at the Autism Genetic Research Exchange (AGRE) for 891 individuals diagnosed with autism and 75 individuals who did not meet the criteria for an autism diagnosis. Our analysis showed that 7 of the 93 items contained in the ADI-R were sufficient to classify autism with 99.9% statistical accuracy. We further tested the accuracy of this 7-question classifier against complete sets of answers from two independent sources, a collection of 1654 individuals with autism from the Simons Foundation and a collection of 322 individuals with autism from the Boston Autism Consortium. In both cases, our classifier performed with nearly 100% statistical accuracy, properly categorizing all but one of the individuals from these two resources who previously had been diagnosed with autism through the standard ADI-R. Our ability to measure specificity was limited by the small numbers of non-spectrum cases in the research data used, however, both real and simulated data demonstrated a range in specificity from 99% to 93.8%. With incidence rates rising, the capacity to diagnose autism quickly and effectively requires careful design of behavioral assessment methods. Ours is an initial attempt to retrospectively analyze large data repositories to derive an accurate, but significantly abbreviated approach that may be used for rapid detection and clinical prioritization of individuals likely to have an autism spectrum disorder. Such a tool could assist in streamlining the clinical diagnostic process overall, leading to faster screening and earlier treatment of individuals with autism.Entities:
Mesh:
Year: 2012 PMID: 22952789 PMCID: PMC3428277 DOI: 10.1371/journal.pone.0043855
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Summary of the data used for both construction and validation of the autism classifier.
| Classifier Data | Validation Data | |||||
| AGRE | SSC | AC | ||||
| Autism | Not Met | Autism | Not Met | Autism | Not Met | |
| Sample Size | 891 | 75 | 1654 | 5 | 322 | 12 |
| Q1 (Age) | 6.44 | 6.38 | 6.75 | 8.38 | 6.50 | 5.42 |
| Median (Age) | 8.06 | 9.24 | 8.75 | 9.75 | 8.50 | 9.50 |
| Q3 (Age) | 10.84 | 11.88 | 11.25 | 12.25 | 11.54 | 13.58 |
| IQR (Age) | 4.4 | 5.5 | 4.5 | 3.88 | 5.04 | 8.17 |
Full sets of answers to the Autism Diagnostic Instrument-Revised questionnaire were downloaded from the Autism Genetic Research Exchange (AGRE), the Simons Simplex Collection (SSC), and the Boston Autism Consortium (AC). The AGRE data were used for training, testing, and construction of the classifier. The SSC and AC data were used for independent validation of the resulting classifier. The table lists the total numbers of spectrum and non-spectrum individuals represented in each of the three data sets with a breakdown of age by quartiles.
The 15 machine learning algorithms used to analyze the Autism Genetic Resource Exchange ADI-R data.
| Classifier Name | Description | FPR | TPR | Accuracy |
| ADTree | An ADTree combines decision trees, voted decision trees, and voted decision stumps.This particular algorithm is based on boosting, which produces accurate predictionsby combining a series of “weak” learners that together, can classify accurately | 0.013 | 1.000 | 0.999 |
| BFTree | The top node of the decision tree is the one that splitsthe data so that the maximum reduction of impurity (misclassified data) is achieved. This is called the “best” node,and it is expanded upon first (unlike in a C4.5 tree, for example, where nodes areexpanded upon according to depth-first) | 0.053 | 0.991 | 0.988 |
| ConjunctiveRule | Within the ConjuctiveRule classifier is a conjunctive rule learner, which can predict forboth numeric and nominal class labels. A rule consists of a series of antecedents joinedby “AND”s | 0.080 | 0.981 | 0.976 |
| DecisionStump | A DecisionStump classifier is a single-level decision tree with one node. The terminalnodes extend directly off of this node, so a classification is made based on asingle attribute | 0.107 | 0.985 | 0.978 |
| FilteredClassifier | FilteredClassifier runs data through an arbitrary classifier after it’s been run through anarbitrary filter. Classifiers are built using training data, and in this case, the filter is alsobuilt based on the training data. This allows the user to skip the pre-processing stepsassociated with transforming the data | 0.040 | 0.993 | 0.991 |
| J48 | J48 is a Java implementation of the C4.5 algorithm; it generates either an unprunedor a pruned C4.5 decision tree. C4.5 uses the concept of information entropy tobuild trees from training data | 0.053 | 0.998 | 0.994 |
| J48graft | This class generates a grafted C4.5 decision tree that can either be prunedor unpruned. Grafting adds nodes to already created decision trees to improveaccuracy | 0.200 | 1.000 | 0.984 |
| JRip | This classifier is an optimized version of Incremental Reduced Error Pruning, andimplements a propositional learner, RIPPER (Repeated Incremental Pruning to ProduceError Reduction). It produces accurate and “readable” rules | 0.053 | 0.997 | 0.993 |
| LADTree | LADTree produces a multi-class alternating decision tree. It has the capability tohave more than two class inputs. It uses the LogitBoost strategy, which performsadditive logistic regression | 0.027 | 1.000 | 0.998 |
| NNge | Nearest neighbor algorithms define a distance function to separate classes. Usinggeneralized exemplars reduce the role of the distance function (relying too heavilyon the distance function can produce inaccurate results) by grouping classestogether | 0.080 | 1.000 | 0.994 |
| OneR | This algorithm finds association rules. It finds the one attribute that classifiesinstances so as to reduce prediction errors | 0.093 | 0.996 | 0.989 |
| OrdinalClassClassifier | This is a meta-classifier (meta-classifiers are likeclassifiers, but have added functionality) used totransform an ordinal class problem to a series of binary class problems | 0.053 | 0.998 | 0.994 |
| PART | A set of rules is generated using the “divide-and-conquer” strategy. From here,all instances in the training data that are covered by this rule get removed and this processis repeated until no instances remain | 0.040 | 0.996 | 0.993 |
| Ridor | This classifier is an implementation of a Ripple-Down Rule Learner. An example ofthis is when the classifier picks a default rule (based on the least weighted error),and creates exception cases stemming from this one | 0.080 | 0.996 | 0.990 |
| SimpleCart | Classification and regression trees are used to construct prediction models for data.They are made by partitioning the data and fitting models to each partition | 0.053 | 0.993 | 0.990 |
These algorithms were deployed using the toolkit WEKA [28]. The false positive rate (FPR) and true positive rate (TPR) are provided together with overall accuracy. The Alternating Decision Tree (ADTree) performed with highest accuracy and was used for further analyses.
Figure 1Performance of 15 machine learning algorithms evaluated for classifying autism cases and non-spectrum controls.
Plot comparing 1-specificity and sensitivity for the 15 different machine learning algorithms used to construct classifiers from the 93-item Autism Diagnostic Interview-Revised (ADI-R) instrument from the Autism Genetic Resource Exchange (AGRE). The best performing algorithm was the alternating decision tree (ADTree), followed by LADTree, PART, and FilteredClassifier. Table 2 summarizes the 15 machine learning algorithms in more detail, and the elements contained in the ADTree classifier are listed in Table 3.
The seven attributes used in the ADTree model.
| ADI-R question | Abbreviation | Description |
| 29 | compsl5 | Comprehension of simple language: answer most abnormal between 4 and 5 |
| 35 | conver5 | Reciprocal conversation (within subject’s level of language): answer if ever (when verbal) |
| 48 | play5 | Imaginative play: answer most abnormal between 4 and 5 |
| 49 | peerpl5 | Imaginative play with peers: answer most abnormal between 4 and 5 |
| 50 | gaze5 | Direct gaze: answer most abnormal between 4 and 5 |
| 64 | grplay5 | Group play with peers: answer most abnormal between 4 and 5 |
| 86 | ageabn | Age when abnormality first evident |
Listed are the number corresponding to the question in the full ADI-R instrument, the question code used by Autism Genetic Research Exchange (AGRE) and a brief description of the question.
Figure 2Decision tree scores and classification of cases with and without a diagnosis of autism.
The Alternating Decision Tree (ADTree) scores of individuals in the both the AC and AGRE data sets versus their age in years. A majority of the ADTree scores were clustered towards greater magnitudes according to their respective classifications, regardless of age.