| Literature DB >> 30498472 |
Charalambos Themistocleous1,2, Marie Eckerström3, Dimitrios Kokkinakis1.
Abstract
While people with mild cognitive impairment (MCI) portray noticeably incipient memory difficulty in remembering events and situations along with problems in decision making, planning, and finding their way in familiar environments, detailed neuropsychological assessments also indicate deficits in language performance. To this day, there is no cure for dementia but early-stage treatment can delay the progression of MCI; thus, the development of valid tools for identifying early cognitive changes is of great importance. In this study, we provide an automated machine learning method, using Deep Neural Network Architectures, that aims to identify MCI. Speech materials were obtained using a reading task during evaluation sessions, as part of the Gothenburg MCI research study. Measures of vowel duration, vowel formants (F1 to F5), and fundamental frequency were calculated from speech signals. To learn the acoustic characteristics associated with MCI vs. healthy controls, we have trained and evaluated ten Deep Neural Network Architectures and measured how accurately they can diagnose participants that are unknown to the model. We evaluated the models using two evaluation tasks: a 5-fold crossvalidation and by splitting the data into 90% training and 10% evaluation set. The findings suggest first, that the acoustic features provide significant information for the identification of MCI; second, the best Deep Neural Network Architectures can classify MCI and healthy controls with high classification accuracy (M = 83%); and third, the model has the potential to offer higher accuracy than 84% if trained with more data (cf., SD≈15%). The Deep Neural Network Architecture proposed here constitutes a method that contributes to the early diagnosis of cognitive decline, quantify the progression of the condition, and enable suitable therapeutics.Entities:
Keywords: MCI; dementia; machine learning; neural network; prosody; speech production; vowels
Year: 2018 PMID: 30498472 PMCID: PMC6250092 DOI: 10.3389/fneur.2018.00975
Source DB: PubMed Journal: Front Neurol ISSN: 1664-2295 Impact factor: 4.003
Age and gender of healthy controls (HC) and participants with Mild Cognitive Impairment (MCI).
| HC | 19 | 11 | 68 (7.6) | 69 (5.7) |
| MCI | 13 | 12 | 72 (5.1) | 70 (5.6) |
Figure 1Waveform, spectrogram, and F0 contour—superimposed on the spectrogram—of an example utterance (upper tier). Shown in the plot is the segmentation of the word havsbottnen “seabed” (middle tier); the individual sounds are shown in the lowest tier. Sound boundaries are indicated with thin vertical lines. The ordinate shows the F0 values whereas the abscissa shows the time in second.
Figure 2Network architecture. We developed 10 different networks with 21 predictors each. The networks differed in the number of hidden layers ranging from 1…10. Each network architecture was evaluated twice using cross-validation and evaluation split. Model comparison measures are reported for each evaluation separately.
Deep neural network architectures with 1 to 10 hidden layers.
| Input layer | Dense 300 (21 Input Dimensions) | ReLU |
| 1…10 hidden layers | Dense 300 | ReLU |
| Output layer | 1 | Sigmoid |
All models employed stochastic gradient descent optimizer with 0.9 Nesterov momentum.
Confusion matrix.
| Predicted condition positive | True positive (TP) | False positive (FP) |
| Predicted condition negative | False negative (FN) | True negative (TN) |
Model M1…M10 mean classification accuracy and mean validation accuracy and the corresponding SD from the 5-fold crossvalidation.
| M1 | 98 | 3 | 75 | 12 |
| M2 | 99 | 3 | 80 | 14 |
| M3 | 99 | 2 | 81 | 15 |
| M4 | 99 | 2 | 82 | 15 |
| M5 | 99 | 2 | 82 | 14 |
| M6 | 99 | 2 | 83 | 15 |
| M7 | 99 | 2 | 83 | 16 |
| M8 | 99 | 2 | 83 | 15 |
| M9 | 99 | 2 | 83 | 16 |
| M10 | 98 | 3 | 83 | 17 |
Figure 3Mean ROC curve and AUC of the 5-fold crossvalidation. Model—M1…M10— are represented by solid line with a different color. The baseline is represented by a dashed gray line. All models provided ROC curves that were over the baseline. The best model is the model whose ROC curve approaches the left upper corner. The shaded area indicates the M10's SD that is the outperforming model both in terms of ROC/AUC (83%) and validation accuracy (83%).
90%/10% validation split results.
| M1 | 67 | 86 | 56 | 63 |
| M2 | 68 | 92 | 56 | 66 |
| M3 | 67 | 100 | 49 | 65 |
| M4 | 68 | 63 | 62 | 62 |
| M5 | 71 | 73 | 71 | 71 |
| M6 | 68 | 73 | 72 | 72 |
| M7 | 75 | 100 | 49 | 65 |
| M8 | 65 | 100 | 49 | 65 |
| M9 | 69 | 100 | 49 | 65 |
| M10 | 66 | 95 | 51 | 64 |
The table shows the accuracy, precision, recall, and f1 score for M1…M10.