| Literature DB >> 32246097 |
Franz M J Pfister1, Terry Taewoong Um2, Daniel C Pichler3,4, Jann Goschenhofer1, Kian Abedinpour3,4, Muriel Lang5, Satoshi Endo5, Andres O Ceballos-Baumann3,4, Sandra Hirche5, Bernd Bischl1, Dana Kulić2, Urban M Fietzek6,7.
Abstract
Patients with advanced Parkinson's disease regularly experience unstable motor states. Objective and reliable monitoring of these fluctuations is an unmet need. We used deep learning to classify motion data from a single wrist-worn IMU sensor recording in unscripted environments. For validation purposes, patients were accompanied by a movement disorder expert, and their motor state was passively evaluated every minute. We acquired a dataset of 8,661 minutes of IMU data from 30 patients, with annotations about the motor state (OFF,ON, DYSKINETIC) based on MDS-UPDRS global bradykinesia item and the AIMS upper limb dyskinesia item. Using a 1-minute window size as an input for a convolutional neural network trained on data from a subset of patients, we achieved a three-class balanced accuracy of 0.654 on data from previously unseen subjects. This corresponds to detecting the OFF, ON, or DYSKINETIC motor state at a sensitivity/specificity of 0.64/0.89, 0.67/0.67 and 0.64/0.89, respectively. On average, the model outputs were highly correlated with the annotation on a per subject scale (r = 0.83/0.84; p < 0.0001), and sustained so for the highly resolved time windows of 1 minute (r = 0.64/0.70; p < 0.0001). Thus, we demonstrate the feasibility of long-term motor-state detection in a free-living setting with deep learning using motion data from a single IMU.Entities:
Mesh:
Year: 2020 PMID: 32246097 PMCID: PMC7125162 DOI: 10.1038/s41598-020-61789-3
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Clinical descriptors of the full cohort and according to the Hoehn & Yahr disease stages two to four.
| parameter | full cohort | HY stage 2 | HY stage 3 | HY stage 4 |
|---|---|---|---|---|
| Gender (male/female) | 20/10 | 9/2 | 9/7 | 2/1 |
| Age (years) | 67.1 ± 10.2 (40–83) | 66.1 ± 8.1 (47–73) | 69.5 ± 11.0 (40–83) | 58.0 ± 9.2 (48–66) |
| Disease duration (years) | 11.0 ± 5.1 (1–21) | 9.9 ± 6.0 (1–18) | 11.2 ± 4.8 (2–21) | 14.3 ± 2.3 (13–17) |
| Levodopa equivalent dose$ | 1109 ± 785 (90–3754) | 1172 ± 1113 (90–3754) | 1053 ± 521 (120–2435) | 1181 ± 806 (675–2110) |
| MDS-UPDRS III (ON) | 21.6 ± 15.3 (2–57) | 16.9 ± 15.8 (2–57) | 25.4 ± 15.6 (5–54) | 19.0 ± 7.9 (13–28) |
AIMS (sum items 1–7) | 2.1 ± 2.4 (0–7) | 1.2 ± 1.8 (0–4.5) | 2.5 ± 2.7 (0–7) | 2.9 ± 2.7 (1.25–6.0) |
| Montreal Cognitive Assessment | 25.7 ± 2.8 (18–30) | 26.1 ± 1.9 (24–30) | 24.9 ± 3.3 (18–29) | 28.0 ± 1.7 (27–30) |
| Body Mass Index (kg/m2) | 25.2 ± 4.8 (12.9–35.4) | 26.9 ± 5.0 (21.6–35.4) | 25.2 ± 3.9 (21.3–35.1) | 19.4 ± 5.8 (12.9–23.7) |
| Duration of motion data recording (hours) | 7.5 ± 3.9 (0.4–13.4) | 6.6 ± 4.0 (0.4–12.5) | 8.3 ± 3.8 (2.8–13.4) | 6.5 ± 5.1 (2.6–12.3) |
| Additional therapy, i.e. deep brain stimulation (DBS) or continuous subcutaneous apomorphine infusion (CSAI) | DBS = 4 CSAI = 2 | DBS = 1 | DBS = 1 CSAI = 2 | DBS = 2 |
Mean ± standard deviation is given with the range in brackets. $Levodopa equivalent dose is calculated according to ref. [68]. Abbreviations: AIMS, abnormal involuntary movement scale; UPDRS, unified PD rating scale; DBS, deep brain stimulation; CSAI, continuous subcutaneous apomorphine infusion.
Figure 1Confusion matrix of the clinical annotation (Reference) and the model prediction (Prediction), and clinimetric results of the CNN prediction for the full (augmented) dataset. The numbers in the matrix indicate corresponding observations.
Comparison of the machine learning (ML) methods with a 4-fold cross-validation.
| Methodology | SVM (linear) | kNN (n = 10) | Random Forest | MLP | CNN |
|---|---|---|---|---|---|
| Balanced Accuracy | 50.28 | 50.41 | 53.73 | 54.02 | 67.39 |
Note that these performance measures were obtained from non-augmented data using 4-fold cross validation. Abbreviations: SVM, support vector machine; kNN, k-nearest neighbor; MLP, multi-layer perceptron; CNN, convoluted neural network.
Figure 2Motor state profiles of four typical patients (A–D). The observed motor state annotations by the expert rater is coded in three background colors along the x-axis; blue, OFF; green, ON; red, DYS). Unsmoothed expCNN point predictions (transparent circles) and LOESS smoothed predicted day curves (drawn line) predict the highly resoluted motor state. In the top section of part A, clinical information on the free-living activity is given that demonstrates the independence of the predicted motor state curve from the concurrent motor activity. Patient A is a severely fluctuating patient who spends almost no time in the ON condition, but experiences several sudden OFF or DYS phases during the recording. The expert evaluated A with an average of 1.36 ± 0.54. Mean expCNN was similar with 1.02 ± 0.81. The high SD identifies A as a severe fluctuator. Patient B is a predominantly bradykinetic patient with no ON time and no dyskinesia. B was admitted to the hospital a couple of weeks after DBS surgery and recorded before readjustment of the stimulation parameters. The expert evaluated B with an average of 0.48 ± 0.20. Mean expCNN was similar with 0.28 ± 0.22. Note that there are missing values in the predictions as the sensor signal was not recorded continuously. Patient C is a patient who is responding well to treatment. Except for the typical OFF in the morning, C spends almost all the time in the ON state. The rater evaluated C with an average of 0.99 ± 0.11. Mean expCNN was similar with 0.91 ± 0.29. The mean around 1 (=ON) and the low SD describe a patient with minor fluctuations mainly in ON state. Patient D is a dyskinetic patient with one OFF period in the afternoon. The expert evaluated him with an average of 1.47 ± 0.34. Mean expCNN was similar with 1.66 ± 0.37. The curves from all patients are available as supplementary material.
Figure 3Correlations of the probabilistic output of the CNN with the bradykinesia and dyskinesia items that were labeled minutely by the expert, but modelled for various time windows. Linear correlation is depicted with 95% confidence interval (grey shading).
Prevalence and CNN performances across seven classes of background activities.
| Activity (%) | Sitting | Walking | Lying | Standing | Testing | Other | Unknown |
|---|---|---|---|---|---|---|---|
| Prevalence | 58.0 | 18.5 | 8.4 | 5.3 | 1.2 | 3.4 | 5.3 |
| Balanced Accuracy | 74.47 | 76.94 | 70.72 | 69.31 | 72.01 | 64.38 | 61.85 |
Figure 4Setup of IMU sensor data acquisition and CNN architecture. The CNN consists of seven convolutional blocks followed by two fully-connected layers. Each block consists of a convolution, batch-normalization[66] and ReLU[67] layer. As we progress through the convolutional blocks, the size of the input vector decreases from 3600 * 3 to 13 * 1, whereas the number of channels increases from 1 to 64. Finally, the 13 * 1 * 64 feature maps are flattened to an 832 * 1 vector and classified by two fully-connected layers which have 512 and 3 nodes, respectively. (Figure drawn by Dr. Pfister).