| Literature DB >> 35847210 |
Chad M Aldridge1, Mark M McDonald1, Mattia Wruble1, Yan Zhuang2, Omar Uribe1, Timothy L McMurry3, Iris Lin4, Haydon Pitchford1, Brett J Schneider1, William A Dalrymple1, Joseph F Carrera5, Sherita Chapman1, Bradford B Worrall1,3, Gustavo K Rohde2,6, Andrew M Southerland1,3.
Abstract
Background: Current EMS stroke screening tools facilitate early detection and triage, but the tools' accuracy and reliability are limited and highly variable. An automated stroke screening tool could improve stroke outcomes by facilitating more accurate prehospital diagnosis and delivery. We hypothesize that a machine learning algorithm using video analysis can detect common signs of stroke. As a proof-of-concept study, we trained a computer algorithm to detect presence and laterality of facial weakness in publically available videos with comparable accuracy, sensitivity, and specificity to paramedics. Methods andEntities:
Keywords: access to care; cerebrovascular disease; computer vision; diagnostic test; infarction; machine learning; stroke
Year: 2022 PMID: 35847210 PMCID: PMC9284117 DOI: 10.3389/fneur.2022.878282
Source DB: PubMed Journal: Front Neurol ISSN: 1664-2295 Impact factor: 4.086
Figure 1Facial landmark extraction and classification. For algorithm training and performance, videos were decomposed into individual frames and facial landmarks were extracted. Normalization was performed by aligning the extracted landmarks of current input sequences to a template to handle different head scales, location, and orientation. The same transformation was applied to the pixel intensity information to remove these variations. A predictive model computed the classification result for each individual frame and a voting classifier reduced the output into discrete classifications comprised of no weakness, left facial weakness, and right facial weakness categories.
Performance metrics of the correct identification of facial weakness and its laterality among all raters.
|
| ||||||
|---|---|---|---|---|---|---|
|
|
|
| ||||
|
|
|
|
|
| ||
| Accuracy | 92.60% | 90.1–94.7% | 88.90% | 83.5–93% | 49% | |
| Sensitivity | 87.80% | 83.9–91.7% | 90.30% | 82.4–95.5% | 100% | |
| Specificity | 99.30% | 98.2–100% | 87.50% | 79.2–93.4% | 0% | |
|
|
|
| ||||
|
|
|
|
|
|
| |
| Accuracy | 94.70% | 91.5–97.9% | 94.20% | 90.5–97.4% | 88.90% | 84.1–93.1% |
| Sensitivity | 95.70% | 91–99% | 89.20% | 82.6–95% | 78.50% | 69.9–86.5% |
| Specificity | 93.80% | 88.5–98% | 99.00% | 96.6–100% | 98.90% | 96.6–100% |
95% Confidence intervals are the result of 10,000 boot-straps via the percentile method. ZeroR shows.
the expected diagnostic performance based on random selection of the presence of facial weakness.
Figure 2ROC curves of facial weakness and laterality detection of the algorithm and each paramedic. The overall performance of paramedics vs. the algorithm in the detection of unilateral facial weakness failed to reach significance (p-value 0.074).
Cross tabulation ratings of presence and laterality of unilateral facial weakness by paramedics as a group vs. the machine learning algorithm.
|
| |||
|---|---|---|---|
|
|
|
| |
|
| |||
| Present | 82 | 1 | 83 |
| Absent | 11 | 95 | 106 |
|
| |||
| Present | 84 | 12 | 96 |
| Absent | 9 | 84 | 93 |
|
| |||
|
|
| ||
| Paramedics | 245 | 13 | |
| Algorithm | 84 | 0 | |
Three independent and blinded board-certified vascular neurologists determined the ground truth labeling for the presence or absence of unilateral facial weakness and the affected side of the face. For the presence of facial weakness, there was no significant differences between Paramedics (as a group) and the algorithm, p-value 0.074. The Paramedics, however, had significantly more laterality errors compared to the algorithm, p-value 0.044. Both comparisons utilized Fisher's exact test.