| Literature DB >> 33723525 |
Brian Stasak1, Zhaocheng Huang1, Sabah Razavi2, Dale Joachim2, Julien Epps1.
Abstract
Currently, there is an increasing global need for COVID-19 screening to help reduce the rate of infection and at-risk patient workload at hospitals. Smartphone-based screening for COVID-19 along with other respiratory illnesses offers excellent potential due to its rapid-rollout remote platform, user convenience, symptom tracking, comparatively low cost, and prompt result processing timeframe. In particular, speech-based analysis embedded in smartphone app technology can measure physiological effects relevant to COVID-19 screening that are not yet digitally available at scale in the healthcare field. Using a selection of the Sonde Health COVID-19 2020 dataset, this study examines the speech of COVID-19-negative participants exhibiting mild and moderate COVID-19-like symptoms as well as that of COVID-19-positive participants with mild to moderate symptoms. Our study investigates the classification potential of acoustic features (e.g., glottal, prosodic, spectral) from short-duration speech segments (e.g., held vowel, pataka phrase, nasal phrase) for automatic COVID-19 classification using machine learning. Experimental results indicate that certain feature-task combinations can produce COVID-19 classification accuracy of up to 80% as compared with using the all-acoustic feature baseline (68%). Further, with brute-forced n-best feature selection and speech task fusion, automatic COVID-19 classification accuracy of upwards of 82-86% was achieved, depending on whether the COVID-19-negative participant had mild or moderate COVID-19-like symptom severity.Entities:
Keywords: Digital medicine; Machine learning; Remote sensing; Respiratory illness
Year: 2021 PMID: 33723525 PMCID: PMC7948650 DOI: 10.1007/s41666-020-00090-4
Source DB: PubMed Journal: J Healthc Inform Res ISSN: 2509-498X
Example of the Center for Disease Control informed self-questionnaire completed by each participant in the SHC dataset
| Q1 | |
| (0) No, I have never had a COVID test | |
| (1) Yes, my test result was positive | |
| (2) Yes, my result was negative | |
| (3) Yes, I am waiting for my test result | |
| (4) Yes, my test was inconclusive | |
| Q2 | |
| (0) I have never had a COVID test | |
| (1) I was tested today | |
| (2) I was tested 1–2 days ago | |
| (3) I was tested 3–6 days ago | |
| (4) I was tested 1–2 weeks ago | |
| (5) I was tested more than 2 weeks ago | |
| Q3 | |
| (0) No | |
| (1) Yes, I have not measured my temperature | |
| (2) Yes, less than 100F | |
| (3) Yes, between 100 and 102 F | |
| Q4 | |
| (0) No | |
| (5) Yes | |
| Q5 | |
| (0) No | |
| (1) Mild fatigue | |
| (2) Severe fatigue—I struggle to get out of bed | |
| Q6 | |
| (0) No | |
| (1) Yes, mild symptoms—slight shortness of breath during ordinary activities | |
| (2) Yes, significant symptoms—breathing is comfortable only at rest | |
| Q7 | |
| (0) No | |
| (5) Yes | |
| Q8 | |
| (0) No | |
| (5) Yes | |
| Q9 | |
| (0) No | |
| (5) Yes | |
| Q10 | |
| (0) No | |
| (5) Yes | |
| Q11 | |
| (0) No | |
| (5) Yes | |
| Q12 | |
| (0) Asthma | |
| (1) Chronic obstructive pulmonary disease (COPD) | |
| (2) Congestive heart failure (CHF) | |
| (3) Other condition that may make it difficult to breathe | |
| (4) None of the above | |
| Q13 | |
| (0) Yes, everyday | |
| (1) Yes, some days | |
| (2) Not currently, but I did in the past | |
| (3) Never smoked | |
| Q14 | |
| (0) Not congested at all | |
| (1) Mild congestion | |
| (2) Moderate congestion | |
| (3) Severe congestion |
These questions are based on several key symptoms reported by many individuals infected with COVID-19 illness
The SHC dataset contains a total of 66 participants. Each participant provided a held vowel, diadochokinetic pataka phrase, and nasal phrase recording
| Participant groups | Metadata | |||
|---|---|---|---|---|
COVID-19-negative | 22 | 10 | 12 | 42.5 ± 13.5 |
COVID-19-negative | 22 | 10 | 12 | 39.5 ± 12.4 |
| COVID-19-positive | 22 | 10 | 12 | 45.6 ± 12.7 |
Fig. 1SHC dataset median symptom severity COVID-19 self-assessment scores per individual question and participant group (see Appendix, Table 5). The dotted circle indicates the median, whereas the plus sign indicates extreme data outliers. In addition, the bar indicates the 25th to 75th percentiles, while the extended thin whisker line represents the remaining outer data values
Fig. 2SHC dataset spectrograms of three similarly aged female participants held vowel task recordings: (a) Cneg, (b) CCneg, and (c) Cpos. To observe the short-term temporal variation, only a 1-s segment of each recording is shown with a 0-16 kHz frequency range. For the particular spectrogram examples given above, there were relatively large differences in participants’ fundamental frequency (F0): Cneg, 199 Hz; CCneg, 155 Hz; and Cpos, 63 Hz. A typical healthy adult female has a F0 of approximately 200 Hz
Fig. 3Experimental design showing both individual systems and their fusion for COVID-19 binary classification (e.g., negative vs. positive). The proposed experimental feature selection method explored in this study herein is indicated by dashed lines
COVID-19 classification accuracy results using a decision tree classifier with leave-one-speaker-out cross validation
| All COVAREP (150) | 55% | 32% | |
| 30% | 32% | ||
| 52% | 52% | ||
| 57% | 52% | 66% | |
| All COVAREP (150) | 52% | 50% | 27% |
| 52% | 50% | ||
| 57% | 39% | 46% | |
| 50% | 48% | 39% | |
The total number of features per feature type is shown in parenthesis
COVID-19 classification accuracy results using automatic brute-forced n-best feature selection and decision tree classifier with leave-one-speaker-out cross validation
| 80% | 75% | ||
| 66% | 61% | ||
| 75% | 73% | ||
| 61% | 64% | 66% | |
| 77% | |||
The total number of features per feature type is shown in parenthesis
COVID-19 classification accuracy results using task-based fusion with automatic brute-forced n-best feature selection (n = 5) with leave-one-speaker-out cross validation. Per row, results are shown for various task/feature combinations (indicated by dots). The selected tasks are indicated by dots for the held vowel (H), pataka phrase (P), and nasal phrase (N) abbreviations
| Glottal | Prosodic | Spectral | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| 75% | 77% | ||||||||||
| 80% | 57% | 64% | |||||||||
| 80% | 82% | ||||||||||
| 75% | 82% | ||||||||||
| 77% | 73% | 82% | |||||||||
| 80% | 71% | 75% | |||||||||
| 75% | 66% | 71% | |||||||||
| 73% | 80% | 77% | |||||||||
| 77% | 71% | ||||||||||
| 75% | 77% | 77% | |||||||||
*Produced measures of 0.91 sensitivity and 0.77 specificity