| Literature DB >> 35453968 |
Tawsifur Rahman1, Nabil Ibtehaz1, Amith Khandakar1, Md Sakib Abrar Hossain1, Yosra Magdi Salih Mekki2, Maymouna Ezeddin1, Enamul Haque Bhuiyan3, Mohamed Arselene Ayari4, Anas Tahir1, Yazan Qiblawey1, Sakib Mahmud1, Susu M Zughaier2, Tariq Abbas5, Somaya Al-Maadeed6, Muhammad E H Chowdhury1.
Abstract
Problem-Since the outbreak of the COVID-19 pandemic, mass testing has become essential to reduce the spread of the virus. Several recent studies suggest that a significant number of COVID-19 patients display no physical symptoms whatsoever. Therefore, it is unlikely that these patients will undergo COVID-19 testing, which increases their chances of unintentionally spreading the virus. Currently, the primary diagnostic tool to detect COVID-19 is a reverse-transcription polymerase chain reaction (RT-PCR) test from the respiratory specimens of the suspected patient, which is invasive and a resource-dependent technique. It is evident from recent researches that asymptomatic COVID-19 patients cough and breathe in a different way than healthy people. Aim-This paper aims to use a novel machine learning approach to detect COVID-19 (symptomatic and asymptomatic) patients from the convenience of their homes so that they do not overburden the healthcare system and also do not spread the virus unknowingly by continuously monitoring themselves. Method-A Cambridge University research group shared such a dataset of cough and breath sound samples from 582 healthy and 141 COVID-19 patients. Among the COVID-19 patients, 87 were asymptomatic while 54 were symptomatic (had a dry or wet cough). In addition to the available dataset, the proposed work deployed a real-time deep learning-based backend server with a web application to crowdsource cough and breath datasets and also screen for COVID-19 infection from the comfort of the user's home. The collected dataset includes data from 245 healthy individuals and 78 asymptomatic and 18 symptomatic COVID-19 patients. Users can simply use the application from any web browser without installation and enter their symptoms, record audio clips of their cough and breath sounds, and upload the data anonymously. Two different pipelines for screening were developed based on the symptoms reported by the users: asymptomatic and symptomatic. An innovative and novel stacking CNN model was developed using three base learners from of eight state-of-the-art deep learning CNN algorithms. The stacking CNN model is based on a logistic regression classifier meta-learner that uses the spectrograms generated from the breath and cough sounds of symptomatic and asymptomatic patients as input using the combined (Cambridge and collected) dataset. Results-The stacking model outperformed the other eight CNN networks with the best classification performance for binary classification using cough sound spectrogram images. The accuracy, sensitivity, and specificity for symptomatic and asymptomatic patients were 96.5%, 96.42%, and 95.47% and 98.85%, 97.01%, and 99.6%, respectively. For breath sound spectrogram images, the metrics for binary classification of symptomatic and asymptomatic patients were 91.03%, 88.9%, and 91.5% and 80.01%, 72.04%, and 82.67%, respectively. Conclusion-The web-application QUCoughScope records coughing and breathing sounds, converts them to a spectrogram, and applies the best-performing machine learning model to classify the COVID-19 patients and healthy subjects. The result is then reported back to the test user in the application interface. Therefore, this novel system can be used by patients in their premises as a pre-screening method to aid COVID-19 diagnosis by prioritizing the patients for RT-PCR testing and thereby reducing the risk of spreading of the disease.Entities:
Keywords: COVID-19; artificial intelligence; cough and breath sounds; crowdsourcing application; deep learning; pre-screening; spectrogram
Year: 2022 PMID: 35453968 PMCID: PMC9028864 DOI: 10.3390/diagnostics12040920
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Figure 1Methodology of the study.
Details of the total Dataset.
| Experiments | Healthy | COVID-19 | ||
|---|---|---|---|---|
| Cambridge | QU | Cambridge | QU | |
| Symptomatic (Cough/Breath) | 264 | 32 | 54 | 18 |
| Asymptomatic (Cough/Breath) | 318 | 213 | 87 | 78 |
| Total | 582 | 245 | 141 | 96 |
Experimental pipelines for this study.
| Pipelines | COVID-19 | Healthy |
|---|---|---|
| Pipeline I |
Cough Breath |
Cough Breath |
| Pipeline II |
Cough Breath |
Cough Breath |
Figure 2Cough and breath sound waveforms and spectrograms for (A) symptomatic and (B) asymptomatic healthy subjects and COVID-19 patients.
Number of mages per class and per fold used for different pipelines.
| Categories | Classes | Total Samples | Training Samples | Validation | Test |
|---|---|---|---|---|---|
| Symptomatic | Healthy | 296 | 213 × 10 = 2130 | 24 | 59 |
| COVID-19 | 72 | 52 × 38 = 1976 | 6 | 14 | |
| Asymptomatic | Healthy | 531 | 383 × 5 = 1915 | 42 | 106 |
| COVID-19 | 165 | 119 × 17 = 2023 | 13 | 33 |
Details of training parameters for classification.
| Training Parameters for Classification | ||||||
|---|---|---|---|---|---|---|
| Batch Size | Learning Rate | Number of Epochs | Epoch Patience | Stopping Criteria | Optimizer | |
| Parameters | 32 | 0.001 | 30 | 15 | 15 | ADAM |
Figure 3Stacking model architecture.
Comparison of different CNN performances for binary classification for symptomatic and asymptomatic patients’ (A) cough and (B) breath sounds.
|
| |||||||
|
|
|
|
|
| |||
|
|
|
|
|
| |||
| Symptomatic | Resnet18 | 93.20 ± 2.57 | 93.65 ± 2.49 | 93.21 ± 2.57 | 93.35 ± 2.55 | 89.94 ± 3.07 | 0.0024 |
| Resnet50 | 95.38 ± 2.14 | 95.41 ± 2.14 | 95.38 ± 2.14 | 95.39 ± 2.14 | 90.47 ± 3.00 | 0.0061 | |
| Resnet101 | 94.29 ± 2.37 | 95.41 ± 2.14 | 94.29 ± 2.37 | 94.53 ± 2.32 | 97.56 ± 1.58 | 0.0108 | |
| Inception_v3 | 90.76 ± 2.96 | 91.53 ± 2.84 | 90.76 ± 2.96 | 91.02 ± 2.92 | 86.19 ± 3.52 | 0.0238 | |
| DenseNet201 | 93.25 ± 2.56 | 93.78 ± 2.47 | 93.21 ± 2.57 | 93.39 ± 2.54 | 90.99 ± 2.93 | 0.0258 | |
| Mobilenetv2 | 90.49 ± 3.00 | 90.78 ± 2.96 | 90.49 ± 3.00 | 90.61 ± 2.98 | 81.92 ± 3.93 | 0.0055 | |
| EfficientNet_B0 | 90.20 ± 2.89 | 90.15 ± 2.90 | 91.30 ± 2.88 | 91.20 ± 2.89 | 78.97 ± 4.16 | 0.0106 | |
| EfficientNet_B7 | 91.30 ± 2.88 | 91.40 ± 2.86 | 91.31 ± 2.88 | 91.35 ± 2.87 | 82.12 ± 3.92 | 0.0428 | |
|
|
|
|
|
|
|
| |
| Asymptomatic | Resnet18 | 96.70 ± 1.33 | 96.68 ± 1.33 | 96.69 ± 1.33 | 96.66 ± 1.33 | 92.29 ± 1.98 | 0.0027 |
| Resnet50 | 94.97 ± 1.62 | 95.12 ± 1.60 | 94.98 ± 1.62 | 94.80 ± 1.65 | 85.07 ± 2.65 | 0.0058 | |
| Resnet101 | 96.84 ± 1.30 | 96.84 ± 1.30 | 96.84 ± 1.30 | 96.84 ± 1.30 | 94.42 ± 1.71 | 0.0121 | |
| Inception_v3 | 96.26 ± 1.41 | 96.30 ± 1.40 | 96.27 ± 1.41 | 96.19 ± 1.42 | 89.65 ± 2.26 | 0.0235 | |
| DenseNet201 | 98.28 ± 0.97 | 98.27 ± 0.97 | 96.28 ± 1.41 | 97.11 ± 1.24 | 99.20 ± 0.66 | 0.0260 | |
| Mobilenetv2 | 98.50 ± 0.90 | 98.30 ± 0.96 | 96.45 ± 1.37 | 97.25 ± 1.21 | 99.20 ± 0.66 | 0.0052 | |
| EfficientNet_B0 | 93.82 ± 1.79 | 93.74 ± 1.80 | 93.82 ± 1.79 | 93.72 ± 1.80 | 85.96 ± 2.58 | 0.0118 | |
| EfficientNet_B7 | 95.40 ± 1.56 | 95.40 ± 1.56 | 95.40 ± 1.56 | 95.31 ± 1.57 | 88.13 ± 2.40 | 0.046 | |
|
|
|
|
|
|
|
| |
|
| |||||||
|
|
|
|
|
| |||
|
|
|
|
|
| |||
| Symptomatic | Resnet18 | 81.49 ± 3.97 | 70.27 ± 4.67 | 82.27 ± 3.90 | 75.80 ± 4.38 | 81.49 ± 3.97 | 0.0027 |
| Resnet50 | 80.66 ± 4.04 | 70.83 ± 4.64 | 81.83 ± 3.94 | 75.93 ± 4.37 | 80.67 ± 4.03 | 0.0060 | |
| Resnet101 | 84.53 ± 3.69 | 73.01 ± 4.54 | 84.01 ± 3.74 | 78.12 ± 4.22 | 84.53 ± 3.69 | 0.0098 | |
| Inception_v3 | 81.49 ± 3.97 | 71.05 ± 4.63 | 82.05 ± 3.92 | 76.15 ± 4.35 | 81.49 ± 3.97 | 0.0254 | |
| DenseNet201 | 83.98 ± 3.75 | 72.43 ± 4.57 | 83.43 ± 3.8 | 77.54 ± 4.26 | 83.98 ± 3.75 | 0.026 | |
| Mobilenetv2 | 87.57 ± 3.37 | 69.50 ± 4.7 | 87.50 ± 3.38 | 77.47 ± 4.27 | 87.57 ± 3.37 | 0.0048 | |
| EfficientNet_B0 | 90.33 ± 3.02 | 70.28 ± 4.67 | 90.28 ± 3.03 | 79.03 ± 4.16 | 90.33 ± 3.02 | 0.0104 | |
| EfficientNet_B7 | 81.77 ± 3.94 | 70.99 ± 4.64 | 81.99 ± 3.93 | 76.09 ± 4.36 | 81.77 ± 3.94 | 0.0434 | |
|
|
|
|
|
|
|
| |
| Asymptomatic | Resnet18 | 66.75 ± 3.50 | 53.95 ± 3.7 | 66.66 ± 3.50 | 59.64 ± 3.64 | 78.54 ± 3.05 | 0.0025 |
| Resnet50 | 66.67 ± 3.50 | 55.45 ± 3.69 | 66.67 ± 3.50 | 60.54 ± 3.63 | 75.27 ± 3.21 | 0.0047 | |
| Resnet101 | 69.72 ± 3.41 | 56.45 ± 3.68 | 69.71 ± 3.41 | 62.38 ± 3.60 | 73.52 ± 3.28 | 0.0118 | |
| Inception_v3 | 67.10 ± 3.49 | 57.10 ± 3.68 | 68.26 ± 3.46 | 62.18 ± 3.60 | 81.25 ± 2.90 | 0.0243 | |
| DenseNet201 | 67.97 ± 3.47 | 55.91 ± 3.69 | 67.97 ± 3.47 | 61.35 ± 3.62 | 79.88 ± 2.98 | 0.0271 | |
| MobileNetv2 | 68.40 ± 3.45 | 53.22 ± 3.71 | 67.10 ± 3.49 | 59.36 ± 3.65 | 78.54 ± 3.05 | 0.0048 | |
| EfficientNet_B0 | 68.30± 3.46 | 57.45 ± 3.67 | 68.62 ± 3.45 | 62.54 ± 3.60 | 76.50 ± 3.15 | 0.0128 | |
| EfficientNet_B7 | 75.60 ± 3.19 | 54.20 ± 3.70 | 72.59 ± 3.31 | 62.06 ± 3.61 | 80.20 ± 2.96 | 0.0511 | |
|
|
|
|
|
|
|
| |
Figure 4ROC curve for healthy and COVID-19 patients’ classification using cough sounds for (A) symptomatic patients and (B) asymptomatic patients, and using breath sounds for (C) symptomatic patients and (D) asymptomatic patients.
Figure 5Confusion matrices for healthy and COVID-19 classification using cough sounds for (A) symptomatic patients and (B) asymptomatic patients, and using breath sounds for (C) symptomatic patients and (D) asymptomatic patients using best performing stacking CNN models.
Figure 6Accuracy vs inference time plot for binary classification using (A) symptomatic cough sound spectrograms, and (B) asymptomatic cough sound spectrograms.
Figure 7Illustration of a generic framework for the QUCoughScope application.
Comparison of the proposed work with similar studies.
| Papers | Dataset | Phenomenon | Reported Method | Performance |
|---|---|---|---|---|
| N. Sharma (2020) | Healthy and COVID-19-positive: 941 | Cough, Breathing, Vowel, and Counting (1–20) | Random forest classifier using spectral contrast, MFCC, spectral roll-off, spectral centroid, mean square energy, polynomial fit, zero-crossing rate, spectral bandwidth, and spectral flatness. | Accuracy: 76.74% |
| C. Brown et al. (2021) | COVID-19-positive: 141, | Cough and Breathing | CNN-based approach using spectrogram, spectral centroid, MFCC. | Accuracy: 80% |
| V. Espotovic (2021) | COVID-19-Positive: 84, COVID-19-Negative: 419 | Cough and Breathing | Ensemble-boosted approach using spectrogram and wavelet. | Accuracy: 88.52% |
| R.Islam (2022) | COVID-19-Positve: 50, | Cough | CNN-based approach using zero-crossing rate, energy, energy entropy, spectral centroid, spectral entropy, spectral flux, spectral roll-offs, MFCC. | Accuracy: 88.52% |
| Proposed Study | COVID-19-Positve: 237, Healthy: 827 | Cough and Breathing | Stacking-based CNN based approach using spectograms | For symptomatic, accuracy: 96.5% and for asymptomatic, accuracy: 98.85% |