| Literature DB >> 35489595 |
Zhiang Chen1, Muyun Li1, Ruoyu Wang1, Wenzhuo Sun1, Jiayi Liu1, Haiyang Li1, Tianxin Wang1, Yuan Lian1, Jiaqian Zhang1, Xinheng Wang2.
Abstract
Scientific evidence shows that acoustic analysis could be an indicator for diagnosing COVID-19. From analyzing recorded breath sounds on smartphones, it is discovered that patients with COVID-19 have different patterns in both the time domain and frequency domain. These patterns are used in this paper to diagnose the infection of COVID-19. Statistics of the sound signals, analysis in the frequency domain, and Mel-Frequency Cepstral Coefficients (MFCCs) are then calculated and applied in two classifiers, k-Nearest Neighbors (kNN) and Convolutional Neural Network (CNN), to diagnose whether a user is contracted with COVID-19 or not. Test results show that, amazingly, an accuracy of over 97% could be achieved with a CNN classifier and more than 85% on kNN with optimized features. Optimization methods for selecting the best features and using various metrics to evaluate the performance are also demonstrated in this paper. Owing to the high accuracy of the CNN model, the CNN model was implemented in an Android app to diagnose COVID-19 with a probability to indicate the confidence level. The initial medical test shows a similar test result between the method proposed in this paper and the lateral flow method, which indicates that the proposed method is feasible and effective. Because of the use of breath sound and tested on the smartphone, this method could be used by everybody regardless of the availability of other medical resources, which could be a powerful tool for society to diagnose COVID-19.Entities:
Keywords: Acoustic analysis; Breath sound; COVID-19; Convolutional Neural Network (CNN); k-Nearest Neighbors (kNN)
Mesh:
Year: 2022 PMID: 35489595 PMCID: PMC9044719 DOI: 10.1016/j.jbi.2022.104078
Source DB: PubMed Journal: J Biomed Inform ISSN: 1532-0464 Impact factor: 8.000
Fig. 1(a) Healthy breath sound in time domain (b) Spectrum of healthy breath sound.
Fig. 2(a) Breath sound with COVID-19 in time domain (no symptom) (b) Spectrum of breath sound with COVID-19.
Fig. 3(a) Breath sound with COVID-19 in time domain (showing symptom) (b) Spectrum of breath sound with COVID-19.
Fig. 4Classification Flow Chart.
Fig. 5(a) Original signal (b) De-noised signal.
Fig. 6(a) MFCCs of healthy breath sound (b) MFCCs of sound with COVID-19.
Fig. 7Feature Selection.
Fig. 8Minimum Classification Error Plot.
Fig. 9Architecture of CNN Model.
Imbalanced Training Data.
| COVID-19 Status | Negative | Positive | Total |
|---|---|---|---|
| Number of Audio Chunks | 3974 | 2774 | 6748 |
| Percentage | 58.9% | 41.1% | 100.0% |
Fig. 10Screenshot of the application.
kNN Model Performance.
| Model | Number of features selected | Accuracy | AUC | TPR | TNR | |||
|---|---|---|---|---|---|---|---|---|
| 1 | 26 | 70.0 | 0.78 | 53.6 | 80.8 | |||
| 2 | 14 | 80.1 | 0.80 | 71.2 | 85.9 | |||
| 3 | 14 | 78.0 | 0.83 | 82.4 | 75.2 |
Fig. 11Validation ROC Curve of (a) Model 1 (b) Model 2 (c) Model 3.
Model Performance with Different Pre-processing Steps.
| Input Feature | Pre-emphasized | Normalized | Test Accuracy | Optimized Epoch Number | AUC | Sensitivity | Specificity | Precision |
|---|---|---|---|---|---|---|---|---|
| Raw Data | N | N | 60.85% | 84 | 0.6877 | 54.45% | 67.32% | 66.85% |
| MFCC | N | N | 97.51% | 90 | 0.9972 | 94.10% | 99.80% | 99.69% |
| MFCC | Y | N | 97.04% | 64 | 0.9969 | 93.62% | 99.59% | 99.41% |
| MFCC | N | Y | 97.57% | 80 | 94.19% | |||
| MFCC | Y | Y | 0.9977 | 99.80% | 99.70% |
Fig. 12Training Loss and Validation Loss with (a) Original 13-dimensional MFCC, and (b) Pre-emphasized and Normalized 13-dimensional MFCC.
Model Performance with or without Noise Reduction.
| Pre-emphasized & Normalized | De-noised | Test Accuracy | Optimized Epoch Number | AUC | Sensitivity | Specificity | Precision | |
|---|---|---|---|---|---|---|---|---|
| Y | N | 97.63% | 61 | 0.9977 | 94.55% | 99.70% | ||
| Y | Y | 99.79% |
Model Performance with Different Lengths of Audio Chunks (ACs).
| Length of ACs (s) | Total Number of ACs | Test Accuracy | AUC | Sensitivity | Specificity | Precision |
|---|---|---|---|---|---|---|
| 1 | 21789 | 96.04% | 0.9944 | 91.93% | 98.99% | 98.50% |
| 2 | 10529 | 96.70% | 0.9964 | 92.85% | 99.36% | 99.01% |
| 3 | 6748 | 0.9978 | 95.26% | 99.79% | 99.71% | |
| 4 | 4943 | 95.95% | 93.39% | 90.98% | ||
| 5 | 3778 | 97.57% | 0.9982 | 93.98% | ||
| 6 | 2970 | 96.64% | 0.9884 | 98.37% | 95.41% | 93.79% |
Model Performance with Different Features (Each Model is Trained 6 Times).
| Features | Test Accuracy | AUC | Sensitivity | Specificity | Precision |
|---|---|---|---|---|---|
| MFCC | 97.87% | 0.9978 | 95.26% | 99.79% | 99.71% |
| 97.04% | 0.9980 | 93.16% | 100.00% | 100.00% | |
| 97.45% | 0.9979 | 94.45% | 99.50% | 99.23% | |
| MFCC | 97.39% | 0.9979 | 93.61% | 100.00% | 100.00% |
| Log-mel Spectrum | 97.21% | 0.9976 | 93.75% | 99.79% | 99.70% |
Fig. 13Test results and lateral flow.