| Literature DB >> 34007303 |
Zahra Nabizadeh-Shahre-Babak1, Nader Karimi1, Pejman Khadivi2, Roshanak Roshandel2, Ali Emami1, Shadrokh Samavi1,3.
Abstract
Coronavirus disease 2019 (COVID-19) was classified as a pandemic by the World Health Organization in March 2020. Given that this novel virus most notably affects the human respiratory system, early detection may help prevent severe lung damage, save lives, and help prevent further disease spread. Given the constraints on the healthcare facilities and staff, the role of artificial intelligence for automatic diagnosis is critical. The automatic diagnosis of COVID-19 based on medical images is, however, not straightforward. Due to the novelty of the disease, available X-ray datasets are very limited. Furthermore, there is a significant similarity between COVID-19 X-rays and other lung infections. In this paper, these challenges are addressed by proposing an approach consisting of a bag of visual words and a neural network classifier. The proposed method can classify X-ray chest images into non-COVID-19 and COVID-19 with high performance. Three public datasets are used to evaluate the proposed approach. Our best accuracy on the first, second, and third datasets is 96.1, 99.84, and 98 percent. Since detection of COVID-19 is important, sensitivity is used as a criterion. The proposed method's best sensitivities are 90.32, 99.65, and 91 percent on these datasets, respectively. The experimental results show that extracting features with the bag of visual words results in better classification accuracy than the state-of-the-art techniques.Entities:
Keywords: Bag of Visual Words; COVID-19; Classifier; Coronavirus
Year: 2021 PMID: 34007303 PMCID: PMC8120450 DOI: 10.1016/j.bspc.2021.102750
Source DB: PubMed Journal: Biomed Signal Process Control ISSN: 1746-8094 Impact factor: 3.880
Fig. 1Block diagram of the test phase of the proposed method.
Fig. 2(a) The average histogram of train images, (b) CDF of image before transformation (blue), CDF of the average histogram (black) and CDF of the image after transformation (red), (c) original image with black background, (d) the output image after transformation.
Fig. 3(a) The original image and it histogram, (b) the image after histogram matching block and its histogram, (c) the image and its histogram after intensity enhancement.
Fig. 4(a) the image of non−COVID19, (b) the image of a COVID-19 case.
The details of dataset.
| Dataset | COVID /Non-COVID Train | COVID /Non-COVID Test | Version |
|---|---|---|---|
| [ | 152/13,482 | 31/200 | 3 |
| [ | 572/1020 | 571/1021 | January-2021 |
| [ | 84/2000 | 100/3000 | January-2021 |
Confusion matrices for tests with and without histogram matching.
| With histogram matching | Predict value | ||
|---|---|---|---|
| Actual value | non-COVID | COVID | |
| non-COVID | 0.965 | 0.035 | |
| COVID | 0.097 | 0.903 | |
Fig. 5Keypoints selected by (a) grid-based method, and (b) detector-based method.
Accuracy, sensitivity and specificity of classification for different values of strongest features.
| Strongest Features | Accuracy | Sensitivity | Specificity |
|---|---|---|---|
| 0.5 | 0.9134 | 0.7096 | 0.945 |
| 0.8 | 0.9523 | 0.8709 | 0.965 |
| 1 | 0.9220 | 0.8387 | 0.935 |
Confusion matrix for different dictionary sizes.
| Dictionary Size 250 | Predicted value | Dictionary Size 5000 | Predicted value | ||||
|---|---|---|---|---|---|---|---|
| Actual value | non-COVID | COVID | Actual value | non-COVID | COVID | ||
| non-COVID | 0.92 | 0.09 | non-COVID | 0.96 | 0.04 | ||
| COVID | 0.23 | 0.77 | COVID | 0.129 | 0.871 | ||
Fig. 6Histograms of the most frequent visual words for (a) two overlapped histograms of inside of the lungs, and (b) outside regions of lungs. Visual words from COVID-19 images shown by red, and non-COVID-19 by blue.
Results produced by different classifiers.
| Classifier type | Accuracy | Sensitivity | Specificity |
|---|---|---|---|
| LogisticRegression | 0.939 | 0.838 | 0.955 |
| LinearSVC | 0.944 | 0.806 | 0.965 |
| KNeighborsClassifier | 0.913 | 0.451 | 0.985 |
| LinearDiscriminantAnalysis | 0.935 | 0.581 | 0.99 |
| GaussianNB | 0.908 | 0.484 | 0.965 |
| DecisionTreeClassifier | 0.757 | 0.419 | 0.81 |
| RandomForestClassifier | 0.922 | 0.516 | 0.985 |
| FCN | 0.957 | 0.903 | 0.965 |
The details of method for some papers in this field.
| paper | #COVID test | #non COVID test | Feature Extraction | Classifier | #Class | Best ACC |
|---|---|---|---|---|---|---|
| [ | 25 | 25 | ResNet50 | SVM | 2 | 95.38 |
| [ | – | – | COVID-CAPS | FCN | 2 | 95.7 |
| [ | 100 | 200 | COVID_Net | FCN | 3 | 93.3 |
| [ | 25 | 50 | Pre-trained CNN Models | SVM | 3 | 95.33 |
| [ | 25 | 50 | GLCM, HOG, LBP, | SVM | 3 | 93.5 |
| [ | 68 | 560 | ResNet | FCN | 2 | 96.1 |
| [ | 68 | 560 | InceptionV3 | FCN | 2 | 95.4 |
| [ | 162 | – | Truncated Inception Net | FCN | 2 | 98.77 |
| [ | 864 | 2686 | InceptionV3 | FCN | 3 | 96 |
| [ | 25 | 200 | Xception | FCN | 3 | 97.4 |
| [ | 455 | 3450 | MobileNetV2 | FCN | 2 | 99.1 |
| [ | 20 | 20 | VGG16, VGG19, ResNet, DenseNet, InceptionV3 | FCN | 2 | 80 |
| [ | 30 | – | Deep CNN | FCN | 2 | 93 |
| [ | 100 | 3000 | ResNet18, ResNet50, SqueezeNet, and DenseNet-121 | FCN | 2 | 94 |
| [ | 44 | 537 | AlexNet, GoogLeNet, SqueezeNet | FCN | 2 | 99.85 |
| Our | 31 | 200 | Bag of Visual Words | SVM, FCN, | 2 | 96.1 |
| Our | 571 | 1021 | Bag of Visual Words | SVM, FCN, | 2 | 99.84 |
| Our | 100 | 3000 | Bag of Visual Words | SVM, FCN, | 2 | 98 |
The article only separates COVID-19 samples from normal samples not all of the Pneumonia samples.
The articles have not mentioned the percentage of test and train images, so the whole number of images is reported.
The article reported number of patients not the number of images.
Results of methods in [2,3,18,19] and the proposed method.
| Dataset | Approach | Accuracy | Sensitivity | Specificity |
|---|---|---|---|---|
| [ | [ | 95.38 | 97.29 | 93.47 |
| [ | 95.7 | 90 | 95.8 | |
| Proposed method | 96.1 | 87.09 | 97.5 | |
| [ | [ | 99.22 | 99.14 | 99.26 |
| Proposed method | 99.84 | 99.65 | 99.95 | |
| [ | [ | 92.29 | 98 | 92.9 |
| Proposed method | 98 | 91 | 98.23 |