| Literature DB >> 36111113 |
Saleh Albahli1, Tahira Nazir2.
Abstract
Machine learning techniques have lately attracted a lot of attention for their potential to execute expert-level clinical tasks, notably in the area of medical image analysis. Chest radiography is one of the most often utilized diagnostic imaging modalities in medical practice, and it necessitates timely coverage regarding the presence of probable abnormalities and disease diagnoses in the images. Computer-aided solutions for the identification of chest illness using chest radiography are being developed in medical imaging research. However, accurate localization and categorization of specific disorders in chest X-ray images is still a challenging problem due to the complex nature of radiographs, presence of different distortions, high inter-class similarities, and intra-class variations in abnormalities. In this work, we have presented an Artificial Intelligence (AI)-enabled fully automated approach using an end-to-end deep learning technique to improve the accuracy of thoracic illness diagnosis. We proposed AI-CenterNet CXR, a customized CenterNet model with an improved feature extraction network for the recognition of multi-label chest diseases. The enhanced backbone computes deep key points that improve the abnormality localization accuracy and, thus, overall disease classification performance. Moreover, the proposed architecture is lightweight and computationally efficient in comparison to the original CenterNet model. We have performed extensive experimentation to validate the effectiveness of the proposed technique using the National Institutes of Health (NIH) Chest X-ray dataset. Our method achieved an overall Area Under the Curve (AUC) of 0.888 and an average IOU of 0.801 to detect and classify the eight types of chest abnormalities. Both the qualitative and quantitative findings reveal that the suggested approach outperforms the existing methods, indicating the efficacy of our approach.Entities:
Keywords: CenterNet; DenseNet; chest X-ray images; deep learning; localization
Year: 2022 PMID: 36111113 PMCID: PMC9469020 DOI: 10.3389/fmed.2022.955765
Source DB: PubMed Journal: Front Med (Lausanne) ISSN: 2296-858X
A comparison of the multi-class chest disease diagnosis.
|
|
|
|
|
|---|---|---|---|
| Ayan and Ünver ( | VGG16 | Accuracy = 0.87% (VGG16) | The accuracy can be improved by combining features from both networks |
| Bhandary et al. ( | Modified AlexNet (MAN) and Haralick and Hu approach | Accuracy = 97.27% | The generalization performance of the model can be enhanced |
| Tataru et al. ( | GoogLeNet, InceptionNet, and ResNet | Accuracy = 80%, F1 score of 0.66 | The performance can be improved by the inclusion of a segmentation approach to allow the network to learn more disease-specific attributes |
| Rajpurkar et al. ( | Novel CNN (121-layer) | F1 score = 43.5% and AUROC = 0.801 | Performance requires further improvement |
| Albahli ( | Novel CNN | Accuracy = 87% | Performance needs improvement |
| Ho and Gwak ( | A hybrid model with a DenseNet-121 network and hand-crafted feature extractor i.e., SIFT, HOG, LBP, GIST, and different ML classifiers such as SVM, KNN, AdaBoost, and others | Accuracy = 0.8462, F1-score = 0.9413, AUC = 0.8097 | Requires improvement in the generalization ability of the model |
| Abiyev and Ma'aita ( | Novel CNN | Accuracy = 92.4% | The model can be made deeper to enhance performance |
| Xu et al. ( | Densenet169 with multi-scale attention network | AUROC = 0.850 | The performance can be improved further |
| AUROC = 0.815 | |||
| Ma et al. ( | Densenet121 and densenet169 with cross attention | AUROC = 0.817 | The model is computationally complex |
| AUROC = 0.775 | |||
| Wang and Xia ( | ResNet-152 with attention network | AUC = 0.781 | The model is computationally complex and suffers from high inference time |
| Ouyang et al. ( | ResNet with a hierarchical visual attention mechanism | AUC = 0.819 | The model is dependent on the availability of box annotations |
| AUC = 0.9166 | |||
| Pan et al. ( | DenseNet and MobileNetV2 | AUROC = 0.924 | The generalizability of the model requires improvement |
| AUROC = 0.900 | |||
| Albahli and Yar ( | ResNet50, NasNetLarge, Xception, InceptionV3, and InceptionResNetV2 | AUC = 96.9, | The images were segmented before the classification |
| Alqudah et al. ( | Novel CNN with softmax classifier, SVM, and KNN | Accuracy = 94%, | Performed classification between Normal vs. Bacterial Pneumonia vs. Viral Pneumonia classes |
| Kim et al. ( | EfficientNetv2 | Accuracy = 82.15%, Sensitivity = 81.40%, Specificity = 91.65% | The evaluation was performed on 4 classes only Pneumonia, Pneumothorax, Tuberculosis, and Normal class |
| Baltruschat et al. ( | ResNet38, ResNet50, ResNet101 | AUC = 0.822 | The performance can be improved further |
| Ibrahim et al. ( | CustomVGG19, ResNet152V2, ResNet152V2-GRU, and ResNet152V2-BiGRU | Accuracy = 98.05%, Recall = 98.05%, Specificity = 99.5%, F1-score = 98.24%, AUC = 99.66% | The model is evaluated only using COVID-19, Pneumonia, Lung Cancer, and Normal classes |
| Ge et al. ( | ResNet and DenseNet with novel multi-loss function | AUC = 0.8398 (ResNet) | The model is evaluated using only four classes |
Figure 1Flow diagram of proposed method namely AI CenterNet CXR.
Figure 2The architectural view of ResNet-101.
Description of DenseNet-41.
|
|
| ||
|---|---|---|---|
|
|
| ||
| Con L1 | 7 × 7 | 2 | |
| Pool L1 | 3 × 3 max_ | 2 | |
| Dense B1 |
| 1 | |
| Transition L1 | Con L2 | 1 × 1 | 1 |
| Pool L2 | 2 × 2 | 2 | |
| Dense B2 |
| 1 | |
| Transition L2 | Con L3 | 1 × 1 | 1 |
| Pool L3 | 2 × 2 | 2 | |
| Dense B3 |
| 1 | |
| Transition L3 | Con L4 | 1 × 1 | 1 |
| Pool L4 | 2 × 2 | 2 | |
| Dense B4 |
| 1 | |
| Classification Layer | 7 × 7 | ||
| FC layer | |||
| SoftMax | |||
Figure 3Dense block and transition layer.
Figure 4A pictorial view of sample information from the NIH Chest X-ray dataset (47).
Figure 5Samples images of NIH CXR dataset.
Figure 6Visual depiction of IOU metric.
Figure 8Pictorial representation of Recall measure.
Figure 9Localization results of the proposed method.
Figure 10Class-wise precision values for the custom CenterNet model.
Figure 11Class-wise AP and recall values for the proposed Custom CenterNet approach.
Figure 12Class-wise F1-Score along with the error rate for CXR diseases classification using custom CenterNet model.
Figure 13Confusion matrix obtained for CXR disease classification with the custom CenterNet.
Figure 14Class-wise accuracy values.
Comparison with base models in terms of the AUC metric.
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|
| AlexNet | 0.64 | 0.69 | 0.66 | 0.60 | 0.56 | 0.65 | 0.55 | 0.74 |
| GoogLeNet | 0.630 | 0.70 | 0.69 | 0.61 | 0.54 | 0.56 | 0.59 | 0.78 |
| VGG16 | 0.63 | 0.71 | 0.65 | 0.59 | 0.51 | 0.65 | 0.51 | 0.63 |
| ResNet50 | 0.71 | 0.81 | 0.74 | 0.61 |
| 0.72 | 0.63 | 0.79 |
| Proposed |
|
|
|
|
|
|
|
|
The bold means highest AUC metric.
Comparative comparison with base models.
|
|
|
|
|
|
|---|---|---|---|---|
| AlexNet | 65.00% | 66.14% | 67.45% | 65.57% |
| GoogLeNet | 69.53% | 71.88% | 70.35% | 70.69% |
| VGG16 | 72.00% | 74.32% | 75.41% | 73.14% |
| ResNet-50 | 77.00% | 75.00% | 77.63% | 75.99% |
| Inception V4 | 79.32% | 75.65% | 79.32% | 79.22% |
| DenseNet-121 | 83.01% | 81.84% | 83.21% | 82.87% |
| EfficientNet | 87.74% | 88.95% | 88.01% | 87.61% |
| Proposed | 89.00% | 91.00% | 92.21% | 89.99% |
Comparison with object detection models.
|
|
|
|
|
|---|---|---|---|
| Fast-RCNN | VGG-16 | 0.65 | 0.28 |
| Faster-RCNN | VGG-16 | 0.77 | 0.25 |
| Mask-RCNN | ResNet-101 | 0.79 | 0.23 |
| RetinaNet | ResNet-101 | 0.63 | 0.27 |
| YOLO | ResNet-50 | 0.76 | 0.22 |
| CenterNet | ResNet-101 | 0.82 | 0.25 |
|
| DenseNet-41 | 0.91 | 0.21 |
Comparison with ML-based classifiers.
|
|
|
|---|---|
| SVM ( | 0.745 |
| KNN ( | 0.721 |
|
|
|
The bold means highest AUC metric.
Comparison of latest approaches in terms of the AUC metric.
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|
| Wang et al. ( | 0.73 | 0.84 | 0.79 | 0.67 | 0.73 | 0.69 | 0.72 | 0.85 |
| Kumar et al. ( | 0.76 | 0.91 | 0.86 | 0.69 | 0.75 | 0.67 | 0.72 | 0.86 |
| Liu et al. ( | 0.79 | 0.87 | 0.88 | 0.69 | 0.81 | 0.73 | 0.75 | 0.89 |
| Seyyed-Kalantari et al. ( | 0.81 | 0.92 | 0.87 | 0.72 | 0.83 | 0.78 | 0.76 | 0.88 |
| Han et al. ( | 0.84 | 0.93 | 0.88 | 0.72 | 0.87 | 0.79 | 0.77 |
|
|
|
|
|
|
|
|
|
| 0.88 |
The bold means highest AUC metric.
Comparison of latest techniques in terms of the IOU metric.
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|
| Wang et al. ( | 0.69 | 0.94 | 0.66 | 0.71 | 0.40 | 0.14 | 0.63 | 0.38 |
| Han et al. ( | 0.72 | 0.96 | 0.88 | 0.93 | 0.74 | 0.45 | 0.65 | 0.64 |
| Li et al. ( | 0.71 | 0.98 | 0.87 | 0.92 | 0.71 | 0.40 | 0.60 | 0.63 |
|
|
|
|
|
|
|
|
|
|
The bold means highest IOU metric.