| Literature DB >> 36042937 |
Sima Sarv Ahrabi1, Alireza Momenzadeh1, Enzo Baccarelli1, Michele Scarpiniti1, Lorenzo Piazzo1.
Abstract
Bidirectional generative adversarial networks (BiGANs) and cycle generative adversarial networks (CycleGANs) are two emerging machine learning models that, up to now, have been used as generative models, i.e., to generate output data sampled from a target probability distribution. However, these models are also equipped with encoding modules, which, after weakly supervised training, could be, in principle, exploited for the extraction of hidden features from the input data. At the present time, how these extracted features could be effectively exploited for classification tasks is still an unexplored field. Hence, motivated by this consideration, in this paper, we develop and numerically test the performance of a novel inference engine that relies on the exploitation of BiGAN and CycleGAN-learned hidden features for the detection of COVID-19 disease from other lung diseases in computer tomography (CT) scans. In this respect, the main contributions of the paper are twofold. First, we develop a kernel density estimation (KDE)-based inference method, which, in the training phase, leverages the hidden features extracted by BiGANs and CycleGANs for estimating the (a priori unknown) probability density function (PDF) of the CT scans of COVID-19 patients and, then, in the inference phase, uses it as a target COVID-PDF for the detection of COVID diseases. As a second major contribution, we numerically evaluate and compare the classification accuracies of the implemented BiGAN and CycleGAN models against the ones of some state-of-the-art methods, which rely on the unsupervised training of convolutional autoencoders (CAEs) for attaining feature extraction. The performance comparisons are carried out by considering a spectrum of different training loss functions and distance metrics. The obtained classification accuracies of the proposed CycleGAN-based (resp., BiGAN-based) models outperform the corresponding ones of the considered benchmark CAE-based models of about 16% (resp., 14%).Entities:
Keywords: BiGAN; CAE; COVID-19 detection; Complexity-vs.-accuracy comparisons; CycleGAN; Hidden feature extraction; Unsupervised-vs.-weakly supervised learning
Year: 2022 PMID: 36042937 PMCID: PMC9411851 DOI: 10.1007/s11227-022-04775-y
Source DB: PubMed Journal: J Supercomput ISSN: 0920-8542 Impact factor: 2.557
Synoptic view of recent work on GAN-based COVID-19 detection
| Ref | Approach | Goal | Training class |
|---|---|---|---|
| [ | CovidGAN | X-ray synthetic augmentation | COVID |
| [ | GAN +transfer learning | X-ray synthetic augmentation | Normal+COVID+Pneumonia |
| [ | Dense GAN+ U-Net | Enhancing the quality of CT | COVID |
| [ | Convolutional GAN | X-ray synthetic augmentation | Normal+COVID+Pneumonia |
| [ | AC-GAN | CT synthetic augmentation | COVID |
| [ | MTT-GAN | X-ray synthetic augmentation | COVID |
| [ | IAGAN | X-ray synthetic augmentation | COVID |
| [ | GAN | X-ray synthetic augmentation | Normal+COVID+Pneumonia |
| [ | Siamese-CapsNet | Image resolution enhancement | COVID+NON-COVID |
| [ | GAN | Image resolution enhancement | COVID+NON-COVID |
| BiGAN & CycleGAN | CT Classification | COVID |
Fig. 1a Four representative samples of lung CT scans from the training sets of size (). b Representative samples of lung CT feature maps extracted by the implemented BiGANs and CycleGAN
Implemented unsupervised and weakly supervised models for COVID-19 detection. HB-DCAE: Histogram-Based DCAE [38]; PB-CAE: PDF-Based CAE [57]; US: Un-Supervised; WS: Weakly Supervised; TRIM: Number of TRaining IMages; TSIM: Number of TeSt IMages
| Model | Model training | Size of the input images | #TRIM | #TSIM | Size of the extract features | Resulting best test accuracy |
|---|---|---|---|---|---|---|
| HB-DCAE | US | 1000 | 1000 | 0.8270 | ||
| PB-CAE | US | 1000 | 1000 | 0.7610 | ||
| CE-BiGAN | WS | 1000 | 1000 | 0.9770 | ||
| W-BiGAN | WS | 1000 | 1000 | 0.9780 | ||
| LS-BiGAN | WS | 1000 | 1000 | 0.9640 | ||
| LS-CycleGAN | WS | 1000 | 1000 |
Bold indicates the best result
Fig. 2Implemented training scheme of BiGAN [55] for feature extraction. x: input feature; y: input data; X: feature space; Y: data space; G: Generator; : Encoder; D: Discriminator; : predicted and extracted feature
The implemented BiGAN architecture. Conv: Convolution; ConvTr: Transposed Convolution; BN: Batch Normalization; LR: Leaky ReLU; DR: Dropout
| (a) Generator | ||
|---|---|---|
| Noise | ||
| Dense+BN+LR | ||
| Reshape | ||
| ConvTr+BN+LR | ||
| ConvTr+BN+LR | ||
| ConvTr+BN+LR | ||
| ConvTr+BN+Tanh | ||
Fig. 3Implemented scheme of CycleGAN for feature extraction. x: Input feature; y: Input data; : Generated and extracted feature; : Generated data; X: Feature space; Y: Data space; G: Generator; : Encoder; : Generator’s discriminator; : Encoder’s discriminator
Fig. 4Proposed inference mechanism for binary classification of test images
Fig. 5Instances of target and test PDFs
Considered inter-PDF distances
| Considered metric distance | Formula |
|---|---|
| Euclidean | |
| Kullback-Leibler divergence | |
| Correlation | |
| Jensen-Shannon |
Main outcomes in binary classification
| Taxonomy | Description |
|---|---|
| True Positive (TP) | COVID-19 image classified as COVID-19 |
| True Negative (TN) | Non-COVID-19 image classified as non-COVID-19 |
| False Positive (FP) | Non-COVID-19 image classified as COVID-19 |
| False Negative (FN) | COVID-19 image classified as non-COVID-19 |
Performance metrics for binary classification [65]
| Performance index | Formula |
|---|---|
| Recall | |
| Precision | |
| F-score | |
| Accuracy |
Implemented solvers for the optimization and related hyper-parameter tuning. LR: Learning Rate; WclipV: Weight clip Value; GclipN: Gradient clip Norm
| Model | Solver | LR | WclipV | GclipN | |
|---|---|---|---|---|---|
| CE-BiGAN | Adam | 0.1 | – | – | |
| W-BiGAN | RMSprop | – | 0.01 | 10 | |
| LS-BiGAN | Adam | 0.1 | – | 10 | |
| LS-CycleGAN | Adam | 0.35 | – | – |
Fig. 6Training loss curves of the considered approaches
Training/test times of the considered models
| Model | Number oftraining iterations | Per-iteration time (s) | Per-image test time (s) |
|---|---|---|---|
| CE-BiGAN | 115,000 | 0.7 | 0.79 |
| W-BiGAN | 70,000 | 1.08 | 0.80 |
| LS-BiGAN | 140,000 | 0.9 | 0.79 |
| LS-CycleGAN | 7000 | 2.23 | 14.38 |
| HB-DCAE | 3125 | 0.27 | 0.05 |
| PB-CAE | 6250 | 0.306 | 0.01 |
Model performance under different distance metrics
| Model | Distance | Accuracy | Precision | Recall | F1-score | Test CTs |
|---|---|---|---|---|---|---|
| CE-BiGAN | ||||||
| Euclidean | 0.9760 | 0.9760 | 0.9765 | 0.9760 | 1000 | |
| Correlation | 0.9770 | 0.9778 | 0.9770 | 1000 | ||
| KL divergence | 0.9740 | 0.9740 | 0.9746 | 0.9740 | 1000 | |
| Jensen-Shannon | 0.9720 | 0.9720 | 0.9728 | 0.9720 | 1000 | |
| LS-BiGAN | ||||||
| Euclidean | 0.9490 | 0.9490 | 0.9492 | 0.9490 | 1000 | |
| Correlation | 0.9630 | 0.9653 | 0.9630 | 1000 | ||
| KL divergence | 0.9560 | 0.9560 | 0.9567 | 0.9560 | 1000 | |
| Jensen-Shannon | 0.9300 | 0.9300 | 0.9314 | 0.9299 | 1000 | |
| W-BiGAN | ||||||
| Euclidean | 0.9770 | 0.9780 | 0.9781 | 0.9780 | 1000 | |
| Correlation | 0.9770 | 0.9770 | 0.0772 | 0.9770 | 1000 | |
| KL divergence | 0.9770 | 0.9770 | 0.9772 | 0.9770 | 1000 | |
| Jensen-Shannon | 0.9780 | 0.9788 | 0.9780 | 1000 | ||
| LS-CycleGAN | ||||||
| Euclidean | 0.9760 | 0.9760 | 0.9766 | 0.9760 | 1000 | |
| Correlation | 0.9700 | 0.9700 | 0.9711 | 0.9700 | 1000 | |
| KL divergence | 0.9860 | 0.9860 | 0.9863 | 0.9870 | 1000 | |
| Jensen-Shannon | 0.9870 | 0.9873 | 0.9870 | 1000 | ||
Bold indicates the best result
Fig. 7Correlation distances between test-PDFs and COVID-PDF of CE-BiGAN; Threshold:
Fig. 8Correlation distances between test-PDFs and COVID-PDF under the LS-BiGAN model; Threshold:
Fig. 9Jensen-Shannon distances between test-PDFs and COVID-PDF under the W-BiGAN model; Threshold:
Fig. 10Jensen-Shannon distances between test-PDFs and COVID-PDF under the LS-CycleGAN model; Threshold:
Fig. 11Confusion matrices for the best setting of model/distance metric
Fig. 12ROCs of the tested models under the corresponding best distance metrics
Details and number of trainable parameters of the considered model architectures. BN: Batch-Normalization; Conv: Convolution; ConvTr: Transposed Convolution
| BiGANs | CycleGAN | HB-DCAE[ | PB-CAE[ | ||
|---|---|---|---|---|---|
| Conv+BN | – | – | – | ||
| ConvTr+BN | 4+4 | – | – | ||
| Dense | 1 | – | – | ||
| Trainable parameters | 82,753,665 | – | – | ||
| Conv + BN | 4+0 | – | – | ||
| Dense | 4 | – | – | – | |
| Trainable parameters | 83,826,561 | – | – | ||
| Conv+BN | 4+4 | 3+3 | 3+2 | 3+2 | |
| Dense | 1 | – | – | 1 | |
| Trainable parameters | 82,515,072 | 374,016 | 376,640 | 31,096,768 | |
Performance evaluation metrics related to some comparisons with other state-of-the-art approaches using the same dataset
| Reference | Model | Accuracy | Precision | Recall | F1-score |
|---|---|---|---|---|---|
| [ | AlexNet | 0.9060 | 0.9060 | 0.9209 | 0.9052 |
| [ | VGG16 | 0.9140 | 0.9140 | 0.9266 | 0.9134 |
| [ | ResNet50 | 0.9245 | 0.9245 | 0.9359 | 0.9302 |
| [ | CovidNet-ct | 0.9790 | 0.9790 | 0.9714 | 0.9752 |
| [ | MERSGAN+ | 0.9765 | 0.9820 | 0.9840 | 0.9830 |
| [ | Random forest | 0.9004 | 0.8917 | 0.9025 | 0.8971 |
| [ | AI-system | 0.9656 | 0.9020 | 0.9710 | 0.9352 |
Fig. 13a Per-model training time versus test accuracy. b Per-model test time versus test accuracy