Literature DB >> 33814762

Comparing a deep learning model's diagnostic performance to that of radiologists to detect Covid -19 features on chest radiographs.

Sabitha Krishnamoorthy1, Sudhakar Ramakrishnan2, Lanson Brijesh Colaco3, Akshay Dias4, Indu K Gopi5, Gautham A G Gowda6, K C Aishwarya6, Veena Ramanan7, Manju Chandran8.   

Abstract

BACKGROUND: Whether the sensitivity of Deep Learning (DL) models to screen chest radiographs (CXR) for CoVID-19 can approximate that of radiologists, so that they can be adopted and used if real-time review of CXRs by radiologists is not possible, has not been explored before.
OBJECTIVE: To evaluate the diagnostic performance of a doctor-trained DL model (Svita_DL8) to screen for COVID-19 on CXR, and to compare the performance of the DL model with that of expert radiologists.
MATERIALS AND METHODS: We used a pre-trained convolutional neural network to develop a publicly available online DL model to evaluate CXR examinations saved in .jpeg or .png format. The initial model was subsequently curated and trained by an internist and a radiologist using 1062 chest radiographs to classify a submitted CXR as either normal, COVID-19, or a non-COVID-19 abnormal. For validation, we collected a separate set of 430 CXR examinations from numerous publicly available datasets from 10 different countries, case presentations, and two hospital repositories. These examinations were assessed for COVID-19 by the DL model and by two independent radiologists. Diagnostic performance was compared between the model and the radiologists and the correlation coefficient calculated.
RESULTS: For detecting COVID-19 on CXR, our DL model demonstrated sensitivity of 91.5%, specificity of 55.3%, PPV 60.9%, NPV 77.9%, accuracy 70.1%, and AUC 0.73 (95% CI: 0.86, 0.95). There was a significant correlation (r = 0.617, P = 0.000) between the results of the DL model and the radiologists' interpretations. The sensitivity of the radiologists is 96% and their overall diagnostic accuracy is 90% in this study.
CONCLUSIONS: The DL model demonstrated high sensitivity for detecting COVID-19 on CXR. CLINICAL IMPACT: The doctor trained DL tool Svita_DL8 can be used in resource-constrained settings to quickly triage patients with suspected COVID-19 for further in-depth review and testing. Copyright:
© 2021 Indian Journal of Radiology and Imaging.

Entities:  

Keywords:  Artificial intelligence; COVID 19; CXR; deep learning

Year:  2021        PMID: 33814762      PMCID: PMC7996677          DOI: 10.4103/ijri.IJRI_914_20

Source DB:  PubMed          Journal:  Indian J Radiol Imaging        ISSN: 0970-2016


Introduction

It is imperative that rapid and sensitive methods to detect infection by SARS-CoV-2 be developed to meet the demands and challenges brought about by the ongoing COVID-19 pandemic. Although COVID-19 infection is typically confirmed by reverse-transcription polymerase chain reaction (RT-PCR), other non-invasive imaging methods can be employed to supplement the diagnosis.[1] CT has demonstrated good, albeit variable, sensitivity (60 to 98%) in detecting the characteristic lung manifestations of COVID-19.[123] However, CT is impractical as a screening modality given the substantial radiation exposure associated with it, lack of easy availability, and higher cost. An additional challenge in the use of CT for initial diagnostic workup of COVID-19 is the need to establish and follow protocols for maintaining corridors for the transport of patients with suspected COVID-19 to the CT scanning room. Chest radiography (CXR) has also been used for the initial diagnosis of COVID-19 and is simpler to implement for this purpose. Its portability is an additional advantage. Moreover, the findings of COVID-19 on CXR mirror those on CT.[45] Though variability may exist, the findings typically include hazy areas of increased opacities, peripheral air space and diffuse lung opacities, and bilateral lower lobe consolidations.[4] Guidelines by the American College of Radiology support the use of portable CXR units to prevent cross-infection during the COVID-19 pandemic.[6] CXR reporting may experience substantial delays due to radiologist staffing issues that may be exacerbated during the pandemic. There may also be a shortage of both general and thoracic radiologists in resource-limited geographic regions. This may lead to misreporting and misdiagnosis by less trained providers.[7] Artificial intelligence (AI) using deep learning (DL) has been applied to various imaging analyses, substantiated by the ImageNet Large Scale Visual Recognition Competition[8] (hereafter referred to as ImageNet). DL has previously been used for detection of pulmonary pathologies such as tuberculosis[9] and more recently has been applied to successfully identify features of COVID-19 on CT.[10] The few existing studies of the application of AI to identify characteristic findings of COVID-19 on CXR focused on the technical aspects of DL, were performed in experimental settings, or used single-center data sets.[1112131415] Physicians with limited technical expertise cannot be expected to readily adopt such models despite the strong need for rapid and real time CXR interpretation in resource-limited settings.[516] For widespread implementation, a DL model must not only be practical and accessible across clinical settings, but also be able to detect COVID-19 features on CXR with reasonable sensitivity not far removed from that of expert radiologists. A central role of physicians in the AI model's curation techniques can improve such models’ generalizability and to facilitate use in a wide variety of clinical contexts. We thus aimed to develop and evaluate the diagnostic performance of a physician-trained DL model to screen for characteristic features of COVID-19 on CXR, comparing the performance of the DL model with that of expert radiologists.

Materials and Methods

This retrospective study was approved by the Institutional Review and the Ethics Committees (Reference Number 11/20/IRC/JMMC&RI). The requirement for informed consent was waived by the Institutional Review Committee. We compiled the CXR examinations used for this study from publicly available datasets of confirmed cases of COVID-19 from numerous countries including Australia, India, Iran, Israel, Italy, Jordan, Pakistan, Qatar, Spain, and the United Kingdom. We also acquired with permission, CXR from case presentations on COVID-19 from three states in India (Kerala, Tamil Nadu, and Karnataka), as well as from patients without COVID-19, from two partnering hospital repositories in India. These latter examinations were performed in 2017 and 2018, before the first known reported case of COVID-19 in December 2019.[17]

Development of the DL model

We curated the DL model using a 42-layer deep convolutional neural network (CNN) based on the Inception-v3 network architecture. We employed transfer learning, pre-training the network initially with ImageNet data. The model was deployed as a publicly available online tool (https://svita.in/). We transferred the knowledge of the ImageNet dataset to another dataset designated as the COVID_COLLECT_TRAINING_SET (CCTS). This served as the primary dataset for curation. CCTS contained 94535 random non-radiograph images obtained from a publicly available collection, listed in Table 1 and 1684 random radiographs that included 1062 CXR examinations [Table 1]. An internist with 15 years of experience in CXR interpretation (SK) and a general radiologist with 15 years of experience (VR) curated the CCTS. These two individuals analyzed all CXR examinations in the primary dataset and assigned them to one of three labels: (1) normal CXR, (2) COVID-19 pattern, and (3) abnormal CXR unlikely to represent COVID-19. The images were randomly split into a test set with 10% of the data, a validation set with another 10% of the data, and a training set with the remaining 80%.
Table 1

COVID_COLLECT_TRAINING_SET (CCTS)

USESOURCES
Stage 1
 DS1Random (non-radiograph) images for trainingtiny-imagenet.herokuapp.com (n=9435)
 DS2Random radiographs for trainingRepository of all types of radiographs belonging to investigators and some images from NIH_imagenet (n=1684)
Stage 2
 DS3Non-CXR radiographs for trainingRepository of Non-CXRs belonging to investigators (n=391)
 DS4“Normal” CXRs for training the DL modelNIH_imagenet (n=77)
montgomery_dataset (n=60)
thrissurnormal_dataset (n=113)
Normal_KVG (n=15)
Normal_NIH (n=15)
Normal_SNH pelvis-osteoporotic 18 (n=17)
 DS5“Abnormal” Mix includes COVID and other chest abnormalities for training the DL modelNIH_imagenet (n=360)
COVID-19 (n=170)
Covidspain_dataset (n=5)
Covid19 cases_cxr (n=31)
Ieee8023-covid-chestxray-dataset (n=97)
Montgomery_dataset (n=18)
SNH (n=84)
Stage 3
 DS6“COVID CXR” for training the DL modelCovidspain_dataset (n=5)
Ieee8023-covid-chestxray-dataset (n=97)
NIH_imagenet (n=121)
Covid19 cases_cxr (n=31)
COVID-19 (n=170)
SNH_dataset (n=42)
 DS7“Non-COVID Abnormal CXR” for training the DL modelNIH_imagenet (n=226)
Montgomery_dataset (n=17)
SNH_dataset (n=39)
The CXRs labeled as COVID-19 exhibited multifocal peripheral and basal lung opacities.[5] The classical consolidation in COVID-19 is bilateral with lower lobe predominance. As the disease progresses, a diffuse distribution of lung opacities develops. CXR examinations labeled as abnormal and unlikely to represent COVID-19, exhibited central and peribronchovascular apical distribution of opacities, hilar lymphadenopathy, cavitation, calcifications, mass lesions, and pleural effusions. CXRs labeled as normal were those without any intrapulmonary opacities, bone lesions, pleural thickening associated with coarse calcifications, or pleural effusions. The DL model was trained to generate an output classifying the radiographs into the three labels. We introduced an expert system classifier that piped the input image through an AI pipeline decision tree. The first stage was a fraud detection system that allowed only a valid chest radiograph to pass to subsequent steps. If the submitted image was deemed to be a valid CXR, the model evaluated if it was normal or abnormal, and if abnormal, if it was likely to represent COVID-19. At each step, the DL model generated a percentage probability of the submitted image satisfying the particular criterion. We set threshold values for the percentage probability at each step for the model to further pipe the image through the multiple stages of the decision tree.[Figure S1] We used a different dataset for validation termed the COVID_COLLECT_REALWORLD_VALIDATON_SET (CCRVS) [Table 2]. CCRVS used three mutually exclusive groups of data from entirely different clinical resources: (1) CXR examinations from patients with confirmed COVID-19 (RT-PCR positive) [DS9, DS11] taken from case presentations by physicians from three states in India and for which the co-investigators (SK, AD, LC) validated the CXR examinations. (2) publicly available CXR images from patients with confirmed COVID-19, collected from open forums from various countries, as has been done previously[14]; (3) CXR images from patients without COVID-19 [DS10] from the two previously noted hospital repositories. The CXR examinations in the final group were either normal, or abnormal due to a non-COVID-19 etiology.
Table 2

Validation study Dataset

COVID_COLLECT_REALWORLD_VALIDATON_SET (CCRVS)REAL WORLD IMAGES TESTTotal No. of images in the datasetImages run through the DL modelImages EXCLUDED (from analysis)Images used for analysis of DL model performance
Dataset compiled by investigators [DS8]RANDOM Non-COVID CXRs [DS10]ADFR60601842
ADFR279791267
ADFR33232428
NKVG1616016
NKVG28817
NKVG32222022
N_SNH4040139
N_NIH3232032
CXRs of RT-PCR confirmed COVID cases [DS11]BK8808
CBE9936
DP5505
MJRI2202
OJ6606
RP3131427
AP5514
Covidspain_ dataset [DS9]CXRs of RT-PCR positive COVID casesSPAIN12912910119
DL Model VALIDATION STUDY on June 10, 2020 using CCRVSTOTAL IMAGES IN CCRVS484
IMAGES EXCLUDED FROM CCRVS54
IMAGES ANALYSED FOR STUDY FROM CCRVS430
The validation study had a comparative cross-sectional design and used 484 CXRs in anteroposterior or posteroanterior projection that were saved in .png or .jpeg format. The radiologists (GG, AC) visually evaluated these CXRs for overall quality and cleared them as interpretable for the study. We excluded lateral chest radiographs as well as any radiographs that either the radiologists or the DL model identified as poor-quality, for example if they were over- or underexposed or showed motion blur artifacts. These exclusions resulted in a final sample of 430 radiographs for validation. Multiple rounds of internal testing and optimization to maximize sensitivity were employed to determine the DL model's classification thresholds for each of the three outcomes (normal, COVID-19, and non-COVID-19 abnormality). For the purpose of this study, the threshold value for classifying a CXR as COVID-19 was set at ≥40% probability. As an example, when we changed the thresholds for detecting COVID-19 from 40% to 60% in the internal trials, the sensitivity of the model decreased to 83%, while specificity increased to 60%. COVID_COLLECT_TRAINING_SET (CCTS) Validation study Dataset

Radiologist review

Two general radiologists with 9 and 12 years of experience (GG, AC) independently reviewed all CXR examinations in CCRVS. These radiologists were blinded to the source of the CXR file, RT-PCR results, as well as other clinical data. The radiologists recorded for each CXR whether or not they suspected COVID-19. The radiologists also performed a subjective post-hoc assessment of imaging findings on CXRs for which the DL model provided a false positive or false negative interpretation.

Statistical analysis

Diagnostic performance of the DL model and of the two radiologists in detecting COVID-19 was assessed using the RT-PCR results as reference standard. True positives, true negatives, false positives, and false negatives were identified. The sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) were computed. The ROC curve was plotted, and the area under the curve (AUC) was calculated. The DL model's overall diagnostic accuracy was calculated. Analysis was performed using statistical software (IBM SPSS (Statistical Package of Social Sciences) version 25).

Results

Performance of the DL model

The DL model achieved sensitivity of 91.5%, specificity of 55.3%. PPV of 60.9%, and NPV of 77.9%. The overall diagnostic accuracy was 70.2%. The AUC was 0.73 (95% CI: 0.69, 0.78). Table 3 provides further classification of the model's performance.
Table 3

Statistical Classification of DL models Performance

Svita_DL8DL model misdiagnoses a normal CXR as COVIDDL model misdiagnoses a COVID CXR as NormalDL model correctly classifies as COVIDDL model correctly classifies as Non-COVIDWhen the DL model predicts COVID positive, how often is it correct? TP/(TP + FP)Recall How many of the COVID cases did DL model identify? TP/(TP + FN)The proportion of CXRs of Actual non-COVID patients who are identified as Non-COVID by the DL model TN/(TN + FP)F1 Overall measure of a model's accuracy: 2*(P*R)/(P + R)Diagnostic Accuracy (TP + TN)/(TP + FP + FN+TN)Reasons are mentioned under exclusion criteria in Discussion
June 10, 2020 test run resultsFalse Positive (FP)False Negative (FN)True Positive (TP)True Negative (TN)Precision (PPV)Sensitivity (SN)Specificity (SP)F1 ScoreDiagnostic AccuracyExcluded from analysis
ADFR230019000.45238100.45238118
ADFR2490018000.26865700.26865712
ADFR320008000.28571400.2857144
COVID_BK017010.87500.9333330.8750
COVID_CBE0060110113
COVID_DP0050110110
COVID_MJRI0020110110
COVID_OJ0060110110
COVID_RP019010.900.9473680.94
COVID_RP15-310116010.94117600.9696970.9411760
COVID_SPAIN012107010.89916000.9469030.89916010
COVID_AP0040110111
NKVG30013000.812500.81250
NKVG20007001011
NKVG315007000.31818200.3181820
N_NIH30029000.9062500.906250
N_SNH00039001011
OVERALL113151621400.5890910.9152540.5533600.7168140.70232654
Statistical Classification of DL models Performance

Comparison of the DL model and radiologists

There was a statistically significant correlation (r = 0.617, P = 0.000) between the interpretation by the DL model and the interpretation of the CXRs by the radiologists in classifying the CXR as a suspected COVID case or not [Figure 1]. The first radiologist's sensitivity was 96%, specificity was 85.7%, PPV 82.5%, NPV 96.8%, AUC was 0.91 (95% CI: 0.88, 0.94).
Figure 1

Comparison of deep learning model to radiologist interpretation

Comparison of deep learning model to radiologist interpretation The second radiologist's sensitivity was 95.5%, specificity was 85.65%, PPV was 84.47%, NPV was 96.21% and AUC 0.88 (95% CI: 0.84, 0.91). The overall diagnostic accuracy of the radiologists was 90% and 90.2%, respectively. Figure 2 shows the distribution of findings on CXR examinations with a false-negative interpretation by the DL model for COVID-19, based on post hoc assessment by the two radiologists. The most common findings were an ARDS like picture and effusion, pneumonia, or dense infiltrates. Figure 3 shows the distribution of findings on CXR examinations with a false-positive interpretation by the DL model for COVID-19, based on post hoc assessment by the two radiologists. The most common findings were an ARDS like picture and a normal radiograph.
Figure 2

Distribution of false-negative CXR interpretations for COVID-19 by the deep learning model

Figure 3

Distribution of false-positive CXR interpretations for COVID-19 by the deep learning model

Distribution of false-negative CXR interpretations for COVID-19 by the deep learning model Distribution of false-positive CXR interpretations for COVID-19 by the deep learning model

Discussion

In this study, we evaluated the performance of a DL model that was trained to detect COVID-19 on CXR using multiple real-world datasets curated by an internist and radiologist. Model validation was performed using CXR data sets from multiple hospitals and outpatient clinics from 10 countries and 3 different states in India. The model had high sensitivity for COVID-19 on CXR in both the training and validation data sets, with sensitivities of 92% and 91.5%, respectively. Of prior studies[1112131415] that used DL models to detect COVID-19 on CXR, one used a class decomposition approach employing a deep CNN architecture (Detrac ResNet). This had obtained a sensitivity of 97.9%, though it was conducted experimentally on a smaller dataset.[13] The other study that used a pre-trained CNN (ResNet 50) included only abnormal CXR examinations and showed an overall accuracy of 89.2%.[14] The techniques presented in these two studies require machine learning expertise and are difficult to incorporate into clinical settings. Another study that explored the diagnostic performance of AI for COVID-19 in a clinical setting used only a small sample of abnormal CXR examinations from a single center.[15] Our DL model and study have several strengths. We not only studied the DL's model's performance characteristics, but also conducted independent analysis of 2 radiologists’ interpretations and compared their performance characteristics with that of the model. We tested the model over a large spectrum of CXRs and demonstrated high sensitivity. This high sensitivity makes the model a potential screening tool to guide patient triage. We used a distinctive curation technique in introducing a step by step classification with an expert system overlay between the steps. The fraud detection system built into the application eliminates “junk” images from entering the model, thereby enabling the model's use in real-world clinical settings. The introduction of variable percentage thresholds at different steps of the expert system helped to change and assess the model's performance before determining the final cut-off [Figure S1]. The expert system thresholds allow users of the model to tailor its sensitivity depending upon the pretest probability for COVID-19 in the setting in which the model is used (e.g., outpatient, inpatient, or ICU). For instance, the threshold can be changed to a higher level in an ICU setting. The DL model was deployed as an online tool (https://svita.in/). Physicians can readily access the model and upload images of CXR examinations. Furthermore, our dataset was collected randomly from various sources, minimizing selection bias. An additional unique feature of the model is that we trained the model using CXR examinations saved in .jpeg or .png format. This facilitates the model being applied in clinical scenarios in which resource constraints exist. For example, the model can assess an image of a CXR taken using a camera of a smart phone on a viewing box. A unique feature of our model is that it has been configured in such a way that it can enable “fan-in” of other pathologies into the AI pipeline without having to reconstruct the system. A limitation is the model's low specificity (55.3%). The false positives may be explained by the fact that the model is not able to differentiate subtle changes on CXRs unlike human radiologists. Therefore, any minimal increase in opacities or haziness due to other etiologies or artifacts were also labelled as COVID-19 by the model. ARDS is a sequela of multiple diseases, and it may be difficult to determine the etiology of ARDS solely by CXR.[18] CXR examinations that show ARDS, as well as pulmonary edema, due to causes other than COVID-19 are difficult to differentiate from those that show ARDS resulting from COVID-19.[19] The CXR examinations from patients without COVID-19 were procured from a tertiary care center with many critically ill patients with other lung pathologies that caused an ARDS-like picture [Figure 4]. These were likely misinterpreted by the model as COVID-19.[19] However, in general, it has to be noted that the reported specificity of radiologists in diagnosing COVID-19 from CXRs is only about 69%.[5] While the model had low specificity, the model is not intended to serve as a standalone tool but rather as a screening tool given its excellent sensitivity. Screened patients considered likely to have COVID-19 on CXR can be triaged to a separate area where an RT-PCR swab test can be performed.
Figure 4

Examples of the interpretation of CXR by the deep learning model

Examples of the interpretation of CXR by the deep learning model Our model aimed to detect COVID-19 characteristics on CXR and not disease extent or severity. Future research and development could supplement the model by introducing an object detection layer, and re-designing, and re-training it to categorize the disease extent.

Conclusion

Our experienced radiologists’ sensitivity and diagnostic performance in detecting COVID -19 characteristics on CXRs is higher than previously reported. The specificity of the DL model is lower than that of experienced radiologists’ to rule out COVID-19 on CXR. This makes it quite clear that it cannot replace expert radiologists. However, its relatively high sensitivity makes it useful as a rapid and real-time screening solution for COVID-19, particularly in resource-constrained environments. Such AI models can help automate some aspects of initial screening, for example, to filter in the CXRs images that need to be flagged for earlier expert radiology assessment.

Financial support and sponsorship

Nil.

Conflicts of interest

There are no conflicts of interest. Expert System for Classification of Outcomes
  13 in total

1.  Radiological services in a hospital emergency department--an evaluation of service delivery and radiograph interpretation.

Authors:  T M Nolan; F Oberklaid; D Boldt
Journal:  Aust Paediatr J       Date:  1984-05

2.  Acute respiratory distress syndrome: the Berlin Definition.

Authors:  V Marco Ranieri; Gordon D Rubenfeld; B Taylor Thompson; Niall D Ferguson; Ellen Caldwell; Eddy Fan; Luigi Camporota; Arthur S Slutsky
Journal:  JAMA       Date:  2012-06-20       Impact factor: 56.272

3.  Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR.

Authors:  Yicheng Fang; Huangqi Zhang; Jicheng Xie; Minjie Lin; Lingjun Ying; Peipei Pang; Wenbin Ji
Journal:  Radiology       Date:  2020-02-19       Impact factor: 11.105

4.  COVID-19 on Chest Radiographs: A Multireader Evaluation of an Artificial Intelligence System.

Authors:  Keelin Murphy; Henk Smits; Arnoud J G Knoops; Michael B J M Korst; Tijs Samson; Ernst T Scholten; Steven Schalekamp; Cornelia M Schaefer-Prokop; Rick H H M Philipsen; Annet Meijers; Jaime Melendez; Bram van Ginneken; Matthieu Rutten
Journal:  Radiology       Date:  2020-05-08       Impact factor: 11.105

Review 5.  Portable chest X-ray in coronavirus disease-19 (COVID-19): A pictorial review.

Authors:  Adam Jacobi; Michael Chung; Adam Bernheim; Corey Eber
Journal:  Clin Imaging       Date:  2020-04-08       Impact factor: 1.605

6.  Radiological Society of North America Expert Consensus Statement on Reporting Chest CT Findings Related to COVID-19. Endorsed by the Society of Thoracic Radiology, the American College of Radiology, and RSNA - Secondary Publication.

Authors:  Scott Simpson; Fernando U Kay; Suhny Abbara; Sanjeev Bhalla; Jonathan H Chung; Michael Chung; Travis S Henry; Jeffrey P Kanne; Seth Kligerman; Jane P Ko; Harold Litt
Journal:  J Thorac Imaging       Date:  2020-07       Impact factor: 3.000

7.  Weakly Labeled Data Augmentation for Deep Learning: A Study on COVID-19 Detection in Chest X-Rays.

Authors:  Sivaramakrishnan Rajaraman; Sameer Antani
Journal:  Diagnostics (Basel)       Date:  2020-05-30

Review 8.  COVID-19 acute respiratory distress syndrome (ARDS): clinical features and differences from typical pre-COVID-19 ARDS.

Authors:  Peter G Gibson; Ling Qin; Ser Hon Puah
Journal:  Med J Aust       Date:  2020-06-22       Impact factor: 12.776

9.  Value of initial chest radiographs for predicting clinical outcomes in patients with severe acute respiratory syndrome.

Authors:  Tai-Nin Chau; Po-Oi Lee; Kin-Wing Choi; Chiu-Man Lee; Ka-Fai Ma; Tak-Yin Tsang; Yuk-Keung Tso; Ming-Chee Chiu; Wing-Lok Tong; Wai-Cho Yu; Sik-To Lai
Journal:  Am J Med       Date:  2004-08-15       Impact factor: 4.965

10.  Using artificial intelligence to read chest radiographs for tuberculosis detection: A multi-site evaluation of the diagnostic accuracy of three deep learning systems.

Authors:  Zhi Zhen Qin; Melissa S Sander; Bishwa Rai; Collins N Titahong; Santat Sudrungrot; Sylvain N Laah; Lal Mani Adhikari; E Jane Carter; Lekha Puri; Andrew J Codlin; Jacob Creswell
Journal:  Sci Rep       Date:  2019-10-18       Impact factor: 4.379

View more
  1 in total

1.  Discovery of a Generalization Gap of Convolutional Neural Networks on COVID-19 X-Rays Classification.

Authors:  Kaoutar Ben Ahmed; Gregory M Goldgof; Rahul Paul; Dmitry B Goldgof; Lawrence O Hall
Journal:  IEEE Access       Date:  2021-05-13       Impact factor: 3.367

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.