Literature DB >> 34054148

Checklist for responsible deep learning modeling of medical images based on COVID-19 detection studies.

Weronika Hryniewska1, Przemysaw Bombiski2, Patryk Szatkowski2, Paulina Tomaszewska1, Artur Przelaskowski1, Przemysaw Biecek1,3.   

Abstract

The sudden outbreak and uncontrolled spread of COVID-19 disease is one of the most important global problems today. In a short period of time, it has led to the development of many deep neural network models for COVID-19 detection with modules for explainability. In this work, we carry out a systematic analysis of various aspects of proposed models. Our analysis revealed numerous mistakes made at different stages of data acquisition, model development, and explanation construction. In this work, we overview the approaches proposed in the surveyed Machine Learning articles and indicate typical errors emerging from the lack of deep understanding of the radiography domain. We present the perspective of both: experts in the field - radiologists and deep learning engineers dealing with model explanations. The final result is a proposed checklist with the minimum conditions to be met by a reliable COVID-19 diagnostic model.
© 2021 The Author(s).

Entities:  

Keywords:  COVID-19; X-ray; computed tomography; deep learning; explainable AI; lungs

Year:  2021        PMID: 34054148      PMCID: PMC8139442          DOI: 10.1016/j.patcog.2021.108035

Source DB:  PubMed          Journal:  Pattern Recognit        ISSN: 0031-3203            Impact factor:   7.740


Introduction

COVID-19 is a fast spreading disease of highly contagious nature caused by the SARS-CoV-2 virus from the coronavirus group. At the end of January 2020, the World Health Organization (WHO) declared a global health emergency and one and a half months later, a pandemic. By September 25, 2020, 32,110,656 confirmed cases and 980,031 deaths had been documented. From a public health perspective, due to the lack of proper medicines, early detection of COVID-19 and patient isolation are crucial. Hospitals are crowded with the exponentially growing number of patients as available resources are limited. Currently, reverse transcription polymerase chain reaction (RT-PCR) is the gold standard used to diagnose COVID-19 infection [1]. However, the results of RT-PCR can be affected by sampling errors and low viral load [1]. As a result, these tests suffer from high rates of false negatives (with sensitivity of 71% [2] or 69% [3]) and may need to be conducted two or more times before the results are finally confirmed [4]. In many articles, chest imaging is considered a suitable tool for early COVID-19 screening [5], [6]. The point is that sensitivity of computed tomography (CT) scan tests can reach 98%, which is much higher than RT-PCR tests [7]. Moreover, due to the fact that on CT images ground-glass opacities are visible earlier than pulmonary consolidation [8], radiologists can assess the stage of COVID-19. Unfortunately, CT scanners are not widely available. Study [9], indicate that distinguishing between COVID-19 and viral pneumonia is a challenging task. However, it is worth noticing that the overall process of undertaking chest imaging to getting the first results is much shorter than in case of RT-PCR. The screening takes approximately 15 s [10] (in terms of X-ray) to 21.5 [11] minutes on average (for CT) to complete. Taking the sample for RT-PCR test is a fast procedure. The difference is not the time of undertaking the test/screening but the time needed to get the first results. In case of RT-PCR, it may take from several hours up to several days [12] as nucleic acid amplification must happen before the start of samples’ analysis. The issue is important as the patient has to be isolated until receiving the test result. In addition, X-rays are cheaper, more available worldwide, and less harmful than CTs because the radiation dose is smaller. Due to the existence of portable devices, X-ray imaging can be performed in isolated rooms, so the risk of infection is significantly decreased [13]. Nonetheless, especially on X-ray images, it is particularly difficult to assess the severity of the pathology, and, thus, only experts in radiology should interpret chest images. In general, this process is faster than waiting for RT-PCR test results. However, after individual patient collection, multiple test samples can be examined by a laboratory assistant simultaneously. Whereas a radiology technician is only able to take an image of one patient at a time, and then such an image must still be analyzed by a radiologist. Recent applications of machine learning (ML) have gained popularity in the medical domain [14], [15]. The performance achieved by neural networks is becoming similar to that reached by medical experts. Deep learning techniques for medical images are present in classification (skin lesions [16], lung cancer [17]), detection (arrhythmia [18], breast cancer [19], pneumonia [20], ADHD [21]), segmentation (lung [22], brain [17]) and imaging reconstruction (magnetic resonance [23], Single Photon Emission Computerized Tomography (SPECT) [24]). Considering the need for a highly accurate and fast diagnosis process, artificial intelligence (AI) can play a significant role in automating the detection of COVID-19 cases. AI solutions are frequently based on complex, so-called black-box models [25]. For this reason, it is difficult to tell what factors lead to a particular model prediction. Such a lack of interpretability may be dangerous, as it may lead to biased results and incorrect decisions in real diagnostic procedures. Recent development in the area of Explainable Artificial Intelligence (XAI) shows the importance of model explanations, which help to avoid erroneous predictions. Nevertheless, surprisingly, in the area of COVID-19 image analysis, there are still only few results concerning the use of XAI for lung image analysis. In this paper, we will summarize recent publications about lung imaging analysis (section II-III), and show how explainable AI techniques have been used in these solutions (section IV). We will confront these approaches with the domain knowledge of radiologists and we will show how many of the assumptions about data, models or explanations made in many of the analyzed studies are not appropriate. Finally, we will construct a checklist to help model developers assess whether they avoided the most common errors. We believe that this criticism, together with the proposed checklist, will contribute to building better models not only for the diagnosis of COVID-19 disease, but also for other applications based on lung images.

Methods

Literature search

This research is based on a systematic literature review. The data was collected between the 1st and 14th of August 2020. The search was performed according to Preferred Reporting Items for Systematic reviews and the Meta-Analyses (PRISMA) statement [26] presented in Fig. 1 in the following academic digital databases: ArXiv, IEEE, Google Scholar, PubMed, Science Direct, Scopus, Web of Science. All studies written in English, regardless of the publication status (preprint, peer-reviewed, or published articles), were included in this review. Studies were identified by the combination of keywords: (XAI COVID-19 OR explainable artificial intelligence COVID-19 OR explainable COVID-19 OR explanations COVID-19 OR interpretable COVID-19 OR interpretations COVID-19 OR transparent COVID-19) and (X-ray OR radiography OR CT OR computed tomography). Then, each study was screened for content relevance.
Fig. 1

PRISMA Flow Diagram shows the flow of information through the different phases of a systematic review including inclusions and exclusions.

PRISMA Flow Diagram shows the flow of information through the different phases of a systematic review including inclusions and exclusions. From 31 collected works, during eligibility checking, 6 studies were dropped due to irrelevant scope or lack of XAI parts. There were 25 studies included for qualitative synthesis. The number of studies considered in the review is vast enough to create a representative set/collection for further investigation. Some of the studies were published as preprints, not as camera-ready articles. Due to the rapidly growing field of tools related to supporting medical practitioners in the fight against pandemics, we included them. They will help to show a variety of considered XAI approaches. Then, on February 26, 2021, it was verified which articles previously available as preprints were published in peer reviewed journals. The content of these articles was again reviewed for changes that had been made. Articles published in journals: [27], [28], [29], [30], [31], [32], [33], [34], [35], [36], [37], [38], preprints: [39], [40], [41], [42], [43], [44], [45], [46], [47], [48], [49], [50], [51].

Types of COVID-19 prediction tasks

The 25 studies identified according to our methodology concern various types of ML tasks. In Fig. 2 , we introduce a taxonomy for these studies. In the following sections, when discussing specific solutions, it is important to remember what kind of problems they are designed for.
Fig. 2

Taxonomy of AI applications in 25 reviewed studies.

Taxonomy of AI applications in 25 reviewed studies. The first breakdown concerns the types of tasks, such as: classification (COVID-19 vs. other classes), severity assessment or segmentation. Among classification tasks, the first group is related to the detection of COVID-19 cases. The goal of the second is to assess how severe the changes caused by COVID-19 are. In the reviewed works, segmentation can be regarded as an image preparation technique for the further classification process. Segmentation was not used in all studies. Lung segmentation is present in works [31], [32], [34], [45], [50] and lesion segmentation in [32], [38], [47]. The lungs are segmented to remove the unnecessary background because, based on medical experience, the lesions caused by COVID-19 are not located outside the lungs. Lesion segmentation, also called infection mask, helps to train the model to recognize infected regions and can be beneficial for further model assessment. Only radiologists can prepare trustworthy and high-quality lung lesion masks, not an automatic tool similar to those in the reviewed works [32], [38]. However, manual mask preparation takes a lot of time and money and requires a high level of consistency among annotators, but, surely, it is the most valuable method for qualitative and quantitative XAI evaluation. There are not any strict guidelines on how many classes the classification should be conducted on. In classification problems, the number of classification classes varies between studies. This distinction is particularly important when comparing model performance. It is worth noting that some studies have verified the performance of the model considering different number of classes [28], [33], [35], [48]. Binary classification task detecting COVID-19 and non-COVID-19 is the most popular [27], [28], [30], [33], [35], [38], [39], [41], [45], [46], [48], [49]. Another frequently used method is three-class classification [28], [29], [33], [35], [36], [37], [47], [48], [51]: no infection/no pneumonia, pneumonia (bacterial or non-COVID-19 viral infection), and COVID-19 (COVID-19 viral infection). In the four-class classification problem, there are two different approaches to splitting images. The first one divides images into: normal, bacterial pneumonia, non-COVID-19 viral pneumonia, and COVID-19 [35], [44], while the second one into: normal, bacterial pneumonia, tuberculosis (TB), and viral pneumonia. In the second approach, study [31] refers to [52] and claims that differentiation between viral pneumonia and COVID-19 is challenging, because similar radiological features exist between them. In classification tasks with the biggest number of classes (5), the authors distinguish five different classes: normal, pneumonia, virus, bacteria, and COVID-19 [40]. In [42] the authors applied multi-label classification. For example, for an image with the lungs of a patient suffering from COVID-19, the correct classification should predict all three labels: pneumonia, viral pneumonia, and COVID-19. In the segmentation for severity assessment, there are two different approaches. One rates how severe the lesions are in parts of the lungs. Another one assigns labels to lesions’ names that point out which changes in the lungs are present in the image. In two studies [34], [50] images or parts of images were divided into classes that correspond to the severity of COVID-19 effects on the lungs. Tabik et al. [34] propose dividing each lung horizontally into 4 parts, and giving them a grade of 1 if it contained any lesions, such as consolidation or ground-glass opacities, or 0 otherwise. Then, the grades from all parts were summed up. Based on grades, the following scale was prepared: Normal-PCR+ 0 (a patient classified by expert radiologist as Normal that has positive RT-PCR test), Mild 1-2, Moderate 3-5, and Severe 6-8. A similar solution was introduced by Signoroni et al. [50]. In this study, each lung was divided horizontally into 3 parts and a 4 grade scale was used for each part. The division of each lung into 3 parts (upper, middle, and lower lung field) resembles natural lung structure and common radiological practice. An example of multi-label classification is described in [43] where 18 outputs of neural network (such as atelectasis, consolidation, infiltration, pneumothorax, edema, emphysema, fibrosis, effusion, pneumonia, pleural thickening, cardiomegaly, nodule, mass, hernia, lung lesion, fracture, lung opacity, and enlarged cardiomediastinum) give information about lesions in lungs. Some of them are characteristic for specific illnesses. This is a very promising approach because with XAI visualizations it should be clearly visible for radiologists if a model learned to recognize proper lesions.

Deep learning on lung image data

Data resources

The COVID-19 virus is a relatively new disease, and, in several articles, the lack of high-quality medical imaging databases is indicated [34], [47]. To verify this, we looked at the quality of the data used to construct the models in selected peer-reviewed research descriptions. The results of this analysis are presented in Table 1 .
Table 1

This table presents the data sources used in studies reviewed in this survey. For each data source, we list articles that use it. The JPEG quality factor (QF) for most images has been set to 75, other cases are indicated. In the case of COVID-Net, please note that it is not a data source, but a study collating 5 datasets. Some other studies refer to it instead of referring to the original source.

No.InstitutionLink to datasetUsed in articleDynamic range of imagesData processingPrepared for scientific experiments
1)University of Waterloogithub.com/lindawangg/COVID-Net[37], [47], [48], [50]
1a)University of Waterloogithub.com/agchung/Figure1-COVID-chestxray-dataset[29], [48]8 bits, 48 casesJPG, PNGX-ray database for research purposes only, continuously growing; Metadata: offset, sex, age. finding, survival temperature, pO2, saturation, view, modality, artifacts/distortion, notes; Categories: covid, pneumonia, no finding
1b)University of Waterloogithub.com/agchung/Actualmed-COVID-chestxray-dataset8 bits, 237 casesPNG, BMPX-ray database for research purposes only, continuously growing; Metadata: finding, view, modality, notes; Categories: covid, no finding
1c)Qatar & Bangladesh Universitieskaggle.com/tawsifurrahman/covid19-radiography-database[29]8 bits, 21165 casesPNG, resizedX-ray database; No metadata; Categories: COVID-19 positive cases (3616), normal (10,192), lung opacity (Non-COVID lung infection - 6,012), viral pneumonia (1,345)
1d)University of Montrealgithub.com/ieee8023/covid-chestxray-dataset[27], [29], [30], [31], [33], [35], [36], [40], [42], [43], [44], [46], [48], [50], [51]8 bits, 951 casesJPG, PNG, resizedX-ray database; Metadata: covid severity scores, sex,age, finding, RT_PCR_positive, survival, intubated, intubation_present, went_icu, in_icu, needed_supplemental_O2, extubated, temperature, pO2_saturation, leukocyte_count, neutrophil_count, lymphocyte_count, clinical_notes, other_notes; Categories: covid, viral, bacterial, fungal, lipoid, aspiration, unknown
1f)National Institutes of Healthkaggle.com/c/rsna-pneumonia-detection-challenge[29], [43], [48], [50]8 bits, 30227 (training)+3000 (test) casesDICOM, resizedX-ray database of Pneumonia Detection Challenge; No metadata; Categories: normal. lung opacity, no lung opacity/not normal
7)National Institutes of Healthnihcc.app.box.com/v/ChestXray-NIHCC[33], [43], [48]8 bits, 112120 casesPNG, resizedX-ray database of Common Thorax Disease; Metadata: finding ROI; Categories: no findings and 14 disease categories (Atelectasis, Cardiomegaly, Effusion, Infiltration, Mass, Nodule, Pneumonia, Pneumothorax, Consolidation, Edema, Emphysema, Fibrosis, Pleural_Thickening, Hernia)
8)National Institutes of Healthkaggle.com/nih-chest-xrays/sample[27]8 bits, Random sample of 5606 from 112,120 images of 30,805 unique patientsPNG, resizedX-ray database; Metadata: finding labels, follow-up, age, gender, view; Categories: Atelectasis, Cardiomegaly, Effusion, Infiltration, Mass, Nodule, Pneumonia, Pleural_Thickening, Hernia, Pneumothorax, Consolidation, Edema, Emphysema, Fibrosis
9)University of Montrealkaggle.com/praveengovi/coronahack-chest-xraydataset[31], [40]8 bits, 5910 cases (normal-1576, covid 58, SARS-4, virus-1493, bacteria 2777, ARDS-2)JPG,PNG-resizedCollection Chest X Ray (anterior-posterior) of Healthy vs Pneumonia (Corona) affected patients infected patients along with few other categories such as SARS (Severe Acute Respiratory Syndrome), Streptococcus & ARDS (Acute Respiratory Distress Syndrome); No metadata
10)University of California San Diegokaggle.com/paultimothymooney/chest-xray-pneumonia[31], [35], [42], [44], [51]8 bits, 5863 casesJPGChest X-ray images (anterior-posterior) were selected from retrospective cohorts of pediatric patients of one to five years old from Guangzhou Women and Childrens Medical Center, Guangzhou. All chest X-ray imaging was performed as part of patients routine clinical care.; Categories: normal and pneumonia; No metadata
11)University of California San Diegogithub.com/UCSD-AI4H/COVID-CT[30], [39], [41], [46]8 bits, 349 casesImages collected (scanned) from covid-related and medical papers in PNG (covid) or JPG (normal)This dataset has 349 CT images containing clinical findings of COVID-19 from 216 patients; Categories: covid and noncovid cases; Metadata: age, gender, location, medical history (unfortunately modest), time after the onset of illness, severity, other diseases
12)University of California San Diegodata.mendeley.com/datasets/rscbjbr9sj/2[36]8 bits, 5233 casesJPG (QF=95 for normal and QF=75 for pneumonia)Collection Chest X Ray; Categories: normal (1349 cases) vs pneumonia (3884 cases) including subcategories of bacteria and virus; No metadata
13)Elazig in Turkeygithub.com/muhammedtalo/COVID-19[27], [29]8 bits, 1125 casesJPG (QF=90, subsampling2x2), PNG (resized)X-Ray Images collection; No metadata; Categories: covid (125 cases), no findings (500 cases), pneumonia (500 cases)
14)National Library of Medicineopeni.nlm.nih.gov/gridquery?it=xg&coll=cxr&m=1&n=100[31], [43]8 bits or full bits, 7470 casesPNG (resized), Full DICOMChest X-rays collection with 3,955 radiology reports; Categories: 14 pulmonary categories; Metadata: time after the onset of illness, severity, other diseases, captions of symptoms as unstructured symptom description
15)Stanford University School of Medicinestanfordmlgroup.github.io/competitions/chexpert[43], [50]8 bits, 224,316 chest radiographs of 65,240 patientsJPGLarge dataset of chest X-rays which features uncertainty labels and radiologist-labeled reference standard evaluation sets; Categories: each report was labeled for the presence of 14 observations (no finding, enlarged cardiom., cardiomegaly, lesion, opacity, edema, consolidation, pneumonia, atelectasis, pneumothorax, pleural effusion, pleural other, fracture, support devices) as positive, negative, or uncertain; Metadata: related to above categories (blank for unmentioned, 0 for negative, -1 for uncertain, and 1 for positive)
16)Hospital San Juan de Alicante - University of Alicantebimcv.cipf.es/bimcv-projects/padchest[43]8 bits, more than 160,000 images from 67,000 patientsPNGPadChest: A large chest x-ray image dataset with multi-label annotated reports; the reports were labeled with 174 different radiographic findings, 19 differential diagnoses, and 104 anatomic locations; a 27% of the reports were manually annotated by trained physicians; Metadata: age, sex
17)Hospital Universitario San Ceciliogithub.com/ari-dasci/OD-covidgr[34]8 bits, 852 imagesJPEG (QF=90)X-ray images: 426 positive covid cases and 426 negative cases; only the posterior-anterior view is considered; Categories: covid severity - normal-PCR+ (76), mild (100), moderate (171), severe (79); General metadata: positive images correspond to patients who have been tested positive with RT-PCR within a time span of at most 24h between the X-ray image and the test; every image has been taken using the same type of equipment and with the same format
18)Beth Israel Deaconess Medical Center in Bostonphysionet.org/content/mimic-cxr/2.0.0[43]full bits, 227,835 imaging studies for 65,379 patientsfull DICOMChest radiographs with metadata: electronic health record data, dicom metadata, free-text radiology reports Categories: 14 pulmonary observations with an additional uncertain category
19)Societ Italiana di Radiologia Medica e Interventisticasirm.org/category/senza-categoria/covid-19[49]8 bitsmostly JPG (QF=95, subsampling2x2)Chest radiographs with free-text radiology and clinical reports, covid confirmation; Metadata includes selected information from electronic health record (e.g. symptoms, lab exams, ARDS, ventilatory assistance, previous exams); Categories: covid confirmation or no with 14 pulmonary observations
20)National Cancer Institutewiki.cancerimagingarchive.net/display/Public/LIDC-IDRI[45]full bits, 1308 casesfull DICOMThe Lung Image Database consists of diagnostic and lung cancer screening thoracic CT scans with marked-up annotated lesions (XML); it includes three categories (”nodule>=3 mm”, ”nodule<3 mm”, and ”non-nodule>=mm”);
21)University of Bresciabrixia.github.io#dataset[50]full bits, 4,707 casesfull DICOMCOVID-19 subjects, acquired with both CR and DX modalities, in AP or PA projection with highly expressive multi-zone COVID-19 severity score, fully annotated; Metadata: the multi-region 6-valued Brixia-score defined for six zones, sex, age
22)open-edit radiology resourceradiopaedia.org[49]8 bits, a significant number of cases, constantly updatedJPG with different QF, resizedDatabase of general radiological purposes; in selected cases free-text radiology and clinical reports, selected; generally, quantitatively and qualitatively differentiated case reports
23)generated using data augmentationkaggle.com/nabeelsajid917/covid-19-x-ray-10000-images[29]8 bits, 104 casesJPEG with different QF, resizedCorona Virus X-ray Dataset; Categories: covid and normal; No metadata
24)offline database or from hospital[28], [32], [38], [45]
This table presents the data sources used in studies reviewed in this survey. For each data source, we list articles that use it. The JPEG quality factor (QF) for most images has been set to 75, other cases are indicated. In the case of COVID-Net, please note that it is not a data source, but a study collating 5 datasets. Some other studies refer to it instead of referring to the original source. For medical imaging, the standard format for representing measured and / or reconstructed data is DICOM (Digital Imaging and Communications in Medicine). Significant features of this format are the ability to faithfully record 16-bit dynamics of data in grayscale (CT, radiography), control of acquisition parameters, and the ability to adapt to presentation conditions at diagnostic stations. The use of full data dynamics, i.e., all information about the imaged objects, taking into account the characteristic properties of the entire measurement and reconstruction process of the data (equipment parameters, filters, parameters of the acquisition process, specificity of pre-processing and forms of image representation) enables the construction of models based on complete measurement information about the examined object. Unfortunately, the vast majority of COVID-19 resources do not retain the source image information on the diagnosed objects. The data is converted from DICOM to typical multimedia image formats (mainly JPEG, PNG, TIFF standards) with the omission of information about the imaging process itself and often with a loss of quality and informative value of the compressed data. Data dynamics is often reduced to the 8 most significant bits, quantized to simplified 8-bits representation or all image information is lossy compressed (JPEG) using standard quantization tables.

Scarcity of publicly available COVID-19 data sources with images in raw DICOM format

It was observed that only one out of five repositories with the DICOM extension presented in Table 1 contains COVID-19 cases. Most databases regarding COVID-19 images are in 8-bit JPG or PNG formats. There are concerns that the quality of shared images is degraded, which may render the trained models less accurate. The quality degradation includes: the Hounsfield unit (HU) values are inaccurately converted into grayscale image data, and the number of bits per pixel and the resolution of images are reduced. An extreme case is the use of digital scans of printed lung images with no standard regarding image size, e.g., images extracted from the manuscripts [49]. Comparative statistical analysis based on the value of systematic measurement errors for the COVID-19 data, including the raw data and the metadata extracted from official reports, showed noticeable and increasing measurement errors [16]. This matter showed the importance of the accuracy, timeliness and completeness of COVID-19 datasets for better modeling and interpretation.

Too few images with low and moderate severity cases

Most studies are based on data sources publicly available on the Internet on a popular sharing platform, such as GitHub and Kaggle. The most commonly used data source with COVID-19 cases was created at the beginning of the epidemic. The first publicly available repository was published on January 15, 2020. In [34] it is stressed that available data sources contain too few images with low and moderate severity cases. Most of the data sources have only class labels without fine-grained pixel-level annotations, for example 3), 4), 8), 9), 10).

Relatively low number of COVID-19 images

Image format is one problem, while the amount of available data is another problem. The median number of COVID-19 images in the considered data resources is 250. With so little data, it is difficult to train a deep neural network (DNN). Table 2 shows the number of cases in particular classes. The last row with aspect ratio shows the proportion of the COVID cases to non-COVID cases.
Table 2

Class balance in the reviewed studies. The class balance is crucial developing an accurate model. In the following rows there are presented: the number of COVID-19 images in the study, the total number of images, the number of classes into which images were divided, and aspect ratio. The aspect ratio is calculated by dividing the number of COVID-19 images by the total number of images and then multiplying it by the number of classes. The biggest collected COVID-19 dataset and the largest total number of images are marked in bold. The smaller the aspect ratio, the less COVID-19 cases participate in the whole study’s dataset, and vice versa. For this reason, the best-balanced dataset (nearest 1) is marked in bold. Studies which do not include full information about the number of cases were excluded.

Study[27][28][29][30][31][32][33][34][35][36][37][38][39][40][42][44][45][47][49][51]
Number of COVID-19 images2502308553455023,389127377112/137763584004005823468829*299120269
Total number of images652346015,959720100421861125754366594913,9757508002800123459411,865*218,5292395 801
Number of classes2/33324235*33/23322412*142323
Aspect ratio0.08/0.111.50.160.9621.360.340.630.92/0.750.040.081.071.000.082.280.050.890.021.000.14

multilabel classification, training slides (106 COVID-19 images), four COVID-19 classes (NormalPCR+: 76 cases, Mild cases: 80, Moderate: 145 cases, Severe: 76 cases) and one Negative (377 cases)

Class balance in the reviewed studies. The class balance is crucial developing an accurate model. In the following rows there are presented: the number of COVID-19 images in the study, the total number of images, the number of classes into which images were divided, and aspect ratio. The aspect ratio is calculated by dividing the number of COVID-19 images by the total number of images and then multiplying it by the number of classes. The biggest collected COVID-19 dataset and the largest total number of images are marked in bold. The smaller the aspect ratio, the less COVID-19 cases participate in the whole study’s dataset, and vice versa. For this reason, the best-balanced dataset (nearest 1) is marked in bold. Studies which do not include full information about the number of cases were excluded. multilabel classification, training slides (106 COVID-19 images), four COVID-19 classes (NormalPCR+: 76 cases, Mild cases: 80, Moderate: 145 cases, Severe: 76 cases) and one Negative (377 cases) The use of imbalanced datasets requires more attention during the model training. Either proper data resampling [32] (oversampling [36], [48], undersampling [35]) should be applied, or an appropriate loss function should be chosen [29], [44], [46], [47], unless acquiring a greater amount of less common data is possible. It is also possible to use micro-base metrics [42]. However, most ML algorithms do not work very well with imbalanced datasets.

The data sources lack descriptions

Data resources: 4), 6), 9), 10), 12), 13), and 23) did not include metadata. At a minimum, the description of the dataset should include the following factors. First of all, the total number of images and the number of images in each class should be given. Additionally, the balance in terms of age and sex is another important factor because of the differences in anatomy. Information about smokers or previous lung diseases is also relevant. For analyzing model responses, the information about concurrent diseases, the severity of COVID-19, and the number of days between the exposure and the acquisition of the image of the chest are also useful.

Mix of CT and X-ray images

The problem that we found in these datasets is the data purity. If we look closer at the images presented in study [46], it appears that CT and X-ray images are mixed in the X-ray dataset. These two techniques are so different that networks for CT and X-ray images should be trained separately.

Inappropriate CT windows

For COVID-related lung analysis, it is essential to have Hounsfield Units equivalent for lung window (width: 1,500, level: -600). Otherwise, the lung structures are obscured or not visible at all, such as some examples in studies [28], [39], [41]. This is a basic, but key, issue because we do not want to assess soft tissues or bones. Photos taken in other windows do not have any real diagnostic value.

Children are not small adults

When we go back to the databases, it appears that, in some cases, e.g. 7), 8), 9) the X-rays of children and adults are mixed. The next problem is related to the mixture of images of patients of different ages. There are crucial differences between the X-rays of children’s and adults’ chests: technical (hands are often located above the head), anatomical (different shape of the heart and mediastinum, as well as bone structures), and pathological (different pathologies). This will also include a different course of infectious diseases, with the most vivid example of round pneumonia mimicking tumors [53]. It is important to mark the age of patients in data resources, and to separate children from adults when preparing data for training.

CT and X-rays images are not in color.

Despite that fact, some databases, e.g. 5) and 11), include images in RGB color space. It introduces redundant information, because values in all channel are the same (R=G=B). This situation leads to increasing the number of input neurons in the neural network by three times. Due to that fact, the number of parameters will also rise, and the training may require more data and time, however, it lacks extra information.

Incorrect categorization of pathologies

We have noticed that some images are incorrectly categorized - into normal or pathologic, e.g. in database 10), 13), and also within the class of pathology, e.g. in database 14). An additional problem is that, from a medical point of view, some images should be multi-categorized. This means that there is more than one pathology in one image. For instance, pneumonia (main class) can manifest itself as lung consolidations, which, however, can also appear with pleural effusion or atelectasis (two additional classes). On the other hand, atelectasis itself, with a mediastinal shift, can be a sign of a different pathology, such as a lung tumor. Thus, databases should be verified by experienced radiologists for proper categorization and maybe a rejection of multi-class images. This, however, would be time-consuming and - what is more important - very difficult or impossible with low-quality images or images without appropriate descriptions.

Lack of information about chest projection for X-ray imaging

This problem is present, for example in 2), 4), 9). There are two main chest projections, see Fig. 3 , Posterior-Anterior (PA) and Anterior-Posterior (AP). The first one is acquired while the patient is standing. The X-ray beam comes through the patient’s chest from its back (posterior) to front (anterior) - i.e., PA projection. The second one is the opposite - the beam enters through the front (anterior) of the chest and exits out of the back (posterior) - i.e., AP projection. This type of examination is mostly conducted in more severe cases, with lying patients, with comorbidities, often in Intensive Care Units. As the X-ray beam is cone-shaped, both projections have one very important difference, which is the size of the heart. In PA projection, the heart is close to the detector, resulting in a similar heart size on the X-ray as in reality. In AP projection, the heart is away from the detector, resulting in a larger heart size on the X-ray, which can be confused with cardiomegaly. In databases, AP and PA images are often mixed, which can cause bias because AP projections are performed on severely ill patients [34]. From a medical point of view, it is impossible to perform chest X-rays in only one projection as this depends on patients condition. However, projection should be specified for every X-ray in dataset, and possible bias in model classification should be evaluated.
Fig. 3

Differences between AP and PA chest projections.

Differences between AP and PA chest projections. To sum up, this section shows that data sources have several weaknesses. First of all, images available for COVID-19 in public databases are in not the most dedicated image format because DICOM images are still rarely available for this disease. Secondly, in the data sources, there is missing data (i.e., chest projection) or poor quality data (i.e., poor image quality, not grayscale images, inappropriate CT window, mixed CT and X-ray images, or incorrect pathology categorization). Lastly, during data preparation, it should be taken into account that DNNs work better when the class balance is maintained.

Image preprocessing methods

The aim of preprocessing is to make the images from different data resources look homogeneous and coherent. This process reduces the possibility of bias via eliminating some artifacts from images, such as captions, annotations, which may deceive the model. The model should learn how to differentiate labels by focusing on image features, not by recognizing from which database the image comes from. During preprocessing, irrelevant image features that are easier to learn are removed. This is because in some databases there are no cases of people suffering from COVID-19, while in others there are, for example, only serious cases. These differences, which are insignificant from a human perspective, must be eliminated. For machines, even the information that images from one data resource are relatively darker might be relevant. However, due to a large amount of data, automation of preprocessing is necessary. Preprocessing cannot introduce any changes in an image which will add or remove some relevant information. Its purpose is to make it impossible to identify the machine or characteristic machine’s calibration parameters, e.g., the dose of exposure. Table 3 lists preprocessing techniques used in the reviewed studies. The most common was resizing an image to the same size. It is the most basic operation needed to train DNN when images have different sizes. Other techniques applied frequently to images were: normalizing pixel intensity, changing color space, eliminating noise, equalizing histogram, and performing image enhancement. Unfortunately, in 9 out of 25 studies there was no information about preprocessing steps provided. [54] also stressed that many studies do not contain sufficient information about preprocessing, such as cropping of images.
Table 3

Image preprocessing techniques in the reviewed studies.

Preprocessing techniqueReference
Resize to the same size[29], [30], [31], [32], [35], [36], [40], [41], [42], [43], [44], [46], [49], [51]
Normalize pixel intensity[29], [35], [42], [43], [44], [49], [50]
Eliminate noise[29], [30], [34], [50]
Use Perona-Malik filter[29]
Limit image intensity[42], [50]
Equalize histogram[29], [31], [40], [50]
Perform image enhancement[29], [30], [31], [32]
Cast data type[29], [31], [50]
Change color space[36], [46], [49]
Crop image[34], [40], [42]
Zoom image / augmentation[41], [49]
Add pixels[34]
Feature encoding[51]
Rotate image[41]
Use 2D wavelet transform[30]
Feature extraction[47], [51]
Lack of preprocessing or description[27], [28], [33], [37], [38], [39], [45], [47], [48]
Image preprocessing techniques in the reviewed studies. Cropping, changing color space, proportionally resizing, or zooming can be helpful to adjust images for training on specific network architecture, or the easiest way to remove some descriptions from the edges of images. If not required, resizing ought to be omitted. Normalizing pixel intensity or equalizing histograms are required to eliminate strong correlations with specific machine settings. Spot changes, such as noise removal, are not desirable. These techniques can be used only very carefully in order not to remove important features, such as lesions or parts of them. To sum up, preprocessing is an important step preceding model training. It should reduce the possibility of bias and guarantee more homogeneous images without the elimination of any medically significant features.

Data augmentation

Data augmentation for ML is a technique that artificially multiplies the number of images through cropping and transforming existing images or creating new synthetic images thanks to generative adversarial networks. This procedure may help to reduce model overfitting and the problem of class imbalance. It helps in achieving a larger training dataset and more robust models. In Table 4 , we summarized data augmentation techniques from the reviewed studies. The most popular augmentation techniques in the reviewed studies are affine transformations, such as rotation, scaling or zooming, flip, and shifting or translation. On the contrary, splitting a radiological image into overlapping patches, or generating new content via a type of Generative Adversarial Network are rarely used.
Table 4

Data augmentation techniques used in studies. Some techniques are parametrizable, so the table indicates the techniques and parameters used. An indentation is used to show the subtypes of the method.

Data augmentation techniqueValues and studies
Affine transformations[28]:
Rotation[35], [37], [45], [48], 5[40], 15[27], [29], [49], 20[44], 25[50]
Scaling / Zooming[37], [44], [48], 10% [40], [46], [50], 20% [35]
Flip[44], [46], [47]
Horizontal[37], [38], [40], [45], [48]
Vertical
Shifting / Translation[37], [44], [48], height 5% [40], 10% [50]
Shearing[35], [36], [44]
Brightness change[28], [37], [48] +/-30 [36], 10% [40]
Crop[38], [45]
Contrast change[28]
Gaussian noise[36]
ZCA whitening transformation[44]
Elastic transformationα=60, σ=12 [50]
Grid distortionsteps=5, limit=0.3 [50]
Optical distortiondistort=0.2, shift=0.05 [50]
Warping10% [46]
Multiple patches from each image[31]
Class-inherent transformations Network*[34]
Augmentation used but parameters are not specified[41]
No augmentation used[30], [32], [33], [39], [42], [43], [51]

* Inspired by generative adversarial networks.

Data augmentation techniques used in studies. Some techniques are parametrizable, so the table indicates the techniques and parameters used. An indentation is used to show the subtypes of the method. * Inspired by generative adversarial networks. However, not all of them are appropriate from a medical point of view. Before an augmentation, it is recommended to consider the ’safety’ for the chosen domain. For example, the rotation should be done carefully, because some parts of the lungs, such as costophrenic recesses, may be placed outside the image. Also, change of brightness or contrast should be performed only in a limited manner, as greater manipulation may obscure lung structure. Moreover, in predicting COVID-19, it is acceptable to crop or proportionally scale/zoom an image to such an extent that it displays only the lungs without a background or other parts of the body. It is also worth noting, that in the case of CT and X-ray images, the augmentation based on rotation or flipping generate photos that cannot naturally appear in real datasets, because the process of taking the photo itself is standardized. Horizontal flips should be done carefully, with some specific limitations. Most pathologies will be present similarly on the left or right lung, except for the change in shape of the heart (like in dextrocardia) or pathologies affecting specific lobes, due to different lung anatomy (like lobar pneumonia or lobar atelectasis). These limitations should be taken into consideration in model design. In general, all augmentation methods should be consulted with radiologists, as domain knowledge is crucial. In every project, it is important to know the field of research to avoid a situation in which instead of solving the problem, bias is accidentally introduced.

Model architecture

In the studies different approaches of modeling were applied. Some benefited from machine learning methods, whereas the rest used deep learning. In the first case, simple classifiers or their ensembles were applied: AdaBoost [41], KNN [41], Naive Bayes [41], SVM [41]. In the reviewed studies, lung-specific model architectures (own models) were relatively often used for classification, whereas the existing architectures were frequently fine-tuned. The following model architectures or their fine-tuned, modified versions were investigated: ResNet [29] (ResNet18 [31], [42], [47], [50], ResNet34 [32], [42], [46], ResNet50 [30], [39], [40], [45], [46]), DenseNet [29], [43], [48], [50] (DenseNet121 [40], [49], DenseNet-161 [42], DenseNet-201 [39], [46], [49]), VGG [29], [50] (VGG-16 [27], [39], [49], [51],VGG-19 [39], [46], [49]), Inception [50], InceptionV3 [35], [42], InceptionResNetV2 [39], [42], [49], MobileNetV2 [39], [49], NASNetMobile [39], [49], EfficientNet-B0 [46], Efficient TBCNN [40], MobileNet [49], NASNetLarge [49], Res2Net [38], Attention-56 [49], ResNet15V2 [39], ResNet50V2 [44], ResNeXt [42], WideResNet [42], Xception [49], own model [28], [33], [34], [36], [37], [41], [46], [50]. It is clearly visible that there are numerous types of neural networks. Different neural networks can catch different dependencies in the data. For solving a problem, many types of model architectures are tested to find the best one for a specific task. Recommendations on how the explanations should look do not depend on the neural network architecture. For segmentation, the following architectures were used: U-net [34], [45], [50], AutoEncoder [47], VGG-16 backbone + enhanced feature module [38], (FC)-DenseNet-103 [31], Nested version of Unet (Unet++) [50], VB-Net [32]. During the segmentation process, it is important that the lungs are accurately segmented. Otherwise, distorted border lines can be an indication of pathology. In study [31], the authors were aware that their segmentation cut pathological changes in lungs. In study [50] segmentation for non-domain experts appears accurate. However, radiologists noticed that also other structures (i.e., bowel loops) were interpreted as lungs in that segmentation. There are multiple purposes for creating new model architectures. The most common is adjusting existing architectures for better explainability or scalability for training on medical COVID-19 imaging [28], [41]. For example, in studies [37], [46], the authors conducted tests and chose the advantages of many architectures while creating their own. The proposed architectures are usually smaller and require a lower number of trainable parameters than in well-known DNN architectures [34], [36]. Six studies published their code on GitHub: [31], [35], [37], [40], [43], [48]. Other studies did not include any reference to their code or model. Often the prediction from multiple models is combined to improve the overall performance. However, surprisingly, in the reviewed studies, there were not many ensemble models: [29], [32], [42], [50].

Transfer learning

Transfer learning is an ML technique about reusing gained knowledge from one problem to a new one. In the reviewed studies, it is commonly used when the neural network has a large number of parameters or the number of collected samples is too small for a specific task. In such a case, fewer training epochs are needed to adjust the model to a particular task. There are several popular image databases: ImageNet and, NoisyStudent for which various architectures of pre-trained neural networks are available. Transfer learning on ImageNet database was utilized in the following studies: [27], [31], [34], [35], [36], [37], [39], [40], [41], [46], [49], [51]. Twelve out of 25 studies decided to use a neural network pre-trained on ImageNet for transfer learning. Therefore, it can be said that this is a very common procedure. However, as [55] shows, it is not clear whether using ImageNet for transfer learning in medical imaging is the best strategy. ImageNet consists of natural images. Meanwhile, medicine is an entirely different field and is completely unrelated. [56] also stressed the fact that the features which are extracted by models pre-trained on ImageNet can introduce bias. Only in three of the reviewed studies was transfer learning conducted on lung images. The chosen datasets included 112,120 in [48], 88,079 in [43], and 951 in [50] non-COVID-19 lung images. The study [29] did not perform any transfer learning because lung images lack colorful patterns, specific geometrical forms, or similar shapes. The amount of redundant information introduced by a network pre-trained with color images may seriously affect the learning process on gray level images. In study [40], the authors discovered that the model has better performance when pre-trained on ImageNet than without it. However, the authors found out that their models pre-trained on ImageNet were using irrelevant markers on lung images while making a prediction. Especially when the model is trained on a small amount of data, the usage of completely irrelevant features from another pre-trained model may increase model accuracy/result. For this reason, it is crucial to find a large database with images similar in domain and appearance to limit the possibility of irrelevant markers that take part in a prediction. It is recommended to train a neural network on this database and then use transfer learning to adjust it to the target task. For transfer learning, it is recommended to take into consideration the following X-ray data sources with DICOM images (consider the fact that, in some of them, children and adults lungs are mixed): U.S. National Library of Medicine1 (7,470 images), Radiological Society of North America2 (29,684 images), Society for Imaging Informatics in Medicine3 (3,209 images), Medical Information Mart for Intensive Care4 (377,110 images). For transfer learning on CT, the following data sources are available: The Reference Image Database to Evaluate Therapy Response5 (15,419 images), A Large-Scale CT and PET/CT Dataset for Lung Cancer Diagnosis6 (260,826 images), The National Lung Screening Trial7 (21,082,502 images).

Training parameters

The selection of hyperparameters has a large impact on model results. Nevertheless, the process of tuning parameters is empirical and depends on the model architecture. For this reason, it is difficult to present a set of parameters adequate for every model architecture. However, there are several tips which can be used for most models. Often the learning rate is decreased during the training process. Sometimes callback functions are used to halt training, when the result of a model is optimum, and during the training process, to save and store the best model and its parameters. The most typically used optimizer is Adam. The batch size of images during model training is between 2 and 81 with the most common value 8. The whole image dataset is typically divided into 3 or 2 sets, most commonly into: training set 80%, validation set 10% and testing set 10% [34], [36], [48], [51]. Proportion 80% to 20% was the most frequently used among divisions into training and testing set respectively [39], [44]. In study [39], the recommendation to conduct external validation is indicated, meaning an evaluation on an independent database. Another public dataset will be the best choice for cross-database validation [31]. However, in the reviewed studies cross-validation is the most frequently used. It is a common choice for training on a small amount of data resources. The problem which may occur during cross-validation is overfitting to the data. For this reason, validation on an external resource is the most trustworthy method.

Model performance

Evaluation metrics are commonly used to compare different models. For DNN image classification, there are many metrics frequently used for model quality assessment. In the reviewed studies, we discovered a large variability in the number of reported metrics. It is a common situation due to the fact that there are no detailed recommendations as to which performance metric should be used. We recommend the instructions presented in [57], but, unfortunately, in almost all the reviewed studies, at least one metric out of these recommendations was missing. Based on the rules described in study [57], there are six evaluation criteria for binary classification: accuracy, precision, recall (sensitivity), F score, specificity, AUC. For multi-class classification, there are eight criteria: average accuracy, error rate, precision, recall, F score, precision, Recall, F score, and for multi-label classification four criteria: exact match ratio, labeling F score, retrieval F score, Hamming loss. There is another important factor which indicates why more than one evaluation metric should be used. It provides the opportunity to compare model architectures and then choose the best one for a given problem. Nevertheless, the models were not trained on the same images. Some databases contained only severe cases which were easier to classify [34]. Even if studies refer to the same data resources, it is possible that the amount of data has increased over time. For this reason, it is rather difficult to make a reliable comparison. The most trustworthy way to compare different model architectures is to look at studies which tested many of them, i.e. [34], [39], [41], [46], [49], [50].

Explainable artificial intelligence

The importance of model explainability

When designing predictive models for healthcare, or any other high-stakes decisions, the explainability of the model is a key part of the solution. The empirical performance of the model is very important, but there can be no responsible modeling if the issue of explainability is not addressed properly for each stakeholder of the system. For physicians, the lack of explainability drastically reduces the confidence in the system. For model developers, it makes it difficult to detect flaws in model behavior and obstructs debugging [58], [59]. For predictive models, two general approaches to explainability are either by using classes of interpretable-by-design models or using post-hoc explanations. Despite the obvious advantages of interpretable-by-design models, their construction requires more domain knowledge linked to the construction of interpretable features. The advantage of post-hoc explanations is that they are constructed after the model has been trained. Thus, the developer can focus on model performance by pouring large volumes of data into a neural network and then deal with model explanations afterwards. In the analyzed studies, the authors used only post-hoc methods which are prevalent in computer vision tasks [60]. Examples of post-hoc explanations are presented in Fig. 4 .
Fig. 4

Examples of explanations for COVID-related models from studies: [34], [36], [39], [42], [43], [44]. The following explanations are used: a) Grad-CAM, b) CAM, c) saliency, d) guided backpropagation, e) integrated gradients, f) LIME. Such explanations can be divided into 4 types: heat maps (image a) - c)), contour lines (d)), points (e)), and image pieces (f)).

Examples of explanations for COVID-related models from studies: [34], [36], [39], [42], [43], [44]. The following explanations are used: a) Grad-CAM, b) CAM, c) saliency, d) guided backpropagation, e) integrated gradients, f) LIME. Such explanations can be divided into 4 types: heat maps (image a) - c)), contour lines (d)), points (e)), and image pieces (f)). Due to the mode of operation, also post-hoc explanation methods can be divided into two groups. The first group consists of input perturbation methods such as Locally Interpretable Model Explanations (LIME), or Occlusion Sensitivity. These methods are based on the analysis of the change of the model response after obscuring, removing, or perturbing some part of the image. The advantage of this class of methods is that they are insensitive to the internal structure of the model. Such so-called model-agnostic approaches assume nothing about the internal structure of the model. By analyzing how a series of input perturbations affect the final prediction, it determines which part of the input is important. The second group are methods based on the analysis of signal propagation through the network, i.e. model-specific methods. This group of methods uses detailed information about the network architecture and the design of subsequent layers to determine the key regions of input for the final prediction. The advantage of such approaches is that usually, one pass through the structure of the network is sufficient to generate explanations. Model specific methods for explanations of CNNs can be organized into a spectrum of solutions, from gradient-based methods to activation map-based methods. For gradient-based methods, the gradient between the output model class and the input image is used to calculate saliency maps. For large networks, such as most of those shown in Table 5 , the gradient information is very noisy, so there have been many modifications to this method that reduce noise by smoothing or thresholding or rescaling. This class of models includes Guided Backpropagation, Layer-wise relevance propagation, and SmoothGrad.
Table 5

The depth, number of parameters and type of layers for neural networks in considered papers. For large networks gradient based explanations are noisy. Some explanation techniques assume specific types of layers.

Model architecturesDepthNo. of parametersLayer types
ResNet18, ResNet34, ResNet50, ResNet15V2, ResNet50V2-11.7M-25.6MZeroPadding2D, Conv2D, BatchNormalization, Activation, MaxPooling2D, Add, GlobalAveragePooling2D, Dense
DenseNet121, DenseNet-161, DenseNet-201121-2018.1M-20.0MZeroPadding2D, Conv2D, BatchNormalization, Activation, MaxPooling2D, Concatenate, AveragePooling2D, GlobalAveragePooling2D, Dense
VGG-16, VGG-1923-26138-144Conv2D, Dense, Flatten, InputLayer, MaxPooling2D
InceptionV315923.9MConv2D, BatchNormalization, Activation, MaxPooling2D, AveragePooling2D, Concatenate, GlobalAveragePooling2D, Dense
InceptionResNetV257255.9MConv2D, BatchNormalization, Activation, MaxPooling2D, AveragePooling2D, Concatenate, Lambda, GlobalAveragePooling2D, Dense
MobileNet884.3MConv2D, BatchNormalization, ReLU, DepthwiseConv2D, ZeroPadding2D, GlobalAveragePooling2D, Reshape, Dropout, Activation
MobileNetV2883.5MConv2D, BatchNormalization, ReLU, DepthwiseConv2D, ZeroPadding2D, Add, GlobalAveragePooling2D, Dense
NASNetMobile, NASNetLarge-5.3M-88.9MConv2D, BatchNormalization, Activation, ZeroPadding2D, SeparableConv2D, Add, MaxPooling2D, AveragePooling2D, Cropping2D, Concatenate, GlobalAveragePooling2D, Dense
EfficientNet-B0-5.3MRescaling, Normalization, ZeroPadding2D, Conv2D, BatchNormalization, Activation, DepthwiseConv2D, GlobalAveragePooling2D, Reshape, Multiply, Dropout, Add, Dense
Efficient TBCNN0.23MConv2D, MaxPool2D, BatchNormalization, GlobalAveragePooling2D, Add, Dense
Attention-5611531.9MConv2D, Lambda, MaxPool2D, UpSampling2D, AveragePooling2D, ZeroPadding2D, Dense, Add, Multiply, BatchNormalization, Dropout
Xception12622.9MConv2D, BatchNormalization, Activation, SeparableConv2D, MaxPooling2D, Add, GlobalAveragePooling2D, Dense
The depth, number of parameters and type of layers for neural networks in considered papers. For large networks gradient based explanations are noisy. Some explanation techniques assume specific types of layers. Methods based on activation maps, such as Class Activation Mapping (CAM) or DeepLIFT, focus on visualizing the relationship between the layer with the feature map (in most cases the penultimate layer of the network) and the model output. Assuming that the feature map stores information about the spatial relevance of features, one can explore what elements of the feature map are most relevant for the final prediction. Such methods often have an assumption about the structure of the network, such as global average pooling before the softmax layer. In our analysis, the most popular solution turned out to be the one that combines both mentioned above approaches, tracing the gradient between the model prediction and the feature map and then analyzing the spatial information of a specific part of the feature map. This group of methods includes the most popular explanation method Grad-CAM and its modifications Guided Backpropagation, Guided Grad-CAM, Grad-CAM++. Using gradient tracking between the feature map and network output is also a more flexible approach in terms of network architecture without enforcing global pooling.

XAI methods used in the reviewed studies

The area of model explanations and the number of methods that can be used for this purpose are increasing rapidly [59]. Such methods differ in properties; they work either for a single image (so-called instance level methods) or globally for the whole dataset. Some of them are based on gradients, others on interpretable features, some are intrinsic or post-hoc, model-specific (class-discriminative, high-resolution, multi-layer) or model-agnostic. Table 6 shows which approach to model explanation was used in each study. The most popular in the reviewed studies was Grad-CAM. Its popularity may be related to the fact that colorful heat maps are easy to implement and seem to be readable. An example of an implementation for Grad-CAM is available online, and its use on melanoma images shows great results.
Table 6

XAI techniques used in considered papers.

Name of the XAI techniqueReference
Grad-CAM (gradient-weighted class activation mapping)[27], [28], [29], [30], [31], [32], [33], [34], [35], [40], [48], [50]
LIME (local interpretable model-agnostic explanations)[39], [46], [50]
CAM (class activation mapping)[32], [36], [39], [44]
Saliency (saliency map)[40], [42], [43], [44]
Guided Backpropagation[42], [44]
LRP (layer-wise relevance propagation)[29], [34]
Occlusion (occlusion sensitivity)[30], [42]
AM (activation mapping)[38]
Attribution maps[47]
DeepLIFT[42]
Feature maps[49]
Grad-CAM+[29]
Guided Grad-CAM[44]
GSInquire[37]
Input X Gradient[42]
Integrated Gradients[42]
XAI techniques used in considered papers. Another very popular method is LIME. As clearly visible in studies [39], [46], some large superpixels include different structures (i.e., lung tissue and chest wall). Therefore, this method is not accurate enough for an interpretable representation of space due to the lack of semantic meaning. Some threads related to the application of XAI in the analyzed publications are questionable. Contrary to what [29] states, explanations of ensemble models are possible. A single best model does not have to be selected for the visualization of the prediction. There are many model-agnostic interpretation methods which do not rely on model architecture and can be easily used for explanations, such as: LIME, or Anchors. In most XAI methods, it is possible to adjust them to suit ensemble models. Moreover, ensemble models usually outperform a single model in terms of accuracy. According to Ozturk et al. [33], the model makes incorrect decisions in poor quality X-rays. This is because the low quality or very low-resolution images do not show enough details even for the models. Such images should be removed while checking the database contents. The study [30] noticed that the region of the lesion is marked correctly, but that model prediction is wrong. Unless we perform a quantitative and qualitative evaluation of XAI results, we will not have the opportunity to assess the trustworthiness of our model. The model may take into greater consideration other image features than it should. To explore this kind of a model mistake, other XAI methods ought to be used to obtain a better comparison possibility.

Domain experts’ evaluation of XAI methods

In most of the reviewed studies, the application of XAI comes down to the series of colorful images without any assessment about how valid these explanations are. Colored explanations obscure the original image, which makes it even more difficult to assess their correctness. In images with XAI heat maps, it is often hard or impossible to see pathologies and guess if the model works well. Raw lung images shall be put next to explanations. Also the explanations should be interpreted or validated by radiologists. Otherwise, they are redundant and contribute nothing to the trustworthiness of the model. Together with the radiologists, we analyzed the explanations from the discussed works. In the following paragraphs, we discuss the most common mistakes or inappropriate explanations. In the first example, in Fig. 5 a), the model focuses on clavicles, scapulas, and soft tissues, which are outside the lungs. Very likely, the model predicts illness based on an improper part of the image. Location of the areas marked by explanation should be inside the chest on the lung tissue because COVID-19 lesions are not located on, e.g., lymph nodes. Moreover, there are some elements that cannot be considered as decision factors like imaging artifacts (cables, breathing tubes, image compression) or embedded markup symbols [37]. To prevent the model from focusing on irrelevant features, in some studies, the lungs were segmented, and their background was removed [31], [32], [34], [45], [50]. However, it may not help when some imaging artifacts are present in the area of the lungs.
Fig. 5

Examples of biased model explanations: a) [36], b) [27], c) [40], d) [47]. Red arrows in the image b) are marked by a radiologist to help locate the lesions. They were not present in the training set.

Examples of biased model explanations: a) [36], b) [27], c) [40], d) [47]. Red arrows in the image b) are marked by a radiologist to help locate the lesions. They were not present in the training set. The second example, in Fig. 5b) shows that the model does not take the lesions into account. The model states that parts of the lungs other than the ones marked by the radiologist are relevant for model prediction. Explanations that roughly indicate the infection location [32] are not acceptable for the robust model. The model should do this with the accuracy of the pixel marked by radiologists as relevant. The third example, in Fig. 5c), visualization is not clear. The study describes a different XAI method than the one present in the image. Moreover, this visualization highlights the whole image, and it is not possible to guess which features took part in the prediction. It is important to point out that some explanation methods can give clearer results for a specific type of DNN and for a specific domain. The last example, in Fig. 5d) is blurred. The image of the lungs is improperly taken, and the process should be repeated. The current image is useless for the accurate diagnosis process. Such images should be removed during data resource verification before model training. If the lung lesions are well described, it will be possible to prepare quantity and quality XAI assessment to score the trustworthiness of the specific model. One possible option would be to create measures for the evaluation of XAI image models based on the measures quoted in study [61]: Intersection-Over-Union and token-level, which presents measures for the evaluation of text models. Evaluation of explanation methods is crucial for confirmation of model trustworthiness. First of all, radiologists should validate a specific model with the help of XAI. They should assess location, size, and shape of marked regions by explanation methods. Their interpretations should contain clear references to structures and lesions in the lungs, such as posterior basal segment, ground-glass opacity, consolidation, frosted glass shadows, etc. The example of a well-prepared XAI interpretation can be found in the study [30].

The checklist for responsible analysis of lung images with deep learning models

In this work, we have shown that development of a model which analyzes lung images is a complex process. Therefore, we prepared the checklist based on the analyzed studies and the errors we found in them. In [62], it is shown that well-prepared checklists significantly improve the quality of the modeling process. They help to avoid, or quickly detect and fix, errors. In the list below, the letter R indicates that the point should be consulted with a field expert / radiologist, and the letter D indicates that the point should be consulted with a model developer. The points in the checklist below are grouped according to the sections’ names discussed in this study. This should assist in finding a detailed description of the problem stated in the checkpoint list in the corresponding section. Data resources Does the data and its associated information provide sufficient diagnostic quality? If images are in DICOM, does the header provide the needed information? If not, is it provided in any other way? Are the low quality images (i.e., blurred, too dark, or too bright) rejected? Is the dataset balanced in terms of sex and age? Does the dataset contain one type of images (CT or X-ray)? Are the lung structures visible (lung window) on CT images? Are images of children and of adults labeled as such within the dataset? Are images correctly categorized in relation to class of pathology? Are AP/PA projections described for every X-ray image? Image preprocessing Is the data preprocessing described? Are artifacts (such as captions) removed? Data augmentation (if needed) Are the lungs fully present after transformations? Are lung structures visible after brightness or contrast transformations? Are only sensible transformations applied? Transfer learning (if used) Is the transfer learning procedure described? Is the applied transfer learning appropriate for this case (i.e.: images of same type and content have been used to train the original model)? Model performance Are at least a few metrics of those proposed in [57] used? Is the model validated on a different database than the one used for training? Domain quality of model explanations Are other structures (i.e., bowel loops) misinterpreted as lungs in segmentation? All the areas marked as highly explanatory are located inside the lungs? Are artifacts (cables, breathing tubes, image compression, embedded markup symbols) misidentified as part of the explanations? Are areas indicated as explanations consistent with opinions of radiologists? Do explanations accurately indicate lesions? According to the prepared checklist, in Tables 7 and 8 , we tried to analyze which points are fulfilled by the reviewed studies and the datasets used in these papers for the neural network training. Due to the possibility of changes in preprints, we only examined papers already published in journals.
Table 7

Summary showing which points from the checklist are fulfilled by the reviewed data resources.

Checklist / Data resource2)3)4)5)6)7)8)9)10)11)12)13)14)17)23)
[D] Does the data and its associated information provide sufficient diagnostic quality?Y?N?NN?N?NN?NN?NNNYN?N
[R] Are the low quality images rejected?NNNNNNNNNn/aYNNY?N
[D] Is the dataset balanced in terms of sex and age?Y???Y?Y?YY?NN???Y??
[R] Does the dataset contain one type of images (CT or X-ray or the same projection)?YYYNYYNNYYYYNYN
[R] Are the lung structures visible (lung window) on CT images?n/an/an/an/an/an/an/an/an/aNn/an/an/an/an/a
[D] Are images of children and of adults labeled as such within the dataset?not allNNY?YYYNYnot allNNNNN
[R] Are images correctly categorized in relation to class of pathology?NNYNNNNNYNYNYN?N
[D] Are AP/PA projections described for every X-ray image?NYNYYYYNYn/aNNYYN
Table 8

Summary showing which points from the checklist are fulfilled by the peer-reviewed studies.

Checklist / Study[27][28][29][30][31][32][33][34][35][36][37][38]
Image preprocessing
[D] Is the data preprocessing described?YNYYYYYYYYYY
[D] Are artifacts (such as captions) removed???YYYn/a?YNNYn/a
Data augmentation (if needed)
[D] Are the lungs fully present after transformations????n/aY?Yn/a????N?
[R] Are lung structures visible after brightness or contrast transformations?n/a?n/an/an/aYn/a?n/a??n/a
[D] Are only sensible transformations applied?Y?Yn/aYYn/a??NNN
Transfer learning (if used)
[D] Is the transfer learning procedure described?Yn/an/aY?Yn/aYYYYYY
[D] Is the applied transfer learning appropriate for this case?Nn/an/aNNn/aNY?NNNN
Model performance
[D] Are at least a few metrics of those proposed in [57] used?YYYYYYYYYYYY
[D] Is the model validated on a different database than the one used for training?NNNNYYNNNNNN
Domain quality of model explanations
[R] Are other structures (i.e., bowel loops) misinterpreted as lungs in segmentation?n/an/an/an/aNY?n/aNn/an/an/aY
[R] All the areas marked as highly explanatory are located inside the lungs?Yn/aYn/aYYY?YNYY
[R] Are artifacts misidentified as part of the explanations?Yn/aNn/an/an/an/aNn/an/an/a
[R] Are areas indicated as explanations consistent with opinions of radiologists?Nn/an/an/an/an/an/an/aYn/an/an/a
[R] Do explanations accurately indicate lesions?Y?n/aY?n/aY?YNNNY?NY
Summary showing which points from the checklist are fulfilled by the reviewed data resources. Summary showing which points from the checklist are fulfilled by the peer-reviewed studies. We applied the following denotements: Y, N mean yes and no respectively (if an answer is probable then the additional ? is added), ? means there is no information provided, n/a signifies that the issue does not apply to a particular publication. Due to the fact that we can only evaluate the information contained in the article, the answers given to some questions from the checklist need to be clarified. The question regarding balance in the dataset (Table 7) has two components. Sometimes, the dataset is balanced concerning only one criterion but not in terms of the second one. In such cases, we put Y? sign. We did similarly in cases where many metadata are missing, but there is a balance considering the existing data. We would like to stress that mixed projections of X-rays are present in databases: 5), 8), 23), both PA and lateral. Databases 5) and 9) contain also CT exam images. Furthermore, dataset 23) includes CT scanogram described as X-ray which is inappropriate for medical use. It is important to note that if the datasets in Table 7 are labeled as containing images incorrectly categorized by pathology class, then in most cases patients should be classified into more than one pathology class. Moreover, in other databases, X-rays marked as pneumonia or other disorders have no radiological findings. In Table 8, for the papers [35], [36], the assessment of the presence of artifacts was made based on the images provided by the authors. In the question regarding the visibility of lungs after data augmentation, we put the value N? because there were no details about applied random cropping. Such an answer reflects the fact that cropping can be risky especially if the parameters are not carefully chosen. In case of [31], we put Y? as the authors honestly pointed out that their model for segmentation sometimes improperly generates masks when there is severe opacity. In such situations, the mask covers lungs only partially. We would like to explain the criteria for assessing data augmentation and transfer learning. In the case of the former, we put N when horizontal flip was applied. In the latter, the main criterion was whether the authors used a model pre-trained on ImageNet dataset. Such behavior is not recommended as natural scene images are significantly different from medical images. The biggest difference is the fact that X-ray and CT images are in grayscale unlike images in ImageNet. In the reviewed papers, there were two approaches to transfer learning: the backbone weights were frozen and the case where the pre-trained weights were just an initialization and were later trained. We decided to put N in both scenarios as both require artificial conversion of grayscale images to RGB. In Table 8, the difficulty to assess the solutions applied in the papers occurred also in the section requiring radiological expert knowledge. In the group Domain quality of model explanations, we were not able to check which of these points were fulfilled by each of the studies, as most publications contain only few or even none images, which could be radiologically evaluated. Nevertheless, although our observations were made on such limited data, we were able to identify many mistakes in the radiologic background.

Conclusion

The sudden outbreak of the COVID-19 pandemic has shown us how we need effective tools to support the physicians. Deep neural networks can offer much in the analysis of lung images, but responsible modeling requires very thorough model validation. Models without explanations create validation debt and explanations without consultations with a radiologist are just an illusion of validation. This work shows a critical analysis of 25 state-of-the-art articles that use deep learning models based on lung images to identify COVID-19. We have selected the most advanced papers in which the authors made extra effort to supplement the models with explanations. But even in this sample, it turns out that only in 7 out of the 25 reviewed studies, the models were consulted with radiologists and regarding the model explanations, they were validated only in three studies [33], [35], [50]. It is important to point out that, for medical examination, the most valuable are large-resolution images, especially in DICOM format. Surprisingly, out of the considered datasets, DICOM COVID-19 cases are available only in one online dataset, and there are only four datasets with DICOM images for other lung diseases. The motivations for explaining models are commendable. Nevertheless, in many works, interpretations of explanations and summaries are missing. The XAI method is not a conclusion in itself. The fact that the model provides correct explanations for a few images does not yet show that the model works properly. It would be good to quantitatively validate XAI methods. For this purpose, the help of clinicians or proper annotations prepared by radiologists beforehand are necessary. The paper mentions a long list of problems in modeling, but this analysis is not intended to criticize any of the mentioned articles, these are state-of-the-art papers often published in prestigious journals. However, the analysis of these articles makes one look critically at standards in AI for healthcare or rather the lack of them. We hope that this paper will initiate the process of development of standards for responsible AI solutions in healthcare. In this paper, we showed that the verification of the XAI solutions for medical images is not only important but it is a must. Following the guidelines proposed in the paper, we created an online GitHub repository8 which can be maintained by the community working on AI models for image analysis in healthcare. This repository is intended to be a starting point for further development of the proposed checklist to meet the evolving challenges in responsible modeling. We believe that if the proposed checklist is taken into account when building models, we will get better models.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  27 in total

1.  COVIDGR Dataset and COVID-SDNet Methodology for Predicting COVID-19 Based on Chest X-Ray Images.

Authors:  S Tabik; A Gomez-Rios; J L Martin-Rodriguez; I Sevillano-Garcia; M Rey-Area; D Charte; E Guirado; J L Suarez; J Luengo; M A Valero-Gonzalez; P Garcia-Villanova; E Olmedo-Sanchez; F Herrera
Journal:  IEEE J Biomed Health Inform       Date:  2020-12-04       Impact factor: 5.772

2.  Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network.

Authors:  Awni Y Hannun; Pranav Rajpurkar; Masoumeh Haghpanahi; Geoffrey H Tison; Codie Bourn; Mintu P Turakhia; Andrew Y Ng
Journal:  Nat Med       Date:  2019-01-07       Impact factor: 53.440

3.  Chest Imaging Appearance of COVID-19 Infection.

Authors:  Weifang Kong; Prachi P Agarwal
Journal:  Radiol Cardiothorac Imaging       Date:  2020-02-13

4.  Systematic review of artificial intelligence techniques in the detection and classification of COVID-19 medical images in terms of evaluation and benchmarking: Taxonomy analysis, challenges, future solutions and methodological aspects.

Authors:  O S Albahri; A A Zaidan; A S Albahri; B B Zaidan; Karrar Hameed Abdulkareem; Z T Al-Qaysi; A H Alamoodi; A M Aleesa; M A Chyad; R M Alesa; L C Kem; Muhammad Modi Lakulu; A B Ibrahim; Nazre Abdul Rashid
Journal:  J Infect Public Health       Date:  2020-07-01       Impact factor: 3.718

5.  CT Imaging Features of 2019 Novel Coronavirus (2019-nCoV).

Authors:  Michael Chung; Adam Bernheim; Xueyan Mei; Ning Zhang; Mingqian Huang; Xianjun Zeng; Jiufa Cui; Wenjian Xu; Yang Yang; Zahi A Fayad; Adam Jacobi; Kunwei Li; Shaolin Li; Hong Shan
Journal:  Radiology       Date:  2020-02-04       Impact factor: 11.105

6.  Automated detection of COVID-19 cases using deep neural networks with X-ray images.

Authors:  Tulin Ozturk; Muhammed Talo; Eylul Azra Yildirim; Ulas Baran Baloglu; Ozal Yildirim; U Rajendra Acharya
Journal:  Comput Biol Med       Date:  2020-04-28       Impact factor: 4.589

7.  Performance of Radiologists in Differentiating COVID-19 from Non-COVID-19 Viral Pneumonia at Chest CT.

Authors:  Harrison X Bai; Ben Hsieh; Zeng Xiong; Kasey Halsey; Ji Whae Choi; Thi My Linh Tran; Ian Pan; Lin-Bo Shi; Dong-Cui Wang; Ji Mei; Xiao-Long Jiang; Qiu-Hua Zeng; Thomas K Egglin; Ping-Feng Hu; Saurabh Agarwal; Fang-Fang Xie; Sha Li; Terrance Healey; Michael K Atalay; Wei-Hua Liao
Journal:  Radiology       Date:  2020-03-10       Impact factor: 11.105

8.  Frequency and Distribution of Chest Radiographic Findings in Patients Positive for COVID-19.

Authors:  Ho Yuen Frank Wong; Hiu Yin Sonia Lam; Ambrose Ho-Tung Fong; Siu Ting Leung; Thomas Wing-Yan Chin; Christine Shing Yen Lo; Macy Mei-Sze Lui; Jonan Chun Yin Lee; Keith Wan-Hang Chiu; Tom Wai-Hin Chung; Elaine Yuen Phin Lee; Eric Yuk Fai Wan; Ivan Fan Ngai Hung; Tina Poy Wing Lam; Michael D Kuo; Ming-Yen Ng
Journal:  Radiology       Date:  2020-03-27       Impact factor: 11.105

9.  Prediction models for diagnosis and prognosis of covid-19: systematic review and critical appraisal

Authors:  Laure Wynants; Ben Van Calster; Gary S Collins; Richard D Riley; Georg Heinze; Ewoud Schuit; Marc M J Bonten; Darren L Dahly; Johanna A A Damen; Thomas P A Debray; Valentijn M T de Jong; Maarten De Vos; Paul Dhiman; Maria C Haller; Michael O Harhay; Liesbet Henckaerts; Pauline Heus; Michael Kammer; Nina Kreuzberger; Anna Lohmann; Kim Luijken; Jie Ma; Glen P Martin; David J McLernon; Constanza L Andaur Navarro; Johannes B Reitsma; Jamie C Sergeant; Chunhu Shi; Nicole Skoetz; Luc J M Smits; Kym I E Snell; Matthew Sperrin; René Spijker; Ewout W Steyerberg; Toshihiko Takada; Ioanna Tzoulaki; Sander M J van Kuijk; Bas van Bussel; Iwan C C van der Horst; Florien S van Royen; Jan Y Verbakel; Christine Wallisch; Jack Wilkinson; Robert Wolff; Lotty Hooft; Karel G M Moons; Maarten van Smeden
Journal:  BMJ       Date:  2020-04-07
View more
  3 in total

1.  FCF: Feature complement fusion network for detecting COVID-19 through CT scan images.

Authors:  Shu Liang; Rencan Nie; Jinde Cao; Xue Wang; Gucheng Zhang
Journal:  Appl Soft Comput       Date:  2022-06-06       Impact factor: 8.263

2.  Pareto optimization of deep networks for COVID-19 diagnosis from chest X-rays.

Authors:  Valerio Guarrasi; Natascha Claudia D'Amico; Rosa Sicilia; Ermanno Cordelli; Paolo Soda
Journal:  Pattern Recognit       Date:  2021-08-09       Impact factor: 7.740

3.  Using deep-learning in fetal ultrasound analysis for diagnosis of cystic hygroma in the first trimester.

Authors:  Mark C Walker; Inbal Willner; Olivier X Miguel; Malia S Q Murphy; Darine El-Chaâr; Felipe Moretti; Alysha L J Dingwall Harvey; Ruth Rennicks White; Katherine A Muldoon; André M Carrington; Steven Hawken; Richard I Aviv
Journal:  PLoS One       Date:  2022-06-22       Impact factor: 3.752

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.