Literature DB >> 33686526

Developing a Training Web Application for Improving the COVID-19 Diagnostic Accuracy on Chest X-ray.

P Menéndez Fernández-Miranda1,2, P Sanz Bellón3,4, A Pérez Del Barrio3,4, L Lloret Iglesias5, P Solís García6, F Aguilar-Gómez5, D Rodríguez González5, J A Vega7,8.   

Abstract

In December 2019, a new coronavirus known as 2019-nCoV emerged in Wuhan, China. The virus has spread globally and the infection was declared pandemic in March 2020. Although most cases of coronavirus disease 2019 (COVID-19) are mild, some of them rapidly develop acute respiratory distress syndrome. In the clinical management, chest X-rays (CXR) are essential, but the evaluation of COVID-19 CXR could be a challenge. In this context, we developed COVID-19 TRAINING, a free Web application for training on the evaluation of COVID-19 CXR. The application included 196 CXR belonging to three categories: non-pathological, pathological compatible with COVID-19, and pathological non-compatible with COVID-19. On the training screen, images were shown to the users and they chose a diagnosis among those three possibilities. At any time, users could finish the training session and be evaluated through the estimation of their diagnostic accuracy values: sensitivity, specificity, predictive values, and global accuracy. Images were hand-labeled by four thoracic radiologists. Average values for sensitivity, specificity, and global accuracy were .72, .64, and .68. Users who achieved better sensitivity registered less specificity (p < .0001) and those with higher specificity decreased their sensitivity (p < .0001). Users who sent more answers achieved better accuracy (p = .0002). The application COVID-19 TRAINING provides a revolutionary tool to learn the necessary skills to evaluate COVID-19 on CXR. Diagnosis training applications could provide a new original manner of evaluation for medical professionals based on their diagnostic accuracy values, and an efficient method to collect valuable data for research purposes.
© 2021. Society for Imaging Informatics in Medicine.

Entities:  

Keywords:  COVID‐19; Chest X-ray; Diagnostic accuracy values; Medical application; Medical education; Training on diagnosis

Year:  2021        PMID: 33686526      PMCID: PMC7939450          DOI: 10.1007/s10278-021-00424-7

Source DB:  PubMed          Journal:  J Digit Imaging        ISSN: 0897-1889            Impact factor:   4.056


Background

In December 2019, a new coronavirus named 2019-nCoV, also known as severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), was isolated in the airway epithelial cells of a cluster of patients with pneumonia of unknown cause in Wuhan, China [1]. Since then, the infection caused by 2019-nCoV has rapidly spread globally [2] affecting more than 4 million people in about 215 countries with more than 280.000 reported deaths to date [3]. On the 11th of March 2020, the World Health Organization (WHO) declared the novel coronavirus outbreak a global pandemic [4]. Most patients with coronavirus disease 2019 (COVID-19) are mild cases, whose symptoms are usually self-limiting and recover within 2 weeks [5]. However, others progress rapidly and develop acute respiratory distress syndrome (ARDS) and septic shock, eventually resulting in multiple organ failure [6]. At the time of this writing, fever and cough were the main clinical manifestations, followed by dyspnea, myalgia or weakness, and chest tightness [7]. Furthermore, a variable percentage of patients report decreased smell function or even anosmia and dysgeusia [8]. Currently, COVID-19 is being diagnosed using molecular detection methods such as the reverse-transcriptase polymerase chain reaction (RT-PCR) test, regarded as the standard of reference [9-11]. Nevertheless, despite RT-PCR COVID-19 testing has a specificity of 100% [11], it shows a potentially high false negative rate deteriorating the sensitivity, and then it is not a definitive diagnostic method [10, 12]. Other techniques for the detection of COVID-19 are being used to increase the efficiency in the diagnosis, especially medical imaging modalities such as chest X-rays (CXR) and computed tomography (CT) scans [9, 10, 13]. In patients with a high level of clinical suspicion of COVID-19 and negative RT-PCR, CXR can be key to identify false negatives to RT-PCR COVID-19 testing as CXR abnormalities may appear before eventually testing positive on RT-PCR [13]. Moreover, in populations around the world with limited access to reliable real-time molecular diagnostic methods, the utilization of CXR for early disease detection also plays a capital role [14]. Additionally, imaging is also critical in assessing the severity and disease progression in a COVID-19 infection [15]. Thus, physicians evaluating COVID-19 on medical images should be aware of the imaging manifestations and radiological features that have been well-described [16]. However, the interpretation of chest radiographs is a challenging task, requiring experience and expertise [17-19]. The American College of Radiology (ACR) recommends that qualified radiologists be available to interpret all radiographs obtained in the Emergency Departments (ED) [20], and previous studies have reported suboptimal performance in the interpretation of CXR by ED physicians compared with expert radiologists [21-23]. Importantly, the number of chest radiographs per ED visit have increased during the last decades [24] which is a practical limitation with regard to the full-time availability of expert radiologists [25]. This situation has deteriorated since the pandemic outbreak appeared in the worldwide clinical scenario. In the context of COVID-19 pandemic and the high demand of CXR reporting, many CXR are being interpreted by nonexpert physicians who have been forced to acquire the competence of detecting and evaluating the radiological features of COVID-19. Being aware of it, we have developed a free Web application to ease the learning process of interpreting COVID-19 CXR for physicians, residents, students, and anybody else interested in acquiring this competence. As far as we know, this application called COVID-19 TRAINING (https://xray.covid.ifca.es/en) is the first available tool that allows a single user to calculate their diagnostic accuracy values: sensitivity, specificity, positive predictive value, negative predictive value, and global accuracy. The COVID-19 TRAINING application can set a precedent for further applications because this is the first time that a physician is considered as a single diagnostic tool by themselves. Usually, diagnostic accuracy values are calculated for diagnostic techniques or collectives of physicians but not for a single professional. In our opinion, providing this kind of tool to physicians can help them to evaluate performance by setting specific metrics, and to follow up their progression in their diagnostic efficiency.

Methods

Patients Recruitment

We collected 196 CXR in our Institution belonging to 33 females (35.48%) and 60 males (64.51%). The average age of these patients was 61.43 years with a standard deviation (SD) of 17.20 years and a range from 19 to 88 (Table 1).
Table 1

Characteristics of the patients included in the application and number of X-rays for each category

AttributeNon-pathologicalPathological compatible with COVID-19Pathological non compatible with COVID-19Total
Gender, n (%)
Female9 (42.86)10 (23.81)14 (46.67)33 (35.48)
Male12 (57.14)32 (76.19)16 (53.33)60 (64.51)
Age, mean ± SD61.04 ± 16.7464.64 ± 15.1654.07 ± 17.6961.43 ± 17.20
Range31–8430–8819–8219–88
CXR, n (%)45 (22.96)121 (61.73)30 (15.31)196 (100)

CXR chest X-ray

Characteristics of the patients included in the application and number of X-rays for each category CXR chest X-ray The chest radiographs were all hand-labeled and classified by four expert thoracic radiologists into 3 categories: non-pathological, pathological compatible with COVID-19, and pathological non-compatible with COVID-19. Classification and inclusion criteria for each category were as follows: (1) non-pathological: (a) the CXR was obtained before the appearance of SARS-CoV-2 virus; and (b) four reports made by the four expert thoracic radiologists confirm the absence of any pathological findings on the CXR; (2) pathological compatible with COVID-19: (a) the CXR has four reports made by the four expert thoracic radiologists indicating pathological findings compatible with COVID-19; (b) the patients have a subsequent confirmation of the disease by a positive RT-PCR COVID-19 testing; and (c) progression of the radiological findings reported previously were observed on further CXR; (3) pathological non-compatible with COVID-19: (a) the CXR was obtained before the appearance of SARS-CoV-2 virus; and (b) the CXR has four reports made by the four expert thoracic radiologists confirming pathological findings. The exclusion criteria for the three categories were as follows: (a) low-quality image, (b) improper alignment of the X-ray tube to the film, and (c) did not meet any of the other inclusion criteria. The chest radiographs were obtained with different equipment (portable and conventional machines), in different projections (posterior-anterior and anterior–posterior) and with different patient positions (standing and supine decubitus), in order to reproduce the most realistic clinical scenario. Data about patient distribution for each category is shown in Table 1. This distribution tries to guarantee that the application shows the user a proportional number of images from each category with a slight predominance of the category pathological compatible with COVID-19, which are the most interesting images for the users. In addition, the application also resembles the real clinical context where females are less frequently infected than males [26].

Data Access and Anonymization

This work was approved by the Ethical Committee of our Institution, and all the data collected to develop the application were fully de-identified before transfer to the development team.

Application Programming and Launching

Subject information and links to the corresponding images are stored in a SQLite database. The application has been implemented using Python 3 [27] for the backend and Flask, a Python framework, for generating the client side or frontend. Some external libraries were used in order to expand the possibilities of the frontend, achieving multilanguage (currently supporting Spanish and English) or more structured styles hierarchy. Some other good practices based on the Clean Code principles are being applied. All the code is publicly available in an open access repository [28]. User management is supported by an Authentication and Authorization Infrastructure based on Open ID Connect standard. A public beta version was published on the 11th of April 2020, and the 24th of April we launched the final version.

Application Usability and Requirements

The COVID-19 TRAINING application was designed to help all doctors, residents, medical students, and anybody else interested in acquiring the skills to recognize the radiological findings of COVID-19 on CXR. It includes CXR with different levels of difficulty so everyone will be able to join. The application is available for free and adapted both for mobile and tablet (Fig. 1) as well as for computer screens (Fig. 2) in Spanish and English languages.
Fig. 1

Welcome page screen, as viewed from a mobile device

Fig. 2

Welcome page screen, as viewed from a desktop device

Welcome page screen, as viewed from a mobile device Welcome page screen, as viewed from a desktop device

Access to the Application

The access and navigation on the application is shown on a video added in the Appendix 1. First step: The first screen that the user finds when opening the application is the welcome page (Figs. 1 and 2). It serves as the starting point of the application and shows a headline and a paragraph describing briefly its functionalities. On this interface two buttons are displayed: “COVID-19 TRAINING” that provides easy access to the login screen, and “Radiological features of COVID-19 on X-Ray,” which opens a summary of the radiological COVID-19 findings that the user should be able to identify on CXR before starting the training. Second Step: Login screen requires the user to accept the terms and conditions and to press the “Login” button to access the authentication interface (Fig. 3a). Authentication can be easily done either just entering a Google Gmail or creating a new account by pressing on the “Register” button in the lowest part of the screen (Fig. 3b). If the user’s preferred option is the second, they will only need to provide their name, surname, email address, and password. An email will be automatically sent with a link to verify the account.
Fig. 3

a Login and b authentication screens, and c, d drop-down menu to select the specialty or professional category of the user, as viewed from a desktop device

a Login and b authentication screens, and c, d drop-down menu to select the specialty or professional category of the user, as viewed from a desktop device Third Step: Finally, a drop-down menu will allow the user to select their specialty or professional category (Fig. 3c, d). Available options are shown on Table 2.
Table 2

Specialties and professional categories available to select on the drop-down menu

Specialty/category
Abdominal radiologist
Neuroradiologist
Breast radiologist
Musculoskeletal radiologist
General radiologist
Interventional radiologist
Pediatric radiologist
Thoracic radiologist
Emergency radiologists
Radiology resident
Resident of other specialty
Medical student
Pulmonology physician
Internal medicine physician
Emergency physician
Intensive care physician
Other kind of physician
Diagnostic Imaging Technician (radiographer)
Other category
Specialties and professional categories available to select on the drop-down menu

Training and Obtaining the Diagnostic Accuracy Values

Once the authentication is done and the profile category is selected, a new screen with CXR will be open. Besides the CXR, patient’s age and gender are also provided (Fig. 4a).
Fig. 4

a Training interface and b the diagnoses available on the drop-down menu, as viewed from a desktop device

a Training interface and b the diagnoses available on the drop-down menu, as viewed from a desktop device On the left side of the screen, the user can choose in a drop-down menu one diagnosis among the three possible categories described before: non-pathological, pathological compatible with COVID-19, and pathological non-compatible with COVID-19 (Fig. 4b). After selecting the desired option, by pressing the “Next” button, the answer will be registered and a new case will be charged. At any time, the user can finish the training session by pressing the “End Test” button (Fig. 4). Each time the user takes a test, a new set of training images is selected. A virtual magnifying glass is also implemented. Users simply need to press on the image to summon the magnifying glass, and they will see a zoomed image within its radius, without disturbing the rest of the page (Fig. 5). To remove it, they only need to press on the image again.
Fig. 5

Virtual magnifying glass, as viewed from a desktop device

Virtual magnifying glass, as viewed from a desktop device The “End Test” button links to the results interface, where the user’s sensitivity, specificity, predictive values and global accuracy are provided after each training (Figs. 6 and 7). If the users scroll down on this interface, those questions that they answered incorrectly will be displayed, indicating the given response and the correct answer, so they will learn from their mistakes. Pathological non compatible with COVID-19 images also include a reference to the real pathology of the patient.
Fig. 6

Screen of results, as viewed from a desktop device

Fig. 7

Screen of results, as viewed from a mobile device

Screen of results, as viewed from a desktop device Screen of results, as viewed from a mobile device

Data Collection and Analysis of the User’s Results

The COVID-19 TRAINING application allows the user to evaluate themselves through the estimation of their diagnostic accuracy values: sensitivity, specificity, positive predictive value, negative predictive value, and global accuracy. Diagnostic accuracy values are calculated for COVID-19 diagnosis. The application registers the answers of the users and classifies them into the following four conventional categories [29, 30]: (1) true positive (TP): the user classifies properly a CXR which belongs to the category of pathological compatible with COVID-19; (2) false positive (FP): the user classifies improperly a CXR which does not belong to the category of pathological compatible with COVID-19, in this category; (3) true negative (TN): the user classifies properly a CXR which does not belong to the category of pathological compatible with COVID-19 in another category; (4) false negative (FN): the user classifies improperly a CXR which belongs to the category of pathological compatible with COVID-19, in another category. The sensitivity represents the user’s ability to determine the COVID-19 cases correctly. It accounts for the proportion of true positives in patient-cases: TP/(TP + FN) [29]. The specificity shows the user’s capacity to rule out COVID-19 correctly. To estimate specificity, the proportion of true negatives in healthy cases should be calculated: TN/(TN + FP) [29]. Positive predictive value (PPV) defines the probability of having COVID-19 when the user classifies the CXR into the category pathological compatible with COVID-19. Therefore, it represents the proportion of COVID-19 patients within the patients with positive CXR for COVID-19 according to the user’s criteria: TP/(TP + FP) [30]. By contrast, negative predictive value (NPV) describes the probability of not having COVID-19 when the user does not classify the CXR into pathological compatible with COVID-19 category. It is defined as the proportion of subjects without COVID-19 within the patients with a negative CXR according to the user’s criteria: TN/(TN + FN) [30]. Finally, global accuracy depicts the ability of the user to differentiate COVID-19 patients and non-COVID-19 patients. It is the proportion of true positives and true negatives in all evaluated cases: (TP + TN)/(TP + TN + FP + FN) [29]. The information gathered by the application includes TP, TN, FP, FN, sensitivity, specificity, PPV, NPV, and global accuracy for each user and each time they take the test. The specialty or professional category selected after the authentication and the answers given to each CXR are also stored. Nevertheless, the specialty or category chosen before starting the training have only been recorded since the launch of the last version on the 24th of April. Beta test did not record it. Engagement with the application is tracked using a customized data capture system and validated with Google Analytics [31]. This program is an effective resource for measuring the diffusion and understanding the geodemographics of users [32]. Statistical analysis was performed with IBM SPSS Statistics [33]. Averages for diagnostic accuracy values were computed following two different approaches: firstly, population diagnostic values were calculated from the totals of TP, TN, FP, and FN collected; differently, the means for the diagnostic values obtained by the users were also estimated. The test used to assess differences in the performances between two groups was chi-squared test. To indicate a statistically significant difference, a p < 0.05 was considered.

Results

After the beta launch on the 11th of April, the application had 431 users within the first 3 days and 704 within the first week, with a total number of answers of 23,130. This version had users in more than 20 countries, according to Google Analytics reports. Figure 8 shows a map with users’ locations.
Fig. 8

Countries of the application users. The first version of the application extended to more than 20 countries as the map shows in blue color: a darker color means more visits

Countries of the application users. The first version of the application extended to more than 20 countries as the map shows in blue color: a darker color means more visits The application registered 16,360 (70.73%) correct answers (TP + TN) and 6770 (29.27%) incorrect answers (FP + FN). Consequently, the odds of answering correctly was 2.42. Positive likelihood ratio was 2.26, 95% CI [2.20, 2.33], and negative likelihood ratio was 0.39, 95% CI [0.38, 0.40]. Each user sent a mean of 32.86 (SD 39.21) answers, ranging from 1 to 434 (Figs. 9 and 10): an average of 23.24 (SD 28.09) of them were correct, and an average of 9.62 (SD 12.66) were incorrect.
Fig. 9

A frequency distribution graph showing the number of answers sent by each user. The x-axis represents the users and the y-axis represents the number of responses

Fig. 10

An histogram depicting the number of answers collected by the first version of the application. The x-axis shows intervals of 16 responses width and the y-axis shows the frequency of the intervals

A frequency distribution graph showing the number of answers sent by each user. The x-axis represents the users and the y-axis represents the number of responses An histogram depicting the number of answers collected by the first version of the application. The x-axis shows intervals of 16 responses width and the y-axis shows the frequency of the intervals Overall, this first version collected the following answers: 9026 (39.02%) TP; 7334 (31.71%) TN; 3545 (15.33%); FP; and 3225 (13.94%) FN. The mean of TP, TN, FP, and FN per user were 12.82 (SD 15.91), 10.42 (SD 12.88), 4.58 (SD 6.31), and 5.04 (SD 7.65), respectively (Table 3).
Table 3

Basic statistics for the user’s answers

Answersn (%)Mean ± SD (per user)Range (per user)
TP9026 (39.02)12.82 ± 15.910–136
FN3225 (13.94)4.58 ± 6.310–91
TN7334 (31.71)10.42 ± 12.880–88
FP3545 (15.33)5.04 ± 7.650–123
Correct responses16,360 (70.73)23.24 ± 28.090–220
Incorrect responses6770 (29.27)9.62 ± 12.660–214
Total responses23,130 (100)32.86 ± 39.211–434

TP true positive, FN false negative, TN true negative, FP false positive

Basic statistics for the user’s answers TP true positive, FN false negative, TN true negative, FP false positive The global population diagnostic accuracy values registered were the following: sensitivity 0.74, 95% CI [0.73, 0.74]; specificity 0.67, 95% CI [0.67, 0.68]; PPV 0.72, 95% CI [0.71, 0.72]: NPV 0.69, 95% CI [0.69, 0.70]; and accuracy 0.71, 95% CI [0.70, 0.71] (Table 4). Differently, the means for the diagnostic values obtained by the users were as follows: sensitivity 0.72, 95% CI [0.71, 0.74]; specificity 0.64, 95% CI [0.62, 0.66]; PPV 0.70, 95% CI [0.69, 0.72]; NPV 0.68, 95% CI [0.66, 0.69]; and accuracy 0.68, 95% CI [0.67, 0.69] (Table 5).
Table 4

Population diagnostic accuracy values

StatisticValue95% CI
Sensitivity.74[.73, .74]
Specificity.67[.67, .68]
Positive predictive value.72[.71, .72]
Negative predictive value.69[.69, .70]
Global accuracy.71[.70, .71]

Values computed from the total answers collected

CI confidence interval

Table 5

Average of user’s diagnostic accuracy values

StatisticMean ± SD95% CI of the mean
Sensitivity.72 ± .21[.71, .74]
Specificity.64 ± .25[.62, .66]
Positive predictive value.70 ± .21[.69, .72]
Negative predictive value.68 ± .22[.66, .69]
Global accuracy.68 ± .16[.67, .69]

Values estimated from the diagnostic values achieved by each user

CI confidence interval

Population diagnostic accuracy values Values computed from the total answers collected CI confidence interval Average of user’s diagnostic accuracy values Values estimated from the diagnostic values achieved by each user CI confidence interval The hardest CXR obtained an odds of answering correctly of 0.32 (Table 6). On the other side, the easiest CXR odds of answering correctly was 18.70 (Table 7). Both CXR belonged to the category pathological compatible with COVID-19 (Figs. 11a and 12).
Table 6

The top 10 hardest chest X-rays

PositionImageTPFNTNFPOdds of answering correctly
1F064714900.32
2F020_20048117.41
3M0535312800.41
4M0340057130.44
5M060056125.45
6M030_25211600.45
7M096612100.55
8F096611800.56
9M032_2006298.63
10M0317010400.67

TP true positive, FN false negative, TN true negative, FP false positive

Table 7

The top 10 easiest chest X-rays

PositionImageTPFNTNFPOdds of answering correctly
1M027187100018.70
2F024001851018.50
3M027_2182100018.20
4F022001631016.30
5F014175110015.91
6M019158110014.30
7M018181130013.92
8F06_3173130013.31
9M02_2151120012.58
10M016163140011.64

TP true positive, FN false negative, TN true negative, FP false positive

Fig. 11

a The most difficult chest X-ray. It is a 66-year-old female with a chest X-ray belonging to the category pathological compatible with COVID-19. On this image, subtle ground-glass opacities on the right lung are visible (arrow): the most evident opacity is in the right upper lobe (circle). b On a subsequent chest X-ray performed to the same patient after three days, this ground-glass opacity in the upper lobe became even clearer (circle). In addition, new opacities appeared in the lower lobes on this chest X-ray (arrows)

Fig. 12

The easiest chest X-ray. It is a 69-year-old male diagnosed of COVID-19. The image shows extensive, multiple and bilateral opacities (arrows) indicating a severe form of COVID-19 pneumonia

The top 10 hardest chest X-rays TP true positive, FN false negative, TN true negative, FP false positive The top 10 easiest chest X-rays TP true positive, FN false negative, TN true negative, FP false positive a The most difficult chest X-ray. It is a 66-year-old female with a chest X-ray belonging to the category pathological compatible with COVID-19. On this image, subtle ground-glass opacities on the right lung are visible (arrow): the most evident opacity is in the right upper lobe (circle). b On a subsequent chest X-ray performed to the same patient after three days, this ground-glass opacity in the upper lobe became even clearer (circle). In addition, new opacities appeared in the lower lobes on this chest X-ray (arrows) The easiest chest X-ray. It is a 69-year-old male diagnosed of COVID-19. The image shows extensive, multiple and bilateral opacities (arrows) indicating a severe form of COVID-19 pneumonia As would be expected, users who achieved a sensitivity equal or higher to the average sensitivity (0.72), also registered less specificity (p < 0.0001). Similarly, the user’s with higher specificity, those who achieved the average specificity (0.64) or more, decrease their sensitivity (p < 0.0001). In addition, the users who sent more answers than the 50th percentile (68 answers) achieved better final average global accuracy (p < 0.0001); excluding users who sent less than 10 answers from the analysis, results were also statistically significant (p = 0.0002) (Table 8).
Table 8

Comparison of the results between groups and p values for chi-squared tests

StatisticGroup 1Group 2Proportion difference95% CI of differenceχ2 valuep value
Sensitivity [TP/(TP + FN)]a.76 [3504/4604].72 [5479/7591]− .04[− .06, − .02]22.814< .0001
Specificity [TN/(TN + FP)]b.72 [3192/4425].64 [4130/6431]− .08[− .10, − .06]74.795< .0001
Global accuracy [(TP + TN)/(TP + TN + FP + FN)]c.70 [10430/14956].73 [5930/8174].03[.02, .04]20.147< .0001
Global accuracy (TP + TN)/(TP + TN + FP + FN)d.70 [9815/13988].73 [5930/8174].02[.01, .04]14.203.0002

CI confidence interval, TP true positive, FN false negative, TN true negative, FP false positive

aGroup 1: users who achieved a specificity lesser than the mean (.64); group 2: users who achieved a specificity equal to or higher than the mean (.64)

bGroup 1: users who registered a lower sensitivity than the mean (.72); group 2: users who registered a sensitivity equal to or higher than the mean (.72)

cGroup 1: users who sent a number of answers equal to the 50th percentile (68 answers) or less; group 2: users who sent more than 68 answers

dGroup 1: users who sent a number of answers between 10 and the 50th percentile (68 answers); group 2: users who sent more than 68 answers

Comparison of the results between groups and p values for chi-squared tests CI confidence interval, TP true positive, FN false negative, TN true negative, FP false positive aGroup 1: users who achieved a specificity lesser than the mean (.64); group 2: users who achieved a specificity equal to or higher than the mean (.64) bGroup 1: users who registered a lower sensitivity than the mean (.72); group 2: users who registered a sensitivity equal to or higher than the mean (.72) cGroup 1: users who sent a number of answers equal to the 50th percentile (68 answers) or less; group 2: users who sent more than 68 answers dGroup 1: users who sent a number of answers between 10 and the 50th percentile (68 answers); group 2: users who sent more than 68 answers Beta version did not ask for the user’s category or specialty, so no data about it was recorded before the final version that was launched on the 24th of April.

Discussion

Several medical educational applications have been developed [33-35], but only a few of them are focused on the diagnosis training, and even less on COVID-19 diagnosis. Currently, the society is facing a health situation which has no precedents [36]. In this context, the role of the medical community is vital [37, 38] and the medical education of those who are taking part of the solution is essential [18, 39]. The application COVID-19 TRAINING was developed to help professionals to acquire the required competencies to diagnose COVID-19 on CXR. This application brings a new manner to ease the learning of the necessary skills to successfully evaluate COVID-19 on CXR. According to us, this is a different educational technique that could set a precedent for developing further applications on the training on the diagnosis of other pathologies. Furthermore, we have introduced an original evaluation method based on the estimation of the users’ diagnostic accuracy values for COVID-19. In our opinion, this mechanism of assessment could prove useful for medical professionals since it provides them with information about their sensitivity, specificity, predictive values, and global accuracy. Armed with that knowledge, they will recognize and improve their weaknesses. The results obtained by the first version of the application showed that those who achieved better sensitivity also decrease their specificity. Similarly, those who registered better specificity also had lower sensitivity. These results evidence that usually, increasing sensitivity entails decreasing specificity, and vice versa. Our application can help the users to find an optimal balance between their sensitivity and specificity and consequently, to achieve their best possible global accuracy. In addition, the analysis and display of the failed answers is also a key feature of this tool. At the end of the test, users can review their incorrect responses so they will become aware about their mistakes and they will recognize their weak points. This feature can be also useful in research to identify the cases missed by numerous users. Later on, the examination of these problematic images may give some clues to find solutions to improvable areas in the medical practice not detected before. Following this approach, we have analyzed the hardest and the easiest CXR trying to understand why users found difficult or obvious the respective cases. The most difficult CXR was a COVID-19 case that had subtle and ill-defined ground-glass opacities in the right lung. The most evident findings on this image were in the right upper lobe, nearly overlapped with the first costochondral junction (Fig. 11). This first costochondral junction is a well-recognized pitfall on CXR and it sometimes mimics a rounded opacity [40]. Perhaps, in the situations where a COVID-19 is suspected and a prominence on either the first costochondral or first costosternal or sternoclavicular junctions is seen on CXR, further evaluation should be performed. On the other side, the easiest CXR was also a COVID-19 patient, but in this case, the image illustrated evident peripheral bilateral opacities, which are one of the most typical radiological finding of this disease [7, 13] (Fig. 12). Overall, this application seems to be helpful. From the beginning, it had a dizzyingly fast diffusion with hundreds of users registered from more than 20 different countries within the first days and more than 20.000 answers sent. Additionally, many medical societies spread and shared this first beta version on their webpages. Results showed that the users who sent more answers achieved better global accuracy. In addition, only 6 users (0.9%) exceeded the number of cases contained in our dataset, so most of them did not repeat answers. Users who sent less than 10 answers were excluded from the analysis to avoid potential biases. These facts could indicate that the application improves the users’ diagnostic skills. Finally, we also found that medical educational applications may be used in research as a new method to collect relevant information. Since the application was officially launched on the 24th of April, data about the users’ specialty or professional category have been recorded. Our intention is to use this data to analyze the difference in the diagnostic values between the users belonging to different specialties or categories and to try to estimate the real utility of the CXR on the evaluation of COVID-19.

Conclusion

In present COVID-19 pandemic, the medical education of the professionals involved in patients care is vital and COVID-19 TRAINING brings a different solution to help them in this purpose. Applications focused on the training on diagnosis could provide a new original manner of evaluation for medical professionals. The assessment of users by estimating their diagnostic accuracy values make them aware of their weak points. In addition, this kind of application also collects valuable information that can be used for research purposes.
  33 in total

1.  Survey of after-hours coverage of emergency department imaging studies by US academic radiology departments.

Authors:  Andrew Sellers; Bruce J Hillman; Max Wintermark
Journal:  J Am Coll Radiol       Date:  2014-04-06       Impact factor: 5.532

2.  Supporting the Health Care Workforce During the COVID-19 Global Epidemic.

Authors:  James G Adams; Ron M Walls
Journal:  JAMA       Date:  2020-04-21       Impact factor: 56.272

3.  App Review: The Radiology Assistant 2.0.

Authors:  Lawrence E Wood; Melissa M Picard; Mark D Kovacs
Journal:  J Digit Imaging       Date:  2018-08       Impact factor: 4.056

4.  Olfactory and gustatory dysfunctions as a clinical presentation of mild-to-moderate forms of the coronavirus disease (COVID-19): a multicenter European study.

Authors:  Jerome R Lechien; Carlos M Chiesa-Estomba; Daniele R De Siati; Mihaela Horoi; Serge D Le Bon; Alexandra Rodriguez; Didier Dequanter; Serge Blecic; Fahd El Afia; Lea Distinguin; Younes Chekkoury-Idrissi; Stéphane Hans; Irene Lopez Delgado; Christian Calvo-Henriquez; Philippe Lavigne; Chiara Falanga; Maria Rosaria Barillari; Giovanni Cammaroto; Mohamad Khalife; Pierre Leich; Christel Souchay; Camelia Rossi; Fabrice Journe; Julien Hsieh; Myriam Edjlali; Robert Carlier; Laurence Ris; Andrea Lovato; Cosimo De Filippis; Frederique Coppee; Nicolas Fakhry; Tareck Ayad; Sven Saussez
Journal:  Eur Arch Otorhinolaryngol       Date:  2020-04-06       Impact factor: 2.503

5.  Frequency and Distribution of Chest Radiographic Findings in Patients Positive for COVID-19.

Authors:  Ho Yuen Frank Wong; Hiu Yin Sonia Lam; Ambrose Ho-Tung Fong; Siu Ting Leung; Thomas Wing-Yan Chin; Christine Shing Yen Lo; Macy Mei-Sze Lui; Jonan Chun Yin Lee; Keith Wan-Hang Chiu; Tom Wai-Hin Chung; Elaine Yuen Phin Lee; Eric Yuk Fai Wan; Ivan Fan Ngai Hung; Tina Poy Wing Lam; Michael D Kuo; Ming-Yen Ng
Journal:  Radiology       Date:  2020-03-27       Impact factor: 11.105

Review 6.  Coronavirus (COVID-19) Outbreak: What the Department of Radiology Should Know.

Authors:  Soheil Kooraki; Melina Hosseiny; Lee Myers; Ali Gholamrezanezhad
Journal:  J Am Coll Radiol       Date:  2020-02-19       Impact factor: 5.532

7.  Stability issues of RT-PCR testing of SARS-CoV-2 for hospitalized patients clinically diagnosed with COVID-19.

Authors:  Yafang Li; Lin Yao; Jiawei Li; Lei Chen; Yiyan Song; Zhifang Cai; Chunhua Yang
Journal:  J Med Virol       Date:  2020-04-05       Impact factor: 2.327

8.  Diagnostic performance between CT and initial real-time RT-PCR for clinically suspected 2019 coronavirus disease (COVID-19) patients outside Wuhan, China.

Authors:  Jian-Long He; Lin Luo; Zhen-Dong Luo; Jian-Xun Lyu; Ming-Yen Ng; Xin-Ping Shen; Zhibo Wen
Journal:  Respir Med       Date:  2020-04-21       Impact factor: 4.582

9.  Coronavirus infections: Epidemiological, clinical and immunological features and hypotheses.

Authors:  Didier Raoult; Alimuddin Zumla; Franco Locatelli; Giuseppe Ippolito; Guido Kroemer
Journal:  Cell Stress       Date:  2020-03-02

10.  WHO Declares COVID-19 a Pandemic.

Authors:  Domenico Cucinotta; Maurizio Vanelli
Journal:  Acta Biomed       Date:  2020-03-19
View more
  2 in total

1.  The Imaging Informatics Response to a Pandemic.

Authors:  Ross W Filice
Journal:  J Digit Imaging       Date:  2021-04       Impact factor: 4.903

2.  A Novel Weighted Consensus Machine Learning Model for COVID-19 Infection Classification Using CT Scan Images.

Authors:  Rohit Kumar Bondugula; Siba K Udgata; Nitin Sai Bommi
Journal:  Arab J Sci Eng       Date:  2021-08-02       Impact factor: 2.807

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.