Literature DB >> 33249683

Derivation With Internal Validation of a Multivariable Predictive Model to Predict COVID-19 Test Results in Emergency Department Patients.

Samuel A McDonald1,2, Richard J Medford2,3, Mujeeb A Basit2,4, Deborah B Diercks1, D Mark Courtney1.   

Abstract

OBJECTIVES: The COVID-19 pandemic has placed acute care providers in demanding situations in predicting disease given the clinical variability, desire to cohort patients, and high variance in testing availability. An approach to stratifying patients by likelihood of disease based on rapidly available emergency department (ED) clinical data would offer significant operational and clinical value. The purpose of this study was to develop and internally validate a predictive model to aid in the discrimination of patients undergoing investigation for COVID-19.
METHODS: All patients greater than 18 years presenting to a single academic ED who were tested for COVID-19 during this index ED evaluation were included. Outcome was defined as the result of COVID-19 polymerase chain reaction (PCR) testing during the index visit or any positive result within the following 7 days. Variables included chest radiograph interpretation, disease-specific screening questions, and laboratory data. Three models were developed with a split-sample approach to predict outcome of the PCR test utilizing logistic regression, random forest, and gradient-boosted decision tree methods. Model discrimination was evaluated comparing area under the receiver operator curve (AUC) and point statistics at a predefined threshold.
RESULTS: A total of 1,026 patients were included in the study collected between March and April 2020. Overall, there was disease prevalence of 9.6% in the population under study during this time frame. The logistic regression model was found to have an AUC of 0.89 (95% confidence interval [CI] = 0.84 to 0.94) when including four features: exposure history, temperature, white blood cell count (WBC), and chest radiograph result. Random forest method resulted in AUC of 0.86 (95% CI = 0.79 to 0.92) and gradient boosting had an AUC of 0.85 (95% CI = 0.79 to 0.91). With a consistently held negative predictive value, the logistic regression model had a positive predictive value of 0.29 (0.2-0.39) compared to 0.2 (0.14-0.28) for random forest and 0.22 (0.15-0.3) for the gradient-boosted method.
CONCLUSION: The derived predictive models offer good discriminating capacity for COVID-19 disease and provide interpretable and usable methods for those providers caring for these patients at the important crossroads of the community and the health system. We found utilization of the logistic regression model utilizing exposure history, temperature, WBC, and chest X-ray result had the greatest discriminatory capacity with the most interpretable model. Integrating a predictive model-based approach to COVID-19 testing decisions and patient care pathways and locations could add efficiency and accuracy to decrease uncertainty.
© 2020 by the Society for Academic Emergency Medicine.

Entities:  

Mesh:

Year:  2020        PMID: 33249683      PMCID: PMC7753649          DOI: 10.1111/acem.14182

Source DB:  PubMed          Journal:  Acad Emerg Med        ISSN: 1069-6563            Impact factor:   5.221


The narrative of COVID‐19 for the acute care provider in the United States has been dominated by testing and the question of likelihood of disease for individual patients. In a matter of months providers have gone from relying on reported classic symptoms based on anecdotal experience to developing their own gestalt based on their own experience to having SARS‐CoV‐2 tests available. Despite very rapid development of viral RNA polymerase chain reaction (PCR) testing in many locations, the question of pretest probability (the probability of disease prior to a test) is still critical for the following reasons: testing throughout the United States is still not fully developed or immediately available, risk of disease may be significantly overestimated resulting in clinician fear and impact on personal protective equipment (PPE) use, expansion of COVID‐19 to at‐risk countries and settings with no testing, , and its role in establishing a posttest probability after result of the test. This latter point has become critical for emergency clinicians, hospital administrators, and epidemiologists trying to limit both community‐ and health care–associated spread and follow disease trajectory. The posttest probability is the probability of disease after a test and is a function of two things: 1) the pretest probability and 2) the diagnostic performance and result of the test. Despite varying tests being used and the uncertainty of their diagnostic performance, the posttest probability is driven to a very large part by pretest probability. Efforts to more scientifically quantify pretest probability would aid individual and health system decision making. , The greatest concern to limiting spread both within the community and importantly within health systems and care facilities is the reported nontrivial rate of false‐negative testing. In the context of high‐risk patients, many are recommending repeat testing, recognizing that quality of nasopharyngeal swab technique, early‐stage illness, and handling of media may contribute to false‐negative tests. , For these reasons, many systems are adopting efforts to risk stratify patients based on gestalt, early internal data, or home‐grown scoring systems. These approaches attempt to meet a significant operational need in the emergency department (ED) in particular as the decision of disposition and early management of suspected COVID‐19 patients typically falls to the emergency physician. However, minimal work has been done to provide a data‐driven method for COVID‐19 risk stratification. Once developed, such a predictive approach would further our ability to provide appropriate recommendations for mitigating disease spread in and out of the hospital, identify patients who warrant repeat testing if the initial test is negative, and inform future research. The purpose of this study was to develop and internally validate a predictive model that could aid in the discrimination and management of patients undergoing investigation for COVID‐19 when presenting to the ED. Specifically, our objective was to use a robust electronic health record (EHR) that captures laboratory, clinical, vital sign, comorbidity, and radiographic data with traditional and novel modeling techniques to predict SARS‐CoV‐2 PCR test results in ED patients.

METHODS

This is a retrospective cohort study of consecutive patients presenting to the ED who were tested for COVID‐19 between March 6 and April 24, 2020, at a single academic quaternary care facility with a typical yearly volume of 50,000 patients. Given the pandemic environment and reduced ED presentations, the volume experienced during the study period was approximately 4,000 unique patient encounters. This quality improvement initiative was prioritized in the early evolution of our response to COVID‐19 to best evaluate how to deploy testing strategies in a resource supply conscious environment and was formally approved by the Department of Emergency Medicine Quality Improvement Committee. Reporting of this derivation has followed guidelines established by TRIPOD statement. We included all patients greater than 18 years presenting to our ED who were tested for COVID‐19 during this index ED visit and had not previously received testing in our health system. Testing decisions during the study period were protocolized by the health system based on a CDC guideline driven testing strategy requiring either a high‐risk exposure or a high‐prevalence location in addition to associated symptoms. Testing was also performed if there was high suspicion from the practicing clinician that fell outside of the protocolized guidelines. No asymptomatic testing was being performed during this period. Patients with repeat visits after the index visit were not included in the analysis because the available prior encounter data within the EHR could differentially affect pretest probability estimation as well as workup of disease relative to an encounter without previous ED testing. The primary outcome of interest was the presence of COVID‐19 infection evaluated by the performance of PCR testing via nasopharyngeal swab obtained at the index or any resulting positive test within the following week. Positive PCR testing was defined by any positive PCR test performed by laboratory technicians without awareness of other clinical data. If initial testing was negative but repeat testing within 1 week was positive, a positive result was attributed as the final outcome. This was done to account for the concern for possible initial false‐negative testing in setting of early presenters. If no additional testing was performed the initial result was considered the final outcome. Initially 44 variable candidates were collected from the clinical data warehouse that are routinely available during an ED encounter when evaluating influenza‐like illnesses. The data set consisted of patient specific information, triage screening questions, initial vital signs, specific laboratory results, and chest x‐ray interpretation. Patient‐specific information included age at encounter; sex, race; and specific comorbid conditions consisting of hypertension, diabetes, end‐stage renal disease, asthma, and COPD. The comorbidities are extracted from SNOMED concept‐based patient registries, defined by a standardized vocabulary of clinical concepts, that are automatically populated within the EHR anytime a clinician adds a diagnosis or problem list item that maps to the specific SNOMED concept. Given the number of patients available for our analysis limits our capacity to use unique comorbid conditions in a predictive model, we quantified the number of comorbid conditions per patient as a unique ordinal variable from 0 to 5. The screening questions performed at triage are a set of standardized general risk stratification questions for developing pathogens with increased specificity to COVID‐19. These questions assessed presence of recent fever, travel to location of high disease prevalence, exposure to someone with COVID‐19, and associated viral symptoms and are treated as dichotomous variables. Continuous variables collected for evaluation consisted of laboratory values and vital signs. Labs selected for evaluation were total white blood cell count (WBC), absolute lymphocyte count and percentage, alanine aminotransferase, aspartate aminotransferase, C‐reactive protein, ferritin, lactate dehydrogenase, high‐sensitivity troponin T (hs‐TnT), and NT‐pro‐BNP. These were selected based on emerging literature reporting value in disease specificity and severity. , Vitals obtained for the visit were collected as additional covariates consisting of blood pressure, pulse, respiratory rate, oxygen saturation, and temperature. To have a uniform collection method the first vitals obtained during the ED visit were used for analysis. Values of each of these covariates consisted of the first value obtained in the encounter for consistency and availability. Finally chest X‐ray (CXR) interpretation was included as a categorical variable consisting of options of consistent with viral pneumonia, low likelihood findings related to viral pneumonia, and negative. This was stored as a structured result within the EHR, classified by the interpreting radiologist. This was a new process implemented specifically for COVID‐19. There was a subset of images that were classified by the study group prior to implementation of that process. These were tagged in a similar fashion by the interpreting radiologist. Listwise elimination was performed for any observation without an associated COVID‐19 order result. Listwise elimination was also performed if no vital sign data or screening questionnaire data were available. This was done given the concern that these data were missing not at random as these points should be present in almost all ED visits. Variables with very low variance, more than 70% missing values, or high correlation with other features were removed prior to modeling. After initial variable selection and division of data, imputation was performed separately on the testing and validation data sets utilizing a random forest‐based nonparametric method that can impute both numeric and categorical variables using R package missRanger. , It is likely that many encounters had missing data for CXR or laboratory values given they were not ill‐appearing and unlikely to be admitted. Clinically, we did not mandate laboratory or CXR in all patients tested for COVID and the resulting data are likely what is to be found in real‐world setting. Data were randomly divided with a split‐sample approach allocating 75% of data for training each model and 25% for internal validation. Utilizing the training data set, three distinct models were developed to predict COVID‐19 test results to maximize our ability to arrive at an optimal method for prediction. A logistic regression model was developed using a backward stepwise variable selection method utilizing Wald chi‐square statistics to evaluate the reduced model. This was performed using R RMS package. Two ensemble decision tree–based approaches were used, random forest using R Ranger package and a gradient‐boosted decision tree method using R XGBoost package utilizing the default recommendations for tuning parameters. These models use the composite prediction of a group of base models to optimize bias and variance of the prediction, improving the prediction of a standard decision tree–based methodology. While there are many options for model types to use for prediction, our goal was to have a model that could be easily implemented within our EHR to provide automated alerting to providers. At this time, our capacity to implement models more complex than generalized linear models or tree‐based models is limited, and we felt that the operational value for testing other model types was low. We present central tendency with mean and associated 95% confidence intervals (CIs). Categorical data are presented as total number and percentages. Model performance is evaluated by examining the area under the receiver operator curve (AUC) for each model with associated 95% CIs in the validation data. To examine the specific threshold that would be used for clinical decision making, we compared the positive predictive values (PPV) at a set negative predictive value of 99%. This cut point was chosen to preferentially minimize false negatives. All analysis was done using R version 3.6.3

RESULTS

Over the study period, 1,060 encounters met the inclusion criteria. There were 11 observations removed due to no associated outcome value and 23 observations were removed due to lack of triage screening data or vitals. Our final study sample consisted of 1,026 unique patients. The prevalence of COVID‐19 in this population was found to be 9.6% (95% CI = 9.9% to 11.6%) with 99 patients testing positive. The cohort had a mean (±SD) age of 52 (±19) years with 56% being female. Other demographic details are found in Table 1.
Table 1

Patient Characteristics and Demographics

Overall (N = 1,026)
Age (years)52 (±19)
Sex, Female573 (55.8)
Race/ethnicity
Asian37 (3.6)
Black or African American328 (32.0)
Other71 (6.9)
Unavailable/unknown55 (5.4)
White, Hispanic120 (11.7)
White, non‐Hispanic415 (40.4)
Comorbidities
HTN601 (58.6)
DM361 (35.2)
COPD188 (18.3)
Asthma329 (32.1)
ESRD89 (8.7)

Data are reported as mean (±SD) or n (%).

COPD = chronic obstructive pulmonary disease; DM = diabetes mellitus; HTN = hypertension; ESRD = end‐stage renal disease.

Patient Characteristics and Demographics Data are reported as mean (±SD) or n (%). COPD = chronic obstructive pulmonary disease; DM = diabetes mellitus; HTN = hypertension; ESRD = end‐stage renal disease. Alanine aminotransferase was found to have most frequent missing data with 29% in training and 31% in validation data subsets, followed by CXR (16 and 18%), lymphocytes (16 and 14%), and WBC (15 and 13%), visualized in Figure 1. The rest of the predictors had less than 3% missingness in both training and validation data. This data set reflects all subjects who were tested. The final selected covariates are presented in Table 2.
Figure 1

Visualization of missing data in training and validation data. AST/ALT = alanine aminotransferase/aspartate aminotransferase; MAP = mean arterial pressure; WBC =white blood cell count.

Table 2

Distribution of Variables Included Within Split by Training and Validation Data Sets

Total (N = 1026)

Mean (SD)/n (%)

Training (N = 770)

Mean (SD)/n (%)

Validation (N = 256)

Mean (SD)/n (%)

p Value
Patient specific
Age (years)52 (±19)52 (±19)52 (±19)0.626
Sex, female573 (55.8)429 (55.7)144 (56.2)0.881
Exposure
Congregate setting24 (2.3)20 (2.6)4 (1.6)0.343
Reported COVID‐19 exposure114(14.5)77 (10)37 (14.5)0.050
Labs
Absolute lymphocytes1.59 (1.78)1.63 (2.00)1.47 (0.88)0.226
ALT38 (49)37 (47)40 (55)0.409
AST/ALT ratio1.317 (0.784)1.333 (0.812)1.267 (0.693)0.330
WBC8.8 (5.8)8.7 (5.4)9.2 (6.8)0.191
Comorbidities1.5 (1.3)1.5 (1.3)1.5 (1.3)0.476
Symptoms
Respiratory680 (66.3)500 (64.9)180 (70.3)0.115
Fever374 (36.5)281 (36.5)93 (36.3)0.962
Vitals
MAP97 (17)97 (17)97 (17)0.922
Pulse90 (19)90 (18)92 (19)0.098
Respirations18 (4)18 (4)19 (4)0.654
SpO2 98 (2)98 (2)98 (3)0.396
Temperature98.5 (1.4)98.5 (1.3)98.5 (1.4)0.830
CXR interpretation
Negative680 (78.6)516 (79.5)164 (75.9)0.491
Typical for COVID113 (13.1)80 (12.3)33 (15.3)
Atypical for COVID72 (8.3)53 (8.2)19 (8.8)
COVID +99 (9.6)69 (9.0)30 (11.7)0.195

Data are reported as mean (±SD) or n (%).

ALT = alanine aminotransferase; AST = aspartate aminotransferase; MAP = mean arterial pressure; WBC = white blood cell count.

Visualization of missing data in training and validation data. AST/ALT = alanine aminotransferase/aspartate aminotransferase; MAP = mean arterial pressure; WBC =white blood cell count. Distribution of Variables Included Within Split by Training and Validation Data Sets Total (N = 1026) Mean (SD)/n (%) Training (N = 770) Mean (SD)/n (%) Validation (N = 256) Mean (SD)/n (%) Data are reported as mean (±SD) or n (%). ALT = alanine aminotransferase; AST = aspartate aminotransferase; MAP = mean arterial pressure; WBC = white blood cell count. Three models were developed for comparison. The logistic regression model was found to have the largest AUC of 0.89 (95% CI = 0.84 to 0.94) and simplistically contained only known exposure to COVID‐19, CXR consistent with viral pneumonia, elevated temperature, and reduced WBC. The two ensemble approaches used did not result in a superior AUC. Random forest method resulted in AUC of 0.86 (95% CI = 0.79 to 0.92) and gradient boosting method demonstrated an AUC of 0.85 (95% CI = 0.79 to 0.91). Associated receiver operator curves are presented in Figure 2. Discrimination of each model can be appreciated in Figure 3. To contextualize performance with respect to the necessary clinical decision making, sensitivity, specificity, and resulting PPV were reported in Table 3 at the prespecified threshold where NPV was 99%. The logistic regression model offered a higher PPV of 0.29 (0.2–0.39) compared to 0.20 (0.14–0.28) for the random forest and 0.22 (0.15–0.3) for the gradient‐boosted method. The selected logistic regression model is presented in Table 4 for review.
Figure 2

Area under the receiver operator curve of the three derived models.

Figure 3

Box plots for the three derived models with predicted probabilities stratified by the true outcome.

Table 3

Test Statistics for the Three Derived Models With Threshold Established at Location Where the Model Had 99% NPV

ModelAUC (95% CI) N Predicted +/True +Predicted –/True –Sensitivity (95% CI)Specificity (95% CI)NPV (95% CI)PPV (95% CI)PLR (95% CI)NLR (95% CI)
Logistic regression0.89 (0.84–0.94)256291550.97 (0.83–1)0.69 (0.62–0.75)0.99 (0.96–1)0.29 (0.2–0.39)3.1 (2.5–3.8)0.05 (0.01–0.3)
Random forest0.86 (0.79–0.92)256291130.97 (0.83–1)0.5 (0.43–0.57)0.99 (0.95–1)0.2 (0.14–0.28)1.9 (1.7–2.2)0.07 (0.01–0.47)
XGBoost0.85 (0.79–0.91)256291220.97 (0.83–1)0.54 (0.47–0.61)0.99 (0.96–1)0.22 (0.15–0.3)2.1 (1.8–2.5)0.06 (0.01–0.4)

AUC = area under the receiver operator curve; NLR = negative liklihood ratio; NPV = negative predictive value; PLR = positive likelihood ratio; PPV = positive predictive value.

Table 4

Presenting the Full Logistic Regression Model With Intercept and Associated Odds Ratios

Intercept and Predictorsβ‐CoefficientOR95% CI
Intercept–36
WBC–0.590.60.4‐0.8
Temperature0.441.61.2‐2.0
Known exposure1.514.52.4‐8.7
Positive CXR1.695.43‐10
Area under the receiver operator curve of the three derived models. Box plots for the three derived models with predicted probabilities stratified by the true outcome. Test Statistics for the Three Derived Models With Threshold Established at Location Where the Model Had 99% NPV AUC = area under the receiver operator curve; NLR = negative liklihood ratio; NPV = negative predictive value; PLR = positive likelihood ratio; PPV = positive predictive value. Presenting the Full Logistic Regression Model With Intercept and Associated Odds Ratios

DISCUSSION

Using data collected in the ED encounter of 1,026 patients presenting for evaluation of COVID‐19, we developed three statistical learning models to aid in discrimination of patients when COVID‐19 testing is not available, not timely, or of questionable result. When comparing the derived models, all had similar discriminative capacity when predicting positive COVID‐19 PCR result, but the logistic regression model offered identical performance at the predefined test threshold (NPV > 99%), the primary aim of this study, with a greater capacity to identify patients positive for COVID‐19 as well (PPV = 29%). We feel the results of this model provide great value to the bedside clinician in offering a relatively simplified and easy to interpret model with enough strength in discrimination to provide risk stratification for patients presenting to the ED for COVID‐19 (Figure 3). The value in this derived model is its accessibility and relevance to an acute care clinician with key features that were identified by the feature selection process—abnormal x‐ray, positive contact with disease, elevated temperature and WBC—clinically meaningful and regularly obtained during an ED encounter. While the ensemble methods had slightly decreased predictive capacity, they also attributed significant importance to these same variables further justifying their use in our final model. As frontline clinicians continue to care for all patients presenting to the ED, the threat of COVID‐19 creates a cognitive burden to ensure accurate diagnosis needed and expected of us to facilitate appropriate recommendations to prevent further spread of this disease. Clinicians in the ED and point of care are being inundated daily with the question of “does this patient have COVID?” while at the same time they are charged with preserving PPE, assigning patients to the right team, and conserving test reagent resources. Models that offer some discrimination among patients provide an easy method for assigning pretest risk stratification. In settings where testing is not easily available or scarce this may offer enough discrimination to reduce the need for testing in low‐risk patients. Even when testing may still be performed it could offer a data‐driven mechanism to reduce PPE consumption. Some of the more accessible tests to emergency providers have reportedly poorer sensitivity causing concern for false‐negative rates. , While not studied here specifically within this cohort, this model could be used to further lower the risk of COVID‐19 after initially receiving a negative point‐of‐care test. This could offer reassurance to care teams, minimize PPE consumption for patients who are admitted, and reduce the need for patients being discharged to perform the self‐isolation that would likely be unnecessary and unwarranted. While none of these models offer near perfect discrimination there is a clear separation in the two distributions in our results (Figures 2 and 3). This offers some flexibility in determination of an appropriate test threshold to divide the model prediction into two classes without significant loss in discrimination, increasing the capacity for utilizing the model for more than its capacity to rule out disease. Increasing the classification threshold could allow one to focus on PPV. While not typically a focus for emergency medicine, utilizing the model in this capacity could provide a method for identifying a high‐risk cohort early in course of care to aid in better allocation of scarce hospital resources such as restricted treatment regimens, negative pressure rooms, and even personal protective equipment. A model of this type could become even more useful in the future when concern for COVID‐19 is not the forefront of the acute care provider thoughts. If implemented appropriately, a provider caring for a patient with a high pretest probability for disease could receive an alert within the EHR that recommends testing and appropriate cohorting to minimize unnecessary staff exposure. With a posttest probability of 30% found utilizing available ED data, with our model would be useful toward these efforts. This is one of the first models to provide a context on the likelihood of COVID‐19 with a population specific to emergency providers in the United States. There has been limited work that remains in preprint from China that attempted to predict COVID‐19 with a goal of minimizing unnecessary testing, specifically CT scans because they were used as a primary decision point in China early in the pandemic. , Unfortunately, many of the features included in some of these early predictive models utilized laboratory tests that are not easily available or would not result during the normal course of an ED visit in the United States. Health care researchers are more frequently utilizing novel techniques for prediction because consensus seems to be that novel “AI” machine learning techniques offer improved prediction and discrimination. , Although allure exists for using some of these newer methods, the benefit of using a logistic regression model, in addition to its improved predictive performance for this use case, is the added interpretability offered. It becomes operationally more lucrative than these black box machine learning methods as it provides the basis for development of a simple decision tool for clinicians to risk stratify patients when a health system lacks the capacity for implementation of a regression model within an EHR. Unlike standard decision trees, the nodes that enable decision making are not easily extrapolated into a single decision rule due to the nature of the composite prediction. Without implementation within an EHR for automated scoring, ensemble methods provide little value to the average clinician.

LIMITATIONS

There are several limitations to our study. First and foremost, this was a retrospective analysis that is prone to bias. In addition, this resulted in much missing data though very few observations had a significant amount of missingness. Our institution's testing capacity, uniquely, has been available to ED physicians since COVID‐19 arrived in the North Texas area. Our broad inclusion criteria, institutional approach to liberal testing, and test availability likely aided to reduce spectrum bias typical in these retrospective derivation studies that overly sample the more ill patients. Another limitation was that this was a single site–derived model in a unique patient population with a presumed inherently high‐risk patient baseline. While generalization of this model would presumably show reduction in discriminatory capacity, the simplicity of the logistic regression model seems less prone to overfitting than a model with a significantly higher number of predictors. One feature that may have higher variance elsewhere would be the categorization of the radiology read because a discrete result was provided by our radiologists following a systemwide protocol. If left to emergency physicians or other radiologists without strict guidance, the predictive nature of the CXR may vary. Additionally, with more data, a more optimally fit model could be constructed with the addition of other high value predictors. Future work intends to externally validate on data from multiple other institutions. It is possible that the information lost due to the necessary discretization of the continuous variables within the tree‐based methods affected their performance when compared to the logistic regression model. Additionally, these ensemble prediction methods are typically “data‐hungry” and could have benefitted from additional observations for training that were unavailable at a single‐site study. It is possible future work with more observations would find that these methods offer better discrimination than we have found in our cohort. Finally, defining an outcome for modeling is difficult when criterion standards are not well validated. The PCR test has been used as criterion standard for diagnosis of COVID, but some of these assays have been found to have a nontrivial false‐negative rate. , While we could not expect all patients to get both a COVID PCR testing and CT chest in normal ED operations in the United States, we attempted to mitigate this false‐negative rate by allowing for repeat testing performed within a reasonable time horizon to their initial visit. Regardless, at present and the foreseeable future, the COVID PCR test is clinically being used as the source of truth.

CONCLUSION

Based on the analysis of a consecutive population of patients arriving to an ED for the evaluation of COVID‐19, the derived and internally validated model has good discriminatory capacity for COVID‐19 disease utilizing four easily obtained variables: exposure history, temperature, WBC, and chest X‐ray result. This model offers an easily interpretable and immediately usable to the average emergency clinician—from a rule‐of‐thumb method to implementation of the model within the EHR—to help drive high‐quality patient and system‐level care for COVID‐19 disease.
  21 in total

1.  MissForest--non-parametric missing value imputation for mixed-type data.

Authors:  Daniel J Stekhoven; Peter Bühlmann
Journal:  Bioinformatics       Date:  2011-10-28       Impact factor: 6.937

Review 2.  Hematologic, biochemical and immune biomarker abnormalities associated with severe illness and mortality in coronavirus disease 2019 (COVID-19): a meta-analysis.

Authors:  Brandon Michael Henry; Maria Helena Santos de Oliveira; Stefanie Benoit; Mario Plebani; Giuseppe Lippi
Journal:  Clin Chem Lab Med       Date:  2020-06-25       Impact factor: 3.694

3.  Pretest probability estimates: a pitfall to the clinical utility of evidence-based medicine?

Authors:  Molly A Phelps; M Andrew Levitt
Journal:  Acad Emerg Med       Date:  2004-06       Impact factor: 3.451

4.  Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): the TRIPOD statement.

Authors:  Gary S Collins; Johannes B Reitsma; Douglas G Altman; Karel G M Moons
Journal:  Ann Intern Med       Date:  2015-01-06       Impact factor: 25.391

5.  Comparison of Abbott ID Now and Abbott m2000 Methods for the Detection of SARS-CoV-2 from Nasopharyngeal and Nasal Swabs from Symptomatic Patients.

Authors:  Amanda Harrington; Brian Cox; Jennifer Snowdon; Jonathan Bakst; Erin Ley; Patricia Grajales; Jack Maggiore; Stephen Kahn
Journal:  J Clin Microbiol       Date:  2020-07-23       Impact factor: 5.948

6.  Improved Molecular Diagnosis of COVID-19 by the Novel, Highly Sensitive and Specific COVID-19-RdRp/Hel Real-Time Reverse Transcription-PCR Assay Validated In Vitro and with Clinical Specimens.

Authors:  Jasper Fuk-Woo Chan; Cyril Chik-Yan Yip; Kelvin Kai-Wang To; Tommy Hing-Cheung Tang; Sally Cheuk-Ying Wong; Kit-Hang Leung; Agnes Yim-Fong Fung; Anthony Chin-Ki Ng; Zijiao Zou; Hoi-Wah Tsoi; Garnet Kwan-Yue Choi; Anthony Raymond Tam; Vincent Chi-Chung Cheng; Kwok-Hung Chan; Owen Tak-Yin Tsang; Kwok-Yung Yuen
Journal:  J Clin Microbiol       Date:  2020-04-23       Impact factor: 5.948

7.  Negative Nasopharyngeal and Oropharyngeal Swabs Do Not Rule Out COVID-19.

Authors:  Poramed Winichakoon; Romanee Chaiwarith; Chalerm Liwsrisakun; Parichat Salee; Aree Goonna; Atikun Limsukon; Quanhathai Kaewpoowat
Journal:  J Clin Microbiol       Date:  2020-04-23       Impact factor: 5.948

8.  Comparison of Abbott ID Now, DiaSorin Simplexa, and CDC FDA Emergency Use Authorization Methods for the Detection of SARS-CoV-2 from Nasopharyngeal and Nasal Swabs from Individuals Diagnosed with COVID-19.

Authors:  Daniel D Rhoads; Sree S Cherian; Katharine Roman; Lisa M Stempak; Christine L Schmotzer; Navid Sadri
Journal:  J Clin Microbiol       Date:  2020-07-23       Impact factor: 5.948

9.  Real-time RT-PCR in COVID-19 detection: issues affecting the results.

Authors:  Alireza Tahamtan; Abdollah Ardebili
Journal:  Expert Rev Mol Diagn       Date:  2020-04-22       Impact factor: 5.225

10.  Multi-tiered screening and diagnosis strategy for COVID-19: a model for sustainable testing capacity in response to pandemic.

Authors:  Michael S Pulia; Terrence P O'Brien; Peter C Hou; Andrew Schuman; Robert Sambursky
Journal:  Ann Med       Date:  2020-05-14       Impact factor: 4.709

View more
  7 in total

Review 1.  Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.

Authors:  Thomas Struyf; Jonathan J Deeks; Jacqueline Dinnes; Yemisi Takwoingi; Clare Davenport; Mariska Mg Leeflang; René Spijker; Lotty Hooft; Devy Emperador; Julie Domen; Anouk Tans; Stéphanie Janssens; Dakshitha Wickramasinghe; Viktor Lannoy; Sebastiaan R A Horn; Ann Van den Bruel
Journal:  Cochrane Database Syst Rev       Date:  2022-05-20

2.  Clinical prediction models for diagnosis of COVID-19 among adult patients: a validation and agreement study.

Authors:  Nadia Dardenne; Médéa Locquet; Anne-Françoise Donneau; Olivier Bruyère; Anh Nguyet Diep; Allison Gilbert; Sophie Delrez; Charlotte Beaudart; Christian Brabant; Alexandre Ghuysen
Journal:  BMC Infect Dis       Date:  2022-05-14       Impact factor: 3.667

3.  CCEDRRN COVID-19 Infection Score (CCIS): development and validation in a Canadian cohort of a clinical risk score to predict SARS-CoV-2 infection in patients presenting to the emergency department with suspected COVID-19.

Authors:  Andrew D McRae; Corinne M Hohl; Rhonda Rosychuk; Shabnam Vatanpour; Gelareh Ghaderi; Patrick M Archambault; Steven C Brooks; Ivy Cheng; Philip Davis; Jake Hayward; Eddy Lang; Robert Ohle; Brian Rowe; Michelle Welsford; Krishan Yadav; Laurie J Morrison; Jeffrey Perry
Journal:  BMJ Open       Date:  2021-12-02       Impact factor: 2.692

4.  XGboost Prediction Model Based on 3.0T Diffusion Kurtosis Imaging Improves the Diagnostic Accuracy of MRI BiRADS 4 Masses.

Authors:  Wan Tang; Han Zhou; Tianhong Quan; Xiaoyan Chen; Huanian Zhang; Yan Lin; Renhua Wu
Journal:  Front Oncol       Date:  2022-03-17       Impact factor: 6.244

5.  Clinical prediction rule for SARS-CoV-2 infection from 116 U.S. emergency departments 2-22-2021.

Authors:  Jeffrey A Kline; Carlos A Camargo; D Mark Courtney; Christopher Kabrhel; Kristen E Nordenholz; Thomas Aufderheide; Joshua J Baugh; David G Beiser; Christopher L Bennett; Joseph Bledsoe; Edward Castillo; Makini Chisolm-Straker; Elizabeth M Goldberg; Hans House; Stacey House; Timothy Jang; Stephen C Lim; Troy E Madsen; Danielle M McCarthy; Andrew Meltzer; Stephen Moore; Craig Newgard; Justine Pagenhardt; Katherine L Pettit; Michael S Pulia; Michael A Puskarich; Lauren T Southerland; Scott Sparks; Danielle Turner-Lawrence; Marie Vrablik; Alfred Wang; Anthony J Weekes; Lauren Westafer; John Wilburn
Journal:  PLoS One       Date:  2021-03-10       Impact factor: 3.240

6.  Managing Pandemics with Health Informatics: Successes and Challenges.

Authors:  Mujeeb A Basit; Christoph U Lehmann; Richard J Medford
Journal:  Yearb Med Inform       Date:  2021-04-21

7.  Derivation With Internal Validation of a Multivariable Predictive Model to Predict COVID-19 Test Results in Emergency Department Patients.

Authors:  Samuel A McDonald; Richard J Medford; Mujeeb A Basit; Deborah B Diercks; D Mark Courtney
Journal:  Acad Emerg Med       Date:  2020-12-22       Impact factor: 5.221

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.