Literature DB >> 35909058

Remote COVID-19 Assessment in Primary Care (RECAP) risk prediction tool: derivation and real-world validation studies.

Ana Espinosa-Gonzalez1, Denys Prociuk1, Francesca Fiorentino2, Christian Ramtale1, Ella Mi1, Emma Mi1, Ben Glampson3, Ana Luisa Neves1, Cecilia Okusi4, Laiba Husain4, Jack Macartney4, Martina Brown5, Ben Browne5, Caroline Warren5, Rachna Chowla6, Jonty Heaversedge7, Trisha Greenhalgh4, Simon de Lusignan4, Erik Mayer1, Brendan C Delaney8.   

Abstract

BACKGROUND: Accurate assessment of COVID-19 severity in the community is essential for patient care and requires COVID-19-specific risk prediction scores adequately validated in a community setting. Following a qualitative phase to identify signs, symptoms, and risk factors, we aimed to develop and validate two COVID-19-specific risk prediction scores. Remote COVID-19 Assessment in Primary Care-General Practice score (RECAP-GP; without peripheral oxygen saturation [SpO2]) and RECAP-oxygen saturation score (RECAP-O2; with SpO2).
METHODS: RECAP was a prospective cohort study that used multivariable logistic regression. Data on signs and symptoms (predictors) of disease were collected from community-based patients with suspected COVID-19 via primary care electronic health records and linked with secondary data on hospital admission (outcome) within 28 days of symptom onset. Data sources for RECAP-GP were Oxford-Royal College of General Practitioners Research and Surveillance Centre (RCGP-RSC) primary care practices (development set), northwest London primary care practices (validation set), and the NHS COVID-19 Clinical Assessment Service (CCAS; validation set). The data source for RECAP-O2 was the Doctaly Assist platform (development set and validation set in subsequent sample). The two probabilistic risk prediction models were built by backwards elimination using the development sets and validated by application to the validation datasets. Estimated sample size per model, including the development and validation sets was 2880 people.
FINDINGS: Data were available from 8311 individuals. Observations, such as SpO2, were mostly missing in the northwest London, RCGP-RSC, and CCAS data; however, SpO2 was available for 1364 (70·0%) of 1948 patients who used Doctaly. In the final predictive models, RECAP-GP (n=1863) included sex (male and female), age (years), degree of breathlessness (three point scale), temperature symptoms (two point scale), and presence of hypertension (yes or no); the area under the curve was 0·80 (95% CI 0·76-0·85) and on validation the negative predictive value of a low risk designation was 99% (95% CI 98·1-99·2; 1435 of 1453). RECAP-O2 included age (years), degree of breathlessness (two point scale), fatigue (two point scale), and SpO2 at rest (as a percentage); the area under the curve was 0·84 (0·78-0·90) and on validation the negative predictive value of low risk designation was 99% (95% CI 98·9-99·7; 1176 of 1183).
INTERPRETATION: Both RECAP models are valid tools to assess COVID-19 patients in the community. RECAP-GP can be used initially, without need for observations, to identify patients who require monitoring. If the patient is monitored and SpO2 is available, RECAP-O2 is useful to assess the need for treatment escalation. FUNDING: Community Jameel and the Imperial College President's Excellence Fund, the Economic and Social Research Council, UK Research and Innovation, and Health Data Research UK.
Copyright © 2022 The Author(s). Published by Elsevier Ltd. This is an Open Access article under the CC BY 4.0 license. Published by Elsevier Ltd.. All rights reserved.

Entities:  

Mesh:

Year:  2022        PMID: 35909058      PMCID: PMC9333950          DOI: 10.1016/S2589-7500(22)00123-6

Source DB:  PubMed          Journal:  Lancet Digit Health        ISSN: 2589-7500


Introduction

Since the start of the COVID-19 pandemic in 2019, a priority in health systems globally has been to rapidly develop and validate data-driven algorithms to predict risk and guide care.1, 2, 3, 4 This includes a requirement to accurately distinguish between patients with COVID-19 who can be safely managed in the community setting and those whose care should be escalated to hospital. In primary care, patients with COVID-19 have been typically categorised into three care strategy categories: reassure and safety net (green risk), monitor (amber risk), and admit to hospital (red risk), on the basis of clinical judgement and consensus using a few clinical parameters, such as heart rate, pulse rate, and peripheral oxygen saturation (SpO2). Several risk prediction models have been used to support the management of COVID-19. However, these were derived and validated in hospital in-patient populations.6, 7 The QCOVID score, based on primary care data, predicts risk of death from COVID-19 based on demographic factors and pre-existing medical conditions, but it does not predict acute deterioration based on current clinical observations. The National Early Warning Score 2 (NEWS2) score is widely used in UK emergency care as an early warning score for sepsis. However, the parameters in NEWS2 (such as tachycardia, fever, and hypotension) are usually very late signs of clinical deterioration, and they have been found to perform poorly in the acute assessment of suspected COVID-19 in hospital in-patients, and have not been evaluated outside hospital settings. Because initial patient contacts with health services are increasingly carried out remotely, a score was needed that could be administered over the telephone or other remote means. Instead of vital signs, a score could be based on clinical symptoms which the patient or a relative could assess (eg, perceived breathlessness or confusion). A Delphi study (with qualitative and survey components) using 112 primary care clinicians and 50 patients derived a set of data items comprising symptoms and vital signs that could be included in a putative remote COVID Assessment in Primary Care prediction (RECAP) model. Templates for collection of these RECAP data elements (known as RECAP-V0), using appropriate Systematized Nomenclature of Medicine (SNOMED) clinical terms, were developed for electronic health record systems so that a model could be derived and then validated. Evidence before this study This study was conceived in March and April 2020, during the first COVID-19 peak in the UK. The review was done according to Cochrane Collaboration standards for rapid reviews; we identified systematic reviews and large-scale observational studies describing the signs and symptoms of COVID-19. We searched for papers published between Jan 1, 2020, and March 31, 2020, and repeated for papers published until April 30, 2020. We searched PubMed and Embase for research on COVID-19, using the search terms “COVID-19”, “SARS-CoV-2”, “novel corona”, and “2019-ncov”, “diagnos$”, “prognos$”, and “prediction model”. We also searched the publicly available publication list of the COVID-19 living systematic review using the same terms. Evidence gathered showed worsening of COVID-19 symptoms around the seventh day of disease and challenges to identify patients with higher likelihood of severity to emphasise their monitoring. To this end, tools such the National Early Warning Score 2 have been used in the UK to assess patients with COVID-19 in primary care, but they do not capture the characteristics of COVID-19 infection and are not suitable for community remote assessment. Several COVID-19 risk scores have been developed. QCOVID provides a risk of mortality considering patients’ existing risk factors but does not include acute signs and symptoms. The International Severe Acute Respiratory and emerging Infection Consortium 4C Deterioration model was specifically developed for hospital settings. In England, the NHS has implemented the Oximetry @home strategy to monitor patients with acute COVID-19 deemed at risk (>64 years old or with comorbidities) by providing pulse oximeters; however, the criteria for monitoring or for escalation have not been validated. There was a need to develop a risk prediction score to establish COVID-19 patients’ risk of deterioration to be used in the community via both face-to-face or remote consultation. Added value of this study We developed and validated two COVID-19-specific risk prediction scores: one to be used in the initial remote assessment of patients with COVID-19 to assess need for monitoring (Remote COVID-19 Assessment in Primary Care-General Practice score; RECAP-GP) and another to assess the need for treatment escalation, which includes peripheral saturation of oxygen as one of the model predictors (RECAP-oxygen saturation score; RECAP-O2). To our knowledge, these scores are the first COVID-19-specific risk prediction scores to assess and monitor COVID-19 patients’ risk of deterioration remotely. These scores will be a valuable resource to complement the use of oximetry in the community when assessing a patient with COVID-19. Implications of all the available evidence To manage pandemic waves and their demand on health care, patients with COVID-19 require close monitoring in the community and prompt escalation of their treatment. Guidance available so far relied on unvalidated tools and clinician judgement to assess deterioration. COVID-19-specific community-based risk prediction scores, such as RECAP, might contribute to reducing the uncertainty in the assessment and monitoring of COVID-19 patients, increase safety in clinical practice, and improve outcomes by facilitating rapid treatment escalation. This study aimed to develop and validate two prediction models: (1) a score incorporating observable vital signs (heart rate, temperature, respiratory rate, and oxygen saturation) and (2) a score for use when these parameters cannot be measured due to an absence of equipment or poor patient familiarity with self-assessment in a virtual consultation. Two cutoff values were derived and validated: one for the need of monitoring (green and amber risk) and the other for hospital admission (amber and red risk) of patients with COVID-19.

Methods

Study design and participants

RECAP was a prospective cohort and observational study. Patients (≥18 years old) with symptoms of COVID-19 who presented to primary care within 14 days of onset of symptoms diagnosed on the basis of clinical judgement who were able to provide informed consent (except for people who used Doctaly Assist where consent was not sought) were eligible for inclusion. Patients were followed for 28 days from the onset of symptoms to determine the occurrence of hospital admissions due to COVID-19. The study protocol and statistical analysis plan have been published previously. The study was sponsored by Imperial College London and approved by the North West - Greater Manchester East Research Ethics Committee and Health Research Authority in May 2020 (Integrated Research Assessment System [IRAS] number 283024; North West - Greater Manchester East Research Ethics Committee and Health Research Authority Research Ethics Committee reference number 20/NW/0266). The study was badged as an Urgent Public Health Study by the National Institute of Health Research in October, 2020. The Northwest London Whole System Integrated Care data analysis was undertaken within a research database that was given favourable ethics approval by the West Midlands Solihull Research Ethics Committee (reference number 18/WM/0323; IRAS project ID 252449). All data used in this paper were fully anonymised before analysis. Imperial clinical analytics, research, and evaluation environment (iCARE) is a Trusted Research Environment (TRE) and provides access to Health Research Authority Research Ethics Committee-approved anonymised data for research (reference number 21/SW/0120; IRAS project ID 282093). Analysis was done within the secure data processing platform of the Oxford Royal College of General Practitioners Clinical Informatics Digital Hub (ORCHID) TRE. Access to Doctaly Assist data was granted under the Control of Patient Information (COPI) regulation (and Confidentiality Advisory Group Resolution 5 after expiration of COPI in March, 2022); therefore, patient consent was not required. However, patients could opt-out using the NHS Digital National Opt-Out policy.

Data

To allow for parallel derivation and validation of the RECAP scores on different cohorts, we used four UK primary care settings (figure 1 ). General practices (GPs) from the northwest London and clinical research network and practices belonging to the Oxford–Royal College of General Practitioners (RCGP) Research and Surveillance Centre (RSC) network,15, 16 which includes practices from all over England, collected data from Oct 1, 2020, to Feb 28, 2021. The COVID-19 Clinical Assessment Service (CCAS), a nationwide COVID-19 remote assessment service, collected patient data from March 15, 2021, to May 23, 2021. We also used data collected from Nov 26, 2020, to Oct 26, 2021 through the Doctaly Assist platform, which was used in southeast London (the main contributor to which was the Lewisham clinical commissioning group) for monitoring patients with COVID-19. We ensured that datasets used for development and validation of each model did not overlap geographically or in time (figure 1). We excluded northwest London practices from the RSC cohort. Patients in southeast London were directly seen in COVID-19 hot hubs during the period of study and onboarded to Doctaly.
Figure 1

Settings used for derivation and validation of the RECAP scores.

(A) Recruitment dates per data source shown. Geographical separation as follows: CCAS was accessible for all patients in England; Doctaly was accessible for patients in southeast London; the RSC network was used by practices all over England (practices in northwest London were excluded), northwest London practices are restricted to northwest London boroughs. (B) Data sources used for derivation and validation of the RECAP scores. CCAS=NHS-111 COVID clinical assessment service. iCARE=Imperial clinical analytics, research, and evaluation environment. NWL=northwest London. ORCHID=Oxford-Royal College of General Practitioners Clinical Informatics Digital Hub environment. RCGP RSC=Royal College of General Practitioners Research and Surveillance Centre. RECAP-GP=Remote COVID-19 Assessment in Primary Care-General Practice score. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score.

Settings used for derivation and validation of the RECAP scores. (A) Recruitment dates per data source shown. Geographical separation as follows: CCAS was accessible for all patients in England; Doctaly was accessible for patients in southeast London; the RSC network was used by practices all over England (practices in northwest London were excluded), northwest London practices are restricted to northwest London boroughs. (B) Data sources used for derivation and validation of the RECAP scores. CCAS=NHS-111 COVID clinical assessment service. iCARE=Imperial clinical analytics, research, and evaluation environment. NWL=northwest London. ORCHID=Oxford-Royal College of General Practitioners Clinical Informatics Digital Hub environment. RCGP RSC=Royal College of General Practitioners Research and Surveillance Centre. RECAP-GP=Remote COVID-19 Assessment in Primary Care-General Practice score. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score. We used primary data on patients’ signs and symptoms collected in the community at the point of consultation linked to secondary data on hospital outcomes. For data collection, the RECAP-V0 electronic template, with selected SNOMED codes, was used in electronic health record systems (EMIS, TPP SystmOne, and Adastra) and completed by clinicians when assessing patients with signs and symptoms of COVID-19. For the Doctaly Assist platform, patients were provided with home pulse oximeters and onboarded to a platform that uses WhatsApp to ask questions based on RECAP-V0 items).17, 19, 20 Additional data on comorbidities, ethnicity (subsequently grouped into two categories White and minority ethnic [or non-White]), age, and sex that were included in the development of the risk prediction model were available from the electronic health record. Since the clinical question that needs to be supported is “does this patient require care escalation?”, hospital admission, defined as a night's hospital stay, was the main outcome to be predicted by the model. To derive outcomes, data were linked with hospital episode statistics via the ORCHID TRE at Oxford University, or with the northwest London SitRep data on COVID-19 admissions in iCARE environment at Imperial College Healthcare NHS Trust. When identifying patients’ COVID-associated admissions in hospital episode statistics data, we searched for COVID-19 International Classification of Diseases-10 codes U071, U072, U073, or U074 as first, second, or third cause of admission. Patients with COVID-19 as the second or third cause of hospital administration were included in this study if the first cause of admission was pneumonia, dyspnoea, pulmonary embolism, or chest pain.

Statistical analysis

We estimated a minimum sample size of 1317 participants for model development and 1400 for model external validation assuming 10% hospitalisation rate for COVID-19, a maximum of 24 predictor variables, a binary outcome (hospital admission), and a minimum 85% model specificity on validation. We aimed to recruit at least 2880 participants per setting assuming 5–6% loss to follow-up. A study-specific SNOMED code carrying the RECAP National Institute of Health research portfolio mumber was used to identify the relevant records for the study in ORCHID and iCARE. A detailed explanation of analysis is provided in the published statistical analysis plan and the appendix (pp 2–3). The extent of missing data for each variable (outcome and predictors) was assessed on degree of missingness and pattern of missingness (at random or not at random) through visual inspection of the distribution. Missing data were then imputed using multiple imputation chain equations, including the outcome endpoint. The imputed datasets were stacked using Rubin's rule before analysis. A probabilistic risk prediction based on a multivariable logistic regression model, including the variables in RECAP-V0 as factors, was done for the RSC dataset (RECAP-GP), and separately for the Doctaly Assist dataset (RECAP-O2). The second model that used only Doctaly Assist data was planned given the predicted availability of larger numbers of oxygen saturation readings in this dataset. The models allow estimation of the likelihood of a particular patient with a COVID-19 diagnosis being admitted to hospital with COVID-19 within 28 days of symptom onset. Variables were checked for independence from each other by including in the model interaction terms between age and respiratory rate when available. Elements of the RECAP-V0 that were constructed using alternative codes representing different levels of severity had that relationship maintained in the models. Using backward elimination, in which the least significant variable was eliminated and the model rerun for each iteration, two models including only the predictor factors that were shown to be statistically significant with a p value of less than 0·05 were obtained. Internal model validation was done using bootstrapping (500 repeats) and then the receiver operating characteristic (ROC) curves were plotted. The model performance in subpopulations in external validation datasets (CCAS, Doctaly 2, and northwest London) by age (≥65 years and <65 years) and by sex was investigated by comparing the diagnostic measures Akaike information criterion (AIC), optimism-corrected C-statistic, and visual inspection of the ROC curves. The effect on the area under the curve (AUC) of adding oxygen saturation at rest as a non-linear term in the RECAP-O2 model was investigated. Selection of upper and lower cutpoints for creating a red, amber, and green categorisation of patients was based on clinical consideration of risk and model performance. External validation of the RECAP-GP model was done using data separately from northwest London primary care practices and CCAS to verify the specificity, sensitivity, negative predictive value, and positive predictive value of the model predictions. The RECAP-O2 model was validated using a subsequent cohort of participants. Software analyses were done with R (version 4.1.2). The main R packages used for the analysis were rms, mice, miceMNAR, xplorerr, tidyverse, ggplot2, pubh, r2mlm, dplyr, tidyr, plotly, mlr3, and data.table.

Role of the funding source

The funders of the study had no role in study design, data collection, data analysis, data interpretation, or writing of the report.

Results

Between Oct 1, 2020, and Oct 26, 2021, 4278 patients (2415 [56·5%] from northwest London and 1863 [43·5%] from RCGP RSC) were recruited by 170 practices (103 [61%] in northwest London and 67 [39%] in RSC); CCAS enrolled 2674 patients; southeast London's Doctaly Assist provided records for 4033 patients (1948 [48·3%] for model development [Doctaly-1 dataset] and 2085 [51·7%] for model validation [Doctaly-2 dataset]; figure 1A and B). Although mean age and sex were similar across the four cohorts (table 1 ), there was a higher proportion of people minority ethnic groups in the northwest London and two Doctaly cohorts. We were interested in assessing whether belonging to a minority ethnic group predicted higher risk of severity, as reported in previous publications. Initially, we had 16 ethnicity groups but, for statistical and policy making purposes, we agreed that grouping ethnicity categories into white and minority ethnic would produce more meaningful results.
Table 1

Summary population characteristics

RCGP RSC (n=1863)NWL (n=2415)CCAS (n=2674)Doctaly-1 (n=1948)Doctaly-2 (n=2085)
Age, years49 (17·7)46 (18·3)42 (15·9)44 (13·1)39 (11·9)
BMI
BMI, kg/m228·8 (6·6)26 (6·5)28·8 (7·1)NANA
Missing data124 (6·6%)84 (3·5%)352 (13·1%)NANA
Obesity (BMI ≥30 kg/m2)633 (34·0%)545 (22·6%)970 (36·3%)NANA
Missing data124 (6·6%)84 (3·5%)358 (13·4%)NANA
Sex
Female1061 (56·9%)1352 (56·0 %)1470 (55·0%)1246 (64·0%)1271 (61·0%)
Male802 (43·1%)1063 (44·0%)1204 (45·0%)702 (36·0%)813 (39·0%)
Ethnicity
Non-White624 (33·5%)1787 (74·0%)2105 (78·7%)1334 (68·5%)1063 (51·0%)
White1239 (66·5%)628 (26·0%)569 (21·3%)614 (31·5%)1022 (49·0%)
Missing data121 (6·5%)207 (8·6%)438 (16·4%)146 (7·5%)2018 (96·8%)
Comorbidity
Diabetes*204 (11·0%)369 (15·3%)157 (5·9%)NANA
Hypertension*465 (25·0%)565 (23·3%)310 (11·6%)NANA
Coronary heart disease*130 (7·0%)72 (3·0%)101 (3·8%)NANA
Chronic kidney disease*130 (7·0%)91 (3·8%)53 (2·0%)NANA
Adverse social circumstances*3 (0·2%)3 (0·2%)NANANA

Data are n (%) or mean (SD). BMI=body-mass index. NA=not available. RCGP RSC=Royal College of General Practitioners Research and Surveillance Centre. NWL=northwest London. CCAS=COVID-19 Clinical Assessment Service.

Data obtained from linked general practitioner electronic health records.

Summary population characteristics Data are n (%) or mean (SD). BMI=body-mass index. NA=not available. RCGP RSC=Royal College of General Practitioners Research and Surveillance Centre. NWL=northwest London. CCAS=COVID-19 Clinical Assessment Service. Data obtained from linked general practitioner electronic health records. The RCGP RSC data and the Doctaly-1 datasets were used for model development (figure 1B). The models were subsequently validated with the northwest London and Doctaly-2 data. Table 2 outlines all the model predictor variables considered and whether they were included in the model following assessment of patterns of missing data. Observations, such SpO2, were mostly missing in the northwest London, RSC, and CCAS data; however, SpO2 was available for 1364 (70·0%) of 1948 patients who used Doctaly. Data on comorbidities were not available for the Doctaly Assist dataset because there was no linkage with GP health record data in southeast London.
Table 2

Predictor variables considered for inclusion in the model and whether they were included following missingness assessment

RCGP RSC (RECAP-GP model; n=1863)
NWL (RECAP-GP external validation; n=2415)
CCAS (RECAP-GP external validation; n=2674)
Doctaly-1 (RECAP-O2 model; n=1948)
Doctaly-2 (RECAP-O2 external validation; n=2085)
CompletenessConsidered in modelCompletenessConsidered in modelCompletenessConsidered in modelCompletenessConsidered in modelCompletenessConsidered in model
Shortness of breath*1602 (86·0%)Yes2101 (87·0%)Yes2567 (96·0%)Yes1792 (92·0%)Yes2085 (100%)Yes
Feeling feverish or shivers*1788 (96·0%)Yes1883 (78·0%)Yes2567 (96·0%)Yes1782 (91·5%)Yes2085 (100%)Yes
Temperature (observed)447 (24·0%)No169 (7·0%)No721 (27·0%)No837 (43·0%)No938 (45·0%)No
Fatigue*1713 (91·9%)Yes2101 (87·0%)Yes2433 (91·0%)Yes1772 (91·0%)Yes2085 (100%)Yes
Acute cognitive decline*1527 (82·0%)Yes1690 (70·0%)Yes0No1772 (91·0%)Yes2085 (100%)Yes
Time from first symptoms (days)1732 (93·0%)No772 (32·0%)No2567 (96·0%)No1928 (99·0%)Yes2085 (100%)Yes
Respiratory rate279 (15·0%)No169 (7·0%)No588 (22·0%)No1402 (72·0%)Yes1855 (89·0%)Yes
Heart rate409 (21·9%)No507 (21·0%)No1123 (42·0%)No1363 (70·0%)No20 (1·0%)No
Oxygen saturation at rest279 (15·0%)No386 (16·0%)No347 (13·0%)No1363 (70·0%)Yes1292 (62·0%)Yes
Oxygen saturation after 40 steps111 (6·0%)No193 (8·0%)No80 (3·0%)No1285 (66·0%)Yes1188 (57·0%)Yes
Muscle aches894 (48·0%)No941 (39·0%)No1390 (52·0%)NoNANoNANo
Trajectory of breathlessness*838 (45·0%)No1267 (52·5%)No2112 (79·0%)No19 (1·0%)yes0Yes
Diabetes1863 (100%)Yes2415 (100%)Yes2674 (100%)YesNANoNANo
Hypertension1863 (100%)Yes2415 (100%)Yes2674 (100%)YesNANoNANo
Coronary health disease1863 (100%)Yes2415 (100%)Yes2674 (100%)YesNANoNANo
Chronic kidney disease1863 (100%)Yes2415 (100%)Yes2674 (100%)YesNANoNANo
Age1863 (100%)Yes2415 (100%)Yes2674 (100·0%)Yes1948 (100·0%)Yes2085 (100%)Yes
Sex1863 (100%)Yes2415 (100%)Yes2674 (100%)Yes1948 (100%)Yes2085 (100%)Yes
Ethnicity1741 (93·5%)Yes2197 (91·0%)Yes2246 (84·0%)Yes1801 (92·5%)Yes62 (3·0%)No

CCAS=COVID-19 Clinical Assessment Service. NA=not available. NWL=Nnorthwest London. RECAP-GP=Remote COVID-19 Assessment in Primary Care-General Practice score. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score. RCGP RSC=Royal College of General Practitioners Research and Surveillance Centre. SNOMED=Systematised Nomenclature of Medicine.

These items contain a proposed severity scale in RECAP-V0 and were captured in the electronic health records templates as a drop-down list of alternatives with appropriate per-item SNOMED coding; this ordering was carried through to the analysis.

Extracted from general practitioners electronic health records systems.

Collected from the caller as part of the NHS-111 pathway.

Predictor variables considered for inclusion in the model and whether they were included following missingness assessment CCAS=COVID-19 Clinical Assessment Service. NA=not available. NWL=Nnorthwest London. RECAP-GP=Remote COVID-19 Assessment in Primary Care-General Practice score. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score. RCGP RSC=Royal College of General Practitioners Research and Surveillance Centre. SNOMED=Systematised Nomenclature of Medicine. These items contain a proposed severity scale in RECAP-V0 and were captured in the electronic health records templates as a drop-down list of alternatives with appropriate per-item SNOMED coding; this ordering was carried through to the analysis. Extracted from general practitioners electronic health records systems. Collected from the caller as part of the NHS-111 pathway. Hospitalisation rates due to COVID-19 were similar in all the datasets except Doctaly-2: 83 (4·4%) of 1863 participants in the RSC, 92 (3·8%) of 2415 in northwest London, 82 (3·1%) of 2674 in CCAS, and 65 (3·3%) of 1948 in Doctaly-1, but only 19 (0·9%) of 2085 in Doctaly-2. All continuous data were found to be sufficiently normally distributed by visual inspection and the pattern of missingness for each variable was random. The pattern of data missingness was assumed to be missing at random because no clear pattern was observed in any of the variables considered for the model. The normal distribution of the variables allowed us to use multiple imputation chain equations for missing data imputation. The RECAP-GP model was built with the RCGP RSC data. The predictor variables used in the final model were sex, age, history of hypertension, degree of breathlessness, and temperature symptoms (table 3 ). Fatigue, confusion, ethnicity, body-mass index (BMI), diabetes, coronary heart disease, and chronic kidney disease were excluded after backward elimination (p>0·05). Age and BMI were included as complex splines fitted through penalised thin plate regression. There was no non-linearity in the fitted splines because dimensionality was optimal at 1 degree of freedom, indicating a linear slope of fit. The model showed good performance to distinguish between risk levels (AUC 0·80 [95% CI 0·76–0·85]; figure 2 ). There was no substantial difference in the performance of the model when stratified for sex (0·81 [0·75–0·86] for men and 0·74 [0·68–0·81] for women) and age (0·70 [0·63–0·78] for patients ≥65 years and 0·73 [0·67–0·79] for those <65 years). The number of patients in the 65 years old or older group was lower (354 [19%] of 1863 patients) than the number of patients younger than 65 years (1509 [81%] patients), which was in line with the English population in 2019 (12·4 million [18·5%] of 67·1 million people).
Table 3

The RECAP-GP model

Coefficientp value
Intercept−6·32<0·0001
Sex
Male1 (ref)..
Female0·560·018
Age (years)0·04<0·0001
Hypertension history0·560·04
Breathlessness
Breathlessness (cannot complete sentences at rest)1·69<0·0001
Breathlessness on mild exertion0·610·025
Breathlessness on moderate exertion0·220·57*
Fever
Temperature (rigors)0·100·98*
Temperature (feeling feverish)0·750·002

Absence of hypertension, breathlessness, and fever are the base coefficients in the logistic regression, set to zero and not shown.

For fever and breathlessness severity all levels were included if one level was significant. RECAP-GP=Remote COVID-19 Assessment in Primary Care-General Practice score.

Figure 2

Receiver operating characteristic curve of the RECAP-GP model

Bootstrapping for internal validation along with model diagnostic measures obtained as part of model calibration and performance assessment was done. Error bars and shaded areas are 95% CIs. AUC=area under the curve. RECAP-GP=Remote COVID-19 Assessment in Primary Care–general practitioner score.

The RECAP-GP model Absence of hypertension, breathlessness, and fever are the base coefficients in the logistic regression, set to zero and not shown. For fever and breathlessness severity all levels were included if one level was significant. RECAP-GP=Remote COVID-19 Assessment in Primary Care-General Practice score. Receiver operating characteristic curve of the RECAP-GP model Bootstrapping for internal validation along with model diagnostic measures obtained as part of model calibration and performance assessment was done. Error bars and shaded areas are 95% CIs. AUC=area under the curve. RECAP-GP=Remote COVID-19 Assessment in Primary Care–general practitioner score. The cutoff points for the green, amber, and red risk groups were chosen by the research team clinicians (BCD, SdL, ErM, EmM, ElM, TG, and AE-G) before validation, using the specificities and sensitivities obtained from the ROC (figure 2). We opted for maximising model sensitivity (90%) for the low-to-moderate risk threshold to ensure all patients who needed monitoring were in the amber group: the associated specificity was 40%, the logit transformed threshold 0·027, and the interval likelihood ratio (LR) 0·16. We maximised specificity (90%) at the moderate-to-high risk threshold to limit the number of unnecessary hospital admissions from the amber group, the associated sensitivity was 40%, the logit transformed threshold 0·098, and the interval LR 6·0. For external validation, the prediction model was run using the northwest London data and the CCAS data separately. Following a data completeness assessment, seven (0·3%) of 2415 participants from northwest London cohort were removed, resulting in a sample size of 2408 with 90 hospitalisations. However, the data were similar to the GP data, and only 82 (3·1%) of 2674 patients were admitted to hospital, which was lower than initially expected admission rates and would have limited power for both model building and validation. Therefore, we used the CCAS data as a validation set only for the RECAP-GP model. We initially planned to use 1317 patients in the development set and 1400 patients in the for-validation set with 10% admitted and a 0·05 margin of error. The selected cutoff points were used to assign risk categories to patients (table 4 ), along with the observed model sensitivity, specificity, negative predictive value, and positive predictive value. Because true negative and true positive cannot be defined for the amber group, only the number of hospitalisations in this group is reported. In the northwest London GP data, the probability of being categorised as a patient at low risk (green) and not needing admission (ie, negative predictive value) was high (1435 [99%] of 1453; 95% CI 98–99), and the probabi lity of being in the high risk (red) group and being admitted (ie, positive predictive value) was low (22 [14%] of 158; 10–19). In the CCAS data, the negative predictive value was (1487 [98%] of 1512; 95% CI 98–99), equivalent to the GP data, but the positive predictive value was lower (12 [6%] of 204; 4–10).
Table 4

Validation of the RECAP-GP model

Green risk groupAmber risk groupRed risk group
Risk group assigned by model in northwest London general practitioner data
Actual hospitalisations18 (1%) of 145350 (6%) of 79722 (14%) of 158
Sensitivity (95% CI)61·9% (59·9–63·9)NA24·4% (16·0–34·6)
Specificity (95% CI)80% (70·2–87·7)NA94·1% (93·1–95·2)
Positive predictive value of red group designation (95% CI)NANA13·9% (9·8–19·4)
Negative predictive value of green group designation (95% CI)98·8% (98·1–99·2)NANA
Risk group assigned by model in CCAS data
Actual hospitalisations25 (2%) of 151245 (5%) of 95812 (6%) of 204
Sensitivity (95% CI)57·4% (55·4–59·3)NA14·6% (7·8–24·2)
Specificity (95% CI)69·5% (54·4–79·2)NA93·0% (92·0–94·0)
Positive predictive value of red group designation (95% CI)NANA5·9% (3·5–9·7)
Negative predictive value of green group designation (95% CI)98·3% (97·7–98·8)NANA

Positive predictive value was calculated as the number of hospitalisations in red group divided by the number of patients in red group. Negative predictive value was calculated as the number of patients non-admitted in green group divided by the number of patients in green group. CCAS=COVID-19 Clinical Assessment Service. RECAP-GP=Remote COVID-19 Assessment in Primary Care-General Practice score. NA=not applicable.

Validation of the RECAP-GP model Positive predictive value was calculated as the number of hospitalisations in red group divided by the number of patients in red group. Negative predictive value was calculated as the number of patients non-admitted in green group divided by the number of patients in green group. CCAS=COVID-19 Clinical Assessment Service. RECAP-GP=Remote COVID-19 Assessment in Primary Care-General Practice score. NA=not applicable. Predictor variables used in the final RECAP-O2 model were age, degree of breathlessness, fatigue, and SpO2 at rest (table 5 ). Sex, ethnicity, temperature, acute cognitive decline, days since onset of symptoms, respiratory rate, and trajectory of breathlessness were excluded after backwards elimination (p>0·05). SpO2 after activity was found to be colinear with SpO2 at rest in the model and was thus excluded from the final model. To explore the potential non-linearity of continuous parameters (age and oxygen saturation at rest) these were modelled as complex splines fitted through penalised thin plane regression. There was no non-linearity found in age, and slight non-linearity found in oxygen saturation at rest, estimated as 1·7 degrees of freedom. The degree of non-linearity was assessed by varying the number of the basis dimension used for spline fitting, and inspecting its influence on the effective number of dimensions in the final fit, with no significant change found. The effect of adding oxygen saturation at rest as a non-linear term in the model was investigated and no significant improvement in performance was found, with a negligible improvement in AIC of 0·15 and no change in McFadden's pseudo R2.
Table 5

The RECAP-O2 model

Coefficientp value
Intercept25·00<0·0001
Age (years)0·040·0002
Breathlessness
Breathlessness (feeling uncomfortable to breathe)0·920·030
Breathlessness when walking* around the room−0·430·26
Fatigue
Fatigue (difficult to wake up)1·500·068
Fatigue (too tired to do usual activities)1·230·0007
Oxygen saturation at rest (0–100%)−0·33<0·0001

Scales of severity are included if one element is significant and only the most severe level is used in the model.

For fatigue and breathlessness severity all levels were included if one level was significant. Absence of hypertension, breathlessness and fatigue are the base coefficients in the logistic regression, set to zero and not shown. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score.

The RECAP-O2 model Scales of severity are included if one element is significant and only the most severe level is used in the model. For fatigue and breathlessness severity all levels were included if one level was significant. Absence of hypertension, breathlessness and fatigue are the base coefficients in the logistic regression, set to zero and not shown. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score. The model was internally validated and calibrated through bootstrapping using 500 repeats. The AUC was 0·84 (95% CI 0·78–0·90; figure 3 ), which suggests good model performance. Of note, the bottom left of the ROC curve has a steeper slope than the RECAP-GP model, indicating better discrimination in patients with more severe disease. There was no significant difference in the performance of the model when stratified for sex (AUC 0·86 [95% CI 0·78–0·95] for men and 0·83 [0·74–0·91] for women) and age (0·86 [0·75–0·98] for patients ≥65 years and 0·83 [0·77–0·90] for those <65 years),
Figure 3

Receiver operating characteristic curve of the RECAP-O2 model

Bootstrapping for internal validation along with model diagnostic measures obtained as part of model calibration and performance assessment was done. Error bars and shaded areas are 95% CIs. AUC=area under the curve. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score.

Receiver operating characteristic curve of the RECAP-O2 model Bootstrapping for internal validation along with model diagnostic measures obtained as part of model calibration and performance assessment was done. Error bars and shaded areas are 95% CIs. AUC=area under the curve. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score. External validation the RECAP-O2 model was run using the Doctaly-2 data. Following data completeness assessment, 411 (19·7%) of 2085 participants were removed from the Doctaly-2 sample, resulting in a sample size of 1674. The cutoff points were chosen by the clinical team based on the ROC curve (figure 3) and were used to assign risk categories to patients. The green–amber risk cutpoint at 90% sensitivity had a specificity of 52%, a logit transformed threshold of 0·013 and an interval LR of 0·19. For the amber–red risk cutpoint, 95% specificity had a sensitivity of 48%, a logit transformed threshold of 0·096, and an interval LR of 10·4. Validation results are shown in table 6 . Although the negative predictive value of patients in the green risk group was slightly higher than in RECAP-GP (1176 [99%] of 1183 participants; 95% CI 99–100), positive predictive value of patients in the red risk group was lower (three [9%] of 34 participants; 95% CI 3–22), which means lower probability of being admitted when categorised as high risk. Performance stratification by age and sex was also assessed in the external validation dataset. However, this was limited by the low number of admissions because 18 (1%) of 2085 participants in Doctaly-2 were hospitalised and there was only one hospitalisation in the subgroup of patients aged 65 years old or older. The results from Doctaly-2 were not representative of the model (appendix p 19).
Table 6

Validation of the RECAP-O2 model Risk group assigned by model in Doctaly-2 validation data

Green risk groupAmber risk groupRed risk group
Actual hospitalisations7 (1%) of 11839 (2%) of 4573 (9%) of 34
Sensitivity (95% CI)71·1% (68·8–73·2)NA15·8% (3·4–39·6)
Specificity (95% CI)63·1% (38·4–83·7)NA98·1% (97·3–98·7)
Positive predictive value of red group designation (95% CI)NANA8·8% (3·1–22·4)
Negative predictive value of green group designation (95% CI)99·4% (98·9–99·7)NANA

Positive predictive value was calculated as the number of hospitalisations in red group divided by the number of patients in red group. Negative predictive value was calculated as the number of patients non-admitted in green group divided by the number of patients in green group. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score. NA=not applicable.

Validation of the RECAP-O2 model Risk group assigned by model in Doctaly-2 validation data Positive predictive value was calculated as the number of hospitalisations in red group divided by the number of patients in red group. Negative predictive value was calculated as the number of patients non-admitted in green group divided by the number of patients in green group. RECAP-O2=Remote COVID-19 Assessment in Primary Care-oxygen saturation score. NA=not applicable.

Discussion

Assessment of severity of COVID-19 in the community is crucial to pandemic management worldwide. Our study provides a derivation and real-world validation of two risk scores specifically designed for patients with COVID-19 in the community. The RECAP-GP model includes degree of breathlessness, temperature symptoms, history of hypertension, sex, and age as hospital admission predictors, and can be used when a pulse oximeter is not available. This model provides a good prediction for non-admission in the lowest risk group. The model performs less well at differentiating amber from red risk groups. When validated in the CCAS data the positive predictive value is slightly lower but performance on the green and amber risk cutoff point is within the northwest London CIs. The RECAP-O2 model included degree of breathlessness, fatigue, SpO2 at rest, and age as predictors. This model can be used if pulse oximeters are available, including in patients at moderate risk who are being monitored. Although the improvement of the slope on the ROC curves indicates an interval LR increasing from 6 to 10 for the amber to red risk cutpoint, the model performs less well on validation. However, its specificity is good (98% [95% CI 97–99]), which supports its use to assess patients at moderate risk. The RECAP models establish an evidence base for the assessment of patients with COVID-19 in the community.5, 23 A recent systematic review identified that monitoring of SpO2 in the community (both at rest and after exercise) was useful to identify risk of patient deterioration, with an SpO2 of 92–94% considered the lower threshold for treatment escalation. In our study, an SpO2 of 92% placed the patient at the threshold for the high risk group. There was colinearity of SpO2 at rest and after exertion. Thus, only SpO2 at rest is used in the RECAP-O2 model; however, degree of breathlessness, when shortness of breath after movement is assessed, is included as a predictor factor. In the UK a national strategy for using home pulse oximetry, COVID Oximetry @home, was established in 2020 to identify silent hypoxia (hypoxia without breathlessness). UK guidance recommends provision of pulse oximeters to monitor patients with symptomatic COVID-19 and individuals older than 65 years or with specific long term conditions putting them at risk of severe COVID-19 (eg, severe liver or kidney disease).19, 24 The use of RECAP in the assessment of patients with COVID-19 is a valuable addition to this strategy. Our results are aligned with the UK guidance since age and SpO2 significantly predict deterioration in our models; additionally, RECAP considers symptom severity, so it can better support clinicians’ judgement on who needs monitoring, particularly for younger patients without comorbidities. Moreover, RECAP-O2 is better at identifying need for treatment escalation compared with SpO2 alone, which is only one factor in the final model. The use of multiple datasets to develop the models is a strength of the study. Moreover, the age distributions, admission rates, ethnicity, and comorbidities are in line with UK population expectations.25, 26 The cohorts from southeast and northwest London contain larger populations of Black British and South Asian ethnicity, whereas the RSC network included a smaller proportion of participants from minority ethnic groups. This supports the external generalisability of our findings. The thresholds used in the two RECAP scores can be adjusted to better suit local circumstances. Althogh the RECAP-V0 template, developed by the Delphi study, contained ten questions, the validated models contained four (RECAP-O2) and five (RECAP-GP) items only, significantly improving their fitness for use in the clinical setting. The choice of logistic regression modelling means that most electronic health record systems using SNOMED codes will be able to recreate the RECAP electronic templates and integrate the score calculator into the system. Moreover, data on hospitalisation was extracted from the NHS databases, which contain information on admission to tertiary NHS care facilities in England and are considered a reliable data source. Even if a patient initially assessed in northwest London was hospitalised in a different region or NHS area, their information would have been captured, and the data linked, in the databases used. The full models are provided as downloadable code on Github. The study has limitations. The RECAP models are founded on assessment of need for hospital admission in an observational dataset. In such a design it is impossible to completely eliminate incorporation bias. We mitigated this by counting admissions as at least one night in hospital, rather than only a review in the emergency department. Admission overnight will be based on investigation in hospital or a need for hospital-based therapy, thus reducing the role of RECAP score elements in the clinical decision that is the outcome variable. Health monitoring devices, such as pulse oximeters, are rarely available in the community. Therefore, these observations were largely missing in the GP and CCAS datasets and could not be included in the RECAP-GP model. By contrast, the Doctaly Assist datasets enabled us to assess the predictive value of resting and after exercise SpO2, heart rate, respiratory rate, and temperature, determining, for the first time, the diagnostic value of SpO2 monitoring in the community. The poor availability of electronic health record data linkage for the Doctaly data prevented the inclusion of comorbidities in the RECAP-O2 model, which might have contributed to the less good calibration of the model at higher risk scores. Moreover, although most of the recruitment was done from October, 2020, to May, 2021, the Doctaly Assist validation dataset was collected from May to October, 2021, at which point the UK's COVID-19 vaccination programme covered more than 70% of the adult population (>18 years). The potential effect of vaccination in hospital admission, along with the possible empirical treatment of COVID-19 in the community,30, 31 and the lower mean age of the Doctaly-2 cohort compared with other cohorts might explain the lower number of hospitalisations in this cohort and the difference in the RECAP-O2 model validation (including the lower performance when assessing performance by age and sex subgroups in the external validation dataset; appendix p 20) because lower admission rates require even better model discrimination to achieve good positive predictive value, and require a larger sample size than we planned. At the time of data collection, vaccination status was not consistently available in the electronic health record and policy was changing rapidly during the study. Our models are based on the associations between symptoms severity, and patient's demographics and comorbidities, and need for hospital admission (primary outcome). As recent publications report, vaccination status might affect the severity of symptoms, but the relationship between symptoms severity and likelihood of hospitalisation is not expected to change. Similarly, severity of symptoms in infections caused by different SARS-CoV-2 variants might also differ, but this will a probably not change the relationship between severity of symptoms and admission probability. We believe that our models are relevant to assess the likelihood of admission based on the patient's symptoms, regardless of the variant identified and vaccination status. However, all models should be subject to ongoing surveillance and calibration to ensure factors considered remain relevant, especially with rapid changes in variants and vaccines. Moreover, vaccination rate and treatment availability in different settings might invoke a change in the population baseline risk of hospitalisation, which might affect the absolute risk (while probably preserving the prediction power of model factors). This might justify recalibration and re-estimation of model intercept in populations with potentially different baseline risk in future studies. Depending on data availability, recalibration could involve the collection of data on predictors and predicted factors in a new population and additional information on vaccination status, PCR result, including type of variant, and COVID-19 treatment received in the community, to assess their predictive power. Despite the need for continuing calibration, the RECAP-GP model showed a good performance for all risk levels, which supports its use by practitioners to decide the need for monitoring patients with signs and symptoms of COVID-19. We suggest that the RECAP-GP score be used remotely in the initial assessment, without need for patient observations. Following this, if the patient is considered moderate to high risk, they could be provided with a home pulse oximeter for calculation of the RECAP-O2 score to detect deterioration. Given the low positive predictive value of the red categorisation on the amber to red cutpoint, the RECAP models will tend to overalert or overestimate (particularly for risk for higher risk groups). Care escalation should be considered with reference to national or local pathways, ability to monitor the patient in the community, hospital capacity, and shared decision making with the patient. Much has evolved since the first wave of the COVID-19 pandemic when we conceptualised this study. Mass vaccination has dramatically changed prognosis and oximetry is now much more widely used in the community than it was in early 2020, with at-home services available in many settings. Yet, new variants are triggering new pandemic waves across the globe, putting services under great strain. We believe that these two scores are likely to be valuable resources to support clinical judgement, reduce uncertainty, and improve safety in triage and monitoring of patients with suspected COVID-19 in health systems worldwide.

Data sharing

Due to data governance limitations, the deidentified patient data used to develop and validate the models cannot be shared. The model code is available in on Github.

Declaration of interests

SdL is Director of the Royal College of General Practitioners Research and Surveillance Centre, the English primary care sentinel system. This work funds part of his academic position, and the network and its Trusted Research Environment were part of this study, funded using the university standard cost template. SdL reports grants through his University from AstraZeneca, Eli Lilly, GSK, Sanofi, Seqirus, and Takeda, none have a direct link to this study and has been a member of Advisory Boards for AstraZeneca, Sanofi, and Seqirus, none have a direct link to this study. All other authors declare no competing of interests.
  20 in total

1.  NEWS 2 - too little evidence to implement?

Authors:  Luke E Hodgson; Jo Congleton; Richard Venn; Lui G Forni; Paul J Roderick
Journal:  Clin Med (Lond)       Date:  2018-10       Impact factor: 2.659

2.  Covid-19: a remote assessment in primary care.

Authors:  Trisha Greenhalgh; Gerald Choon Huat Koh; Josip Car
Journal:  BMJ       Date:  2020-03-25

3.  Assessing the safety of home oximetry for COVID-19: a multisite retrospective observational study.

Authors:  Jonathan Clarke; Kelsey Flott; Roberto Fernandez Crespo; Hutan Ashrafian; Gianluca Fontana; Jonathan Benger; Ara Darzi; Sarah Elkin
Journal:  BMJ Open       Date:  2021-09-14       Impact factor: 3.006

4.  Risk factors for SARS-CoV-2 among patients in the Oxford Royal College of General Practitioners Research and Surveillance Centre primary care network: a cross-sectional study.

Authors:  Simon de Lusignan; Jienchi Dorward; Ana Correa; Nicholas Jones; Oluwafunmi Akinyemi; Gayatri Amirthalingam; Nick Andrews; Rachel Byford; Gavin Dabrera; Alex Elliot; Joanna Ellis; Filipa Ferreira; Jamie Lopez Bernal; Cecilia Okusi; Mary Ramsay; Julian Sherlock; Gillian Smith; John Williams; Gary Howsam; Maria Zambon; Mark Joy; F D Richard Hobbs
Journal:  Lancet Infect Dis       Date:  2020-05-15       Impact factor: 25.071

5.  Risk stratification of patients admitted to hospital with covid-19 using the ISARIC WHO Clinical Characterisation Protocol: development and validation of the 4C Mortality Score.

Authors:  Stephen R Knight; Antonia Ho; Riinu Pius; Iain Buchan; Gail Carson; Thomas M Drake; Jake Dunning; Cameron J Fairfield; Carrol Gamble; Christopher A Green; Rishi Gupta; Sophie Halpin; Hayley E Hardwick; Karl A Holden; Peter W Horby; Clare Jackson; Kenneth A Mclean; Laura Merson; Jonathan S Nguyen-Van-Tam; Lisa Norman; Mahdad Noursadeghi; Piero L Olliaro; Mark G Pritchard; Clark D Russell; Catherine A Shaw; Aziz Sheikh; Tom Solomon; Cathie Sudlow; Olivia V Swann; Lance Cw Turtle; Peter Jm Openshaw; J Kenneth Baillie; Malcolm G Semple; Annemarie B Docherty; Ewen M Harrison
Journal:  BMJ       Date:  2020-09-09

Review 6.  Systemic and organ-specific immune-related manifestations of COVID-19.

Authors:  Manuel Ramos-Casals; Pilar Brito-Zerón; Xavier Mariette
Journal:  Nat Rev Rheumatol       Date:  2021-04-26       Impact factor: 20.543

7.  Royal College of General Practitioners Research and Surveillance Centre (RCGP RSC) sentinel network: a cohort profile.

Authors:  Ana Correa; William Hinton; Andrew McGovern; Jeremy van Vlymen; Ivelina Yonova; Simon Jones; Simon de Lusignan
Journal:  BMJ Open       Date:  2016-04-20       Impact factor: 2.692

8.  Emergence of a Novel Coronavirus (COVID-19): Protocol for Extending Surveillance Used by the Royal College of General Practitioners Research and Surveillance Centre and Public Health England.

Authors:  Simon de Lusignan; Jamie Lopez Bernal; Maria Zambon; Oluwafunmi Akinyemi; Gayatri Amirthalingam; Nick Andrews; Ray Borrow; Rachel Byford; André Charlett; Gavin Dabrera; Joanna Ellis; Alex J Elliot; Michael Feher; Filipa Ferreira; Else Krajenbrink; Jonathan Leach; Ezra Linley; Harshana Liyanage; Cecilia Okusi; Mary Ramsay; Gillian Smith; Julian Sherlock; Nicholas Thomas; Manasa Tripathy; John Williams; Gary Howsam; Mark Joy; Richard Hobbs
Journal:  JMIR Public Health Surveill       Date:  2020-04-02

9.  Inhaled budesonide for COVID-19 in people at high risk of complications in the community in the UK (PRINCIPLE): a randomised, controlled, open-label, adaptive platform trial.

Authors:  Ly-Mee Yu; Mona Bafadhel; Jienchi Dorward; Gail Hayward; Benjamin R Saville; Oghenekome Gbinigie; Oliver Van Hecke; Emma Ogburn; Philip H Evans; Nicholas P B Thomas; Mahendra G Patel; Duncan Richards; Nicholas Berry; Michelle A Detry; Christina Saunders; Mark Fitzgerald; Victoria Harris; Milensu Shanyinde; Simon de Lusignan; Monique I Andersson; Peter J Barnes; Richard E K Russell; Dan V Nicolau; Sanjay Ramakrishnan; F D Richard Hobbs; Christopher C Butler
Journal:  Lancet       Date:  2021-08-10       Impact factor: 79.321

10.  Characteristics of and Important Lessons From the Coronavirus Disease 2019 (COVID-19) Outbreak in China: Summary of a Report of 72 314 Cases From the Chinese Center for Disease Control and Prevention.

Authors:  Zunyou Wu; Jennifer M McGoogan
Journal:  JAMA       Date:  2020-04-07       Impact factor: 56.272

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.