| Literature DB >> 33762237 |
Elvira D'Andrea1, Lydia Vinals2, Elisabetta Patorno1, Jessica M Franklin1, Dimitri Bennett3,4, Joan A Largent5, Daniela C Moga6, Hongbo Yuan7, Xuerong Wen8, Andrew R Zullo9,10, Thomas P A Debray11,12, Grammati Sarri13.
Abstract
OBJECTIVE: To determine whether assessment tools for non-randomised studies (NRS) address critical elements that influence the validity of NRS findings for comparative safety and effectiveness of medications.Entities:
Keywords: clinical pharmacology; epidemiology; public health; qualitative research; statistics & research methods
Year: 2021 PMID: 33762237 PMCID: PMC7993210 DOI: 10.1136/bmjopen-2020-043961
Source DB: PubMed Journal: BMJ Open ISSN: 2044-6055 Impact factor: 2.692
Glossary of terms
| Term | Definition |
| Active comparator design | A study design that compares the effect of the drug of interest with another drug used in clinical practice instead of non-use. |
| Adjustment for causal intermediaries | Adjustment for an intermediate variable (or a descending proxy for an intermediate variable) on a causal path from exposure to outcome. |
| Case–control design | A study design in which cases (patients with outcomes) are identified and compared with controls (patients without outcomes) with respect to the exposure of interest. |
| Cohort design | A study design in which a group of patients (a cohort) is identified and followed to ascertain the occurrence of an outcome. |
| Confounding | A mixing of effects that arises when patients with different baseline risks are compared; the resulting effect measure is a mix of drug effects and risk factor effects. |
| Depletion of susceptibles | Selection bias that occurs when the initiation of exposure to a drug is associated with an early increased incidence rate of the study outcome, followed by a decreased incidence rate with longer duration of exposure (eg, users of new drugs are compared with users of older drugs). |
| Detection or surveillance bias | Bias that occurs when the degree of outcome surveillance (or an associated symptom) is related to exposure and is differential among the exposure groups. |
| Immortal time bias | Time-related bias that derives from including a period of follow-up during which, by design, outcomes cannot occur. |
| Time-window bias | Time-related bias, in the context of a case–control study nested in a cohort, that derives from the use of time-windows of different lengths between cases and controls to define time-dependent exposures. |
| Incorrect outcome model specification | Misspecification of a statistical model that leads to biased outcome results. Common causes are omission of a relevant variable, inclusion of an unnecessary variable, adopting the wrong functional form, incorrect specification of the error term, uncertainty about what the true model is and reciprocal causation. |
| Loss to follow-up bias | Bias that occurs when there is difference in retention during the follow-up period after enrolment that are related to exposure status and outcome. |
| New-user design | A study design that starts following patients at the time they initiate a new drug (also known as incident-user design) |
| Non-contemporaneous comparator bias | Bias generated by differences in the timing of selection of comparator group(s) within a study influence exposures and outcomes resulting in biased estimates. |
| Reverse causation (or reverse causality) | Bias due to direction of cause and effect contrary to a common presumption, or a two-way causal relationship between exposure and outcome. |
| Recall bias | Bias that occurs when participants do not remember previous events or experiences accurately or omit details (not for claims-based studies). |
| Selection bias | Bias that occurs when selection of participants or follow-up time is related to both intervention and outcome (eg, prevalent users of a drug are compared with non-users or incident users). Our framework has a separate subdomain that refers to selection bias due to lack of generalisability, applicability or transferability to patients who were excluded from the study. |
Methodological challenges addressed by the included assessment tools
| Domains | Cohort tools*, | Case–control tools, n=12 | NRS tools†, | Total, |
| 1. Methods for selecting participants, n (%) | 11 (100) | 12 (100) | 20 (95) | 43 (98) |
| Sampling strategies to correct selection bias | 4 (36) | 6 (50) | 9 (42) | 19 (43) |
| Inclusion and exclusion criteria of target population | 6 (55) | 8 (67) | 13 (61) | 27 (61) |
| Depletion of susceptibles | 1 (9) | 0 (0) | 0 (0) | 1 (2) |
| External validity of target population | 6 (55) | 6 (50) | 9 (43) | 21 (48) |
| Others‡ | 11 (100) | 12 (100) | 18 (86) | 41 (93) |
| 2. Measurement of exposure, outcomes, covariates and follow-up, n (%)§ | 11 (100) | 12 (100) | 19 (90) | 42 (95) |
| Measurement of exposure§ | 11 (100) | 11 (92) | 18 (81) | 40 (91) |
| Measurement of outcomes§ | 11 (100) | 11 (92) | 18 (81) | 40 (91) |
| Measurement of covariates | 4 (36) | 4 (33) | 4 (19) | 12 (27) |
| Measurement of follow-up | 9 (82) | 3 (25) | 5 (24) | 17 (39) |
| 3. Design-specific sources of bias, n (%) | 11 (100) | 10 (83) | 19 (90) | 40 (91) |
| New-user design | 0 (0) | 0 (0) | 2 (10) | 2 (5) |
| Active comparator design | 0 (0) | 0 (0) | 0 (0) | 0 (0) |
| Immortal time bias or time-window bias | 0 (0) | 0 (0) | 3 (14) | 3 (7) |
| Detection or surveillance bias | 1 (9) | 2 (17) | 1 (5) | 4 (9) |
| Loss to follow-up bias | 9 (82) | 1 (8) | 12 (57) | 22 (50) |
| Non-contemporaneous comparator bias | 0 (0) | 1 (8) | 5 (24) | 6 (14) |
| Reverse causation | 5 (45) | 1 (8) | 2 (10) | 8 (18) |
| Recall bias¶ | 1 (9) | 4 (33) | 1 (5) | 6 (14) |
| Interviewer or observer bias¶ | 1 (9) | 3 (25) | 7 (35) | 11 (25) |
| Ascertainment bias¶ | 0 (0) | 1 (8) | 1 (5) | 2 (5) |
| General item/question on bias¶ | 3 (27) | 3 (25) | 3 (14) | 9 (20) |
| Other biases** | 0 (0) | 2 (17) | 5 (24) | 7 (16) |
| 4. Confounding, n (%) | 11 (100) | 11 (92) | 18 (86) | 40 (91) |
| Study design used to minimise confounding | 6 (55) | 7 (58) | 13 (62) | 26 (59) |
| Confounders measured and included in statistical analyses | 10 (91) | 10 (83) | 18 (86) | 38 (86) |
| Potential unmeasured confounding addressed in the analysis (eg, proxy analysis and IV analysis) | 1 (9) | 1 (8) | 3 (14) | 5 (11) |
| 5. Lack of appropriateness of statistical analyses (with specific mention of overadjustment and/or incorrect outcome model specification), n (%) | 2 (18) | 3 (25) | 10 (48) | 15 (34) |
| 6. Methods for assessing statistical uncertainty in the findings (eg, CIs reported for each analysis), n (%) | 7 (64) | 6 (50) | 8 (38) | 21 (48) |
| 7. Methods for assessing internal validity (eg, sensitivity analysis addressing potential confounding, measurement errors or other biases), n (%) | 3 (27) | 3 (25) | 9 (43) | 15 (34) |
| 8. Methods for assessing external validity (eg, post hoc subgroup analysis and comparison with other populations), n (%) | 4 (36) | 3 (25) | 4 (19) | 11 (25) |
*Two tools evaluated both cohort and RCTs together; one tool evaluated both cohort and cross-sectional studies together.
†NRS tools refer to a single tool built to evaluate both cohort and case–control studies or a tool built to evaluate additional NRS (eg, cross-sectional studies and before–after studies) together with cohort and case–control studies. Eight NRS tools included also the evaluation of RCTs.
‡'Others’ refers to items not included in our evaluation framework but included in the reviewed tools to investigate selection bias (eg, population characteristics sufficiently described to determine the applicability of the research question, sample size justification and power description, and ethical considerations).
§Items or questions on exposure misclassification and/or outcome misclassification are counted in this domain and relative subdomains.
¶Design-specific biases not included in the evaluation framework but addressed by the reviewed tools.
**Other design-specific biases not included in the evaluation framework but addressed by a few tools (eg, bias due to missing data, patients' blinding, different length of follow-up between groups, Berkson’s bias and protopathic bias).
IV, instrumental variable; NRS, non-randomised studies; RCTs, randomised controlled trials.
Individual characteristics of the tools included in the systematic review
| Tool identified* | Year | Type of tool | Scope of the tool | Study design evaluated | tems |
| RELEVANT | 2019 | Checklist | Critical appraisal and reporting | NRS | 21 |
| RAMboMAN - GATE-EPIQ | 2019 | Rating scale+summary judgement | Critical appraisal | Coh (+RCTs), CC | Coh (+RCTs) 21, CC 18 |
| MMAT | 2018 | Checklist | Critical appraisal | NRS | 5 |
| CASP | 2018 | Checklist | Critical appraisal | Coh, CC | Coh 12, CC 11 |
| SURE | 2018 | Checklist+summary judgement | Critical appraisal | Coh, CC | Coh 13, CC 11 |
| JBI | 2017 | Checklist+summary judgement | Critical appraisal | Coh, CC | Coh 11, CC 10 |
| ROBINS-I | 2016 | Checklist+summary judgement | Critical appraisal | NRS | 34 (+8 optional question) |
| ISPOR-AMCP-NPC† | 2014 | Checklist+summary judgement | Critical appraisal | Coh CC | 32 |
| GRACE† | 2014 | Checklist+summary judgement | Critical appraisal | Coh CC | 11 |
| NIH–NHLBI | 2014 | Checklist+summary judgement | Critical appraisal | Coh (+CSS), CC | Coh (+CSS) 14, CC 12 |
| HEBW | 2014 | Checklist+summary judgement | Critical appraisal | Coh | 18 |
| RoBANS | 2013 | Rating scale | Critical appraisal | NRS | 6 |
| RTI-Item Bank | 2013 | Checklist | Critical appraisal | NRS | 13 |
| Newcastle-Ottawa | 2013 | Rating scale +summary judgement | Critical appraisal | Coh, CC | Coh 8, CC 8 |
| SIGN - V.3.0 | 2012 | Checklist+summary judgement | Critical appraisal | Coh, CC | Coh 14, CC 11 |
| Montreal | 2011 | Checklist | Critical appraisal | Coh CC (+RCTs) | 10 |
| EPHPP | 2011 | Rating scale | Critical appraisal | Coh CC (+RCTs) | 17 |
| STROBE – V.4 | 2007 | Checklist | Reporting | Coh, CC | Coh 22, CC 22 |
| TREND | 2004 | Checklist | Reporting | NRS | 22 |
| Margetts | 2002 | Checklist | Reporting | Coh CC | 11 |
| Zaza | 2000 | Checklist | Critical appraisal | Coh CC | 15 |
| Downs-Black | 1998 | Rating scale | Critical appraisal and reporting | Coh CC (+RCTs) | 27 |
| Elwood | 1998 | Checklist | Critical appraisal | Coh CC (+RCTs) | 20 |
| Hadorn | 1996 | Checklist | Critical appraisal | Coh (+RCTs) | 7 |
| London | 1996 | Checklist | Critical appraisal | Coh CC | 33 |
| Avis | 1994 | Rating scale+summary judgement | Critical appraisal and reporting | Coh CC (+RCTs) | 24 |
| Durant | 1994 | Checklist | Critical appraisal | CC | 23 |
| Levine | 1994 | Checklist | Critical appraisal | Coh CC (+RCTs) | 10 |
| Gyorkos | 1994 | Checklist | Critical appraisal | Coh, CC | Coh 6, CC 5 |
| Cho† | 1994 | Rating scale+summary judgement | Critical appraisal | NRS (+RCTs) | 24 |
| COEH | 1991 | Checklist | Critical appraisal | NRS | 54 |
| Fowkes-Fulton | 1991 | Checklist+summary judgement | Critical appraisal and reporting | NRS (+RCTs) | 6 |
| Lichtenstein | 1987 | Checklist | Critical appraisal | CC | 20 |
| Gardner | 1986 | Checklist | Critical appraisal | NRS | 12 |
| Horwitz | 1979 | Checklist | Critical appraisal and reporting | CC | 12 |
Nine tools from our bibliographic search provided two separate instruments to assess cohort or case–control studies. Thus, the overall number of included records is 35, while the number of included assessment tools is 44.
*Tool name or first author name, if the tool does not have an assigned name, and it was published in peer-review journals.
†Tool developed to assess NRS on the comparative safety and effectiveness of medications.
CASP, The Critical Appraisals Skills Programme; CC, case–control study; COEH, Centre for Occupational and Environmental Health of The University of Manchester; Coh, cohort study; CSS, cross-sectional study; EPHPP, Effective Public Health Practice Project Quality Assessment Tool; GRACE, The Good ReseArch for Comparative Effectiveness; HEBW, Health Evidence Bulletins Wales; ISPOR-AMCP-NPC, International Society for Pharmacoeconomics and Outcomes Research – Academy of Managed Care Pharmacy – National Pharmaceutical Council; JBI, The Joanna Briggs Institute; MMAT, Mixed Methods Appraisal Tool; NIH–NHLBI, The National Institute of Health - The National Heart, Lung, and Blood Institute; NRS, non-randomised studies; RAMboMAN, GATE-EPIQ, Recruitment Allocation Maintenance blind objective Measurements Analyses, Graphic Approach To Epidemiology – Effective Practice, Informatics and Quality Improvement; RCTs, randomised controlled trials; RELEVANT, The REal Life EVidence AssessmeNt Tool; RoBANS, Risk of Bias Assessment tool for Non-randomized Studies; ROBINS-I, Risk Of Bias In Non-randomized Studies of Interventions; RTI-Item Bank, Research Triangle Institute Item Bank; SIGN, The Scottish Intercollegiate Guidelines Network; STROBE, STrengthening the Reporting of OBservational studies in Epidemiology; SURE, Specialist Unit for Review Evidence; TREND, Transparent Reporting of Evaluations with Non-randomized designs.
General characteristics of the assessment tools included in the systematic review
| Characteristics | All, | Cohort*, | Case–control, | NRS†, |
| Publication year, n (%) | ||||
| 1979–1989 | 3 (7) | 0 (0) | 2 (17) | 1 (5) |
| 1990–1999 | 12 (27) | 2 (18) | 2 (17) | 8 (38) |
| 2000–2009 | 5 (11) | 1 (9) | 1 (8) | 3 (14) |
| 2010–2019 | 24 (55) | 8 (73) | 7 (58) | 9 (43) |
| Type of tool, n (%) | ||||
| Checklist | 22 (50) | 4 (36) | 6 (50) | 12 (57) |
| Checklist+summary judgement | 13 (30) | 5 (45) | 4 (33) | 4 (19) |
| Rating scale | 3 (7) | 0 (0) | 0 (0) | 3 (14) |
| Rating scale+summary judgement | 6 (14) | 2 (18) | 2 (16) | 2 (9) |
| Scope of the tool, n (%) | ||||
| Critical appraisal | 35 (80) | 9 (81) | 10 (83) | 16 (76) |
| Reporting | 4 (9) | 2 (18) | 1 (8) | 1 (5) |
| Critical appraisal and reporting | 5 (11) | 0 (0) | 1 (8) | 4 (19) |
| Tools designed for CER, n (%) | 3 (7) | 0 (0) | 0 (0) | 3 (14) |
| Number of items, median (IQR) | 13 (10.3–21.8) | 13 (9.5–16) | 11.5 (10.8–18.5) | 17 (11–24) |
*Two tools evaluated both cohort and RCTs together; one tool evaluated both cohort and cross-sectional studies together.
†NRS tools refer to a single tool built to evaluate both cohort and case–control studies or a tool built to evaluate additional NRS (eg, cross-sectional studies and before–after studies) together with cohort and case–control studies. Eight NRS tools included also the evaluation of RCTs.
CER, Comparative Effectiveness research; NRS, non-randomised studies; RCTs, randomised controlled trials.