Literature DB >> 31023756

Uniformity in measuring adherence to reporting guidelines: the example of TRIPOD for assessing completeness of reporting of prediction model studies.

Pauline Heus1,2, Johanna A A G Damen1,2, Romin Pajouheshnia2, Rob J P M Scholten1,2, Johannes B Reitsma1,2, Gary S Collins3, Douglas G Altman3, Karel G M Moons1,2, Lotty Hooft1,2.   

Abstract

To promote uniformity in measuring adherence to the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) statement, a reporting guideline for diagnostic and prognostic prediction model studies, and thereby facilitate comparability of future studies assessing its impact, we transformed the original 22 TRIPOD items into an adherence assessment form and defined adherence scoring rules. TRIPOD specific challenges encountered were the existence of different types of prediction model studies and possible combinations of these within publications. More general issues included dealing with multiple reporting elements, reference to information in another publication, and non-applicability of items. We recommend our adherence assessment form to be used by anyone (eg, researchers, reviewers, editors) evaluating adherence to TRIPOD, to make these assessments comparable. In general, when developing a form to assess adherence to a reporting guideline, we recommend formulating specific adherence elements (if needed multiple per reporting guideline item) using unambiguous wording and the consideration of issues of applicability in advance. © Author(s) (or their employer(s)) 2019. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Keywords:  adherence; prediction model; reporting guideline; risk score; tripod

Mesh:

Year:  2019        PMID: 31023756      PMCID: PMC6501951          DOI: 10.1136/bmjopen-2018-025611

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


The original 22 items of the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) statement were transformed into a systematic and transparent adherence assessment form including scoring rules, for use by anyone evaluating adherence to TRIPOD. During the development, the adherence assessment form was extensively discussed, piloted and refined. Recommendations for developing and using a standardised form for measuring adherence to a reporting guideline were formulated based on challenges encountered.

Background

Incomplete reporting of research is considered to be a form of research waste.1 2 To eventually implement research results in clinical guidelines and daily practice, one needs sufficient details regarding the research to critically appraise the methods and interpret study results in the context of existing evidence.3–6 To improve the reporting of health research, many reporting guidelines have been developed for various types of studies, such as the CONsolidated Standards Of Reporting Trials (CONSORT) statement, STAndards for Reporting of Diagnostic Accuracy (STARD) statement, Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement, STrengthening the Reporting of OBservational studies in Epidemiology (STROBE) statement, REporting recommendations for tumour MARKer prognostic studies (REMARK) and the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) statement.7–15 A large number of reporting guidelines can be found on the website of the Enhancing the QUAlity and Transparency Of health Research (EQUATOR) Network, an international collaboration that supports the development and dissemination of reporting guidelines in order to achieve accurate, complete and transparent health research reporting (www.equator-network.org).4 5 Publishing a reporting guideline followed by some form of recommendation or journal endorsement is not enough for researchers to adhere to reporting guidelines - a more active implementation is usually required.5 In their guidance for developers of health research reporting guidelines, Moher and colleagues proposed 18 steps to be taken in the development of a reporting guideline, including several post-publication activities.6 One of these activities is to evaluate the actual adherence and thus use of a reporting guideline over time, as has been carried out for CONSORT, STARD and PRISMA.16–23 In multiple evaluations of the same guideline, different approaches to extract, score and record adherence to items of the guideline were seen, making comparisons difficult.17 21–23 For example, a systematic review of studies assessing adherence to STARD found that the number of items assessed was inconsistent and the criteria required for considering the reporting of an item to be complete differed between adherence evaluations. In addition, not all studies performed quantitative scoring, preventing an objective comparison of adherence between studies.17 A systematic adherence-scoring-system is needed to enhance objectivity and to ensure consistent measurement of adherence to a reporting guideline. A unique assessment form for adherence evaluations would reduce variation in the number of items being evaluated, how multicomponent items are being handled, and the scoring rules applied (on item level and overall adherence), and thereby facilitate comparison of reporting between different fields and over time. As the TRIPOD statement was only recently published (2015), its impact has not been assessed yet. However, recently a baseline measurement was performed to evaluate the extent to which prediction model studies before the introduction of TRIPOD reported each of the TRIPOD items.24 Based on this, the TRIPOD steering committee aimed to develop a systematic and transparent adherence-scoring-system to be used by other researchers to facilitate and ensure uniformity in measuring adherence to TRIPOD in future studies. We also provide general recommendations on developing an adherence assessment form for other reporting guidelines.

Developing the TRIPOD adherence assessment form

Our adherence assessment form contains all 22 main items of the original TRIPOD statement. Ten of these TRIPOD items actually comprise two (items 3, 4, 6, 7, 14, 15 and 19), three (items 5 and 13) or five (item 10) sub items (denoted by a, b, c, etc; see box 1).15 25 For our TRIPOD adherence assessment form, we further specified these original TRIPOD items (main or sub items, hereafter referred to as items) into so-called adherence elements. When a TRIPOD item contains multiple elements to report, multiple adherence elements were used. For example, for TRIPOD item 5a ‘Specify key elements of the study setting (eg, primary care, secondary care, general population) including number and location of centres.’ we defined three adherence elements to record information regarding (1) setting, (2) number and (3) location of centres. We further distinguished four types of prediction model studies: model development, external validation, incremental value of adding one or more predictor(s) to an existing model, or a combination of development and external validation of the same model. Six TRIPOD items only apply to development of a prediction model (10a, 10b, 14a, 14b, 15a and 15b) and six only to external validation (10 c, 10e, 12, 13 c, 17 and 19a) (box 1).15 25 All TRIPOD items, except for TRIPOD item 17, were considered applicable to incremental value reports. As not all TRIPOD items apply to all four types of prediction model studies, we defined four versions of the adherence assessment form, depending on whether a report described model development, external validation, a combination of these, or incremental value. If a report addressed both the development and external validation of the same prediction model, the reporting of either was assessed separately and subsequently combined for each adherence element. There were several stages in the process of developing the adherence assessment form (figure 1). All authors commented on the first version of the form. A revised version was then piloted by four authors representing the TRIPOD steering committee (JBR, GSC, DGA and KGMM). Based on their experiences adaptations to the form were made, mainly in the number and wording of the adherence assessment elements. Subsequently, the form was piloted by a group of various end-users consisting of PhD students, junior researchers, assistant and associate professors, professors and senior editors (n=16). Thereafter, three other authors (PH, JAAGD and RP) used the next version of the form when assessing six studies in duplicate. Items that led to disagreement or uncertainty more than once (items 2, 4b, 5a, 5c, 6a, 6b, 7b, 8, 10a, 10b, 10d, 11, 13a, 13b, 19 and 20) were discussed within the entire author team, leading to the final version of the form that was used to assess adherence to TRIPOD in a set of 146 publications.24 The form was also used by another group assessing adherence to TRIPOD in prognostic models for diabetes (publication in preparation). Challenges encountered and discussions held in this stage, only led to textual refinements to the form. Our final adherence assessment form, including considerations and guidance regarding scoring and calculations, is summarised in online supplementary file 1. It can also be found on the website of the TRIPOD statement (www.tripod-statement.org).
Figure 1

Process of developing the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) adherence assessment form with the aim of reducing unnecessary variation in scoring quality of reporting of prediction model studies based on TRIPOD.

Title (D; V): identify the study as developing and/or validating a multivariable prediction model, the target population and the outcome to be predicted. Abstract (D; V): provide a summary of objectives, study design, setting, participants, sample size, predictors, outcome, statistical analysis, results and conclusions. Background and objectives: (D; V) Explain the medical context (including whether diagnostic or prognostic) and rationale for developing or validating the multivariable prediction model, including references to existing models. (D; V) Specify the objectives, including whether the study describes the development or validation of the model or both. Source of data: (D; V) Describe the study design or source of data (eg, randomised trial, cohort or registry data), separately for the development and validation data sets, if applicable. (D; V) Specify the key study dates, including start of accrual, end of accrual and if applicable, end of follow-up. Participants: (D; V) Specify key elements of the study setting (eg, primary care, secondary care or general population) including number and location of centres. (D; V) Describe eligibility criteria for participants. (D; V) Give details of treatments received, if relevant. Outcome: (D; V) Clearly define the outcome that is predicted by the prediction model, including how and when assessed. (D; V) Report any actions to blind assessment of the outcome to be predicted. Predictors: (D; V) Clearly define all predictors used in developing or validating the multivariable prediction model, including how and when they were measured. (D; V) Report any actions to blind assessment of predictors for the outcome and other predictors. Sample size (D; V): explain how the study size was arrived at. Missing data (D; V): describe how missing data were handled (eg, complete-case analysis, single imputation or multiple imputation) with details of any imputation method. Statistical analysis methods: (D) Describe how predictors were handled in the analyses. (D) Specify type of model, all model-building procedures (including any predictor selection) and method for internal validation. (V) For validation, describe how the predictions were calculated. (D; V) Specify all measures used to assess model performance and, if relevant, to compare multiple models. (V) Describe any model updating (eg, recalibration) arising from the validation, if done. Risk groups (D; V): provide details on how risk groups were created, if done. Development vs. validation (V): for validation, identify any differences from the development data in setting, eligibility criteria, outcome and predictors. Participants: (D; V) Describe the flow of participants through the study, including the number of participants with and without the outcome and, if applicable, a summary of the follow-up time. A diagram may be helpful. (D; V) Describe the characteristics of the participants (basic demographics, clinical features and available predictors), including the number of participants with missing data for predictors and outcome. (V) For validation, show a comparison with the development data of the distribution of important variables (demographics, predictors and outcome). Model development: (D) Specify the number of participants and outcome events in each analysis. (D) If done, report the unadjusted association between each candidate predictor and outcome. Model specification: (D) Present the full prediction model to allow predictions for individuals (ie, all regression coefficients and model intercept or baseline survival at a given time point). (D) Explain how to use the prediction model. Model performance (D; V): report performance measures (with CIs) for the prediction model. Model-updating (V): if done, report the results from any model updating (ie, model specification and model performance). Limitations (D; V): discuss any limitations of the study (such as non-representative sample, few events per predictor, missing data). Interpretation: (V) For validation, discuss the results with reference to performance in the development data and any other validation data. (D; V) Give an overall interpretation of the results, considering objectives, limitations, results from similar studies and other relevant evidence. Implications (D; V): discuss the potential clinical use of the model and implications for future research. Supplementary information (D; V): provide information about the availability of supplementary resources, such as study protocol, Web calculator and data sets. Funding (D; V): give the source of funding and the role of the funders for the present study. D; V: item relevant to both development and external validation; D: item only relevant to development; V: item only relevant to external validation Process of developing the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) adherence assessment form with the aim of reducing unnecessary variation in scoring quality of reporting of prediction model studies based on TRIPOD.

Using the TRIPOD adherence assessment form

Scoring adherence per TRIPOD item

First, one has to judge for each adherence element whether the requested information is available in a report. The elements are formulated as statements that can be answered with ‘yes ‘or ‘no’ (see online supplementary file 1). For some elements it may be acceptable if authors in their report make explicit reference to another publication (ie, explicitly mention that the information of that adherence element is described somewhere else). This is denoted by the answer option ‘referenced’. For adherence elements that do not apply to a specific situation (for example reporting of follow-up (item 4b) might be not relevant in a diagnostic prediction model study), there is the answer option ‘not applicable’. The next step is to determine the adherence of a report per TRIPOD item. In general, if the answer to all adherence elements of a particular TRIPOD item is scored ‘yes’ or ‘not applicable’, the TRIPOD item is considered as adhered. In some situations a different scoring rule is used, which is described in the adherence assessment form for the corresponding items.

Overall adherence to TRIPOD

A report’s overall TRIPOD adherence score is calculated by dividing the sum of the adhered TRIPOD items by the total number of applicable TRIPOD items. Since some TRIPOD items are not applicable to all four types of prediction model studies, this total varies. The total number of applicable TRIPOD items for development is 30, for external validation 30, for incremental value 35 and for development and external validation of the same model 36. In addition, five TRIPOD items (5c, 10e, 11, 14b and 17) might not be applicable for specific reports (online supplementary file 1). If one reviews multiple prediction model studies on their adherence to TRIPOD, overall adherence per TRIPOD item can be calculated by dividing the number of studies that adhered to a specific TRIPOD item by the number of studies in which the specific TRIPOD item was applicable.

Recommendations for developing and using a standardised form for assessing adherence to a reporting guideline

As described earlier, during the process of designing this adherence assessment form we extensively discussed, piloted and refined our methods. One issue specific to TRIPOD we discussed, were the different types of prediction model studies (development, external validation and incremental value) that can be found in various combinations within publications. As not all TRIPOD items apply to all types of prediction model studies, overall adherence scores need to be calculated per type of prediction model study. A more general issue is how to deal with items containing several reporting elements. For TRIPOD we decided to determine adherence to a specific item by requiring complete information on all elements of that item. Hence, we created multiple adherence elements per TRIPOD item, as necessary. Another issue with regard to scoring adherence is how to handle (elements of) TRIPOD items that were not applicable for a specific prediction model study. This not only concerns the judgements at the level of adherence elements, but also the calculations of adherence per TRIPOD item and of the overall adherence. Overall adherence, in the form of a percentage of items adhered to, requires a clear denominator of total number of items one can adhere to. One has to decide whether to take items that are considered not applicable into account in the numerator as well as in the denominator. Determining applicability is subjective and requires interpretation. In our experience, items for which interpretation was needed, sometimes indicated by phrases like ‘if relevant’ or ‘if applicable’, were the most difficult ones to score and these items are a potential threat to inter-assessor agreement. We present our recommendations for developing and using a standardised form for measuring adherence to a reporting guideline in box 2. Decide which items are applicable to the set of publications of which you are going to measure adherence to the reporting guideline. Split items of a reporting guideline that consist of several sub items and elements into separate adherence elements to enable more detailed judgement of reporting. Pay attention on explicit wording of adherence elements, to make them as objective as possible. Determine for which items reference to information in another publication (instead of explicit reporting of that information) is acceptable for adherence. Define how to handle items that are not applicable to a specific report: Agree on which items this may concern and in what specific situations an adherence element or item could be considered as not applicable. Decide how to incorporate the ‘not applicable’ scores’ in determining adherence, per item as well as overall. Provide the final tailored adherence assessment form with clear guidance about the procedure and pilot the document in a small number of studies with several assessors: If there is poor agreement, discuss and refine the document. With good agreement, complete the assessment for all publications. Abstract and document information separately for each adherence element. This creates flexibility, as one is able to decide post hoc which elements to incorporate in calculating adherence per item, and thus overall adherence.

Concluding remarks

Evaluation of the impact of a reporting guideline should be as standardised and uniform as possible. However, this is not straightforward as reporting guidelines are usually not developed as an instrument to measure completeness of reporting. We present an adherence assessment form that facilitates uniformity in measuring adherence to TRIPOD. The form is provided in online supplementary file 1 and on the website of the TRIPOD statement (www.tripod-statement.org). Although, when developing the form, we had researchers evaluating quality of reporting in mind as target users, it can also be used by others interested in assessing adherence to TRIPOD, like authors, journal reviewers and editors. We would like to emphasise that our form should be used for assessing adherence to TRIPOD and not for assessing quality of prediction model studies (for which the Prediction model study Risk Of Bias Assessment Tool (PROBAST) was developed; www.probast.org). We did not perform formal user testing or reliability assessments, however we refined our adherence assessment form based on extensive discussions and pilot assessments within the author team, as well as by other potential users. We advise developers of reporting guidelines to consider adherence issues and impact evaluation early in the process of guideline development, as also recommended by Moher and colleagues.6 More specifically, attention should be paid to explicit wording of items, to make them as objective as possible and facilitate the interpretation of applicability and relevance.
  25 in total

1.  CONSORT 2010 statement: updated guidelines for reporting parallel group randomized trials.

Authors:  Kenneth F Schulz; Douglas G Altman; David Moher
Journal:  Ann Intern Med       Date:  2010-03-24       Impact factor: 25.391

Review 2.  Reporting in randomized clinical trials improved after adoption of the CONSORT statement.

Authors:  Robert L Kane; Jye Wang; Judith Garrard
Journal:  J Clin Epidemiol       Date:  2006-10-02       Impact factor: 6.437

Review 3.  Reporting quality of diagnostic accuracy studies: a systematic review and meta-analysis of investigations on adherence to STARD.

Authors:  Daniël A Korevaar; W Annefloor van Enst; René Spijker; Patrick M M Bossuyt; Lotty Hooft
Journal:  Evid Based Med       Date:  2013-12-24

Review 4.  Transparent and accurate reporting increases reliability, utility, and impact of your research: reporting guidelines and the EQUATOR Network.

Authors:  Iveta Simera; David Moher; Allison Hirst; John Hoey; Kenneth F Schulz; Douglas G Altman
Journal:  BMC Med       Date:  2010-04-26       Impact factor: 8.775

Review 5.  Towards complete and accurate reporting of studies of diagnostic accuracy: The STARD Initiative.

Authors:  Patrick M Bossuyt; Johannes B Reitsma; David E Bruns; Constantine A Gatsonis; Paul P Glasziou; Les M Irwig; Jeroen G Lijmer; David Moher; Drummond Rennie; Henrica C W de Vet
Journal:  Ann Intern Med       Date:  2003-01-07       Impact factor: 25.391

6.  Reducing waste from incomplete or unusable reports of biomedical research.

Authors:  Paul Glasziou; Douglas G Altman; Patrick Bossuyt; Isabelle Boutron; Mike Clarke; Steven Julious; Susan Michie; David Moher; Elizabeth Wager
Journal:  Lancet       Date:  2014-01-08       Impact factor: 79.321

7.  STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies.

Authors:  Patrick M Bossuyt; Johannes B Reitsma; David E Bruns; Constantine A Gatsonis; Paul P Glasziou; Les Irwig; Jeroen G Lijmer; David Moher; Drummond Rennie; Henrica C W de Vet; Herbert Y Kressel; Nader Rifai; Robert M Golub; Douglas G Altman; Lotty Hooft; Daniël A Korevaar; Jérémie F Cohen
Journal:  BMJ       Date:  2015-10-28

8.  Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement.

Authors:  David Moher; Alessandro Liberati; Jennifer Tetzlaff; Douglas G Altman
Journal:  PLoS Med       Date:  2009-07-21       Impact factor: 11.069

9.  A systematic scoping review of adherence to reporting guidelines in health care literature.

Authors:  Zainab Samaan; Lawrence Mbuagbaw; Daisy Kosa; Victoria Borg Debono; Rejane Dillenburg; Shiyuan Zhang; Vincent Fruci; Brittany Dennis; Monica Bawor; Lehana Thabane
Journal:  J Multidiscip Healthc       Date:  2013-05-06

10.  REporting recommendations for tumour MARKer prognostic studies (REMARK).

Authors:  L M McShane; D G Altman; W Sauerbrei; S E Taube; M Gion; G M Clark
Journal:  Br J Cancer       Date:  2005-08-22       Impact factor: 7.640

View more
  19 in total

1.  TRIPOD statement: a preliminary pre-post analysis of reporting and methods of prediction models.

Authors:  Amir H Zamanipoor Najafabadi; Chava L Ramspek; Friedo W Dekker; Pauline Heus; Lotty Hooft; Karel G M Moons; Wilco C Peul; Gary S Collins; Ewout W Steyerberg; Merel van Diepen
Journal:  BMJ Open       Date:  2020-09-18       Impact factor: 2.692

2.  [Differences of characteristics and outcomes between cancer patients and patients with no active cancer hospitalised for a SARS-CoV-2 infection].

Authors:  Charlotte Fenioux; Yves Allenbach; Aurore Vozy; Joe-Élie Salem; Georgina Maalouf; Matheus Vieira; Alexandre Le Joncour; Olivier Benveniste; David Saadoun; Corinne Frère; Luca Campedel; Pierre Salem; Joseph Gligorov; Christian Funck-Brentano; Patrice Cacoub; Paul Gougis
Journal:  Bull Cancer       Date:  2021-03-20       Impact factor: 1.276

3.  Predicting falls in community-dwelling older adults: a systematic review of prognostic models.

Authors:  Gustav Valentin Gade; Martin Grønbech Jørgensen; Jesper Ryg; Johannes Riis; Katja Thomsen; Tahir Masud; Stig Andersen
Journal:  BMJ Open       Date:  2021-05-04       Impact factor: 2.692

4.  Completeness of reporting of clinical prediction models developed using supervised machine learning: a systematic review.

Authors:  Constanza L Andaur Navarro; Johanna A A Damen; Toshihiko Takada; Steven W J Nijman; Paula Dhiman; Jie Ma; Gary S Collins; Ram Bajpai; Richard D Riley; Karel G M Moons; Lotty Hooft
Journal:  BMC Med Res Methodol       Date:  2022-01-13       Impact factor: 4.615

5.  Clinical predictive models of invasive Candida infection: A systematic literature review.

Authors:  Adriana M Rauseo; Abdullah Aljorayid; Margaret A Olsen; Lindsey Larson; Kim L Lipsey; William G Powderly; Andrej Spec
Journal:  Med Mycol       Date:  2021-11-03       Impact factor: 3.747

6.  Assessment of Value of Neighborhood Socioeconomic Status in Models That Use Electronic Health Record Data to Predict Health Care Use Rates and Mortality.

Authors:  Alejandro Schuler; Liam O'Súilleabháin; Gina Rinetti-Vargas; Patricia Kipnis; Fernando Barreda; Vincent X Liu; Oleg Sofrygin; Gabriel J Escobar
Journal:  JAMA Netw Open       Date:  2020-10-01

7.  The International Collaboration for Cancer Classification and Research.

Authors:  Ian A Cree; Blanca Iciar Indave Ruiz; Jiri Zavadil; James McKay; Magali Olivier; Zisis Kozlakidis; Alexander J Lazar; Chris Hyde; Stefan Holdenrieder; Ros Hastings; Nasir Rajpoot; Arnaud de la Fouchardiere; Brian Rous; Jean Claude Zenklusen; Nicola Normanno; Richard L Schilsky
Journal:  Int J Cancer       Date:  2020-10-09       Impact factor: 7.396

8.  Protocol for a systematic review on the methodological and reporting quality of prediction model studies using machine learning techniques.

Authors:  Constanza L Andaur Navarro; Johanna A A G Damen; Toshihiko Takada; Steven W J Nijman; Paula Dhiman; Jie Ma; Gary S Collins; Ram Bajpai; Richard D Riley; Karel Gm Moons; Lotty Hooft
Journal:  BMJ Open       Date:  2020-11-11       Impact factor: 2.692

9.  Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies.

Authors:  Myura Nagendran; Yang Chen; Christopher A Lovejoy; Anthony C Gordon; Matthieu Komorowski; Hugh Harvey; Eric J Topol; John P A Ioannidis; Gary S Collins; Mahiben Maruthappu
Journal:  BMJ       Date:  2020-03-25

10.  Machine learning for initial insulin estimation in hospitalized patients.

Authors:  Minh Nguyen; Ivana Jankovic; Laurynas Kalesinskas; Michael Baiocchi; Jonathan H Chen
Journal:  J Am Med Inform Assoc       Date:  2021-09-18       Impact factor: 4.497

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.