Literature DB >> 33543082

The Template for Intervention Description and Replication as a Measure of Intervention Reporting Quality: Rasch Analysis.

Marcel P Dijkers1, Scott R Millis1,2.   

Abstract

OBJECTIVE: To determine whether the 12 items of the Template for Intervention Description and Replication (TIDieR) can be combined into a single summary score reflecting intervention reporting completeness and quality.
DESIGN: Systematic review and reanalysis of published data. After a systematic search of the published literature, 16 review articles were retrieved with 489 sets of 12 TIDieR ratings of experimental intervention, comparator, or the 2 combined as reported in primary studies. These 489 sets were recoded into a common format and analyzed using Rasch analysis for binary items.
SETTING: Not applicable. PARTICIPANTS: Not applicable.
INTERVENTIONS: Not applicable. MAIN OUTCOME MEASURES: Psychometric qualities of a Rasch Analysis-based TIDieR summary score.
RESULTS: The data fit the Rasch model. Infit and outfit values were generally acceptable (range, 0.70-1.45). TIDieR was reasonably unidimensional in its structure. However, the person (here: study) separation ratio was 1.25 with a corresponding reliability of 0.61. In addition, the confidence interval around each estimate of reporting completeness was wide (model standard error of 0.78).
CONCLUSION: Several Rasch indicators suggested that TIDieR is not a strong instrument for assessing the quality of a researcher's reporting on an intervention. It is recommended that it be used with caution. Improvements in TIDieR itself may make it more helpful as a reporting tool.
© 2020 The Authors.

Entities:  

Keywords:  CONSORT, Consolidated Standards of Reporting Trials; Guideline adherence; MNSQ, mean square; N/A, not applicable; NM, no modifications; RCT, randomized controlled trial; Rehabilitation; Reproducibility of results; Research report; TIDieR, Template for Intervention Description and Replication

Year:  2020        PMID: 33543082      PMCID: PMC7853349          DOI: 10.1016/j.arrct.2020.100055

Source DB:  PubMed          Journal:  Arch Rehabil Res Clin Transl        ISSN: 2590-1095


Complaints about the quality of the reporting of the results of medical and health care research were first published in the early 1980s,1, 2, 3, 4, 5 and the stream of critique has not abated. Poor reporting is relevant to science because it interferes with research users’ ability to distinguish poor research (with presumably less reliable or even irrelevant findings) from good research. In a worst-case scenario, well-performed groundbreaking research is reported so poorly that it never has any effect. One solution to poor reporting that has been attempted is the creation of checklists that tell authors which elements they should include in their papers, in what specific places. Some journals have made obligatory using and submitting the checklist appropriate for one’s study design. Checklists for randomized controlled trials (RCTs) were the first to be developed. The initial version of the Consolidated Standards of Reporting Trials (CONSORT) was published in 1997 and was followed by 18 subsidiary CONSORT standards, for example, for types of study design, classes of interventions, and report components. For biobehavioral research investigating the effects of complex interventions, the CONSORT standards, premised on drug versus placebo studies, turned out to be deficient, and a CONSORT version for nonpharmacological trials was published in 2008 and updated in 2017. Even CONSORT version for nonpharmacological trials was judged to be insufficient to guide authors, and in 2014 the TIDieR checklist and guide, was published, offering guidance for describing rehabilitation, psychotherapy, and all other treatments that cannot be simply communicated with a drug name and dosage. TIDieR offers a list of 12 items (table 1) which (added to CONSORT or to SPIRIT [Standard Protocol Items: Recommendations for Interventional Trials], another parent to TIDieR) together are expected to “improve the reporting of interventions and make it easier for authors to structure accounts of their interventions, reviewers and editors to assess the descriptions, and readers to use the information”(p1)
Table 1

The TIDieR checklist and Rasch analysis results for items

ItemDifficultyModel SEInfit MNSQOutfit MNSQ
T1 BRIEF NAME: Provide the name or a phrase that describes the intervention.−3.720.250.950.95
T2 WHY: Describe any rationale, theory, or goal of the elements essential to the intervention.−1.620.121.241.45
T3 WHAT: Materials: Describe any physical or informational materials used in the intervention, including those provided to participants or used in intervention delivery or in training of intervention providers. Provide information on where the materials can be accessed (eg, online appendix, URL).0.340.101.021.24
T4 WHAT: Procedures: Describe each of the procedures, activities, and/or processes used in the intervention, including any enabling or support activities.−0.920.111.041.00
T5 WHO PROVIDED: For each category of intervention provider (eg, psychologist, nursing assistant), describe their expertise, background, and any specific training given.0.560.100.960.90
T6 HOW: Describe the modes of delivery (eg, face-to-face or by some other mechanism, such as internet or telephone) of the intervention and whether it was provided individually or in a group.0.050.100.970.99
T7 WHERE: Describe the type(s) of location(s) where the intervention occurred, including any necessary infrastructure or relevant features.0.380.101.031.04
T8 WHEN and HOW MUCH: Describe the number of times the intervention was delivered and over what period of time including the number of sessions, their schedule, and their duration, intensity, or dose.−0.150.100.830.76
T9 TAILORING: If the intervention was planned to be personalized, titrated, or adapted, then describe what, why, when, and how.1.060.111.061.03
T10 MODIFICATIONS: If the intervention was modified during the course of the study, describe the changes (what, why, when, how).2.450.161.011.10
T11 HOW WELL—Planned: If intervention adherence or fidelity was assessed, describe how and by whom, and if any strategies were used to maintain or improve fidelity, describe them.1.040.120.900.83
T12 HOW WELL—Actual: If intervention adherence or fidelity was assessed, describe the extent to which the intervention was delivered as planned.0.530.110.990.94
Mean0.000.121.001.02
SD1.480.040.100.18
The TIDieR checklist and Rasch analysis results for items Even though CONSORT et al were created to assist authors in writing research reports, it did not take long for researchers to use these tools as measures to evaluate the reporting quality of entire batches of published reports. There are now dozens of papers in the literature reporting on the quality of published papers, in general or before and after the publication of CONSORT or another landmark reporting checklist. The same fate has befallen TIDieR—there now are more than 3 dozen published papers that assess, using TIDieR, the literature in a certain area. One of these papers, by Yamato et al, used a sample of 200 reports of RCTs to assess the completeness of physical therapy research reporting, separately for the experimental intervention and the comparator. The authors concluded that reporting was typically incomplete, for both arms. In a more recent paper, they reanalyzed their data, exploring whether the 12 TIDieR items can be summed to create an interval scale of reporting completeness. They argued that a simple summary score would be helpful in synthesizing the results of individual critical reviews and facilitating the evaluation of strategies to improve intervention reporting. They rated each TIDieR item on a 0-1-2 scale, assigning 2 points if both intervention and comparator were described adequately, 1 point if either was, and 0 points if neither was. Then they conducted a Rasch analysis of the 2400 evaluations (12 ratings for 200 papers) and found that the data fit the Rasch model, targeted the sample well, with the items progressing in a logical order. Even so, they concluded, “The TIDieR summary score requires validation in an independent data set and, this could be carried out using the data generated in other evaluations of cohorts of articles using the TIDieR checklist.”(p34) The purpose of this study is to do so, using TIDieR ratings for papers in a variety of health care areas, including various rehabilitation disciplines.

Methods

In July 2019, we searched PubMed, Embase, PsycINFO, Web of Science, and CINAHL for any paper that in title or abstract used the term Template for Intervention Description and Replication or TIDieR. All abstracts were screened for the likelihood of authors using TIDieR as a tool to rate the reporting quality of full-text articles published in the peer-reviewed literature. The full texts of these were examined for reporting, either in a table in the text, or as supplemental appendix S1 (available online only at http://www.archives-pmr.org), ratings on all or most TIDieR items, for multiple papers. We found 16 studies with ratings of altogether 489 individual papers (table 2).
Table 2

Key information on the 16 studies contributing data

Author NameDomain of InterventionStudy Designs IncludedYears of Publication of Papers ExtractedNumber of Experimental Intervention Arms RatedNumber of Comparator Intervention Arms RatedNumber of Combined Experimental and Comparator Arms RatedNumber of Original TIDieR Items UsedTIDieR Rating Criterion
Baron et al18Skin-care self-management interventions for people with spinal cord injuryRandomized and nonrandomized trials1974-2016170012Fully reported
Bartholdy et al19Exercise for knee osteoarthritisAny design1982-20121330012Completeness of reporting sufficient for replication
Comer et al20Non-pharmacological interventions for non-inflammatory multi-joint painRandomized and non-randomized trials, pre- post designs2006-201540012Not stated
Nascimento et al21Interventions for non-specific low back painRCTs1998-20181818012Completeness of description
Grudniewicz et al22Printed educational materials to improve primary care physicians’ knowledge, behavior, and patient outcomesRCTs, quasi-randomized trials, controlled pre-post studies, interrupted time series1983-2014320012Completeness of reporting and replicability
Gspörer and Schrems23Long-term care nursing for elderly peopleAny design2015002212Completeness of reporting
Hacke et al24Exercise for hypertensionRCTs1980-2010002312Completeness of intervention description
Howlett et al25Physical activity interventions for inactive healthy adultsRCTs1998-20162619012Not stated
Knols et al26Exercise for lung transplant recipientsRCTs2003-2001770012Not stated
Liljeberg et al27Oral nutritional supplementsRCTs2002-20155817112Completeness of reporting
Mackie et al28Promotion of family involvement in the care of hospitalized patientsAny design, including qualitative2003-2014110012(Sufficiently) reported
Picariello et al29Social-psychological interventions for fatigue in end-stage kidney diseaseRCTs and quasi-RCTs2000-2015160012Adequately reported
Ross et al30Exercise for posterior tibial tendon dysfunction in adultsRCTs2008-201530012Completely reported
Stevens et al31Advice for low back pain selfcareRCTs1987-20152900Completeness of reporting
Guidelines2003-201414009
Sykes et al32Audit and feedback to improve dementia careAny longitudinal design2002-201513008Not stated
Zandstra et al33Shared decision making and care for women with heavy menstrual bleedingAny design1999-200780012Quality of description and reproducibility
Key information on the 16 studies contributing data Most of these secondary studies applied TIDieR to the experimental intervention, although that was not always clearly stated. In most instances, communication with first authors allowed us to clarify what arm(s) TIDieR had been applied to; see supplemental appendix S1. All authors used essentially a reported [1] vs. not [adequately] reported [0] rating scheme, with the exception of Picariello et al, who used a partial rating in addition. We recoded the partial scores to 0. Other than Yamato et al, we could not create a 0-1-2 coding scheme, because with 2 exceptions, only scores for 1 arm (generally, the treatment arm) were reported by the 16 authors. In addition, we considered that in future studies of the quality of intervention reporting in the literature, it would be useful to have a summary score for the comparator arm, separate from one for the intervention arm. The primary studies used many codes of not applicable (N/A), especially for TIDieR items 9 (tailoring) and 10 (modifications). We decided that if there was no tailoring or no modification in a primary study, the authors ought to say so explicitly in their report, and a code of 1 could be given. (If there is individualization or revision, TIDieR instructs to report the details; see table 1.) A similar decision was made with respect to the code ?, which occurred in about 10% of the ratings for items 10, 11 (strategies to improve fidelity and fidelity assessment methods), and 12 (actual fidelity). Arguing that a description that is unclear enough to make the rater question what rating to give is an insufficient description, we recoded these too to 0. Blank cells in the lists of published ratings, reflecting that authors had not used a particular TIDieR item, were left blank in the data file we created, and in the Rasch analysis (see below) these did not contribute information to the analysis for the item in question. Supplemental appendix S1 lists for each of the 16 papers what the nature of the published data was and how we recoded information to create a single file containing all 489 sets of up to 12 TIDieR ratings. This data file is available in supplemental appendix S1. Winstepsa software was used to perform Rasch analysis. The TIDieR items had 2 categories: 0 (not [adequately] reported) and 1 (adequately reported). We therefore used the dichotomous Rasch model to evaluate the psychometric characteristics of TIDieR used as a scale applied to studies (rather than the traditional persons) assessing its unidimensionality, reliability, and targeting. To assess structure and unidimensionality, we examined infit and outfit mean squares (MNSQs). Infit and outfit MNSQ values in the range of 0.6-1.4 are considered acceptable for rating scales and surveys. We also assessed unidimensionality by using infit and outfit values, as well as Rasch principal component analysis. Local dependency was assessed using Yen’s Q3 statistic, where highly locally dependent items will have correlations exceeding 0.70. We evaluated the internal consistency of person (here: study) and item performance by examining separation reliability estimates and separation ratio. Separation reliability for persons (here, papers) refers to the consistency of a paper’s responses across items, whereas the separation reliability for items refers to the consistency of TIDieR item performances across papers. Targeting was addressed using visual inspection of the person-item map. A measure with good targeting for the cases to be rated results in a map symmetric along the vertical axis with items and persons (papers) clustered in a similar fashion with a similar range. We considered conducting, in addition, separate Rasch analyses of TIDieR ratings of experimental interventions, of comparators, and of treatments and their comparators combined, but decided that the numbers available for the latter 2 categories (37 and 22, respectively) were too small to result in reliable findings as to differences in Rasch results between the 3 (eg, regarding difficulty of individual items). The measure (ie, TIDieR reporting adequacy score) reported by Winsteps for each of the 489 papers was merged back with data extracted from the 16 papers literature and further processed using SPSSb to calculate mean scores by secondary study and trial arm.

Results

The 16 secondary studies from which data were used covered a variety of health care research areas, research designs, years of publication, and number of primary studies evaluated (see table 2). Rasch analysis infit and outfit MNSQs along with item difficulty (measure) for each item are reported in table 1. Item 1 was the most adequately reported TIDieR item, whereas Item 10 was the least frequently endorsed. Infit and outfit values were generally acceptable (range, 0.70-1.45), suggesting that the items fit the Rasch model expectation of unidimensionality. Examination of the Rasch residual-based principal components showed that the total variance explained by the measures was only 35%. However, the expected variance (the variance that would likely be explained if the data fit the Rasch model exactly) also was 35%. In addition, the first contrast eigenvalue was 1.99 and the disattenuated correlations of person measures on different clusters of items were 1.0. Taken together, this pattern of findings suggests that TIDieR is reasonably unidimensional in its structure. We found a person (paper) separation ratio of 1.25 with a corresponding reliability of 0.61. The assumption of local item independence did not appear to be violated. None of the correlations between items and standardized residuals exceeded 0.70. The item-person (paper) map is shown in fig 1. The map is symmetric along the vertical axis with items and persons (papers) clustered in a similar fashion with a similar range, indicating good targeting.
Fig 1

Person-item map of TIDieR items and primary studies. Notes: # and . represent studies. Each # is 7 studies; each . is 1-6 studies. T1-T12 represent the 12 TIDieR items.

Person-item map of TIDieR items and primary studies. Notes: # and . represent studies. Each # is 7 studies; each . is 1-6 studies. T1-T12 represent the 12 TIDieR items. Table 3 provides mean TIDieR summary scores (measure, in the Winsteps report) and mean model SE (also from Winsteps) by study and (if applicable) by TIDieR ratings subgroup (experimental treatment only, comparator only, both) in a study. The last line indicates that the average TIDieR summary score of the 489 studies was 0.19, with a standard deviation of 1.30. However, the score of the average study had a model SE of 0.78, indicating that the confidence interval for each estimate was wide (running from −0.59 [0.19-0.78] to 0.97 [0.19+0.78]).
Table 3

Mean and standard deviation for measure and model SE, by secondary study and subgroup

Study/SubgroupMeasure Mean ± SDModel SE Mean ± SDn
Baron et al exp180.28 ±0.880.70 ±0.0717
Bartholdy et al exp19−0.56 ±0.770.73 ±0.07133
Comer et al exp20−0.72 ±1.280.77 ±0.144
Nascimento et al exp210.14 ±0.640.68 ±0.0418
Nascimento et al comp21−0.23 ±0.930.71 ±0.1418
Grudniewicz et al exp22−0.03 ±1.090.72 ±0.1332
Gspörer and Schrems both231.88 ±1.330.94 ±0.4122
Hacke et al exp241.09 ±0.900.74 ±0.1323
Howlett et al exp250.66 ±0.830.70 ±0.0626
Howlett et al comp25−0.22 ±1.130.73 ±0.1519
Knols et al exp26−1.18 ±0.590.79 ±0.097
Liljeberg et al exp270.21 ±1.000.71 ±0.1076
Mack et al exp280.22 ±1.270.73 ±0.0911
Picariello exp290.34 ±1.020.71 ±0.0616
Ross et al exp302.20 ±2.701.24 ±0.603
Stevens et al exp312.08 ±1.221.26 ±0.4729
Stevens et al cpg31−1.48 ±1.070.93 ±0.1814
Sykes et al exp321.19 ±1.430.99 ±0.3113
Zandstra et al exp331.38 ±0.680.74 ±0.088
Both experimental and comparator arm are described1.88 ±1.330.94 ±0.4122
Comparator arm only is described-0.22 ±1.020.72 ±0.1437
Experimental arm only is described0.14 ±1.260.77 ±0.22430
Total0.19 ±1.300.78 ±0.23489

Abbreviations: both, experimental and comparator arms rated together; comp, comparator arm; cpg, clinical practice guideline; exp, experimental arm.

Mean and standard deviation for measure and model SE, by secondary study and subgroup Abbreviations: both, experimental and comparator arms rated together; comp, comparator arm; cpg, clinical practice guideline; exp, experimental arm. There was a great difference in mean TIDieR scores between studies or groups, from 2.08 for Stevens’ primary studies, to −1.48 for Stevens’ clinical practice guidelines. At the bottom of table 3, mean scores are provided for ratings of treatment, comparator and combined arms. This information suggests that in the average paper, the report of the experimental arm was somewhat better (0.14, on average) than that of the comparator (−0.22).

Discussion

Our analysis confirms the findings of the Yamato et al study, that the TIDieR items satisfy the assumptions of Rasch analysis and can be combined into an interval level summary score. This confirmation occurred despite the fact that we made some major changes in the methodology, including scoring individual arms (rather than a combination of the 2 arms in an RCT) and assigning a score reflecting no (adequate) report to all entries of N/A, NM, or ? in the 16 studies that contributed data. However, based on the various indicators reported (person [paper] separation ratio of 1.25, paper reliability of 0.61, and model SE value of 0.78, on average), we conclude that the Rasch-calculated score is a poor measure of reporting quality and completeness; it basically allows one to distinguish 2 strata of papers only, good and poor ones. Yamato did not come to this conclusion; they decided that their Rash analysis “supports the use of a summary TIDieR score as an indicator of completeness of reporting.”(p34) Because they used a trichotomy for scoring items and we a dichotomy, and the fact that they used 11 items only (item 10 was omitted as being a constant across 200 papers), the numerical results are not directly comparable. However, some of the outfit MNSQs they report, of 0.47 (for item 11) and 0.19 (for item 1) suggest problematic items. So do the disordering of thresholds and a person (paper) reliability index of 0.62. They note the same restriction as we do (“the summary score may only be able to discriminate between the least and most detailed reports”(p31)), but they are much more optimistic that this should not stand in the way of use of TIDieR as a rating scale. The poor Rasch results we obtained may be the consequence of a number of factors, including the following: (1) the nature of TIDieR, designed as a checklist rather than a measurement tool per se; (2) our decisions on how to handle scores of ?, N/A, and NM; (3) differences in rater severity between the various secondary studies; and (4) some secondary studies not using all 12 TIDieR items, and this pattern possible being associated with rater severity and/or the application of N/A, and so on. This all does not mean that TIDieR as developed originally, a checklist to assist authors of intervention reports, is a poor tool; converting it to a quality scoring measure is, as of now, problematic. As Yamato stated, reviewers who want to assess the intervention reporting quality of the research literature in a particular area, or determine whether a particular intervention (eg, requiring the submission of a reporting checklist along with a research paper) has resulted in improved reporting, may want to have a simple single summary score, rather than working with the 12 TIDieR items separately. However, we doubt that such a summary score ever is sufficient, even if it can be created. Any such analysis will want to focus on the specifics, we think: what particular reporting areas are strong, which ones have improved the least, etc. If a reliable summary score can be created, it presumably will primarily have a supplemental use. The ease with which Rasch analysis produces scores on an interval measure may seduce one into believing that the TIDieR checklist is reliable as a measure and has utility in differentiating, with a high level of precision, between multiple strata of reporting completeness and quality. As a checklist, TIDieR can be improved by providing explicit guidance to authors for what to do in an instance of something not being planned or observed: they should be told to write something like “there was no tailoring; no modification; no ….” Then readers and report assessors do not need to guess what happened, or read the tea leaves of ambiguous statements. Many secondary study authors (whether included in our analysis, or not,) have distinguished subitems under various TIDieR items, to make sure that they determined whether issues that could be distinguished within an item were reported or not. There now is a tool similar to TIDieR, Consensus on Exercise Reporting Template, that has a lot of overlap in terms of items, but uses subitems for a few of them. Distinguishing smaller items makes it easier for raters to evaluate an intervention and for rating teams to determine causes of (dis)agreement with an independent second rater. A finer grained TIDieR checklist might result in a better Rasch measure, but it would increase the burden on authors to report what was done or not done and how exactly it was done. Consensus on Exercise Reporting Template was designed for the assessment of the completeness of exercise studies, and there now also is a TIDieR version for public health intervention studies. Presumably, additional specialized versions will follow, and authors of studies applying them to published intervention research are advised to take a careful look at subitems and how they should be scored—following the rule that each subitem needs to be satisfied before the parent item is considered satisfied, or otherwise. A last issue is the satisfactoriness of TIDieR per se in assisting authors in describing what is or are the active ingredient(s) in their interventions. For duplicators, it is nice to know what the name of the treatment is they are duplicating (item 1), but that knowledge by itself does not affect the participants in the replication intervention study. Nor do the descriptions of What-procedures and What-materials (items 4 and 3, respectively). The Hawthorne studies should be a reminder of the fact that often we have little idea of what actually is the cause of changes in clients or patients. Various research teams have started investigations of the (classification of) active ingredients in complex interventions (eg, Michie et al40, 41, 42 and Dijkers et al43, 44, 45), but the authors of TIDieR seem not to have relied on this work. Whyte et al46, 47, 48, 49 most recently published a series of papers on the Rehabilitation Treatment Specification System that make clear how involved and painfully slow is the task of identifying and adequately describing a single intervention, let alone a bundling of treatments into what we traditionally call an intervention (or comparator).46, 47, 48, 49

Study limitations

We followed Yamato et al in assuming that TIDieR is a reflective measure, and that therefore conducting a Rasch analysis makes sense to begin with, which is not necessarily the case. Research reports may not have an innate trait that can be called reportiveness and that determines with great certainty the satisfactoriness of answers to the 12 TIDieR items. If the quality of the text of a paper is determined by a large number of random factors (including journal page limitations, standards applied by editor and peer reviewers, authors’ urgency to publish or perish) and TIDieR constitutes a measure created with effect indicators, Rasch analysis has no role, and other means of determining a reporting quality total score need to be used. We were dependent on sets of TIDieR ratings we found in the literature using a systematic search, but know that there are at least 20 more studies that did not publish their raw TIDieR scores. Some of the available reviews were very small (in terms of the number of primary studies rated), and much of the information as to what was rated (experimental treatment only, or experimental and comparator), and how, was missing or at least ambiguous. If we erred in our decisions that the TIDieR ratings in a particular study applied to the experimental treatment only (and not to the intervention and comparator combined), that does not invalidate the results. Some may disagree with the method we selected to handle scores such as N/A, NM, and ?, and an alternative approach may result in somewhat different results. Last, in our analysis we did not take into account the clustering of primary studies within secondary ones, because the Winsteps program does not offer such an option. It should be considered that only 2 papers (Nascimento et al and Howlett et al) contributed to the comparator appraisal subgroup, so it is possible that rater severity in these 16 studies was confounded with arm rated. The same holds true for the average report on both intervention and control arm combined, which is based on just 1 secondary study (Gspörer and Schrems).

Conclusions

Our analysis of scores on the 12 TIDieR items reported for 489 primary studies in 16 reviews suggests that the unidimensionality assumption underlying Rasch analysis are satisfied. However, multiple indicators suggest that the TIDieR summary score does not offer a strong instrument for rating the overall quality of reporting on an intervention. Consequently, rehabilitation researchers should be careful in depending on this TIDieR summary score for an analysis of the quality of intervention reporting, over time or comparing between various areas of complex interventions. Improvement in TIDieR itself (subitems, better instructions for reporting absences) may enhance it as a checklist helping authors.

Suppliers

Winsteps; Winsteps Software Technologies. SPSS; IBM Corp.
  46 in total

1.  Extending the CONSORT statement to randomized trials of nonpharmacologic treatment: explanation and elaboration.

Authors:  Isabelle Boutron; David Moher; Douglas G Altman; Kenneth F Schulz; Philippe Ravaud
Journal:  Ann Intern Med       Date:  2008-02-19       Impact factor: 25.391

2.  Development of a theory-driven rehabilitation treatment taxonomy: conceptual issues.

Authors:  John Whyte; Marcel P Dijkers; Tessa Hart; Jeanne M Zanca; Andrew Packel; Mary Ferraro; Theodore Tsaousides
Journal:  Arch Phys Med Rehabil       Date:  2014-01       Impact factor: 3.966

3.  Consensus on Exercise Reporting Template (CERT): Explanation and Elaboration Statement.

Authors:  Susan C Slade; Clermont E Dionne; Martin Underwood; Rachelle Buchbinder
Journal:  Br J Sports Med       Date:  2016-10-05       Impact factor: 13.800

4.  Do Exercise Trials for Hypertension Adequately Report Interventions? A Reporting Quality Study.

Authors:  Claudia Hacke; David Nunan; Burkhard Weisser
Journal:  Int J Sports Med       Date:  2018-08-21       Impact factor: 3.118

Review 5.  When is audit and feedback effective in dementia care? A systematic review.

Authors:  Michael J Sykes; Jennifer McAnuff; Niina Kolehmainen
Journal:  Int J Nurs Stud       Date:  2017-11-10       Impact factor: 5.837

Review 6.  [Transparency and replicability of nursing intervention studies in long-term care: A selective literature review].

Authors:  Irene Gspörer; Berta M Schrems
Journal:  Z Evid Fortbild Qual Gesundhwes       Date:  2017-12-19

7.  CONSORT extension for reporting N-of-1 trials (CENT) 2015 Statement.

Authors:  Sunita Vohra; Larissa Shamseer; Margaret Sampson; Cecilia Bukutu; Christopher H Schmid; Robyn Tate; Jane Nikles; Deborah R Zucker; Richard Kravitz; Gordon Guyatt; Douglas G Altman; David Moher
Journal:  J Clin Epidemiol       Date:  2015-08-10       Impact factor: 6.437

8.  SPIRIT 2013 statement: defining standard protocol items for clinical trials.

Authors:  An-Wen Chan; Jennifer M Tetzlaff; Douglas G Altman; Andreas Laupacis; Peter C Gøtzsche; Karmela Krleža-Jerić; Asbjørn Hróbjartsson; Howard Mann; Kay Dickersin; Jesse A Berlin; Caroline J Doré; Wendy R Parulekar; William S M Summerskill; Trish Groves; Kenneth F Schulz; Harold C Sox; Frank W Rockhold; Drummond Rennie; David Moher
Journal:  Ann Intern Med       Date:  2013-02-05       Impact factor: 25.391

Review 9.  Self-management interventions for skin care in people with a spinal cord injury: part 2-a systematic review of use of theory and quality of intervention reporting.

Authors:  Justine S Baron; Katrina J Sullivan; Jillian M Swaine; Arlene Aspinall; Susan Jaglal; Justin Presseau; Dalton Wolfe; Jeremy M Grimshaw
Journal:  Spinal Cord       Date:  2018-05-25       Impact factor: 2.772

10.  From Theory-Inspired to Theory-Based Interventions: A Protocol for Developing and Testing a Methodology for Linking Behaviour Change Techniques to Theoretical Mechanisms of Action.

Authors:  Susan Michie; Rachel N Carey; Marie Johnston; Alexander J Rothman; Marijn de Bruin; Michael P Kelly; Lauren E Connell
Journal:  Ann Behav Med       Date:  2018-05-18
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.