Literature DB >> 34129173

Between-group minimally important change versus individual treatment responders.

Ron D Hays1, John Devin Peipert2.   

Abstract

PURPOSE: Estimates of the minimally important change (MIC) can be used to evaluate whether group-level differences are large enough to be important. But responders to treatment have been based upon group-level MIC thresholds, resulting in inaccurate classification of change over time. This article reviews options and provides suggestions about individual-level statistics to assess whether individuals have improved, stayed the same, or declined.
METHODS: Review of MIC estimation and an example of misapplication of MIC group-level estimates to assess individual change. Secondary data analysis to show how perceptions about meaningful change can be used along with significance of individual change.
RESULTS: MIC thresholds yield over-optimistic conclusions about responders to treatment because they classify those who have not changed as responders.
CONCLUSIONS: Future studies need to evaluate the significance of individual change using appropriate individual-level statistics such as the reliable change index or the equivalent coefficient of repeatability. Supplementing individual statistical significance with retrospective assessments of change is desirable.
© 2021. The Author(s), under exclusive licence to Springer Nature Switzerland AG.

Entities:  

Keywords:  Meaningful change; Minimally important difference; Reliable change index; Responder

Mesh:

Year:  2021        PMID: 34129173      PMCID: PMC8204732          DOI: 10.1007/s11136-021-02897-z

Source DB:  PubMed          Journal:  Qual Life Res        ISSN: 0962-9343            Impact factor:   4.147


Introduction

In health-related quality of life (HRQOL) research, the significance of group-level change is evaluated to assess treatment efficacy and effectiveness. In addition, group-level minimally important change (MIC) thresholds are used because trivial mean change can be statistically significant if the sample size is large enough. The MIC indicates if statistically significant group mean differences are large enough to be important or meaningful to patients and clinicians. Identifying those who improve (“responders” to treatment) provides important supplemental information to group-level change. This paper reviews approaches for assessing MIC and estimating responders to treatment. We note that while group-level MIC thresholds have been used to identify responders to treatment in HRQOL studies [1, 2], other approaches are more appropriate.

Estimating the MIC

MIC estimates rely on anchors to provide an external indication of the level of underlying change. The variety of possible anchors makes a single MIC estimate problematic. It is advisable to use multiple anchors whenever possible, but the most used anchor is a retrospective rating of change question such as: How is your health now compared to 6 weeks ago? Much better A little better About the same A little worse Much worse This example item refers to change in “health”. Depending on the context and measure being evaluated, the anchor might be worded more specifically such as “physical functioning”, “pain”, “getting along with family”, etc. The choice of words is likely to result in different MIC estimates. In addition, there are known limitations of retrospective ratings of change, which include a tendency to reflect the patient’s current state more than change, potentially due to recall bias [3, 4]. Change on the target measure should be correlated and have a monotonic association with change indicated on the anchor. The mean group change on the target measure should be larger for the subgroup of people who report they are much better on the anchor than mean change for the other subgroups. And those who report no change on the anchor should have no more than minimal change on the target measure [5]. The mean group changes on a HRQOL measure for those who report being “a little better” (improvement) or “a little worse” (decrement) are the basis for MIC estimates. But sometimes investigators fail to limit the MIC to those who changed a little and include all those with any change on the anchor. This was the case in a sample of 123 adult surgical patients with spinal deformity [6] and in a study of 223 patients with chronic low back pain [7]. Including all those who change rather than focusing on those with minimal but important change means that the MIC thresholds are too large.

Identifying responders to treatment

Individual-level variation and change can be estimated using simulation modeling for time series data, but it requires a minimum of 10 observation in the data stream [8]. Similarly, Moinpour et al. [9] estimated mixed effect models and noted that the PROMIS fatigue computer adaptive test would need 15 total assessments to obtain 0.90 reliability of change. Because of limits on research budgets and concerns about respondent burden, nearly all longitudinal HRQOL studies are limited to a few waves of assessment (e.g., two time points). Guidance for identifying responders to treatment for this environment are needed. Hence, we review approaches for estimating individual change from baseline to a single post-baseline assessment. Table 1 lists several formulae previously proposed for estimating the significance of individual change that are analogous to t-tests for within group change [10, 11]. All the formulae include individual change in the numerator and “error” in the denominator. The different methods vary in how they estimate error–for example, the time 1 standard deviation (SD), standard error of measurement (SEM; , standard error of estimation and the standard error of prediction . The is used for the reliable change index (RCI). Note that these formulae have between group variance in the denominators that may not be representative of variance in individual change.
Table 1

Formula for Evaluating Individual Change

StatisticFormula
Standard deviation index(X2X1)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$/{\mathrm{SD}}_{1}$$\end{document}/SD1
Standard error of measurement (SEM)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\mathrm{SD}}_{1}\sqrt{1-\mathrm{reliability}}$$\end{document}SD11-reliability
95% Confidence interval around SEMX1 ± \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1.96 \; \mathrm{SEM}$$\end{document}1.96SEM
Standard error of estimation\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\mathrm{SD}}_{1} \sqrt{\mathrm{reliability} (1-\mathrm{reliability})}$$\end{document}SD1reliability(1-reliability)
Standard error of prediction\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\mathrm{SD}}_{1} \sqrt{1-{\mathrm{reliability}}^{2}}$$\end{document}SD11-reliability2
Reliable change index(X2X1)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$/\sqrt{2 }\; \mathrm{SEM}$$\end{document}/2SEM
Coefficient of repeatabilitya\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1.96 \sqrt{2 }\, \mathrm{SEM}$$\end{document}1.962SEM
Reliable change index (practice effects)(X2X1 - practice effects)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$/\sqrt{2 \; \mathrm{SEM}}$$\end{document}/2SEM
Reliable change index (item response theory)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${(X}_{2 }-{X}_{1})/\sqrt{{\mathrm{SE}}_{1}^{2}+{\mathrm{SE}}_{2}^{2}}$$\end{document}(X2-X1)/SE12+SE22
Regression-based\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\left(X_{2}-{X}_{2{\text{p}}}^{\text{b}}\right)/\sqrt{{S}_{1}^{2}+{S}_{2}^{2}}\sqrt{1-\mathrm{reliability}}$$\end{document}X2-X2pb/S12+S221-reliability

aAlso known as the smallest detectable change, minimally detectable change, or smallest real difference. A group-level version of the coefficient of repeatability has been proposed by dividing the formula by the

bPrediction of time 2 from time 1 variables

Note: X2 = score at time 2; X1 = score at time 1; SD1 = standard deviation at time 1; SE1 = standard error at time 1; SE2 = standard error at time 2; S12 = variance at time 1; S22 = variance at time 2 used instead of SD1. The SD of change can be used instead of SD1. The estimated true score (mean + reliability (X1-mean)) may be used instead of X1 to account for regression to the mean

Formula for Evaluating Individual Change aAlso known as the smallest detectable change, minimally detectable change, or smallest real difference. A group-level version of the coefficient of repeatability has been proposed by dividing the formula by the bPrediction of time 2 from time 1 variables Note: X2 = score at time 2; X1 = score at time 1; SD1 = standard deviation at time 1; SE1 = standard error at time 1; SE2 = standard error at time 2; S12 = variance at time 1; S22 = variance at time 2 used instead of SD1. The SD of change can be used instead of SD1. The estimated true score (mean + reliability (X1-mean)) may be used instead of X1 to account for regression to the mean Following the conventional p < 0.05 threshold for group-level research, responders are usually defined by an RCI of 1.96 or larger. A variant of the RCI used for cognitive measures corrects for practice effects [12], though caution has been raised about its use [13]. The denominator of the RCI for item response theory (IRT) calibrated measures uses IRT standard errors at time 1 and time 2 [14, 15]. The coefficient of repeatability indicates the amount of change necessary to be significant on the RCI and is, therefore, equivalent to it. This coefficient is also known as the minimally detectable change, smallest real difference, and the smallest detectable change [16]. Variations to these methods have been proposed to account for regression to the mean (see Table 1). Regression-based approaches compare observed scores at time 2 with regression predicted scores based on time 1 score and other time 1 variables. This can be useful clinically because time 2 status is compared to what would be expected based on time 1 characteristics.

MIC thresholds should not be used to identify responders to treatment

There are two major problems with applying group-based MIC methods to categorize individual patients as having changed or not: one conceptual and one statistical. The conceptual issue regards using averages derived from groups that may not be relevant to any one patient. MIC estimates are averages from distributions of individual MICs; small changes may be meaningful for some and large changes for others [17, 18]. Even if such MIC estimates are derived from patient-reported anchors representing the construct of interest, these averages may not represent change that is meaningful to individuals. For example, an individual patient may consider only a large magnitude improvement in physical function to be meaningful and be uninterested in achieving the average improvement, since the average value falls below that individual’s perception of meaningful change. The statistical issue is that group-based MIC methods drastically underestimate the amount of change needed to be significant at the individual level due to the large measurement error around individual change scores [19]. “Any inspection of measured data reveals an order of magnitude difference between the variability in group versus individual changes” [20]. Thus, group-based MIC estimates will often be indistinguishable from individual measurement error [21]. Abu et al. [22] is a recent example of using MIC thresholds to identify whether patients improved or declined on the Atrial Fibrillation Effect QualiTy-of-Life (AFEQT) Questionnaire. A five-point change threshold was used as the threshold for “clinically meaningful change”. This threshold was based on group-level MIC estimates from a prior study of the AFEQT MIC that used physician assessment of functional status [23]. The authors concluded that 22% declined and 40% improved from baseline to 1 year later in a sample of 1097 older adults with atrial fibrillation. Table 2 shows the standard deviations, internal consistency reliabilities, and coefficients of repeatability for the four AFEQT scores. The coefficients of repeatability are two-to-three times larger than the 5-point change threshold the authors used. Ironically, Abu et al. could have adopted the more appropriate SDC estimates (equal to the coefficient of repeatability) reported by Spertus et al. [23].
Table 2

Amount of change in atrial fibrillation effect on QualiTy-of-Life (AFEQT) scores needed for significant individual change (coefficient of repeatability)

Overall scoreSymptomsDaily activitiesTreatment concerns
Standard deviation17.817.524.519.3
Internal consistency reliability0.90*0.950.940.90
Coefficient of repeatability15.610.816.616.9

*Exact reliability not reported in Abu et al. [22] so we estimated this from prior work[23]

Amount of change in atrial fibrillation effect on QualiTy-of-Life (AFEQT) scores needed for significant individual change (coefficient of repeatability) *Exact reliability not reported in Abu et al. [22] so we estimated this from prior work[23] The Abu et al. [22] paper is one where the MIC, derived from group-based estimates, falls well below the coefficient of repeatability. When this is the case, Kemmler et al. [21] suggest increasing the MIC thresholds to the coefficients of repeatability. Terwee et al. [16] recommend looking to see how measurement error might be reduced by: (1) increasing homogeneity of the study sample’s scores at the first measurement timepoint and thereby reducing the SD; and/or (2) increasing the reliability of the measure. Both options are made difficult if the amount of SD reduction or reliability increase is not trivial. Using the Abu et al. [22] example, we calculated and plotted the SD’s needed at 0.90, 0.95 and 0.99 reliability on the AFEQT. Figure 1 uses the approximate SD (~ 17.5) and coefficient of repeatability (~ 15) observed for the AFEQT overall scale at 90% as a starting point so that scenarios under which the reliability is increased or the SD is decreased can be examined. As seen in the plot, at 0.90 reliability the SD must drop to about 5 for the coefficient of repeatability to equal the MIC. If the reliability were 0.99, SDs under 17.5 would result in a coefficient of repeatability at or less than the MIC. This example demonstrates the types of conditions required for an instrument’s coefficient of repeatability to equal its MIC. Many instruments will not achieve such low SD’s or high reliabilities under any circumstances. Note that the coefficient of repeatability does not provide information about the amount of change that is meaningful to an individual.
Fig. 1

Needed change in measurement error parameters for coefficient of repeatability equal to MIC on Atrail fibrillation effect on QuailTy-of-Life (AFEQT)

Needed change in measurement error parameters for coefficient of repeatability equal to MIC on Atrail fibrillation effect on QuailTy-of-Life (AFEQT)

Statistically significant and meaningful individual change

A clinician or researcher might also regard relative standing on the measure at the follow-up time point to be important. In some areas of medicine, change in clinical status alone is enough to be important. For example, COVID-19 patients who changed to a more positive level on a six-point ordinal scale (not hospitalized; hospitalized but not requiring supplemental oxygen; hospitalized, requiring supplemental oxygen; hospitalized, requiring nasal high-flow oxygen therapy, non-invasive ventilation, or both; hospitalized, requiring invasive mechanical ventilation, extracorporeal membrane oxygenation, or both; dead) were regarded as improved in one study [24]. Or, a primary care physician might be interested in whether a patient ends up within the normal blood pressure range following initiation of high blood pressure medicine. Similarly, a rehabilitation clinician might want to know if a patient with impaired physical functioning at the beginning of treatment ends up functioning as well as other people with a similar condition. The FDA has suggested that meaningful change needs to be assessed in addition to significant individual change [1]. Some contend that any individual change that is significant at p < 0.05 is substantial and likely to be meaningful to patients [10, 25]. Several years ago, Jacobson and Truax [26] made suggestions for how both significant and clinically meaningful change might be used together. They classified change as (1) “recovered” (statistically significant and clinically significant); (2) improved (statistically significant but not clinically significant); (3) unchanged (not statistically significant), and (4) deteriorated (statistically significant decrement). In one study, responders were those with significant individual improvement on the Functional Disability Inventory (FDI) and improvement in the FDI severity level (no/minimal disability, moderate disability, severe disability) [27]. These change categories offered by Jacobson and Truax may be more appealing than use of either statistically significant change (coefficient of repeatability) or the MIC alone.

Secondary analysis considering significant and meaningful individual change

To illustrate how significant individual change and meaningful individual change can be presented together, we conduct a secondary analysis of the Impact Stratification Score (ISS) administered in a prospective comparative effectiveness clinical trial of 750 active-duty U.S. military personnel [28]. The average age of the sample was 31; 76% were males and 67% white. Most of the participants reported low back pain for more than 3 months. The ISS was proposed for use with chronic low back pain patients by a National Institutes of Health Pain Consortium research task force. The ISS is the sum of the PROMIS-29 v2.1 physical function, pain interference and pain intensity scores [29]. The ISS has a possible range of 8 (least impact) to 50 (greatest impact). Physical function (4 items with response options ranging from without any difficulty = 1 to unable to do = 5) and pain interference (4 items with response options ranging from not at all = 1 to very much = 5) each contribute from 4 to 20 points, and the pain intensity item contributes from 0 to 10 points. The task force proposed three categories of ISS severity: 8–27 (mild), 28–34 (moderate), and 35–50 (severe). Following guidelines by de Vet et al. [30], Dutmer et al. [7] estimated a SEM of 5.2 for the ISS based on test–retest reliability. But test–retest reliability estimates can be problematic. Test–retest reliability can underestimate reliability when there is true underlying change. Reeve et al. [31] noted that: ISOQOL respondents agreed that as a minimum standard a multi-item PRO measure should be assessed for internal consistency reliability… However, they did not support as a minimum standard that a multi-item PRO measure should be required to have evidence of test–retest reliability. They noted practical concerns regarding test–retest reliability; primarily that some populations studied in PCOR are not stable and that their HRQOL can fluctuate. This phenomenon would reduce estimates of test–retest reliability, making the PRO measure look unreliable when it may be accurately detecting changes over time. In addition, memory effects will positively influence the test–retest reliability when the two survey points are scheduled close to each other (p. 1895). We estimated a much smaller SEM of 2.4 using an internal consistency reliability estimate from another study [28]. In this dataset, we examine significance of individual change on the ISS between baseline and 6 weeks later using the coefficient of repeatability (= 6.6). In addition, we compare the significance of change with self-reports on a retrospective rating of change item administered at 6 months: “Compared to your first visit, your low back pain is: much worse, a little worse, about the same, a little better, moderately better, much better or completely gone”? Thirty-seven percent of the sample improved significantly on the ISS over these 6 weeks and 59% reported on the retrospective change item that they were better (16% a little better, 14% moderately better, 23% much better, and 6% completely gone). Among those who improved significantly on the ISS, 89% reported they were better on the retrospective rating item. Thirty-three percent of the sample improved significantly and reported improvement on the retrospective change item (statistically and clinically meaningful), 4% improved significantly but did not report that they were better on the retrospective change item (statistically but not clinically meaningful), 26% did not improve significantly but reported improvement on the change item, and 37% did not improve significantly or report improvement on the change item. Extending this application to further illustrate how group-based methods of estimating MICs can underestimate significant individual change, we compared two alternative ways of defining improvement on a retrospective rating of change item to identify optimal cut points on the ISS. The first way is more inclusive in that improvement from baseline to 6 weeks later included those who reported on the retrospective change item at 6 weeks that their back pain was either a little better, moderately better, much better or completely gone. The second way is more restrictive as improvement was limited to those who reported their back pain was moderately better, much better or completely gone on the retrospective change item. The Youden index [32], (sensitivity + specificity) −1, suggested an optimal cut point of 5 points for change on ISS from baseline to 6 weeks later for the first definition of improvement: sensitivity of 65%, specificity of 82%, negative predictive value of 62%, and positive predictive value of 84%. For the second definition of improvement, the Youden index indicated an optimal cut point of 7 points for ISS change: sensitivity of 66%, specificity of 85%, negative predictive value of 77%, and positive predictive value of 76%. The group-level thresholds estimated for the second definition that excluded those who said they were a little better from the improvement group were closer to the coefficient of repeatability.

Discussion

In contrast to significant group-level change that can be trivial in magnitude if the sample size is large, significant individual change is substantial and worth noting regardless of whether the patient reports that they have improved. As suggested by Jacobson and Truax [26], researchers and clinicians may also be interested in whether those who have significantly improved on a HRQOL measure perceive that they have done so. One can separate people who improved significantly and report at time 2 that they have improved since time 1 from those who do not perceive they have improved. One could also note who reaches a desirable status such as becoming symptom free or ending up within the “normal” range at time 2. Using group-level estimates of meaningful change (group means) to classify individuals as responders to treatment is inappropriate. Doing so results in overoptimistic estimates of the number of people who improve (i.e., too many will be classified as improved). Ironically, a MIC estimate might yield similar numbers of responders as individual-level significance tests if the estimate erroneously includes those who changed by more than a minimally important amount [33]. In our secondary analysis we observed that the optimal cut-point on the ISS using one way of classifying improvement (i.e., those who reported that they were moderately better, much better or their back pain was completely gone) over 6 weeks was similar to the coefficient of repeatability for individual change. But including people who felt they were a little better as improvers resulted in an overoptimistic number of responders. Future work is needed to investigate whether group-level threshold estimates based on retrospective ratings of more than a little improvement converge with appropriate individual-level significance tests. A fundamental criterion for a responder is that the individual improves significantly (i.e., individual change is greater than estimated measurement error). Individual-level statistical indices such as the RCI or the equivalent coefficient of repeatability have been available for decades. These or parallel item response theory approaches [15, 34] that allow reliability to vary across the true score continuum need to be used to determine if patients have stayed the same, deteriorated, or improved. Some may argue for using a significance level other than p < 0.05 to identify individual change that doesn’t meet the conventional cutoff used for group-level comparisons. Individual differences may be important even if they do not equal or exceed the conventional 0.05 significance level. One possible strategy is to use a combination of one-tailed and two-tailed tests of significance and report five levels of change: definitely worse (two-tailed), probably worse (one-tailed), same (one-tailed), probably better (one-tailed), and definitely better (two-tailed). This classification preserves more information and, therefore, helps to address to some extent concerns about missing noteworthy individual change Others might favor even more liberal significance levels to capture more potential responders. Indeed, Donaldson [20] entertained focusing on likely instead of unlikely values and classifying individuals into categories such as: almost certainly changed, quite likely changed, and probably stayed the same. Others [35] have suggested using a Bayes factor to indicate the evidence in the data for or against true change but note that it requires at least three data points. More work is needed about what cutoffs should be used to identify important individual change. Variance between individuals was used in previous research and in the examples in this paper. As noted earlier, it would be ideal if individual-level variation was available, but the number of observations required to do this is prohibitive for most research studies [8]. The SD of change is more consistent epistemologically with the assessment of individual change and is analogous to the denominator used for the standardized response mean and the responsiveness statistic when evaluating responsiveness of measures [36]. Future research is needed that compares SD between individuals with the SD of change. In addition, different ways of identifying meaningful change need to be examined and group-level threshold cutoff approaches (e.g., area under the curve) compared with use of individual ratings of change. We recommend that investigators conducting clinical trials and observational studies routinely report responders to treatment using the significance of individual change. Presenting individual statistical significance and whether the individual feels that they have improved together could be especially useful. We also suggest that clinicians evaluate statistically significance and meaningful change in their patients.
  32 in total

Review 1.  Principles of defining reliable change indices.

Authors:  G H Maassen
Journal:  J Clin Exp Neuropsychol       Date:  2000-10       Impact factor: 2.475

2.  Combining clinical relevance and statistical significance for evaluating quality of life changes in the individual patient.

Authors:  C B Terwee; B Terluin; D L Knol; H C W de Vet
Journal:  J Clin Epidemiol       Date:  2011-12       Impact factor: 6.437

Review 3.  Evidence-based indicators of neuropsychological change in the individual patient: relevant concepts and methods.

Authors:  Kevin Duff
Journal:  Arch Clin Neuropsychol       Date:  2012-02-29       Impact factor: 2.813

4.  Evaluating the statistical significance of health-related quality-of-life change in individual patients.

Authors:  Ron D Hays; Marc Brodsky; M Francis Johnston; Karen L Spritzer; Ka-Kit Hui
Journal:  Eval Health Prof       Date:  2005-06       Impact factor: 2.651

5.  Identifying treatment responders and predictors of improvement after cognitive-behavioral therapy for juvenile fibromyalgia.

Authors:  Soumitri Sil; Lesley M Arnold; Anne Lynch-Jordan; Tracy V Ting; James Peugh; Natoshia Cunningham; Scott W Powers; Daniel J Lovell; Philip J Hashkes; Murray Passo; Kenneth N Schikler; Susmita Kashikar-Zuck
Journal:  Pain       Date:  2014-03-17       Impact factor: 6.961

6.  Methodological problems in the retrospective computation of responsiveness to change: the lesson of Cronbach.

Authors:  G R Norman; P Stratford; G Regehr
Journal:  J Clin Epidemiol       Date:  1997-08       Impact factor: 6.437

7.  Report of the NIH Task Force on research standards for chronic low back pain.

Authors:  Richard A Deyo; Samuel F Dworkin; Dagmar Amtmann; Gunnar Andersson; David Borenstein; Eugene Carragee; John Carrino; Roger Chou; Karon Cook; Anthony DeLitto; Christine Goertz; Partap Khalsa; John Loeser; Sean Mackey; James Panagis; James Rainville; Tor Tosteson; Dennis Turk; Michael Von Korff; Debra K Weiner
Journal:  J Pain       Date:  2014-04-29       Impact factor: 5.820

8.  Can Methods Developed for Interpreting Group-level Patient-reported Outcome Data be Applied to Individual Patient Management?

Authors:  Madeleine T King; Amylou C Dueck; Dennis A Revicki
Journal:  Med Care       Date:  2019-05       Impact factor: 2.983

9.  Minimum clinically important difference of the health-related quality of life scales in adult spinal deformity calculated by latent class analysis: is it appropriate to use the same values for surgical and nonsurgical patients?

Authors:  Selcen Yuksel; Selim Ayhan; Vugar Nabiyev; Montse Domingo-Sabat; Alba Vila-Casademunt; Ibrahim Obeid; Francisco Sanchez Perez-Grueso; Emre Acaroglu
Journal:  Spine J       Date:  2018-08-27       Impact factor: 4.166

10.  A linear prognostic score based on the ratio of interleukin-6 to interleukin-10 predicts outcomes in COVID-19.

Authors:  Oliver J McElvaney; Brian D Hobbs; Dandi Qiao; Oisín F McElvaney; Matthew Moll; Natalie L McEvoy; Jennifer Clarke; Eoin O'Connor; Seán Walsh; Michael H Cho; Gerard F Curley; Noel G McElvaney
Journal:  EBioMedicine       Date:  2020-10-08       Impact factor: 8.143

View more
  2 in total

1.  Inferring meaningful change in quality of life with posterior predictive distribution: an alternative to standard error of measurement.

Authors:  Yuelin Li
Journal:  Qual Life Res       Date:  2022-09-09       Impact factor: 3.440

2.  Comparative responsiveness and minimally important difference of Fatigue Symptom Inventory (FSI) scales and the FSI-3 in trials with cancer survivors.

Authors:  Catherine E Mosher; Ekin Secinti; Shelley A Johns; Kurt Kroenke; Laura Q Rogers
Journal:  J Patient Rep Outcomes       Date:  2022-07-23
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.