Literature DB >> 25652323

Meta-analysis of test accuracy studies: an exploratory method for investigating the impact of missing thresholds.

Richard D Riley1, Ikhlaaq Ahmed2, Joie Ensor3, Yemisi Takwoingi3, Amanda Kirkham3, R Katie Morris4, J Pieter Noordzij5, Jonathan J Deeks3.   

Abstract

BACKGROUND: Primary studies examining the accuracy of a continuous test evaluate its sensitivity and specificity at one or more thresholds. Meta-analysts then usually perform a separate meta-analysis for each threshold. However, the number of studies available for each threshold is often very different, as primary studies are inconsistent in the thresholds reported. Furthermore, of concern is selective reporting bias, because primary studies may be less likely to report a threshold when it gives low sensitivity and/or specificity estimates. This may lead to biased meta-analysis results. We developed an exploratory method to examine the potential impact of missing thresholds on conclusions from a test accuracy meta-analysis.
METHODS: Our method identifies studies that contain missing thresholds bounded between a pair of higher and lower thresholds for which results are available. The bounded missing threshold results (two-by-two tables) are then imputed, by assuming a linear relationship between threshold value and each of logit-sensitivity and logit-specificity. The imputed results are then added to the meta-analysis, to ascertain if original conclusions are robust. The method is evaluated through simulation, and application made to 13 studies evaluating protein:creatinine ratio (PCR) for detecting proteinuria in pregnancy with 23 different thresholds, ranging from one to seven per study.
RESULTS: The simulation shows the imputation method leads to meta-analysis estimates with smaller mean-square error. In the PCR application, it provides 50 additional results for meta-analysis and their inclusion produces lower test accuracy results than originally identified. For example, at a PCR threshold of 0.16, the summary specificity is 0.80 when using the original data, but 0.66 when also including the imputed data. At a PCR threshold of 0.25, the summary sensitivity is reduced from 0.95 to 0.85 when additionally including the imputed data.
CONCLUSIONS: The imputation method is a practical tool for researchers (often non-statisticians) to explore the potential impact of missing threshold results on their meta-analysis conclusions. Software is available to implement the method. In the PCR example, it revealed threshold results are vulnerable to the missing data, and so stimulates the need for advanced statistical models or, preferably, individual patient data from primary studies.

Entities:  

Keywords:  Diagnostic test; Imputation; Meta-analysis; Missing data; Multiple thresholds; Sensitivity analysis

Mesh:

Year:  2015        PMID: 25652323      PMCID: PMC4417327          DOI: 10.1186/2046-4053-4-12

Source DB:  PubMed          Journal:  Syst Rev        ISSN: 2046-4053


Background

Medical tests are used to inform screening, diagnosis and prognosis in medicine. Meta-analysis methods are increasingly used to synthesise the evidence about a test’s accuracy from multiple studies, to produce summary estimates of sensitivity and specificity [1-4]. When the test is measured on a continuous scale, many studies report test performance at multiple thresholds, each relating to a different choice of threshold above which test results are classed as ‘positive’ and below which test results are classed as ‘negative’. Unfortunately, most primary studies do not report the same set of thresholds. For example, in an evaluation of the spot protein:creatinine ratio (PCR) for detecting significant proteinuria in pregnancy, Morris et al. [5] extracted tables for 23 different thresholds across 13 studies; eight of the thresholds were considered by just one study, but the other 15 thresholds were considered in two or more studies (Table 1), with a maximum of six studies for any threshold. In this situation, meta-analysts generally either utilise the results for just one of the thresholds per study or utilise results for all reported thresholds but perform a separate meta-analysis for each of the thresholds independently [6]. However, an approach that considers meta-analysis for each threshold independently will omit any studies that do not report the threshold of interest and thus also ignore information from other thresholds that are available in those studies.
Table 1

PCR results for each threshold in each of the 13 studies of Morris et al. [5]

First authorThreshold ID, t Threshold value, x TPFPFNTNTotalHigh proteinuriaNormal proteinuria
Al Ragib10.13355149518539146
60.1833426104
70.1933396107
80.231388108
220.49292310123
Durnwald30.1515635121722016852
80.2152271625
150.3136233229
190.39123144538
200.4120124840
230.510696243
Dwyer30.1554282321165660
50.175125535
70.195018642
120.244181552
140.283731957
190.393102560
Leonas150.327775638927282645
Ramos230.5251120472621
Robert150.3274238712942
Rodriguez20.1469340351386969
30.156834135
40.166826143
50.176525444
60.186224745
70.196221748
80.26019950
90.216017952
Saudan80.214270591001486
130.251314172
150.3137179
180.35124282
200.4113383
210.45100486
Schubert30.1593031596
40.169204
Shahbazian80.2352341813843
Taherian20.146776201007327
30.15673624
40.16651826
50.17641926
60.186301027
80.25901427
Wheeler90.2159139451266858
Yamasmit70.1929607422913
90.2129508
100.2229409
110.23283110
120.24282111
130.25281112
140.28271212
160.31261312
170.32251412

ID ordered identification number, TP true positives, FP false positives, TN true negatives, FN false negatives.

PCR results for each threshold in each of the 13 studies of Morris et al. [5] ID ordered identification number, TP true positives, FP false positives, TN true negatives, FN false negatives. In this article, we propose an exploratory method (a sensitivity analysis) to help researchers examine the potential impact of missing thresholds on their meta-analysis conclusions about a test’s accuracy. The method first imputes results in studies where missing thresholds are bounded between a pair of known thresholds; missing results are also bounded because as the threshold value increases, sensitivity must decrease and specificity must increase. The imputed results are then added to the meta-analysis, and this allows researchers to evaluate whether their original conclusions are robust. This is especially important when thresholds are prone to selective reporting bias; that is, they are less likely to be reported when they give lower values of sensitivity and/or specificity. In this situation, meta-analysis may otherwise produce summary sensitivity and specificity results that are too high (i.e. biased). The article is structured as follows. In the “Motivating example” section, we describe the motivating PCR dataset in detail. In the “Methods” section, we describe our imputation method, explain its assumptions and perform an empirical evaluation. The “Results” section applies it to the PCR data, and the “Discussion” section concludes by considering the strengths and limitations of the method and further research.

Motivating example: identification of significant proteinuria in patients with suspected pre-eclampsia

Pre-eclampsia is a major cause of maternal and perinatal morbidity and mortality and occurs in 2%–8% of all pregnancies [7-10]. The diagnosis of pre-eclampsia is determined by the presence of elevated blood pressure combined with significant proteinuria (≥0.3 g per 24 h) after the 20th week of gestation in a previously normotensive, non-proteinuric patient [11]. The gold-standard method for detection of significant proteinuria is the 24-h urine collection, but this is cumbersome, time consuming and inconvenient, to patients as well as hospital staff. There is therefore a need for a rapid and accurate diagnostic test to identify significant proteinuria to allow more timely decision-making. The spot PCR has been shown to be strongly correlated with 24-h protein excretion and thus is a potential diagnostic test for significant proteinuria. Morris et al. [5] performed a systematic review and meta-analysis to assess the diagnostic accuracy of PCR for the detection of significant proteinuria in patients with suspected pre-eclampsia. Thirteen relevant studies were identified, and in each study, the reference standard was proteinuria greater than or equal to 300 mg in urine over 24 h. Across the 13 studies, 23 different threshold values were considered for PCR, ranging from 0.13 to 0.50. Five studies provided diagnostic accuracy results (i.e. a two-by-two table showing the number of true positives, false positives, false negatives and true negatives) for just one threshold, but the other eight studies reported results for each of multiple thresholds, up to a maximum of nine thresholds (Yamasmit study). Eight of the 23 thresholds were considered by just one study, but the other 15 thresholds were considered in two or more studies, up to a maximum of six studies (for threshold 0.20). The studies and thresholds are summarised in Table 1. Meta-analysis is important here to summarise the diagnostic accuracy of PCR at each threshold from all the published evidence, to help ascertain whether PCR is a useful diagnostic test and, if so, which threshold is the most appropriate to use in clinical practice. However, this is non-trivial given the number of thresholds available, the variation in how many studies report each threshold and the likely similarity between neighbouring threshold results. The PCR data is thus an ideal dataset to motivate and apply the statistical methods developed during the remainder of the paper.

Methods

We now propose our exploratory method for examining the impact of missing thresholds in meta-analysis of test accuracy studies.

Exploratory method to examine the potential impact of missing thresholds

Let there be i = 1 to m studies that measure a continuous test result on n1 diseased patients and n0 non-diseased patients, whose true disease status is provided by a reference standard. In each study, at a particular threshold value, x, each patient’s measured test value is classed as either ‘positive’ (≥ x) or ‘negative’ ( When results for a particular threshold are missing but other thresholds above and below are available, then the missing threshold has sensitivity and specificity results constrained between these values. For example, consider the Al Ragib study (Table 1), which has threshold values of 0.13 and 0.18 available, but not 0.14 to 0.17. The number of true positives for the missing thresholds must be constrained between the other threshold values of 35 and 33. Similarly, the missing false positives must be within 42 and 51, the missing false negatives within 4 and 6 and the missing true negatives within 95 and 104. Rather than ignoring missing thresholds that are bounded between known thresholds, our exploratory method imputes the missing results under particular assumptions, so that they can be included in the meta-analysis. The aim is to ascertain whether the original meta-analysis conclusions (obtained without imputed data) are robust to the inclusion of imputed data. For example, does the summary test accuracy at each threshold remain similar, and is the choice of best threshold the same? The exploratory method is a two-step approach, as now described.

Step 1: imputation of missing bounded thresholds in each study

In each study separately, for each threshold that is missing but bounded between known thresholds, the missing results are imputed by assuming each 1-unit increase in threshold value corresponds to a constant reduction in logit-sensitivity (y1), and also a constant increase in logit-specificity (y0). Thus, imputation is on a straight line between pairs of observed points in logit receiver operating characteristic (ROC) space. This piece-wise linear approach is illustrated graphically in Figure 1. So the key assumption here is a constant change in logit values for each 1-unit change in threshold value between each pair of observed threshold results. The linear slope can be different between each pair of thresholds, and so no single trend is assumed across all thresholds, with the fitted lines forced to go through the observed points. Linear relationships on the logit scale are often used in diagnostic test analyses when considering the ROC curve, especially in meta-analysis [12], and it is a straightforward approach for this exploratory analysis.
Figure 1

Graphical illustration of the imputation approach for the Al Ragib study.

Graphical illustration of the imputation approach for the Al Ragib study. Once the imputed logit values are obtained, one can back transform to compute the corresponding imputed true and false positives and negatives. For example, let TP1 be the true positive number at a threshold value of 0.13 and TP2 be the true positive number at threshold 0.18. Then, in the Al Ragib study there are 5 threshold units from 0.13 to 0.18. The imputed logit-sensitivity at threshold 0.14 is y1 + (y1 - y1)/5, and at threshold 0.15 the imputed logit-sensitivity is y1 + (2(y1 - y1)/5), and so on (Table 2, Figure 1). It is then straightforward to calculate the number of true positives, false positives, true negatives and false negatives that are necessary to produce these values. For example, for threshold 0.15, the imputed logit-sensitivity is 1.983, and so the imputed sensitivity is 0.879, and given there are 39 patients truly with high proteinuria, the imputed number of true positives and false negatives is 34.28 and 4.72, respectively (Table 2).
Table 2

Actual and imputed results for the Al Ragib study between thresholds 0.13 and 0.18

First authorThreshold ID, t Threshold value, x Imputed?TPFPFNTNTotalHigh proteinuriaNormal proteinuria
Al Ragib10.13No355149518539146
20.14Yes34.749.14.396.9
30.15Yes34.347.34.798.7
40.16Yes33.945.55.1100.5
50.17Yes33.543.75.5102.3
60.18No33426104

The imputation is undertaken on the logit-scale, and then the values are back transformed to calculate the corresponding imputed raw data values.

ID ordered identification number, TP true positives, FP false positives, TN true negatives, FN false negatives.

Actual and imputed results for the Al Ragib study between thresholds 0.13 and 0.18 The imputation is undertaken on the logit-scale, and then the values are back transformed to calculate the corresponding imputed raw data values. ID ordered identification number, TP true positives, FP false positives, TN true negatives, FN false negatives. Note that we neither impute beyond the highest available threshold nor impute below the lowest available threshold in each study. Further assumptions would be necessary to do this, but here we only work within the limits of the observed data available. Thus, for studies with only 1 threshold reported, no imputation was used. Similarly, we do not impute for any new threshold values which were not considered by any of the available studies. A STATA ‘do’ file to fit the imputation method is available in Additional file 1, and we aim to release an associated STATA module in the near future. It provides the original and imputed values within a few seconds, for any number of studies and any number of thresholds.

Step 2: meta-analysis at each threshold separately using actual and imputed data

The imputation in step 1 borrows strength from available thresholds to allow a larger set of threshold data to be available from each study for meta-analysis. For ease of language, let us order the thresholds of interest and refer to the ordered value as t (e.g. t = 1 to 23 in the PCR example, Table 1). Each threshold t now has (i) one or more studies with observed results and potentially (ii) some studies with imputed results. A separate meta-analysis of each threshold separately can now be considered, using the observed and imputed results. A convenient model is the bivariate meta-analysis of Chu and Cole [2]. This approach is recommended by the Cochrane Screening and Diagnostic Test Methods Group and is commonly used in diagnostic accuracy meta-analyses. It utilises the exact binomial within-study distribution, thereby avoiding the need for any continuity corrections, and accounts for any between-study correlation in sensitivity and specificity, as follows: β1 and β0 give the average logit-sensitivity and average logit-specificity at threshold t, respectively, and these can be transformed to give the summary sensitivity and summary specificity from the meta-analysis for each threshold. The between-study covariance matrix is given by Ω, containing the between-study variances ( and ) and the between-study correlation in logit-sensitivity and logit-specificity (ρ10); if the latter is zero, the model reduces to a separate univariate analysis for each of sensitivity and specificity. Indeed, ρ10 will often be poorly estimated at +1 or -1 [13], and so it may be sensible to adopt two separate univariate models here [14], as follows: Models (1) and (2) can be estimated using adaptive Gaussian quadrature [15], for example using PROC NLMIXED in SAS [16], or the xtmelogit command in STATA [17]. A number of quadrature points can be specified, with increasing estimation accuracy as the number of points increases, but at the expense of increased computational time. We generally chose 5 quadrature points for our analyses, as this gave estimates very close to those when using >10 points but in a faster time. Successful convergence of the optimization procedure was assumed when successive iteration estimates differed by <10-7, resulting in parameter estimates and their approximate standard errors based on the second derivative matrix of the likelihood function.

Empirical evaluation of the imputation method

To empirically evaluate the imputation method, we utilised individual participant data (IPD) from six studies examining the ability of parathyroid hormone (PTH) to correctly classify which patients will become hypocalcemic within 24-h after a thyroidectomy [18]. The percentage decrease in PTH (from pre-surgery to 6-h post-surgery) was used as the test, and thresholds of 40%, 50%, 60%, 65%, 70%, 80% and 90% were examined. As IPD were available, the results for all thresholds were available for all studies. Thus, for each threshold separately, we could fit model (1) using the complete set of data from the six studies. However, model (1) often poorly estimated the between-study correlations at +1 or -1 and gave summary test accuracy results very similar to model (2). Thus we focus here only on model (2) results, and these provided our ‘complete data’ meta-analysis results, for when the thresholds are all truly available from all studies.

Generation of missing thresholds and imputation

To replicate missing data mechanisms, we considered two scenarios: Scenario (i): thresholds missing at random We took the complete set of threshold results for each study and randomly assigned some to be missing, with each having a 0.5 probability of being omitted. This provided a new meta-analysis dataset of up to six studies with missing threshold results. We repeated this process until 1,000 such meta-analysis datasets had been produced. In each dataset, we applied our imputation approach, and then for each threshold, we fitted model (2) to (i) each generated dataset without including the imputed results and (ii) each generated dataset with the addition of the imputed results. We then compared the average meta-analysis estimates and standard errors from the 1,000 analyses of (i) and (ii) with the true meta-analysis results when the complete data were available (Table 3).
Table 3

Empirical evaluation results—scenario (i), thresholds missing at random

% PTH decreaseEstimate of interestMeta-analysis of the complete dataMeta-analysis of the datasets with missing threshold results, not including imputed resultsMeta-analysis of the datasets with missing threshold results, including imputed results
True estimateMean estimate across the 1,000 datasetsMedian estimate across the 1,000 datasetsMean estimate across the 1,000 datasetsMedian estimate across the 1,000 datasets
40Summary Sensitivity0.870.880.900.880.90
0.000.040.000.040.00
s.e. (logit sensitivity)0.400.700.620.700.62
Summary specificity0.520.520.520.520.52
0.000.050.000.050.00
s.e. (logit specificity)0.180.250.240.250.24
50Summary sensitivity0.870.880.900.870.87
0.000.040.000.040.00
s.e. (logit sensitivity)0.400.690.620.540.47
Summary specificity0.620.630.630.640.63
0.000.050.010.040.00
s.e. (logit specificity)0.180.270.260.220.21
60Summary sensitivity0.870.880.900.870.86
0.000.040.000.050.00
s.e. (logit sensitivity)0.400.690.620.520.43
Summary specificity0.780.780.790.780.77
0.000.050.000.050.00
s.e. (logit specificity)0.210.310.290.240.23
65Summary sensitivity0.850.870.890.850.85
0.000.040.000.050.00
s.e. (logit sensitivity)0.380.690.620.490.39
Summary specificity0.800.810.820.810.80
0.000.050.000.040.00
s.e. (logit specificity)0.220.330.320.260.23
70Summary sensitivity0.850.860.880.830.84
0.000.070.000.050.00
s.e. (logit sensitivity)0.380.680.640.470.39
Summary specificity0.850.860.860.850.85
0.000.040.000.040.00
s.e. (logit specificity)0.250.360.350.300.26
80Summary sensitivity0.730.730.760.720.72
0.000.040.000.050.00
s.e. (logit sensitivity)0.300.490.450.380.35
Summary specificity0.910.920.920.910.91
0.000.040.000.040.00
s.e. (logit specificity)0.300.500.460.410.35
90Summary sensitivity0.550.510.550.510.55
0.000.050.000.050.00
s.e. (logit sensitivity)0.270.460.360.460.36
Summary specificity0.960.960.960.960.96
0.000.040.000.040.00
s.e. (logit specificity)0.460.690.590.690.59

NB All meta-analyses used model (2), as model (1) often poorly estimated ρ 10 as +1 or -1.

Empirical evaluation results—scenario (i), thresholds missing at random NB All meta-analyses used model (2), as model (1) often poorly estimated ρ 10 as +1 or -1. Scenario (ii): thresholds selectively missing We took the complete set of threshold results for each study and assigned some to be missing through a selective (not missing at random) mechanism, based on the observed sensitivity estimate. All thresholds with the observed sensitivity ≥ 90% were always included; however, those with sensitivity <90% had a 0.5 probability of being omitted. This reflects a realistic situation where researchers are more likely to report those thresholds where sensitivity is observed to be high. We repeated this process until 1000 such meta-analysis datasets had been produced. In each dataset, we applied our imputation approach, and then for each threshold, we fitted model (2) to (i) each generated dataset without including the imputed results and (ii) each generated dataset with the addition of the imputed results. We then compared the average meta-analysis estimates and standard errors from the 1,000 analyses of (i) and (ii) with those true meta-analysis results when the complete data were available (Table 4).
Table 4

Empirical evaluation results—scenario (ii), thresholds selectively missing

% PTH decreaseEstimate of interestMeta-analysis of the complete dataMeta-analysis of the datasetswith missing threshold results, not including imputed resultsMeta-analysis of the datasets with missing threshold results, including imputed results
True estimateMean estimate across the 1,000 datasetsMedian estimate across the 1,000 datasetsMean estimate across the 1,000 datasetsMedian estimate across the 1,000 datasets
40Summary sensitivity0.870.890.880.890.88
0.000.000.000.000.00
s.e. (logit sensitivity)0.400.520.480.520.48
Summary specificity0.520.520.520.520.52
0.000.050.040.050.04
s.e. (logit specificity)0.180.250.240.250.24
50Summary sensitivity0.870.890.900.880.87
0.000.000.000.010.00
s.e. (logit sensitivity)0.400.530.520.460.44
Summary specificity0.620.640.640.640.64
0.000.040.040.010.00
s.e. (logit specificity)0.180.260.260.210.20
60Summary sensitivity0.870.890.900.870.87
0.000.000.000.030.00
s.e. (logit sensitivity)0.400.520.480.440.40
Summary specificity0.780.790.790.770.77
0.000.030.000.010.00
s.e. (logit specificity)0.210.280.260.230.21
65Summary sensitivity0.850.870.860.850.85
0.000.060.000.040.00
s.e. (logit sensitivity)0.380.540.490.410.38
Summary specificity0.800.820.820.810.80
0.000.010.000.000.00
s.e. (logit specificity)0.220.310.280.240.23
70Summary sensitivity0.850.870.860.830.84
0.000.060.000.080.00
s.e. (logit sensitivity)0.380.540.490.410.38
Summary specificity0.850.860.850.840.84
0.000.000.000.000.00
s.e. (logit specificity)0.250.330.310.260.25
80Summary sensitivity0.730.750.750.710.71
0.000.020.000.050.00
s.e. (logit sensitivity)0.300.450.410.370.34
Summary specificity0.910.900.900.910.91
0.000.000.000.020.00
s.e. (logit specificity)0.300.410.390.360.35
90Summary sensitivity0.550.580.560.580.56
0.000.020.000.020.00
s.e. (logit sensitivity)0.270.410.370.410.37
Summary specificity0.960.950.960.950.96
0.000.000.000.000.00
s.e. (logit specificity)0.460.570.590.570.59

NB All meta-analyses used model (2), as model (1) often poorly estimated ρ 10 at +1 or -1.

Empirical evaluation results—scenario (ii), thresholds selectively missing NB All meta-analyses used model (2), as model (1) often poorly estimated ρ 10 at +1 or -1.

Results

Results of empirical evaluation

In the empirical evaluation, there is no imputation for the lowest and highest threshold of 40% and 90%, and so meta-analysis results are identical for these thresholds regardless of whether imputed data is included or not. However, for other thresholds, there is always potential for imputation. For thresholds between 40% and 90% in scenario (i), where thresholds were missing at random, the mean and median estimates tend to be slightly closer to the complete data results when using the imputed data (Table 3). For example, at threshold 65%, the true sensitivity estimate from complete data is 0.85, whilst the mean/median without using imputed data is 0.87/0.89, but when using the imputed data, it is pulled back to 0.85/0.85. Further, the meta-analyses including the imputed data give substantially smaller standard errors than those from the meta-analyses excluding imputed data. For example, for threshold 65%, the standard error of the summary logit-sensitivity is 0.69 when ignoring imputed data and 0.49 when including it, a gain in precision of almost 30%. The gain in standard error reflects the additional information being used from the imputed results. As estimates are close to the true estimates and standard errors are considerably reduced, the mean-square error of estimates is therefore improved. The imputation approach also gives standard errors of estimates that are also closer (but not smaller) to those from the true complete data meta-analysis. For scenario (ii), where thresholds are selectively missing based on the value of observed sensitivity, the summary meta-analysis results without the imputed data are again slightly larger than the true estimated values, especially for sensitivity. The meta-analysis results using the imputed data generally reduce this bias and give more conservative estimates. For example, for the 50% threshold, the true estimated value for sensitivity is 0.87, whilst the median meta-analysis estimate without imputation data is 0.90, but the median estimate including imputed data is pulled back down to 0.87. Occasionally, the imputation method over-adjusted, so that the estimate was pulled down too far. For example, for the 80% threshold, the true estimated value for sensitivity is 0.73, but the median estimate from the without imputation data is 0.75 and the median estimate from the imputed data is 0.71. However, even here the absolute magnitude of bias is the same (0.02) with and without imputed data. For all thresholds between 40% and 90%, there is again considerable reduction in the standard error of meta-analysis estimates following the use of imputed data. In summary, the empirical evaluation shows that the imputation method performs well, with summary test accuracy estimates generally moved slightly closer to the true estimates based on complete data. This finding, combined with smaller standard errors and thus smaller mean-square error of estimates, suggests the imputation approach is useful as an exploratory analysis.

Application to the PCR example

Our imputation approach was applied to the PCR studies introduced in the “Motivating example” section, and missing threshold results could be imputed in 6 of the 13 studies. For 21 of the 23 different thresholds, the imputation approach increased the number of studies providing data for that threshold (Table 5), and in total, an additional 50 thresholds results were imputed, substantially increasing the information available for meta-analysis. For example, at a PCR threshold of 0.22, the imputation increased the available studies from 1 to 5, and at a threshold of 0.3, the available studies increased from 4 to 7.
Table 5

Summary meta-analysis results following application of model (2) with and without the imputed data included

Without imputed dataWith imputed data
Threshold value, x No. studies with this thresholdSummary estimate95% CITauNo. studies with this thresholdSummary estimate95% CITau
LowerUpper
LowerUpper
Sensitivity
0.1310.8970.7560.9610.00010.8970.7560.9610.000
0.1420.9100.8410.9510.00030.9550.8010.9911.147
0.1550.9440.9010.9690.06760.9370.9090.9570.001
0.1630.9600.8310.9910.85060.9250.8900.9500.091
0.1730.9090.8600.9420.00050.9110.8790.9350.000
0.1830.8730.8160.9140.00050.8940.8600.9200.000
0.1940.9020.8510.9360.00060.8890.8550.9170.096
0.260.8750.8280.9100.23480.8860.8390.9210.312
0.2130.8920.8340.9310.00070.8820.8480.9090.000
0.2210.9830.7820.9990.00050.8990.7610.9610.669
0.2310.9500.7860.9900.00050.8700.7660.9320.534
0.2420.8770.5750.9740.97350.8500.7560.9120.473
0.2520.9530.8320.9880.00050.8500.7590.9100.442
0.2820.8180.5310.9470.83750.8180.7150.8900.473
0.340.9380.8290.9790.97570.8930.7730.9541.076
0.3110.8830.7130.9590.00050.7800.6890.8510.362
0.3210.8500.6750.9390.00050.7810.6870.8530.363
0.3510.8330.5620.9510.00040.7330.6410.8090.296
0.3920.6620.5300.7720.32040.6990.6070.7780.278
0.420.7200.6500.7800.00030.7240.6610.7790.000
0.4510.7000.4360.8760.00030.6910.6270.7480.000
0.4910.8420.7740.8930.00020.6570.5900.7180.000
0.520.8440.4330.9751.23020.8440.4330.9751.230
Specificity
0.1310.6510.5700.7240.00010.6510.5700.7240.000
0.1420.6710.5970.7360.00030.6240.5240.7140.250
0.1550.5620.3660.7400.79560.5830.4210.7280.717
0.1630.8030.4990.9431.02660.6610.4620.8160.910
0.1730.7650.4630.9251.04250.6770.4650.8340.922
0.1830.8560.4360.9791.50350.7260.4580.8931.188
0.1940.7080.6530.7580.00060.7200.5220.8580.961
0.260.8180.5970.9311.24580.7750.6090.8841.031
0.2130.7500.6720.8150.00070.7070.6350.7710.332
0.2210.6920.4090.8800.00050.7050.5990.7930.438
0.2310.7690.4780.9240.00050.7380.6230.8280.508
0.2420.8630.7640.9250.00050.7620.6360.8550.592
0.2520.8480.7640.9070.00050.7980.6550.8920.724
0.2820.9450.8630.9790.00050.8450.6810.9330.947
0.340.9170.7030.9811.51570.9160.7930.9691.311
0.3110.9230.6090.9890.00050.8860.7070.9621.183
0.3210.9230.6090.9890.00050.8920.7160.9641.194
0.3510.9480.8760.9790.00040.8980.7010.9711.239
0.3920.9800.0921.0003.12540.9330.7130.9871.556
0.420.9030.6750.9770.98230.8720.7150.9490.796
0.4510.9940.9151.0000.00030.9440.5870.9951.883
0.4910.8420.7740.8930.00020.8380.7790.8830.000
0.520.8630.7640.9250.00020.8630.7640.9250.000

NB All meta-analyses used model (2), as model (1) often poorly estimated ρ 10 at +1 or -1.

Summary meta-analysis results following application of model (2) with and without the imputed data included NB All meta-analyses used model (2), as model (1) often poorly estimated ρ 10 at +1 or -1. Meta-analysis model (1) was applied to each threshold’s data separately, but the between-study correlations were often estimated poorly as -1 in these analyses, so we decided to rather fit model (2) (i.e. ρ10 was set to zero for all analyses, allowing a separate analysis for sensitivity and specificity at each threshold) [14]. The summary meta-analysis results when including or ignoring the imputed data are shown for each threshold in Table 5 and Figure 2.Importantly, the results when including the imputed data are often very different to when ignoring it. In particular, the summary estimates of sensitivity and specificity are generally reduced when using the imputed data, as can be seen visually in the summary ROC space (Figure 2). For example, when imputed data were included, the summary specificity at a PCR threshold of 0.16 reduced from 0.80 to 0.66 and the summary sensitivity at a PCR threshold of 0.25 reduced from 0.95 to 0.85.
Figure 2

Summary meta-analysis results presented in ROC space, comparing the summary meta-analysis results shown in Table 5 , with and without inclusion of imputed thresholds. To help compare approaches, summary estimates for the same threshold are shown connected.

Summary meta-analysis results presented in ROC space, comparing the summary meta-analysis results shown in Table 5 , with and without inclusion of imputed thresholds. To help compare approaches, summary estimates for the same threshold are shown connected. The points in ROC space tend to move down and to the right after including imputed data, revealing lower sensitivity and specificity than previously thought. Thus, it appears that the results when ignoring imputed data may be optimistic, potentially due to biased availability of thresholds when they give higher test accuracy results. In both analyses (assuming sensitivity and specificity are equally important), the best threshold appears to be between 0.25 and 0.30; however, test accuracy at these thresholds is lower after imputation. The dramatic change in results for some thresholds suggests that individual patient data are needed to obtain a complete set of threshold results from each study and thereby remove the suspected reporting bias in primary studies. We also attempted to use the advanced statistical modelling framework of Hamza et al. [12] to reduce the impact of missing thresholds by jointly synthesising all thresholds in one multivariate model; however, this approach failed to converge, most likely due to the amount of missing data. The multiple thresholds model of Putter et al. [19] also required complete data for all thresholds, whilst the method of Dukic and Gatsonis [20] was not considered suitable, as it produces a summary ROC curve but does not give meta-analysis results for each threshold.

Discussion

Primary study authors often do not use the same thresholds when evaluating a medical test and will predominately report those thresholds that produce the largest (optimal) sensitivity and specificity estimates [21]. This may lead to optimistic and misleading meta-analysis results based only on reported thresholds. We have proposed an exploratory method for examining the impact of missing threshold results in meta-analysis of test accuracy studies and shown its potential usefulness through an applied example and empirical evaluation. The imputation method is applicable when studies use the same (or similarly validated or standardised) methods of measuring a continuous test (e.g. blood pressure or a continuous biomarker, like prostate-specific antigen). It is deliberately very simple, so that applied researchers can still implement standard meta-analysis methods and examine the potential impact of missing thresholds on meta-analysis conclusions. For example, our application to the PCR data showed how the imputation method revealed lower diagnostic test accuracy results than a standard meta-analysis of each threshold independently, but conclusions about the best choice of threshold appeared robust. Other more sophisticated methods are also available to deal with multiple thresholds, but all have limitations. Hamza et al. [12] propose a multivariate random-effect meta-analysis approach and apply it when all studies report all of the thresholds of interest. It models the (linear) relationship between threshold value and test accuracy within each study but is prone to convergence problems (as we experienced for the PCR example), prompting Putter et al. [19] to propose an alternative survival model framework for meta-analysing the multiple thresholds. However, this also requires the multiple thresholds to be available in all studies. Others have also considered the multiple threshold issue [20, 22–27]. A well-known method by Dukic and Gatsonis [20] only produces a summary ROC curve, rather than summary results for each threshold of interest. We recently proposed a multivariate-normal approximation to the Hamza et al. approach [27], which produces both a summary ROC curve and summary results for each threshold and easily accommodates studies with missing thresholds. However, the multivariate-normal approximation to the exact multinomial likelihood is a potential limitation. Our exploratory method is not a competitor to these more sophisticated methods. Rather, it is an exploratory tool aimed at researchers (usually non-statisticians) conducting systematic reviews of test accuracy studies. The method is practical and easy to implement without advanced statistical expertise and so can quickly flag whether researchers should be concerned about missing thresholds in their meta-analysis. This was demonstrated in the PCR example, where the method flagged major concerns that original conclusions were optimistic. In this situation, researchers should be stimulated to put resources toward undertaking the aforementioned advanced statistical methods or, ideally, obtaining individual participant data to calculate missing threshold results directly. The key reason that we label our method as ‘exploratory’ is that it only considers single imputation. Single imputation of missing values usually causes standard errors of estimates to be too small, since it fails to account for the uncertainty in the imputed values themselves, and multiple imputations would help address this [28]. In particular, imputed data between two thresholds close together (e.g. imputing data for a threshold of 0.24 using available thresholds 0.23 and 0.25) should have less uncertainty than imputing data between two thresholds far apart (e.g. imputing at a threshold of 0.24 using thresholds 0.13 and 0.50), but this is not currently accounted for in our approach. Further research may consider extension to multiple imputations. Also, our imputation assumes a linear relationship between threshold value and logit-sensitivity and logit-specificity; although this linear relationship is commonly used in meta-analysis of test accuracy studies, it is of course an assumption. Thus, our imputation method is a sensitivity analysis: it shows, under the assumptions made, how vulnerable the original meta-analysis conclusions are to the missing threshold results. The focus is therefore on how the method modifies the original summary meta-analysis estimates; less attention should be paid to the standard errors and confidence intervals it produces, as these may be artificially small and narrow. The method is thus similar in spirit to how others have evaluated the potential impact of (biased) missing data in meta-analysis of randomised trials, such as trim and fill [29] and adjustments based on funnel plot asymmetry [30]. For example, trim and fill imputes missing studies assuming asymmetry is caused by publication bias and Peters et al.[31] conclude it ‘can help to reduce the bias in pooled estimates, even though the performance of this method is not ideal … we recommend use of the trim and fill method as a form of sensitivity analysis.’ Similarly, our method can help to reduce bias and mean-square error in pooled meta-analysis results.

Conclusion

We have proposed an exploratory analysis that allows researchers to examine the potential impact of missing thresholds on the conclusions of a test accuracy meta-analysis. Currently, most researchers ignore this issue, but our PCR example shows that this may be naive, as conclusions are susceptible to selective threshold reporting in primary studies. STATA code to fit the imputation approach is available in the Additional file 1, and an associated STATA module will be released in the near future. Additional file 1: A STATA ‘do’ file to fit the proposed imputation method. (ZIP 6 KB)
  26 in total

1.  Meta-analysis of ROC curves.

Authors:  A D Kester; F Buntinx
Journal:  Med Decis Making       Date:  2000 Oct-Dec       Impact factor: 2.583

2.  Trim and fill: A simple funnel-plot-based method of testing and adjusting for publication bias in meta-analysis.

Authors:  S Duval; R Tweedie
Journal:  Biometrics       Date:  2000-06       Impact factor: 2.571

3.  A hierarchical regression approach to meta-analysis of diagnostic test accuracy evaluations.

Authors:  C M Rutter; C A Gatsonis
Journal:  Stat Med       Date:  2001-10-15       Impact factor: 2.373

4.  A latent normal distribution model for analysing ordinal responses with applications in meta-analysis.

Authors:  Wai-Yin Poon
Journal:  Stat Med       Date:  2004-07-30       Impact factor: 2.373

5.  Meta-analysis of diagnostic test accuracy assessment studies with varying number of thresholds.

Authors:  V Dukic; C Gatsonis
Journal:  Biometrics       Date:  2003-12       Impact factor: 2.571

6.  Geographic variation in the incidence of hypertension in pregnancy. World Health Organization International Collaborative Study of Hypertensive Disorders of Pregnancy.

Authors: 
Journal:  Am J Obstet Gynecol       Date:  1988-01       Impact factor: 8.661

7.  A general regression methodology for ROC curve estimation.

Authors:  A N Tosteson; C B Begg
Journal:  Med Decis Making       Date:  1988 Jul-Sep       Impact factor: 2.583

8.  Fetal and uteroplacental haemodynamics during short-term atenolol treatment of hypertension in pregnancy.

Authors:  S Montan; H Liedholm; G Lingman; K Marsál; N O Sjöberg; T Solum
Journal:  Br J Obstet Gynaecol       Date:  1987-04

9.  Guidelines for meta-analyses evaluating diagnostic tests.

Authors:  L Irwig; A N Tosteson; C Gatsonis; J Lau; G Colditz; T C Chalmers; F Mosteller
Journal:  Ann Intern Med       Date:  1994-04-15       Impact factor: 25.391

Review 10.  The classification and diagnosis of the hypertensive disorders of pregnancy: statement from the International Society for the Study of Hypertension in Pregnancy (ISSHP).

Authors:  M A Brown; M D Lindheimer; M de Swiet; A Van Assche; J M Moutquin
Journal:  Hypertens Pregnancy       Date:  2001       Impact factor: 2.108

View more
  9 in total

1.  How can we define and analyse drug exposure more precisely to improve the prediction of hospitalizations in longitudinal (claims) data?

Authors:  Andreas D Meid; Andreas Groll; Ulrich Schieborr; Jochen Walker; Walter E Haefeli
Journal:  Eur J Clin Pharmacol       Date:  2016-12-24       Impact factor: 2.953

2.  Selective Cutoff Reporting in Studies of Diagnostic Test Accuracy: A Comparison of Conventional and Individual-Patient-Data Meta-Analyses of the Patient Health Questionnaire-9 Depression Screening Tool.

Authors:  Brooke Levis; Andrea Benedetti; Alexander W Levis; John P A Ioannidis; Ian Shrier; Pim Cuijpers; Simon Gilbody; Lorie A Kloda; Dean McMillan; Scott B Patten; Russell J Steele; Roy C Ziegelstein; Charles H Bombardier; Flavia de Lima Osório; Jesse R Fann; Dwenda Gjerdingen; Femke Lamers; Manote Lotrakul; Sonia R Loureiro; Bernd Löwe; Juwita Shaaban; Lesley Stafford; Henk C P M van Weert; Mary A Whooley; Linda S Williams; Karin A Wittkampf; Albert S Yeung; Brett D Thombs
Journal:  Am J Epidemiol       Date:  2017-05-15       Impact factor: 4.897

Review 3.  The clinical utility of circulating tumour cells in patients with small cell lung cancer.

Authors:  Victoria Foy; Fabiola Fernandez-Gutierrez; Corinne Faivre-Finn; Caroline Dive; Fiona Blackhall
Journal:  Transl Lung Cancer Res       Date:  2017-08

4.  Biochemical tests of placental function versus ultrasound assessment of fetal size for stillbirth and small-for-gestational-age infants.

Authors:  Alexander Ep Heazell; Dexter Jl Hayes; Melissa Whitworth; Yemisi Takwoingi; Susan E Bayliss; Clare Davenport
Journal:  Cochrane Database Syst Rev       Date:  2019-05-14

5.  The Effects of Early Bispectral Index to Predict Poor Neurological Function in Cardiac Arrest Patients: A Systematic Review and Meta-Analysis.

Authors:  Chun-Yu Chang; Chien-Sheng Chen; Yung-Jiun Chien; Po-Chen Lin; Meng-Yu Wu
Journal:  Diagnostics (Basel)       Date:  2020-04-30

6.  Quantifying how diagnostic test accuracy depends on threshold in a meta-analysis.

Authors:  Hayley E Jones; Constantine A Gatsonsis; Thomas A Trikalinos; Nicky J Welton; A E Ades
Journal:  Stat Med       Date:  2019-09-30       Impact factor: 2.373

7.  Meta-analysis of diagnostic accuracy studies in mental health.

Authors:  Yemisi Takwoingi; Richard D Riley; Jonathan J Deeks
Journal:  Evid Based Ment Health       Date:  2015-10-07

8.  Modelling multiple thresholds in meta-analysis of diagnostic test accuracy studies.

Authors:  Susanne Steinhauser; Martin Schumacher; Gerta Rücker
Journal:  BMC Med Res Methodol       Date:  2016-08-12       Impact factor: 4.615

9.  Meta-analysis of test accuracy studies using imputation for partial reporting of multiple thresholds.

Authors:  J Ensor; J J Deeks; E C Martin; R D Riley
Journal:  Res Synth Methods       Date:  2017-11-22       Impact factor: 5.273

  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.