Savannah Bergquist1, Gabriel A Brooks2, Mary Beth Landrum3, Nancy L Keating3, Sherri Rose4. 1. Haas School of Business, University of California, Berkeley, Berkeley, California, USA. 2. The Dartmouth Institute for Health Policy and Clinical Practice, Geisel School of Medicine, Hanover, New Hampshire, USA. 3. Department of Health Care Policy, Harvard Medical School, Boston, Massachusetts, USA. 4. Department of Health Policy and Center for Health Policy, Stanford University, Stanford, California, USA.
Abstract
The difficulty in identifying cancer stage in health care claims data has limited oncology quality of care and health outcomes research. We fit prediction algorithms for classifying lung cancer stage into three classes (stages I/II, stage III, and stage IV) using claims data, and then demonstrate a method for incorporating the classification uncertainty in survival estimation. Leveraging set-valued classification and split conformal inference, we show how a fixed algorithm developed in one cohort of data may be deployed in another, while rigorously accounting for uncertainty from the initial classification step. We demonstrate this process using SEER cancer registry data linked with Medicare claims data.
The difficulty in identifying cancer stage in health care claims data has limited oncology quality of care and health outcomes research. We fit prediction algorithms for classifying lung cancer stage into three classes (stages I/II, stage III, and stage IV) using claims data, and then demonstrate a method for incorporating the classification uncertainty in survival estimation. Leveraging set-valued classification and split conformal inference, we show how a fixed algorithm developed in one cohort of data may be deployed in another, while rigorously accounting for uncertainty from the initial classification step. We demonstrate this process using SEER cancer registry data linked with Medicare claims data.
Understanding generalizability across populations is important for answering applied health questions,
especially in the context of prediction algorithms.
In many cases, prediction tools are developed for a specific cohort or population but have limited practical applicability because it is difficult to rigorously assess the uncertainty that comes from deploying a fixed prediction algorithm across settings. Population‐level oncology research in the United States is an example of this phenomenon. There is great interest in predicting cancer stage using health care claims data because stage—an indicator of disease severity not included in claims data—is a key variable in understanding whether patients received appropriate treatments and for assessing health outcomes, including survival.
,
The availability of valid algorithms to identify cancer stage using only claims data would allow for timely assessments of quality and outcomes as well as comparative effectiveness research in non‐clinical trial settings.However, in applied classification settings, it can be a particular challenge to convey uncertainty. Often, a classification or risk prediction tool is developed with a focus on in‐sample and validation sample performance measures, but little or no consideration is given to the use of the predicted values in subsequent estimation tasks.
,
,
,
,
Predicted values are commonly treated as observed data in downstream analyses. Depending on the specific applied context, this can lead to bias, erroneously deflated SEs, and decreased interpretability.
,To the best of our knowledge, there is no method that incorporates cancer stage classification uncertainty in survival estimation. The goal of this study is to develop a method that bridges between presenting raw probabilities—which may reflect our best sense of epistemic uncertainty albeit difficult to act upon in practice—and treating predicted class labels as ground truths. Using SEER registry data linked to Medicare claims data, we develop prediction algorithms to classify patients as having stage I/II, stage III, or stage IV lung cancer. After evaluating predictive performance, we then examine survival outcomes stratified by these stage classes. Motivated by this real‐world applied problem of identifying stage for use in survival estimation, we propose an approach that leverages set‐valued classification, split conformal inference, and resampling to produce sets of labels that simplify underlying classification uncertainty and translate it into outcomes estimation. We describe this procedure and compare it to a naive classification and outcomes estimation method without consideration of prediction uncertainty, and a naive method with uncertainty incorporated via resampling. In a simulation study, we explore conditions under which the standard practice approach differs from the bootstrap‐based alternatives, and show how uncertainty may vary across class labels. Both bootstrap methods are an improvement over standard practice in terms of taking steps to more comprehensively communicate uncertainty in applied settings. Our applied data example shows that in practice the simpler, naive bootstrap tool may perform as well as the more complex weighted labeling bootstrap procedure.
Related work on prediction and uncertainty
Multiple imputation and the measurement error modeling literatures are frequently applied in settings where a key variable of interest is missing for some or all subjects. Multiple imputation commonly seeks to generate values when a variable is missing for a subset of subjects in a data set. Covariate error measurement models often rely on some portion of observed data or external information pertaining to the variables measured with error and require assumptions about the nature of the error (eg, independent and additive). In cancer stage classification with claims data—and many other settings across medicine and health services research—the variable to be predicted is systematically missing for the entire analysis sample. In our proposed approach we avoid the assumption that errors are independent of the predicted values, and we do not require downstream users to observe any values for the variable to be predicted.Wang et al
propose a method for correcting estimates, SEs, and test statistics in settings where the research question of interest focuses on the relationship between a covariate and outcome, but the outcome must first be predicted. They term inference with predicted outcomes “post‐prediction inference.” Their correction is based on the relationship between the observed and predicted outcome in a test data set. This post‐prediction inference method is flexible but it is not developed for cases where the predicted variable is not the subsequent target outcome.In clinical risk prediction, a popular method for communicating uncertainty is generating an uncertainty score to accompany the predicted values.
,
The aim of this work is often to provide a measure of uncertainty to facilitate clinical decision‐making, rather than using the predicted values and associated uncertainty measure in statistical analysis. However, these methods share a high‐level goal with our proposed method and post‐prediction inference: improve the effectiveness of prediction tools for applied problem solving.
Conformal inference and related methods
The conformal inference framework was developed by Vovk et al
to provide an unbiased solution for obtaining a prediction interval around a new observation drawn from the same distribution as a set of training data. Note that a naive solution to creating a prediction interval based on ranking the residuals from the training data is in practice unlikely to obtain target coverage, meaning the probability that a new test point lies in the prediction interval may be less than 1‐, where is the specified error level. Full conformal inference is computationally expensive because it requires refitting the prediction algorithm, and recalculating and reordering the residuals for every new data point.Split conformal inference separates the fitting and ranking steps so that the fitting only occurs as frequently as the data are split (ie, with a single split, the prediction algorithm is fit once and the residuals are ranked once). Papodopoulos,
Vovk,
and Lei et al
show that if the model is correctly specified, split conformal inference provides near‐optimal conditional coverage; in cases where it is not, meaningful, marginal coverage may still be obtained. Health care data sets are typically subject to some gradual shifts in treatment practices and population demographics,
and these types of drifts may or may not threaten assumptions about the data distribution over limited time periods. However, investigator knowledge is critical for understanding events that can cause dramatic shifts and violate exchangeability (eg, changes in medical coding regulations or the introduction of a blockbuster drug). Although split conformal inference can lead to wider prediction intervals than methods that use a greater number of splits, it is less computationally expensive and has been developed to work with classification outcomes.
,
In this article, we deploy split conformal inference to leverage these characteristics and extend it for purposes of generalizability.Although they are not directly applicable to our real‐world data application of cancer stage classification and survival estimation because they do not produce label sets, we briefly describe two further, related approaches in the literature that produce prediction intervals for regression settings. Cross conformal inference
,
is a natural extension of split conformal inference: instead of dividing the data in two halves, the data are divided into folds of equal size. Cross conformal inference has no coverage guarantees when is large. Jackknife+ is a recently introduced method that uses leave‐one‐out predictions for the prediction interval.
Barber et al
also propose a method called CV+, which yields larger intervals than jackknife+ but requires less computation. We did not pursue adapting these approaches in this study because we are focused on an applied classification setting.
METHODS
We will compare three methods, which are summarized conceptually in Figure 1. In the naive standard practice approach, the prediction algorithm is fit in the development data and applied to a validation data set, yielding a single label per observation that is used in subsequent outcomes analysis. In our data setting, the naive approach is to apply a fitted classification algorithm in the validation data, assign lung cancer stage based on the greatest predicted class probability, and then use this stage label to estimate survival. Under this naive approach, no special consideration is given to the uncertainty from predicting class labels. In the naive bootstrap method, uncertainty is incorporated by bootstrapping the validation data. This resampling allows for bootstrap‐based confidence intervals around classification performance measures and outcomes estimation. Finally, the weighted labeling bootstrap procedure divides the development data into two halves to leverage split conformal inference (discussed in more detail below). The fitted algorithm and label thresholds are then applied to the validation data to assign a set of plausible labels for each individual observation. The validation data are resampled and a single label is selected from each label set and used in outcomes estimation. Thus the weighted labeling approach captures classification and outcomes estimation uncertainty via set‐valued classification (the label sets) and resampling.
FIGURE 1
Conceptual overview of methods
Conceptual overview of methods
LABEL classifier and weighted bootstrap
Our data structure is defined as , where is a feature vector and is the mutually exclusive label space. In the multiclass setting, we seek to apply a label to each observation based on . We set ; in our applied data analysis, lung cancer stage I/II, stage III, and stage IV map to labels 1‐3, respectively. Although each observation has a single true label, our goal is to assign a set of plausible labels in order to provide more information about the uncertainty of the classification process. To do so, we employ a set‐valued classifier proposed by Sadinle et al,
the least ambiguous with bounded error levels (LABEL) classifier. In combination with split conformal inference, the LABEL classifier yields distribution‐free, finite sample coverage levels with minimal label ambiguity when the assumption of exchangeability holds. To estimate each , we require an initial estimate of the conditional probability function, , and class‐specific thresholds for converting probabilities to labels, . Any conventional estimator of can be plugged in to the LABEL classifier; in our simulation study we consider a multinomial logistic regression and in our data analysis we include a range of estimators.We desire equal coverage across all three lung cancer stage classes, and therefore pre‐specify individual class error levels of . As described in Sadinle et al,
in the optimal procedure, the set‐valued classifier, , is a subset of the label space and can be represented by a collection of sets mapping the feature inputs to labels: . The estimated sets for each class label, , are based on the estimated conditional probabilities, , and estimated thresholds, .The coverage for class is based on the collection of estimated sets, , that contain the true class label:
where indexes independent observations from . The threshold is estimated:
In an applied, finite sample setting with split conformal inference, the data are split in two halves, indexed by and . For our application, the data are first divided into a development and validation cohort, and the development cohort is split into and . The subset is used to estimate . We partition into groups corresponding to each label class: . Within the given partition , thresholds are estimated according to pre‐specified error levels :
where indexes over for comparison of each predicted probability to the ranked predicted probabilities. We then obtain the sets so that is the split conformal set‐valued classifier based on the plug‐in estimator . Instances where more than one label is assigned to an observation are called ambiguous sets, and those with no assigned labels are null sets. With set equally across classes, a higher threshold for a given class indicates a greater number of higher predicted probabilities generated by . We apply the classifier—based on the plug‐in estimator fit in and the thresholds obtained in —to the observations in the validation cohort. Separating the creation of the classifier from the validation cohort allows the classifier to be used in data settings where the true class is unobserved.Our end goal is to use the class labels in outcomes estimation. Thus, we force a single label per observation in a bootstrap procedure: for all resamples of the validation data, a single label, , is randomly selected with equal probability from all classes where . For null sets, all class labels are considered with equal probability (eg, in settings where the classification algorithm has high average accuracy and there is little uncertainty, we may expect higher classification thresholds and thus null sets for difficult‐to‐classify observations). After a single label is randomly selected among those belonging to the sets , outcome estimation can be performed based on . We refer to this approach as “weighted” labeling with bootstrap, or weighted bootstrap, because the probability of label assignment for a resampled observation is weighted by the set obtained prior to the resampling procedure. For example, in the case of , the set of possible label assignment weights is .The bootstrap procedure allows the class labels assigned to a single observation to vary, thus incorporating classification uncertainty in both classification evaluation and outcomes estimation. Additionally, the resampling permits us to calculate confidence intervals for a variety of metrics of interest, including label coverage, counts of ambiguous label sets, conventional classification evaluation measures, and outcome measures. Because there is no sample mean for metrics that depend on a single class per observation (eg, classification accuracy), we report bootstrap percentile‐based averages and confidence intervals.
Naive comparators: Most probable class, with and without resampling
As comparators, we implement a naive classification strategy where a single class label is assigned to each observation based on the most probable predicted class. Both naive comparators implement this classification approach. We use the same conditional probability estimator as in the weighted labeling approach, , but fit it on the entire development cohort. The fitted classification algorithm is then applied to the validation cohort, and class labels are assigned based on the class with the highest predicted probability. These class labels are treated as known in subsequent outcomes estimation exercises. We use this approach, when applied in a single iteration of an analysis sample, as our standard practice comparator. A step beyond implementing this naive standard practice strategy is to incorporate bootstrap resampling. The classification algorithm fitting and prediction steps remain unchanged, but a bootstrap procedure is introduced for the classification performance evaluation and outcome estimation: the validation cohort is resampled and bootstrap‐based confidence intervals may be calculated for classification and outcome measures. We include the naive bootstrap approach as a more nuanced method than the naive standard practice, but a less complex alternative to the weighted labeling bootstrap procedure introduced above. Depending on the real‐world data context, the simpler, naive bootstrap may be preferable to the weighted labeling method.
Classification performance measures
For a given , we assess classification performance via accuracy, sensitivity, specificity, and positive predictive value (see Table A1 for definitions). To calculate these measures, we tabulate the per‐class proportions of true positives, true negatives, false positives, and false negatives. We then present what we call the macro‐averaged versions of each measure by averaging across the classes. Because the split conformal inference‐based LABEL classifier does not provide any coverage guarantees in this resampling setting, we empirically verify marginal coverage and examine estimated thresholds and ambiguous labels for each class. In our data application, we also plot the proportion of observations in a given stage by ordered predicted probability to assess calibration.
Survival analysis
Survival outcomes are central to many strands of oncology research. Cancer stage indicates disease severity and is critical to assessing not only survival but also common health services research questions, such as examining spending outcomes and determining whether or not patients received appropriate treatment. Summarizing disease severity into a single variable simplifies both survival estimation and interpretability. To examine how our method compares to the naive approaches, we implement a common form of survival estimation, stratified Kaplan‐Meier analysis. We select Kaplan‐Meier estimation for simplicity and ease of exposition, but many alternative methods for survival outcome estimation could be used in practice, including standard parametric survival regression models, the Cox proportional hazards regression,
random survival forests,
and combinations of parametric, semiparametric, and nonparametric approaches in a stacked survival model.
,
,
To evaluate survival performance we examine bias and determine how the median survival time and 90 and 365‐day survival probabilities based on predicted class compare to those based on observed class. We report average bias and 95% confidence intervals.
SIMULATION STUDY
In this section, we describe a simulation study to compare the methods for classification and outcome estimation outlined above. The goal is to highlight settings where the relative performance of a naive, single iteration classification and outcome estimation approach (“standard practice”) varies from a similar naive approach that incorporates a bootstrap procedure and from our proposed novel approach of weighted labeling with bootstrap, which leverages set‐valued classification. Although we construct our simulation data to be similar to our real‐world lung cancer data setting, the simulation study allows us to explore methodological performance under differing data scenarios. To reflect this, throughout the simulation study we refer to the classification outcome variable groups as “classes” rather than stage.Our simulation sample is observations and we generate patient‐level covariates for each patient , where . To roughly approximate the sociodemographic and clinical characteristics of the lung cancer study, we simulate six continuous, seven binary, and two count variables; Table B2 provides additional details on covariate distributions. The class label outcome where , is generated from a multinomial distribution, , where the class label is assigned based on the greatest class probability. Class balance is approximately 37% class 1, 49% class 2, and 13% class 3. See Appendix B for additional details on the construction of the probabilities underlying .To perform survival analysis, we use a Weibull distribution and simulate right censored event times: , where is simulated survival time influenced by time‐independent covariate with effect parameter , , is the Weibull shape parameter, and the scale parameter. We set , , and is the class label simulated for the classification exercise. We set the length of observation time to 365 days and generate a variable to capture observed censored event times and an indicator if the event occurred or was censored. After simulating all variables, the data are stratified on the class label and divided into equally sized development and validation cohorts, each of size 1000. For the weighted labeling with bootstrap method, the development data are again stratified on and further divided into two halves of 500 observations each.We implement three scenarios: (1) accurate and certain classification, (2) accurate and uncertain classification, and (3) inaccurate and uncertain classification. Table B2 compares the covariates used to fit the multinomial logistic regression for label classification in each scenario; scenarios differ only by which covariates are used to predict class label probabilities; all other simulation parameters are the same. We characterize the scenarios based on how accurately the classification regression predicts class labels with the given covariates, and based on the empirical distributions of the predicted probabilities (the uncertain scenarios have a higher proportion of predicted probabilities near 0.5). After fitting the conditional probability estimation algorithms and applying labels in the validation data, Kaplan‐Meier survival estimation is performed in the validation data, stratified by observed and predicted class labels. We draw 500 bootstrap resamples for the bootstrap‐based methods and we perform 1000 simulation repetitions for each scenario.Across the simulation scenarios, we compare how each of the three methods performs in terms of classification and survival analysis. For classification performance we examine accuracy, sensitivity, specificity, and positive predictive value (PPV), as well as the underlying counts of true positive, true negatives, false positives, and false negatives. We present mean measures and bootstrap‐based 95% confidence intervals where applicable. For all methods, we report coverage and 95% confidence intervals for each class. We also report the class‐specific thresholds and ambiguity for the weighted labeling classifier (recall the naive methods do not incorporate threshold‐based label assignment and only assign a single label per observation so there is no label ambiguity). Our primary survival analysis performance measure is bias. We examine the average bias across simulations and report percentile 95% confidence intervals around bias estimates.
Simulation results
Class‐specific coverage results across simulation scenarios are presented in Table 1. The weighted bootstrap method yields 90% coverage for all three classes across settings. Note that although we do not fill in null sets until the bootstrap procedure, because the number of null sets is low (described below), our classifier is still able to obtain the target coverage level. The naive classification method of assigning class labels based on the highest probability produces coverage ranging from 1% to 94% across all three classes, and coverage declines within each class from scenario 1 (accurate and certain class prediction) to scenario 2 (accurate and uncertain) to scenario 3 (inaccurate and uncertain). The naive classification methods do not target coverage, while class‐specific target coverage is an explicit input of the LABEL classifier in the weighted bootstrap method. The two naive methods are identical at this stage because coverage is calculated in sample prior to the bootstrap procedure.
TABLE 1
Coverage and Monte Carlo SEs
Method
Class 1
Class 2
Class 3
Scenario 1
Naive labeling
0.94
0.87
0.41
(0.01)
(0.01)
(0.00)
Weighted labeling
0.90
0.90
0.90
(0.01)
(0.01)
(0.01)
Scenario 2
Naive labeling
0.82
0.80
0.21
(0.01)
(0.01)
(0.01)
Weighted labeling
0.90
0.90
0.91
(0.01)
(0.01)
(0.01)
Scenario 3
Naive labeling
0.54
0.62
0.00
(0.02)
(0.02)
(0.00)
Weighted labeling
0.90
0.90
0.91
(0.01)
(0.01)
(0.01)
Coverage and Monte Carlo SEs
Classification performance
The classifier thresholds in the weighted labeling with bootstrap method tend to decrease within class as uncertainty and inaccuracy are introduced across simulation scenarios (Table B3). Recall the thresholds are class‐specific cutoff probabilities for determining if an observation receives a given class label. Class 2 is the most prevalent class (approximately 49% of the sample), class 1 is approximately 37% of the sample, and class 3 comprises approximately 13%. In scenario 1, the most accurate and certain setting, class 1 has the highest threshold, illustrating that the most prevalent class will not necessarily generate the highest label threshold.Label set ambiguity provides a more complete picture as to how the classifier operates (Figure 2). As the accuracy and certainty declines across simulation scenarios, we see the number of null and single label sets decrease and the number of double and triple label sets increase. Examining ambiguous label sets by true class, we see where the distribution of label sets differs by class. For example, in scenario 1 (accurate and certain), about 75% of all observations receive a single label set. But for class 1, over 95% of observations have a single label set, while in class 2 only 67% have a single label and in class 3 the proportion drops to 41%. Under scenario 2 (accurate and uncertain), 49% of all observations receive a single label set, but again this differs by true class: 78% of class 1 observations have a single label, 34% of class 2, and 21% of class 3. Under the inaccurate and uncertain scenario 3, the distribution becomes more even, with 2% of class 1 observations having a single label, 3% of class 2, and 2% of class 3. Label set ambiguity can be helpful as a diagnostic tool, particularly in the multiclass setting where uncertainty levels may vary across classes.
FIGURE 2
Simulation study label ambiguity: Share of sample by number of assigned labels in label set
Simulation study label ambiguity: Share of sample by number of assigned labels in label setFigure 3 reports average classification accuracy, sensitivity, specificity, and PPV across simulation scenarios. Overall, the weighted labeling approach yields slightly poorer performance across metrics compared to the naive methods. Small differences can be seen for some metrics. For example, in scenario 2, weighted labeling has an accuracy of 77% while the naive methods have an accuracy of 82%. Such differences may or may not be meaningful depending on the application. Table B4 reports class‐specific classification accuracy. Results for class 1 are similar across the methods, but the weighted labeling performs worse compared to the naive methods in classes 2 and 3 by 4‐7 percentage points.
FIGURE 3
Simulation study classification discrimination: Average accuracy, sensitivity, specificity, and PPV (for visual clarity, 95% confidence intervals less than 0.05 are not displayed)
Simulation study classification discrimination: Average accuracy, sensitivity, specificity, and PPV (for visual clarity, 95% confidence intervals less than 0.05 are not displayed)Tables B5‐B7 contain class‐specific classification sensitivity, specificity, and PPV. The small differences between the naive methods and the weighted bootstrap are much more pronounced at the class‐level. For example, the weighted bootstrap produces lower sensitivity for class 2 (by 12 percentage points in scenario 1, and about 25 percentage points in scenarios 2 and 3), but performs substantially better for class 3. In scenario 1, the weighted bootstrap sensitivity estimates are a 20 percentage point improvement over the naive methods, and an approximately 30 percentage point improvement in scenarios 2 and 3 (in scenario 3, the naive methods yield 0% sensitivity). Tables B8‐B11 present true positive, true negative, false positive, and false negative counts by class across simulation scenarios. The conditional probability estimates for class 3 are generally low compared to classes 1 and 2, leading to few class 3 labels under the naive approaches. Across all settings, the weighted bootstrap method generates fewer true positives than the naive methods for classes 1 and 2, but yields a substantially larger number for class 3. The practical import of these class‐based differences will vary by applied context, especially considering our primary outcome of interest here is survival.
Survival analysis
Bias for median survival is shown in Figure 4. Overall, the weighted bootstrap produces smaller bias about 40% of the time compared to the naive methods, but in each of these cases the confidence intervals are compatible with equal bias across methods. The weighted bootstrap and naive methods have similarly sized confidence intervals across scenarios for classes 1 and 2. For class 3, the weighted bootstrap has a much smaller confidence interval than the naive methods, particularly so in scenario 3. The results for 90 and 365‐day survival (Figures B1 and B2) run counter to what one might expect: rather than yielding smaller SEs or less biased average survival probabilities than the naive or weighted bootstrap methods, standard practice produces similar results and often larger confidence intervals, with the noted exception of class 3 in the inaccurate and uncertain scenario.
FIGURE 4
Simulation study median survival days bias (For visual clarity, 95% confidence intervals less than 31 days are not displayed)
Simulation study median survival days bias (For visual clarity, 95% confidence intervals less than 31 days are not displayed)
DATA ANALYSIS
We use SEER cancer registry data linked with Medicare claims data as our real‐world data application that motivated the development of these methods.
The SEER data provide information on all cancers diagnosed among individuals living in areas covered by SEER registries, including cancer stage.
SEER data are abstracted from medical records and contain validated staging information at the time of diagnosis, and thus are the source of our “gold standard” cancer stage labels: stage I/II, stage III, and stage IV. We combine stages I and II into a single label to accommodate sample size constraints and similarity of clinical outcomes.
Fee‐for‐service Medicare claims data contain detailed information on treatments received as well as health care visits and comorbidities. Medicare enrollment data provide information on patient age, race/ethnicity, vital status, and information on zip‐code level measures of socioeconomic status.Our study cohort includes individuals aged 65 and older who were enrolled in fee‐for‐service Medicare and were diagnosed with stage I‐IV lung cancer between 2010 and 2013 who received chemotherapy within 6 months of diagnosis. We divide the data into two cohorts based on timing of diagnosis: a development cohort (2010‐2011 diagnoses) and a validation cohort (2012‐2013 diagnoses). In practice, a fixed classification algorithm is likely to be applied to individuals diagnosed in time periods successive to the training sample, thus we aim to approximate this likely real‐world scenario. Table 2 shows the similarity between the two cohorts in terms of basic summary statistics; the similarity of our development and validation samples can be considered a “best case” scenario in terms of generalizability and prediction.
TABLE 2
Cohort summary statistics
Characteristic
Development cohort 2010‐2011 diagnosis
Validation cohort 2012‐2013 diagnosis
N
14 760
14 620
Age (mean)
72.1
71.9
Documented sex female (%)
45.4
46.8
Race/ethnicity (%)
White
82.7
81.4
Black
8.8
8.9
Hispanic
4.2
4.4
Other
4.2
5.3
Region (%)
Northeast
20.3
20.2
Midwest
13.6
13.2
West
34.2
34.5
South
31.9
32.2
Stage at diagnosis (%)
I/II
15.1
14.8
III
34.0
33.1
IV
50.8
52.0
Cohort summary statisticsFor stage classification, our input features are 94 variables derived from or linked to the Medicare claims data in the period 3 months before or after an individual receives their first lung cancer chemotherapy. Variables include patient demographic characteristics, visits and hospitalizations, chemotherapy drugs, surgeries and procedures, radiation, comorbidities, and lung cancer anatomic site and malignancy diagnosis codes. A full list of features for this data set is described in Brooks et al.Prior work classifying lung cancer stage for patients receiving chemotherapy focuses on a binary split of early (stages I‐III) vs late (stage IV).
,
Because it is unclear which single algorithm will perform best in the multiclass setting (stages I/II, stage III, and stage IV), we implement 7 algorithms that are multiclass versions of the most promising discrete binary prediction algorithms: main terms multinomial logistic regression;
penalized regressions (lasso, ridge, an elastic net with overall penalty selected via internal cross‐validation, and a balanced elastic net penalty set at 0.5),
generalized additive regression with cubic splines and a smoothing parameter set to 0.6;
random forests with node size 250 and 500 trees;
and gradient boosting with a maximum tree depth of 3, learning rate set to 1, and 2 fitting rounds.We examine patient survival to illustrate how our classification method can be used in practice with a common form of health outcomes estimation. Survival outcomes are estimated based on the number of days from first chemotherapy to death, and we follow patients for a one year period. For the bootstrap‐based methods we report average bias, survival estimates, and percentile‐based confidence intervals. For the naive standard practice approach we report bias, survival estimates, and Greenwood confidence intervals for the survival estimates. (Note that unlike in the simulation study, naive standard practice approach Monte Carlo‐based SEs are not available for the data analysis.)
Data analysis results
In the weighted labeling approach, all algorithms provide at least nominal coverage (90%) across all three stage classes. Under the naive method, for stage I/II, the average coverage is about 31%, stage III 60%, and stage IV about 84%. Table C15 contains the average coverage and 95% confidence intervals across bootstrap samples for each algorithm and method.The thresholds for the weighted labeling approach vary across algorithms and classes (Table C16). All of the algorithms produce the lowest thresholds for stage I/II and the highest thresholds for stage IV. The stage IV thresholds are similar across all algorithms (ranging from 0.31 to 0.40), but the random forests generates the lowest thresholds for stage I/II (0.00 vs 0.07‐0.11) and stage III (0.07 vs 0.21‐0.23).None of the algorithms produce a null label set (multinomial logistic regression and random forests presented in Figure C3; results for other algorithms omitted due to similarity with multinomial logistic regression). Across all validation observations, random forests produces the fewest single label sets (about 24% of validation observations) and the most triple label sets (30%). In contrast, the multinomial logistic regression produces more single (35%) and double label sets (44%) across all classes. Examining ambiguity by true label class, we see the algorithms produce fewer single label sets for observations with a true class of stage I/II, reflecting the lower thresholds for this class. True stage IV is most likely to belong to a single label set, while true stages I/II and III are more likely to belong to a double label set, and all true classes have similar proportions of triple label sets.To assess prediction calibration for each algorithm, we plot ordered predicted probabilities against the percent of observations belonging to a given stage (Figure C4). Recall the naive prediction methods use the entire development sample for algorithm fit, while the weighted bootstrap approach uses only half of the development sample for algorithm fitting, and the other half to set labeling thresholds. The average predicted probabilities across the distributions are nearly identical between the two methods for all stages and algorithms, with slight exceptions in gradient boosting and random forests. Overall, the algorithms are most poorly calibrated for stage I/II, show slightly better calibration for stage III, and perform best for stage IV.Within the naive methods, the algorithms produce nearly identical classification performance results for several measures of discrimination (Figure 5). There are some small variations across algorithms within the weighted bootstrap method, but in general the classification performance of all algorithms—except the random forests—is within 1‐2 percentage points. Tables C17‐C20 present class‐specific measures. The weighted bootstrap results are an improvement compared to the naive methods for stage I/II where sensitivity is higher (by 12‐19 percentage points), and sensitivity and PPV estimates are higher for stage IV. However, the weighted bootstrap yields slightly lower accuracy across all three stage classes compared to the naive methods. Overall, in Figures C4 and 5, we see that measures of calibration and classification are similar, although in some cases there may be small differences, and performance by stage group is more variable.
FIGURE 5
Data analysis classification discrimination: Average accuracy, sensitivity, specificity, and PPV (for visual clarity, 95% confidence intervals less than 0.05 are not displayed)
Data analysis classification discrimination: Average accuracy, sensitivity, specificity, and PPV (for visual clarity, 95% confidence intervals less than 0.05 are not displayed)For 90‐day survival, the weighted bootstrap is close to 0% bias across all stages, whereas the naive methods show greater variation (Figure 6). The weighted bootstrap is approximately 5 percentage points better than the naive methods in stage I/II, about 2 percentage points better in stage III, and only about 2 percentage points worse than the naive methods for stage IV. The naive and weighted labeling bootstrap approaches yield similarly sized bootstrap‐based confidence intervals. For 365‐day survival, the weighted labeling method continues to generate the smallest bias for stages I/II and III, and in some algorithms also produces the smallest bias in stage IV (Figure C5). The bias estimates are most different between the naive methods and the weighted label bootstrap for stage I/II, which is also the stage group that saw the greatest difference in classification results (Tables C17‐C20). The observed median survival for stage I/II exceeds 1 year, so we do not estimate it here. For stage III, both naive methods tend to overestimate median survival compared to the true class, although they generate little bias for stage IV. The weighted labeling bootstrap method slightly underestimates median survival in both stage III and IV (Figure C6).
FIGURE 6
Data analysis: 90‐day survival probability percent bias
Data analysis: 90‐day survival probability percent biasThe naive standard practice and naive bootstrap methods generally produce similarly sized confidence intervals around the survival estimates, although there are some small variations in relative width across algorithms and class (see Figures C7‐C9). Across all survival estimates, the weighted labeling bootstrap method produces similar results to the naive methods but does generate slightly narrower confidence intervals for the stage I/II 365‐day survival estimates. As in the simulation study, we find the standard practice survival estimates do not necessarily yield smaller confidence intervals or less biased average survival probabilities than the naive bootstrap method. Instead, they tend to produce similar results that vary by predicted class and the algorithm used to estimate conditional probabilities.
DISCUSSION
In this article, we studied conveying uncertainty in applied classification settings, and proposed a procedure leveraging set‐valued classification, split conformal inference, and resampling. Our proposed method uses bootstrap resampling from the sets of plausible labels generated in the classification step, and then performs outcomes estimation based on the selected labels. In our real‐world data example, we developed fixed multiclass prediction algorithms for labeling lung cancer stage. The weighted labeling procedure yielded the smallest bias for survival estimates in stages I/II and stage III.Our development and validation samples had similar observed characteristics and were drawn from sequential time periods: development cohort patients were diagnosed with lung cancer from 2010 to 2011 and validation cohort patients in 2012‐2013. As such, the naive bootstrap approach may in practice be preferable to the weighted labeling bootstrap in such scenarios due to the simplicity of implementation without dramatic losses in performance. However, had our validation sample been drawn from a very different population—for example, patients diagnosed in 2019‐2020—we might find much higher levels of uncertainty and lower levels of prediction accuracy due to changes in treatment patterns and patient characteristics. In this case, the weighted labeling approach may provide a more nuanced picture of the label uncertainty and impact on outcomes estimation by generating label sets for each observation.Simulations show that our method outperforms the naive methods in terms of bias for some classes and scenarios, and for others yields similar survival estimates and confidence intervals. The simulation study also illustrates how label uncertainty can vary across classes and be translated into poor outcomes estimation performance; for example, class 3 yielded a higher number of ambiguous label sets and the greatest amount of bias across all simulation scenarios. Further, we note that our method helps highlight the issues with making predictions in uncertain or inaccurate data scenarios: reporting label ambiguity in addition to multiple performance metrics emphasizes poor prediction characteristics and algorithms that should not be used in downstream analyses.The goal of this study was to propose an approach for characterizing uncertainty from a prediction exercise and demonstrate how this uncertainty can be incorporated in downstream survival analysis. As noted in the text, while we implemented a simple Kaplan‐Meier analysis, there are more sophisticated time‐to‐event estimation approaches that could be deployed. Similarly, there are a range of other outcomes that may be of interest, including adjusted quality of care measures or cost‐effectiveness estimates for oncology treatments.In summary, the weighted labeling with bootstrap method is one approach for incorporating uncertainty in applied classification and estimation problems. The proposed naive bootstrap procedure is also an improvement over simply ignoring prediction uncertainty. Depending on the data setting, the naive bootstrap may be implemented as an uncomplicated alternative to the weighted labeling method. As classification and risk prediction algorithms become more commonplace in medical and health services settings, we must think beyond prediction evaluation and implement tools to effectively communicate uncertainty.
CONFLICT OF INTEREST
The authors declare no potential conflict of interests.Data S1: Supporting informationClick here for additional data file.
1. For development data:
a. Split development data into equal halves: ℐ1 and ℐ2
b. Fit conditional probability estimator P(y|x) on ℐ1
c. Apply P^(y|x) to ℐ2 to obtain thresholds t^y
2. For validation data:
a. Apply P^(y|x) and t^y to obtain sets C^y and calculate coverage
b. Draw nboot resamples:
i. Select a single label y^boot from each resampled observation's label set
ii. Calculate classification performance measures based on y^boot
iii. Estimate outcome of interest based on y^boot
c. Report statistics of interest based on validation sample and validation bootstrap resamples
Authors: Michael J Hassett; Hajime Uno; Angel M Cronin; Nikki M Carroll; Mark C Hornbrook; Debra Ritzwoller Journal: Med Care Date: 2017-12 Impact factor: 2.983
Authors: Daina B Esposito; Leo Russo; Dina Oksen; Ruihua Yin; Vibha C A Desai; Jennifer G Lyons; Patrice Verpillat; Jose L Peñalvo; Francois-Xavier Lamy; Stephan Lanes Journal: Cancer Epidemiol Date: 2019-05-22 Impact factor: 2.984
Authors: Debra P Ritzwoller; Michael J Hassett; Hajime Uno; Angel M Cronin; Nikki M Carroll; Mark C Hornbrook; Lawrence C Kushi Journal: J Natl Cancer Inst Date: 2018-03-01 Impact factor: 13.506
Authors: Nancy L Keating; Shalini Jhatakia; Gabriel A Brooks; Amanda S Tripp; Inna Cintina; Mary Beth Landrum; Qing Zheng; Thomas J Christian; Roberta Glass; Van Doren Hsu; Colleen M Kummet; Susannah Woodman; Carol Simon; Andrea Hassol Journal: JAMA Date: 2021-11-09 Impact factor: 56.272
Authors: Lindsey Enewold; Helen Parsons; Lirong Zhao; David Bott; Donna R Rivera; Michael J Barrett; Beth A Virnig; Joan L Warren Journal: J Natl Cancer Inst Monogr Date: 2020-05-01