Literature DB >> 35474515

Penalized likelihood estimation of a mixture cure Cox model with partly interval censoring-An application to thin melanoma.

Annabel Webb1, Jun Ma1, Serigne N Lô2,3,4.   

Abstract

Time-to-event data in medical studies may involve some patients who are cured and will never experience the event of interest. In practice, those cured patients are right censored. However, when data contain a cured fraction, standard survival methods such as Cox proportional hazards models can produce biased results and therefore misleading interpretations. In addition, for some outcomes, the exact time of an event is not known; instead an interval of time in which the event occurred is recorded. This article proposes a new computational approach that can deal with both the cured fraction issues and the interval censoring challenge. To do so, we extend the traditional mixture cure Cox model to accommodate data with partly interval censoring for the observed event times. The traditional method for estimation of the model parameters is based on the expectation-maximization (EM) algorithm, where the log-likelihood is maximized through an indirect complete data log-likelihood function. We propose in this article an alternative algorithm that directly optimizes the log-likelihood function. Extensive Monte Carlo simulations are conducted to demonstrate the performance of the new method over the EM algorithm. The main advantage of the new algorithm is the generation of asymptotic variance matrices for all the estimated parameters. The new method is applied to a thin melanoma dataset to predict melanoma recurrence. Various inferences, including survival and hazard function plots with point-wise confidence intervals, are presented. An R package is now available at Github and will be uploaded to R CRAN.
© 2022 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd.

Entities:  

Keywords:  asymptotic variance; constrained optimization; direct likelihood maximization; mixture cure model; penalized likelihood

Mesh:

Year:  2022        PMID: 35474515      PMCID: PMC9544451          DOI: 10.1002/sim.9415

Source DB:  PubMed          Journal:  Stat Med        ISSN: 0277-6715            Impact factor:   2.497


INTRODUCTION

The prognosis of patients diagnosed with thin primary cutaneous melanomas, defined as a Breslow tumor thickness 1.0 mm, is generally excellent. The 10‐year melanoma‐specific survival rates range between and . Based on the presence of a large number of long‐term survivors in published Kaplan‐Meier survival curves, one can reasonably consider that some individuals will never experience cancer recurrence and therefore are cured. , Being able to identify cured and non‐cured patients is of high importance to thin melanoma patients for their management decisions and appropriate follow‐up schedules. Multiple studies have analyzed death due to melanoma to evaluate the prognosis or predictive value of patient and tumor characteristics at the time of diagnosis. However, very few studies have analyzed time to melanoma recurrence in this context. To our view, one reason for why time to recurrence is underused is because the exact time of recurrence is challenging to determine (interval‐censored). To alleviate this issue, the time at which recurrence is diagnosed is usually analyzed instead, but the interpretation is still made in terms of the time to recurrence. Alternative methods that model the recurrence time need to be developed and made accessible to the community for more accurate prognospis and predictive analyses. In standard survival analysis, patients who do not experience the event of interest (eg, death or recurrence) during the follow‐up time are simply right censored. However, in some circumstances, some individuals might be cured and therefore would never experience the event no matter the length of follow‐up. Those patients constitute a cured fraction and should be considered as such. In general, a cured fraction is never observed directly; its information is contained in the right‐censored times of a dataset. A common feature for a dataset having a cured fraction is the presence of a group of long‐term survivors. Ignoring the cured fraction, and thus treating all the observed right censored times as true right censoring of event times, may lead to biased inferences. The most popular approach to deal with a cured fraction is called the mixture cure model, and was first proposed by Farewell. This model considers the population of interest to be a mixture of two sub‐populations, one of which is susceptible to the event of interest, and another which is cured. Models for these two components are called latency and incidence models, respectively. The incidence is most commonly modeled using a logistic regression as in Reference 3. For the latency there has been consideration of parametric survival models, as well as much focus on the use of semi‐parametric survival models such as the Cox model. , , , , Such a model is appropriate in cases where there is evidence of a possible cured fraction in the data, based on biological feasibility and the presence of a large number of long‐term survivors in the sample even after a lengthy follow‐up time (see discussion in, for instance, Reference 9). When fitted to a sample without a cured fraction, the mixture cure model reduces to a standard Cox proportional hazards model. In this article, we propose to extend the mixture cure Cox model to accommodate data with a more general partly interval censoring scheme that consists of exact event times and left, interval and right censoring times. We will develop a maximum penalized likelihood (MPL) method to estimate the model parameters, including the logistic regression parameters, the Cox regression coefficients and nonparametric baseline hazard. A penalty function is used for a smooth estimate of the baseline hazard. Parameter estimation for a mixture cure Cox model is difficult to carry out using Cox's partial likelihood. A variety of mixture cure Cox model estimation methods for event times with right censoring have been proposed, including References 10, 11, 12. In more recent work, there has also been consideration of interval censoring. , , , , , Previous research on this topic has only rarely considered a smooth estimate of the baseline hazard function. The method in Reference 18 allows for a piecewise linear approximation to the cumulative baseline hazard of a mixture cure model for interval‐censored data. Alternatively, Corbière et al consider a smooth estimate of the baseline hazard function directly via the use of a penalized likelihood, but was limited to right censored samples only. This article sits within a wider body of research concerning penalized likelihood estimation of proportional hazards models. , , Generally, this approach introduces a roughness penalty term to the likelihood to produce a smooth baseline hazard function estimate that is approximated using a set of non‐negative basis functions. Ma et al demonstrate that a MPL approach can easily incorporate partly interval censored data. To date, there has been extremely limited consideration of methods for fitting a proportional hazards mixture cure model to partly interval censored data that provides a smooth estimate of the baseline hazard function. This article aims to address this gap by drawing on the MPL method for fitting a proportional hazards model to partly interval censored data. The rest of this article is organized as follows. In Section 2, the mixture cure Cox model is introduced. It also explains an approximation of the baseline hazard function using M‐spline basis functions, and presents the penalized likelihood function. In Section 3, we lay out the simultaneous estimation of the regression parameters and baseline hazard function using an alternating algorithm for the constrained MPL estimation. An automatic smoothing parameter estimation procedure is also presented in this section. Section 4 details the asymptotic properties of the estimates. These asymptotic results enable inferences on both regression parameters and survival quantities without computationally intensive methods such as bootstrapping. In Section 5, we report and discuss the results of two simulation studies, and in Section 6, an application to the thin melanoma study is illustrated. Finally, concluding remarks are included in Section 7.

MIXTURE CURE COX MODELS AND PENALIZED LIKELIHOOD

Let be a random variable denoting the time to the event of interest of individual (), where it is possible that some individuals are not susceptible to experiencing the event. Under a scenario of partly interval censored survival data, we may observe event times and also right, left and interval censoring times. In this article, the recorded survival time for each individual will be denoted as a random vector containing , where , and we may have for a right‐censored time and for a left‐censored time. If the event time is observed directly, then we have . Let be an unobserved random variable where indicates that individual is susceptible to the event or is non‐cured (ie, patient will experience the event of interest given a sufficient follow‐up), and indicates otherwise (ie, patient is cured after treatment and will never experience the event of interest regardless of the length of his/her follow‐up). Using a Cox proportional hazards regression model, we can denote the hazard function of in the non‐cured fraction as where is an unknown baseline hazard function, is a vector for the values of of covariates, and is a ‐vector of proportional hazards regression parameters. We may denote simply as when there is no confusion. Additionally, we can model the probability of having using a logistic regression model, such that where is the probability of being non‐cured, is a set of covariates that may be identical to, or be completely different from . It is also possible that share some components with . In (2), is a ‐vector for the logistic regression parameters. Now, we can specify the survival function for the whole population, consisting of both the cured and non‐cured fractions, as Clearly, when (ie, all members of the population are susceptible to the event and there is no cured fraction) the above survival function reduces to that of a standard Cox model. Also, we may denote the conditional survival function simply as when its meaning is clear in the context. For our derivations, it is convenient to denote the information of event, left, right, and interval censoring times using a set of indicator values for each . Let , and be, respectively, the indicators for event time, left, right, and interval censoring time for . Thus, for each , the set of observed information available is . Note that for the survival times corresponding to event, left or interval censoring we have . Conversely, for the right‐censored times their values are unknown. Similar to the method of sieves (eg, Reference 24), the nonparametric baseline hazard function can be approximated using some non‐negative basis functions, where is a positive integer, such that where each coefficient and each is a non‐negative basis function. Let the vector of be denoted by . Here, the baseline hazard function will be approximated using cubic M‐splines, following from previous work on MPL estimation of a proportional hazards model such as References 20 and 23. One convenience of using M‐splines to approximate is that ensuring the estimate of the baseline hazard function is non‐negative requires only that the spline coefficient vector be non‐negative. Another benefit is that it makes the computation of the cumulative baseline hazard function straightforward: , where are I‐splines. M‐splines and their corresponding I‐splines are readily available in, for example, the R splines2 package. The proposed method in this article is to estimate the three parameter vectors, , , and simultaneously by maximizing a penalized likelihood function. Let . Under independent interval censoring (eg, Reference 26), the log‐likelihood is given by Direct maximization of for estimation of is not ideal. This is because: (1) is usually a smooth function and this information should be incorporated into the estimation of ; and (2) it is possible that knots selected to approximate may not be optimal, where particularly some knots may be unnecessary so that their corresponding 's should be zero. We use a penalized log‐likelihood to obtain a smooth estimate for and to force the unnecessary 's close to zero. We adopt a roughness quadratic penalty function and the penalized likelihood is now given by where is a roughness penalty function and is a smoothing parameter. The roughness penalty is given by . Given is now given by (4), we can conveniently express this roughness penalty as , where is an matrix with the th element given by . We comment that in general the above roughness penalty is effective for imposing smoothness on but less ideal for constraining unnecessary to zero. A composite penalty employing both quadratic and ‐norm (equivalent to lasso) penalty could be more efficient. Further investigations are necessary to explore this option.

PENALIZED LIKELIHOOD ESTIMATION

A constrained optimization algorithm

The MPL estimate of , denoted by , is obtained by Given the constraint that , we have the following Karush‐Kuhn‐Tucker (KKT) conditions for a constrained optimal solution: These conditions are solved iteratively using an algorithm similar to the Newton‐MI algorithm of Reference 23. This algorithm requires the score vector and the Hessian matrix for and , but for it only demands its score vector. Details of score vectors and Hessian matrices can be found in the Supplementary Materials. Before describing this algorithm we first introduce some notations. Let , , and be, respectively, the estimates of , , and at iteration . Also, for any function , we let and be respective the positive and negative components of , so that . Iteration of our algorithm is obtained in a three‐step process as follows. First, obtain using a modified Newton algorithm: where is the line search step size used to ensure that . The value of the line search step size can be determined by using, for instance, Armijo's rule. Second, we compute using again a modified Newton algorithm: where is defined similarly to . Finally, we get the update using the multiplicative‐iterative algorithm: where is defined similarly to and and is a diagonal matrix with elements for , and where Referring to the Supplementary Materials for the score vector of , we can see that for the problem considered in this article, becomes Note that is a small constant included simply to avoid the numerical issue of a zero denominator in the calculation of and this value does not have any impact on the final solution for .

Estimation of the smoothing parameter

A marginal likelihood method for the automatic selection of the smoothing parameter, previously outlined in, for example, References 23 and 28, can be implemented to the model of this article. In this method, the penalty function is related to a normal prior distribution for the vector with , where . We can then obtain the log‐posterior: The marginal likelihood for may be difficult to obtain directly, and as such we can approximate it using Laplace's method. Applying the Laplace approximation and substituting in the MPL estimates of , , and , we can obtain the approximated log‐marginal likelihood for : where is the negative Hessian matrix from evaluated at the MPL estimates , , and , and An approximate maximum marginal likelihood solution for is: where , which can be considered as equivalent to the model degrees of freedom. Given that the estimates of , , and depend on , this approximate solution for allows for the development of an iterative procedure with two steps. First, with a current , the corresponding MPL estimates for , , and are obtained. Then, is updated using the current , and the just obtained , , and on the right‐hand side of (13). These two steps are repeated until is stabilized, such as the difference between two consecutive values is less than 1.

ASYMPTOTIC PROPERTIES AND INFERENCE

Development of the asymptotic properties of the proposed model allows for large sample inference to be conducted without reliance on bootstrapping or other computationally intensive methods. Following from Reference 23, it is possible to demonstrate asymptotic consistency for the MPL estimates of both sets of regression parameters, and , and the baseline hazard function . We adopt , , and to denote the true model parameters (ie, the parameters that gave rise to the observed data). Theorem 1 states this asymptotic property; the proofs can be found in the Supplementary Materials. Assume that conditions B1 to B4 (see the Supplementary Materials) hold. Assume that is bounded and has some number derivatives over the interval . Assume that , where . Then, when , almost surely, almost surely, and almost surely. Additionally, it is desirable to develop asymptotic normality results for all three parameters, , , and as this allows for inference to be made not only on regression parameters but also on other quantities, such as survival probabilities. In order to develop these results, however, it is necessary to restrict to be a finite number, similar to References 23 and 29. Note that this fixed is not predetermined as it depends on the given sample size . Usually, a practical guide for is . where denotes the non‐right censored samples size. Another issue we face when developing asymptotic normality results is that we must take into account the possibility of encountering active constraints in the estimation of . This is particularly likely to occur when the number of knots is larger than strictly necessary, or some knots are placed at non‐important locations. The penalty function will push the corresponding to zero. Ignoring active constraints often leads to undesirable results, such as negative variances. Recall that we have defined the parameter vector , which has a length of , and that we can express the penalized likelihood function in terms of such that Let be the MPL estimate of . Let the true value of be represented by . Without loss of generality, we assume that the estimates of first elements of are zero, and so that they are actively constrained. Define where is a matrix of zeros, is an identity matrix. Clearly, is satisfied. Theorem 2 states the asymptotic normality results, with relevant proofs in the Supplementary Materials. Let . Assume that and that we have the first active constraints in the MPL estimate of . Define matrix as above. Let where matrix is defined in ( ). Under these conditions, when , converges in distribution to , where . In order to implement the result of Theorem 2, it is necessary to define a method for identifying active constraints when they arise in the MPL estimation of . The method used here follows that proposed by Ma et al. Active constraints can be identified by inspecting both the value of and the corresponding gradient for each . After the Newton‐MI algorithm has reached convergence, some may be exactly zero with negative gradients, and thus are clearly subject to an active constraint. Furthermore, there may be some that are very close to, but not exactly, zero. For these , a corresponding negative gradient value is indicative that they are also subject to an active constraint. In practice, active constraints are defined where, for a given , and the corresponding gradient is less than , where is a positive threshold value such as . After the indices associated with active constraints are identified, obtaining the matrix is a very straightforward computation. The matrix is obtained by removing the rows and columns of associated with the active constraints. The result is then inverted, and then padded with zeros in the deleted rows and columns to obtain . To make use of these asymptotic results for inference on finite samples, it is necessary to approximate the distribution for when is large. Doing so also incorporates nonzero values for the smoothing parameter into the inference on the parameter estimates. The necessary results are presented below in Corollary 1. Assume that the smoothing parameter . Define Then, when is large, the distribution for the MPL estimate can be approximated by a multivariate normal distribution having mean zero and covariance matrix These results allow for inferences to be made not only on both sets of regression parameters but also on quantities associated with the baseline hazard function; see the results report in Section 5.

SIMULATION STUDIES

The performance of the proposed method is illustrated via two Monte Carlo experiments. The first simulation study evaluates the performance of our proposed MPL method for partly interval censored data with a comparison to the generalized odds rate mixture cure model proposed by Zhou et al, implemented in the GORCure R package, which includes the mixture cure Cox model as a special case. This method uses an expectation‐maximization (EM) algorithm with a gamma‐Poisson data augmentation for the regression parameters, and a spline approximation to a transformed cumulative hazard function. The second simulation study compares the EM based method of Reference 12 with our approach as the former is already implemented in the R package smcure. This method uses an EM algorithm for both Cox and logistic regression parameter estimation and a Breslow estimator for the baseline survival function. However, smcure can only be implemented to right‐censored survival datasets, so this simulation considers only right censoring.

Simulation study design and data generation

In order to generate event or censoring time(s) for individual , we first computed the non‐cured fraction probability given by the following logistic model: where and are logistic regression covariates. Note that the value of controls the size of the cured fraction in the sample. Then, the cured indicator value was obtained according to where . Note that means individual was in the non‐cured fraction (those who will experience the event of interest), and otherwise was in the cured fraction. For in either study, we first simulated a follow‐up time , where could be adjusted to control the extent of right censoring. Afterward, we simulated an event time from a specified distribution. In Study 1, event times were drawn from one of three possible distributions (Weibull, exponential, and log‐logistic). In the smaller Study 2, only a Weibull distribution was considered. For , we simulated a follow‐up time , where has been defined before. Afterwards, an observed survival time, denoted by , was simulated using the procedure described below. For , we took and recorded it as a right censoring time. For , was generated depending on Study 1 or Study 2. In Study 2 (right censoring), was simply generated using . In Study 1 (partly interval censoring), was generated based on a sequence of simulated “examination times” as described below. We first generated a number of “scheduled examinations” from a Poisson(8) (ie, mean 8) distribution, and denote this number by . Next, we simulated uniform random numbers from , and denoted these numbers by . Then, a series of “scheduled examination” times were then given by: . However, the follow‐up time meant not all of these examination times might be used. In fact, when was less than one of the examination times, the examination process would be terminated at and the corresponding final examination times formed time intervals and we then checked if the simulated fell into one of these time intervals. If yes, then was interval censored and interval censoring times were given by the two end‐points of this interval. If no, could be on the left or right of these intervals: if on the left then the minimum examination time was a left censoring time; if on the right then the maximum examination time was a right censoring time. Moreover, for an interval censoring, if the length of the interval was less than some then we recorded corresponding as an event time rather than an interval‐censoring time. Details of all simulation scenarios considered can be found in Table 1 for Study 1 (partly interval censoring) and Table 2 for Study 2 (right censoring). We will report in this article the simulation results associated with the Weibull distribution with sample sizes and and cured fractions sizes 0.2 and 0.6. Results for other distributions, the other sample size (), and the other cured fraction size (0.05) can be found in the Supplementary Materials.
TABLE 1

Simulation study 1 (partly interval censoring) specifications, and the consequent cured and censoring proportions

Scenario 1Scenario 2Scenario 3
Baseline hazard h0(t)=3t2 h0(t)=t h0(t)=4.5t/(1+t2)
Event times distributionWeibullExponentialLog‐logistic
Sample sizes n = 50, 200, 1000 n = 50, 200, 1000 n = 50, 200, 1000
Covariates xi=[xi1,xi2] xi=[xi1,xi2] xi=[xi1,xi2]
zi=[1,zi1,zi2] zi=[1,zi1,zi2] zi=[1,zi1,zi2]
xi1Bern(0.5) xi1Bern(0.5) xi1Bern(0.5)
xi2N(0,1) xi2N(0,1) xi2N(0,1)
zi1Bern(0.5) zi1Bern(0.5) zi1Bern(0.5)
zi2N(0,1) zi2N(0,1) zi2N(0,1)
Parameters γ1=0.5,γ2=1 γ1=0.5,γ2=1 γ1=0.5,γ2=1
β0=3,1.5,0.5 β0=1.5,0.5 β0=1.5,0.5
β1=1,β2=0.5 β1=1,β2=0.5 β1=1,β2=0.5
Cured fraction 1π(z)=0.05,0.2,0.6 1π(z)=0.2,0.6 1π(z)=0.2,0.6
Event proportion in uncured fraction πE=0.65,0.38,0.0 πE=0.65,0.38,0.0 πE=0.65,0.38,0.0
TABLE 2

Simulation study 2 (right censoring) specifications, and the consequent cured and censoring proportions

Scenario 1
Baseline hazard h0(t)=3t2
Event times distributionWeibull
Sample sizes n = 200, 1000
Covariates xi=[xi1,xi2]
zi=[1,xi1,xi2]
xi1Bern(0.5)
xi2U(0,1)
Parameters γ1=0.5,γ2=1
β0=1,0.5
β1=0.5,β2=0.5
Cured fraction 1π(z)=0.27,0.61
Event proportion in uncured fraction πE=0.85,0.5
Simulation study 1 (partly interval censoring) specifications, and the consequent cured and censoring proportions Simulation study 2 (right censoring) specifications, and the consequent cured and censoring proportions We adopted cubic M‐splines (with some knots) to approximate the baseline hazard function. Define and as the minimum and maximum observed survival times, respectively. The observed survival times included interval, left, and right censoring times. External knots for the M‐splines were placed at and . Define and as the respective minimum and maximum of a set of times (called pseudo times) consisting of event times, the mid‐point of any left censoring intervals, and the mid‐point of any interval censoring intervals. The internal knots were placed at equal quantiles across the 5th and 95th percentiles of these pseudo times between and . Following Reference 23, we calculated the number of (internal) knots using a rough guideline of the cubic root of the number non‐right censored individuals. Ma et al remarked that the MPL method is fairly robust (due to the penalty function) to the number of knots as long as the smoothing parameter is appropriate. The smoothing parameter was selected automatically using the marginal likelihood function laid out above in Section 3. Note that we have presented results based on estimates of the marginal (non‐cured fraction only) baseline survival function rather than the marginal baseline hazard function, as quantities of the marginal baseline survival function are more readily available in the GORCure and smcure packages. Computing the marginal baseline survival function and associated asymptotic and Monte Carlo standard errors for the MPL method is straightforward. At time , the MPL estimate of can be computed by taking , where denotes a vector of I‐spline values at . The asymptotic variance estimate can be produced using the delta method, such that , where is the first derivative of with respect to . The Monte Carlo variance can be obtained simply by replacing with the Monte Carlo covariance matrix for in the equation for .

Study 1 (partly interval censoring) results

Table 3 shows the results for the estimation of and for Study 1 (partly interval censoring) for sample sizes and , for the MPL method and the GOR method, using a Weibull distribution for the baseline hazard. It includes, for each parameter, the absolute bias, the relative bias (in brackets beneath the absolute bias), the asymptotic standard error estimate, the Monte Carlo standard error estimate (in brackets beneath the asymptotic standard error), and the asymptotic coverage probability. The MPL method appears to perform reasonably in a variety of the scenarios considered, with small biases, good agreement between the asymptotic and Monte Carlo standard errors, and reasonable coverage probabilities. In particular, when the sample size is small, the MPL estimates for the proportional hazards regression parameters ( and ) consistently have smaller biases than the GOR estimates. This is especially noticeable in the scenarios, that is, when there is a larger number of cured individuals in the sample. Additionally, across all scenarios, the GOR estimate of the has a large negative bias and has very low coverage probabilities, while the MPL estimate of that parameter is reasonable. This equates to an overestimation of the cured fraction size by the GOR method.
TABLE 3

Study 1 (partly interval censoring): Cox and logistic regression parameters for and for Scenario 1 (Weibull baseline)

1π(z)=0.2 1π(z)=0.6
πE=0.65 πE=0.38 πE=0.0 πE=0.65 πE=0.38 πE=0.0
BiasSECPBiasSECPBiasSECPBiasSECPBiasSECPBiasSECP
n=200 β0 MPL0.0490.2110.960.0700.2140.960.0250.2100.97 0.0050.1600.95 0.0210.1600.95 0.0010.1590.95
(0.033)(0.208)(0.046)(0.224)(0.017)(0.210)(0.010)(0.163)(0.041)(0.160)(0.001)(0.163)
GOR 0.4570.2370.50 0.4680.2380.49 0.4970.2420.45 0.5130.2350.42 0.4970.2340.42 0.5100.2140.01
(0.305)(0.227)(0.312)(0.236)(0.332)(0.254)(1.026)(0.242)(0.995)(0.207)(1.019)(0.267)
β1 MPL0.0330.4070.960.0510.4130.940.0430.4060.950.0440.3210.950.0260.3200.960.0170.3200.96
(0.033)(0.411)(0.051)(0.461)(0.043)(0.413)(0.044)(0.335)(0.026)(0.329)(0.017)(0.315)
GOR 0.0140.3940.940.0530.3970.970.0450.4060.950.0210.3160.91 0.0210.3140.960.0180.2870.86
(0.014)(0.422)(0.053)(0.381)(0.045)(0.419)(0.021)(0.336)(0.021)(0.308)(0.018)(0.316)
β2 MPL 0.0200.2000.95 0.0060.2010.95 0.0130.2000.93 0.0220.1690.95 0.0140.1670.94 0.0110.1670.96
(0.041)(0.206)(0.013)(0.201)(0.026)(0.212)(0.044)(0.170)(0.029)(0.170)(0.022)(0.172)
GOR 0.0100.1970.95 0.0040.1970.94 0.0140.2020.94 0.0140.1660.95 0.0100.1650.97 0.0120.1510.86
(0.019)(0.203)(0.008)(0.209)(0.029)(0.213)(0.027)(0.174)(0.019)(0.144)(0.024)(0.175)
γ1 MPL 0.0040.1650.95 0.0100.1650.960.0030.1670.95 0.0080.2380.92 0.0080.2390.930.0010.2370.94
(0.008)(0.167)(0.020)(0.162)(0.006)(0.174)(0.016)(0.254)(0.015)(0.252)(0.001)(0.250)
GOR0.0330.1720.960.0560.1760.95 0.0210.1800.950.0100.2470.940.0080.2560.96 0.4530.2500.71
(0.066)(0.174)(0.112)(0.174)(0.041)(0.183)(0.020)(0.255)(0.016)(0.247)(0.907)(0.171)
γ2 MPL 0.0090.1000.92 0.0020.1010.92 0.0040.1010.93 0.0020.1420.900.0010.1420.890.0020.1400.88
(0.009)(0.108)(0.002)(0.112)(0.003)(0.109)(0.002)(0.162)(0.001)(0.168)(0.002)(0.169)
GOR0.0500.1130.940.0430.1180.960.0320.1170.96 0.0170.1600.930.0090.1640.97 0.9020.1650.07
(0.050)(0.119)(0.043)(0.119)(0.032)(0.112)(0.017)(0.227)(0.009)(0.164)(0.902)(0.325)
n=1000 β0 MPL0.0180.0920.940.0170.0920.960.0140.0920.94 0.0040.0710.940.0050.0700.96 0.0010.0710.95
(0.012)(0.094)(0.011)(0.089)(0.009)(0.092)(0.007)(0.074)(0.009)(0.068)(0.001)(0.072)
GOR 0.4860.1050.02 0.4910.1040.01 0.4840.1070.01 0.4940.1040.00 0.5070.1040.00 0.5040.1040.00
(0.324)(0.109)(0.327)(0.108)(0.323)(0.104)(0.987)(0.096)(1.013)(0.100)(1.008)(0.110)
β1 MPL0.0190.1770.950.0100.1770.95 0.0060.1770.950.0120.1410.950.0050.1410.960.0010.1410.95
(0.019)(0.174)(0.010)(0.176)(0.006)(0.178)(0.012)(0.137)(0.005)(0.133)(0.001)(0.141)
GOR0.0020.1720.950.0020.1720.940.0080.1760.950.0010.1390.950.0010.1390.96 0.0020.1390.93
(0.002)(0.167)(0.002)(0.174)(0.008)(0.178)(0.001)(0.133)(0.001)(0.138)(0.002)(0.141)
β2 MPL 0.0080.0880.95 0.0030.0880.950.0010.0880.95 0.0020.0740.92 0.0050.0740.950.0010.0730.95
(0.015)(0.090)(0.006)(0.088)(0.001)(0.090)(0.004)(0.079)(0.009)(0.075)(0.001)(0.073)
GOR 0.0090.0860.930.0050.0860.980.0020.0880.94 0.0100.0730.96 0.0080.0730.970.0010.0730.93
(0.018)(0.092)(0.011)(0.082)(0.003)(0.089)(0.019)(0.074)(0.015)(0.065)(0.001)(0.073)
γ1 MPL0.0030.0750.94 0.0050.0750.95 0.0060.0760.970.0020.1070.95 0.0010.1070.94 0.0100.1080.95
(0.006)(0.078)(0.009)(0.078)(0.013)(0.075)(0.003)(0.110)(0.001)(0.110)(0.019)(0.109)
GOR0.0050.0750.950.0010.0760.93 0.0010.0770.970.0010.1060.950.0110.1090.93 0.3470.1150.29
(0.009)(0.077)(0.002)(0.082)(0.001)(0.076)(0.001)(0.104)(0.022)(0.107)(0.693)(0.241)
γ2 MPL 0.0080.0460.93 0.0070.0460.94 0.0050.0460.92 0.0080.0630.91 0.0020.0630.92 0.0130.0640.92
(0.008)(0.049)(0.007)(0.046)(0.005)(0.050)(0.008)(0.072)(0.002)(0.073)(0.018)(0.071)
GOR0.0110.0480.970.0010.0500.960.0070.0500.95 0.0140.0670.93 0.0010.0690.97 0.6940.0790.28
(0.011)(0.045)(0.001)(0.051)(0.007)(0.051)(0.014)(0.072)(0.001)(0.063)(0.694)(0.468)

Abbreviations: CP, coverage probability; SE, standard error.

Study 1 (partly interval censoring): Cox and logistic regression parameters for and for Scenario 1 (Weibull baseline) Abbreviations: CP, coverage probability; SE, standard error. Table 4 shows the results for the estimation of the baseline survival function for Study 1 (partly interval censoring) for sample sizes and , for the MPL method and the GOR method, using a Weibull distribution for the baseline hazard. For the MPL method it reports the bias, asymptotic and (Monte Carlo) standard errors, and asymptotic and (Monte Carlo) coverage probabilities. For the GOR method, it reports the bias, (Monte Carlo) standard errors and (Monte Carlo) coverage probabilities, as asymptotic estimates for the standard error are unavailable from the GORCure package. The ability to make fast and efficient inferences on survival model quantities using asymptotic standard errors can be considered as a key strength of the proposed MPL method when compared with the existing alternatives. The estimated survival function was evaluated at three time points, , , and , corresponding to the first, second, and third quartile of the observed survival times, excluding 0 and . Across all scenarios, the bias for the MPL estimate of the baseline survival function was small. There is a fair agreement between the asymptotic and Monte Carlo standard error estimates for the MPL method across the scenarios. We note that in some instances, the coverage probabilities produced by the MPL method are low; specifically, this tends to occur at the later time point where event times are sparser. The baseline survival function estimates from the GOR method tend to have larger biases than the MPL estimates, particularly when . The Monte Carlo standard errors for the GOR estimates are generally larger than both the asymptotic and Monte Carlo standard error estimates from the MPL method, indicating that the MPL method is less variable.
TABLE 4

Study 1 (partly interval censoring): Baseline survival function estimation for and for Scenario 1 (Weibull baseline)

1π(z)=0.2 1π(z)=0.6
πE=0.65 πE=0.38 πE=0.0 πE=0.65 πE=0.38 πE=0.0
BiasSECPBiasSECPBiasSECPBiasSECPBiasSECPBiasSECP
n=200 t1 MPL 0.0050.0290.93 0.0020.0290.93 0.0070.0290.93 0.0020.0370.92 0.0070.0360.90 0.0070.0360.90
(0.033)(0.95)(0.039)(0.98)(0.040)(0.96)(0.038)(0.90)(0.047)(0.93)(0.053)(0.96)
GOR 0.010 0.0330.0660.004 0.0100.397
(0.103)(0.86)(0.102)(0.90)(0.071)(0.96)(0.107)(0.90)(0.100)(0.88)(1.100)(0.89)
t2 MPL0.0020.0200.930.0010.0210.930.0020.0200.93 0.0010.0380.900.0010.0360.900.0030.0360.92
(0.016)(0.89)(0.024)(0.98)(0.031)(0.97)(0.029)(0.90)(0.046)(0.96)(0.129)(1.00)
GOR 0.0260.0480.082 0.0270.050 0.253
(0.113)(0.88)(0.128)(0.87)(0.125)(1.00)(0.150)(0.86)(0.157)(0.89)(0.953)(0.89)
t3 MPL0.0010.0020.890.0010.0020.870.0010.0020.880.0010.0070.830.0020.0060.840.0010.0060.86
(0.001)(0.84)(0.001)(0.88)(0.003)(0.88)(0.004)(0.82)(0.007)(0.83)(0.025)(1.00)
GOR 0.005 0.0060.009 0.014 0.018 0.233
(0.016)(0.94)(0.018)(0.91)(0.021)(0.94)(0.043)(0.88)(0.051)(0.90)(0.889)(0.90)
n=1000 t1 MPL 0.0020.0240.98 0.0010.2110.99 0.0010.0380.94 0.0020.1180.95 0.0010.0200.92 0.0030.1480.96
(0.018)(0.95)(0.031)(1.00)(0.030)(1.00)(0.030)(1.00)(0.037)(1.00)(0.025)(0.98)
GOR0.0260.0130.0820.015 0.003 0.043
(0.113)(0.87)(0.114)(0.91)(0.063)(0.95)(0.104)(0.88)(0.107)(0.91)(0.402)(0.99)
t2 MPL0.0010.0050.920.0010.0050.930.0010.0060.930.0010.0100.91 0.0010.0100.930.0010.0110.92
(0.003)(0.95)(0.005)(0.99)(0.006)(0.97)(0.006)(0.92)(0.013)(0.97)(0.008)(0.94)
GOR0.001 0.0090.047 0.006 0.0190.040
(0.036)(0.83)(0.052)(0.93)(0.053)(0.98)(0.070)(0.90)(0.069)(0.86)(0.343)(0.99)
t3 MPL0.0010.0010.930.0010.0010.940.0010.0010.930.0010.0010.870.0010.0010.860.0010.0010.91
(0.001)(0.89)(0.001)(0.92)(0.001)(0.91)(0.001)(0.88)(0.001)(0.89)(0.007)(0.88)
GOR0.0010.0010.0010.0010.001 0.012
(0.001)(0.94)(0.002)(0.96)(0.002)(0.95)(0.005)(0.97)(0.003)(0.93)(0.322)(0.99)

Abbreviations: CP, coverage probability; SE, standard error.

Study 1 (partly interval censoring): Baseline survival function estimation for and for Scenario 1 (Weibull baseline) Abbreviations: CP, coverage probability; SE, standard error. We have included in the Supplementary Materials some additional simulation study results for partly interval censored data. These include results from scenarios where we have used different distributions for the baseline hazard function (namely, exponential and log‐logistic), scenarios where the sample size was very small (), and scenarios where the cured fraction was very small (approximately of the sample). The exponential and log‐logistic baseline hazard distribution scenarios have generally produced comparable results to those discussed above. When the sample size is very small, both methods tend to produce larger biases in the regression parameter estimates. However, biases and coverage probabilities appear to be more reasonable for the MPL estimates, particularly for the proportional hazards regression parameter estimates ( and ) when the cured fraction in the sample is larger. When the cured fraction was very small, the MPL estimates of regression parameters and the baseline survival function were still reasonable.

Study 2 (right censoring) results

Table 5 exhibits the results for the estimation of and for Study 2 (right censoring) with and for both the MPL and EM methods. It includes, for each parameter, the absolute bias, relative bias (in brackets below absolute bias), the average standard error (asymptotic and bootstrap for, respectively, MPL and EM), the Monte Carlo standard error (in brackets below average SE), and the coverage probability. Note that the EM method in the smcure R package does not provide directly an asymptotic standard error, instead it provides a computationally intensive bootstrapping standard error.
TABLE 5

Study 2 (right censoring): Cox and logistic regression parameters for and for MPL and EM methods

1π(z)=0.27 1π(z)=0.61
πE=0.85 πE=0.5 πE=0.85 πE=0.5
BiasSECPBiasSECPBiasSECPBiasSECP
n=200 β0 MPL0.0460.2510.930.0800.7950.81 0.0080.1720.940.0960.4070.94
(0.046)(0.250)(0.080)(0.836)(0.016)(0.181)(0.192)(0.394)
EM0.0290.4140.95 0.9100.3360.28 0.2620.1940.70 0.2650.4190.89
(0.029)(0.430)(0.910)(0.430)(0.524)(0.253)(0.530)(0.469)
β1 MPL0.0420.5000.95 0.3290.8970.940.0180.3440.960.1280.4620.94
(0.084)(0.462)(0.658)(1.024)(0.035)(0.360)(0.256)(0.469)
EM0.0190.4150.96 0.4240.5460.900.0050.3540.960.0670.4050.95
(0.038)(0.394)(0.847)(0.503)(0.010)(0.358)(0.134)(0.416)
β2 MPL 0.0500.6680.950.3730.6080.68 0.0240.1930.94 0.0930.7180.96
(0.100)(0.714)(0.747)(1.231)(0.048)(0.200)(0.186)(0.731)
EM 0.0170.6950.950.6730.3530.44 0.0010.1990.940.0600.7020.97
(0.034)(0.684)(1.347)(0.303)(0.003)(0.198)(0.120)(0.677)
γ1 MPL 0.0130.1910.930.1640.7950.91 0.0370.1720.760.0490.4010.93
(0.026)(0.199)(0.329)(0.523)(0.074)(0.288)(0.098)(0.392)
EM 0.0200.2070.940.1570.3950.92 0.0290.3150.96 0.0210.3660.96
(0.040)(0.202)(0.315)(0.402)(0.059)(0.290)(0.042)(0.357)
γ2 MPL0.0190.3320.93 0.1930.8970.990.0150.3441.000.1030.6940.92
(0.019)(0.358)(0.193)(0.480)(0.015)(0.176)(0.206)(0.694)
EM0.0280.3650.95 0.2800.2690.830.0100.2000.970.0330.6490.95
(0.028)(0.363)(0.280)(0.309)(0.010)(0.180)(0.066)(0.640)
n=1000 β0 MPL0.0130.0840.950.1320.3250.92 0.0040.0760.940.0370.1050.94
(0.013)(0.085)(0.132)(0.384)(0.008)(0.079)(0.074)(0.134)
EM0.0010.1710.940.8030.1970.02 0.2600.0820.140.2530.1740.66
(0.001)(0.179)(0.803)(0.206)(0.520)(0.109)(0.506)(0.195)
β1 MPL0.0160.1670.930.0500.3420.910.0100.1520.940.0070.1740.96
(0.032)(0.174)(0.099)(0.418)(0.020)(0.153)(0.014)(0.175)
EM0.0140.1670.93 0.3420.2180.570.0070.1520.92 0.0070.1670.96
(0.028)(0.174)(0.683)(0.255)(0.015)(0.154)(0.014)(0.159)
β2 MPL0.0010.2870.95 0.0410.2160.89 0.0170.0840.940.0190.2930.94
(0.002)(0.285)(0.083)(0.339)(0.034)(0.091)(0.038)(0.301)
EM0.0060.2880.950.5450.1250.10 0.0110.0850.940.0260.2910.93
(0.012)(0.285)(1.090)(0.174)(0.021)(0.091)(0.052)(0.296)
γ1 MPL0.0090.0860.940.0040.3251.000.0130.0760.74 0.0080.1590.96
(0.018)(0.088)(0.008)(0.157)(0.026)(0.134)(0.016)(0.145)
EM0.0150.1540.930.0910.1590.900.0050.1280.94 0.0030.2590.97
(0.030)(0.090)(0.181)(0.159)(0.010)(0.137)(0.006)(0.143)
γ2 MPL0.0020.1490.930.0020.3391.00 0.0070.1521.000.0250.2760.96
(0.002)(0.158)(0.002)(0.118)(0.007)(0.075)(0.050)(0.258)
EM0.0150.1540.93 0.1730.0990.560.0050.0800.95 0.0030.2590.97
(0.015)(0.161)(0.173)(0.119)(0.005)(0.078)(0.006)(0.248)

Abbreviations: CP, coverage probability; SE, standard error.

Study 2 (right censoring): Cox and logistic regression parameters for and for MPL and EM methods Abbreviations: CP, coverage probability; SE, standard error. The results from Table 5 indicate the two methods appear largely equivalent under the scenario with cured fraction and , with both methods producing small biases in the parameter estimates, having good agreement between the estimated (asymptotic for MPL and bootstrap for EM) and Monte Carlo standard errors, and giving reasonable coverage probabilities. When the cured fraction is but , the bias for most parameter estimates is fairly large for both methods, although it is generally larger for the EM estimates compared to the MPL estimates, and the EM coverage probabilities for this scenario are particularly low. However, when the sample size is large, the MPL estimates for the proportional hazards regression parameters ( and ) have small biases, while the equivalent estimates for the EM method produce very large biases. However, for these MPL estimates the coverage probabilities are high, reflecting the larger asymptotic SE estimates compared to the Monte Carlo estimates. When the simulated data contained a larger cured fraction, both methods similarly yield little bias in the parameter estimates with the exception of the estimate for . The EM estimate of has a large negative bias and that is less severe in the MPL estimate. We observe from the results that in general the MPL standard error matches well the Monte Carlo standard error for the higher censoring scenario, but less well for the scenario with less censoring, especially for the proportional hazards parameters and . For a large sample size (ie, ), the asymptotic standard error produced by the MPL method are generally close to with the bootstrap standard errors from the EM method, except . However, computations of MPL asymptotic standard errors are much faster than the EM bootstrap standard errors. Table 6 reports the biases, Monte Carlo standard errors and Monte Carlo coverage probabilities for the estimated baseline survival function from both the MPL and EM methods. The estimated survival function was evaluated at three time points, , , and , corresponding to the first, second, and third quartile of the observed event times, excluding 0 and . Note that for comparison purposes only the Monte Carlo standard errors and Monte Carlo coverage probabilities are reported here for both MPL and EM since the smcure package does not provide asymptotic nor bootstrapped standard error estimates for the baseline survival function. From Table 6, it is clear that the MPL method provides better estimates of the survival function in all scenarios than the EM method. MPL gives smaller biases across all the selected time points regardless of scenarios and samples sizes. Also, MPL provides smaller standard errors and better coverage probabilities than EM.
TABLE 6

Study 2 (right censoring): Baseline survival function estimation for and for MPL and EM estimation methods

1π(z)=0.2 1π(z)=0.6
πE=0.65 πE=0.38 πE=0.65 πE=0.38
BiasSECPBiasSECPBiasSECPBiasSECP
n=200 t1 MPL0.0020.0320.94 0.0030.0150.95 0.0020.1001.000.0130.0520.94
EM0.0490.0540.86 0.0180.0210.85 0.0530.1200.990.0440.0850.92
t2 MPL0.0020.0390.95 0.0140.0560.95 0.0040.1261.000.0200.0740.92
EM0.0840.0800.81 0.1060.0710.67 0.0560.1190.980.0750.1320.91
t3 MPL0.0030.0360.95 0.0160.1070.940.0010.0190.980.0270.0950.94
EM0.0880.0830.81 0.1970.1090.55 0.0050.0130.930.0900.1470.91
n=1000 t1 MPL 0.0010.0140.95 0.0010.0050.94 0.0020.0881.000.0060.0230.96
EM0.0490.0240.450.0170.0100.61 0.0700.1021.000.0500.0340.68
t2 MPL 0.0010.0170.960.0010.0210.960.0020.0481.000.0110.0360.96
EM0.0840.0370.38 0.1040.0330.11 0.0270.0320.930.0840.0560.69
t3 MPL 0.0010.0160.950.0090.0470.940.0010.0010.990.0150.0440.96
EM0.0910.0390.34 0.1940.0490.03 0.0010.0010.960.0910.0630.73

Abbreviations: CP, coverage probability; SE, standard error.

Study 2 (right censoring): Baseline survival function estimation for and for MPL and EM estimation methods Abbreviations: CP, coverage probability; SE, standard error.

APPLICATION TO MELANOMA RECURRENCE DATA

Our new method was applied to the real dataset introduced in Section 1. Data from a cohort of 2968 patients diagnosed with thin melanoma (defined as patients diagnosed with a Breslow thickness 1 mm) between January 1992 and December 2014 were extracted from the prospectively maintained research database at the Melanoma Institute Australia (MIA). Information collected included baseline characteristics (age, sex, and body site of lesion), pathological factors (Breslow thickness, ulceration, and mitoses count), date of follow‐up visit, date of diagnosis of first recurrence (either local, regional, or distant), and survival status of patient at last contact. All patients with known date of recurrence or last survival status were analyzed. All patients had given consent for use of their de‐identified information for research purposes. Ethical approval was provided by the Sydney Local Health District Ethics Committee. The primary clinical outcome was time to first melanoma recurrence calculated from the date of the initial primary diagnosis. Patients who experienced melanoma recurrence were subject to interval censoring, as the exact time of recurrence was unknown. The time interval in which the recurrence occurred was denoted where was the preceding follow‐up visit date of the recurrence, also called the last known recurrence free time, and was the date recurrence was first diagnosed. Some patients did not have recorded values of and , but had a status at their last follow up of either “alive with melanoma” or “dead, melanoma,” indicating that they had experienced a tumor recurrence at some point and thus left‐censored. These patients therefore had a left censoring interval of (0, ), where and denote, respectively, the starting and last follow‐up time. All other patients were right‐censored at the time of their last follow up. Six categorical covariates were considered in this analysis and coded as: Breslow thickness (0.8 mm; 0.8 mm), tumor ulceration (yes; no), age group (50; 50), sex (male; female), tumor mitoses (yes; no), and site of tumor (arm; head & neck; leg; trunk). Thin melanoma recurrence was rare, with only 6.69% of the sample having experienced the event at some point during the follow‐up time. The remaining individuals were right‐censored. This prevalence of right‐censored observations suggested that there was likely a cured fraction. A mixture cure Cox model was fitted with all six covariates in the latency model and the incidence model. The incidence and latency models were then reduced using backwards step‐wise selection with ‐values, until all covariates left in the model were significant at the 5% level. Table 7 shows the odds ratios (from the incidence logistic regression model) and hazard ratios (from the latency proportional hazards model), 95% confidence intervals (CIs) and ‐values after implementing our proposed method. Results from the incidence model indicate that the odds of thin melanoma recurrence increase significantly in patients with Breslow thickness 0.8 mm, with mitoses, or who are male. Conversely, the odds of thin melanoma recurrence are significantly lower in patients who had a tumor on the leg or the trunk, compared to on the arm (the reference category). The most striking feature of the latency model for non‐cured patients was that Breslow thickness was not significantly associated with recurrence. Tumor ulceration and having a tumor on the head & neck, the leg or the trunk instead of the arm significantly increased the risk of melanoma recurrence. Sex was also significantly associated with risk of recurrence, with males in the non‐cured population having a significantly lower risk of melanoma recurrence than females. Age was not significant in either the incidence or the latency model.
TABLE 7

Model fitting results for the thin melanoma data

MPL mixture cure modelGOR mixture cure modelStandard Cox model
CovariateOR/HR95% CI P‐valueOR/HR95% CI P‐valueHR95% CI P‐value
Incidence model
Breslow thickness: >0.8 mm1.98(1.12, 3.48)0.019
Ulceration: Yes3.01(2.34, 3.88)0.015
Sex: Male7.57(3.76, 15.22) <0.001
Mitoses: Yes2.86(1.66, 4.95)0.001
Body site: Leg0.28(0.12, 0.62)0.002
Body site: Trunk1.93(1.27, 2.95)0.002
Latency model
Breslow thickness: >0.8 mm2.24(1.30, 3.86)0.0041.59(1.05, 2.42)0.029
Ulceration: Yes2.38(1.20, 4.73)0.0142.38(1.14, 4.97)0.022
Sex: Male0.15(0.09, 0.28) <0.001
Mitoses: Yes3.29(1.95, 5.54) <0.0012.25(1.46, 3.47) <0.001
Body site: Head & Neck2.80(1.34, 5.86)0.0062.92(1.57, 5.44) <0.0012.13(1.23, 3.66)0.007
Body site: Leg4.98(1.94, 12.82) <0.001
Body site: Trunk2.46(1.34, 4.48)0.0031.89(1.25, 2.86)0.003

Abbreviations: CI, confidence interval; HR, hazard ratio; OR, odds ratio.

Model fitting results for the thin melanoma data Abbreviations: CI, confidence interval; HR, hazard ratio; OR, odds ratio. Interpretation of the outputs of a mixture cure model demands some extra attention. Particularly, notice should be paid to the fact that the latency model is a conditional model, only relevant to the non‐cured sub‐population. For example, the latency model of this melanoma example suggests that, if we consider at the non‐cured (ie, melanoma recurrence) sub‐population, males have lower risk of recurrence than females. However, the results from the incidence model suggest males have higher odds of being classified into the non‐cured sub‐population. These two interpretations of sex in the incidence and latency model do not contradict each other. Corbière et al similarly found that some parameter estimates for the same covariates had different signs between the latency and the incidence models in their analysis of tonsil tumor recurrence data. Table 7 also contains the results of fitting a proportional hazards mixture cure model using the GORCure package, and the Cox model results without a cured fraction (standard Cox model), fitted using the survivalMPL R package. The same six covariates were considered, and the same backwards step‐wise selection was carried out to select significant covariates. There are substantial differences between the MPL and GOR fitted models. There was no overlap between the MPL and GOR incidence models in terms of significant predictors of recurrence susceptibility, and only one (body site: head & neck) shared between the MPL and GOR latency model for risk of recurrence. Notably, according to the GOR model there was no significant difference between males and females for either incidence or latency, while sex was significant in both the latency and incidence MPL models. It is of interest to compare the results of the standard Cox model results with the latency model from the mixture cure model. For the standard Cox model, there was no significant difference between males and females in terms of melanoma recurrence risk, which contradicts the latency model result. Also, for the standard Cox model, Breslow thickness and mitoses are significant, but were not significant in the latency model. The magnitudes of the hazard ratios and the significance of the body site categories also differ between the latency sub‐model and the standard Cox model. Age group was not significant in either the latency or the standard Cox model. The estimate of the baseline hazard function for the melanoma recurrence sub‐population can be seen in Figure 1. There is a notably rapid drop in the risk of tumor recurrence over the first 3 to 4 years, after which the risk of recurrence decreases slowly. Between 20 and 25 years there appears to be a small increase in the risk, although the point‐wise confidence intervals are wide for this time period. Figure 2 exhibits the baseline survival function with 95% point‐wise CIs for the melanoma recurrence sub‐population. For this group of individuals, their baseline survival function decreases faster over the first five years, and then more slowly for the remainder of the follow‐up period.
FIGURE 1

Estimated baseline hazard function for the non‐cured population (with point‐wise 95% confidence intervals)

FIGURE 2

Estimated conditional baseline survival function for susceptible sub‐population (with point‐wise 95% confidence interval)

Estimated baseline hazard function for the non‐cured population (with point‐wise 95% confidence intervals) Estimated conditional baseline survival function for susceptible sub‐population (with point‐wise 95% confidence interval) The ability to make predictions based on an estimated mixture cure survival function is a key strength of the method presented in this article. In fact, values of the mixture cure survival function can be easily computed from our regression parameter estimates (both incidence and latency) and the estimate for the conditional baseline hazard function. For example, the estimated probability of a person with a Breslow thickness 0.8 mm (all other covariates are fixed at their mean values) having no recurrence for 5 years is 0.95 (95% CI: 0.93, 0.96), while this probability for a person with a Breslow thickness 0.8 mm is down to 0.91 (95% CI: 0.86, 0.94). Furthermore, the probabilities of a patient having no recurrence for 10 years are 0.93 (95% CI: 0.91, 0.94) and 0.88 (95% CI: 0.82, 0.93), respectively, for Breslow thickness 0.8 mm and 0.8 mm. Similarly, based on the fitted mixture cure model, we can also easily compute entire predictive survival functions, and their point‐wise CIs. For example, Figure 3 displays mixture cure predictive survival functions illustrating the effect of Breslow thickness (top panel) and ulceration (bottom panel) with 95% point‐wise CIs. These plots explain that, at the population level, those with a Breslow thickness 0.8 mm are more likely to be free of recurrence at any time than those with a Breslow thickness 0.8 mm, when all other covariates are set to the sample mean values. Similarly at the population level, those with no ulceration are more likely to be free of recurrence at any time than those with ulceration. However, the survival differences between these groups may not be significantly different, as the associated point‐wise CIs slightly overlap.
FIGURE 3

Estimated mixture survival functions for Breslow thickness 0.8 mm vs 0.8 mm and ulceration vs no ulceration, with point‐wise 95% confidence intervals

Estimated mixture survival functions for Breslow thickness 0.8 mm vs 0.8 mm and ulceration vs no ulceration, with point‐wise 95% confidence intervals We calculated again the predicted survival curves using GOR model from GORCure as well as the standard Cox model (so there was no cured fraction). The mixture cure survival functions from the MPL and GOR models and the standard Cox survival curves are displayed in Figure 4, comparing again the Breslow thickness groups (top panel) and the ulceration groups (bottom panel). The estimated survival functions from the GOR model are noticeably higher than their MPL counterparts. As seen in Section 5, the GOR model produces consistently negative bias in the estimation of the intercept of the incidence logistic regression model, which is equivalent to over‐estimating the size of the cured fraction and underestimating the risk of the event in the whole population. As a result, it appears that the survival function estimates produced by the GOR mixture cure model are too high. There are also clear discrepancies between the Cox and mixture cure predictive survival functions. Clearly, failing to account for the presence of a cured fraction also leads to an overestimation of survival probabilities in this thin melanoma example.
FIGURE 4

Estimated survival functions for Breslow thickness 0.8 mm vs 0.8 mm and ulceration vs no ulceration using the MPL Cox mixture cure model, GOR Cox mixture cure model, and a standard Cox model

Estimated survival functions for Breslow thickness 0.8 mm vs 0.8 mm and ulceration vs no ulceration using the MPL Cox mixture cure model, GOR Cox mixture cure model, and a standard Cox model

DISCUSSION AND CONCLUDING REMARKS

In this article, we propose a new penalized likelihood estimation method for fitting a mixture cure Cox model where observations are assumed partly interval‐censored. Our approach finds the MPL estimation of regression parameters for the latency and incidence models as well as estimation of the latency baseline hazard function which is approximated using M‐spline basis functions. The baseline hazard estimate is constrained to be non‐negative. One advantage of our method, when compared with the existing EM methods, is that it also yields an asymptotic covariance matrix for all the parameters, and hence allowing inference on both regression parameters and survival quantities. Another advantage of our method is that it produces smooth baseline hazard estimates. The results of simulation studies indicate that this method produces regression parameter and baseline hazard function estimates that perform well in terms of bias, variance and coverage probability. Our method avoids computationally intensive methods like bootstrapping which is adopted by the competitor EM‐algorithm for computing variances of the estimates. A package to implement the proposed method is available on Github, and we intend to upload the package to R CRAN in the near future. Existing methods for checking the diagnostics of mixture cure models are sparse, particularly for cases involving partly interval censored data. Schoenfeld residuals are inappropriate for mixture cure models as the marginal hazards of any mixture cure model will be nonproportional. Wileyto et al developed pseudo‐residuals modeled on Schoenfeld to assess the fit of the latency model in the non‐cured fraction, but considered parametric mixture cure models for right censored data only. Peng and Taylor developed a modified martingale residual and a modified Cox‐Snell residual appropriate for checking the latency model of a mixture cure model, but again only considered right censored samples. Scolas et al developed a Cox‐Snell residual applicable to a parametric mixture cure model for interval censored data, which can be used to check both the uncured and mixture population survival distributions. These authors also developed a deviance residual appropriate for checking the linearity of both the incidence and latency parts of the model. However, the methods developed by Scolas et al are only appropriate for cases where the data is entirely right‐ or interval‐censored. Evidently, more work is required to derive appropriate residuals for diagnostic checks of the mixture cure Cox model proposed here. We plan to extend the approach discussed in the article to other mixture cure survival models, particularly mixture cure additive hazards (AH) models and mixture cure accelerated failure time (AFT) models, where partly interval censoring times will be adopted. MPL estimations of AH and AFT models from partly interval censored survival times have been explored by Li and Ma, , and these computational algorithms will be extended to their mixture cure counterparts. Data S1 Supplementary Materials Click here for additional data file.
  22 in total

1.  Hazard regression for interval-censored data with penalized spline.

Authors:  Tianxi Cai; Rebecca A Betensky
Journal:  Biometrics       Date:  2003-09       Impact factor: 2.571

2.  Parametric versus non-parametric methods for estimating cure rates based on censored survival data.

Authors:  A B Cantor; J J Shuster
Journal:  Stat Med       Date:  1992-05       Impact factor: 2.373

3.  A penalized likelihood approach for mixture cure models.

Authors:  Fabien Corbière; Daniel Commenges; Jeremy M G Taylor; Pierre Joly
Journal:  Stat Med       Date:  2009-02-01       Impact factor: 2.373

4.  Application of the theory of finite mixtures for the estimation of 'cure' rates of treated cancer patients.

Authors:  N H Gordon
Journal:  Stat Med       Date:  1990-04       Impact factor: 2.373

5.  Comparison of parametric and non-parametric survival methods using simulated clinical data.

Authors:  J W Gamel; R L Vogel
Journal:  Stat Med       Date:  1997-07-30       Impact factor: 2.373

6.  Semiparametric transformation models for interval-censored data in the presence of a cure fraction.

Authors:  Chyong-Mei Chen; Pao-Sheng Shen; Wei-Lun Huang
Journal:  Biom J       Date:  2018-11-25       Impact factor: 2.207

7.  Development and Validation of Nomograms to Predict Local, Regional, and Distant Recurrence in Patients With Thin (T1) Melanomas.

Authors:  Mary-Ann El Sharouni; Tasnia Ahmed; Alexander H R Varey; Sjoerd G Elias; Arjen J Witkamp; Vigfús Sigurdsson; Karijn P M Suijkerbuijk; Paul J van Diest; Richard A Scolyer; Carla H van Gils; John F Thompson; Willeke A M Blokx; Serigne N Lo
Journal:  J Clin Oncol       Date:  2021-02-18       Impact factor: 44.544

8.  Computationally Efficient Estimation for the Generalized Odds Rate Mixture Cure Model with Interval-Censored Data.

Authors:  Jie Zhou; Jiajia Zhang; Wenbin Lu
Journal:  J Comput Graph Stat       Date:  2018-02-01       Impact factor: 2.302

9.  Penalized likelihood estimation of a mixture cure Cox model with partly interval censoring-An application to thin melanoma.

Authors:  Annabel Webb; Jun Ma; Serigne N Lô
Journal:  Stat Med       Date:  2022-04-26       Impact factor: 2.497

View more
  1 in total

1.  Penalized likelihood estimation of a mixture cure Cox model with partly interval censoring-An application to thin melanoma.

Authors:  Annabel Webb; Jun Ma; Serigne N Lô
Journal:  Stat Med       Date:  2022-04-26       Impact factor: 2.497

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.