Literature DB >> 34819988

Joint Modelling Approaches to Survival Analysis via Likelihood-Based Boosting Techniques.

Colin Griesbach1, Andreas Groll2, Elisabeth Bergherr1.   

Abstract

Joint models are a powerful class of statistical models which apply to any data where event times are recorded alongside a longitudinal outcome by connecting longitudinal and time-to-event data within a joint likelihood allowing for quantification of the association between the two outcomes without possible bias. In order to make joint models feasible for regularization and variable selection, a statistical boosting algorithm has been proposed, which fits joint models using component-wise gradient boosting techniques. However, these methods have well-known limitations, i.e., they provide no balanced updating procedure for random effects in longitudinal analysis and tend to return biased effect estimation for time-dependent covariates in survival analysis. In this manuscript, we adapt likelihood-based boosting techniques to the framework of joint models and propose a novel algorithm in order to improve inference where gradient boosting has said limitations. The algorithm represents a novel boosting approach allowing for time-dependent covariates in survival analysis and in addition offers variable selection for joint models, which is evaluated via simulations and real world application modelling CD4 cell counts of patients infected with human immunodeficiency virus (HIV). Overall, the method stands out with respect to variable selection properties and represents an accessible way to boosting for time-dependent covariates in survival analysis, which lays a foundation for all kinds of possible extensions.
Copyright © 2021 Colin Griesbach et al.

Entities:  

Mesh:

Substances:

Year:  2021        PMID: 34819988      PMCID: PMC8608498          DOI: 10.1155/2021/4384035

Source DB:  PubMed          Journal:  Comput Math Methods Med        ISSN: 1748-670X            Impact factor:   2.238


1. Introduction

First suggested by [1], joint models were established as a valuable tool for analysing data where event times are measured alongside a longitudinal outcome. One naive approach of evaluating such frequently occurring data structures would be separate modelling, i.e., fitting appropriate models independently for longitudinal and time-to-event data. However, separate modelling neither corrects for event-dependent dropout in longitudinal analysis nor quantifies the relation between a time-dependent covariate and the risk for an event in survival analysis [2]. Various approaches have been proposed to address these issues, one being the Andersen-Gill model [3] for time-varying covariates in survival analysis or two-stage approaches, where longitudinal model fits are included as fixed covariates in time-to-event regression. It has been shown, however, that these methods tend to produce biased results [4, 5]. One solution therefore is combining both the survival and longitudinal models within one single joint likelihood. A wide introduction to this joint modelling framework is presented in [4] including the JM package [6]. Moreover, an evolution of joint model progression up to the year 2004 is provided in [7], and in addition, several Bayesian approaches have been carried out [8-10]. Current joint modelling estimation methods, however, lack clear concepts for proper variable selection and good performance regarding prediction. Moreover they are not feasible for high-dimensional data, in particular where the number of covariates exceeds the number of observations, i.e., p > n problems. In order to overcome these hindrances, an algorithm was initially proposed, where joint models are fitted with gradient boosting techniques, which are known for addressing exactly these issues [11]. Evolved from machine learning as an approach to classification problems originally proposed in [12], gradient boosting deals with high-dimensional data and the component-wise updating scheme offers implicit variable selection. The boosting algorithm for joint models was extended [13], but when wanting to lay more focus on the survival side of the model, gradient boosting proved to struggle with time-varying covariates in time-to-event analysis. This has also been observed for pure survival models in [14]. Hence, this work focuses on likelihood-based boosting. First introduced in [15], likelihood-based boosting is designed to directly maximize a given likelihood where concepts, which gradient boosting implicitly offers, are reproduced artificially for regular optimization methods like Newton algorithms or Fisher scoring. The method was further developed for flexible semiparametric mixed models [16] and for several classes of generalised mixed models [17-20]. The R package GMMBoost [21] covers most of these approaches. Likelihood-based boosting has also been proved useful for survival analysis with time-varying effects [22], and a general overview is given in [23]. Since the random structure plays an important role as a connector between longitudinal and time-to-event data, we additionally incorporate a novel correction step within the estimation procedure for the random effects, which was first suggested in [24, 25] and reduces possible bias arising from wrongly identified random effects. The contribution of this work is the novel lbbJM boosting algorithm for joint models, which offers the first boosting-based regularization approach for time-dependent covariates in survival analysis and in addition new variable selection mechanics for joint models with focus on time-to-event analysis. The remainder of this manuscript is structured as follows: Section 2 highlights the overall concepts of both joint modelling and likelihood-based boosting to give a sufficient understanding of the methods used in the following parts. Section 3 then contains a detailed description of the considered joint model together with the proposed boosting algorithm and its computational details. Sections 4 and 5 deal with applying the algorithm to different setups of simulated data as well as to the AIDS dataset [26] included in the JM package. Results and possible extensions are discussed in the final section.

2. Backgrounds

Before the algorithm is presented and discussed in detail, we briefly highlight the concepts of both joint modelling and likelihood-based boosting.

2.1. Joint Models

In general, a joint model consists of two parts, one longitudinal and one survival submodel. A popular view on joint modelling is to choose one model as the main model, whereas the other model then features the analysis occurring in the main model. With the primary outcome being longitudinal data, a survival model can be used to correct for event-dependent dropout in longitudinal analysis. For time-to-event data as outcome of interest, additional longitudinal modelling reduces measurement error on the one hand and, on the other hand, extrapolates only on single time points observed longitudinal data to continuous functions which are then included in survival analysis. We will from now on focus on joint models with time-to-event data as primary outcome. The longitudinal submodel takes the form: where longitudinal outcome y is described by the longitudinal predictor function ηl depending on time t and a set of covariates x. Although t can be included in x, we will highlight it in the context of joint models, as the role of t is of greater importance. In the survival submodel, the hazard is modelled by a baseline hazard λ0(t) with multiplicative effects described by the survival predictor function ηs. In addition, the longitudinal predictor ηl is reappearing in the survival model, this time scaled by a factor α. The parameter α thus quantifies the association of the two submodels and is therefore called the association parameter. It can be interpreted as the impact a time-varying longitudinal covariate has on the hazard for an event. Parameter estimation for such joint models can be done in various ways. Two common methods are two-stage and joint likelihood approaches, respectively. In the former, the longitudinal model is estimated with the estimation method of choice leading to the model fit , which is then carried forward as fixed covariate into survival analysis. In the latter, longitudinal and survival submodels are combined in a single joint likelihood. Let i = 1, ⋯, n denote clusters and j = 1, ⋯n the repeated measurements. Assuming independent data generating processes for both submodels, the joint likelihood can then be written as with densities fl and fs for the longitudinal and survival submodels and time-to-event outcome (T, δ) = (T, δ). Regular inference is now done by maximizing (3) using appropriate maximization methods, the most prominent one being EM algorithms.

2.2. Likelihood-Based Boosting

We intend to give a short description of the underlying mechanics used in the following section. The overall concept of likelihood-based boosting is to create an iterative and component-wise updating scheme, which eventually converges to a maximum likelihood estimator but is stopped early in order to prevent overfitting. Let β model be the effect of p covariates. Likelihood-based boosting maximizes a given log-likelihood l(β) by component-wise Fisher scoring in the following way: For each covariate r ∈ {1, ⋯, p} consider the subvector β containing only the coefficients referring to the rth covariate. We compute the score vector and Fisher matrix as and obtain a possible update for the rth component. Now, we determine the best performing covariate with respect to likelihood maximization, i.e., find the component where the corresponding update yields the biggest improvement of the likelihood. One receives a new model fit by weakly updating this best performing component, i.e., by scaling with a factor ν, the so called step length: The step length ν is controlling the weakness of the update to prevent overfitting and give every covariate a chance for selection. A popular choice in the literature is ν = 0.1. Repeating this updating process for a sufficiently large number of iterations leads to the regular maximum likelihood estimator for β. But instead the algorithm is stopped early to gain better prediction performance and variable selection. The optimal amount of iterations actually is a tuning parameter of the method and can be determined via cross-validation or by focusing on information criteria like AIC or BIC [27, 28].

3. Boosting Joint Models

3.1. The Model

Before we introduce the boosting algorithm, we describe the specific joint model. With i = 1, ⋯, n denoting the individual and j = 1, ⋯, n a single specific measurement, the longitudinal submodel is given by where y is modelled by ηl depending on specific measurement times t and longitudinal time-independent covariates xl ∈ ℝ. This represents a standard linear mixed model with intercept β0 and fixed linear effects β and βl of time and baseline covariates as well as individual specific random effects γ0 and γ with (γ0, γ) ~ 𝒩⊗2(0, Q). The error terms ε are assumed to follow a normal distribution with 𝔼[ε] = 0 and Var(ε) = σ2 > 0. Please note that the model can be additionally extended to interaction effects of time t with baseline covariates x ∈ ℝ by including the term βxt in (8). This results in slightly adjusted integrals in the survival part and is omitted in the following for the sake of better readability. In the survival part, the individual hazard is modelled with the survival predictor ηs(xs) = βsxs containing additional linear effects βs of baseline covariates xs ∈ ℝ. To execute a full likelihood approach, the baseline hazard is chosen to be piecewise-constant depending on the number of segments K and their exact locations I = [t, t) with t0 = 0 and t = max(T) for k = 1, ⋯, K. The collection of values for the baseline hazard is denoted in λ = (λ). Later, we will choose K between 7 and 10 in order to guarantee substantially more flexibility than a constant baseline hazard without becoming computationally too demanding. Given two formulas (8) and (9), we can now calculate the joint log-likelihood. Let y = (y) denote the collection of all longitudinal measurements. Assuming the time-to-event process is conditionally independent from the longitudinal random structure, the joint likelihood can be decomposed into a longitudinal and a survival term. Set γ = (γ) with γ = (γ0, γ) and ϑl≔(β0, βl, β, γ). Furthermore, τ contains information on variance-covariance components σ2 and Q. The unpenalized longitudinal log-likelihood is where ϕ(·∣m, v) denotes the density of a normal distribution with mean m and variance v. Laplace approximation follows [29] and then leads to an additional quadratic penalty term for the random effects yielding the penalized log-likelihood: Note that for the penalized log-likelihood τ substitutes σ2 as an argument, since the penalized log-likelihood additionally contains information of the variance matrix Q. On the other hand, for given survival outcome (T, δ) = (T, δ) with event times T and censoring indicator δ, the survival log-likelihood takes the well-known form: with ϑs≔(λ, α, βs). For ϑ≔(ϑl, ϑs), we finally receive the penalized and unpenalized joint log-likelihood:

3.2. The Algorithm

The following lbbJM algorithm (likelihood-based boosting for joint models) describes a way to fit the formulated joint model by likelihood-based boosting methods discussed in Section 2.2.

3.3. Computational Details of the Algorithm

In general, the algorithm carries out a hybrid between a two-stage and a joint likelihood approach. For one single tuple (ml, ms), the fitting procedure goes as follows: In a first step, the longitudinal submodel is boosted up to ml iterations using the lbbLMM boosting algorithm [25]. The received estimates are carried forward into the survival model, where another boosting process up to ms iterations takes place. This fitting process is carried out for any tuples (ml, ms) with ml < mmax,l, ms < mmax,s, where mmax,l and mmax,s are prespecified maximum numbers of iterations per submodel. For every of these possible combinations of stopping iterations, the corresponding estimates are evaluated based on the joint likelihood using test data, which can be achieved via cross-validation or bootstrapping. Hence, the algorithm uses two-stage fitting but joint likelihood evaluation. We give a detailed description for both parts. Exact formulas for all appearing variants of score vectors and Fisher matrices can be found in the supplementary material. Please note that due to the component-wise updating scheme in both submodels, the lbbJM algorithm works with arbitrarily high numbers of candidate variables and is therefore not confined to low-dimensional data structures. For starting values, the parameters, which actually underlie the boosting process, are necessarily set to zero, thus . The baseline hazard is initialized with the intercept estimator . The remaining values are extracted from a standard linear mixed model for time and random effects by using, e.g., the function lme from the R package nlme. For boosting longitudinal fixed effects, for convenience, we omit the iteration index as well as the hat indicating estimated values in the following subsections. In the first step of the longitudinal part, the effects βl follow the classical component-wise likelihood-based boosting procedure. In each iteration, an update for every single covariate r together with intercept β0 and time effect β is computed, leading to pl three-dimensional updates. Selection of the best performing component is then performed either by selecting the component yielding the optimal likelihood maximization or lowest information criteria like AIC or BIC, which minimizes the model complexity rather than residuals. The linear effect β is excluded from the selection process, as it holds a very important role in a joint model and should always be included. For updating random effects, after the best performing fixed effect from βl was updated, in every iteration, an update for the random effects is executed separately. This means that the score vector and Fisher matrix have to be derived in order to execute the update The matrix C is a correction matrix which prevents from potential correlations between the random effects estimates and any observed covariates [25], and its derivation can be traced in more detail in the supplementary material. For updating variance-covariance components, the covariance matrix Q of the random effects is updated with an approximate EM algorithm using the posterior curvatures F of the random effects model [30]. Receive an update by computing The current longitudinal model error is obtained by setting Var(y − ηl). For boosting the association parameter and survival effects, once the longitudinal part was updated in up to ml iterations, the algorithm proceeds to boost the effects βs and α. Although being of different structures, the association parameter α is boosted alongside the effects βs, meaning the algorithm decides whether the association or some baseline survival effect is updated, based on which parameter leads to the best likelihood improvement. This means, the linear effect of the whole longitudinal trajectory is also scaled by the step length ν when being updated within the selection step, which minimizes the chance of potential overfitting also for the association parameter. An alternative method would be choosing just from the effects in βs and updating α in an additional step by optimizing the current likelihood. This approach was used in [11] and treats α as a nuisance parameter, which might not be satisfactory with regards to the importance of α. Again, only the update for α and βs is scaled by the step length νs. The baseline hazard λ receives a full update. For step lengths, apart from the stopping iterations, the step lengths are tuning parameters of the boosting algorithm. Although there is some effort in focusing on adaptive step lengths recently, we chose to set both step lengths to the constant value νl = νs = 0.1. The exact choice of the step length factor is of minor importance as long as it is sufficiently small to ensure proper performance. Setting it to 0.1 is an established choice in the boosting literature [31, 32]. For stopping iterations, since the step lengths are chosen to be constant, the tuple (ml, ms) is the main tuning parameter of the boosting algorithm. In regular boosting with only one iteration index, it is convenient to check for every single iteration and take the estimates from the estimation count leading to the best prediction. In the present two-dimensional case, this would mean finding with ℳ≔{1, ⋯, mmax,l} × {1, ⋯, mmax,s}, denoting the vector of estimates received via the tuple (ml, ms) of total iterations and 𝒳test a complete set of test data for evaluation. Problem (19) is then solved via k-fold cross-validation. But since checking for every single tuple (ml, ms) ∈ ℳ implies a very high computational effort, we suggest to coarsen the grid and check for fewer possible stopping iterations in the longitudinal part, e.g., ml ∈ {10,20,30, ⋯, mmax,l}. Because of the two-stage-approach nature of the algorithm, we still can check for every single ms ∈ {1, ⋯, mmax,s} without gaining additional computational effort. Furthermore, parallel computing can be executed in order to minimize computational demand.

4. Simulations

We evaluate the lbbJM algorithm with a simulation study. The aim is to assess estimation and shrinkage characteristics in general as well as variable selection properties and performance in high dimensional, i.e., p > n settings. The lbbJM algorithm is included in two variants. While lbbJMa executes the full approach as depicted in Section 3.2, lbbJMb performs a shortened two-stage procedure where the longitudinal submodel is fitted in advance using regular maximum likelihood inference and does not underlie any regularization. The exact lbbJMb algorithm is depicted in detail in the supplementary material. We additionally include the JM package as state of the art for convenient estimation of joint models as well as the glmnet package, which offers elastic net penalization for start-stop-data and therefore an alternative approach for regularization of time-dependent covariates in survival analysis. None of the competitors are completely suitable for a benchmark comparison and are viewed as reference points for the specific objectives addressed by the lbbJM algorithm. Regarding glmnet, as an alternative approach to regularization of time-dependent covariates, shrinkage and variable selection properties are of interest, although it focuses solely on survival analysis. JM in addition offers unregularized effect estimates with corresponding significance indicator but is neither suitable for high-dimensional setups nor offers variable selection.

4.1. Setup

The simulations are executed according to the model described in Section 3.1 with n ∈ {100,500} and n = 5 using inversion sampling, which is explained in detail in the supplementary material. The prespecified true parameter values are with variance components The entries of the covariate vectors xl and xs are drawn independently from the standard normal distribution 𝒩(0, 1). In addition to the informative covariates with effects βl and βs, the covariate vectors xl and xs are expanded with noninformative noise variables until the chosen numbers pl and ps of total dimensions are reached. These additional noise variables are included to evaluate variable selection properties and robustness of the approach in case of a misspecified model. The baseline hazard is chosen as λ0(t) = 2.5t1.5 and given the censoring mechanism described in the supplementary material, the chosen parameter values result in an average censoring rate of ≈50%. Overall, we consider two scenarios. One low-dimensional setup with n = 500 and pl = ps = 9 mimicking a more common data structure and one high-dimensional setup, where the number of covariates included in the survival submodel exceeds the number of individuals so that conventional approaches like JM fail to return results. For the computation, JM and lbbJM use K = 10 with equidistant knot placement. The grid ℳ≔{25, 50, ⋯, 500} × {1, 2, ⋯, 1000} is specified for possible tuples of stopping iterations and the optimal regularization parameter for glmnet is determined by the function cv.glmnet() based on 10-fold cross-validation.

4.2. Results

Since the compared estimation routines follow different approaches targeting various objectives from regular maximum likelihood estimation in joint models to regularization in pure time-to-event analysis, we focus on plain coefficient estimates averaged over 100 independent simulation runs in order to asses estimation characteristics. Variable selection properties are evaluated by considering share of true positives (TP) and false discovery rate (FDR). For low-dimensional setup (n = 500, ps = 9), Table 1 depicts the results for the low-dimensional setup. In the longitudinal submodel, lbbJMa has small shrinkage and therefore offers variable selection with a rather low false discovery rate of 0.23 but still selects all informative variables. The time effect β receives comparatively high shrinkage, since time, as a cluster-varying variable, adds more information to the model. Overall, the longitudinal submodel is boosted up to 108.25 iterations on average yielding rather strongly shrunk coefficient estimates. Please note that the results for lbbJMb are simply obtained by lme() and very similar to JM. In the survival part, both boosting approaches substantially outperform glmnet in terms of variable selection while again receiving also more shrinkage. Due to the comparatively rough baseline hazard depicted in Figure 1, all full likelihood approaches, i.e., JM and lbbJM, receive additional shrinkage which is unaffected by possible regularization. As glmnet uses the partial likelihood, there are no estimates for the baseline hazard function available and the small elastic net penalty of λ∗ = 0.003 on average also results in weaker performance regarding the rate of false positives. Overall, the lbbJM approaches yield satisfactory results regarding both regularization and variable selection. The effect estimates clearly reflect the true values approximately obtained by JM while simultaneously receiving sufficiently large shrinkage for decent performance of identifying influential covariates in both the longitudinal and survival submodels.
Table 1

Shrinkage and variable selection properties regarding longitudinal and survival outcomes averaged over 100 simulation runs of the low-dimensional scenario.

β t (sd) β l1 (sd) β l2 (sd) β l3 (sd)TPFDR m l

True2121
JM1.9980.9942.0081.002
(0.03)(0.07)(0.07)(0.07)
lbbJMa1.7600.9141.9220.9231.000.23108.25
(0.08)(0.07)(0.07)(0.07)
lbbJMb1.9920.9942.0081.002
(0.03)(0.07)(0.07)(0.07)

α (sd) β s1 (sd) β s2 (sd) β s3 (sd)TPFDR m s

True0.512-2
JM0.4570.9031.807-1.800
(0.04)(0.08)(0.12)(0.12)
lbbJMa0.3900.7281.521-1.5161.000.27209.2
(0.03)(0.06)(0.07)(0.07)
lbbJMb0.3730.7131.500-1.4951.000.22196.9
(0.03)(0.06)(0.07)(0.08)
glmnet0.4270.9091.833-1.8231.000.51
(0.03)(0.07)(0.11)(0.10)
Figure 1

Piecewise-constant baseline hazard estimates with K = 10 by JM, lbbJMa and lbbJMb averaged over 100 simulation runs of the low dimensional scenario.

For high-dimensional setup (n = 100, ps = 100), Table 2 depicts the results for the high-dimensional setup. As expected, estimates in the high-dimensional setup are regularized stronger and therefore experience more shrinkage. Again, the boosting approaches contained in lbbJM yield better variable selection properties and slightly more regularized coefficient estimates, although results seem to align with increasing dimensions. Note that JM is not capable of handling high-dimensional data structures and is therefore not included in the high-dimensional setup at all.
Table 2

Shrinkage and variable selection properties regarding the longitudinal and survival outcomes averaged over 100 simulation runs of the high-dimensional scenario.

β t (sd) β l1 (sd) β l2 (sd) β l3 (sd)TPFDR m l

True2121
lbbJMa1.7480.8681.8430.8751.000.36124.2
(0.20)(0.14)(0.14)(0.14)
lbbJMb1.9911.0081.9821.011
(0.08)(0.13)(0.16)(0.15)

α (sd) β s1 (sd) β s2 (sd) β s3 (sd)TPFDR m s

True0.512-2
lbbJMa0.3070.5121.242-1.2161.000.70136.7
(0.06)(0.12)(0.16)(0.13)
lbbJMb0.2850.4981.215-1.1911.000.67127.0
(0.05)(0.13)(0.15)(0.13)
glmnet0.2930.6271.449-1.4221.000.83
(0.08)(0.18)(0.30)(0.28)
For computational effort, Table 3 shows estimates for elapsed computation time of each routine. Times are measured in seconds and depict the computation time for one single model fit which was carried out on a 2 × 2.66 GHz-6-Core Intel Xeon CPU (64 GB RAM). As expected, the full boosting approach executed in lbbJMa comes with high computational costs similar to [11]. Note that the runtimes are higher in the low-dimensional scenario as the overall number of clusters is higher (n = 500) leading to an increased burden in the already time-consuming longitudinal boosting procedure. The reduced approach lbbJMb, however, runs considerably faster and is therefore more desirable as long as research focus lies solely on the time-to-event analysis.
Table 3

Averaged computation times for one single model fit (in seconds).

SetupJMglmnetlbbJMalbbJMb
Low110.00149.1515776.1643.76
High 156.444072.80248.08

5. Application

We showcase the lbbJM algorithm by applying it to the 1994 AIDS data [26]. The study is aimed at comparing the two antiretroviral drugs, didanosine (ddI) and zalcitabine (ddC), based on a collective of 467 patients infected with human immune deficiency virus (HIV) who were either intolerant to or failed a previous treatment with Zidovudine (AZT). Alongside several baseline covariates, the square root CD4 cell count was recorded at study entry and in multiple follow-ups after 2, 6, 12, and 18 months, respectively. The CD4 cells are attacked by the virus and thus decrease over time for infected patients; hence, they are a widely used surrogate for disease progression. In addition to the longitudinal outcome, 188 patients died during the time of the study leading to 188 observed and 279 censored events. The structure of the data is depicted in Table 4.
Table 4

Structure of the dataset with primary outcomes for the joint analysis in the three columns on the left.

y T δ t DrugGenderAZTprevOIID
10.6716.9700ddCMaleIntoleranceAIDS1
8.4316.9706ddCMaleIntoleranceAIDS1
9.4316.97012ddCMaleIntoleranceAIDS1
6.3219.0000ddIMaleIntolerancenoAIDS2
8.1219.0006ddIMaleIntolerancenoAIDS2
4.5819.00012ddIMaleIntolerancenoAIDS2
5.0019.00018ddIMaleIntolerancenoAIDS2
3.4618.5300ddIFemaleIntoleranceAIDS3
3.6118.5302ddIFemaleIntoleranceAIDS3
6.1618.5316ddIFemaleIntoleranceAIDS3
We formulate the joint model The CD4 cell count CD4(t) for i = 1, ⋯, 467 is described by a linear mixed model with random intercepts, random slopes for time t and linear effects of time, drug and an additional interaction between time and drug. The true, i.e. modelled by ηl(t), underlying profile of the CD4 cell count is then included together with the remaining baseline covariates in the Cox full likelihood model, thus the model error ε(t) is eliminated. Here, drug is a dummy for ddI, gender for female gender, AZT for failure of Zidovudine therapy, and prevOI for prevalence of AIDS. The number of segments for the baseline hazard was chosen to be K = 7. We fit model (22) with the same methods as already used in the simulation study. The tuning parameters of the boosting algorithm were chosen to be νl = νs = 0.1 for step lengths and ℳ≔{5, 10, ⋯, 100} × {1, 2, ⋯, 250} for the grid of possible tuples of stopping iterations for lbbJMa and ℳ≔{1, 2, ⋯, 250} for lbbJMb. Again, glmnet was tuned using the cv.glmnet() function and all regularization approaches are based on 10-fold cross validation. For lbbJMa, m∗,l = 10 and m∗,s = 33 formed the best performing tuple of stopping iterations. The two-stage approach lbbJMb used m∗,s = 40 and the optimal penalization parameter for glmnet turned out as λpen∗ = 0.0014. The corresponding coefficient estimates are shown in Table 5. Overall, the results reflect what was already observed in the simulation study. While glmnet shows quite conservative shrinkage properties where every variable is included in the final model, the lbbJM approaches tend to stop rather early yielding bigger shrinkage and in addition effects, which did not get selected at all. In order to give some point of reference regarding variable selection properties, the p values computed by JM are included in Table 5. The selected effects align quite nicely with being significant according to JM. While lbbJMa only includes the variable drug with only half the effect size, lbbJMb additionally sees a quite tiny impact of female gender.
Table 5

Shrinkage and variable selection properties by the different packages for model (22).

β 0 β l β t1 β t2 α β s1 β s2 β s3
JM6.970.49−0.18<0.01−0.240.310.090.66
lbbJMa6.950.26−0.050−0.13000.73
lbbJMb6.950.48−0.16−0.02−0.180.0300.61
glmnet −0.150.310.090.81

p value (JM)<0.010.26<0.010.98<0.010.230.61<0.01
The corresponding coefficient progressions for the survival submodel are visualized in Figure 2. Both algorithms update the coefficients referring to the longitudinal CD4 cell profile and the variable prevOI right away and therefore see a strong connection between the risk for death and the CD4 cell count as well as whether or not a patient has AIDS. Due to early stopping, lbbJMa selects neither of the two remaining covariates into the final model. lbbJMb includes one variable more, gender, which however only has a very small effect.
Figure 2

Coefficient progression in the survival part for lbbJMa ((a), with m∗,l = 10) and lbbJMb (b).

6. Outlook and Discussion

Overall, the lbbJM algorithm introduces a novel boosting-based regularization scheme to joint models focusing on survival analysis as well as to Cox models with time-dependent covariates. The method fits in well among alternative routines and especially stands out with respect to variable selection properties. Due to its clear advantage regarding computational effort as depicted in Table 3, lbbJMb is the preferred routine when research interest clearly lies on time-to-event analysis, whereas lbbJMa is capable of regularizing both submodels simultaneously. Besides the good results regarding variable selection, it is also expected that the proposed boosting methods can improve the predictive power of a joint model, since boosting algorithms are, due to their model tuning based on test errors, primarily used for prediction. A thorough investigation of improving and evaluating prediction performance of a joint model by several boosting techniques remains an interesting task. Still, the presented foundation is of a comparatively simple nature and possible extensions include more flexible modelling, e.g., based on P-splines which allow smooth effects of time-dependent as well as time-independent covariates and can additionally include time-varying effects as well as possibly time-varying association structures. As the current algorithm is confined to a time-constant association parameter α, an extension to α(t) would increase the flexibility of the model. While it is usually difficult to disentangle a time-dependent association parameter from the longitudinal trajectory for parameter estimation, the lbbJM algorithm could potentially avoid these identification issues due to its clear separation in a longitudinal and a survival boosting process. Similar regularization-based approaches have been proved useful for time-to-event settings [22, 33], and it can be assumed that the presented method could benefit from these concepts as well. Moreover, the presented work only lays a foundation to several extensions addressing known issues for both boosting and joint modelling. It represents an accessible way to boosting for time-dependent covariates in survival analysis. Gradient boosting has known limitations in this matter, and although efforts for a framework to overcome these issues are made [34], things rather quickly become technical and the proven robustness of likelihood-based boosting represents a flexible and far more intuitive alternative. Further developments in this direction could include multiple time-dependent covariates based on the two-stage approach of lbbJMb, where additional effort is necessary in order to provide a fair competition between time-dependent and time-independent covariates within the selection procedure. Furthermore, the component-wise updating process is capable of including allocation mechanisms into the algorithm. While the lbbJM algorithm is due to its two-stage fitting process fairly robust to estimate models, where one candidate variable is assigned to both submodels, this kind of specification is not advised in general as identification issues may arise and a proper interpretation of the resulting model might be challenging. However, it is usually tricky to decide, whether a covariate should be included in the longitudinal or the survival part of the model and these decisions are often made using prior knowledge. An allocation routine based on likelihood maximization could therefore greatly improve joint model inference and would additionally eliminate the two-stage nature of the lbbJM algorithm. This would not only decrease the computational burden but also provide a far more flexible algorithm allowing for regularization, variable selection, and allocation.
  13 in total

1.  Generalized additive modeling with implicit variable selection by likelihood-based boosting.

Authors:  Gerhard Tutz; Harald Binder
Journal:  Biometrics       Date:  2006-12       Impact factor: 2.571

2.  A joint model for survival and longitudinal data measured with error.

Authors:  M S Wulfsohn; A A Tsiatis
Journal:  Biometrics       Date:  1997-03       Impact factor: 2.571

3.  Flexible Bayesian additive joint models with an application to type 1 diabetes research.

Authors:  Meike Köhler; Nikolaus Umlauf; Andreas Beyerlein; Christiane Winkler; Anette-Gabriele Ziegler; Sonja Greven
Journal:  Biom J       Date:  2017-08-10       Impact factor: 2.207

4.  Joint analysis of longitudinal and survival AIDS data with a spatial fraction of long-term survivors: A Bayesian approach.

Authors:  Rui Martins; Giovani L Silva; Valeska Andreozzi
Journal:  Biom J       Date:  2017-05-02       Impact factor: 2.207

5.  Regularization for generalized additive mixed models by likelihood-based boosting.

Authors:  A Groll; G Tutz
Journal:  Methods Inf Med       Date:  2012-03-01       Impact factor: 2.176

6.  The evolution of boosting algorithms. From machine learning to statistical modelling.

Authors:  A Mayr; H Binder; O Gefeller; M Schmid
Journal:  Methods Inf Med       Date:  2014-08-12       Impact factor: 2.176

7.  Bayesian joint modeling of bivariate longitudinal and competing risks data: An application to study patient-ventilator asynchronies in critical care patients.

Authors:  Montserrat Rué; Eleni-Rosalina Andrinopoulou; Danilo Alvares; Carmen Armero; Anabel Forte; Lluis Blanch
Journal:  Biom J       Date:  2017-08-11       Impact factor: 2.207

8.  A comparative trial of didanosine or zalcitabine after treatment with zidovudine in patients with human immunodeficiency virus infection. The Terry Beirn Community Programs for Clinical Research on AIDS.

Authors:  D I Abrams; A I Goldman; C Launer; J A Korvick; J D Neaton; L R Crane; M Grodesky; S Wakefield; K Muth; S Kornegay
Journal:  N Engl J Med       Date:  1994-03-10       Impact factor: 91.245

9.  Joint modelling of longitudinal and time-to-event data with application to predicting abdominal aortic aneurysm growth and rupture.

Authors:  Michael J Sweeting; Simon G Thompson
Journal:  Biom J       Date:  2011-08-10       Impact factor: 2.207

10.  Addressing cluster-constant covariates in mixed effects models via likelihood-based boosting techniques.

Authors:  Colin Griesbach; Andreas Groll; Elisabeth Bergherr
Journal:  PLoS One       Date:  2021-07-09       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.