Literature DB >> 35799329

CAPITAL: Optimal subgroup identification via constrained policy tree search.

Hengrui Cai1, Wenbin Lu2, Rachel Marceau West3, Devan V Mehrotra3, Lingkang Huang4.   

Abstract

Personalized medicine, a paradigm of medicine tailored to a patient's characteristics, is an increasingly attractive field in health care. An important goal of personalized medicine is to identify a subgroup of patients, based on baseline covariates, that benefits more from the targeted treatment than other comparative treatments. Most of the current subgroup identification methods only focus on obtaining a subgroup with an enhanced treatment effect without paying attention to subgroup size. Yet, a clinically meaningful subgroup learning approach should identify the maximum number of patients who can benefit from the better treatment. In this article, we present an optimal subgroup selection rule (SSR) that maximizes the number of selected patients, and in the meantime, achieves the pre-specified clinically meaningful mean outcome, such as the average treatment effect. We derive two equivalent theoretical forms of the optimal SSR based on the contrast function that describes the treatment-covariates interaction in the outcome. We further propose a constrained policy tree search algorithm (CAPITAL) to find the optimal SSR within the interpretable decision tree class. The proposed method is flexible to handle multiple constraints that penalize the inclusion of patients with negative treatment effects, and to address time to event data using the restricted mean survival time as the clinically interesting mean outcome. Extensive simulations, comparison studies, and real data applications are conducted to demonstrate the validity and utility of our method.
© 2022 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd.

Entities:  

Keywords:  constrained policy tree search; optimal subgroup identification; personalized medicine

Mesh:

Year:  2022        PMID: 35799329      PMCID: PMC9544117          DOI: 10.1002/sim.9507

Source DB:  PubMed          Journal:  Stat Med        ISSN: 0277-6715            Impact factor:   2.497


INTRODUCTION

Personalized medicine, a paradigm of medicine tailored to a patient's characteristics, is an increasingly attractive field in health care. Its ultimate goal is to optimize an outcome of interest by assigning the right treatments to the right patients. The success of personalized medicine relies on using baseline covariates to identify a subgroup of patients that benefit more from the targeted treatment than other comparative treatments. The resulting identification strategy is referred to as a subgroup selection rule (SSR). If used properly, subgroup analysis can lead to more well‐informed clinical decisions and improved demonstration of the efficacy of a treatment. Various data‐driven methods for subgroup identification , , , , , , , have been developed during the past decade. Song and Pepe considered using a selection impact curve to evaluate treatment policies for a binary outcome based on a single baseline covariate. Foster et al and Cai et al developed notable methods for detection of subgroups with enhanced treatment effects based on multiple baseline covariates. Foster et al's method, virtual twins (VT), is a two‐stage method, first predicting the counterfactual outcome for each individual under both the test and control treatments, and then using tree‐based methods to infer the responding subgroups. Cai et al's method proposed instead to use a parametric scoring system to rank treatment effects, using this ranking to identify patients who benefit more from the new treatment. A useful tutorial and literature review of some commonly used subgroup identification methods is provided in Lipkovich et al. Recently, VanderWeele et al considered selecting the optimal subgroup under different constraints, including constrained resources, unconstrained resources, and in the presence of side effects and costs, aiming to maximize effect heterogeneity. This idea of constrained subgroup optimization is becoming increasingly of interest, with many newly proposed methods focusing on performing subgroup selection while simultaneously minimizing patient risk or cost. One such approach was proposed by Wang et al, who uses outcome weighted learning to generate an individualized optimal decision rule that maximizes the clinical benefit for patients while controlling the risk of adverse events. Guan et al proposed to estimate the optimal dynamic treatment regime under a constraint on the cost function by leveraging nonparametric Bayesian dynamics modeling with policy search algorithms. Zhou et al extended the constrained optimal treatment regime approach to competing risk data, using a penalized value search method to handle the trade‐off between the primary event of interest and the time to a severe treatment side effect. Most recently, Doubleday et al proposed two methods to identify risk‐controlled individualized treatment rules that maximize benefit while controlling risk at a pre‐specified threshold. While these methods make important contributions to finding optimal subgroups under constraints and while balancing risks, they all focus on optimizing the mean outcome of interest without considering the size of the subgroup. In such, they usually yield a smaller, and thus less satisfactory group of selected patients. Identifying the largest possible subgroup of patients that benefit from a given treatment at or above some clinically meaningful threshold can be critical both for the success of a new treatment and more importantly for the patients who may rely on a treatment for their health and survival. When too small of a subgroup is selected, the erroneously unselected patients may suffer from suboptimal treatments. For a test treatment, this reduced subgroup size can further lead to problems with regulatory approvals and may even halt compound development and availability. Postapproval accessibility can also be hindered by a lackluster subgroup size, especially in countries with all‐or‐nothing reimbursement markets where the seemingly small proportion of benefiting patients leads to reduced reimbursements that may not be financially sustainable for continued treatment manufacturing. Though these points are important, the crucial point remains that a subgroup learning approach that selects as many patients with evidence of clinically meaningful benefit from treatment as possible is desirable to ensure that more patients can receive the treatment that is best for them. Further, most of the existing optimization approaches with constraints use complex decision rules. It is hard to search within an interpretable class of decision rules using these methods since the loss functions in both outcome weighted learning and value search methods are defined based on the whole sample. Forming an interpretable decision rule helps regulators, doctors, and patients make sense of and ensure proper prescriptions. In this article, we develop a constrained policy tree search algorithm (CAPITAL) to optimize subgroup size while maintaining a prespecified clinical threshold within the selected subgroup(s). Our contributions can be summarized as follows. First, we derive two equivalent theoretical forms of the optimal SSR based on the contrast function that describes the treatment‐covariates interaction in the outcome. Second, we transform the loss function of the constrained optimization into individual rewards defined at the patient level. This enables us to identify the patients with a larger mean outcome and develop a decision tree to generate an interpretable subgroup using the policy tree algorithm proposed by Athey and Wager. Third, we extend our proposed method to the framework with multiple constraints, for example, penalizing the inclusion of patients with negative treatment results, and to time to event data, using the restricted mean survival time as the clinically interesting mean outcome. Extensive simulations, comparison studies, and real data applications are conducted to demonstrate the validity and utility of our method. The source code, implemented in the R language, is publicly available at our repository at https://github.com/HengruiCai/CAPITAL. The rest of this article is organized as follows. We first formulate our problem in Section 2. In Section 3, we establish the theoretical optimal SSR that achieves our objective, and then propose CAPITAL to solve the optimal SSR. We extend our work to multiple constraints and survival data in Section 4. Simulation and comparison studies are conducted to evaluate our methods in Section 5, followed by the real data analysis in Section 6. In Section 7, we discuss and conclude our paper. All the technical proofs and additional simulation results are provided in the Appendix.

PROBLEM FORMULATION

Let denote a ‐dimensional vector containing individual's baseline covariates with the support , and denote the binary treatment an individual receives. After a treatment is assigned, we observe the outcome of interest with support . Let and denote the potential outcomes that would be observed after an individual receives treatment 0 or 1, respectively. Define the propensity score function as the conditional probability of receiving treatment 1 given baseline covariates , denoted as . Denote as the sample size. The sample consists of observations independent and identically distributed (I.I.D.) across . As standard in the causal inference literature, we make the following assumptions: Stable unit treatment value assumption (SUTVA): , Ignorability: , Positivity: for all . Based on assumptions (A1) and (A2), we define the contrast function as that describes the treatment‐covariates interaction in the outcome. Under Assumptions (A1) to (A3), the contrast function is estimable from the observed data. Define the subgroup selection rule (SSR) as that assigns the patient with baseline covariates to the subgroup () or not (). Denote the class of the SSR as . The goal is to find an optimal SSR that maximizes the size of the subgroup and also maintains a desired mean outcome such as the average treatment effect (), with a theoretical version as follows, where is a prespecified threshold of clinically meaningful average treatment effect. Based on assumptions (A1) and (A2), the constraint in (1) can be represented by Combining this with (1), we can obtain the empirical version of the optimization objective: where is an indicator function, and is some estimator of the contrast function .

METHOD

In this section, we first establish the theoretical optimal SSR that achieves our objective, and then propose CAPITAL to solve the optimal SSR.

Theoretical optimal SSR

We first derive the theoretical optimal SSR that solves the theoretical objective in (1). Given the pre‐specified threshold , we denote a cut point associated with the contrast function such that the expectation of the contrast function larger than achieves , that is, By introducing , when we are maximizing the subgroup size, the treatment effect of each patient in the subgroup is ensured to meet the minimum acceptable beneficial effect size. We illustrate the density function of the contrast function with a cut point for the prespecified threshold in Figure A1. The yellow area in Figure A1 contains the patients whose contrast functions are larger than and thus satisfy (3). Intuitively, the theoretical optimal SSR should choose the patients whose contrast functions fall into the yellow area in Figure A1, that is, those whose treatment effects are larger than , to maximize the size of the subgroup. Without loss of generality, we consider the class of the theoretical SSRs as Here, for a given , the SSR selects a patient into the subgroup if his / her contrast function is larger than . The following theorem gives the theoretical optimal SSR.
FIGURE A1

Illustration of the density function of the contrast function with a cut point for the prespecified threshold

(Theoretical optimal SSR) Assuming (A1) and (A2), the optimal subgroup selection rule is Equivalently, the optimal subgroup selection rule is The proof of Theorem 1 consists of two parts. First, we show the optimal SSR is , where satisfies (3), within the class . Second, we derive the equivalence between (4) and (5). See the detailed proof of Theorem 1 provided in the Appendix. From Theorem 1 and the definition of the cut point , the optimal SSR can be found based on the density of the contrast function. Since the density function is usually unknown to us in reality, we use the estimated contrast function for each patient, that is, the individual treatment effect, to approximate the density function. A constrained policy tree search algorithm is provided to solve the optimal SSR in the next section.

Constrained policy tree search algorithm

In this section, we formally present CAPITAL. First, we transform the constrained optimization in (2) into individual rewards defined at the patient level. This enables us to identify patients more likely to benefit from treatment. Then, we develop a decision tree to partition these patients into the subgroups based on the policy tree algorithm proposed by Athey and Wager. In this article, we focus on the SSR in the class of finite‐depth decision trees. Specifically, for any , a depth‐ decision tree is specified via a splitting variable , a threshold , and two depth‐ decision trees , and , such that if , and otherwise. Denote the class of decision trees as . We illustrate a simple decision tree with splitting variables and in Figure A2. This decision tree has mathematical form . Define as the difference between the contrast function and the desired average treatment effect . Under (A1)‐(A3), we can estimate the contrast function as , using the random forest method and out‐of‐bag prediction (see eg, Lu et al ). Define . It is immediate that a patient with larger is more likely to be selected into the subgroup based on Figure A1. We sort the estimates as This sequence gives an approximation of the density of . We further define the cumulative mean based on the above sequence as With sufficiently large sample size, converges to the average treatment effect minus the desired effect , within the selected patients whose contrast function is larger than the upper quantile of the density of , that is, where is the upper quantile of the density of when goes to infinity. As long as is larger than zero, the selected subgroup satisfies the condition in (2) based on the theoretical optimal SSR in (5) from Theorem 1. Therefore, we need to select patients with positive and maximize the subgroup size to solve (2). To do this, we define the reward of the th individual based on the sign of as follows:
FIGURE A2

Illustration of a simple decision tree with splitting variables and

Reward 1: where is the rank of in the sequence or the sequence , and “” is the sign operator such that if , if , and if . Given is positive, the reward is 1 if the patient is selected to be part of the subgroup, and is 0 otherwise. Likewise, supposing is negative, the reward is if the patient is selected to be in the subgroup, that is, , and is 0 otherwise. This is in accordance with the intuition that we should select patients with larger than zero. To encourage the decision tree to include patients who have a lager treatment effect, we also propose the following reward choice based on the value of directly: Reward 2: The optimal SSR is searched within the decision tree class to maximize the sum of the individual rewards defined in (6) or (7). Specifically, the decision tree allocates each patient to the subgroup or not, and receives the corresponding rewards. We use the exhaustive search to estimate the optimal SSR that optimizes the total reward, using the policy tree algorithm proposed in Athey and Wager. It is shown in the simulation studies (Section 5) that the performances are very similar under these two reward choices. We denote the estimated optimal SSR that maximizes the size of the subgroup and also maintains the desired average treatment effect as . The proposed algorithm not only results in an interpretable SSR (see more discussion in Section 5), but also is flexible to handle multiple constraints and survival data, as discussed in detail in the next section.

EXTENSIONS

In this section, we discuss two main extensions of CAPITAL for solving (2). We first address multiple constraints on the average treatment effect in Section 4.1, and then handle the time to event data with the restricted mean survival time as the clinically interesting mean outcome in Section 4.2.

Extension to multiple constraints

In addition to the main constraint described in (2), in reality there may exist secondary constraints of interest. For instance, besides a desired average treatment effect, the individual treatment effect for each patient should be greater than some minimum beneficial value. Under such multiple constraints, the optimal SSR is defined by where is a prespecified minimum beneficial value. In the rest of this article, we focus on the case with , that is, the individual treatment effect for each patient should be nonnegative so that the treatment is beneficial to the patients in the selected group. Following similar arguments in (2), we can derive the empirical version of the optimization objective in (8) as The above objective function can be solved by modifying CAPITAL presented in Section 3.2. Specifically, we define the reward of the th individual based on (9) and (7) as follows. Reward 3: where is the nonnegative penalty parameter that represents the trade‐off between the first and the second constraint. When , the reward defined in (10) reduces to (7). Here, we only add the penalty on the reward when the estimated contrast function is negative, that is, . This prevents the method from selecting patients with a negative individual treatment effect.

Extension to survival data

We next consider finding the optimal SSR for a survival endpoint. Let and denote the survival time of interest and the censoring time, respectively. Assume that and are independent given baseline covariates and the treatment. Then, the observed dataset consists of independent and identically distributed triplets, , where and . The goal is to maximize the size of the subgroup with a pre‐specified clinically desired effect , that is, where is the maximum follow up time, which is pre‐specified or can be estimated based on the observed data. Denote and as the restricted mean survival time for groups with treatment 0 and 1, respectively, given baseline covariate , where and are survival functions in the control and treatment groups, respectively. To estimate and , we first fit a random forest on the survival functions in the control and treatment groups, respectively, and obtain estimates as and . Then, the estimated restricted mean survival time for groups with treatment 0 and 1, denoted as and , are calculated by integrating the estimated survival functions to the minimum of the maximum times over the 2 arms. Similarly, the empirical version of optimization in (11) is Define to capture the distance from the estimated contrast function to the desired difference in restricted mean survival time for the th individual. It is immediate that an individual with larger is more likely to be selected into the subgroup. We sort the estimates as and define the cumulative mean as . The reward for the constrained policy tree search can be defined following similar arguments as in (6) and (7).

SIMULATION STUDIES

Evaluation and comparison with average treatment effect

Data generation

Suppose baseline covariates , the treatment information , and the outcome are generated from the following model: where is the baseline function of the outcome, is the contrast function, is the random error. We set the dimension of covariates as and consider the following three scenarios respectively. Scenario 1: Scenario 2: Scenario 3: The true average treatment effect can be calculated as 0 under all scenarios. We illustrate the density of for Scenarios 2 and 3 in Figure S1 in the Supplementary Material. Note the density of for Scenario 1 is just a uniform distribution on interval . Based on Figure S1, we consider the clinically meaningful treatment effect for all scenarios, with the corresponding optimal subgroup sample proportions as listed in Table A2. Let the total sample size be chosen from the set .
TABLE A2

Empirical results of subgroup analysis under the estimated optimal SSR by CAPITAL with reward in (6) and the VT‐C method

Scenario 1Scenario 2Scenario 3
Method r=10 n=200 n=500 n=1000 n=200 n=500 n=1000 n=200 n=500 n=1000
CAPITAL δ=0.7 Proportion 65% 67% 75%
pr{D^(X)} 0.62 (0.16)0.63 (0.08)0.65 (0.05)0.42 (0.23)0.51 (0.11)0.56 (0.05)0.72 (0.15)0.74 (0.08)0.77 (0.05)
ATE(D^) 0.66 (0.28)0.72 (0.17)0.69 (0.10)0.72 (0.47)0.96 (0.20)0.86 (0.11)0.66 (0.34)0.67 (0.18)0.61 (0.11)
RCD0.83 (0.10)0.91 (0.05)0.93 (0.03)0.62 (0.15)0.81 (0.08)0.87 (0.03)0.83 (0.08)0.89 (0.03)0.90 (0.01)
RPI0.78 (0.13)0.80 (0.09)0.78 (0.06)0.74 (0.15)0.88 (0.08)0.86 (0.06)0.67 (0.10)0.67 (0.06)0.65 (0.04)
δ=1.0 Proportion 50% 50% 63%
pr{D^(X)} 0.46 (0.16)0.48 (0.09)0.50 (0.06)0.21 (0.17)0.32 (0.12)0.40 (0.06)0.56 (0.16)0.59 (0.09)0.62 (0.06)
ATE(D^) 0.90 (0.27)1.00 (0.15)0.99 (0.11)0.83 (0.63)1.31 (0.27)1.17 (0.11)1.02 (0.37)1.00 (0.20)0.94 (0.15)
RCD0.84 (0.11)0.91 (0.05)0.94 (0.03)0.62 (0.12)0.79 (0.11)0.88 (0.05)0.79 (0.07)0.85 (0.03)0.87 (0.01)
RPI0.88 (0.11)0.94 (0.06)0.94 (0.05)0.75 (0.19)0.95 (0.05)0.97 (0.03)0.78 (0.11)0.78 (0.06)0.77 (0.05)
δ=1.3 Proportion 35% 37% 51%
pr{D^(X)} 0.30 (0.16)0.31 (0.11)0.34 (0.08)0.09 (0.09)0.14 (0.10)0.25 (0.09)0.41 (0.15)0.44 (0.09)0.48 (0.06)
ATE(D^) 1.05 (0.33)1.28 (0.18)1.29 (0.14)0.66 (0.73)1.58 (0.59)1.48 (0.24)1.36 (0.40)1.38 (0.24)1.29 (0.15)
RCD0.81 (0.10)0.88 (0.07)0.92 (0.04)0.67 (0.07)0.74 (0.08)0.82 (0.06)0.78 (0.08)0.83 (0.03)0.86 (0.02)
RPI0.93 (0.12)0.99 (0.02)1.00 (0.01)0.69 (0.21)0.91 (0.13)0.95 (0.04)0.86 (0.10)0.89 (0.06)0.88 (0.04)
VT‐C δ=0.7 Proportion 65% 67% 75%
pr{D^(X)} 0.31 (0.12)0.34 (0.09)0.35 (0.08)0.15 (0.10)0.19 (0.09)0.22 (0.08)0.29 (0.10)0.30 (0.06)0.30 (0.06)
ATE(D^) 1.11 (0.20)1.27 (0.17)1.30 (0.15)0.85 (0.61)1.46 (0.38)1.53 (0.32)1.76 (0.36)1.82 (0.23)1.81 (0.21)
RCD0.66 (0.12)0.69 (0.09)0.70 (0.08)0.43 (0.08)0.51 (0.09)0.55 (0.09)0.54 (0.10)0.55 (0.06)0.55 (0.06)
RPI0.97 (0.06)0.99 (0.03)1.00 (0.01)0.77 (0.17)0.95 (0.09)0.97 (0.08)0.95 (0.07)0.98 (0.03)0.98 (0.03)
δ=1.0 Proportion 50% 50% 63%
pr{D^(X)} 0.21 (0.13)0.24 (0.10)0.26 (0.07)0.07 (0.06)0.09 (0.07)0.14 (0.07)0.23 (0.09)0.24 (0.06)0.25 (0.05)
ATE(D^) 1.19 (0.21)1.37 (0.18)1.45 (0.13)1.01 (0.74)1.67 (0.49)1.78 (0.38)1.94 (0.34)2.02 (0.23)2.00 (0.18)
RCD0.70 (0.12)0.74 (0.10)0.76 (0.07)0.54 (0.06)0.59 (0.07)0.64 (0.07)0.60 (0.08)0.62 (0.06)0.62 (0.05)
RPI0.98 (0.05)1.00 (0.02)1.00 (0.00)0.81 (0.20)0.96 (0.09)0.98 (0.08)0.97 (0.05)0.99 (0.02)0.99 (0.01)
δ=1.3 Proportion 35% 37% 51%
pr{D^(X)} 0.12 (0.11)0.11 (0.11)0.16 (0.11)0.03 (0.04)0.03 (0.04)0.07 (0.05)0.17 (0.09)0.18 (0.06)0.20 (0.05)
ATE(D^) 1.25 (0.23)1.43 (0.18)1.50 (0.12)1.11 (0.81)1.81 (0.61)1.98 (0.42)2.12 (0.37)2.24 (0.23)2.19 (0.20)
RCD0.74 (0.09)0.76 (0.11)0.81 (0.11)0.65 (0.03)0.66 (0.04)0.69 (0.05)0.65 (0.09)0.67 (0.06)0.69 (0.05)
RPI0.99 (0.04)1.00 (0.01)1.00 (0.00)0.83 (0.21)0.95 (0.13)0.98 (0.07)0.99 (0.03)1.00 (0.01)1.00 (0.00)

Results under CAPITAL

We apply CAPITAL to find the optimal SSR. The policy is searched within based on the R package “policytree”. , For better demonstration, we focus on decision trees. To illustrate the interpretability of the resulting SSR, we highlight the results of three specific simulation replicates (denoted as replicate No.1, No.2, and No.3) under Scenario 2 with using the first reward (7) for . The estimated SSR under these three selected replicates are shown in Figure A3, with the splitting variables and their splitting thresholds reported in Table A1. We summarize the selected sample proportion under the estimated SSR, the average treatment effect of the estimated SSR, and the rate of making correct subgroup decisions by the estimated SSR, using Monte Carlo approximations. Finally, we visualize the density function of within the subgroup selected by the estimated SSR and compare to that of unselected patients for three replicates in Figure A4.
FIGURE A3

The estimated optimal subgroup selection tree by CAPITAL under Scenario 2 with and . Upper left panel: for Replicate No.1. Upper right panel: for Replicate No.2. Lower middle Panel: for Replicate No.3

TABLE A1

Results of estimated optimal subgroup selection tree for three particular replicates under Scenario 2 with and (where the optimal subgroup sample proportion is ) under CAPITAL

SimulationReplicate No.1Replicate No.2Replicate No.3
pr{D^(X)} 44.5%49.2%55.0%
ATE(D^) 1.111.000.90
Rate of correct decision91.85%92.01%94.45%
DT2 Split variable (split value) X(1)(0.12) X(2)(0.26) X(2)(0.03)
DT1(Left) Split variable (split value) X(2)(0.18) X(1)(0.13) X(1)(0.29)
DT1(Right) Split variable (split value) X(2)(0.28) X(1)(0.02) X(1)(0.12)
FIGURE A4

The density function of within or outside the subgroup under Scenario 2 with and . Left panel: for Replicate No.1. Middle panel: for Replicate No.2. Right Panel: for Replicate No.3

Over 200 replicates under Scenario 2 with , the rate of correctly identifying important features and under the estimated SSRs is 70.8% with , increasing to 95.8% with , and 100.0% with . It can be seen from both Figure A3 and Table A1 that the estimated SSRs under the proposed method identify the important features that determine the outcome, and , for all three replicates. In Scenario 2, and have identical roles in the contrast function, so the resulting optimal tree can either use or as the first splitting variable. Replicate No.3 over‐selects the subgroup and therefore yields a lower average treatment effect, while replicate No.1 under‐selects the subgroup and achieves a higher average treatment effect, as shown in Table A1. This finding is in line with the trade‐off between the size of the selected subgroup and its corresponding average treatment effect discussed in the introduction. Moreover, all these three replicates have a high rate () of making correct subgroup decisions under the estimated SSRs, supported by both Table A1 and Figure A4.

Comparison studies

We compare the proposed method with two popular methods with variations: two variants of the VT method and two variants of the policy tree search method using the adjusted value estimator. While the VT method can theoretically be used for both binary and continuous outcomes, the current R package “aVirtualTwins” only deals with binary outcomes in a two‐armed clinical trial. To address the continuous outcomes in Scenarios 1‐3, following the VT method we fit the estimated individual treatment effect on features via a regression tree. We consider two subgroup selection rules based on the VT method. VT‐A: Denote the average treatment effect within a terminal node as where is the size of the terminal node. The final subgroup is formed as the union of the terminal nodes where the predicted values are greater than . VT‐C: Denote . Then each terminal node is classified into the subgroup based on a majority vote within the node by . The final subgroup is defined as the union of the terminal nodes with . The second method considered for comparison finds subgroups by maximizing the ‐quality adjusted value estimator , via a policy tree search. Specifically, we treat or as the individual value and select the th individual into the treatment benefitting subgroup if its value is positive. The optimal decision tree can be found by maximizing ‐quality adjusted value estimators, that is, or for the or individual rewards, respectively. , We apply the proposed method, the VT‐A and VT‐C methods, and the policy tree search methods based on different ‐quality adjusted value estimators, with 200 replications on Scenarios 1‐3. Under the estimated SSR, we summarize the following estimates and corresponding standard deviations, aggregated using Monte Carlo approximations over 200 replications in Table A2: the selected sample proportion as , the average treatment effect as , the rate of making correct subgroup decisions (RCD, the number of correct subgroup decisions divided by the total sample size), and the rate of positive individual treatment effect within the selected subgroup (RPI, the number of positive individual treatment effects divided by the size of the selected subgroup). Since the performance of our method under reward (6) and reward (7) are similar, and the VT‐A and VT‐C methods have nearly identical results, we do not report the proposed method with reward (7) and the VT‐A method in the empirical results Table A2. These results can be found in Table S1 in the Supplemental Material. The results of policy tree search by different ‐quality adjusted value estimators are included in Table S2 in the Supplemental Material using Scenario 1 as an illustration. Based on Tables A2, S1, and S2, it is clear that the proposed method has a better performance than the VT methods and the ‐quality adjusted policy tree search methods in all cases. In Scenario 1 with , our method achieves a selected sample proportion of 65% for (the optimal is 65%), 50% for (the optimal is 50%), and 34% for (the optimal is 35%), with corresponding average treatment effects close to the true values. In Scenario 2, the selected sample proportion of the proposed method is moderately underestimated due to the fact that the density function of is concentrated around 0, as illustrated in the left panel of Figure S1 in the Supplementary Material. The proposed method performs well under small sample sizes, but with slightly lower selected sample proportion, and gets better performance in selected subgroups as the sample size increases. In contrast, in most cases, the VT methods identify subgroups that are barely half of the desired optimal subgroup size. According to Table S2, simply using or as the individual reward in policy tree with ‐quality adjusted estimator method , cannot achieve the desired optimal subgroup proportion and underestimate/overestimate the average treatment effect. The results with as the individual value are very close to the results under the VT‐C method, where we use as the individual reward with the majority vote for subgroup tree search. Though the policy tree can find a relatively larger subgroup when using as the individual value, the corresponding average treatment effects are much smaller than both our proposed method and the VT methods in all cases.

Evaluation of multiple constraints

In this section, we further investigate the performance of the proposed method under multiple constraints. Specifically, we aim to solve the objective in (9) with the penalized reward defined in (10). We define four cases based on penalty terms , where corresponds to (7), that is, a single constraint. We use the same setting as described in Section 5.1 with under Scenarios 1 to 3 and apply CAPITAL to find the optimal SSR within . The empirical results are reported in Table A3 under the different penalty terms over 200 replications. It can be observed from Table A3 that for all cases, as the penalty term increases, the rate of positive individual treatment effect within the selected subgroup increases while the rate of making correct subgroup decisions slightly decreases. This reflects the trade‐off between two constraints in our theoretical objective in (8).
TABLE A3

Empirical results of optimal subgroup selection tree by CAPITAL with the penalized reward in (10)

Scenario 1Scenario 2Scenario 3
r=10 n=200 n=500 n=1000 n=200 n=500 n=1000 n=200 n=500 n=1000
δ=0.7 Proportion 65% 67% 75%
λ=0 pr{D^(X)} 0.63 (0.16)0.63 (0.08)0.65 (0.05)0.44 (0.24)0.51 (0.11)0.57 (0.06)0.72 (0.15)0.75 (0.07)0.77 (0.04)
ATE(D^) 0.67 (0.30)0.72 (0.17)0.70 (0.11)0.71 (0.48)0.95 (0.20)0.85 (0.11)0.67 (0.35)0.66 (0.17)0.60 (0.10)
RCD0.84 (0.10)0.91 (0.05)0.93 (0.03)0.62 (0.15)0.81 (0.08)0.87 (0.03)0.83 (0.08)0.89 (0.03)0.91 (0.01)
RPI0.78 (0.13)0.80 (0.09)0.78 (0.06)0.74 (0.16)0.88 (0.09)0.85 (0.07)0.67 (0.10)0.67 (0.06)0.65 (0.04)
λ=0.5 pr{D^(X)} 0.55 (0.12)0.56 (0.06)0.57 (0.04)0.39 (0.21)0.48 (0.10)0.53 (0.05)0.63 (0.13)0.65 (0.07)0.66 (0.05)
ATE(D^) 0.83 (0.23)0.86 (0.11)0.86 (0.08)0.77 (0.48)1.01 (0.17)0.93 (0.10)0.89 (0.30)0.88 (0.16)0.86 (0.11)
RCD0.84 (0.09)0.90 (0.05)0.91 (0.03)0.61 (0.15)0.79 (0.08)0.85 (0.04)0.81 (0.09)0.86 (0.04)0.87 (0.03)
RPI0.86 (0.11)0.88 (0.07)0.88 (0.05)0.76 (0.15)0.91 (0.07)0.90 (0.05)0.74 (0.09)0.74 (0.06)0.74 (0.04)
λ=1 pr{D^(X)} 0.52 (0.11)0.54 (0.05)0.54 (0.04)0.37 (0.20)0.46 (0.09)0.51 (0.05)0.57 (0.13)0.60 (0.07)0.61 (0.05)
ATE(D^) 0.88 (0.20)0.91 (0.11)0.91 (0.07)0.79 (0.48)1.05 (0.16)0.97 (0.10)1.00 (0.29)0.99 (0.16)0.98 (0.12)
RCD0.83 (0.09)0.88 (0.05)0.89 (0.04)0.60 (0.15)0.78 (0.08)0.83 (0.05)0.78 (0.10)0.83 (0.05)0.84 (0.04)
RPI0.88 (0.09)0.90 (0.06)0.91 (0.05)0.77 (0.15)0.92 (0.06)0.92 (0.05)0.78 (0.09)0.78 (0.05)0.78 (0.04)
λ=2 pr{D^(X)} 0.49 (0.11)0.52 (0.05)0.52 (0.04)0.33 (0.19)0.43 (0.09)0.48 (0.05)0.52 (0.12)0.55 (0.07)0.55 (0.05)
ATE(D^) 0.93 (0.19)0.95 (0.11)0.96 (0.07)0.83 (0.51)1.10 (0.15)1.03 (0.10)1.12 (0.30)1.11 (0.16)1.11 (0.12)
RCD0.81 (0.10)0.86 (0.05)0.87 (0.04)0.58 (0.15)0.76 (0.08)0.81 (0.05)0.74 (0.10)0.78 (0.06)0.79 (0.04)
RPI0.91 (0.09)0.92 (0.06)0.94 (0.05)0.78 (0.16)0.94 (0.05)0.94 (0.04)0.81 (0.09)0.82 (0.05)0.83 (0.04)

Evaluation of survival data

The data is generated by a similar model as (13): We set the dimension of covariates as , and define the survival time as . Consider the following scenario: Scenario 4: Here, for the random noise component we consider three cases: (i) Case 1 (normal): ; (ii) Case 2 (logistic): ; (iii) Case 3 (extreme): . The censoring times are generated from a uniform distribution on , where is chosen to yield the desired censoring level 15% and 25%, respectively, each applied for the three choices of noise distributions for a total of 6 settings considered. We illustrate in Figure S2 in the Supplemental Material. The clinically meaningful difference in restricted mean survival time is summarized in Table A4. Each setting was selected to yield a selected sample proportion of 50%. We report the empirical results in Table A4 with the second choice of reward (7), including the selected sample proportion under the estimated SSR, the average treatment effect of the estimated SSR, and the rate of making correct subgroup decisions by the estimated SSR, over 200 replications using Monte Carlo approximations with standard deviations presented in parentheses.
TABLE A4

Empirical results of optimal subgroup selection tree by CAPITAL for the survival data under Scenario 4 (where the optimal subgroup sample proportion is )

Censoring level 15% Censoring level 25%
n=500 n=1000 n=500 n=1000
Case 1 (normal)True δ 1.070.86
pr{D^(X)} 0.45 (0.17)0.47 (0.12)0.46 (0.16)0.48 (0.11)
ATE(D^) 1.07 (0.31)1.11 (0.24)0.87 (0.22)0.87 (0.16)
RCD0.84 (0.11)0.88 (0.07)0.84 (0.09)0.90 (0.06)
Case 2 (logistic)True δ 1.340.87
pr{D^(X)} 0.57 (0.26)0.56 (0.18)0.52 (0.24)0.52 (0.18)
ATE(D^) 0.94 (0.49)1.06 (0.36)0.63 (0.31)0.75 (0.24)
RCD0.72 (0.13)0.80 (0.10)0.74 (0.13)0.82 (0.09)
Case 3 (extreme)True δ 0.730.54
pr{D^(X)} 0.44 (0.18)0.46 (0.12)0.41 (0.18)0.44 (0.12)
ATE(D^) 0.76 (0.21)0.78 (0.15)0.57 (0.15)0.58 (0.11)
RCD0.84 (0.11)0.89 (0.08)0.83 (0.12)0.88 (0.08)
Table A4 shows that the proposed method performs reasonably well under all three considered noise distributions. Both the selected sample proportion and average treatment effect under the estimated SSR get closer to the truth, and the rate of making correct subgroup decisions increases as the sample size increases. The selected sample proportion is slightly underestimated for Cases 1 and 3 where has a more concentrated density function, while it is marginally overestimated for Case 2 where the density function of is more spread out. All these findings are in accordance with our conclusions in Section 5.1.

REAL DATA ANALYSIS

In this section, we illustrate our proposed method by application to the AIDS Clinical Trials Group Protocol 175 (ACTG 175) data as described in Hammer et al and a Phase III clinical trial in patients with hematological malignancies from Lipkovich et al.

Case 1: ACTG 175 Data

There were 1046 HIV‐infected subjects enrolled in ACTG 175, randomized to two competing antiretroviral regimens: zidovudine (ZDV) + zalcitabine (zal) (denoted as treatment 0), and ZDV + didanosine (ddI) (denoted as treatment 1). Patients were randomized in equal proportions, with 524 patients randomized to treatment 0 and 522 patients to treatment 1, with constant propensity score . We consider baseline covariates: (1) four continuous variables: age (years), weight (kg), CD4 count (cells/mm) at baseline, and CD8 count (cells/mm) at baseline; and (2) eight categorical variables: hemophilia (0 = no, 1 = yes), homosexual activity (0 = no, 1 = yes), history of intravenous drug use (0 = no, 1 = yes), Karnofsky score (four levels on the scale of 0‐100, as 70, 80, 90, and 100), race (0 = white, 1 = nonwhite), gender (0 = female), antiretroviral history (0 = naive, 1 = experienced), and symptomatic status (0 = asymptomatic). The outcome of interest () is the CD4 count (cells/mm) at 20 5 weeks. A higher CD4 count usually indicates a stronger immune system. We normalize by its mean and standard deviation. Our goal is to find the optimal subgroup selection rule that optimizes the size of the selected subgroup and achieves the desired average treatment effect. In real analysis, the thresholds of average treatment effects, , should be specified to exceed the whole group estimated average treatment effects by some extent while staying below the maximum individual treatment effects, to allow for identification of a reasonably sized benefiting subgroup. In practice, one can specify any clinical meaningful average treatment effect as the threshold and apply our method. In this example we use the estimated contrast density, along with the estimated mean contrast difference of 0.228, to select thresholds and , corresponding to optimal subgroup sample proportions of approximately 40% and 70%, respectively. The density of the estimated contrast function for the ACTG 175 data is provided in Figure S3 in the Supplementary Material. We apply the proposed CAPITAL method and the virtual twin method (VT‐C) (because the results under the VT‐C and VT‐A methods as well as the ‐quality adjusted policy tree search , have nearly identical performances, as shown in the simulation studies), using the same procedure as described in Section 5.1. The estimated SSRs under the proposed method are shown in Figure A5. To evaluate the proposed method and VT‐C method in the ACTG 175 data, we randomly split the whole data, with 70% of the data as a training sample to find the SSR and 30% as a testing sample to evaluate its performance. Here, we consider CAPITAL without penalty, with small penalty, and with large penalty on negativity of average treatment effect, respectively. The penalty term is chosen from , where encourages a positive average treatment effect in the selected group. Here, the magnitude of the (non‐zero) penalty terms is selected based on the size of the average treatment effects. In Table A5, we summarize the selected sample proportion , the average treatment effect under the estimated SSR , the average treatment effect outside the subgroup , the difference of the average treatment effect within the subgroup and outside the subgroup , and the rate of positive individual treatment effect within the selected subgroup (RPI), aggregated over 200 replications with standard deviations presented in parentheses, under different for two methods.
FIGURE A5

The estimated optimal subgroup selection tree using CAPITAL under the ACTG 175 data. Left panel: for . Right panel: for

TABLE A5

Evaluation results of the subgroup optimization using CAPITAL and the subgroup identification (using virtual twins ) under the ACTG 175 data

Threshold δ=0.35 δ=0.40
CAPITAL pr{D^(X)} 92.8% (0.023)82.8% (0.029)
without penalty ATE(D^) 0.250 (0.015)0.270 (0.016)
ATE(D^c) 0.107 (0.069)0.004 (0.038)
ATE(D^)ATE(D^c) 0.357 (0.068)0.266 (0.038)
RPI83.0% (0.021)85.1% (0.022)
CAPITALPenalty λ= 420
with small penalty pr{D^(X)} 52.7% (0.052)34.2% (0.034)
ATE(D^) 0.327 (0.022)0.385 (0.021)
ATE(D^c) 0.113 (0.021)0.142 (0.017)
ATE(D^)ATE(D^c) 0.214 (0.027)0.243 (0.026)
RPI91.5% (0.029)96.2% (0.017)
CAPITALPenalty λ= 20100
with large penalty pr{D^(X)} 35.6% (0.035)19.5% (0.051)
ATE(D^) 0.381 (0.021)0.414 (0.032)
ATE(D^c) 0.139 (0.017)0.180 (0.017)
ATE(D^)ATE(D^c) 0.242 (0.025)0.234 (0.033)
RPI95.9% (0.017)96.9% (0.025)
Virtual twins pr{D^(X)} 22.1% (0.063)10.5% (0.029)
ATE(D^) 0.462 (0.043)0.556 (0.050)
ATE(D^c) 0.159 (0.021)0.187 (0.014)
ATE(D^)ATE(D^c) 0.302 (0.037)0.368 (0.047)
RPI97.8% (0.019)99.6% (0.010)
As illustrated in Figure A5, the estimated SSRs based on the proposed method under both and rely on the weight and age of patients. For instance, for a desired average treatment effect of 0.35, younger patients ( years old) who weigh less than 91.2 kg or those years old weighting 91.2 kg may not benefit from treatment 1 (ZDV + ddI) and thus are not selected in the subgroup, while those older should be included into the subgroup of patients with enhanced effects from treating with ZDV + ddI. From Table A5, it is clear that the selected sample proportion under our method is much larger than that under the VT method in all cases. Specifically, our method yields a selected sample proportion at for , and at for , with a single constraint. Under a penalty on negativity of average treatment effect, the size of the identified subgroup using the proposed method is reduced to with small penalty and to with large penalty under , and further decreases to with small penalty and to with large penalty under . With a large penalty, our proposed method can achieve the desired average treatment effect at 0.381 (vs ) and at 0.414 (vs ). In contrast, the VT method identifies less than a quarter of the patients (22.1%) in the case of , and nearly a tenth of patients for , with overestimated average treatment effects of 0.462 and 0.556, respectively. These imply that the proposed method could largely increase the number of benefitting patients to be selected in the subgroup while also maintaining the desired clinically meaningful threshold.

Case 2: Phase III trial for hematological malignancies

Next, we consider a Phase III randomized clinical trial in 599 patients with hematological malignancies. We exclude seven subjects with missing records and use the remaining 592 complete records, for a final analysis dataset consisting of 301 patients receiving the experimental therapy plus best supporting care (as treatment 1) and 291 patients only receiving the best supporting care (as treatment 0). We use the same baseline covariates selected by Lipkovich et al. (2017): (1) 12 categorical variables: gender (1 = Male, 2 = Female), race (1 = Asian, 2 = Black, 3 = White), Cytogenetic markers 1 through 9 (0 = Absent, 1 = Present), and outcome for patient's prior therapy (1 = Failure, 2 = Progression, 3 = Relapse); and (2) two ordinal variables: Cytogenetic category (1 = Very good, 2 = Good, 3 = Intermediate, 4 = Poor, 5 = Very poor), and prognostic score for myelodysplastic syndromes risk assessment (IPSS) (1 = Low, 2 = Intermediate, 3 = High, 4 = Very high). These baseline covariates contain demographic and clinical information that is related to baseline disease severity and cytogenetic markers. The primary endpoint in the trial was overall survival time. Our goal is to find the optimal subgroup selection rule that maximizes the size of the selected group while achieving the desired clinically meaningful difference in restricted mean survival time in the survival data. Based on the estimated contrast function and the estimated difference of mean survival time of 44 days, we consider and days, such that the corresponding optimal subgroup sample proportions are approximately around 40% and 70%. The density of the estimated contrast function for the hematological malignancies data is provided in the Supplementary Material. We apply the proposed method and the virtual twin method using the procedure described in Sections 5.3 and 6.1. The estimated SSRs under the proposed method are shown in Figure A6. The evaluation results for the hematological malignancies data are summarized in Table A6 for varying under the proposed method with and the virtual twin method. For Case 1 with ACTG 175 data in Section 6.1, we normalize the outcome by its mean and standard deviation, so the corresponding penalty term is relatively larger compared to Case 2 with the hematological malignancies data in this section, which directly uses the survival time as the outcome. Our estimated SSRs are shown in Figure A6, both using the IPSS score and the outcome for the patient's prior therapy as the splitting features in the decision tree. With a desired average treatment effect of , patients who had a relapse during prior therapy and IPSS larger than 3, as well as those who had no relapse with IPSS larger than 2, are selected into the subgroup with enhanced treatment effect of the experimental treatment plus best supporting care. From Table A6, we can also observe that our proposed method has a much better performance compared to the virtual twins method. To be specific, the selected sample proportion under the proposed method is much larger than that under the virtual twins method for all cases, with estimated treatment effect sizes closer to and over the desired clinically meaningful difference in restricted mean survival time as the penalty term increases. All these findings conform with the results in Section 6.1.
FIGURE A6

The estimated optimal subgroup selection tree using CAPITAL under the hematological malignancies data. Left panel: for . Right panel: for

TABLE A6

Evaluation results of the subgroup optimization using CAPITAL and the subgroup identification (using virtual twins ) under the hematological malignancies data

δ=84 δ=108
CAPITAL pr{D^(X)} 79.3% (0.031)43.2% (0.057)
without penalty ATE(D^) 69.5 (5.0)101.2 (9.8)
ATE(D^c) 53.7 (17.7)1.2 (7.2)
ATE(D^)ATE(D^c) 123.2 (16.9)100.0 (9.3)
RPI87.0% (0.028)94.5% (0.034)
CAPITALPenalty λ= 22
with small penalty pr{D^(X)} 71.7% (0.061)33.9% (0.060)
ATE(D^) 74.6 (6.5)108.4 (9.9)
ATE(D^c) 34.1 (15.8)11.5 (8.7)
ATE(D^)ATE(D^c) 108.7 (13.2)96.8 (9.0)
RPI89.2% (0.027)97.2% (0.034)
CAPITALPenalty λ= 44
with large penalty pr{D^(X)} 51.9% (0.119)30.8% (0.032)
ATE(D^) 87.2 (13.2)112.6 (7.0)
ATE(D^c) 2.6 (15.9)13.9 (6.5)
ATE(D^)ATE(D^c) 89.9 (10.9)98.7 (8.9)
RPI92.2% (0.039)99.1% (0.015)
Virtual twins pr{D^(X)} 38.1% (0.043)12.9% (0.117)
ATE(D^) 113.8 (6.2)151.4 (29.2)
ATE(D^c) 1.4 (7.2)29.7 (13.9)
ATE(D^)ATE(D^c) 112.4 (7.9)121.7 (21.4)
RPI99.5% (0.010)99.9% (0.003)

DISCUSSION

In this article, we proposed a constrained policy tree search method, CAPITAL, to address the subgroup optimization problem. This approach identifies the theoretically optimal subgroup selection rule that maximizes the number of selected patients under the constraint of a pre‐specified clinically desired effect. Our proposed method is flexible and easy to implement in practice and has good interpretability. Extensive simulation studies show the improved performance of our proposed method over the popular virtual twins subgroup identification method, with larger selected benefitting subgroup sizes and estimated treatment effect sizes closer to the truth. We further demonstrated the broad usage of our methods for multiple use cases, different trait types, and varying constraint conditions. The key idea of our proposed algorithm is to transform the constraints defined at the population level into the individual rewards at the patient level. This enables us to identify the patients via policy tree search using rewards as a function of the estimated contrast function . In our numerical studies, we estimate based on the random forest method and out‐of‐bag prediction (Lu et al ), which is shown to have low bias and variance. We further consider the doubly robust (DR) learner in step 2 of algorithm 1 in Kennedy to estimate the conditional average treatment effect with or without additional regression in equation 2 of algorithm 1 in Kennedy. The corresponding results are provided in Table S3 in the Supplementary Material based on CAPITAL for Scenario 1 as an illustration, aggregated over 200 replicates. It can be shown from Table S3 that CAPITAL with the DR‐learner plus additional regression achieves comparable results as CAPITAL with the out‐of‐bag prediction. Yet, without regression, the doubly robust pseudo‐outcome in equation 1 of algorithm 1 in Kennedy is not consistent at the individual level and thus fails to find the optimal SSR. There are several possible extensions we may consider in future work. First, we only consider two treatment options in this article, while in clinical trials it is not uncommon to have more than two treatments available for patients. Thus, a more general method applicable to multiple treatments or even continuous treatment domains is desirable. Second, we only provide the theoretical form of the optimal SSR. It may be of interest to develop the asymptotic performance of the estimated SSR such as the convergence rate. Appendix S1 Supplemental Material Click here for additional data file.
  17 in total

1.  Tutorial in biostatistics: data-driven subgroup identification and analysis in clinical trials.

Authors:  Ilya Lipkovich; Alex Dmitrienko; Ralph B
Journal:  Stat Med       Date:  2016-08-03       Impact factor: 2.373

2.  Optimal treatment regimes for survival endpoints using a locally-efficient doubly-robust estimator from a classification perspective.

Authors:  Xiaofei Bai; Anastasios A Tsiatis; Wenbin Lu; Rui Song
Journal:  Lifetime Data Anal       Date:  2016-08-01       Impact factor: 1.588

3.  Precision Medicine.

Authors:  Michael R Kosorok; Eric B Laber
Journal:  Annu Rev Stat Appl       Date:  2019-03       Impact factor: 5.810

4.  Evaluating markers for selecting a patient's treatment.

Authors:  Xiao Song; Margaret Sullivan Pepe
Journal:  Biometrics       Date:  2004-12       Impact factor: 2.571

5.  A regression tree approach to identifying subgroups with differential treatment effects.

Authors:  Wei-Yin Loh; Xu He; Michael Man
Journal:  Stat Med       Date:  2015-02-05       Impact factor: 2.373

6.  Learning Optimal Personalized Treatment Rules in Consideration of Benefit and Risk: with an Application to Treating Type 2 Diabetes Patients with Insulin Therapies.

Authors:  Yuanjia Wang; Haoda Fu; Donglin Zeng
Journal:  J Am Stat Assoc       Date:  2017-03-31       Impact factor: 5.033

7.  Bayesian Nonparametric Policy Search with Application to Periodontal Recall Intervals.

Authors:  Qian Guan; Brian J Reich; Eric B Laber; Dipankar Bandyopadhyay
Journal:  J Am Stat Assoc       Date:  2019-10-09       Impact factor: 5.033

8.  Selecting Optimal Subgroups for Treatment Using Many Covariates.

Authors:  Tyler J VanderWeele; Alex R Luedtke; Mark J van der Laan; Ronald C Kessler
Journal:  Epidemiology       Date:  2019-05       Impact factor: 4.822

9.  CAPITAL: Optimal subgroup identification via constrained policy tree search.

Authors:  Hengrui Cai; Wenbin Lu; Rachel Marceau West; Devan V Mehrotra; Lingkang Huang
Journal:  Stat Med       Date:  2022-07-07       Impact factor: 2.497

10.  On restricted optimal treatment regime estimation for competing risks data.

Authors:  Jie Zhou; Jiajia Zhang; Wenbin Lu; Xiaoming Li
Journal:  Biostatistics       Date:  2021-04-10       Impact factor: 5.899

View more
  1 in total

1.  CAPITAL: Optimal subgroup identification via constrained policy tree search.

Authors:  Hengrui Cai; Wenbin Lu; Rachel Marceau West; Devan V Mehrotra; Lingkang Huang
Journal:  Stat Med       Date:  2022-07-07       Impact factor: 2.497

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.