Literature DB >> 28361667

Combination therapy design for maximizing sensitivity and minimizing toxicity.

Kevin Matlock1, Noah Berlow2, Charles Keller2, Ranadip Pal3.   

Abstract

BACKGROUND: Design of personalized targeted therapies involve modeling of patient sensitivity to various drugs and drug combinations. Majority of studies evaluate the sensitivity of tumor cells to targeted drugs without modeling the effect of the drugs on normal cells. In this article, we consider the individual modeling of drug responses to tumor and normal cells and utilize them to design targeted combination therapies that maximize sensitivity over tumor cells and minimize toxicity over normal cells.
RESULTS: The problem is formulated as maximizing sensitivity over tumor cell models while maintaining sensitivity below a threshold over normal cell models. We utilize the constrained structure of tumor proliferation models to design an accelerated lexicographic search algorithm for generating the optimal solution. For comparison purposes, we also designed two suboptimal search algorithms based on evolutionary algorithms and hill-climbing based techniques. Results over synthetic models and models generated from Genomics of Drug Sensitivity in Cancer database shows the ability of the proposed algorithms to arrive at optimal or close to optimal solutions in significantly lower number of steps as compared to exhaustive search. We also present the theoretical analysis of the expected number of comparisons required for the proposed Lexicographic search that compare favorably with the observed number of computations.
CONCLUSIONS: The proposed algorithms provide a framework for design of combination therapy that tackles tumor heterogeneity while satisfying toxicity constraints.

Entities:  

Keywords:  Combination drug design; Lexicographic search; Toxicity constraints

Mesh:

Year:  2017        PMID: 28361667      PMCID: PMC5374708          DOI: 10.1186/s12859-017-1523-1

Source DB:  PubMed          Journal:  BMC Bioinformatics        ISSN: 1471-2105            Impact factor:   3.169


Background

Design of drug therapies for cancer have primarily been considered from the perspective of sensitivity prediction using genetic characterizations as the predictor variables [1-3]. The genetic characterization based methodologies have severe limitations when the cancer type shows numerous aberrations among the samples and consequently predicting sensitivity based on similar steady state genetic characterizations provide limited accuracy. We have recently considered the modeling of tumor sensitivity using functional drug response data [4, 5], along with a functional and genetic characterization based integrated modeling [6]. Models were designed based on the in vitro tumor response to a set of drugs with known targets. However, the combination therapy design was based on the model reflecting the average behavior of the tumor tissue [7]. In this article, we incorporate heterogeneity by considering the effect of a drug on various parts of the tumors and incorporate toxicity by considering the effect of the drug on normal cell types. Consider a solid tumor tissue where the biopsy sample can be divided into separate samples to explore the heterogeneity. We can pass each biopsy sample through a drug screen to create a probabilistic target inhibition map (PTIM) [5] model. Let M T 1,M T 2,⋯,M T denote the k models corresponding to the k spatial tumor biopsies. For toxicity the affect of the drug is not limited to the same organ that the tumor resides in. To solve this we can pass normal cell cultures from different organs of the body through drug screens to create separate models of different organs to assess toxicity of the drugs. Note that, normal cells from kidney, lungs etc. of a specific cancer patient may not be readily available and thus, the response to drugs can be approximated by using drug screens on normal human based cell lines of kidney, lungs and other organs. The assumption is that variations in normal cell response to different drugs over different patients are smaller compared to tumor cell response over different patients. The normal cell response to different drugs can vary significantly for cells belonging to different organs in the body. Let M N 1,M N 2,⋯,M N denote p models corresponding to p different normal cell types. The goal of the combination therapy design will be to select a set of drugs that will maximize the sensitivity over heterogeneous tumor models M T 1,M T 2,⋯,M T and minimize sensitivity over normal cell type models M N 1,M N 2,⋯,M N . Note that currently available combination therapy design techniques are model free and require multiple experimental iterations to arrive at the optimal strategy [8-15]. This article considers model based combination therapy design over multiple models of tumor and normal cell lines. We utilize the constrained structure of tumor proliferation models to design an accelerated lexicographic search algorithm for generating the optimal solution. For comparison purposes, we also designed two suboptimal search algorithms based on evolutionary algorithms and hill-climbing based techniques. We test the performance of our algorithms on synthetic models and models generated from Genomics of Drug Sensitivity in Cancer (GDSC) database [16]. Utilizing the model structure in the search process allows us to arrive at optimal or close to optimal solutions in significantly lower number of steps as compared to exhaustive search. The article also presents the theoretical analysis of the expected number of comparisons required for the proposed optimal Lexicographic search that compare favorably with the observed number of computations. The paper is organized as follows: The model representation is discussed in “Methods” section. “Algorithms” section discusses the proposed lexicographic search algorithm along with suboptimal Genetic algorithm and Hill climbing approaches. “Results and discussion” section presents the results followed by Conclusions in “Conclusions” section.

Methods

Model type

In this section, we provide a brief review of the model used to represent each spatial tumor biopsy or normal cell line. A Probabilistic Target Inhibition Map (PTIM) model provides an estimate of sensitivity for all possible target inhibitions. Consider the example PTIM model with 3 targets k 1, k 2, k 3 shown in Fig. 1 where the values for each cell represent the sensitivity corresponding to that specific inhibition. For instance, inhibition of k 3 alone will produce a sensitivity of 0.75. Since the commonly used targeted drugs inhibit oncogenes, we consider the targets to be all oncogenes and inhibition of more oncogenes can only cause the sensitivity to remain same or increase. For instance, since the inhibition of k 3 alone produces a sensitivity of 0.75, all supersets of that inhibition ([k 3,k 1], [k 3,k 2], [k 3,k 1,k 2]) will have sensitivity ≥0.75. Similarly, any subset of known inhibition will have sensitivity less than or equal to the observed value. Based on these two biological constraints and limited drug perturbation experiments, we can arrive at an inferred PTIM model that can provide an estimate of sensitivity for all possible target inhibitions. The details of the model are available at [4-6] along with biological validation at [17]. Note that a PTIM can also be approximately represented as a tumor proliferation circuit as shown in Fig. 2 where the tumor proliferation can be restricted by inhibiting at least one series block. For instance, inhibition of the [k 1,k 2] block will provide a sensitivity of 0.95 whereas inhibition of k 3 will provide a sensitivity of 0.75. Inhibiting more than the minimum will produce higher sensitivities that are given by the original map shown in Fig. 1.
Fig. 1

PTIM example. An example PTIM model with 3 targets k 1, k 2 and k 3

Fig. 2

PTIM circuit. A circuit representation of a PTIM model in Fig. 1

PTIM example. An example PTIM model with 3 targets k 1, k 2 and k 3 PTIM circuit. A circuit representation of a PTIM model in Fig. 1

Structure of tumor and normal cell models

Based on the previously discussed model structure, each of the k tumor models will be represented as a probabilistic target inhibition map that can also be approximated by a circuit representation of series of parallel blocks as shown in Fig. 3.
Fig. 3

PTIM block diagram. Representation of k tumor models and p normal models as series of parallel target blocks

PTIM block diagram. Representation of k tumor models and p normal models as series of parallel target blocks In Fig. 3, the number of blocks for models M T for i=1,⋯,k and M N for j=1,⋯,p are denoted by n and n respectively. Every model is composed of five blocks connected in series. Each block, b, contains a set of targets T (up to a maximum of 5 targets), that are connected in parallel. Thus each model can have up to 25 targets.

Optimization objectives

For optimization, we consider both the worst case and best expected scenario. Let O(M T ,ϕ) denote the sensitivity of Tumor model i for i=1,⋯,k with inhibition ϕ. Let O(M N ,ϕ) denote the sensitivity of normal model i for i=1,⋯,p with inhibition ϕ. Worst case optimization (WCO): We desire high sensitivity over the tumor cell lines and low sensitivity over the normal cells which can be formulated in the worst case scenario as maximizing the minimum sensitivity over the tumorous cells while maintaining the maximum sensitivity over the normal cells below a certain threshold θ 1. i.e. maxϕ(mini[O(M T ,ϕ)]) while maxi [ O(M N ,ϕ) ] ≤θ 1 Best expected optimization (BEO): In this scenario, our goal will be to maximize the average sensitivity over the tumorous cells while maintaining the average sensitivity over the normal cells below a threshold θ 2 i.e. while .

Algorithms

Lexicographic search algorithm

In order to find the optimal target set for our problem, we can exhaustively search through all possible target combinations for a given toxicity threshold. Normally, for T targets this would require searching through 2 combinations, which is not computationally feasible for large T. However, we can utilize the monotonic relationships of PTIM models to our advantage to reduce the number of search steps. Given a set of targets S 1 whose toxicity (i.e. sensitivity over normal cell lines) is greater than a given threshold θ 1, then all possible supersets of S 1 will also have a toxicity ≥θ 1 and thus, there is no need to search through the supersets of S 1. Note that this is only valid when all the targets are oncogenes. To take advantage of this property, we perform a branching Lexicographical Search of the solution space. We can view the solution space as a directed graph where each node of the graph is our target set represented as a binary string with T bits. Each edge of the graph corresponds to turning on one bit to the right of the least significant bit, creating a superset of that node. If the toxicity at a node exceeds the threshold, then there is no need to continue along the associated edges and we should instead trace back to the previous node. A recursive algorithm to perform this search is shown in Algorithm 1. A demo using four targets is shown in Fig. 4. Note that in Fig. 4, we are assuming that the sensitivity over normal cell lines exceed the given threshold for target set [1100] and thus its supersets consisting of [1110], [1101] and [1111] marked by dotted lines are excluded from the search process.
Fig. 4

Lexicographic search example. Lexicographical Search for 4 Targets. Utilizing the superset rule, sets surrounded by dotted lines are excluded from the search process when toxicity of [1100] ≥θ 1

Lexicographic search example. Lexicographical Search for 4 Targets. Utilizing the superset rule, sets surrounded by dotted lines are excluded from the search process when toxicity of [1100] ≥θ 1

Lexicographic search analysis

In this section, we consider the stochastic analysis of the proposed Lexicographic Search to generate the expected number of searches. Let C D F(l,θ)=f(l,θ) denote the probability that a normal cell model will have sensitivity ≤θ for a random inhibition of l targets. We then define g (l,θ) as the probability that we will not exceed sensitivity θ while targeting l random inhibitions in a set of P cells. For the worst case scenario, is the probability that the maximum sensitivity over the normal cells is below threshold θ. Considering, independence of the normal cell sensitivities, we have: For the best expected scenario, let us consider the probability density function of observing a sensitivity of θ after l inhibitions . Let X denote the random variable with P D F(l,θ) and Z denote the sum of P such random variables. The probability density function of Z denoted by q (l,θ) can be calculated by repeatedly convolving P D F(l,θ) with itself for P−1 times and is given by Let Y denote the random variable denoting the average sensitivity over P cell lines with l random inhibitions. The probability density function of Y is given by We can then estimate the cumulative density function of Y, by integrating across θ:

Expected savings

Define A to denote the event that the sensitivity over P normal models with i random inhibitions ≥θ, i.e. P r(A )=1−g (i,θ). Let L denote the event of stoping at level i of the Lexicographic Search where i represents the number of bits we are searching through. The probability of event L is given by: We note that: By applying Bayes’ theorem, we can simplify further: By combining Eqs. 3, 4 and 5, we have: To find the expected savings, we note that by stopping at L , we search through combinations. Thus, the expected savings E(S) is given by :

Genetic algorithm based search

Pareto optimality

We consider a multi-objective optimization scenario where we maximize sensitivity over tumor cells and minimize sensitivity over normal cells. For worst case optimization scenario, if therapies ϕ 1 and ϕ 2 satisfy the following relation: m i n 1≤[ O(M T ,ϕ 1)]≥m i n 1≤[O(M T ,ϕ 2)] and m a x 1≤[ O(M N ,ϕ 1)]≤m a x 1≤[ O(M N ,ϕ 2)] with either m i n 1≤[O(M T , ϕ 1)]>m i n 1≤[O(M T ,ϕ 2)] or m a x 1≤[O(M N ,ϕ 1)]

Algorithm

Genetic Algorithms (GA) are inspired by evolutionary theory where strong species have a higher opportunity to pass their genes to offspring via reproduction and weaker chromosomes are eliminated by natural selection [18, 19]. Each generation or population consists of diverse individuals or chromosomes and in our Genetic Algorithm based Combination Therapy design (GACT), each therapy ϕ is regarded as a chromosome comprised of different target inhibitions. These target inhibitions are binary variables with values of 0 (non-inhibited) or 1 (completely inhibited). In order to select the best solutions (therapies) for the next generation, the fitness of each solution is computed. The therapies with the best fitness (our Pareto front) will be selected as the parents of the next generation. During each reproduction process, crossover and mutation operators are applied for the purpose of generating new solutions from existing ones. Mutation is performed by randomly flipping inhibition values of our targets. Crossover is performed by randomly picking values between two different target sets. For example, if we take the two target sets [a 1,a 2,a 3] and [b 1,b 2,b 3] a crossover between them can be performed by considering: Based on our starting set of targets (M), we form the initial population P 0 of N random target inhibition profiles. After calculating the fitness functions for the existing population, we calculate different Pareto front layers according to their dominance relationships. The top Pareto optimal points are selected to pairwise conduct crossover and mutations to form offsprings. Here we have set the number of offspring to be at least twice the number of points in our Pareto fronts with a minimum of O f f m i n=1000 and a maximum of n O f f m a x=15,000 offsprings. After merging these offsprings with their parent population P , we extract top N therapies to generate population P . We iterate our algorithm until we have achieved totalG generations or the number of offspring is greater than nOffmax. Note that evolutionary algorithms like GA will not guarantee convergence of the Pareto front but the performance of our therapies will improve if the Pareto front moves towards our desired direction with subsequent GA iterations. The detailed procedure for multi-objective GACT is shown as Algorithm 2. Figure 5 illustrates how the algorithm moves our pareto front towards better solutions across successive iterations. After running the GACT, we consider the final Pareto front and pick the target set that provides the maximum sensitivity over tumor cell lines when the toxicity is below threshold θ 1.
Fig. 5

GACT example. Pareto fronts converge for subsequent genetic algorithm iterations

GACT example. Pareto fronts converge for subsequent genetic algorithm iterations

Random restart hill climbing

We consider an additional suboptimal algorithm based on Hill Climbing to search the target space. Hill Climbing is an iterative method for finding the local maximum for any arbitrary function. Given a starting point, the algorithm considers all the nearest neighbors and then selects the neighbor that provides the best solution for the given optimization criteria. These steps are subsequently repeated until there are no neighbors that provide a better solution or a maximum number of iterations have been reached. While simple and effective in finding local optimum, Hill Climbing will rarely find the global optimum for non-convex functions. In order to overcome this handicap, we will randomly restart our search to a new random position whenever the algorithm converges to a local optimum. In our case, the starting criteria will be a random set of targets chosen using latin hypercube numbers and each neighbor will be created by inhibiting or un-inhibiting a single target. As shown in Algorithm 3, our optimization criteria will change depending on the toxicity of our current set. If the toxicity is greater than the given threshold then we choose the neighbor with the least toxicity, otherwise we pick the neighbor with the highest sensitivity whose toxicity is below the threshold. When no improvements can be found among the neighbors we randomly choose a new set of targets. This is repeated until we have completed maxIter iterations.

Results and discussion

To evaluate the performance of our algorithms, we considered both synthetic models and models based on experimental datasets.

Synthetic model generation

The synthetic models are simulated using a proliferation network structure based on probabilistic target inhibition maps [5, 7]. Each cellular pathway, i, representing either a tumor or normal cell model is modelled by connecting a set of blocks in series. The number of blocks for models M T for i=1,⋯,k and M N for j=1,⋯,p are denoted by n and n respectively. Within each block, b, contains a set of targets T (up to a maximum of 5 targets), that are connected in parallel. Since the targets are in parallel, the effective inhibition for each block given a set of target inhibitions, ϕ, is the minimum inhibition of the given targets within the block. Thus, the effective inhibition of block b in model M T with target inhibition ϕ is given by λ(M T ,b,ϕ)= min(∀ϕ∈T ). Each block is also given a score, S , randomly using a uniform distribution with a minimum of 0.5 and maximum of 1. Finally, the overall sensitivity of the pathway can be computed using the following equation where we assume independence between the series blocks: Similary for normal cells: A representation of k tumor and p normal models as series of parallel target blocks is shown in Fig 3. The synthetic model set consists of a total of 1000 synthetic pathways, 500 normal and 500 cancerous. A total of 25 targets are examined and all targets are equiprobable in both the cancer and normal pathways. We group the pathways into 100 groups where each group has 5 normal pathways and 5 cancerous pathways. From every group, we consider nNormal normal pathways and nTumor cancerous pathways.

GDSC data

In order to test our algorithms on biological functional data, we have utilized the GDSC database [16] to generate a set of PTIM models for 20 different cell lines. These cell lines were segregated into groups of 10, the first group is composed of breast-cancer cell lines and the second group is B-cell lymphoma cancer cell lines. A list of the cell lines is shown in Table 1. For each of the cell lines, we considered the IC50 values for 32 drugs and combined with the corresponding drug panels generated a PTIM model. The drug panels contained the K values for 404 targets and 62 of these targets were found to correspond to the PTIM model of at least one of the cell lines, 42 targets in the breast cell lines, 49 in the lymphoma and 27 targets where found in both the breast and lymphoma cell lines.
Table 1

Cell lines used in GDSC dataset

BreastLymphoma
BT474A3KAW
COLO824A4FUK
DU4475BC1
EVSATCTB1
HCC1187DOHH2
HCC2157HT
HCC2218KARPAS422
MFM223MC116
MRKnu1RL
OCUBMTUR
Cell lines used in GDSC dataset For the generation of sets of tumor and normal models, we generate 100 groups and for each group we randomly assign one type of cell lines (breast or lymphoma) to be the Normal cells and the other type to be the Tumor cells. We then randomly pick nNormal cell lines from the corresponding group and assign them as Normal cell models. Likewise, we pick nTumor random cell lines from the other group and assign them as our Tumor cell models.

Performance comparisons

For both the synthetic and GDSC cases, we select the solution that provides the maximum cancer cell sensitivity while keeping the normal cell sensitivity below a threshold of θ=0.1 using both the best expected scenario and the worst case optimization method. The maximum cancer sensitivity is then averaged across all 100 groups.

GACT parameter selection

In order to find an optimal r a t e and r a t e for the GACT, we set n T u m o r,n N o r m a l=5 and t o t a l G=100. We then repeated the WCO GACT on the synthetic dataset, varying either r a t e or r a t e on each run. The results are shown in Table 2. Based on these results, we have selected a r a t e of 0.2 and r a t e of 0.6. We then start varying nGen and the corresponding results are recorded in Table 3. From these results we can see that past 400 iterations, there is no significant improvement in the GACT.
Table 2

Worst case optimization tumor cell sensitivity using GACT for sythetic models

r a t e c
0.60.70.80.9
r a t e m 0.020.58260.58910.59330.5643
0.050.59030.58420.58770.5753
0.10.57700.58290.57570.5728
0.20.59380.59230.58480.5812
0.50.56690.56730.57620.5796
Table 3

WCO GACT for synthetic models with varying nGen

nGen Sensitivity
1000.5938
2000.6009
3000.6027
4000.6094
5000.6074
Worst case optimization tumor cell sensitivity using GACT for sythetic models WCO GACT for synthetic models with varying nGen

Hill climbing parameter selection

We perform a similar operation on the Hill Climbing algorithm where we keep n T u m o r=n N o r m a l=5 and vary nIter and measuring the best expected outcome in contrast to worst case optimization in previous case. The results are shown in Table 4. Since, there are no significant changes in performance after 15,000 iterations, we use m a x I t e r=15000 for running our Hill Climbing algorithm.
Table 4

Hill climbing BEO performance with varying number of iterations

maxIter Sensitivity
100000.8029
150000.8136
200000.8214
250000.8248
300000.8275
Hill climbing BEO performance with varying number of iterations

WCO results

The results for worst case optimization for the three approaches on synthetic and biological models are shown in Tables 5, 6, 7, 8, 9 and 10. For the synthetic dataset, we achieved minimum cancer sensitivities >0.60 using the LS (Table 5) and GACT (Table 6) algorithms while the Hill Climbing algorithm could only achieve a sensitivity of 0.55 (Table 7). Since the LS algorithm is the optimal approach, it produces the best performance followed closely by the GACT algorithm for synthetic dataset. However, for the GDSC dataset, the GACT was only able to achieve a sensitivity of 0.33 which is significantly worse than the LS which was able to achieve a sensitivity of 0.45. For the GDSC dataset, it appears that GACT is close to the optimal solution for smaller number of normal cells and increasing the number of normal cells reduces its performance (Table 9). Hill Climbing has reasonable performance for GDSC data based models (Table 10).
Table 5

LS worst case optimization synthetic data

nNormal
12345
nTumor 10.99340.98280.96330.96810.9096
20.98670.96030.90990.92550.7852
30.97940.94540.88360.81100.7331
40.96810.92550.84870.77280.6858
50.96070.90050.81130.71940.6157
Table 6

GACT worst case optimization synthetic data

nNormal
12345
nTumor 10.99340.98280.96330.93840.9096
20.98620.95990.90850.85940.7843
30.97720.94390.88110.80070.7268
40.96470.92000.84300.76790.6717
50.95610.89380.80560.70070.6097
Table 7

Hill climbing worst case optimization synthetic data

nNormal
12345
nTumor 10.99260.98170.96140.93690.9086
20.98050.95500.90310.84490.7635
30.96250.92700.86500.79290.7016
40.94020.89230.80890.73410.6241
50.91370.85140.76300.65730.5536
Table 8

LS worst case optimization GDSC data

nNormal
12345
nTumor 10.97920.95360.92180.85050.8282
20.96010.92260.87180.78530.7556
30.93260.87670.82070.70170.6270
40.91630.82800.74050.61260.4880
50.90450.79780.68090.55350.4539
Table 9

GACT worst case optimization GDSC data

nNormal
12345
nTumor 10.97890.95360.87380.75320.6056
20.95980.91240.81210.69000.4951
30.93200.86880.80620.64010.4843
40.91520.82100.71300.55250.3458
50.90280.79050.65510.47350.3348
Table 10

Hill climbing worst case optimization GDSC data

nNormal
12345
nTumor 10.97880.93900.88590.84890.8230
20.95880.87620.81630.73990.7074
30.93570.82830.72830.62150.5724
40.91480.77730.65330.52860.4683
50.89590.74220.59360.45840.3908
LS worst case optimization synthetic data GACT worst case optimization synthetic data Hill climbing worst case optimization synthetic data LS worst case optimization GDSC data GACT worst case optimization GDSC data Hill climbing worst case optimization GDSC data

BEO results

The results for best expected optimization scenario for the three algorithms on synthetic and biological models are shown in Tables 11, 12, 13, 14, 15 and 16. For the synthetic dataset, we achieved expected sensitivities over cancer models higher than 0.85 using the LS (Table 11) and GACT (Table 12) algorithms and >0.81 for Hill Climbing Approach (Table 13). For the GDSC based models, performance close to LS (Table 14) is observed with GACT (Table 15) followed by Hill Climbing (Table 16).
Table 11

LS best expected optimization synthetic data

nNormal
12345
nTumor 10.99340.98280.96330.93840.9096
20.99160.97560.94680.91700.8744
30.99010.97340.94610.91100.8698
40.98830.97020.93550.90490.8662
50.98720.96680.93050.89840.8596
Table 12

GACT best expected optimization synthetic data

nNormal
12345
nTumor 10.99340.98280.96320.93840.9095
20.99120.97550.94510.91560.8741
30.98890.97160.94430.91020.8664
40.98610.96790.93250.90240.8640
50.98300.96290.92700.89450.8549
Table 13

Hill climbing best expected optimization synthetic data

nNormal
12345
nTumor 10.99230.98170.96210.93770.9069
20.98530.97080.94090.90850.8524
30.97790.95970.93050.89000.8496
40.96840.94850.91380.87830.8268
50.96030.93610.90160.86280.8123
Table 14

LS best expected outcome GDSC data

nNormal
12345
nTumor 10.97920.95850.93150.86810.8572
20.97830.96150.93600.89090.8813
30.97370.95680.93660.89100.8653
40.97350.95110.92630.88510.8508
50.97360.94980.92190.88260.8544
Table 15

GACT best expected optimization GDSC data

nNormal
12345
nTumor 10.97890.95840.89160.84740.8129
20.97810.96110.93570.86090.8672
30.97340.94780.89960.87090.8246
40.97320.95020.89750.87710.8328
50.97330.94020.91240.87330.8407
Table 16

Hill climbing best expected optimization GDSC data

nNormal
12345
nTumor 10.97770.94470.90300.86000.8464
20.97380.93600.89950.85890.8286
30.97340.93690.89560.85430.8267
40.97310.93700.89560.85010.8152
50.97310.93800.89340.85020.8177
LS best expected optimization synthetic data GACT best expected optimization synthetic data Hill climbing best expected optimization synthetic data LS best expected outcome GDSC data GACT best expected optimization GDSC data Hill climbing best expected optimization GDSC data

Computational complexity

Tables 17 and 18 shows the average number of searches and runtime for each algorithm for synthetic and GDSC datasets respectively. All results are for n N o r m a l=n T u m o r=5. Each program is written using MATLAB and ran on a Inspiron 15 laptop with a Core i5-6300HQ processor with 8 GB of RAM. For the synthetic models, GACT was the fastest, taking less than a second for both WCO and BEO. The second fastest was the Hill Climbing (HC) algorithm that takes around 20 s per group while the LS algorithm takes the longest at over 125 s per group. It should be noted that the GACT is able to perform a proportionally larger number of searches in a shorter amount of time than both the LS and HC algorithm. This is because the GACT uses a small number of iterations but performs a large number of searches per iteration and is thus able to be vectorized more than the other algorithms.
Table 17

Computational complexity synthetic data

WCOBEO
SearchesAvg.SearchesSearchesAvg.Searches
Time (s)per sTime (s)per s
LS1033500132.278141033500126.37934
GACT5010000.60528278255010000.5806862900
HC38013020.111890237894016.6922705
Table 18

Computational complexity GDSC data

WCOBEO
SearchesAvg.SearchesSearchesAvg.Searches
Time (s)per sTime (s)per s
LS8790.095925320320.131715429
GACT1464304.3233896308501.0130433
HC30065013.822174929958011.4126256
Computational complexity synthetic data Computational complexity GDSC data In contrast for the GDSC dataset, the LS algorithm performs the fastest due to the small number of searches required. The GACT is now the second fastest and HC is the slowest. It should be noted that the GACT performs significantly slower in the GDSC dataset than the synthetic case. This is because we are unaware of the underlying circuit models of the PTIMs in the GDSC dataset as the sensitivities are computed using a lookup table and cannot be vectorized as efficiently as the synthetic data. In the next section, we explain the smaller number of searchers required for the LS approach for GDSC data as compared to Synthetic data based on the the specific structures of g(l,θ).

Estimated number of searches

In this section, we empirically generate the distributions of searches required for LS technique for random sets of normal and cancerous cells for both synthetic and GDSC data and compare them to theoretical estimates. In order to generate f(l,θ), we randomly select l targets and record the measured sensitivity for each cell line following inhibition of the selected l targets. This process is repeated 300,000 times for l=1 to nTargets to generate the entire CDF. The distributions for Synthetic, Breast cancer and Lymphoma cell lines are shown in Figs. 6, 7 and 8 respectively.
Fig. 6

CDFs for synthetic cells and WCO

Fig. 7

CDFs for breast cells and WCO

Fig. 8

CDFs for lymphoma cells and WCO

CDFs for synthetic cells and WCO CDFs for breast cells and WCO CDFs for lymphoma cells and WCO The corresponding distributions for Synthetic, Breast cancer and Lymphoma cell lines are shown in Figs. 9, 10 and 11 respectively.
Fig. 9

CDFs for synthetic cells and BEO

Fig. 10

CDFs for breast cells and BEO

Fig. 11

CDFs for lymphoma cells and BEO

CDFs for synthetic cells and BEO CDFs for breast cells and BEO CDFs for lymphoma cells and BEO We then use the theoretical estimate outlined in “Lexicographic search analysis” section to predict the number of searches required by the LS algorithm. In all the scenarios, we are assuming P=5. The predicted value is plotted against the actual value fitted to a normal distribution and the plots are shown in Fig. 12 for synthetic data and Fig. 13 for the GDSC data. We note that the theoretical estimates are very close to the mean of the distributions confirming our savings estimate.
Fig. 12

Estimated vs. actual searches for synthetic dataset

Fig. 13

Estimated vs. actual searches for GDSC dataset

Estimated vs. actual searches for synthetic dataset Estimated vs. actual searches for GDSC dataset To explain the limited number of searches required for GDSC models as compared to synthetic models, note that and have gradual decreases with l for synthetic data as shown in Figs. 6 and 9 respectively whereas and for GDSC models have sharp decreases with l as shown in Figs. 7, 8, 10 and 11. Due to the sharp change in the probabilities, the expected number of searches required for LS for GDSC models (in the 600 to 2200 range) is significantly lower than the expected number of searches required for the synthetic models (in the 106 range).

Conclusions

In this paper, we have formulated the combination therapy design problem of maximizing efficacy while minimizing toxicity as an algorithmic search problem to find the optimal target set that maximally inhibit spatially heterogeneous cancer cell models while maintaining the effect along multiple normal cell models below a certain threshold. Our cell proliferation models are based on probabilistic target inhibition map (PTIM) framework [4-7] that consists of a series of blocks where each block contains a set of targets connected in parallel. To find the ideal target inhibition profile, we proposed a lexicographic search method to effectively search through all possible solutions. This method takes advantage of the properties of the PTIM to significantly reduce the number of potential solutions that we have to search through. We compare the performance and computational complexity of this method with other commonly used algorithms such as Genetic algorithms and Hill Climbing. We demonstrated the effectiveness of our algorithms using both synthetic models and models generated from Drug Sensitivity of Cancer Database. A theoretical analysis of the expected number of steps required by the Lexicographic Search process is included that was shown to provide a close approximation to actual observed search steps.
  16 in total

Review 1.  Multi-target therapeutics: when the whole is greater than the sum of the parts.

Authors:  Grant R Zimmermann; Joseph Lehár; Curtis T Keith
Journal:  Drug Discov Today       Date:  2006-11-28       Impact factor: 7.851

2.  The right combination.

Authors: 
Journal:  Nature       Date:  2006-01-26       Impact factor: 49.962

3.  Closed-loop control of cellular functions using combinatory drugs guided by a stochastic search algorithm.

Authors:  Pak Kin Wong; Fuqu Yu; Arash Shahangian; Genhong Cheng; Ren Sun; Chih-Ming Ho
Journal:  Proc Natl Acad Sci U S A       Date:  2008-03-20       Impact factor: 11.205

4.  An Integrated Approach to Anti-Cancer Drug Sensitivity Prediction.

Authors:  Noah Berlow; Saad Haider; Qian Wan; Mathew Geltzeiler; Lara E Davis; Charles Keller; Ranadip Pal
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2014 Nov-Dec       Impact factor: 3.710

5.  A kinase inhibition map approach for tumor sensitivity prediction and combination therapy design for targeted drugs.

Authors:  Ranadip Pal; Noah Berlow
Journal:  Pac Symp Biocomput       Date:  2012

6.  Algorithmic guided screening of drug combinations of arbitrary size for activity against cancer cells.

Authors:  Ralph G Zinner; Brittany L Barrett; Elmira Popova; Paul Damien; Andrei Y Volgin; Juri G Gelovani; Reuben Lotan; Hai T Tran; Claudio Pisano; Gordon B Mills; Li Mao; Waun K Hong; Scott M Lippman; John H Miller
Journal:  Mol Cancer Ther       Date:  2009-03-10       Impact factor: 6.261

7.  Systematic discovery of multicomponent therapeutics.

Authors:  Alexis A Borisy; Peter J Elliott; Nicole W Hurst; Margaret S Lee; Joseph Lehar; E Roydon Price; George Serbedzija; Grant R Zimmermann; Michael A Foley; Brent R Stockwell; Curtis T Keith
Journal:  Proc Natl Acad Sci U S A       Date:  2003-06-10       Impact factor: 11.205

8.  Enhanced stochastic optimization algorithm for finding effective multi-target therapeutics.

Authors:  Byung-Jun Yoon
Journal:  BMC Bioinformatics       Date:  2011-02-15       Impact factor: 3.169

9.  Genomics of Drug Sensitivity in Cancer (GDSC): a resource for therapeutic biomarker discovery in cancer cells.

Authors:  Wanjuan Yang; Jorge Soares; Patricia Greninger; Elena J Edelman; Howard Lightfoot; Simon Forbes; Nidhi Bindal; Dave Beare; James A Smith; I Richard Thompson; Sridhar Ramaswamy; P Andrew Futreal; Daniel A Haber; Michael R Stratton; Cyril Benes; Ultan McDermott; Mathew J Garnett
Journal:  Nucleic Acids Res       Date:  2012-11-23       Impact factor: 16.971

10.  Search algorithms as a framework for the optimization of drug combinations.

Authors:  Diego Calzolari; Stefania Bruschi; Laurence Coquin; Jennifer Schofield; Jacob D Feala; John C Reed; Andrew D McCulloch; Giovanni Paternostro
Journal:  PLoS Comput Biol       Date:  2008-12-26       Impact factor: 4.475

View more
  6 in total

1.  In silico drug combination discovery for personalized cancer therapy.

Authors:  Minji Jeon; Sunkyu Kim; Sungjoon Park; Heewon Lee; Jaewoo Kang
Journal:  BMC Syst Biol       Date:  2018-03-19

2.  Deep Functional and Molecular Characterization of a High-Risk Undifferentiated Pleomorphic Sarcoma.

Authors:  Noah E Berlow; Catherine S Grasso; Michael J Quist; Mingshan Cheng; Regina Gandour-Edwards; Brian S Hernandez; Joel E Michalek; Christopher Ryan; Paul Spellman; Ranadip Pal; Lynn S Million; Mark Renneker; Charles Keller
Journal:  Sarcoma       Date:  2020-05-04

3.  Predicting effective drug combinations using gradient tree boosting based on features extracted from drug-protein heterogeneous network.

Authors:  Hui Liu; Wenhao Zhang; Lixia Nie; Xiancheng Ding; Judong Luo; Ling Zou
Journal:  BMC Bioinformatics       Date:  2019-12-09       Impact factor: 3.169

4.  Multiobjective optimization identifies cancer-selective combination therapies.

Authors:  Otto I Pulkkinen; Prson Gautam; Ville Mustonen; Tero Aittokallio
Journal:  PLoS Comput Biol       Date:  2020-12-28       Impact factor: 4.475

5.  Heterogeneity Aware Random Forest for Drug Sensitivity Prediction.

Authors:  Raziur Rahman; Kevin Matlock; Souparno Ghosh; Ranadip Pal
Journal:  Sci Rep       Date:  2017-09-12       Impact factor: 4.379

6.  Suppression of stress induction of the 78-kilodalton glucose regulated protein (GRP78) in cancer by IT-139, an anti-tumor ruthenium small molecule inhibitor.

Authors:  Suzanne J Bakewell; Daisy F Rangel; Dat P Ha; Jyothi Sethuraman; Richard Crouse; Emma Hadley; Tara L Costich; Xingliang Zhou; Peter Nichols; Amy S Lee
Journal:  Oncotarget       Date:  2018-07-03
  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.