| Literature DB >> 34708337 |
Emeric Sibieude1,2, Akash Khandelwal3, Pascal Girard2, Jan S Hesthaven4, Nadia Terranova5.
Abstract
A fit-for-purpose structural and statistical model is the first major requirement in population pharmacometric model development. In this manuscript we discuss how this complex and computationally intensive task could benefit from supervised machine learning algorithms. We compared the classical pharmacometric approach with two machine learning methods, genetic algorithm and neural networks, in different scenarios based on simulated pharmacokinetic data. Genetic algorithm performance was assessed using a fitness function based on log-likelihood, whilst neural networks were trained using mean square error or binary cross-entropy loss. Machine learning provided a selection based only on statistical rules and achieved accurate selection. The minimization process of genetic algorithm was successful at allowing the algorithm to select plausible models. Neural network classification tasks achieved the most accurate results. Neural network regression tasks were less precise than neural network classification and genetic algorithm methods. The computational gain obtained by using machine learning was substantial, especially in the case of neural networks. We demonstrated that machine learning methods can greatly increase the efficiency of pharmacokinetic population model selection in case of large datasets or complex models requiring long run-times. Our results suggest that machine learning approaches can achieve a first fast selection of models which can be followed by more conventional pharmacometric approaches.Entities:
Keywords: Deep learning; Genetic algorithm; Model-informed drug discovery and development; Neural network; Pharmacometrics; Population PK/PD
Mesh:
Year: 2021 PMID: 34708337 PMCID: PMC8940812 DOI: 10.1007/s10928-021-09793-6
Source DB: PubMed Journal: J Pharmacokinet Pharmacodyn ISSN: 1567-567X Impact factor: 2.745
Main features (structural + residual error) of models in the considered library approaches
| Input function | # compartments | Enterohepatic circulation | Output function | Residual error |
|---|---|---|---|---|
Bolus 0 order 0 order + Tlag 0 order 1 order + Tlag 1 order + transitory compartment 0 order + 1 order | 1 2 3 | No Yes | Linear Michaelis–Menten Linear + Michaelis–Menten | Additive Proportional Combined 1 Combined 2 |
T lag time
Summary of simulated datasets investigated using PMX and GA approaches
| Dataset | Input model | # compartments | Output model | Error model |
|---|---|---|---|---|
| Dataset 1 | Transit compartment + 1 order | 1 | Linear | Proportional |
| Dataset 2 | First order + 0 order | 1 | Linear | Combined 1 |
| Dataset 3 | Bolus | 2 | Michaelis–Menten + linear | Combined 1 |
| Dataset 4 | Tlag + 1st order | 2 | Linear | Additive |
| Dataset 5 | Bolus | 3 | Linear | Combined 1 |
GA genetic algorithm, PMX pharmacometric, T lag time
Fig. 1General workflow—the first step was the data and model library generation, followed by the investigation of the three approaches selected (PMX, GA, and NN). GA genetic algorithm, NN neural network, PMX pharmacometric
Fig. 2Hybrid GA—the hybrid component makes the GA convergence faster by performing an exhaustive search around the best models. GA genetic algorithm. N is a parameter (integer) set by the user for the GA
Example of generated population of size 3 in GA for A structural and residual error genes, and B statistical models for random effects
| Model | Absorption | Circulation | Compartments | Elimination | Error |
|---|---|---|---|---|---|
| A | |||||
| 1 | 001 (0 order) | 0 (no) | 10 (3 comp) | 00 (linear) | 01 (proportional) |
| 2 | 011 (1 order) | 0 (no) | 01 (2 comp) | 10 (mixed) | 00 (constant) |
| 3 | 000 (Bolus) | 1 (yes) | 00 (1 comp) | 01 (Michaelis–Menten) | 11 (combined 1) |
CL clearance, GA genetic algorithm, Ka 1 order absorption, Km and Vm Michaelis–Menten elimination, PK pharmacokinetic Q2 inter-compartmental clearance, TK0 0 order absorption, T lag time, V volume for central compartment, V2 volume for second compartment, V3 volume for third compartment. Parameters depend on the generated structural model
Fig. 3Example of training output parameters for the two NN tasks. CL clearance, Fr bioavailability, Ka 1 order absorption, Km and Vm Michaelis–Menten elimination, Ktr transitory compartment, Mtt transitory compartment, NN neural network, Q2 and Q3 inter-compartmental clearance, TK0 0 order absorption, T lag time, V volume for central compartment, V2 volume for second compartment, V3 volume for third compartment. Note For the regression task, individual pharmacokinetic parameters constitute the output to be predicted (top table). Data for the classification task can be derived from this by combining parameters into model components binarily labeled according to their presence or absence (bottom table)
Summary of GA-based model selection
| Model # | True model | Selected model | Generation # | Shrinkage-related penalty | Hybrid | Runtime (h) | Fitness | Δ-fitness |
|---|---|---|---|---|---|---|---|---|
| 1 | 1cmt, 1_abs, transit_cmt, lin_elim, prop_err | 1cmt, 1_abs, lin_elim, prop_err | 20 | 0 | No | 18.2 | 2370.6 | – 65 |
| 1cmt, 1_abs,, lin_elim, prop_err | 11 | 0 | Yes | 11.1 | 2370.6 | − 65 | ||
| 1cmt, 1_abs,, lin_elim, prop_err | 20 | 100 | No | 17.3 | 2376.5 | − 150.7 | ||
| 1cmt, 1_abs,, lin_elim, comb1_err | 11 | 100 | Yes | 17.7 | 2376.7 | − 150.5 | ||
| 2 | 1cmt, 1_abs, 0_abs, lin_elim, comb1_err | 1cmt, 1_abs, lin_elim, comb2_err | 20 | 0 | No | 22.8 | 2504.4 | − 887.7 |
| 1cmt, 1_abs, lin_elim, comb1_err | 11 | 0 | Yes | 11.2 | 2505.7 | − 886.4 | ||
| 1cmt, 1_abs, lag, lin_elim, comb1_err | 20 | 100 | No | 19.1 | 2522.2 | − 856 | ||
| 1cmt, 1_abs, lin_elim, comb2_err | 11 | 100 | Yes | 10.7 | 2592.4 | − 785.8 | ||
| 3 | 2_cmt, bolus, lin_elim, MM_elim, comb1_err | 1_cmt, bolus, lin_elim, MM_elim, comb_err | 20 | 0 | No | 16.2 | 1632.1 | − 346.8 |
| 1_cmt, bolus, lin_elim, MM_elim, comb1_err | 11 | 0 | Yes | 14.2 | 1633.4 | − 345.5 | ||
| 1_cmt, bolus, lin_elim, MM_elim, comb2_err | 20 | 100 | No | 17.3 | 1634.1 | − 364.4 | ||
| 1_cmt, bolus, lin_elim, MM_elim, comb2_err | 11 | 100 | Yes | 15.6 | 1638.2 | − 360.3 | ||
| 4 | 2_cmt, 1_abs, lag, lin_elim, add_err | 1_cmt, bolus, lin_elim, add_err | 20 | 0 | No | 20.5 | 4918.1 | − 94.1 |
| 1_cmt, bolus, lin_elim, MM_elim, prop_err | 11 | 0 | Yes | 14.6 | 4928.4 | − 83.8 | ||
| 1_cmt, bolus, lin_elim, add_err | 20 | 100 | No | 27.9 | 4921.5 | − 201.4 | ||
| 1_cmt, bolus, lin_elim, add_err | 11 | 100 | Yes | 8.7 | 4921.5 | − 201.4 | ||
| 5 | 3_cmt, bolus, lin_elim, comb1_err | 1_cmt, bolus, lin_elim, comb1_err | 20 | 0 | No | 14.2 | 2522.9 | − 101.2 |
| 1_cmt, bolus, lin_elim, comb1_err | 11 | 0 | Yes | 12.2 | 2522.9 | − 101.2 | ||
| 1_cmt, bolus, lin_elim, comb1_err | 20 | 100 | No | 18.1 | 2526.3 | − 109.3 | ||
| 1_cmt, bolus, lin_elim, comb1_err | 11 | 100 | Yes | 9.5 | 2526.3 | − 109.3 |
Δ delta, GA genetic algorithm, h hours, 1_cmt one compartment, 2_cmt two compartment, 3_cmt three compartment, 1_abs 1st order absorption, 0_abs 0 order absorption, lag lag time, transit_cmt transit compartments, lin_elim linear elimination, MM_elim Michaelis–Menten elimination, add_err additive error model, prop_err proportional error model, comb1_err combined1 error model, comb2_err combined2 error model. GA selection was considered successful if the best model in the last generation (selected model) had a fitness value smaller than the true model (negative Δ-fitness)
Fig. 4NN train and test MSE obtained for regression, A during the learning phase for the global NN, and B if 14 independent NN were trained for each of the parameters. On panel A, train and test MSE obtained during the learning phase for the global NN are shown in dashed and solid lines, respectively, for the full NN (red) and for the NN without prediction of Km and V3 (blue). On panel B, train and test MSE obtained during the learning phase are shown in dashed and solid lines, respectively, for 14 independent NN trained for each of the parameters. CL clearance, Fr bioavailability, Ka 1 order absorption, Km Michaelis–Menten elimination, Ktr transitory compartment, Mtt transitory compartment, MSE mean squared error, NN neural network, Q2 and Q3 inter-compartmental clearance, TK0 0 order absorption, T lag time, V volume for central compartment, V2 volume for second compartment, V3 volume for third compartment, Vm Michaelis–Menten elimination. Note Various NNs for regression were trained (Color figure online)
Fig. 5Evolution of the percentage of the label correctly predicted in the NN classification task. NN, neural network. Note: NN classification results are shown for scenario 1 (random split) where the test set was randomly selected (red curves), and for scenario 2 (non-random) where all observations of two models not included in the training set were selected to compose the test set (Color figure online)