| Literature DB >> 30816929 |
Alejandro F Villaverde1, Fabian Fröhlich2,3, Daniel Weindl2, Jan Hasenauer2,3, Julio R Banga1.
Abstract
MOTIVATION: Kinetic models contain unknown parameters that are estimated by optimizing the fit to experimental data. This task can be computationally challenging due to the presence of local optima and ill-conditioning. While a variety of optimization methods have been suggested to surmount these issues, it is difficult to choose the best one for a given problem a priori. A systematic comparison of parameter estimation methods for problems with tens to hundreds of optimization variables is currently missing, and smaller studies provided contradictory findings.Entities:
Mesh:
Year: 2019 PMID: 30816929 PMCID: PMC6394396 DOI: 10.1093/bioinformatics/bty736
Source DB: PubMed Journal: Bioinformatics ISSN: 1367-4803 Impact factor: 6.937
Main features of the benchmarks.The model IDs follow the nomenclature in Villaverde and Fröhlich )
| ID | B2 | B3 | B4 | B5 | BM1 | BM3 | TSP |
|---|---|---|---|---|---|---|---|
| Original ref. | |||||||
| Organism | Chinese hamster | Generic | Mouse | Human | Generic | ||
| Description | Metabolic | Metabolic | Metabolic | Signaling | Signaling | Signaling | Metabolic |
| level | & transcription | ||||||
| Parameters | 116 | 178 | 117 | 86 | 383 | 219 | 36 |
| Upper bounds | varying | ||||||
| Lower bounds | varying | ||||||
| Dynamic states | 18 | 47 | 34 | 26 | 104 | 500 | 8 |
| Observed states | 9 | 47 | 13 | 6 | 12 | 5 | 8 |
| Experiments | 1 | 1 | 1 | 10 | 1 | 4 | 16 |
| Data points | 110 | 7567 | 169 | 960 | 120 | 105 | 2688 |
| Data type | Measured | Simulated | Simulated | Simulated | Measured | Measured | Simulated |
| Noise level | Real | No noise | Variable | Uniform | Real | Real |
Noise levels are unknown as real measurement data are used.
Noise levels differ for readouts.
Noise level is constant (); the data values generated by this model are between 0 and 1 by construction.
Noise levels are proportional to the signal intensity.
Classification of the hybrid optimization methods considered in the benchmarking
| Global strategy | Local method & gradient calculation | Parameter scaling | |||
|---|---|---|---|---|---|
| FMINCON-ADJ | NL2SOL-FWD | DHC | None | ||
| MS | MS-FMINCON-ADJ-LOG | MS-NL2SOL-FWD-LOG | MS-DHC-LOG | – | LOG |
| MS-FMINCON-ADJ-LIN | MS-NL2SOL-FWD-LIN | MS-DHC-LIN | – | LIN | |
| eSS | eSS-FMINCON-ADJ-LOG | eSS-NL2SOL-FWD-LOG | eSS-DHC-LOG | eSS-NOLOC-LOG | LOG |
| eSS-FMINCON-ADJ-LIN | eSS-NL2SOL-FWD-LIN | eSS-DHC-LIN | eSS-NOLOC-LIN | LIN | |
Notes: These methods result from the combination of two global strategies with three local methods and two types of scaling for the search space. Additionally, we tested a global metaheuristics optimization method, Particle Swarm Optimization (PSO) both in logarithmic and linear scale (PSO-LOG, PSO-LIN). The abbreviations are defined in Sections 2.3 and 2.4.
Fig. 1.Illustration of performance criteria. (A) Convergence curves for three different methods. Shaded areas show the range of all runs, while solid lines represent their median. The dashed horizontal line is the value to reach (VTR), that is the maximum objective function value that can be considered a successful result. The dashed vertical line is the maximum time allowed (MAXT). (B) Dispersion plot of objective value after the maximum time allowed and the derived success rates (SR). The SR is the area under the curve where objective VTR. (C) Success rate and computation time. Points indicate individual methods. The Pareto front is the set of non-dominated methods. Methods to the right or above the Pareto front are dominated by other methods with either shorter computation time or higher success rate. Filled areas show the average computation time required to obtain a successful run for the respective method. For algorithms with a success rate of zero, meaning that no optimization run reached the VTR, 1/success rate is set to infinity
Fig. 2.Results of performance evaluation. (A) Convergence curves of the different methods for benchmark TSP. Results for the remaining benchmarks are reported in the Supplementary Material. (B) Average computation time of each method versus the inverse of its success rate for benchmark TSP. Methods with zero success rate are not shown. Results for the remaining benchmarks are reported in the Supplementary Material. (C) Cumulative overall efficiency: Each method is represented by a stack of the OEs observed for the individual benchmark problems. The maximum possible score is the same as the number of benchmarks, i.e. seven. (D) Successful methods for each benchmark are shown in color; methods which never succeeded for a given problem are shown in white. A, B and D use the thresholds VTR C and MAXT A as defined in the Supplementary Table S1. Panel C shows the average OE across all considered VTRs and MAXTs