| Literature DB >> 33265683 |
Liang Yan1, Xiaojun Duan1, Bowen Liu1, Jin Xu1.
Abstract
Bayesian optimization (BO) based on the Gaussian process (GP) surrogate model has attracted extensive attention in the field of optimization and design of experiments (DoE). It usually faces two problems: the unstable GP prediction due to the ill-conditioned Gram matrix of the kernel and the difficulty of determining the trade-off parameter between exploitation and exploration. To solve these problems, we investigate the K-optimality, aiming at minimizing the condition number. Firstly, the Sequentially Bayesian K-optimal design (SBKO) is proposed to ensure the stability of the GP prediction, where the K-optimality is given as the acquisition function. We show that the SBKO reduces the integrated posterior variance and maximizes the hyper-parameters' information gain simultaneously. Secondly, a K-optimal enhanced Bayesian Optimization (KO-BO) approach is given for the optimization problems, where the K-optimality is used to define the trade-off balance parameters which can be output automatically. Specifically, we focus our study on the K-optimal enhanced Expected Improvement algorithm (KO-EI). Numerical examples show that the SBKO generally outperforms the Monte Carlo, Latin hypercube sampling, and sequential DoE approaches by maximizing the posterior variance with the highest precision of prediction. Furthermore, the study of the optimization problem shows that the KO-EI method beats the classical EI method due to its higher convergence rate and smaller variance.Entities:
Keywords: K-optimal design; bayesian optimization; design of experiments; gaussian processes
Year: 2018 PMID: 33265683 PMCID: PMC7513107 DOI: 10.3390/e20080594
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Four situations considering the acquisition function and K-optimality simultaneously.
|
|
| |
|---|---|---|
|
|
|
|
|
|
|
|
Figure 1Demonstration of the effect of c on the shape of .
Means and standard deviations of the leave-one-out cross validation error (LOO-CV), the integrated posterior variance (IPV), the root mean squared error (RMSE) and the condition number (CN) based on 4 DoEs in Example 1, where the number in the parentheses is the standard deviation and the bold numbers represent the best outcomes.
| LOO-CV | IPV | RMSE | CN | |
|---|---|---|---|---|
| MC | 0.0778 (0.1100) | 0.1644 (0.1734) | 1.2651 × 10 | |
| LHS | 0.7557 (0.2217) | 0.0566 (0.1032) | 0.0837 (0.1354) | 3.2924 (0.9513) |
| MPV | 0.6242 (0.1916) | 0.0104 (0.0044) | 0.0254 (0.0078) | 2.5248 (0.5519) |
| BKO | 0.6771 ( |
Means and standard deviations of the LOO-CV, IPV, RMSE, and CN based on 4 DoEs in Example 2 where the number in the parentheses is the standard deviation. Bold numbers represent the best outcomes.
| LOO-CV | IPV | RMSE | CN | |
|---|---|---|---|---|
| MC | 0.1832 (0.1416) | 300.0422 (219.0879) | 23.0570 (8.8187) | 7.4361 × 10 |
| LHS | 0.3032 (0.2120) | 216.2945 (223.1399) | 20.6949 (9.0519) | 1.5387 × 10 |
| MPV | 0.1825 (0.1806) | 109.4216 (103.9399) | 12.8602 ( | 195.8292 (27.0753) |
| BKO |
The input variables of the Borehole function and their usual distributions.
| Input | Distribution | Unit | |
|---|---|---|---|
|
| radius of borehole |
| m |
|
| radius of influence |
| m |
|
| transmissivity of upper aquifer |
| m |
|
| potentiometric head of upper aquifer |
| m |
|
| transmissivity of lower aquifer |
| m |
|
| potentiometric head of lower aquifer |
| m |
|
| length of borehole |
| m |
|
| hydraulic conductivity of borehole |
| m/yr |
Means and standard deviations of the LOO-CV, IPV, RMSE, and CN based on 4 DoEs in Example 3, where the numbers in parentheses are the standard deviations and the bold numbers are the best results.
| LOO-CV | IPV | RMSE | CN | |
|---|---|---|---|---|
| MC | 0.0024 (0.0012) | 8.5170 (1.0624) | 2.0212 (0.3855) | 1 + 1.0808 × 10 |
| LHS | 1 + 1.0808 × 10 | |||
| MPV | 0.0033 (0.0028) | 11.2356 (2.8431) | 2.7163 (0.5043) | 1 + 6.3612 × 10 |
| BKO | 0.0026 (0.0013) | 8.8565 ( | 2.2822 (0.2895) |
Figure 2Comparisons of the expected improvement (EI) and K-optimal enhanced EI (KO-EI) algorithms. The above subfigure of each figure illustrates the predictions (darker line) and corresponding standard deviations (lighter line), and the one beloowdemonstrates the values of acquisition functions. The red and blue squares represent the current samples for the EI and KO-EI algorithm respectively. The red dots are the best points to be sampled according to the EI criterion, while the blue pentagrams were collected w.r.t the KO-EI algorithm.
Figure 3Comparison of convergence rate between the EI and KO-EI algorithms. The solid dash line represents the mean value of the minimal observed objective, while the transparent region represents the standard deviation w.r.t 100 independent simulations. (a) Viana function; (b) Branin function.
Figure 4Comparison of accuracy between the EI and KO-EI algorithms on the logistic regression classification on MNIST data. (a) Demonstration of the MNIST data; (b) The solid dash line represents the mean value of the classification accuracies while the transparent region represents the standard deviation w.r.t 50 independent simulations.