| Literature DB >> 33265298 |
Hossein Foroozand1, Valentina Radić2, Steven V Weijs1.
Abstract
Recently, the Entropy Ensemble Filter (EEF) method was proposed to mitigate the computational cost of the Bootstrap AGGregatING (bagging) method. This method uses the most informative training data sets in the model ensemble rather than all ensemble members created by the conventional bagging. In this study, we evaluate, for the first time, the application of the EEF method in Neural Network (NN) modeling of El Nino-southern oscillation. Specifically, we forecast the first five principal components (PCs) of sea surface temperature monthly anomaly fields over tropical Pacific, at different lead times (from 3 to 15 months, with a three-month increment) for the period 1979-2017. We apply the EEF method in a multiple-linear regression (MLR) model and two NN models, one using Bayesian regularization and one Levenberg-Marquardt algorithm for training, and evaluate their performance and computational efficiency relative to the same models with conventional bagging. All models perform equally well at the lead time of 3 and 6 months, while at higher lead times, the MLR model's skill deteriorates faster than the nonlinear models. The neural network models with both bagging methods produce equally successful forecasts with the same computational efficiency. It remains to be shown whether this finding is sensitive to the dataset size.Entities:
Keywords: EEF method; ENSO; El Niño; bagging; bootstrap aggregating; bootstrap neural networks; ensemble model simulation criterion; entropy ensemble filter; neural network forecast; sea surface temperature; tropical Pacific
Year: 2018 PMID: 33265298 PMCID: PMC7512722 DOI: 10.3390/e20030207
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Examples of studies on machine learning algorithms with bagging methods and summary of their discussion on computational efficiency.
| Authors (Year) | Machine Learning Method * | Computational Efficiency of the Bagging Method |
|---|---|---|
| Wan et al. (2016) [ | HANN and BBNN | The bootstrapped NN training process is extremely time-consuming. The HANN approach is nearly 200 times faster than the BBNN approach with 10 hours runtime. |
| Liang et al. (2016) [ | BNN and BMH | The bootstrap sample cannot be very large for the reason of computational efficiency. |
| Zhu et al. (2016) [ | BNN | The proposed improvement in accuracy comes at the cost of time-consumption during the network training. |
| Gianola et al. (2014) [ | GBLUP | Bagging is computationally intensive when one searches for an optimum value of BLUP-ridge regression because of the simultaneous bootstrapping. |
| Faridi et al. (2013) [ | ANN | Each individual network is trained on a bootstrap re-sampling replication of the original training data. |
| Wang et al. (2011) [ | ANN and GPR | Subagging gives similar accuracy but requires less computation than bagging. This advantage is especially remarkable for GPR since its computation increases in cubic order with the increase of data. |
| Mukherjee and Zhang (2008) [ | BBNN | Dividing the batch duration into fewer intervals will reduce the computation effort in network training and batch optimisation. However, this may reduce the achievable control performance … |
| Yu and Chen (2005) [ | BNN, SVM, and MLE-GP | Fully Bayesian methods are much more time consuming than SVM and MLE- GP. |
| Rowley et al. (1998) [ | ANN | To improve the speed of the system different methods have been discussed, but this work is preliminary and is not intended to be an exhaustive exploration of methods to optimize the execution time. |
* ANN (artificial neural network), BNN (Bayesian neural network), HANN (hybrid artificial neural network), BBNN (bootstrap-based neural network), BMH (bootstrap Metropolis-Hastings), GPR (Gaussian process regression), GBLUP (genomic best linear unbiased prediction), MLE-GP (maximum likelihood estimation-based Gaussian process), SVM (support vector machine) and V-SVM (virtual support vector machine).
Figure 1The flowchart of Entropy Ensemble Filter (EEF) method applied in the study.
Figure 2Spatial patterns (eigenvectors) of the first five PCA modes for the SST anomaly field. The percentage variance explained by each mode is given in the panel titles.
Figure 3Correlation skill of predictions of the five leading principal components of the SST fields at lead times from 3 to 15 months for all 9 models.
Figure 4Weighted mean correlation and computational time for all models.
Figure 5Correlation skills (left column) and RMSE scores (right column) of SST anomaly forecasts at lead times of 3–15 months for the Niño 4, Niño 3.4, Niño 3 and Niño 1+2 regions.
Figure 6Forecast performance (correlation) per pixel of the forecast reconstructed from 5 leading principal components at lead times of 3–15 months for the period 1979–2017. Top row: BNNE model, middle and bottom rows: improvement of performance of NNE and MLRE over BNNE.