Literature DB >> 32976990

Multi-step ahead meningitis case forecasting based on decomposition and multi-objective optimization methods.

Matheus Henrique Dal Molin Ribeiro1, Viviana Cocco Mariani2, Leandro Dos Santos Coelho3.   

Abstract

Epidemiological time series forecasting plays an important role in health public systems, due to its ability to allow managers to develop strategic planning to avoid possible epidemics. In this paper, a hybrid learning framework is developed to forecast multi-step-ahead (one, two, and three-month-ahead) meningitis cases in four states of Brazil. First, the proposed approach applies an ensemble empirical mode decomposition (EEMD) to decompose the data into intrinsic mode functions and residual components. Then, each component is used as the input of five different forecasting models, and, from there, forecasted results are obtained. Finally, all combinations of models and components are developed, and for each case, the forecasted results are weighted integrated (WI) to formulate a heterogeneous ensemble forecaster for the monthly meningitis cases. In the final stage, a multi-objective optimization (MOO) using the Non-Dominated Sorting Genetic Algorithm - version II is employed to find a set of candidates' weights, and then the Technique for Order of Preference by similarity to Ideal Solution (TOPSIS) is applied to choose the adequate set of weights. Next, the most adequate model is the one with the best generalization capacity out-of-sample in terms of performance criteria including mean absolute error (MAE), relative root mean squared error (RRMSE), and symmetric mean absolute percentage error (sMAPE). By using MOO, the intention is to enhance the performance of the forecasting models by improving simultaneously their accuracy and stability measures. To access the model's performance, comparisons based on metrics are conducted with: (i) EEMD, heterogeneous ensemble integrated by direct strategy, or simple sum; (ii) EEMD, homogeneous ensemble of components WI; (iii) models without signal decomposition. At this stage, MAE, RRMSE, and sMAPE criteria as well as Diebold-Mariano statistical test are adopted. In all twelve scenarios, the proposed framework was able to perform more accurate and stable forecasts, which showed, on 89.17% of the cases, that the errors of the proposed approach are statistically lower than other approaches. These results showed that combining EEMD, heterogeneous ensemble and WI with weights obtained by optimization can develop precise and stable forecasts. The modeling developed in this paper is promising and can be used by managers to support decision making.
Copyright © 2020 Elsevier Inc. All rights reserved.

Entities:  

Keywords:  Ensemble empirical mode decomposition; Ensemble learning models; Meningitis; Multi-objective optimization; Time series forecasting

Mesh:

Year:  2020        PMID: 32976990      PMCID: PMC7507988          DOI: 10.1016/j.jbi.2020.103575

Source DB:  PubMed          Journal:  J Biomed Inform        ISSN: 1532-0464            Impact factor:   6.317


Introduction

Meningitis is an inflammation that has several classifications with specific causes and symptoms, and, unfortunately, it is still a major public health problem because it causes irreversible health damages and has kept mortality rates high [1]. Besides, this disease could be a result of viral, bacterial, or fungal infection of the fluid surrounding the brain and spinal cord [2]. Early diagnosis and the immediate initiation of the treatment are fundamental for a good prognosis of the disease. Concerning the number of meningitis cases registered in Brazilian states located in the South or Southeast such as Parana (PR), São Paulo (SP), Minas Gerais (MG), and Rio de Janeiro (RJ) presented over ten thousand cases in 2018 [3]. Within this context, the development of efficient public policies and preventive campaigns are the key to change this scenario. The development of efficient predictive models plays an important role in epidemiological modeling. Predictive mathematics models are widely used to design the epidemiological scenario of specific diseases. These approaches are adopted for different purposes such as comparing, implementing, evaluating, prevention, therapy, and the development of public policies [4]. Usually, the epidemic models are based on parameters related to susceptibility (S), infected (I), and removed (R), as well as exposed (E) individuals can be considered which leads to the SIR or SEIR models. Each variation of these models has its particularity, and different factors can be considered in these approaches to provide knowledge about the disease spread. Nowadays, these approaches have been proposed to understand the spread of new coronavirus [5], [6]. Also, different mathematical approaches are proposed to mitigate the effects of several diseases such as ebola [7], influenza [8], and malaria [9]. In the last years, a computer science field called artificial intelligence (AI) able to recognize patterns of historical data and support the decision making has received attention to solving problems from the commerce [10], and industry [11]. Machine learning models, an AI sub-field, becomes the kernel of data analysis, once dealing with classification [12], data clustering [13], and regression tasks [14]. Nonetheless, when it comes to matters of diseases that plague the Brazilian public health system, such as dengue, malaria, and others, there are limited discussions as regards the effectiveness of machine learning models to develop predictive models. Some of these studies aimed to define the incidence of diseases such as ventriculitis and meningitis [15], as well as map the transmission risk of Zika [16], and make probabilistic forecasting of influenza [17]. Considering the purpose of knowing the number of future cases for any disease, once the datasets contain temporal information, time series forecasting should be used for this task. Therefore, time series forecasting aims to use past data to know future values with the purpose of, for example, making strategic planning to improve the knowledge in the domains that are inserted and help develop public policies. In this aspect, developing an efficient model is desirable; moreover, techniques such as the ensemble learning and decomposition may be used for this purpose. These strategies can be employed to deal with nonlinearity, nonstationarity, and cyclicity inherent to time series. Faced with this, the ensemble learning is an approach that is applicable in regression [18] and/or classification [19] tasks which have the objective of improving the model’s predictive accuracy. The main aspect of this approach lies in training several base (weak) models and combining their predictions to build an efficient model [20]. It is believed that this improvement occurs because each base model learns the different characteristics of the data and adds this information to the final results. Indeed, this methodology has proven effective in forecasting tasks in different fields of knowledge such as agribusiness [21], chemometrics [22], ecology [23], and medicine [24], [25]. Also, an additional approach usually adopted to improve the models’ performance is the ensemble empirical mode decomposition (EEMD) [26]. By employing EEMD, it is possible to separate the original signal into components (intrinsic mode functions — IMF and residual signals) with different amplitudes and frequencies added to the noise, aiming to extract relevant data information. The EEMD approach has been proven effective in then forecasting field of dealing with applications in several domains of knowledge, such as aquaculture [27], economy [28], and energy [29]. The EEMD steps consist of data decomposition and aggregating the components or aggregating the predictions of each component to obtain the original signal. In respect of this, each component is treated as an input set, and they can be trained separately using various algorithms. One strategy considers the same weight for all components (direct strategy — DI), which penalizes those that explain more data variability by attributing the same importance for all components, while another strategy employs different weights (weighted strategy – WI) for each component. Because the disease series show up nonlinear and nonstationary on most of the cases, the EEMD is efficient to analyze series related to meningitis cases. Considering the aforementioned, because of the use of the EEMD approach, two questions emerge. The first one lies in which algorithm should be employed to train and forecast each component. The second refers to which approach will be used to reconstruct the decomposed signal, that is, which strategy should be adopted to obtain an ensemble of components. To deal with these questions, this paper aims to propose a hybrid framework that was developed in three steps for multi-month-ahead forecasting (one, two and three-month-ahead) meningitis cases in the Brazilian states of MG, SP, PR, and RJ. As regards to this, in the first, the original signal is split into five components (four IMF and one residual). In the sequence, each component is used as the input of techniques: Bayesian Regularized Neural Networks (BRNN) [30], Cubist (CUBIST) [31], Gradient Boosting Machine (GBM) [32], Partial Least Squares (PLS) [33], and Quantile Random Forest (QRF) [34]. Following, predictions for all components are obtained by each model. Different combinations of models to predict the components are developed, and, for each combination, the results are weighted integrated to formulate a heterogeneous ensemble (HTE), considering that at least two different models are used in each combination. Besides, multi-objective optimization (MOO) using elitist Non-Dominated Sorting Genetic Algorithm — version II (NSGA-II) [35] is employed to find a Pareto Front (composed by a set of candidate’s weight), and Technique for Order of Preference by similarity to Ideal Solution (TOPSIS) [36] is applied to choose the adequate set of weights from Pareto Front. Finally, considering all the combinations developed, the most adequate model is the one with the best generalization of out-of-sample capacity in terms of mean absolute error (MAE), relative root mean squared error (RRMSE), and symmetric mean absolute percentage error (sMAPE). The contributions of this paper lie in four-folds, such as: The development of a unique hybrid framework based on data decomposing, ensemble, and MOO can improve simultaneously the accuracy and stability of forecasts of meningitis cases. Performing multi-step-ahead forecast, once that in papers related to diseases incidence forecasting, single forecasting horizon is used; An investigation of the effectiveness of the multi-objective procedure for choosing weights used to aggregate the forecasts of EEMD components. In this regard, this paper seeks to add discussions in this field, taking into account most of the debates regarding single–objective optimization for components aggregation, such as Du et al. [37] and Zhang and Wang [38]; Another contribution lies in the use of a diversified set of algorithms, concerning algorithms structure, to train and forecast each component after the EEMD performance, where these different algorithms can capture their inherent variability. The remainder of this paper is structured as follows: Section 2 presents some related works on the use of machine learning to forecasting disease cases. Section 3 describes the datasets adopted, while Section 4 details the methods employed in this paper for such an approach. Section 5 lists the steps for modeling. Section 6 presents the results and discussions. Finally, Section 7 concludes and presents the proposals of future research for the theme hereby adopted. Diseases related and its adopted modeling. Study area representation and datasets behavior.

Related works

This section presents a summary of some recent developments in time series forecasting for epidemic diseases using machine learning and general models. Considering the matter above mentioned, there were sixteen relevant papers, in the last years, found in the literature. With regard to diseases type, most of the papers aimed to perform comparisons either between machine learning models on the task of forecasting future cases or pandemic risk for dengue, malaria, and influenza. In these cases, one paper focuses on the diseases of dengue and malaria, simultaneously. This information is summarized in Table A.1 contained in Appendix A which presents a review of related papers with the use of machine learning and general models for forecasting disease cases.
Table A.1

Related works to machine learning and general models for the epidemiologic time series forecasting.

PublicationDiseaseModelsObjectiveFinding
Ch et al. [88 ]MalariaSVR-FFA, ANNs and ARMAPredicting the malaria incidences in India using SVR and FFA.Proposed approach outperform compared models.

Guo et al. [89 ]DengueSVR, GBM,LASSO and GAMUse the state-of-the-art machine learning algorithms to develop dengue forecasting in China.The SVR model had the consistently smallest prediction error rates for tracking the dynamics of dengue and forecasting the outbreaks.

Scavuzzo et al. [90 ]DengueSVR, ANNs, KNN and DTTemporal modeling of the oviposition activity (measured weekly on 50 ovitraps in a north Argentinean city) of Aedes aegypti (Linnaeus)These new tools perform better than linear approaches, in particular KNN performs better than compared approaches.

Chen et al. [91 ]Dengue, MalariaLASSOAccess how the LASSO method may be useful in providing useful forecasts for different pathogens in countries with different climates.Short-term predictions generally perform better than longer-term predictions, suggesting public health agencies may need the capacity to respond at short-notice to early warnings.

Chekol and Hagras [92 ]MalariaANFIS and SVRComparing ANFIS and SVR approaches on the task of predict the malaria epidemic, in Ethiopia, up to three-months-ahead.Similar results were observed for both approaches and they can be used in parallel.

Poirier et al. [93 ]InfluenzaRF, SVR and ENETCompared internet and electronic health records data and by statistical models to identify the best approach for influenza estimates in real time.For national and Brittany region influenza incidence rate, SVR model is the best approach.

Liang et al. [94 ]InfluenzaSVRExplore the application of the SVR model for forecasting influenza data in Liaoning, China.It is observed the feasibility of using internet search engine query data and the efficiency of SVR model in tracking the influenza.

Mollalo et al. [95 ]TuberculosisMLP and LRInvestigated the applicability of MLP to predict the tuberculosis incidence.Among the developed models, single hidden layer MLP had the best test accuracy.

Soliman et al. [17 ]InfluenzaDL, LASSO, MARS and ARIMAInvestigate utility of DL to forecast influenza in Dallas County, USA.DL and multi-model ensemble of forecasts yield a similar competitive performance.

Chen et al. [14 ]InfluenzaLASSO-GP, LR, ANNs, SVR and SARIMADeveloping a non-parametric model based on LASSO-GP regression for influenza prediction considering meteorological effect.Proposed approach outperform compared models.

Guo et al. [32 ]DengueEnsembleDevelop an ensemble penalized regression algorithm for initializing near-real time forecasts of the dengue epidemic trajectory.The proposed algorithm had the best performance in comparison with regression models using single penalties.

Shirmohammadi et al. [96 ]BrucelossisRF, SVR and MARSInvestigate and compared the performance of three data mining techniques to predict monthly brucellosis.Results indicated that the RF model outperformed the SVM and MARS and it can be utilized to diagnose the behavior of brucellosis over time.

Venna et al. [97 ]InfluenzaARIMA, EAKF, LSTMPropose a novel data-driven machine learning method using long short-term memory-based multi-stage for influenza forecasting.Proposed method performs better than the existing well-known influenza forecasting methods (ARIMA and EAKF) and the results offer a promising direction to improve influenza forecasting.

Thakur and Dharavath [98 ]MalariaANNsDetermine the malaria abundances using clinical and environmental variables with Big Data by ANNs.The results vary from area to area based on clinical variables and rainfall in the prediction model corresponding to areas.

Chakraborty et al. [99 ]DengueNNAR, ARIMA, LSTM and SVRProposed a hybrid ARIMA-NNAR for dengue forecasting on three regions.The ARIMA-NNAR give better forecasting accuracy in comparison to the state-of-the-art.

Su et al. [100 ]InfluenzaSARIMA, LASSO, LSTM and XGBoostDevelop a self-adaptive model by integrating the SARIMA and XGBoost approaches for real-time estimation of influenza in Chongqing, China.Compared with LASSO and LSTM approaches, the proposed framework reach better improvement than this approaches for the adopted task.
In the context of the employed techniques, the adopted approaches may be split into different classes, such as: (i) Artificial neural networks (ANNs): Multilayer perceptron (MLP), deep learning (DL), neural network autoregressive (NNAR) model, and long-short-term memory (LSTM); (ii) Regression trees: Decision trees (DT), random forests (RF), gradient boosting machine (GBM), and eXtreme gradient boosting (XGBoost); (iii) General regressions: Least absolute shrinkage and selection operator (LASSO), support vector regression (SVR), multivariate adaptive regression splines (MARS), elasticnet (ENET), and linear regression (LR); (iv) Remain approaches: Ensembled adjustment Kalman filter (EAKF), Gaussian process (GP), generalized additive model (GAM), autoregressive integrated moving average (ARIMA), and seasonal autoregressive integrated moving average (SARIMA). Fig. 1 associates the diseases and adopted modeling.
Fig. 1

Diseases related and its adopted modeling.

Adjacent to the above mentioned, as well the presented in Appendix A, some gaps in relation to the developed approaches can be found and are stated as follows: Considering the disease’s type, around of 93.75% of the papers focused on malaria, dengue or influenza. Hence, there is a lack of discussion concerning the predictive capacity of machine learning-based approaches for diseases such as measles, meningitis, and chikungunya on the forecast task; In the modeling aspect, only four papers focused on ensemble approaches such as bagging and boosting or models combined by average. One paper used an optimization approach of the swarm intelligence field called firefly algorithm (FFA) for hyperparameters tuning, and no paper adopted signal decomposition or MOO with the purpose of building ensembles. It is well known that the combination of these strategies can help on the improvement of the model’s accuracy and, therefore, out-of-sample generalization; In general, to the forecast horizon, most of the papers focused on a single horizon, that is, one week or one-month-ahead. The discussions are narrow for the multi-step-ahead forecast. Indeed, the use of other horizons could be important because they can be useful for managers to develop strategic plains with the goal of preventing the population from having these diseases, as well as developing strategies and public policies so that these diseases will affect as few people as possible. Therefore, faced with the aforementioned aspects, this study seeks to fill the pointed-out gaps, by proposing a hybrid framework to forecast the number of meningitis cases, which considers EEMD, heterogeneous ensemble, and MOO approaches to reconstruct the decomposed signal.

Material

The datasets adopted in this paper refer to the monthly number of confirmed cases of meningitis, recorded on a disease information system at the Brazilian Ministry of Health. The adopted period of analysis ranges from January of 2007 to December of 2018, and the data is available at the Department of Informatics of the Unified Health System (Departamento de Informática do Sistema Único de Saúde, in Portuguese, DATASUS) [3]. The data ranged from January of 2007 to December of 2017 is used in the training process and the remaining information is used to test the proposed framework’s performance, respectively. In this context, the adopted data refers to information on the number of confirmed cases of meningitis in the Brazilian states PR, SP, MG, and RJ taking into account that these states presented the four largest notifications number in 2018. In this paper, to verify the generalization ability of the proposed methodology, the above-described datasets are used. We did not fit the proposed model on the time series of a single region, and then validate the model by evaluating the prediction of new cases of another region because each state has different features, such as demography, geography, population density, economic and human development index, as well as public health policies. Therefore, to avoid erroneous conclusions, the proposed model is trained (training set) and validated (test set, out-of-sample forecasting) in the same Brazilian state, but in different splitting setups of the datasets. In this way, we tried to accommodate the data variability for each state. Fig. 2 shows the study areas, the behavior of the number of notified cases by state and its descriptive measures, as well as the autocorrelation function (ACF) for the MG(A), PR(B), RJ(C) and SP(D) series. Secondly, as highlighted by Fig. 2, for the state of PR, there is a greater number of notifications than presented for other states, as reported by the descriptive measures (number of observations — n, minimum — Min, average, maximum — Max, and standard deviation — SD). The Augmented Dickey–Fuller (DF) test shows that the four time series are non-stationary (DF = −5.35 - −3.32, p-value 0.05). Aiming at evaluating the presence of seasonality within the data, the Kruskal–Wallis test is performed. In this case, for the MG, SP, and RJ time series, there is no evidence of seasonality ( = 13.32 - 15.50, p-value 0.05), while the series related to the state of PR present evidence of seasonality ( = 33.07, p-value 0.05) [39]. Additionally, the autocorrelation measures suggest, due to lags, that up to the first four observations are correlated and can be used as inputs for the data modeling.
Fig. 2

Study area representation and datasets behavior.

Methods

The objective of this section is to present each technique employed in this paper.

Ensemble Empirical Mode Decomposition

The EEMD technique was proposed by Wu and Huang [26]. It is an extension of the Empirical Mode Decomposition (EMD) algorithm and should be used with the purpose of data de-noising. The EMD approach allows the analyzing of nonlinear and non-stationary signals. The steps followed in this technique are data decomposition, and later data aggregation. First, in the decomposition stage, the EMD separates the original signal into coexisting oscillatory components (IMF and residual signal) [40]. Next, after the modeling of each component, it is necessary to perform the aggregation of results to obtain the original signal. When the EMD is used, the main drawback is named “mode mixing”, that is, each single IMF consists of signals with dramatically disparate scales or a signal of the same scale appears in different IMF components [26]. Seeking to solve this problem, the noise assisted EEMD was proposed. In this approach, EMD is performed k times, and different white noise (a random signal that follows a normal distribution with zero mean and constant variance) is added to the data in each trial. Two disadvantages can be stated for EEMD, as follows: (i) extra noise exists in the reconstructed signal and (ii) it needs more computational resources than EMD [41]. Initially, the EMD approach is presented and generalized for EEMD. In fact, the main steps are stated as follows: Input the original time series; Add white noise on the original time series; Obtain the local maximum and minimum values of time series of step 2.; Generate envelops for the new time series, taking into account its boundaries; Calculate the average of the boundaries of step 4.; Extract from the generated data the average value obtained in step 5.; After performing each step 2.–6., each component is named IMF [42] if it has the following characteristics: (i) in the entire dataset, the number of extreme and zero crossings must either be equal or different, at most, by one, and (ii) at any point, the mean value of the envelopes is defined by the local maximal and the local minimal which must be zero; Defining that the component of the step 7. is an IMF, the residual component is computed as the difference between the IMF and the remaining data. The steps 2.–7. are performed until that residual component becomes a monotonic function or has one local extreme point from which no more IMF can be extracted. The EEMD performed the steps 2.–8. k times and for each trial there are n components (IM, …, IM, Residual). Therefore, each final component is treated as the average of the respective components of the k executions; the number of repeated procedures is called the ensemble number. After the end of the decomposition process, the original time series can be expressed as the sum of IMF and residual components. In this stage, DI or WI strategies should be used. The DI consists of assuming the same weight for each component, while WI considers different weights for each component. When a decomposition technique is adopted, there are two possibilities for data analysis, as follows: (i) datasets are decomposed and components are aggregated and used as input in some forecasting method, see the example developed by Jiang and Liu [43], and (ii) each component obtained during the decomposition stage is used as an input by some forecast method, and the forecasting of each component is used in the aggregation stage, see the example developed by Wang et al. [44] or Wu et al. [45]. In this paper, the second possibility and WI are used for data modeling. The parameters used for the decomposition task are presented in Section 5, item 1..

Bayesian Regularized Neural Network

The BRNN is a type of feedforward neural network, composed by one input and one hidden layer, which uses the Bayesian methods, such as Empirical Bayes, for parameter estimation, with the purpose of avoiding overfitting [46]. The mathematical modeling can be stated as follows: in which is ith output value , s is the number of neurons, is an input value, , is the vector of weights and biases, is the weight of kth neuron , is the bias of kth neuron, is the weight of jth input to the network , and is the activation function. In this paper, the activation function is the hyperbolic tangent. According to the procedure proposed by MacKay [46], the elements of are estimated by using an empirical Bayesian approach, according to two steps described as follows, Minimization of in which, F(.) is a function of , and are the observed and estimated values, respectively. Also, and are the variance of errors, weights and biases, respectively; Updating the variance components and by maximizing a conditional probability () given the observed values according to the variances, i.e., . When the early expression is not present in the closed form, an approximation for the marginal log-likelihood is adopted. In Eq. 2., the variances are regularization parameters, where the trade-off between goodness-of-fit and smoothing can be controlled. In this approach, only the neurons number should be determined.

CUBIST

The CUBIST is a rule-based model, which performs predictions following the regression of trees principle [47]. Giving a regression tree, for each leaf is constructed a linear regression model, or in this case, a rule, associated with the information contained in each leaf. Once all rules are constructed, the final predictions are based on the linear combination of developed rules. The main difference between a simple regression tree and a CUBIST model is how the models make predictions within the nodes [31]. In order, to attempt an improvement in the model’s accuracy, this rule-based model generates a set of rules, named a committee, as happens with the boosting approach. This is developed to correct the predictions of the previous members. Also, given a set of new features, the CUBIST can adjust the model prediction using samples from the training set by employing the neighborhood concept, like the -nearest-neighbors approach [48]. Aiming at finding the number of neighbors, a distance measure, in this case, Manhattan is applied. Finally, once the set of committees and number of neighbors are defined, the final rule prediction is the simple average of committees predictions.

Generalized boosted regression

The generalized boosted regression modeling or gradient boosting machine (GBM) model is based on the boosting principle, proposed by Friedman [49], which seeks to find an additive model that minimizes the loss function. By using a gradient descent approach, the GBM builds models in the negative sense of the partial derivative of the loss function regarding the set of predictors. The GBM performs iteratively so that a regression tree model is fitted to the data, and the residuals from the fitted model are obtained. A new model is adjusted for previous residuals, a new prediction is obtained, to which the initial forecast is added, and then a new residual measure is obtained. This process is performed iteratively until a convergence criterion is reached, and the final prediction is obtained by the average in regression problems. Adjacent to the improvement that can be reached by the GBM structure, this process tends to generate overfitting. To avoid this problem, weights are assigned for observations whose errors are greater. In this paper, for the adopted GBM approach, control hyperparameters such as number of boosting iterations, maximum tree depth, shrinkage and minimum terminal node size are used. The best combination of these parameters should be found; this is a challenge since an inappropriate choice can result in a computationally costly and inefficient process.

Quantile random forests

The quantile random forests (QRF) [50] approach is an extension of the RF ensemble learning model [51]. It provides information about the full conditional distribution of the response variable, and not only about the conditional mean. In this approach, the use of conditional quantile is to enhance the RF performance, which makes this a consistent approach [50]. The main assumption about QRF lies on the weighted observations that can be used for estimating the conditional average [52]. Additionally, while the RF approach retains the result’s information about the average of notifications number of the leaves, the QRF keeps all notifications contained in the leaves to estimate, where in which the left side of the equation represents the conditional probability of the number of notifications conditioned to the predictors, being the notifications in a horizon and the predictors’ number, and (t = ). The right side refers to the conditional distribution function (CDF) of the average notification numbers regarding the predictors. Considering that QRF uses the quantiles in the prediction process, the -quantile of CDF is stated as the probability that the number of notifications is lower than if the given is equal to , where the estimate of is stated as follows: in which is the -quantile estimated relation to the predictor  [50]. For the QRF architecture adopted in this paper, only the number of randomly selected predictors should be determined.

Partial least squares

The PLS regression approach is a technique to analyze multivariate data, in which the aim is to relate one or two output variables () with several inputs (). For this purpose, given a linear model, the problem that often arises is the matrix of inputs being singular. Faced with this, to deal with this problem, the PLS decomposes into orthogonal scores and loadings , in which . Regarding this approach, and values are chosen in such a way that the covariance between inputs and outputs is maximized. In other words, the PLS finds components that maximize the variation of the predictors while simultaneously requiring these components to have a maximum correlation with the response. In a more general way, the PLS finds linear combinations of the predictors, named components that are like the principal component analysis. Additionally, the PLS in its classical form is based on the nonlinear iterative partial least squares [53], [54], and the number of components should be defined.

Multi-objective optimization

On some problems, it is necessary to minimize (or maximize) multiple objectives to achieve a preferable solution. Naturally, MOO should be adopted [55], and it is performed in three steps. First, it is necessary to define the multi-objective problem (MOP); second, some algorithms are used to optimize the objectives; third, it is essential to choose the most appropriate result for the formulated problem, appointed as multicriteria decision making (MCDM) [56]. In time series analysis, the multi-objective approach is used to develop a model that makes predictions with small errors which should vary slightly, in other words, it is adopted to deal with bias–variance trade-off [57] where forecast errors and errors variation are minimized. According to Marler and Arora [55], in the first step, the MOP is defined, regarding which decision variables, constraints and objectives are stated. The MOP for two objectives can be stated as follows: subject to inequality constraints in which , () is a vector of decision variables, L and U are the lower and upper boundaries for each decision variable, and is the kth objective to be minimized or maximized. In this respect, during the MOO step, an optimization algorithm is applied to find the Pareto Front approximation (PF). This set is composed of non-dominated solutions for which there is no other permissible solution that simultaneously improves all the objective functions without sacrificing at least one of the other objective function. Each set of decision variables associated with each element of PF makes up the Pareto Set (PS) [55]. As regards to the MOO approach, most of the algorithms proposed in the literature are based on evolutionary computation and are named multi-objective evolutionary algorithms (MOEA). An evolutionary algorithm is based on nature’s laws, and it uses mechanisms such as crossover (an operation used to combine the genetic information of parents to produce a new offspring), mutation (an operator used to maintain the diversity of the population by introducing another level of randomness) and selection (an operation applied to select the fittest individuals according to the objectives to be optimized) [58]. Basically, in an MOEA, the fittest members of a parent in a population will survive and propagate their fitness along with the evolution of its populations of offspring. In the literature, new approaches to deal with MOO are constantly being proposed. Nevertheless, in this paper NSGA-II is adopted [35], considering it is a classical MOO approach and has already shown good results (in terms of accuracy, in general for regression analysis, lower error) to deal with MOP on several fields of knowledge [59], [60]. The NSGA-II is a derivative-free method which employs a non-sorting ordering operator and crowding distance criteria in the optimization process. The first operation is used to find a set of equally good solutions closest to the Pareto-optimal front, while the second is employed to promote diversity among members of the solution population. The main idea behind the crowding distance is finding the Euclidean distance between neighboring individuals in the m-dimension space [35], [61]. The NSGA-II parameters adopted in this paper are exposed in Section 5, item 4b. Lastly, in the MCDM step, it is possible to find a preferable set of decision variables (weights in this paper) that allows dealing with the trade-off between the objectives. It means finding the best compromise between the optimized objectives. An approach used for this purpose is TOPSIS, which was proposed by Hwang and Yoon [36]. The TOPSIS was meant to determine the best alternative based on the concepts of the compromised solution. In other words, the chosen solution should be as close to the positive ideal solution as possible, and as far away from the negative ideal solution as possible, considering a measure of similarity. The distance of ideal and negative ideal solutions can be obtained by the Euclidean distance [62], [63]. The TOPSIS weights used for each objective are defined in Section 5, item 4c. Flowchart of proposed approach.

Methodology

Performing EEMD for the datasets and obtaining four IMF and one residue component. The parameters, number of ensemble components, number of components (IMF and residual components) and amplitude of white noise are defined as 100, 5 and 6.4 × 10−7, respectively. These values are obtained according to Wu and Huang [26]. This paper does not focus on the discussion of the model’s performance according to the parameter setting; For each component, the ACF analysis showed, for most cases, that up to four lags are suitable to be used as predictors. Without loss of generality, for all components and states studied, this configuration is used as inputs of each adopted model (BRNN, CUBIST, GBM, PLS, and QRF); Training each IMF and residual components using the models mentioned in step 2. by leave-one-out cross-validation with a time slice window (LOOCV-TS), according to and forecast the meningitis cases according to the recursive method, as given by Eq. (8), in which is a function related to the adopted model in the training process, is the forecast value for kth component obtained in the decomposition stage (k = 1,…,5) on time and forecast horizon (), are the previously notified cases lagged in and is the random error which follows a normal distribution () with zero mean and constant variance. The recursive method, also known as an iterated method, can lead to poor accuracy in long forecasting horizons. According to Pouzols and Barros [64], and Veloz et al. [65], the recursive strategy uses forecasting values as a model’s input to forecasting the next predicted values. Its main disadvantage is to accumulate the previous forecasting errors in the recursive process. However, the advantage of the recursive method lies in the use of one model for all processes, i.e., train one model to forecast a one-step-ahead horizon, and use it for multi-step-ahead forecasting task. On the other hand, the direct method uses only past values to predict the future, which is its advantage, as it does not accumulate prediction errors. However, its disadvantage lies in the necessity to fit or training a new model for each forecasting horizon. Therefore, this disadvantage makes the process complex and computationally intensive. Due to separated models are used to forecasting two points, there is no opportunity of handling the dependency between two consecutive predictions. According to Ma and Fildes [66], comparative studies between Recursive and Direct strategies have shown contradictory results as to which strategy achieves better forecasting performances than others. While Rana and Rahman [67] achieved better results for direct method over the recursive method, Xue et al. [68] provided evidence of better results from the recursive method over the direct method. There is a trade-off between computational cost and accuracy in the choice of a multi-step-ahead forecasting method. Machine learning models, sometimes, have high training time, either due to the use of different training strategies such as cross-validation (k-fold or LOOCV-TS), or even due to the number of parameters to be tuned. In this context, because in this paper several models are evaluated, to find an efficient ensemble learning forecasting model to study meningitis cases, the recursive forecasting strategy is adopted. Moreover, the forecasting horizons are defined as one, two, and three months-ahead, which are considered short-term. Therefore, even though the recursive method could lead to high forecasting errors, it is used in this paper due to its lower computational cost. However, the direct method can be considered for this study. In this paper, the hyperparameters of the adopted models in step 2. are obtained during the LOOCV-TS by a grid-search (GS) and each situation is presented in Table B.1 of Appendix B. Giving a set of combinations of the hyperparameters values, GS consists of searching exhaustively throughout the combinations which is the best set of hyperparameters that minimize/maximize any criteria; in this paper, root mean squared error is considered. The LOOCV-TS consists of using an initial slice of training data as an estimation set, and then it validates the performance of the trained model over a desired number of horizons. The next step consists of using the data for validation while incorporating it into the training set, and the process is performed until the entire training set is used. The model performance is obtained as the average criteria used in each iteration [69].
Table B.1

Control Hyperparameters obtained by GS during cross-validation process for each adopted model.

StateComponentBRNNCUBIST
GBM
QRFPLS
NeuronsCommittesInstancesBoosting IterationsMax Tree DepthRandomly Selected PredictorsComponents
IMF1(2,3,4)(1,1,1)(5,5,5)(50,150,50)(2,2,2)(4,4,4)(2,2,2)
IMF2(1,1,1)(1,1 ,1)(1,0,0)(150,150,150)(2,3,2)(3,3,3)(3,3,2)
MGIMF3(3,3,3)(20,20,20)(9,9,9)(50, 150,50)(2,1,2)(3,3,3)(3,3,3)
IMF4(3,2,3)(10,10,10)(1,0,0)(150,150,150)(3,2,3)(4,4,4)(3,3,3)
Residual(2,1,2)(10,10,10)(1,0,0)(150,50,150)(3,2,3)(4,3,4)(3,3,2)
Non-decomposed(1,1,1)(10,10,10)(9,9,9)(50, 50,50)(1,3,2)(4,3,2)(2,2,2)

IMF1(1,4,4)(20,20,10)(9,9,9)(150,50,50)(3,3,3)(4,4,4)(2,2,2)
IMF2(1,4,4)(1,1,1)(0,0,0)(150,150,150)(2,2,2)(4,4,4)(3,3,3)
SPIMF3(1,4,4)(20,20,20)(0,0,0)(100,50,50)(1,1,1)(4,4,4)(3,3,3)
IMF4(1,4,4)(1,1,1)(0,0,0)(150,150,150)(2,3,3)(4,4,4)(3,3,3)
Residual(1,2,2)(20,20,20)(5,5,5)(150,150,150)(3,3,3)(2,2,2)(3,3,3)
Non-decomposed(3,4,4)(10,10,10)(5,5,5)(150,100,100)(3,2,3)(4,4,4)(3,3,3)

IMF1(2,3,4)(10,10,10)(0,0,0)(150,50,50)(2,1,2)(4,3,4)(1,1,1)
IMF2(1,4,4)(1,1,1)(0,0,0)(150,150,150)(3,2,3)(4,4,4)(3,3,3)
PRIMF3(1,2,3)(1,1,1)(5,5,5)(100,100,100)(2,2,2)(3,2,3)(3,3,3)
IMF4(2,2,2)(20,20,20)(0,0,0)(150,150,150)(3,2,3)(2,2,2)(3,3,3)
Residual(1,4,4)(1,1,1)(9,9,9)(150,150,150)(3,3,3)(4,4,4)(3,3,3)
Non-decomposed(1,2,3)(1,5,5)(5,5,5)(50,50,50)(2,2,3)(4,2,3)(3,3,3)

IMF1(3,3,4)(20,20,10)(9,9,9)(50,100,50)(1,1,2)(3,3,4)(2,2,1)
IMF2(3,4,4)(10,10,1)(0,0,0)(150,150,150)(2,3,3)(3,4,4)(2,2,3)
RJIMF3(3,4,3)(1,1,1)(5,5,5)(150,150,100)(3,2,2)(3,4,3)(3,3,3)
IMF4(4,3,2)(10,10,20)(0,0,0)(150,150,150)(3,2,3)(4,3,2)(3,3,3)
Residual(4,4,4)(20,20,1)(5,5,9)(100,150,150)(2,1,3)(4,4,4)(3,3,3)
Non-decomposed(2,2,3)(10,10,1)(0,0,5)(50,50,50)(1,1,3)(2,2,2)(1,1,3)
To aggregate the EEMD’s components, the forecasting obtained in step 3. are used. To choose the order of the models used to train and predict the results of each component, a set of 3125 models is evaluated. These models are obtained by permuting the five forecasting models QRF, PLS, BRNN, GBM, and CUBIST (in this order) for the five EEMD components. Therefore, the methodology used to select the models’ order is the grid-search. Table 1 shows a sample of 3 out of 3125 ensemble learning models, randomly selected, where the order of models for each component is detailed.
Table 1

Randomly selected ensembles learning models and their order with respect to the EEMD components.

Grid-search indexIMF1IMF2IMF3IMF4Residual
295CUBISTGBMPLSBRNNQRF
1831QRFPLSGBMCUBISTBRNN
2639GBMBRNNQRFPLSCUBIST
After defining the weights for each model-component, suppose that the ensemble learning model number 295 has better accuracy for out-of-sample forecasting, then the order of the models used for each component is the presented in Table 1. For each combination, the optimization process is conducted according to what is described as follows: In the MOP, the cost function, for each combination of models and components, is stated as follows: in which is the predicted value on time and h horizon, and are the predictions of each component and is the weight vector to be estimated, in which  [70]. Considering the bias–variance framework [57], the objectives are defined as follows: in which is the ith predicted value through the aggregation of components and is the ith observed value. The error is computed through the mean squared error criterion, and the stability by the variance of the forecasting errors. In the sequence, the NSGA-II is applied (one run) and the PF approximation is obtained. In this paper, the parameters used for this algorithm, population size (number of candidate solutions), maximum number of generations (stopping criterion), crossover and mutation probability, crossover and mutation distribution, are defined as 100, 100, 0.9, 0.1, 0.7 and 0.2, respectively. The crossover rate is set to 0.9 because it allows new structures to be introduced into the population at a faster rate, whereas the mutation rate is set to 0.1 because it prevents a given position from becoming stationary in a set of values for the parameters to be optimized [35]; To find the best set of weights, between all candidates, the TOPSIS approach is employed, in which the weights for bias (error) and variance objectives are 50% and 50%, respectively; Forecasting the meningitis cases, out-of-sample, according to where is the estimated weight; Computing the performance measures MAE, RRMSE and sMAPE given by, where represents the number of observations, and are the ith observed and predicted values, respectively. The best model is the one with the best capacity of out-of-sample generalization in terms of MAE, RRMSE, and sMAPE. Aiming to compare the forecast errors of two models, the Diebold–Mariano (DM) test [71] is applied. In this paper, the lower tail priori hypothesis is given by Eq. (16), and statistic of DM test is given by Eq. (17), in which L is a loss function that can estimate the accuracy of each model, is the error of the proposed model, is the error of the compared model, and is an estimate for the variance of . By using the hypothesis defined for the DM test, the interest lies in knowing if the errors for the proposed model are lower than the compared model. If the null hypothesis is rejected, it is possible to say that there is statistically evidence that there is a reduction in the errors of the EEMD-HTE-MOO model regarding the compared model at the level of significance. In addition, Fig. 3 presents the modeling process.
Fig. 3

Flowchart of proposed approach.

Randomly selected ensembles learning models and their order with respect to the EEMD components. The results presented in Section 6 are generated using the processor Intel(R) Core(TM) i5-4200U central processing unit of 1.6 Hz in Windows 10 operating system with 8GB of random access memory. The R software [72] is adopted to perform the modeling. The packages hht [73], caret [74], mco [75] and MCDM [76] are used on the steps 1., 3., 4b and 4c, respectively. Moreover, the codes adopted in the proposed methodology are available at https://github.com/MRibeiro2107/-JBI-19-1008.

Results and discussions

This section presents the results and discussions of the information obtained from the developed experiments. Table B.1 contained in Appendix B presented the control hyperparameters of the adopted models for the task of training and predicting each component of the decomposition stage, as well as the data trained without EEMD decomposition. In the context of the structure of the hybrid, the framework proposed Table C.1 contained in Appendix C shows the set of weights, and the respective models used to train and predict each IMF and residue. Considering all the forecast horizons, for the developed models, the GBM model is the most used to train the components, while QRF is the least employed for this task. In both cases, GBM and QRF are ensemble learning approaches; however, the first model is based on boosting, while the second one on bagging. The objective of Sections 6.1, 6.2, 6.3 is to present the main results obtained with the proposed approach against various classes of models. Faced with this, three different comparisons were developed and performance on the out-of-sample forecast (for test set) is presented. Through these comparisons, the prediction accuracy of the proposed framework is effectively evaluated. In Tables 2 to 4, the best results are presented in bold.
Table C.1

Weights obtained by MOO approach, which are assigned for each model adopted in the signal reconstruction and models used in the structure of proposed hybrid framework.

Forecast HorizonWeightMG
SP
M1M2M3M4M5Proposed (Models)M1M2M3M4M5Proposed (Model)
θ11.15601.30871.43701.13191.10221.4536 (GBM)1.10031.14981.24331.02041.04511.1457 (CUBIST)
θ20.93800.89530.95070.86861.03140.9076 (CUBIST)1.00840.93310.89771.00611.08780.9404 (CUBIST)
One-monthθ30.77980.75560.80980.77230.87720.8034 (BRNN)0.90841.02871.07120.95961.05241.0561 (CUBIST)
θ41.05791.06221.08971.04851.10011.0877 (QRF)1.04281.22181.10901.03141.32021.2278 (GBM)
θ50.99540.99680.95440.99501.03730.9917 (GBM)0.87220.97450.88740.98330.87740.9804 (GBM)

θ11.22231.29381.25981.20651.12021.3093 (GBM)0.98721.13021.18641.15501.04491.1903 (CUBIST)
θ20.97500.90450.94481.01741.03651.0226 (BRNN)0.91210.85080.90040.99991.04860.9527 (PLS)
Two-monthsθ30.75390.75440.80830.82090.88630.7738 (BRNN)0.93621.05051.15120.97561.04401.1667 (BRNN)
θ41.01661.06301.10711.02891.08081.1105 (CUBIST)1.09421.22051.08001.02621.33201.2825 (GBM)
θ50.98810.98650.95410.98350.99090.9891 (BRNN)0.98120.96910.97800.89140.98290.9999 (GBM)

θ11.15001.37041.31301.02711.11251.3541 (GBM)0.87561.09921.09451.41390.98961.1008 (CUBIST)
θ20.91110.89500.97290.64650.96660.9625 (GBM)0.99510.89800.77340.63231.14300.9084 (CUBIST)
Three-monthsθ30.79990.82140.77250.67580.80980.7603 (BRNN)0.84410.92911.22800.95441.05580.9713 (BRNN)
θ41.08351.10001.03231.03611.08891.0578 (QRF)1.01061.12201.04681.03581.24801.0956 (GBM)
θ51.00241.00200.99561.00111.09661.0179 (QRF)0.91220.96810.87430.90100.90401.0354 (GBM)

Forecast HorizonWeightPR
RJ
M1M2M3M4M5Proposed (Models)M1M2M3M4M5Proposed (Model)

θ10.89061.99951.13490.74801.08601.0751 (QRF)1.01371.11761.46461.03151.12161.4744 (GBM)
θ20.80280.62840.84720.76221.00301.0158 (QRF)0.93310.94550.94980.91491.05660.9518 (GBM)
One-monthθ30.92980.93330.82440.95031.03201.0065 (CUBIST)1.03280.97991.04891.02151.02721.0502 (PLS)
θ40.81200.77440.89830.78630.97590.9686 (PLS)1.08291.09861.06131.06481.03671.0807 (GBM)
θ50.97410.93970.98220.96881.00230.9972 (BRNN)0.99580.99750.97060.99700.99700.9983 (CUBIST)

θ10.81851.99880.62240.62411.06290.9358 (QRF)0.89161.20571.59270.81981.17030.9121 (BRNN)
θ20.75540.56250.68770.74070.93140.8234 (CUBIST)0.91980.94250.94890.85071.08860.9051 (GBM)
Two-monthsθ30.96520.85430.76560.81751.07280.9067 (BRNN)0.96340.95501.00310.99671.04780.9746 (PLS)
θ40.76830.60670.56620.54880.98811.1394 (GBM)1.09531.03931.05541.00761.05691.1191 (CUBIST)
θ50.96940.76240.75650.73070.97890.9959 (GBM)0.99711.00160.99290.99621.00180.9959 (BRNN)

θ10.83431.99980.87950.55391.07130.8521 (QRF)1.06790.97241.60171.20631.11901.0491 (BRNN)
θ20.80850.64580.70490.61580.96020.8281 (CUBIST)0.96060.93740.91820.72451.11780.7348 (PLS)
Three-monthsθ30.83050.99810.69170.83211.09910.8287 (BRNN)1.03100.99630.90850.98151.00730.9590 (PLS)
θ40.65000.79180.53450.52670.92921.2443 (BRNN)1.09941.10061.11811.01591.02741.0157 (CUBIST)
θ50.77220.94200.71220.69800.97020.9997 (GBM)0.99620.89940.96280.99920.96891.001 (QRF)
Table 2

Performance measures of proposed and decomposed homogeneous optimized ensemble learning models.

StateModelForecasting Horizon
One-month-ahead
Two-months-ahead
Three-months-ahead
MAEsMAPERRMSEMAEsMAPERRMSEMAEsMAPERRMSE
EEMD-MOO-BRNN1011.34%14.04%12.5814.38%15.79%11.8313.62%14.75%
EEMD-MOO-CUBIST9.0810.39%12.93%8.7510.14%13.13%10.7512.37%14.30%
MGEEMD-MOO-GBM910.48%13.36%9.6711.03%13.84%7.838.77%11.75%
EEMD-MOO-PLS10.5011.86%14.81%13.2515.14%16.94%12.3314.17%15.68%
EEMD-MOO-QRF9.1710.16%13.96%12.5014.37%17.31%11.6713.20%16.04%
Proposed7.338.35%10.83%7.929.02%10.63%6.677.37%10.38%

EEMD-MOO-BRNN151.4226.86%28.01%104.9216.86%21.55%137.3322.76%27.24%
EEMD-MOO-CUBIST92.5014.62%18.89%103.8316.44%21.64%115.8318.44%23.78%
SPEEMD-MOO-GBM146.5023.52%30.02%127.9219.60%26.74%161.7526.12%34.24%
EEMD-MOO-PLS106.5017.02%21.72%157.5027.18%28.56%172.5828.96%33.65%
EEMD-MOO-QRF16127.62%30.89%127.6720.43%24.67%182.5831.73%32.35%
Proposed67.589.90%15.14%80.6712.02%17.04%87.0812.77%17.21%

EEMD-MOO-BRNN15.4210.95%14.24%16.3311.41%15.06%33.8328.72%28.43%
EEMD-MOO-CUBIST15.8311.19%15.18%27.1721.45%25.55%15.2510.86%14.67%
PREEMD-MOO-GBM17.7512.41%17.52%38.8333.13%34.74%4743.93%39.73%
EEMD-MOO-PLS15.8310.77%15.14%35.5030.08%31.50%40.1735.06%33.82%
EEMD-MOO-QRF11.588.03%12.84%15.3310.37%19.80%17.5011.52%17.42%
Proposed11.507.57%12.10%12.508.61%11.95%11.757.70%13.40%

EEMD-MOO-BRNN6.337.47%8.57%7.087.99%9.12%4.835.46%7.21%
EEMD-MOO-CUBIST7.338.71%9.71%8.259.36%10.58%10.7512.64%13.91%
RJEEMD-MOO-GBM6.257.13%8.11%8.589.73%11.06%6.927.89%8.43%
EEMD-MOO-PLS7.338.29%9.32%7.428.32%9.19%66.76%7.99%
EEMD-MOO-QRF6.758.13%10.64%8.429.87%12.49%89.20%10.21%
Proposed4.424.98%5.89%6.587.37%8.52%4.084.52%6.49%
Table 4

Performance measures for proposed and decomposed heterogeneous direct integrated ensemble learning models.

StateModelForecasting Horizon
One-month-ahead
Two-months-ahead
Three-months-ahead
MAEsMAPERRMSEMAEsMAPERRMSEMAEsMAPERRMSE
MGProposed7.338.35%10.83%7.929.02%10.63%6.677.37%10.38%
EEMD-HTE-DI8.929.93%12.51%8.7510.14%13.13%8.589.43%12.51%

SPProposed67.589.90%15.14%80.6712.02%17.04%87.0812.77%17.21%
EEMD-HTE-DI101.9215.31%20.18%102.2515.49%20.52%132.8320.57%25.87%

PRProposed11.507.57%12.10%12.508.61%11.95%11.757.70%13.40%
EEMD-HTE-DI12.678.73%13.43%1610.81%19.85%17.7511.68%17.48%

RJProposed4.424.98%5.89%6.587.37%8.52%4.084.52%6.49%
EEMD-HTE-DI5.926.72%8.12%7.838.68%10.51%5.836.65%8.05%
Performance measures of proposed and decomposed homogeneous optimized ensemble learning models. Performance measures for proposed and non-decomposed models. Performance measures for proposed and decomposed heterogeneous direct integrated ensemble learning models. Average standard deviation of errors obtained by each model in forecast out-of-sample (test set forecast).

Comparisons I: Evaluation of proposed ensemble learning and decomposed homogeneous optimized ensemble learning models

The comparisons I is designed to verify the forecast performance of the proposed hybrid framework by comparing it with five models based on decomposition, which uses the same model for all components and MOO in the aggregation step, namely EEMD-BRNN-MOO, EEMD-CUBIST-MOO, EEMD-GBM-MOO, EEMD-QRF-MOO, and EEMD-PLS-MOO. The comparisons are shown in Table 2 and additional discussion are also presented. These results shows that the approach considered as a heterogeneous ensemble of components is more accurate than other approaches that consider the same model for all IMF and residual, to forecast the number of meningitis cases one, two and three-month-ahead of time. According to Zhang et al. [77], this effectiveness is associated with the diversity used by the heterogeneous ensemble approach, which is an efficient and simple way to perfect forecasting accuracy and stability (lower standard deviation of the errors). Faced with this, it makes the predictive model more robust. Concerning the one-month-ahead forecast, for all states, the proposed framework achieves better accuracy than the compared models. Regarding MAE criterion, the compared models increasing the forecasting errors regarding proposed methodology, which ranges between 22.73% and 36.36%, 36.87% and 138.22%, 0.72%, and 54.35%, as well as 41.51% and 66.04% for the Brazilian states MG, SP, PR, and RJ, respectively. Regarding the sMAPE the compared models increasing the forecasting errors regarding proposed ensemble learning model, which is ranged between 1.81%-3.51, 4.72%–17.72%, 0.46%–5.96% and 2.16%–3.32% for the states of MG, SP, PR, and RJ, respectively. Considering the RRMSE, the compared models increasing the forecasting errors regarding proposed ensemble learning model, which is ranged between 2.09%–3.98%, 3.76%–15.75%, 0.74%–6.58% and 2.22%–4.75% for the states of MG, SP, PR, and RJ, respectively. The same behavior is reproduced when the forecasting horizon is two and three-month-ahead of time and in some situations with a greater magnitude of the reduction. In this scenario, the largest and smallest reductions in these measures occurred when the proposed approach and EEMD-QRF-MOO for the states of SP and PR are compared. This is associated with the fact that in the proposed framework, for these states, there is the use of two and four different models, respectively, to compose the heterogeneous ensemble approach. This diversity allows the developed model to achieve the best performance, regarding RRMSE and sMAPE, associated to the fact that the manipulation of some algorithms to build the ensemble is an important strategy to get an efficient model [78]. In this respect, each model learns different patterns of the data, and when the results are combined, they build an effective model. These results corroborate with the findings of Heinermann and Kramer, Xiang et al. [79], [80], proving that using different models for the decomposed components is promising for time series forecasting.

Comparisons II: Evaluation of proposed ensemble learning and non-decomposed models

The comparison II is designed to verify the forecasting performance of the proposed hybrid framework by comparing it with five models which do not consider EEMD for signal decomposition, namely BRNN, CUBIST, GBM, QRF, and PLS. The comparisons are shown in Table 3 and additional discussion is presented. Regarding these results, it is observed that the use of signal decomposition, specifically in the case of the EEMD approach, can enhance the model’s performance. This shows that the EEMD approach is suitable for decomposing the time series of meningitis case for all states. For this reason, the idea of decomposition and ensemble is feasible, and the proposed framework can overcome the drawbacks of individual models by removing the noise of the original data.
Table 3

Performance measures for proposed and non-decomposed models.

StateModelForecasting Horizon
One-month-ahead
Two-months-ahead
Three-months-ahead
MAEsMAPERRMSEMAEsMAPERRMSEMAEsMAPERRMSE
BRNN15.4717.56%19.38%13.8315.79%18.14%13.6715.60%17.66%
CUBIST14.9917.05%20.14%14.6716.73%20.12%13.0814.96%18.57%
MGGBM16.9719.21%21.88%14.6716.72%19.53%15.7417.88%20.99%
PLS15.7217.82%19.65%13.9215.89%18.18%13.9315.88%17.99%
QRF15.5817.60%20.05%15.3317.30%19.50%12.6714.48%16.93%
Proposed7.338.35%10.83%7.929.02%10.63%6.677.37%10.38%

BRNN120.3717.99%24.60%135.3320.39%27.14%160.5623.82%30.24%
CUBIST122.9418.62%23.92%151.3322.79%28.65%153.8222.93%27.96%
SPGBM117.9817.86%23.37%143.8321.26%29.62%178.9126.14%32.15%
PLS110.8016.76%22.53%11717.97%23.41%160.2624.12%28.83%
QRF129.8319.12%26.82%148.8321.97%30.39%171.9625.69%32.34%
Proposed67.589.90%15.14%80.6712.02%17.04%87.0812.77%17.21%

BRNN20.3614.49%22.52%22.5816.24%25.05%23.2116.38%24.29%
CUBIST20.5814.91%24.88%2216.08%28.90%25.4218.08%26.75%
PRGBM21.4115.15%25.21%23.6717.14%27.91%25.7618.39%27.87%
PLS19.1313.71%22.22%20.5814.82%24.06%22.9216.15%24.18%
QRF22.1715.74%23.75%24.6717.85%28.19%25.7518.40%25.89%
Proposed11.507.57%12.10%12.508.61%11.95%11.757.70%13.40%

BRNN11.7913.25%15.46%11.5813.06%15.98%10.6612.09%14.75%
CUBIST11.6012.99%15.26%11.4212.87%15.90%10.8612.29%15.24%
RJGBM10.4511.81%14.34%9.5010.80%13.39%9.9911.36%14.00%
PLS10.5011.83%13.93%10.5011.83%14.49%10.1111.42%14.24%
QRF11.9613.40%14.83%1213.59%15.93%11.5013.04%15.43%
Proposed4.424.98%5.89%6.587.37%8.52%4.084.52%6.49%
Considering the average of each criterion for each forecast horizons, there is a reduction in the criteria for the proposed approach compared to models that do not consider signal decomposition. The increasing on MAE criterion compared models regarding proposed ensemble learning model is ranged between 95.03%–116.18%, 64.90%–91.48%, 75.20%–99.88%, 98.52%–125.62% for the states of MG, SP, PR, and RJ. In its turn, for sMAPE and RRMSE, the same behavior is observed. The smallest reduction is visualized by comparing the proposed framework to the PLS model, and the largest when it is performed comparison with QRF, for both criteria. The results of one-month-ahead up to three-month-ahead forecast for meningitis cases were obtained. The comparisons between the proposed approach and the models without decomposition showing that the performance measures of the proposed model are lower than others obtained from the classic single models. For this reason, the proposed model is more adequate than the single models without decomposition to forecasting meningitis cases. The EEMD presents itself as being the more robust and accurate model, after the nonlinear and non-stationary signals are separated into a series of components. Consequently, this effectiveness allows an improvement of the model’s performance by removing the noise from the time-series [81], [82].

Comparisons III: Evaluation of proposed ensemble learning and decomposed heterogeneous direct integrated ensemble learning models

The comparison III is designed to verify the forecasting performance of the proposed hybrid framework by comparing it with the same structure of EEMD-HTE which uses DI to aggregate the EEMD components, namely EEMD-HTE-DI. The comparisons are shown in Table 4, and additional discussion is presented. In accordance, with these results, it can be stated that the use of MOO in the WI aggregation allows an improvement of the model’s accuracy. Based on the evaluation of all criteria for the EEMD-HTE-MOO, its approach achieves more optimum accuracy than the EEMD-HTE-DI approach. Thus, the proposed methodology can generate robust models that have enhanced forecasting performance with this model combined with it based on decomposition, ensemble, and MOO. The EEMD-HTE-MOO presents a better generalization capacity than the EEMD-HTE-DI, considering that the nonlinearities are accommodated by the use of different weights for each component. In general, this shows that the proposed ensemble model can combine the advantages of the single forecasting models, through the use of MOO, making it possible to add more accuracy and stability to the model [83]. Also, the effectiveness of the proposed approach, according to Bui et al. [84], can be attributed to the process of construction of the combined model using the optimization. Consequently, the MOO allows dealing with the trade-off between bias–variances while assigning different weights for each model. By minimizing the forecast and standard deviation errors it becomes possible to obtain an efficient model with both accuracy and stability improved simultaneously presenting a balance between the two optimized objectives. As presented in Section 5, the results presented in Tables 2 to 4 are obtained through the recursive method to forecasting multi-step-ahead meningitis cases. Through this method, the errors can be accumulated according to the growth of the forecasting horizon. This phenomenon can be observed in MG, PR, and RJ states when are compared the first two forecasting horizons (regarding to MAE and sMAPE), and for SP state in the three forecasting horizons. However, in some cases, the proposed model reduces the forecasting errors, even with the increase of the forecasting horizon. Nevertheless, according to Taieb et al. [85], there is greater evidence of errors accumulation in long-term forecasting horizons, once the input vector adopted by the predictive model can be formed only by previously forecasting, instead of observed values. However, this is not the case of this paper, once the forecast horizon is short (h equals to 1, 2, and 3). In its turn, as three values (=4) are adopted as input to the models, there are always -h+1 observed values in the input vector. Therefore, the forecasting errors can reduce according to the adopted forecasting horizons. Additionally, this information is supported by the results presented in Table 5, in which the error’s standard deviation is presented for each model throughout the twelve study scenarios. The best results are presented in bold.
Table 5

Average standard deviation of errors obtained by each model in forecast out-of-sample (test set forecast).

ModelMGSPPRRJModelMGSPPRRJ
Proposed5.8496.9717.686.36EEMD-MOO-QRF9.06126.6923.419.72
EEMD-MOO-BRNN6.29104.1920.407.45GBM12.75189.4336.7312.79
EEMD-MOO-CUBIST6.3697.2521.417.96PLS12.96166.8232.6212.81
EEMD-MOO-PLS7.41113.2723.998.10BRNN13.67182.6433.2414.05
EEMD-HTE-DI7.47124.9523.518.16CUBIST13.97177.9036.5513.98
EEMD-MOO-GBM7.82145.5427.338.24QRF14.29198.9934.6614.23
By considering the behavior of the meningitis monthly incidence, there is no clear trend in the time series. In this respect, once the forecasting value of two months-ahead is used as models’ input to forecasting three-months-ahead, the information provided in the previous forecasting value can be beneficial to the next forecasting, even though they are different of observed values. In this context, the forecasting errors of the next step-ahead can be reduced, which is observed in Tables 2 to 4. The same is observed in the results achieved by Xiong et al. [86] when the hybrid model which combines seasonal and trend decomposition using locally estimated scatterplot smoothing with extreme learning machine was applied to forecast the vegetable market, and Moreno et al. [87] when the authors proposed a hybrid model based on singular spectral analysis, variational mode decomposition, and long short-term memory to forecasting multi-step-ahead wind speed. Diebold–Mariano test results. 1% significance level. 5% significance level. 10% significance level.

Statistical tests and graphical analysis.

In this subsection, the results of DM tests are used to evaluate the forecasting validity of the proposed model, further verifying the performance of this framework. The EEMD-HTE-MOO errors are compared to the models presented in Sections 6.1, 6.2, 6.3 , and the results of DM test statistics for the errors during the task of forecasting the out-of-sample meningitis cases are presented in Table 6.
Table 6

Diebold–Mariano test results.

ModelMG
SP
One-month-aheadTwo-months-aheadThree-months-aheadOne-month-aheadTwo-months-aheadThree-months-ahead
EEMD-MOO-QRF−1.32−2.91⁎⁎⁎−2.48⁎⁎−2.88⁎⁎⁎−2.06⁎⁎−3.05⁎⁎⁎
EEMD-MOO-PLS−3.08⁎⁎⁎−7.02⁎⁎⁎−2.97⁎⁎⁎−2.91⁎⁎⁎−3.57⁎⁎⁎−4.23⁎⁎⁎
EEMD-MOO-BRNN−2.12⁎⁎−4.07⁎⁎⁎−1.65−3.77⁎⁎⁎−2.18⁎⁎−2.78⁎⁎
EEMD-MOO-GBM−2.02⁎⁎−1.58−1.22−2.62⁎⁎−1.90⁎⁎−2.43⁎⁎
EEMD-MOO-CUBIST−1.99⁎⁎−1.35−2.55⁎⁎−2.63⁎⁎−2.08⁎⁎−2.37⁎⁎
QRF−3.62⁎⁎⁎−2.84⁎⁎⁎−2.11⁎⁎−1.96⁎⁎−2.39⁎⁎−1.96⁎⁎
PLS−3.51⁎⁎⁎−4.52⁎⁎⁎−2.40⁎⁎−2.17⁎⁎−2.07⁎⁎−3.48⁎⁎⁎
BRNN−3.54⁎⁎⁎−5.99⁎⁎⁎−2.42⁎⁎−2.08⁎⁎−2.23⁎⁎−1.87⁎⁎
GBM−3.82⁎⁎⁎−2.71⁎⁎⁎−2.03⁎⁎−1.83⁎⁎−37.76⁎⁎⁎−2.01⁎⁎
CUBIST−2.81⁎⁎⁎−2.41⁎⁎⁎−2.02⁎⁎−2.09⁎⁎−2.65⁎⁎−2.48⁎⁎

ModelPR
RJ
One-month-aheadTwo-months-aheadThree-months-aheadOne-month-aheadTwo-months-aheadThree-months-ahead

EEMD-MOO-QRF−0.58−1.07−1.39−1.55−1.84⁎⁎−1.72⁎⁎
EEMD-MOO-PLS−2.26⁎⁎−2.95⁎⁎⁎−12.61⁎⁎⁎−2.50⁎⁎−0.59−2.94⁎⁎⁎
EEMD-MOO-BRNN−2.43⁎⁎−1.91⁎⁎−6.89⁎⁎⁎−2.16−4.00⁎⁎⁎−1.56
EEMD-MOO-GBM−1.67−2.35⁎⁎−2.74⁎⁎⁎−2.96⁎⁎⁎−1.10−0.95
EEMD-MOO-CUBIST−1.59−2.64⁎⁎−0.80−2.35⁎⁎−2.08⁎⁎−3.76⁎⁎⁎
QRF−1.31−1.24−1.91⁎⁎−2.97⁎⁎⁎−2.35⁎⁎−3.69⁎⁎⁎
PLS−1.24−1.28−1.40−2.41⁎⁎−1.45−1.51
BRNN−1.29−1.30−1.40−3.02⁎⁎⁎−1.81⁎⁎−2.56⁎⁎
GBM−1.30−1.33−1.55−2.12⁎⁎−1.43−5.19⁎⁎⁎
CUBIST−1.33−1.26−1.42−3.14⁎⁎⁎−1.65−2.13⁎⁎

1% significance level.

5% significance level.

10% significance level.

Performance analysis of proposed framework in MG state for one-month-ahead forecast. Performance analysis of proposed framework in SP state for one-month-ahead forecast. Performance analysis of proposed framework in PR state for one-month-ahead forecast. Performance analysis of proposed framework in RJ state for one-month-ahead forecast. Related works to machine learning and general models for the epidemiologic time series forecasting. Control Hyperparameters obtained by GS during cross-validation process for each adopted model. The comparisons are developed according to three situations. First, the proposed framework and the structures that use a homogeneous ensemble of components and MOO are compared. According to this, for the one-month-ahead forecast, the results of Table 6 highlights that in 95% of the comparisons, the errors of the proposed approach are statistically smaller than errors of the compared approaches. In the cases of EEMD-QRF-MOO and EEMD-GBM-MOO, for the states of PR and RJ respectively, the proposed approach presents only smaller errors than these approaches. Considering two-month-ahead or three-month-ahead forecasting, in 85% and 95% of the comparisons, respectively, there is a statistical difference between the errors. For this scenario, the biggest differences are found when the proposed model and EEMD-PLS-MOO approaches are compared with two and three-month-ahead forecast in the states of MG and PR, respectively. It is possible to perceive that the proposed approach shows better performance in a statistical context, in 91.67% of the cases, allowing an inference that makes the developed framework efficient for the proposed task. Second, when the errors of the developed structure and the models that do not employ signal decomposition are compared, for all cases, there is a statistical difference. By comparing these results, Table 6 illustrates that the biggest differences are −3.82, −37.76 and −3.69 in MG and SP, states on the task of one, two and three-month-ahead forecast, respectively, for models GBM and QRF. Additionally, the smallest differences are −1.24, −1.24 and −1.40 for the state of PR, for the three forecast horizons, for models PLS, QRF, and BRNN, respectively. Therefore, these results showed that the use of signal decomposition to forecasting meningitis cases allows the enhancement of the models’ accuracy. Also, the results presented in Table 6, by comparing the EEMD-HTE-MOO and EEMD-HTE-DI, in 86.67% of the cases, there is a statistical difference between the errors. Moreover, on the task of forecasting the meningitis cases in one and two-month-ahead for the state of PR, and three-month-ahead for the state of RJ, there is no statistical difference between the errors. Moreover, the smallest and biggest differences are observed on three-month-ahead for the state of MG (−1.32) and two-month-ahead forecast for the state of RJ (−1.92), respectively. Regarding what has previously been mentioned, it can be stated that the proposed approach showed a significant improvement concerning the use of the DI strategy for the task of grouping the EEMD components. Weights obtained by MOO approach, which are assigned for each model adopted in the signal reconstruction and models used in the structure of proposed hybrid framework. Therefore, regarding the results presented in Sections 6.1, 6.2, 6.3, 6.4 , the proposed approach showed better accuracy than the compared models. In parallel, the proposed framework achieves excellent performance in 83.33% and 25% of the cases, and good results in 16.67% and 25% of the cases for sMAPE and RRMSE, respectively Li et al. [101]. This shows that the use of signal decomposition, heterogeneous ensemble components and MOO combined can enhance the model’s performance, taking into account the combination of each methodology’s expertise. In addition to the presented analysis, Fig. 4, Fig. 5, Fig. 6, Fig. 7 illustrate the Pareto Front (a) of the MOO step, the observed and predicted time series (b) and the residual autocorrelation function of the trained models (c) for the task of one-month-ahead forecasting for the states of MG, SP, PR, and RJ, respectively.
Fig. 4

Performance analysis of proposed framework in MG state for one-month-ahead forecast.

Fig. 5

Performance analysis of proposed framework in SP state for one-month-ahead forecast.

Fig. 6

Performance analysis of proposed framework in PR state for one-month-ahead forecast.

Fig. 7

Performance analysis of proposed framework in RJ state for one-month-ahead forecast.

Considering what is shown by Figs. 4(a), 5(a), Fig. 6, Fig. 7 evidence of a trade-off among the objectives adopted on MOO can be seen, in other words, depending on the weights used to create the ensemble obtained from the EEMD, there is a bias increase while the variance decreases. In this aspect, the use of MOO is adequate, because it allows the obtaining of an efficient model that is able to reach small forecast errors and lower standard deviation errors. The same behavior is replicated for the other two forecast horizons. During this round, Figs. 4(b), 5(b), Fig. 6, Fig. 7 show that the data behavior is learned by the models, in most of the cases, which allows predictions compatible with the observed values. That is, the meningitis cases forecasted are close to the observed values. The good performance regarding accuracy obtained in the training phase persists in the test stage, indicating that the hybrid framework is robust to reach the developed predictions. The overfitting phenomenon occurs when the model has great generalization in the training set, but not in test set or out-of-sample forecasting. To avoid it, two approaches were considered. First, each adopted model was trained using a cross-validation procedure, as described in the methodology section, to prevent overfitting. And second, when the bias and variance are objectives adopted in multi-objective optimization, the trade-off between these measures is considered which leads to overfitting treatment. Also, by the illustrated trough of predicted and observed values (Figs. 4(b), 5(b), Fig. 6, Fig. 7), once a similar performance is observed in training and test sets, there is no evidence of overfitting. Finally, according to what is shown in Figs. 4(c), 5(c), Fig. 6, Fig. 7, it is possible to evaluate the residuals (autocorrelation function) of final models adopted in each case study, from training the set results. The Box–Ljung test is applied to the residuals from the final ensemble models to determine whether residuals are random. In both cases, this test shows that the first 11 lag autocorrelation among the residuals are zero ( = 0.84 - 9.28, p-value 0.05), indicating that the residuals are random and that the model provides an adequate fit to the data [39]. These results are supported by the Figures of residuals ACF since all lags are between confidence interval extremes.

Conclusion and future works

In this paper, a hybrid framework was proposed which combines signal decomposition, a heterogeneous ensemble of EEMD components and WI strategy do aggregate the decomposed components. The MOO approach used to find the PF and TOPSIS is employed to find the best set of weights used during the WI stage. This approach is applied to forecasting three different horizons (one, two, and three-month-ahead) of the meningitis cases for the states of MG, SP, PR, and RJ, Brazil. The performance of the proposed framework was compared to the performance of the models that do not employ signal decomposition, models that apply the same model to forecast all components obtained in the decomposition stage combined with MOO to signal reconstruction, and finally with an approach that employs signal decomposition, heterogeneous ensemble of components and DI strategy to reconstruct the original signal. The MAE, sMAPE, and RRMSE criteria, as well as the DM test, were adopted to evaluate the performance of the developed approach. Finally, a residual analysis was conducted by the ACF plot and Ljung–Box [39] test to validate the proposed hybrid model. According to the forecasting results, it is possible to conclude that: Employing signal decomposition improves the final results on not applying decomposition; The combination of EEMD, HTE, and WI strategy for result aggregation with weights select by NSGA-II TOPSIS allowing the enhancement of the model’s performance for the adopted time series; The use of HTE models for components allowing to improve the framework accuracy in relation to the use of a homogeneous ensemble of components; The use of WI strategy to combine the components is better than the use of DI strategy when optimization is employed to finding the set of weights; The proposed approach reaches better accuracy and stability (lower standard deviation of the errors) than the compared models; In 89.17% of the cases, there is a statistical difference between the errors of the EEMD-HTE-MOO framework and the compared structures. For future works, the adoption of other decomposition techniques, such as variational mode decomposition and singular spectral analysis, is intended to reconstruct the decomposed signal through the clustering of techniques, and using artificial neural networks, adopting stacking ensemble approaches after the decomposition performance, using coyote [102], [103], owls [104], and falcon [105] metaheuristics to tune hyperparameters of adopted forecasting models, using other MOEA to find the weights of WI strategy, as well as make use of the assembling of MCDM techniques. Also, is desirable to compare the recursive and direct methods to perform multi-step-ahead forecasting for the proposed task.

CRediT authorship contribution statement

Matheus Henrique Dal Molin Ribeiro: Conceptualization, Methodology, Formal analysis, Validation, Writing - original draft, Writing review & editing. Viviana Cocco Mariani: Conceptualization, Writing - review & editing. Leandro dos Santos Coelho: Conceptualization, Writing - review & editing.
  26 in total

1.  Partial least squares and logistic regression random-effects estimates for gene selection in supervised classification of gene expression data.

Authors:  Arief Gusnanto; Alexander Ploner; Farag Shuweihdi; Yudi Pawitan
Journal:  J Biomed Inform       Date:  2013-06-07       Impact factor: 6.317

Review 2.  Current meningitis outbreak in Ghana: Historical perspectives and the importance of diagnostics.

Authors:  Alexander Kwarteng; John Amuasi; Augustina Annan; Samuel Ahuno; David Opare; Michael Nagel; Christof Vinnemeier; Jürgen May; Ellis Owusu-Dabo
Journal:  Acta Trop       Date:  2017-01-22       Impact factor: 3.112

3.  Modeling Dengue vector population using remotely sensed data and machine learning.

Authors:  Juan M Scavuzzo; Francisco Trucco; Manuel Espinosa; Carolina B Tauro; Marcelo Abril; Carlos M Scavuzzo; Alejandro C Frery
Journal:  Acta Trop       Date:  2018-05-16       Impact factor: 3.112

4.  A novel bagging C4.5 algorithm based on wrapper feature selection for supporting wise clinical decision making.

Authors:  Shin-Jye Lee; Zhaozhao Xu; Tong Li; Yun Yang
Journal:  J Biomed Inform       Date:  2017-11-11       Impact factor: 6.317

5.  Healthcare-associated ventriculitis and meningitis in a neuro-ICU: Incidence and risk factors selected by machine learning approach.

Authors:  Ivan Savin; Ksenia Ershova; Nataliya Kurdyumova; Olga Ershova; Oleg Khomenko; Gleb Danilov; Michael Shifrin; Vladimir Zelman
Journal:  J Crit Care       Date:  2018-03-23       Impact factor: 3.425

6.  A comparison of three data mining time series models in prediction of monthly brucellosis surveillance data.

Authors:  Nasrin Shirmohammadi-Khorram; Leili Tapak; Omid Hamidi; Zohreh Maryanaji
Journal:  Zoonoses Public Health       Date:  2019-07-15       Impact factor: 2.702

7.  COVID-ABS: An agent-based model of COVID-19 epidemic to simulate health and economic effects of social distancing interventions.

Authors:  Petrônio C L Silva; Paulo V C Batista; Hélder S Lima; Marcos A Alves; Frederico G Guimarães; Rodrigo C P Silva
Journal:  Chaos Solitons Fractals       Date:  2020-07-07       Impact factor: 9.922

8.  Forecasting influenza activity using self-adaptive AI model and multi-source data in Chongqing, China.

Authors:  Kun Su; Liang Xu; Guanqiao Li; Xiaowen Ruan; Xian Li; Pan Deng; Xinmi Li; Qin Li; Xianxian Chen; Yu Xiong; Shaofeng Lu; Li Qi; Chaobo Shen; Wenge Tang; Rong Rong; Boran Hong; Yi Ning; Dongyan Long; Jiaying Xu; Xuanling Shi; Zhihong Yang; Qi Zhang; Ziqi Zhuang; Linqi Zhang; Jing Xiao; Yafei Li
Journal:  EBioMedicine       Date:  2019-08-30       Impact factor: 8.143

9.  Forecasting influenza epidemics by integrating internet search queries and traditional surveillance data with the support vector machine regression model in Liaoning, from 2011 to 2015.

Authors:  Feng Liang; Peng Guan; Wei Wu; Desheng Huang
Journal:  PeerJ       Date:  2018-06-25       Impact factor: 2.984

View more
  3 in total

1.  A novel general-purpose hybrid model for time series forecasting.

Authors:  Yun Yang; ChongJun Fan; HongLin Xiong
Journal:  Appl Intell (Dordr)       Date:  2021-06-05       Impact factor: 5.086

2.  A Highly Discriminative Hybrid Feature Selection Algorithm for Cancer Diagnosis.

Authors:  Tarneem Elemam; Mohamed Elshrkawey
Journal:  ScientificWorldJournal       Date:  2022-08-09

Review 3.  Fault Prediction Based on Leakage Current in Contaminated Insulators Using Enhanced Time Series Forecasting Models.

Authors:  Nemesio Fava Sopelsa Neto; Stefano Frizzo Stefenon; Luiz Henrique Meyer; Raúl García Ovejero; Valderi Reis Quietinho Leithardt
Journal:  Sensors (Basel)       Date:  2022-08-16       Impact factor: 3.847

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.