Literature DB >> 35499039

An incremental learning approach to prediction models of SEIRD variables in the context of the COVID-19 pandemic.

E Camargo1, J Aguilar2,3,4, Y Quintero2, F Rivas5,3, D Ardila2.   

Abstract

Several works have proposed predictive models of the SEIRD (Susceptible, Exposed, Infected, Recovered, and Dead) variables to characterize the pandemic of COVID-19. One of the challenges of these models is to be able to follow the dynamics of the disease to make more precise predictions. In this paper, we propose an approach based on incremental learning to build predictive models of the SEIRD variables for the COVID-19 pandemic. Our incremental learning approach is a dynamic ensemble method based on a bagging scheme that allows the addition of new models or the updating of incremental models. The article proposes an incremental learning architecture composed of two components. The first component carries out an analysis of the interdependencies of the SEIRD variables and the second component is an incremental learning model that builds/updates the predictive models. The paper analyses the quality of the predictive models of our incremental learning approach using data of the COVID-19 from Colombia, and shows interesting results about the predictions of the SEIRD variables. These results are compared with an incremental learning approach based on random forests.
© The Author(s) under exclusive licence to International Union for Physical and Engineering Sciences in Medicine (IUPESM) 2022.

Entities:  

Keywords:  COVID-19; Machine learning; Prediction model

Year:  2022        PMID: 35499039      PMCID: PMC9035346          DOI: 10.1007/s12553-022-00668-5

Source DB:  PubMed          Journal:  Health Technol (Berl)        ISSN: 2190-7196


Introduction

The well-known and most used mathematical models to study the spread of infections/epidemics are the classical ordinary differential equations, such as the SIR (Susceptible, Infectious, Recovered) and SEIRD (Susceptible, Exposed, Infectious, Recovered, and Dead) models, which are widely used in epidemiology. Since the outbreak of COVID-19, the SIR model and its variations have increased their popularity. For the COVID-19, these models have been proposed to study its dynamic in different works [1-3]. For example, Zhong et al. [4] have proposed a SIR model for predicting the COVID-19 using China’s first reported data. Also, Yang and Wang [5] presented an extended SEIR model for COVID-19 with time-varying transmission rates by considering the environmental effects. On the other hand, Zhu et al. [6] present a stochastic-based method for modeling and analysis of COVID-19 spread using a SEIR-Re-infected and Deceased-based Social Distancing model, called SEIR(R)D-SD, to consider the re-infection rate and social distancing factor into the traditional SEIRD model. In particular, the SEIRD model is the most adopted to characterize the COVID-19 pandemic because, from it, the effectiveness of various measures to attack the outbreak can be evaluated, which seems to be a difficult task for general statistical methods. Other works have considered alternative approaches based on machine learning techniques to exploit the large amount of data that has been collected on COVID-19, with the aim of developing predictive models of the behavior of the SEIRD variables [7, 8]. For example, Quintero et al. [9] have proposed predictive models using classical machine learning techniques for the SEIRD variables based on historical data collected about them, and the contextual variables where the model was applied. Particularly, the contextual variables considered in this work are: total population, the number of people over 65, poverty index, morbidity rates, average age and population density. For the construction of the SEIRD predictive models, in Quintero et al. [9], an analysis of the dependencies of these variables among themselves, and with the context variables, was carried out. A first work that uses deep learning related to COVID-19 is presented by Pelaez and Loayza [10]. They propose the use of a convolutional neural network for helping in the diagnosis of COVID-19 using chest X-ray images, giving as output the probability of identifying pneumonia caused by this virus. Finally, Chen et al. [11] develop a data-driven workflow to extract, process, and develop models of the COVID-19 epidemic using deep learning methods. However, in the context of the pandemic, a relevant issue for data-based models is that they must adjust to the data that is being captured in real-time. In this sense, new online learning approaches are required in order to keep updating the models. Approaches such as incremental learning, online learning, active learning, are necessary. In this work, we focus on incremental learning techniques [12]. Incremental learning is a machine learning method where the current input data is continuously used to update the knowledge model. The aim of incremental learning is to adapt the model to new data without forgetting its existing knowledge. Many traditional machine learning algorithms inherently support incremental learning. Other algorithms can be adapted to support incremental learning. Examples of incremental algorithms include [12-14]: decision trees, decision rules, some artificial neural networks models (ex. Fuzzy ART and TopoART) or the incremental Support Vector Machine (SVM). In the case of incremental learning works for the health area, an approach in medical diagnosis is proposed by Casalino, et al. [15], which consists in a Dynamic Incremental Semi-supervised Fuzzy Clustering to detect Bipolar Disorder Episodes. In the same direction of medical diagnosis using incremental learning, Braccioni, et al. [16] used a Forest-Tree Machine Learning approach for lung transplant recipients to study symptoms such as dyspnea, muscle effort and muscle pain, and their relationship with cardiac and pulmonary function parameters during an incremental exercise testing. Also, Schiel et al.[17] defined an approach for improving the prediction in electromyography signals for controlling prosthesis using incremental learning in a sparse Gaussian Process. Wang et al. [18] presented a proposal for the multiple percept detection problems in image sequences, in order to include new data from not previously identified categories, using the fast approximate nearest neighbor search tree-based filter using incremental learning. Another approach for using incremental learning in deep learning is presented by Su et al. [19]. They have proposed a framework for an incremental learning approach with a deep convolutional neural network in an anthropomorphic robot manipulator. Finally, Neto et al. [20] proposed an incremental LSTM model to tackle evolving data stream problems. Now, to the best of our knowledge, there are no works on incremental learning approaches applied in the field of prediction for COVID-19. In this paper, we propose an incremental learning approach for prediction models based on a dynamic ensemble method that uses a bagging scheme that allows the addition of new models or the updating of incremental learning models. In this way, when the degradation of the ensemble method prediction is detected, then our approach prepares a training dataset with the current data to build new predictive models or update the incremental learning models. The new models are built with different machine learning techniques, and are considered by the ensemble method using a bagging scheme. Specifically, this article defines an incremental learning architecture that has two components. A first component carries out an analysis of the interdependencies of the SEIRD variables similar to the work [9], and a second component is the incremental learning model that builds/updates the predictive models of the SEIRD variables for the COVID-19. Additionally, this work compares the prediction results of our approach with an incremental learning approach based on random forests. The remainder of the paper is organized as follows. Section 2 defines incremental learning and Sect. 3 describes our incremental learning architecture. Section 4 presents the experiments and results. Finally, Sect. 5 contains the conclusions.

Incremental learning

For a long time in the history of machine learning, there has been an implicit assumption that a “good” training set in a domain is available a priori. The training set then is “good” if it contains all necessary knowledge that once learned, can be reliably applied to any new examples in the domain. Consequently, emphasis is put on learning as much as possible from a fixed training set. Unfortunately, many real-world applications cannot match this ideal case, such as in dynamic control systems, web mining, and time series analysis, where the training examples are often fed to the learning algorithms over time, i.e., the learning process is continuous. So, the incremental learning approaches can handle these problems and yield adjustments in the obtained models that can include newly collected data. The main idea of incremental learning approaches is to create models that can be dynamically adjusted without losing previously found knowledge. It can include new data sets and also new classes or other information that can be incorporated into the models. This is a very good option for big data systems because it can include the data as time goes, without needing to stop and spend time creating new models. The three main conditions for using incremental learning are: The complete data set is not available at the moment of creating the model. The learning approach should be able to incorporate the new data without losing the knowledge previously obtained. It must find a balance between stability (the ability to not forget the already acquired knowledge) and plasticity (the ability to adjust the new data presented). According to Chefrour [12], the diverse incremental learning algorithms can be classified into two classes: Adaptive systems: In this case, the structure of the system remains fixed, and the parameters are adjusted according to the new data received. Evolving systems: In this case, the structure can be modified by including new classes or behavior according to system changes over time. There are some problems in incremental learning, Gepperth and Hammer [21] consider that the most important challenges are the online-adaptation of the model parameters, the concept drift (changes in data statistics that occur over time), the stability-plasticity dilemma, among others. In general, there are three different kinds of incremental tasks: Example-Incremental Learning Tasks (E-IL Tasks) New training examples are provided after a learning system is trained. For example, a face recognition system can gradually improve its accuracy by incorporating new face images of registered users as they use it, without reconfiguring and/or retraining the entire system. Class-Incremental Learning Tasks (C-IL Tasks): New output classes are provided after training a learning system. For example, if a new user is added to the group of registered users in the facial recognition system mentioned above, the system should be able to recognize the new user without reconfiguring or retraining the entire system. Attribute-Incremental Learning Tasks (A-IL Tasks): New input attributes are provided after training a learning system. For example, if the camera used in the aforementioned facial recognition system is changed from a grayscale camera to a color camera, the system should be able to use the additional color features without reconfiguring or retraining the entire system. A suitable incremental learning algorithm must be carefully designed. In some applications, incremental learning algorithms should automatically distinguish previous cases, and take actions to improve the representation of the current concepts as more examples become available, or respond to changes in the definition of the concepts when incoming examples become inconsistent with the learned concepts. Schlimmer and Granger [22] proposed three criteria to measure the usefulness and effectiveness of an incremental learning method: (1) the number of observations (examples) needed to obtain a ‘stable’ concept description, (2) the cost of updating memory, and (3) the quality of learned concept descriptions. Some years later, [13] claimed that in order to measure an incremental learning algorithm, two new questions need to be answered: (1) How much better is a learned model at step n + i than another model obtained before step n? (2) Can an incremental algorithm recover in the next incremental step(s), if it goes drastically off the “actual” concept at any stage? Consequently, they proposed three criteria for the evaluation of the robustness of incremental learning algorithms: (1) Stability – the prediction accuracy on the test set should not vary wildly at every incremental learning step; (2) Improvement – there should be an improvement in the prediction accuracy as the training progresses and the learning algorithm sees more training examples; and (3) Recoverability – the learning method should be able to recover from its errors, i.e., if the performance drops at a certain learning step, then the algorithm should be able to recover to the previous best performance. Another frequently used criterion is the learning curve. An incremental algorithm may start learning from scratch and gain knowledge gradually with an increasing number of training examples. Consequently, the quality of the learned model shows a curve of gradual improvement over time, which is called the learning curve. Usually, the learned model is not very reliable at the early stage of the curve. Decisions can be made according to the learning curve on how valuable the output of the incremental learner might be at a certain stage. However, in practice, it is often difficult to determine the point at which the model has learned “enough” to be reliable. Generally, a typical “good” learning curve should increase rapidly to a relatively steady high level.

Our incremental learning approach

In this work, we propose an incremental learning architecture whose main objective is to predict the future behavior of SEIRD variables in a time window. This architecture allows updating the predictive models of the SEIRD variables with the most recent information (determined by the time window) on the behavior of COVID-19. The architecture consists of two components (see Fig. 1): A dependency analysis process, which defines the variables that will be considered by the predictive models, and the incremental learning model that predicts the SEIRD variables.
Fig. 1

Incremental Learning Architecture

Incremental Learning Architecture Specifically, the first component performs a feature engineering process to select the variables to build the predictive models. In particular, this process consists of a dependency analysis to find the relationships between the SEIRD variables and the predictor variables, that is, to find the variables that are relevant for the construction of the predictive models. The second component builds the predictive models using our incremental learning model. The incremental learning process invokes the dependency analysis process from time to time to update the relationships between the SEIRD variables and the predictor variables. Thus, this architecture that considers two vital processes for the construction of accurate predictive models, one of feature engineering and the other of incremental learning, is the first contribution of this article.

Variable dependence analysis for the SIERD model for incremental learning

The variables considered important as predictor variables for the incremental learning model were obtained through a time series analysis that considers other variables as predictors. In this work, a time series analysis is carried out to detect the relationships between the variables due to the data set has a chronological order. The model can be seen in Eq. 1. where: βo and γi are real constants, and ηI is an ARIMA process. The models were adjusted with the 5 variables of the SEIRD model as predictor variables, to which the delayed effects of up to 7 days were considered. In this way, for the 5 variables is analyzed the dependence with the other variables and themselves over time. To evaluate the method, the dataset is divided into training and testing, and to evaluate the predictions made is used the Mean Absolute Percentage Error (MAPE) of the predictions of the dataset (see Eq. 2). where: is the real value of the variable in time , is the predicted value of the variable in time n is the total amount of data in the training set, and N is the total amount of data in the test set. To determine the best predictors, a genetic algorithm (GA) was implemented, which searches for the best descriptors for each target variable. Figure 2 shows the process followed by the method.
Fig. 2

General procedure for selection/extraction of the best features

General procedure for selection/extraction of the best features Figure 2 describes the general procedure of detection/selection of the best descriptors that are highly associated with the SEIRD variables, using the GA with ARIMA time series models. The inputs of the GA are: the variable of interest (SEIRD) and the set of predictors (SEIRD lagging). Within the GA there is a selection and variation process in which different subsets of predictors are obtained. With each combination of predictors, several ARIMA models are adjusted, varying the parameters (p, d, q), and the best model is chosen according to the MAPE metric (see Eq. 2). At the end, the GA provides the best adjusted ARIMA model and the subset of predictors. Thus, GA determines the best combination of lagging variables as descriptors, optimizing MAPE.

Incremental machine learning model

We propose an incremental learning approach based on the Ensemble Learning paradigm. Our goal is to create models that constantly learn from COVID-19 data, to generate more accurate predictions of the SEIRD model of the COVID-19. The general process is shown in Fig. 3. In general, our approach uses a bagging scheme like the ensemble method [23, 24]. It initially trains a set of predictive models, which use different machine learning techniques (gradient boosting, random forest, etc.) and are trained using a random subset of the data. These subsets of data (random samples) are defined using a 0.632 Bootstrap strategy. During the operation, our approach, in the first stage, makes online predictions using the predictive models that have been previously built. Then, the MAPE of each model is calculated and the model with the lowest MAPE is selected. If the prediction of the best predictive model is very bad, then it builds new prediction models (maybe with new machine learning techniques) and/or updates the current incremental learning models. For that, it builds a new dataset with the most recent data, and using a 0.632 Bootstrap strategy, defines the training data to update the incremental learning models and/or create new predictive models. The newly trained models are included in the bagging scheme to become part of the set of predictive models. Thus, our approach keeps its predictive capability.
Fig. 3

Our incremental learning approach

Our incremental learning approach Prediction of the Susceptible variable with our Incremental learning approach Prediction of the Exposed variable with our Incremental learning approach Prediction of the Infectious variable with our Incremental learning approach Prediction of the Recovered variable with our Incremental learning approach Prediction of the Death variable with our Incremental learning approach Our approach uses powerful machine learning techniques for building predictive models, like Gradient boosting, Random Forest, Bootstrap Aggregation, linear regression, and the Limited-memory BFGS (L-BFGS or LM-BFGS) neural network [23, 25]. Also, our approach proposes a dynamic ensemble method that adds new predictive models but also, can update the incremental learning models. In this way, when the degradation of the ensemble method prediction is detected, then our approach prepares a training dataset with the current data to build new predictive models or update the incremental learning models. Our approach has two parameters: the threshold beyond which the predictions are considered not good,and the backward time interval that is used to calculate the prediction error (deferred prediction evaluation). Both must be optimized according to the dataset used.

Experimentation

Experimental context

For the development of this work, the dataset used to predict the SEIRD variables is from Colombia, and it contains the daily information of these variables for the COVID-19. This dataset can be obtained from the official website of the National Institute of Health of Colombia (INS). In the INS dataset can be found the next variables: Date: timestamp; Exposed: number of people who have been exposed, which is estimated with the number of tests performed; Infected: number of people infected; Recovered: number of people who have recovered; and Dead: number of people who have died. The Susceptible variable is calculated as the total number of inhabitants in Colombia minus the variables exposed, recovered, infected and deaths. For the variables exposed, infected, recovered and dead, a summary is presented in Table 1, which shows the average, minimum (min), maximum (max) and standard deviation (sd) of each of these variables. The susceptible variable is not in Table 1, since it is calculated as the difference between the total number of inhabitants of the country and the other variables.
Table 1

Variables statistics

ExposedInfectedRecoveredDead
Average16.0973.9293.284135
min258550
max41.43413.05612.295438
sd12.3284.0103.670132
Variables statistics The dataset contains information for 300 days for each of the variables mentioned above. For the susceptible variable, there is no meaning in calculating these statistics. Table 1 shows that the total number of infected persons per day is between 5 and 13.056, with an average of 3.929 infected persons per day. On the other hand, there were days in which there were no deaths from COVID-19, and sometimes a maximum of 438 deaths, daily on average there were 135 deaths, while on average the daily total of people recovered was 3.284, and on some days this figure rose to 12.295 people recovered. In addition, an average of 16.097 people were exposed to the virus daily, reaching a maximum of 41.434. For this study, the target variables were Susceptible, Exposed, Infected, Recovered and Dead. The predictive models of the target variables were built with the Gradient Boosting Regressor, with the following hyperparameters: the loss function to be optimized was least squares regression, n_estimators = 500, learning_rate = 0.01, criterion = "MSE" and test_size = 0.2; Random Forest with the following hyperparameters:n_estimators = 500, criterion = "MSE", max_depth = 100, min_samples_leaf = 2 and test_size = 0.2; Linear regression with the following hyperparameters:test_size = 0.2 and random_state = 10;and L-BFGS with the following hyperparameters: Relu was used as the activation function, the optimization function was Adam and the loss function was the MSE,hidden_layer_sizes = (25), solver = 'lbfgs' and alpha = 1e-5. All the experimentation was done with data that are between March and December of 2020. For the analysis of dependencies, the MAPE was used as the metric to evaluate the predictive quality of the model. The quality metrics used to evaluate each model were MAPE, Mean Square Error (MSE), and coefficient of determination (R2).

Experimental cases

Variable dependence analysis

Two experiments were carried out for the analysis of dependencies, one for August and the other for September. With the development of this study, it was not only determined the best model with the set of variables that compose it, but it was also possible to find the time window to predict the future behavior. Table 2 shows the average MAPE for the 4 weeks of August for the intervals of 1, 2, 3 and 4 days. The reason for using a 4-day interval to make predictions is that up to this number of days is where we have obtained the best performance of the predictive models for the interval of delayed of 7 days analyzed.
Table 2

Average MAPE predictions for August 2020

Variables1 day (%)2 day (%)3 day (%)4 day (%)
Susceptible0.0250.0380.0970.009
Exposed5.4847.7757.7037.931
Infected6.8166.5297.1628.738
Recovered6.3066.8917.7147.956
Deaths11.65117.56618.34816.565
Average MAPE predictions for August 2020 The MAPE values presented in Table 2 are the mean of the MAPE values calculated for each week in the prediction intervals (1, 2, 3 and 4 days). According to the results of the susceptible and dead variables, the best prediction is obtained on days 1 and 4, i.e., on average on days 1 and 4 the predictions of these variables are closer to reality. The infected and recovered variables are closer to reality, on average, on days 1 and 2, while the exposed variable is closer to reality on days 1 and 3. The average MAPE for the 4 weeks of September is shown below, at intervals of 1, 2, 3 and 4 days (see Table 3).
Table 3

Average MAPE predictions for September 2020

Variables1 day (%)2 day (%)3 day (%)4 day (%)
Susceptible0.7631.0460.7261.020
Exposed6.9326.7666.5467.077
Infected6.6137.7007.4789.128
Recovered8.0058.1468.5547.585
Dead13.61912.92514.21013.469
Average MAPE predictions for September 2020 In the second case, the quality of the models and the time window for September 2020 are shown in Table 3. As for August 2020, the values presented are the mean of the MAPE values calculated for the 4 weeks in the prediction intervals (1, 2, 3 and 4 days). In this case, 3 of the 5 predicted variables are closer to reality on day 3 (susceptible, exposed and infected); however, susceptible and infected are also closer to reality on day 1, while exposed on day 2. On the other hand, the recovered and dead variables have in common that on day 4 the predictions made for them are closer to reality, and the second-best predicted day for them are 1 and 2, respectively. Tables 2, 3 show the average MAPE values for the 4 weeks of each month, for 4 days (time window), since predictions beyond these days are very far from reality (MAPE values higher than 50%), and result in unreliable values in the predictions of the machine learning models. Now, the prediction of the death variable is not so good, regardless of the time window. Considering the dependency analysis performed, the variables that help to better predict the behavior of the SEIRD variables are given in Table 4, and are the ones considered to build the machine learning models.
Table 4

Target and Predictor variables for machine learning models

TargetPredictor Variables
Susceptibleexposed(t-5), exposed(t-6), infected(t-5), recovered(t-5), recovered(t-7)
Exposedsusceptible(t-6), infected(t-7), deaths(t-5), dead(t-7)
Infectedexposed(t-6), exposed(t-7), deaths(t-5), dead(t-6)
Recoveredsusceptible(t-6), susceptible(t-7), exposed(t-5), exposed(t-6), exposed(t-7), infected(t-5), infected(t-7), dead(t-5)
Dead

susceptible(t-5), exposed(t-5), exposed(t-6), infected(t-5),

recovered(t-5), recovered(t-7)

Target and Predictor variables for machine learning models susceptible(t-5), exposed(t-5), exposed(t-6), infected(t-5), recovered(t-5), recovered(t-7)

Prediction models

Table 5 shows the performance of our approach predicting the SEIRD variables based on the analysis of the time dependence that determined a 4-day predictive interval, where each SEIRD variable has the dependence defined in Table 4. Table 5 presents an example of the performance of each technique used by our incremental learning approach for the SEIRD variables for three different moments (), using MAPE, MSE and R2 metrics. is the backward time interval (deferred prediction evaluation), which was determined as 7-day after a hyperparameter optimization process (we have used a Bayesian hyperparameter optimization approach). Also, during the hyperparameter optimization process was determined the threshold values equal to 0.9. Finally, we have used an incremental learning version of random forest.
Table 5

Quality of the models used to predict learning incremental the SEIRD variables for Colombia

Target VariableMachine Learning Techniques in our incremental approach\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$t-1$$\end{document}t-1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$t-2$$\end{document}t-2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$t-3$$\end{document}t-3
R2MSEMAPER2MSEMAPER2MSEMAPE
SGradient boosting0.9900.00171.5540.9993.1e-0570.2710.9930.00170.770
Random forest0.9220.00371.5810.8220.00971.4150.8920.00872.180
Linear regressor0.9980.00176.6640.9960.00176.7610.9110.00275.890
LM-BFGS0.9330.00177.1050.9940.00177.0540.9220.00178.200
EGradient boosting0.7880.00677.4240.4810.01570.2710.6510.02176.770
Random forest0.8420.00671.7860.9870.00171.5020.8870.0770.118
Linear regressor0.7880.00676.0640.8990.00376.6810.8090.00375.889
LM-BFGS0.9500.00577.9050.8460.00677.0540.8010.01178.020
IGradient boosting0.8720.00474.5640.9900.00170.0710.8100.00270.077
Random forest0.8580.00575.6810.4430.01071.1020.3430.02171.618
Linear regressor0.8600.00576.6640.8240.00476.7810.8400.00475.989
LM-BFGS0.8330.00677.8030.8630.00477.1540.7910.00778.120
RGradient boosting0.9790.00171.5540.9950.00170.9710.8450.01370.177
Random forest0.9550.00570.9230.2290.03071.1020.2460.03171.118
Linear regressor0.8840.00676.6640.9800.00176.7820.9800.00175.189
LM-BFGS0.9380.00278.3020.9910.00177.1540.9810.00478.120
DGradient boosting0.9190.00770.4320.9680.00170.8710.9550.00370.177
Random forest0.9330.00870.5810.1880.03671.5020.2310.03071.118
Linear regressor0.8290.00975.3420.9590.00276.6810.9400.00375.189
LM-BFGS0.8630.00676.9030.9350.00277.0540.9120.00378.120
Quality of the models used to predict learning incremental the SEIRD variables for Colombia In general, our approach has created a new predictive model for E and I variables because the initial predictive models are not good (see Table 5, column. ), R2s are smaller than 0.9). Also, we can see that the best predictive model of our approach can change over time. For example, for the variable S, Gradient boosting is the best at the beginning (R2 = 0.993, MAPE = 71.554 and MSE = 0.0001), and after, the linear model is the best. Finally, the only incremental learning technique implemented, besides our approach, was random forest, and sometimes it gives very bad results after retraining. For example, for the variable S, but it improves for the variables I, R and D. For the other techniques, in each step , our approach adds randomly new predictive models based on them to the set of predictive models, trained with a new dataset using a 0.632 Bootstrap strategy. Figures 4, 5, 6, 7, 8 show the prediction of our incremental learning model (remember that it selects the best prediction as output) of the SEIRD variables for the interval t-4. In these Figures, the predictions of the SEIRD variables are made with a 95% confidence interval, and the predicted values by our incremental learning approach are the blue line. Based on these results, each variable has a low error predicting its value. Thus, our incremental learning approach follows all variables quite well. In some abrupt behaviors, it has problems to predict them (but very slight, see Figs. 6, 7, 8), while when it is very homogeneous the behavior, then it can follow it perfectly (see susceptible variable).
Fig. 4

Prediction of the Susceptible variable with our Incremental learning approach

Fig. 5

Prediction of the Exposed variable with our Incremental learning approach

Fig. 6

Prediction of the Infectious variable with our Incremental learning approach

Fig. 7

Prediction of the Recovered variable with our Incremental learning approach

Fig. 8

Prediction of the Death variable with our Incremental learning approach

Finally, we consider three metrics of the incremental learning area (see section II.2) to assess the quality of our approach to incremental learning: Stability, Improvement and Recoverability. With respect to the Stability metric, our approach keeps the prediction accuracy on the test set because it does not vary in every incremental learning step. Concerning the Improvement metric, also our approach keeps the improvement in the prediction accuracy as the training progresses and the learning algorithm receives more training examples; Finally, for the Recoverability metric, our learning method can recover very well because it can keep the performance drops at a certain time using the bagging approach such as it selects the best prediction model, and when it determines that the result is very bad, then it retrains (for the predictive models based on incremental learning techniques) and/or learns new models.

Conclusions and future works

The SEIRD model is a mathematical model based on dynamical equations, which has been widely used for characterizing the pandemic of COVID-19. Several works have proposed to predict these variables using the large amount of data that has been collected on the disease. However, these models need to be updated online considering the dynamics of the disease, particularly, to be able to characterize each of the waves that the disease has had. In this work, we have proposed a novel incremental learning architecture, which can predict the SEIRD variables. This architecture has a first component to analyze the temporal inter-dependence and intra-dependence of the SEIRD variables (it carries out a feature engineering process). Also, the architecture has a second component, an incremental learning model based on the Ensemble Learning paradigm composed of two steps: In the first step, the incremental learning model follows a bagging scheme to make predictions based on different predictive models that have been built using different machine learning techniques. Then, this model selects the best prediction as output. If the prediction is not good, then it builds new prediction models (maybe with new machine learning techniques) or/and it updates the current models (for the predictive models developed with incremental learning techniques). The incremental learning model is the second contribution of this work. With respect to the obtained results, they are of different types. A first result is about the analysis of dependencies, which defines the relationships between the SEIRD variables and the predictor variables, and the interval of delay to be used. The first component of our architecture determines both of these things quite efficiently. A second result is the predictive models of the SEIRD variables. In general, our approach can create new predictive models or update the current models developed with incremental learning techniques. The second component of our architecture carries out these tasks quite well according to the values obtained in the performance metrics. Finally, our incremental learning architecture accomplishes the typical criteria of an incremental learning approach, namely, stability, improvement and recoverability criteria. In general, the predictive results of our incremental learning approach are very good because, by the bagging approach, we can obtain good predictions with the predictive models for the current data; and when it considers necessary a new predict model, then it generates it with the new online data. By combining individual models, the ensemble model tends to be very flexible (less biased) and less data-sensitive (less variance). Thus, our approach can follow the different behaviors of the SEIRD variables with a low error. In each case, it considers a different predictive model to be used. Future works require testing the model to see how it behaves in countries that have already had more than two waves, in such a way as to determine if the proposed approach can perform well in inter-wave predictions, understanding that the intra-wave transition will require of training phases to generate the appropriate models for these new waves, which our approach allows perfectly. Also, our model is based on the MAPE metric for the data dependency analysis, and in the MAPE, MSE and R2 metrics for the estimation of the quality of the prediction. Here, again, it is possible to test with other metrics, in order to test the sensibility of our approach. Moreover, it is possible to extend the machine learning algorithms used in other areas [26, 27], to develop specialized incremental learning approaches for specific domains. Finally, the addition of ontological information during the prediction process will be analyzed to introduce contextual information like reasoning [28], or the introduction of an online feature engineering process [29, 30] inside of our incremental learning approach.
  11 in total

1.  Early Prediction of the 2019 Novel Coronavirus Outbreak in the Mainland China Based on Simple Mathematical Model.

Authors:  Linhao Zhong; Lin Mu; Jing Li; Jiaying Wang; Zhe Yin; Darong Liu
Journal:  IEEE Access       Date:  2020-03-09       Impact factor: 3.367

Review 2.  Artificial Intelligence (AI) applications for COVID-19 pandemic.

Authors:  Raju Vaishya; Mohd Javaid; Ibrahim Haleem Khan; Abid Haleem
Journal:  Diabetes Metab Syndr       Date:  2020-04-14

3.  A modified SEIR model to predict the COVID-19 outbreak in Spain and Italy: simulating control scenarios and multi-scale epidemics.

Authors:  Leonardo López; Xavier Rodó
Journal:  Results Phys       Date:  2020-12-25       Impact factor: 4.476

4.  Machine learning models for the prediction of the SEIRD variables for the COVID-19 pandemic based on a deep dependence analysis of variables.

Authors:  Yullis Quintero; Douglas Ardila; Edgar Camargo; Francklin Rivas; Jose Aguilar
Journal:  Comput Biol Med       Date:  2021-05-24       Impact factor: 4.589

5.  A mathematical model for the novel coronavirus epidemic in Wuhan, China.

Authors:  Cha Yu Yang; Jin Wang
Journal:  Math Biosci Eng       Date:  2020-03-11       Impact factor: 2.080

6.  The effect of control strategies to reduce social mixing on outcomes of the COVID-19 epidemic in Wuhan, China: a modelling study.

Authors:  Kiesha Prem; Yang Liu; Timothy W Russell; Adam J Kucharski; Rosalind M Eggo; Nicholas Davies; Mark Jit; Petra Klepac
Journal:  Lancet Public Health       Date:  2020-03-25

7.  Modified SEIR and AI prediction of the epidemics trend of COVID-19 in China under public health interventions.

Authors:  Zifeng Yang; Zhiqi Zeng; Ke Wang; Sook-San Wong; Wenhua Liang; Mark Zanin; Peng Liu; Xudong Cao; Zhongqiang Gao; Zhitong Mai; Jingyi Liang; Xiaoqing Liu; Shiyue Li; Yimin Li; Feng Ye; Weijie Guan; Yifan Yang; Fei Li; Shengmei Luo; Yuqi Xie; Bin Liu; Zhoulang Wang; Shaobo Zhang; Yaonan Wang; Nanshan Zhong; Jianxing He
Journal:  J Thorac Dis       Date:  2020-03       Impact factor: 3.005

8.  Artificial intelligence and machine learning to fight COVID-19.

Authors:  Ahmad Alimadadi; Sachin Aryal; Ishan Manandhar; Patricia B Munroe; Bina Joe; Xi Cheng
Journal:  Physiol Genomics       Date:  2020-03-27       Impact factor: 3.107

9.  Dyspnea, effort and muscle pain during exercise in lung transplant recipients: an analysis of their association with cardiopulmonary function parameters using machine learning.

Authors:  Fausto Braccioni; Daniele Bottigliengo; Andrea Ermolao; Marco Schiavon; Monica Loy; Maria Rita Marchi; Dario Gregori; Federico Rea; Andrea Vianello
Journal:  Respir Res       Date:  2020-10-15
View more
  1 in total

1.  Analysis of the socioeconomic impact due to COVID-19 using a deep clustering approach.

Authors:  Yullys Quintero; Douglas Ardila; Jose Aguilar; Santiago Cortes
Journal:  Appl Soft Comput       Date:  2022-09-05       Impact factor: 8.263

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.