Literature DB >> 35936352

Improved seagull optimization algorithm of partition and XGBoost of prediction for fuzzy time series forecasting of COVID-19 daily confirmed.

Sidong Xian1,2, Kaiyuan Chen2, Yue Cheng1.   

Abstract

The establishment of fuzzy relations and the fuzzification of time series are the top priorities of the model for predicting fuzzy time series. A lot of literature studied these two aspects to ameliorate the capability of the forecasting model. In this paper, we proposed a new method(FTSOAX) to forecast fuzzy time series derived from the improved seagull optimization algorithm(ISOA) and XGBoost. For increasing the accurateness of the forecasting model in fuzzy time series, ISOA is applied to partition the domain of discourse to get more suitable intervals. We improved the seagull optimization algorithm(SOA) with the help of the Powell algorithm and a random curve action to make SOA have better convergence ability. Using XGBoost to forecast the change of fuzzy membership in order to overcome the disadvantage that fuzzy relation leads to low accuracy. We obtained daily confirmed COVID-19 cases in 7 countries as a dataset to demonstrate the performance of FTSOAX. The results show that FTSOAX is superior to other fuzzy forecasting models in the application of prediction of COVID-19 daily confirmed cases.
© 2022 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  COVID-19; COVID-19, Corona Virus Disease 2019; FST, Fuzzy time series; Fuzzy time series; ISOA, Improved seagull optimization algorithm; Improved seagull optimization algorithm; XGBoost; XGBoost, Extreme Gradient Boosting Tree

Year:  2022        PMID: 35936352      PMCID: PMC9340105          DOI: 10.1016/j.advengsoft.2022.103212

Source DB:  PubMed          Journal:  Adv Eng Softw        ISSN: 0965-9978            Impact factor:   4.255


Introducation

Time series prediction is crucial in many fields, but typical prediction approaches are ineffective for some time series containing fuzzy information. The concept of the fuzzy set, as an extension of set theory, was proposed by Zadeh [1]. Fuzzy sets and their variants are used to deal with problems with fuzzy information. Zeng [2] proposed an intuitionistic fuzzy social network hybrid MCDM model for an assessment of digital reforms in the manufacturing industry in China. The social network multiple-criteria decision-making approach for evaluating unmanned ground delivery vehicles under the Pythagorean fuzzy environment was proposed by Zeng [3]. The fuzzy time series(FTS) was first proposed by Song and Chissom [4] on the basis of fuzzy sets to work out time series problems related to fuzzy information. Establishing fuzzy relations and fuzzing the time series are vital steps of the forecasting model in FTS. Fuzzing original time series is the start of predicting model. With the development of fuzzy time series, there are two main means for dealing with the fuzzification of time series. Partitioning the discourse of the dataset into some intervals is the first method. The fuzzy membership function can transform the original data into the fuzzy membership of intervals, which is the main procedure of this method. The critical point of this method is the length of intervals, and many papers have studied how to split the universe of discourse into appropriate intervals. Some researchers [5], [6], [7], [8] explored the influence of lengths of intervals and the method of partitioning intervals to improve the performance of forecasting model. Chen[9] proposed the optimal weighting vectors to find optimal partitions. Bose [10] researched the data partitioning and rule selection technique for FTS with the effect of interval length. Lu [11] used interval information granules to improve forecasting in fuzzy time series. Some literature uses intelligent optimization algorithms to partition intervals. Nizam [12] proposed an improved model based on FTS and PSO for forecasting blood glucose level. An improved genetic algorithm was proposed by Bas[13] for predicting fuzzy time series. Tinh [14] researched the prediction of fuzzy time series with particle swam optimization. In order to partition intervals, an improved artificial fish swarm optimization algorithm was proposed by Xian [15]. In recent years, some new optimization algorithms have emerged, such as seagull optimization algorithm(SOA). They have been shown to have better performance, but have not been used in FTS forecasting models. The second method of fuzzification is based on fuzzy clustering. The fuzzy membership of varieties can be obtained by fuzzy clustering. On the basis of fuzzy clustering, Aliyev [16] proposed a novel fuzzy time series method for hotel occupancy forecasting. A fuzzy time series forecasting model was proposed by Vovan [17] based on an improved fuzzy function and cluster analysis problems. On the basis of Gustafson-Kessel fuzzy clustering Fan [18] proposed a long-term intuitionistic fuzzy time series predicting model. The length of intervals obtained by the optimization method and fuzzy clustering is often different because the distribution of the sample data on its universe is not uniform, and there has a complex internal structure. The appropriate interval length can better reflect the data structure and improve the prediction accuracy. The establishment of fuzzy relations is the core step of the fuzzy forecasting model. Some literature further studied the fuzzy relation for increasing the accuracy of the forecasting model. Abhishekh [19] researched the fuzzy relations in the forecasting model. Kocak [20] used an ARMA-type recurrent Pi-Sigma artificial neural network to replace fuzzy relations in high-order fuzzy time series. Cheng [21] forecasted the financial market with the weighted association rule and fuzzy time series model. Some other literature uses regression methods or Markov chains instead of fuzzy relations. Alyousifi [22] proposed a fuzzy time series model with Markov chains. A novel forecasting method in fuzzy series with stochastic seasonal was proposed by Guney [23]. On the basis of fuzzy c-regression, Dincer [24] proposed a novel means to predict fuzzy time series. Zhang [25] proposed a novel predicting method in fuzzy time series based on time series clustering and multiple linear regression. A method that can better describe the trend of time series is critical to improving the accurateness of the predicting model. Fuzzy relation, hidden Markov chain, and linear regression are all trying to find out the changing time series trend. There are two problems with the fuzzy series forecasting model. One is the optimization algorithms of partition used in FTS are outdated. Some new optimization algorithms such as SOA have been proved to be more accurate, but they are not used in FTS. The other is the result of fuzzy relations is not accurate enough, and linear regression performs poorly on nonlinear time series. Correspondingly, we propose two methods to solve the above two problems. Firstly, Dhiman [26] proposed SOA and it has better accuracy than traditional optimization algorithms, such as PSO. Inspired by SOA, an improved SOA(ISOA) is proposed by enhancing the convergence ability with the Powell algorithm and a random curve action. In this paper, ISOA is used to partition the domain of discourse to get more suitable intervals. Secondly, since Chen [27] proposed Extreme Gradient Boosting Tree(XGBoost), XGBoost has become one of the most popular machine learning methods and obtains impressive achievements in numerous algorithm competitions. XGBoost is a nonlinear model and has outstanding performance in a variety of tasks, which is why we use XGBoost instead of fuzzy relations to forecast the change of fuzzy membership. With the advantages of ISOA and XGBoost, this paper has the following innovations. The ISOA is put forward for better accuracy and convergence ability and applied to get appropriate intervals. It is the first application of XGBoost on fuzzy time series to forecast the change of fuzzy membership in the literature. A new fuzzy time series forecasting model(FTSOAX) is proposed based on ISOA and XGBoost. Meanwhile, FTSOAX is applied to forecast the daily confirmed of COVID-19. The rest of the paper is as follows. In Section 2, we introduce the preliminary knowledge and basic concepts of the fuzzy set and the fuzzy time series. Then, we proposed an improved seagull optimization algorithm in Section 3. In Section 4, we introduce the symmetric triangular fuzzy membership function and describe the steps of FTSOAX. Finally, we give an application to convince that FTSOAX has better performance than other models in the application of prediction of COVID-19 daily confirmed cases and summarize the contribution of this paper as a conclusion in Section 5 and Section 6, respectively.

Preliminaries

Fuzzy time series

To deal with some problems with fuzzy information, Zadeh first defined the concept of fuzzy set. Song and Chissom proposed fuzzy time series based on fuzzy sets to deal with time series with fuzzy information. In the coming, we will review the concept of fuzzy sets and fuzzy time series. Let and is universe of discourse. Fuzzy set in is the following. . is the membership function of the fuzzy set , and denotes the grade of the membership of in the fuzzy set . Let , a subset of real numbers, be the universe of discourse on which fuzzy sets is defined, and . Then is called a fuzzy time series defined on Both and are fuzzy set. Let be the fuzzy relations defined from to and , where stands for composite operation, then is said to be derived from by fuzzy relation , which can be expressed by fuzzy logic relations . The relations is called a first-order fuzzy relations defined on . Let , then a fuzzy logic relations(FLR) can be used to represent the relations between two continuous observations and . and indicate the current state and next state of the fuzzy relations, respectively. Assume fuzzy logical relations(FLRs) such that, , so these fuzzy logical relations(FLRG) can form a whole as . There is an obvious relations between and , then represents -th order fuzzy time series forecasting model, where is a multivariable function.

Classical fuzzy time series forecasting model

According to the above definition, the building of the forecasting model in fuzzy time series can be split into four steps. Taking the enrollment of the University of Alabama as the experimental dataset, we will review the classical forecasting model in fuzzy time series. Step 1. Determine the fuzzy set and fuzzy membership function according to the training set and partition the universe of discourse. First, we need to find the maximum and minimum values of the training data, and determine the scope of the universe of discourse . So the universe of discourse should be , because we can know that the minimum and maximum values are 13055 and 19337 in the data set, respectively. In addition, the range of the is often rounded down and up for the convenience of discussion and calculation. In this example, we define . The next stage is to partition the defined universe of discourse . Generally speaking, the range of the divided universe of discourse should not be too narrow, because of the fuzziness of this problem. In this case, we take 1000 as the interval length, and partition into 7 subsets, , such as . Next, we need to define the fuzzy membership function. There are many functions for us to choose from, such as trapezoidal, Gaussian, and triangular fuzzy membership functions, etc. Step 2. Establish the fuzzy relation of the sample data in order according to the training data The corresponding fuzzy logic relations can be determined according to the data of two adjacent samples. For example, fuzzy relations and can be obtained from the data from 1971 to 1972 and 1974 to 1975, respectively. In this training data that the University of Alabama’s enrollments, 21 fuzzy relations are found. Step 3. Obtain the relations matrix from all the fuzzy relations. Song and Chissom proposed the following formula to obtain the relations matrix from all the fuzzy relations. and is an operator that get the maximum element in the matrix . Step 4. Defuzzification of forecasting value according to the relations matrix and the given forecasting rule. We can get a vector by normalizing the membership of each fuzzy set, then the final forecasting value can be calculated by and the relations matrix . For example, the membership of training data in 1972 is , and is (0.6324,0.3460,0.0206,0,0,0,0) by normalizing this membership. The maximum value is the first in , and there is a maximum of 1 in the first row of the relations matrix , which falls on both the first and second columns. So the forecasting values of 1973 are the central value of the interval from the fuzzy set and . It is .

An improved seagull optimization algorithm

The swarm intelligence algorithm is a machine learning method that optimizes problems by simulating group behavior. For example, the ant colony optimization algorithm is a representative swarm intelligence algorithm. The seagull optimization algorithm(SOA) has the property of a swarm intelligence optimization algorithm. In this section, an improved seagull optimization algorithm will be proposed by improving the SOA.

Seagull optimization algorithm

Seagulls are seabirds all over the world, living in groups and using their wisdom to find and attack prey. Migration and aggressive behavior are significant features of seagulls. Seagulls migrate from one place to another according to the change of seasons, looking for the richest source of food to get sufficient vitality. During the migration, each seagull is located in a different position to avoid colliding. In a group, seagulls change their position by moving in the orientation of the best location. Dhiman proposed a seagull optimization algorithm based on seagull behavior. The core process of the algorithm is the following. Step 1. Migration (global search) SOA simulates how a flock of seagulls moves from one place to another and avoids collisions in the migration process. In addition, SOA calculates the new position with an additional variable to prevent clashes with neighbors (other seagulls). indicates the present iteration, and is the seagull’s present location. indicates a new location without collision with other seagulls and the movement behavior of seagulls in feasible space is expressed as . indicates the maximum size of iteration. The value of variable is controlled by , and linearly reduces from 2 to 0. represents the best position in all seagulls, and is the orientation of the . At the same time, in order to balance local search and global search, the parameter B is added to control them. The seagulls reach the new position by moving in the direction of the best position after avoiding overlap. indicates the new location. Step 2. Attack (local search) During the migration, seagulls can unceasingly change the speed of attack and angle. And they keep their height with their wings and weight. When attacking, seagulls move spirally in the sky. The following is the movement action in the , , and planes. is a random angle number in the range of and is the radius of each helix. represents the seagull’s attack position.

An improved seagull optimization algorithm

In reality, when migratory birds migrate, they often do not walk straight direction, mainly because migration routes require the appropriate feeding site. In addition, with hot air rising over the land, it is possible to save energy. Therefore, the birds generally along a curve of movement over the ground, not straight flying over the ocean. This is also true for seagulls. To enhance the global search capabilities of SOA, we give seagulls a random curve action in migratory behavior as follows. represents the random curve action and is a random number of . The random curve action enhances the variety of the seagull population in the global search, and its value will decrease as the size of the iteration rise. To strengthen the local search capability of SOA, we add the Powell algorithm [28] to the attacking behavior of seagulls. Powell algorithm is simple to calculate and has solid local searchability. However, the Powell algorithm is exceptionally dependent on the initial point configuration. The selection of the initial value directly affects whether the algorithm can converge to the global minimum and even causes the algorithm to fail. Therefore, the position information optimized by the SOA algorithm is used as the initial value of the Powell algorithm to avoid Powell search failure. It is obvious that the Powell algorithm will increase the time complexity of SOA. So Powell algorithm is used for local search after every iteration. We proposed an improved seagull optimization algorithm(ISOA) based on the above two aspects. The ISOA is shown in Algorithm 1 .
Algorithm 1

An improved Seagull optimization algorithm.

An improved Seagull optimization algorithm.

Experiments of improved SOA

The experiment is necessary to convince the advantage of the ISOA. Firstly, we compare the original SOA with the improved SOA on two standard test functions in Table 1 .
Table 1

Standard Test Function.

FunctionInitial rangeDimension nFmin
F1(x)=i=1nxi2[100,100]300
F2(x)=i=1n|xi|+i=1n|xi|[10,10]300
F3(x)=i=1n(0.5+xi)2[100,100]300
F4(x)=i=1n(1010cos(2πxi)+xi2)[5.12,5.12]300
F5(x)=exp(1ni=1ncos(2πxi))20exp(0.21ni=1nxi2)+20+e[32,32]300
F6(x)=(i=1n0.5ixi)4+(i=1n0.5ixi)2+i=1nxi2[100,100]300
Standard Test Function. We ran SOA and ISOA 100 times on the test function to better demonstrate their performance because their results have a certain degree of randomness. The amount of iterations there is 100, and the size of the population is 20 for each algorithm. The result of the comparison is shown in Fig. 1 . ”(a)” and ”(b)” are the results of and , respectively. ”Pre-improved SOA” is that only uses the random curve action , and ”Improved SOA” is based on random curve action and the Powell algorithm. Comparing ISOA with the original SOA, it is explicitly revealed that it has better accuracy and convergence ability. Furthermore, their main results are shown in Table 2 . In Table 2, we compare the best, mean, and standard deviation of different SOA’s results. Through comparison, we can intuitively realize that the improved SOA has a better performance than the original SOA in terms of mean, best, the standard deviation of results.
Fig. 1

Comparison between SOA and ISOA.

Table 2

Comparison between SOA and ISOA in and .

AlgorithmIndF1(x)F2(x)
Original SOABEST4.995.80
MEAN13.4910.09
STD3.191.80
Pre-improved SOABEST3.027.29
MEAN7.474.24
STD2.351.28
Improved SOABEST0.000.36E-3
MEAN9.12E-330.18E-2
STD1.51E-320.17E-2
Comparison between SOA and ISOA. Comparison between SOA and ISOA in and . To show the performance of the improved SOA, we select four optimization algorithms as a comparison of four test functions. Yolcu [29] proposed a hybrid fuzzy time series model with single particle swarm optimization(PSO). Saremi [30] proposed Grasshopper Optimisation Algorithm(GOA). Zhang [31] proposed a fuzzy time series model based on the Genetic Algorithm(GA), and a hybrid forecasting system based on the Differential Evolution(DE) is proposed by Jiang [32]. The above four optimization algorithms are compared with the original SOA and improved SOA. Functions in Table 1 are the standard test function as the evaluation criteria of these optimization algorithms. The number of iterations, population, and dimension are 300, 10, and 20, respectively. The number of dimensions of each test function is in Table 1. Fig. 1 depicts the results of the comparison of optimization algorithms. ”(a)”, ”(b)”, ”(c)”, ”(d)” represent the results of , , , , respectively. At the same time, we run these optimization algorithms 100 times on four test functions. Comparing the results among them, we take three indicators BEST, MEAN, and STD as shown in Table 3 .
Table 3

Comparison of indicators between ISOA and other optimization algorithms.

AlgorithmIndicatorF3(x)F4(x)F5(x)F6(x)
Original SOABEST5.0028.982.370.12E-2
MEAN12.029.873.470.96
STD3.210.120.250.39
Improved SOABEST0.000.35E-070.15E-30.00
MEAN0.000.13E-20.54E-30.00
STD0.000.37E-20.28E-30.00
PSOBEST2147.00170.7010.370.14E-2
MEAN4423.28230.0012.552.36
STD1331.3525.501.031.74
GOABEST2655.00226.3212.010.23E-11
MEAN14341.22318.4018.490.13
STD6530.8040.252.190.34
GABEST1256.00106.887.490.40E-3
MEAN2113.88150.128.881.51
STD416.5516.960.661.54
DEBEST93.0078.873.720.21E-7
MEAN145.8098.484.860.67E-2
STD38.408.250.430.02
Comparison of indicators between ISOA and other optimization algorithms. Through the comparison in Fig. 2 , we can find that SOA has the best performance than other optimization algorithms. Especially in ”(c)” of Fig. 2, we can notice that both the original SOA and the improved SOA fall into a locally optimal point at the 50th iteration. However, the improved SOA successfully jumped out of the local optimal point in the following iteration, while the original SOA converged at the local optimal point all the time. It shows that the improved SOA has better searchability than the original SOA. Furthermore, the results of improved SOA are better than other algorithms in terms of MEAN, BEST, STD in Table 3. Through the comparison above, we think that the improved SOA has better convergence accuracy.
Fig. 2

Comparison between ISOA and other optimization algorithms.

Comparison between ISOA and other optimization algorithms.

Computational complexity of ISOA

The complexity of an algorithm is a critical criterion for evaluating its performance. All of the optimization algorithms mentioned above require time to initialize. where is the dimension of the objective function, and represents the population size. The time complexity of SOA, PSO, and GOA is . It takes time to simulate the entire procedure, and is the maximum number of iterations. The time complexity of computing the objective function is denoted by . The GA and DE algorithms have a time complexity of , where and are the crossover and mutation operators, respectively. The time complexity of ISOA is , where represents the time complexity of running Powell’s algorithm and is a hyperparameter ranging from 0 to 1. The space complexity of the algorithm is the highest amount of space used at any given point in time. All algorithms mentioned in this work have a space complexity of .

Discussion of ISOA in terms of convergence accuracy

According to the results of the experiments, ISOA has a noticeable advantage in terms of convergence accuracy. In this section, the reason why ISOA outperforms other optimization methods in terms of convergence accuracy is explored. Population diversity is thought to be a significant component influencing the convergence accuracy of swarm intelligence optimization algorithms. Experiments [33], [34], [35] demonstrate that excellent population diversity can considerably increase the convergence accuracy of PSO, GA, and DE. The variance of positions is an important metric for measuring population diversity. The high variance of the population position can effectively avoid falling into a local optimum and improve the convergence accuracy. According to formula (7), a random curve action is added to SOA in the global search period to improve its population diversity. and represent the seagull’s positions of ISOA, SOA respectively. It is obvious that the ISOA seagull’s positions are produced by adding to the SOA seagull positions. denotes the data’s variance. It is simple to prove that the variance of the ISOA seagull’s positions is increased by adding to the variance of the SOA seagull’s positions. To graphically demonstrate ISOA’s population diversity, formula (10) is solved with ISOA and SOA, and compares the population position of their iterative process. For brevity, the dimension in formula (10) is 2, and it is solved in the range . The population size for both ISOA and SOA is 10. In Fig. 3 , ’(a)’, ’(b)’, ’(c)’ represent the population positions of ISOA and SOA with a maximum number of iterations of 100, 500 and 1000, respectively. The blue dot indicates SOA and the yellow dot indicates ISOA. The points in Fig. 3 represent the positions explored by ISOA’s SOA during the iteration. There is no doubt that the number of regions reached by the ISOA and SOA increases with the increase of the maximum number of iterations. Furthermore, it is noticeable that ISOA has reached a wider range of areas than SOA. In Fig. 4 , ’(a)’, ’(b)’, ’(c)’ represent the heat map of the SOA’s population positions at the maximum iterations of 100, 500, and 1000, respectively. ’(d)’, ’(e)’, ’(f)’ represent the heat map of the ISOA’s population positions at the maximum iterations of 100, 500, and 1000, respectively. In the heatmap, it can be more clearly found that the ISOA’s population is spread out across a larger area than the SOA’s population. It is considered that the population diversity of ISOA has been enhanced on the basis of SOA, resulting in a better convergence accuracy.
Fig. 3

Positions of ISOA and SOA over multiple iterations.

Fig. 4

Heatmap of positions of ISOA and SOA over multiple iterations.

Positions of ISOA and SOA over multiple iterations. Heatmap of positions of ISOA and SOA over multiple iterations. In the local search, the Powell algorithm is used to accelerate the rate of convergence, according to the formula (8). It’s worth mentioning that the conjugate gradient or quasi-Newton methods aren’t used to speed up local searches because they require gradient information. The Powell algorithm is simple and does not need to calculate the gradient. The convergence rate of PSO is , according to Quan [36]. The evolutionary algorithms(GA, DE) are considered to converge at the rate of [37]. The convergence rates of both PSO and evolutionary algorithms are sublinear. The exact convergence rate of the Powell algorithm is unclear, but it is thought to have a linear convergence rate [38], hence its convergence rate is better than that of the PSO and evolutionary algorithms. Because the Powell algorithm is extremely sensitive to the initial point, ISOA is employed to search for the good initial region globally. In general, the benefits of ISOA come from two factors. The first is that the random curve behavior increases population diversity, which aids in global search. The second is that the Powell algorithm improves the local convergence rate, which helps in local search.

A novel fuzzy time series forecasting model

This section will propose a new fuzzy time series forecasting model(FTSOAX) based on an improved seagull optimization algorithm and XGBoost. We employ the improved SOA to partition the universe of discourse into several intervals . The original data can then be turned into fuzzy data using the symmetric triangular fuzzy membership function. Each interval’s fuzzy membership is represented by the fuzzy data. To handle issues where fuzzy relations are insufficiently accurate, we employ XGBoost rather than fuzzy relations to anticipate the change in fuzzy membership of each interval. Finally, using the inverse operation of the symmetric triangular fuzzy membership function, the anticipated fuzzy data will be turned into real data. The procedure of FTSOAX is expressed as follows to illustrate the details of FTSOAX. Step1. Determine the parameters of the model : is the number of intervals, and its range is usually in (5,20). : represents the -th order fuzzy time series forecasting model, and its range is usually in (1,5). : represents the population number of improved SOA, and its range is usually in (5,50). : is the dimension of improved SOA, and its range is usually in (1,50) : The Maximum number of iterations of improved SOA, and its range is usually in (50,1000) : represents the Powell algorithm is executed every iterations in improved SOA, and its range is usually in (1,100) : A constant of the regularization term of XGBoost, and its range is usually in (0,1). : A constant of the regularization term of XGBoost, and its range is usually in (0,1) Step2. Partition the universe of discourse into intervals with the ISOA Let be the universe of discourse, and should be appropriately partitioned into intervals with the ISOA (Algorithm 1). The output of ISOA is split points. The centers of each interval are . We apply RMSE (Root Mean Squard Error) as the fitness function of ISOA as follows. represents the forecasted data as a result of XGBoost and is the original data. Step3. Fuzzify the training time series There is a problem with the traditional fuzzy time series forecasting model that the fuzzification and defuzzification are not inverse operations. This problem reduces the accuracy of the forecasting model. We use the symmetric triangular fuzzy membership function to deal with this problem. This fuzzy membership function can realize the accurate conversion between time series and fuzzy time series. Let as the all intervals of time series, and as the center of each interval. and are the start point of and end point of , respectively. Assuming the number of intervals is 5, this fuzzy membership function can be illustrated in Fig. 5 . This fuzzy membership function has the following three advantages.
Fig. 5

The symmetric triangular fuzzy membership function.

The symmetric triangular fuzzy membership function. : When is in the center of a certain interval , the fuzzy membership equals 1. : When is in the boundary of two intervals and , both the fuzzy membership and equal 0.5. : If the range of the actual data can be found by two non-zero fuzzy membership and and the range of actual data is , then the actual data can be obtained by the inverse operation of the fuzzy membership function and fuzzy data, which means that the defuzzification of fuzzy time series is also straightforward and simple. The symmetric triangular fuzzy membership function is as follows. After partitioning discourse into intervals, we can apply the symmetric triangular fuzzy membership function to fuzzify the training time series. Assuming , the fuzzy time series . Step4. Train XGBoost with fuzzy time series XGBoost is a highly automated model, so we don’t need to care about the details of the training process. In the training phase, as a -th order fuzzy time series forecasting model with the training data , XGBoost is trained according to the process of Algorithm2. In forecasting phase, the input of XGBoost is and the output of XGBoost is . Step5. Defuzzify the forecasted fuzzy time series The result of XGBoost is fuzzy data, so we need to defuzzify fuzzy time series to get actual forecasted time series. There are two kinds of fuzzy data. One is that the sum of two adjacent fuzzy membership is close to 1, which determines the actual forecasted data is in the range of . Other is that or is close to 1, which means the actual forecasted data is in interval or . After determining the range of the actual forecasted data, the actual forecasted data can be calculated by the inverse operation of Equation . For example, if the number of intervals is 5, and is as follows. So we can know that the actual data is in the range of , where and are the center of and , respectively. Assuming and , the actual forecasted data can be calculated by the inverse operation of Equation . The main procedure of FTSOAX is illustrated in Fig. 6 .
Fig. 6

The main procedure of FTSOAX.

The main procedure of FTSOAX.

Application

In this section, we will give an application to illustrate the performance of FTSOAX. Recently, COVID-19 has become the center of discussion all over the world. We got the dataset that the COVID-19 daily confirmed cases in 7 countries viz. USA, India, Russia, Iran, Norway, UK, and Japan from a GitHub repository of CSSE. The COVID-19 daily confirmed cases from June.09.2020 to June.22.2021 are used to be training data in the proposed model, and test data are the COVID-19 daily confirmed cases from June.23.2021 to July.29.2021. Meanwhile, several fuzzy time series forecasting models will be compared with FTSOAX to prove that FTSOAX has better performance.

Application of forecasting the COVID-19 daily confirmed in training phase

In the training phase, we execute our model FTSOAX and other models in the COVID-19 daily confirmed in 7 countries from June.07.2020 to June.22.2021, and the range of daily confirmed is . The details of the experiment are as follows. The number of intervals and the order of fuzzy time series are 7 and 2, respectively. The population and iteration of the ISOA are 10 and 50, respectively. The Powell algorithm runs once every 10 iterations. The number of iterations of XGBoost is 300. Compare FTSOAX with the following models, and we can convince that FTSOAX has better performance. Chen [39] proposed first order conventional fuzzy time series. Efendi [40] proposed first order improved weighted fuzzy time series. Sadaei [41] proposed first order exponentially weighted fuzzy time series. Naresh [42] used a fuzzy time series model combined with particle swarm optimization to forecasting COVID-19 confirmed cases. Kumar [43] proposed a novel hybrid fuzzy time series model for the prediction of COVID-19 infected cases and deaths in India. Those models run on the dataset of COVID-19 daily confirmed cases to compare FTSOAX’s training result. Firstly, according to the steps of FTSOAX in the previous section, we need to use ISOA to partition the universe of discourse . Taking the upper and lower bounds (8635,414188) of as input, we partition the into 7 intervals. We compare the partition results of the 50th iteration of ISOA with the results of the traditional partition method. is the intervals of ISOA, and represents the results of the traditional partition method. The diversity of the length of intervals of is greater than . The traditional partition method tries to make each interval the same length, making the accuracy of the prediction results worse. Because the density of samples in the same length intervals may be different. There are even no samples in some intervals. The importance of each sample is also different. For example, the sample close to the predicted time is more important than the other samples, and the volume of each sample is difficult to express. So we use ISOA as a tool for partitioning. In this way, we do not need to consider the density and importance of the sample, and satisfactory intervals can be obtained by continuous iteration. After partitioning the into 7 intervals, we can get the fuzzy time series with the help of the symmetric triangular fuzzy membership function. For example, Taking two samples as original data, the corresponding fuzzy data can be obtained. Let is the interval’s center, upper and lower limits, and both and are in the range (34667,103583). So The fuzzy data and can be calculated by Equation . Converting all-time series into fuzzy time series, we can use XGBoost to predict the trend of fuzzy membership of each interval. For example, the order of fuzzy time series is 2, so we can get . By analogy, we can get all forecasted fuzzy time series. Training XGBoost and getting all fuzzy data, the next step is the defuzzification, and its details are shown in the previous section. So we get the actual forecast data and take the RMSE of actual forecast data and original data as the fitness function of ISOA. The general flow of FTSOAX is described in Fig. 6, and the main result of FTSOAX in this application is as follows. Aiming to illustrate the performance of FTSOAX, we give two crucial performance indicators, which reveal that FTSOAX has better performance in those indicators in the training phase. We select RMSE(Root Mean Square Error), and SMAPE (Symmetric Mean Absolute Percentage Error) as the evaluation criteria. Both of them are classic evaluation criteria, and the smaller the indicators, the better. The results of all indicators are as Table 4 Comparing FTSOAX with other models in Table 4. FTSOAX’s RMSE are the best in all countries, and the second best results are behind it by in 7 countries. The SMAPE of FTSOAX are the best in all countries, and the second best results are behind it by in 7 countries. We can explicitly uncover that FTSOAX has better performance and is superior to other models in the training phase.
Table 4

Comparison between FTSOAX and other models in the training phase.

ChenEfendiSadaeiKumarNareshFTSOAX
RMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPE
USA24131.7526.8721763.2623.9824518.6132.7618879.4318.7129838.736.310236.8315.18
India15489.530.0114332.229.4613831.1624.0210915.5010.1727169.6944.776476.967.81
Russia2704.4318.691957.3611.662133.6713.751786.007.802326.9816.661240.493.62
Iran4217.3434.432556.5324.652671.3618.42674.7710.983413.1333.31735.818.08
Norway185.0267.15133.8240.69141.4952.33130.536.28139.1954.9676.7931.63
UK4111.2354.213167.4643.443202.6836.42797.5816.997875.282.921343.0114.53
Japan1323.1457.91883.250.46900.6540.46919.8124.061285.4464.82465.1315.45
Comparison between FTSOAX and other models in the training phase. We have chosen the daily confirmed cases of India as a graph to more vividly show the difference between FTSOAX and other models. The results of these models with India’s data are clearly illustrated in Fig. 7 . ”(a)”, ”(b)”,”(c)”, ”(d)”, ”(e), and ”(f)” in Fig. 7 are the results of Chen, Efendi, Sadaei, Kumar, Naresh, FTSOAX, respectively. Fig. 8 depicts the process of iteration in the FTSOAX training phase to better describe the effect of the improved SOA algorithm. The blue and red lines in Fig. 7 are the training data and the forecasted data, respectively. Comparing the result of FTSOAX and the result of other models, it is easy to perceive that the result of FTSOAX in Fig. 7 more coincides with training data than the results of other models. By comparing the figures, we can directly find that FTSOAX has better performance in the training phase.
Fig. 7

Forecasting of daily confirmed cases of India in the training phase.

Fig. 8

The process of iterations of ISOA.

Forecasting of daily confirmed cases of India in the training phase. The process of iterations of ISOA. By comparing figures and tables, there is no doubt that FTSOAX has better performance in the training phase. To make this conclusion more persuasive, we run FTSOAX and other models on test data and compare the results of these models.

Application of forecasting the COVID-19 daily confirmed in test phase

In the test phase, we execute FTSOAX, and other models in test data of the COVID-19 daily confirmed cases from June.23.2021 to July.29.2021. Test data are forecasted by trained models in the previous subsection, and as in the training phase, we chose a dataset from India as the graph to show the results of models in Fig. 9 . ”(a)”, ”(b)”,”(c)”, ”(d)”, ”(e), and ”(f)” are the results of Chen, Efendi, Sadaei, Kumar, Naresh, FTSOAX, respectively. Evaluation indicators of the test phase are shown in Table 5 .
Fig. 9

Forecasting of daliy confirmed cases of India in the test phase.

Table 5

Comparison between FTSOAX and other models in the test phase.

ChenEfendiSadaeiKumarNareshFTSOAX
RMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPE
USA29479.4965.8925142.554.5925619.557.8428688.2454.0746384.5386.4423865.8252.36
India12333.4425.2113355.9527.6713526.9728.04822.589.9911708.423.674285.907.98
Russia4450.108.614349.997.424425.848.034380.587.044360.85.784255.55.46
Iran7216.6930.766893.424.956856.6724.777765.3229.136292.9725.386092.6721.32
Norway247.7270.8156.7555.18139.4647.32131.1047.21141.9347.59133.6946.53
UK7869.7116.117727.1815.577949.3616.597184.6214.698545.8515.587108.1514.03
Japan1420.6738.531397.5136.551371.7734.501520.732.901473.8034.881332.3131.17
Forecasting of daliy confirmed cases of India in the test phase. Comparison between FTSOAX and other models in the test phase. FTSOAX’s RMSE are the best in all countries except Norway, and the second best result are behind it by in 6 countries. FTSOAX’s SMAPE are the best in all countries, and the second best results are behind it by in 7 countries. It is explicit that the result of FTSOAX is better than any result of other models through the comparison of the figure and table. The forecasted data line of FTSOAX in Fig. 9 is more coinciding with the test data line than other models. And FTSOAX is superior to other models in most indicators. The above comparisons show that FTSOAX has an absolute advantage in the test phase. Combined with the results of the training phase and the test phase, the FTSOAX is proven to have better performance in this application.

Comparison of generalization between FTSOAX and other models

FTSOAX is compared with other models in the performance of COVID-19’s confirmed prediction in the subsections above. The results show that the performance of FTSOAX is better than other models in COVID-19’s confirmed prediction. However, the results of the application are not enough to show the advantages of FTSOAX. For this reason, in this subsection, we will further discuss the generalizability of FTSOAX and other models. For simplicity of discussion, the Indian dataset is chosen as the training set, but the difference from the previous subsection is that the length of the test set is increased from 37 to 180 to explore the generalization of these models. The results of FTSOAX and other models on the extended test set are shown in Fig. 10 and Table 6 . Fig. 10 shows the SMAPE of the models from 30 days to 180 days in the test set. Table 6 displays the RMSE and SMAPE of FTSOAX and other models in the test set over multiple periods.
Fig. 10

SMAPE of FTSOAX and other models in the test set.

Table 6

Comparison between FTSOAX and other models in multiple periods.

ChenEfendiSadaeiKumarNareshFTSOAX
RMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPE
40 days12665.3625.9513613.6728.2813796.5828.634925.4210.1711212.4222.673532.317.31
100 days13736.9128.7215950.5642.9015787.3440.665100.1811.6815729.0733.483148.616.89
180 days15854.2354.1312368.0637.3612128.5334.854409.2315.6028355.2073.202785.2210.77
SMAPE of FTSOAX and other models in the test set. Comparison between FTSOAX and other models in multiple periods. The test set is divided into 3 periods of time, 40 days, 100 days, and 180 days in Table 6. The RMSE and SMAPE of these models were calculated for each period. It is clear from Table 6 that FTSOAX is superior to the other models in all periods. Further, the stability of FTSOAX is also the best among these models. Its 40 days and 180 days SMAPE differ by 3.46, which is smaller than that of other models. Fig. 10 shows the change in SMAPE of FTSOAX and other models as the length of the test set varies. Intuitively, the line indicating FTSOAX is below the other lines at all times, and it fluctuates less than the other lines. It is clear at a glance that the accuracy and stability of FTSOAX are superior to other models in the Indian dataset. Analyzing Table 6 and Fig. 10, it is obvious that FTSOAX is generalizable and its accuracy is always better compared to other models in this case.

Comparison of robustness between FTSOAX and other models

Robustness is a very important role in evaluating the quality of a model. In this subsection, the robustness of FTSOAX is compared with other models by adding noise and outliers to the training set. As in the previous subsection, to simplify the discussion, daily confirmed of COVID-19 in India is selected as the data set. The noise is added to each element of the training set, and the noise follows a Gaussian distribution with mean and variance . There is a hyperparameter whose value is in the range [0,1] and it controls how much noise is added. The and of the Gaussian distribution of the noise is 0 and the variance of the training set in the actual experiment, respectively. The training set is severely deformed, which means that the experiment is of little significance when is greater than 0.10. Therefore was decided to be 4 values [0.01,0.02,0.05,0.10]. Table 7 and Table 8 present the results of the models on the training and test sets, respectively. Fig. 11 illustrates the SMAPE of the models under different in the test set with added noise. By comparison, it can be found that the SMAPE and RMSE of models increase with the increase of noise. FTSOAX outperforms other models in the training set and test set, judging from the tables and picture, which indicates that FTSOAX has better resistance to noise than other models in this case.
Table 7

Comparison between FTSOAX and other models in the training set with noise.

ChenEfendiSadaeiKumarNareshFTSOAX
factorRMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPE
0.0115489.5030.0114332.2029.4613831.1624.0210756.6110.1916124.8234.026230.188.18
0.0215248.7128.6814642.2429.5914169.2227.8911675.7611.0628239.3148.218342.719.32
0.0515133.2726.9514888.0432.1514316.3321.6512475.5114.4514633.6530.6110135.6411.80
0.1016848.1934.9316035.7634.5716503.1029.7214119.5520.6318575.2336.9711508.1517.34
Table 8

Comparison between FTSOAX and other models in the test set with noise.

ChenEfendiSadaeiKumarNareshFTSOAX
factorRMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPE
0.0112333.4425.2113355.9527.6713526.9722.915064.7110.1610666.2918.684335.918.66
0.0210711.8721.5711302.5622.8311344.0718.195548.2111.1317027.7115.854826.609.20
0.057270.7014.145970.8211.409978.6619.976715.5913.1810462.7921.625122.429.96
0.1012909.9833.5113438.2535.407973.4328.007286.1115.639179.9533.756537.0214.38
Fig. 11

SMAPE of FTSOAX and other models in the test set with noise.

Comparison between FTSOAX and other models in the training set with noise. Comparison between FTSOAX and other models in the test set with noise. SMAPE of FTSOAX and other models in the test set with noise. Adding outliers to the dataset is a common way to check the robustness of a model. Outliers are added to the training set according to formula (12). means to turn an element of the training set into times. denotes a function that selects random subscripts from 0 to . In a word, there are random elements in the training set that become times themselves. In actual experiments, equal to 3 is suitable, because there is no difference in the results of models with smaller , and the results of models will become very poor with larger . is [3,10,20], indicating that 3,10,20 elements are randomly selected from the training set. Table 9 and Table 10 show the results of the models on the training and test sets with outliers, respectively. Fig. 12 illustrates the SMAPE of the models under different in the test set with outliers. As can be seen from the tables and figure, FTSOAX outperforms other models under all . Further, FTSOAX is more stable. Its SMAPE in the test set is the smallest difference between and , which is 8.91. The results of models prove that FTSOAX outperforms other models in data with outliers in this case.
Table 9

Comparison between FTSOAX and other models in the training set with outliers.

ChenEfendiSadaeiKumarNareshFTSOAX
MRMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPE
337757.2848.6219953.629.5719657.6125.1518743.3913.046710.9345.5212111.1110.24
1052549.8958.0831439.5131.8331568.3528.3530820.1821.6796610.8797.1320495.9118.10
20111165.48101.9344516.7934.1944900.9135.2347600.9527.750624.8972.0733641.8518.74
Table 10

Comparison between FTSOAX and other models in the test set with outliers.

ChenEfendiSadaeiKumarNareshFTSOAX
MRMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPERMSESMAPE
345763.4782.5513409.3738.3812710.1136.076109.7418.1122474.8463.216245.3316.13
1054084.2586.2313677.9538.3112676.2436.206076.8120.62116712.41140.227996.8119.57
20117202.3139.9121044.8553.119899.5452.5314654.330.1461384.98112.2812075.7425.04
Fig. 12

SMAPE of FTSOAX and other models in the test set with outliers.

Comparison between FTSOAX and other models in the training set with outliers. Comparison between FTSOAX and other models in the test set with outliers. SMAPE of FTSOAX and other models in the test set with outliers. The perspective of the noise experiment and outlier experiment reveals that FTSOAX has better fault tolerance and robustness compared with other models to a certain extent in this case.

Conclusions

In this paper, a novel fuzzy time series forecasting model(FTSOAX) was proposed by combining ISOA and XGBoost, and it is an extension of the fuzzy time series model. It is a worthwhile attempt to apply it to COVID-19 prediction. Based on a random curve action and the Powell algorithm, we enhanced the current SOA and proposed the improved SOA(ISOA). The ISOA is used to partition the universe of discourse into suitable intervals. Furthermore, it is the first application of XGBoost on fuzzy time series to forecast the change of fuzzy membership in the literature. We compared FTSOAX and the other models with an application on COVID-19 to demonstrate that FTSOAX outperformed them. The results of the experiments reveal that FTSOAX beats other models in terms of forecasting COVID-19 daily confirmed cases. Finally, there are some issues, such as time consumption, that require more investigation and development.

CRediT authorship contribution statement

Sidong Xian: Ideas; Formulation or evolution of overarching research goals and aims; Conceptualization; Writing-review and editing; Supervision; Project administration; Funding acquisition. Kaiyuan Chen: Conceptualization; Methodology; Creation of models;Formal analysis; Data curation; Computational; Writing-Original draft preparation. Yue Cheng: Testing of existing code components; Verification.

Declaration of Competing Interest

The authors acknowledged that there is no conflict of interest in this work.
  2 in total

1.  A novel hybrid fuzzy time series model for prediction of COVID-19 infected cases and deaths in India.

Authors:  Niteesh Kumar; Harendra Kumar
Journal:  ISA Trans       Date:  2021-07-06       Impact factor: 5.911

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.