Literature DB >> 35992191

Tensor based stacked fuzzy neural network for efficient data regression.

Jie Li1, Jiale Hu1, Guoliang Zhao1,2, Sharina Huang3, Yang Liu1.   

Abstract

Random vector functional link and extreme learning machine have been extended by the type-2 fuzzy sets with vector stacked methods, this extension leads to a new way to use tensor to construct learning structure for the type-2 fuzzy sets-based learning framework. In this paper, type-2 fuzzy sets-based random vector functional link, type-2 fuzzy sets-based extreme learning machine and Tikhonov-regularized extreme learning machine are fused into one network, a tensor way of stacking data is used to incorporate the nonlinear mappings when using type-2 fuzzy sets. In this way, the network could learn the sub-structure by three sub-structures' algorithms, which are merged into one tensor structure via the type-2 fuzzy mapping results. To the stacked single fuzzy neural network, the consequent part parameters learning is implemented by unfolding tensor-based matrix regression. The newly proposed stacked single fuzzy neural network shows a new way to design the hybrid fuzzy neural network with the higher order fuzzy sets and higher order data structure. The effective of the proposed stacked single fuzzy neural network are verified by the classical testing benchmarks and several statistical testing methods.
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2022, Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Entities:  

Keywords:  Extreme learning machine (ELM); Random vector functional link network (RVFL); Tensor stacked fuzzy neural network (TSFNN); Tensor-based type-2 extreme learning machine (TT2-ELM); Tensor-based type-2 random vector functional link network (TT2-RVFL)

Year:  2022        PMID: 35992191      PMCID: PMC9382627          DOI: 10.1007/s00500-022-07402-3

Source DB:  PubMed          Journal:  Soft comput        ISSN: 1432-7643            Impact factor:   3.732


Introduction

The random vector functional link (RVFL) and extreme learning machine (ELM) are two popular randomized single layer forward learning networks, which provide us a unified framework for both regression and multi-class classification with single layer. Then the semi-supervised RVFL and ELM networks can be merged into a joint optimization framework, it shows that the algorithm is efficient in moderate scale data classification (Peng et al. 2020). The parameters could be regularized when ridge regression is used (Yildirim and Revan Özkale 2019). When singular value decomposition (SVD) is used for algorithm iterative solution searching, the SVD update algorithm scales better and works faster than SVD computed from scratch (Grigorievskiy et al. 2016). Multi-label learning method could also use the multi-label radial basis function neural network and Laplacian ELM (Xu et al. 2019), in this algorithm, clustering algorithm determines the number of hidden nodes, and the center of the activation function could be determined by the data itself, then the output is solved by a Laplacian ELM. Inspired by biological intelligent systems, bio-inspired learning model blooms a lot recently (Huang and Chen 2016; Alencar et al. 2016; Christou et al. 2019), it can be applied to many areas, such as, anomalous trajectory classification (Sekh et al. 2020), long-term time series prediction (Grigorievskiy et al. 2014), T-S fuzzy model identification (Wei et al. 2020), dictionary learning-based image classification (Zeng et al. 2020), anomaly detection (Hashmi and Ahmad 2019), HRV recognition (Bugnon et al. 2020), energy system (Yaw et al. 2020), mislabeled samples detection (Akusok et al. 2015), concept drift detection (Yang et al. 2020), etc. Although ELM and RVFL have been applied in many fields, their disadvantages are found during verification. ELM has the extremely fast training speed, however, its performance is not stable. For example, to the structural risk minimization approach, ELM easy performs worse in terms of stability, generalization performance and sparsity perspectives. ELM also tends to be over-fitting. The performance of RVFL is more stable than ELM, and the training speed is slower than ELM. In addition, the over-fitting risk of RVFL could be reduced by the enhancement nodes of RVFL. Generalization performance is the main concern for the learning algorithms, balancing computational complexity and generalization ability have been extended via ELM (Ragusa et al. 2020); with the designed data and modeled parallel ELMs, large-scale learning tasks could be tackled by ELM (Ming et al. 2018), moreover, a tradeoff should be made among efficiency and scalability, the algorithm should have complementary advantages. With the aid of graph learning and adaptive unsupervised/semi-supervised clustering method, flexible and discriminative data embedding could be achieved (Zeng et al. 2020; Zheng et al. 2020). By using the regularized correntropy criterion and half-quadratic optimization technique, convergence speed and performance are both showed superiorities than the original (Yang et al. 2020), and the robust type algorithm has been studied (Yang et al. 2020). When inverse-free recursive algorithm is used to update the inverse of the networks’ Hermitian matrix, efficient inverse-free algorithm is designed to update the regularized pseudo-inverse, which has been proposed for ELMs (Zhu and Wu 2020). To the above work, the performance of ELM is enhanced via adding the regularization method which can also be used to avoid the over-fitting risk. The regularization methods are the commonly used and recognized strategy to improve the performance of artificial neural networks (ANN). However, the effect of the regularization method is limited for the performance, when artificial neural networks are used to handle the complex and huge data. In order to make the ANN models suitable for dealing with the complex data structures, two strategies can be considered. The first strategy is to add methods that can optimize the ANN structure and mechanism. These group of methods are the parameter (i.e. hidden nodes numbers) selection method, hidden layer pruned method, fuzzy structure, and so on. When big data environment is encountered, a fast parameter selection scheme for modeling the large amount of data is needed, alternating direction method and maximally splitting method could be applied to the algorithms to minus the number of the sub-model and coefficients training (Lai et al. 2020). Concerning the credit probability for network output, probabilistic output from the original architecture of ELM is proposed, iterative way of learning is eliminated, and the merits of ELM is preserved (Wong et al. 2020). Using Bayesian inferences, multiple-instance learning-based ELM has proven to be efficient in classification problems (Wang et al. 2020). Optimally pruned ELM (Miche et al. 2010) is presented to both regression and classification problems, the proposed algorithm could counter the effect of noise. To move forward, an regularization penalty applied to the optimally pruned ELM, and a double-regularized ELM using LARS and Tikhonov regularization is proposed (Miche et al. 2011). Missing data case for the regression problem is studied (Yu et al. 2013). When training sample selection method is designed based on the fuzzy C-means clustering algorithm, and the proposed small training samples selection-based hierarchical ELM could reduce the computational time (Xu et al. 2020). Random vector functional link networks (RVFL) (Zhang and Suganthan 2016) could also use the techniques mentioned above, such as ridge regression (Zhang and Suganthan 2017), and its extended version, the new learning paradigm is named as RVFL plus (RVFL+) (Zhang 2020), it has been used in neuro-imaging-based parkinson’s disease diagnosis (Xue et al. 2018; Shi et al. 2019). The type-2 fuzzy set is commonly used for their excellent ability in modeling uncertain information. The generalization performance and uncertain modeling ability of ELM-like models can be extended via the type-2 fuzzy set. Motivated by the ELMs and RVFLs, generalized Moore–Penrose inverse and triangular type-2 fuzzy sets are used to extend the ELM, and tensor-based ELM has been proposed (Huang et al. 2019). RVFL network has been also expanded to tensor case (Zhao and Wu 2019), the type-reduction method for general type-2 fuzzy sets are removed in this type of network. The second strategy is to improve the structure of ANN. For example, the deep network technology has been studied maturely, the performance of ANN could be significantly enhanced by the multi-hidden layers. It is noted that the computational complexity is increased, and the training speed is slowed gradually with the increasing of the hidden layers. In the recent decade, deep networks have been used in the control system of the real world. The uncertain information and big data problem are existed in control system, especially in the robot control system. In Lu et al. (2019), residual errors and state variables of system are handled via RNN model. The time-varying under-determined linear system with double bound limits for real robot system are simulated and verified. Li et al. (2021), neural networks are employed to test the advantages and disadvantages of robot parameter calibration technology. Moreover, the kinematic parameters of robot arms are also modeled via neural network (Li et al. 2022). Six regularization schemes combine with least square method and LM algorithm are used to solve over-fitting problems, and the robot calibration parameters are identified with these techniques. Besides the usual fuzzy sets, the pythagorean fuzzy set (PFS) is also used in control problems that is proposed in Ejegwa et al. (2022), which can be used to compute the correlation between PFSs, modeling of the complex control system and pattern recognition via fuzzy theory is the aim of this work. Feng et al. (2018), drive-response memristive neural networks (MNNs) is used to dealt with problem of the time delay, control widths and rest widths in intermittent control system. The problem of asymptotic synchronization of MNNs is solved via quantized intermittent control and weighted double-integral inequalities. The strategy of constructing network with composite structure via serval neural networks is another way to improve the structure of ANN. At present, the mainstream method is to divide a part of neurons into the network’s hidden layers to train other networks. These networks are used instead of the original fully connected part in the main network. This way has been used in single or deep networks. However, the performance of this composite structure has limited the ability of the main network. As everyone knows, the tensor is the higher order structure, which can be unfolded into serval matrixes. In a word, tensor space is a higher order array space, which includes matrix space. Therefore, tensor structure can be used to fuse serval lower order structure. Huang et al. (2019) and Zhao and Wu (2019), the type-2 fuzzy set is combined into the 4-D tensor space. Motivated by the above-mentioned work, we have noticed that the type-2 fuzzy sets and tensor structure provide a new way to model the complex data, whether the ELMs, RVFLs or the neuro-fuzzy systems could be used under the tensor structure. Our target is to unveil the links or laws behind the data, and a new tensor-based stacked neural network for efficient data regression is studied. To get the merits of the algorithms, a good way is to fuse the algorithms into one frame, the balance of performance and structure simplicity would be achieved. The tensor-based stacked neural network can be considered as a new composite structure of merging several models, which also can be treated as a novel neural network fused structure. To the best of our knowledge, little related research work has studied with type-2 fuzzy sets under tensor structure. To fuse different concept and techniques into the algorithm, it is inevitable to extract the different aspect of the data, then different view results of the data can be obtained, then the original proposed algorithms could be used to minimize the testing error. Go back to the type-2 fuzzy sets, it could map the data with different parameter-specified fuzzy membership functions with at least three type-1 fuzzy membership functions, then the multi-view on the data could be obtained. A question follows this is how to fuse the results into one data structure, tensor is the suitable choice for these types of learning methods, this is the motivation of the work. The main contributions of this work are provided as below:The structure of the rest of paper is as follows: Sect. 2 introduces the three algorithms that are used by stacking system, Sect. 2.1 introduces the tensor-based type-2 RVFL, Sect. 2.2 presents the tensor-based type-2 ELM, and Sect. 2.3 is the introduction to the TROP-ELM, it is used to compare algorithms’ performance. In Sect. 3, the structure of tensor-based hybrid single fuzzy neural networks, that is, a stacked single fuzzy neural network is presented. Simulation results and discussions are given in Sect. 4. Finally, conclusions are inferred in Sect. 5. A new neural network stacked model based on tensor structure is proposed. Three neural network structures, including two type-2 fuzzy networks (TT2-ELM and TT2-RVFL) and an ordinary neural network (TROP-ELM) which can be considered as a type-1 fuzzy network, are the members of the proposed tensor-based fuzzy stacked neural network (TSFNN). They are fused into a 3-D tensors as members of the stacked neural network. The hidden layer of three member networks of the proposed TSFNN are stacked in a 3-D tensor. In this network stacking process, the mapping results of the three member networks are fused into the generated 3-D tensor or type-2 fuzzy set. The 3-D tensor combines the contents of one type-1 fuzzy set and two type-2 fuzzy sets, which can be considered as a new type-2 fuzzy set. Therefore, TSFNN is a new high-order fuzzy modeling method. The tensor regression is realized via a tensor unfolded algorithm. The 3-D tensor of the proposed model is unfolded into three matrices. The information of tensor space can be reduced into three matrix spaces. The matrix regression with Tikhonov regularization of three matrices can be used to solve the consequent problem. In this process, the problem of tensor space is solved in matrix space, and the Tikhnonv regularization method is used in the tensor regression process.

Preliminary

In this section, the tensor-based type-2 RVFL, tensor-based type-2 ELM and Tikhonov regularized OP-ELM are introduced.

Tensor-based type-2 RVFL

The RVFL usually adopts the activation functions to construct the network, for example the Radbas (y = ) functions. In the hidden layers of the network, where y and s are defined as the output and input, respectively. The enhancement nodes of tensor-based RVFL are replaced with IT2 fuzzy sets. The structure of tensor-based type-2 RVFL (TT2-RVFL) is represented in Fig. 1.
Fig. 1

Structure of the TT2-RVFL network

Structure of the TT2-RVFL network TT2-RVFL has three layers, input layer is the first, the second is hidden layer which includes fully connected part, that is the green nodes in Figs. 1 and 3, and enhancement part, the third is output layer. To the hidden layer, the enhancement part is extended via interval type-2 fuzzy set and the expanded enhancement part is stacked into a 3-D tensor structure. Activation function Radbas of RVFL is extended to interval type-2 fuzzy set IT2Radbas, and the extended RVFL is constructed using IT2Radbas.
Fig. 3

Structure of the TT2-ELM network

Figure 2 shows the standard formulation of the IT2 fuzzy set. IT2 fuzzy set is constituted via lower membership function (LMF) and upper membership function (UMF), the secondary membership function is a constant. In the light of the front view of the Fig. 2, the upper line represents UMF and lower one represents LMF. The area between LMF and UMF is the footprint of uncertainty (FOU) that includes the information of IT2 fuzzy set. The incentive function is named IT2Radbas, and the interval type-2 fuzzy set could be constructed with this type of activation functions.
Fig. 2

Structure of the IT2 fuzzy set

Structure of the IT2 fuzzy set The membership function (MF) of type-2 fuzzy sets in TT2-RVFL is defined as follows:where . Given testing data set , where , and = , and . For a lower MF matrix can be structured with the following matrices:where is bias and (; ) is input weights, respectively, and they are randomly generated. By the definition of the IT2 fuzzy sets’ lower MF, the relationship between input and expected output could be approximated by the lower MF matrix. The principal MF matrix and the upper MF matrix can also be structured similarly, with the following formulas:It also forms the upper membership functions. The UMFs are used to filling tensor. The slices of tensor are:where is a random generated weighted vector constructed by tensor.

Remark 1

The uncertain weight method (Runkler et al. 2018) is applied to compute the principal MF matrices , which reflects the impact of MF value on the whole defuzzification result of the set. The mapping results are obtained as follows:where measures the uncertainty of lower membership value on type reduction results through the formula . The uncertainty weight method expands the simple method to obtain the mean value of upper and lower MF values. The legible output can be given as follows:Formula (4) shows that for , the defuzzification results increase linearly with uncertainty; the defuzzification results are less than linear when ; the defuzzification results would be greater than linear when .

Remark 2

IT2 fuzzy set can be used to expand the enhancement node of RVFL. The original enhancement part of RVFL is constructed by type-1 MF or type-1 fuzzy set. In a word, RVFL can be considered as a type-1 fuzzy network. IT2 fuzzy set can extend RVFL via updating the enhancement node. In this process, type-1 fuzzy set is extended to IT2 fuzzy set. After the extension of IT2 fuzzy set, the enhancement part of RVFL has interval type-2 fuzzy structure. When formula (3) is used in the defuzzification step, which represents the above extension process of IT2 fuzzy set to RVFL. The new RVFL with the type-2 fuzzy set can be named as interval type-2 random vector function link (IT2-RVFL) network. Finally, a 3-tensor is established by the three foregoing membership functions , and . Thereinafter, because of the relevant usage of in tensor equations, will be changed to another capital letter in the next section. It can be known from the relevant content of the tensor equation that in the enhanced node of TT2-RVFL, the weighting matrix is , so the weighting matrix of TT2-RVFL can be defined as:To TT2-RVFL, the output model can be fused into one matrix by the following equation:where is called the equilibrium coefficient of TT2-RVFL, is denoted as mapping consequences of non-linear interval type-2 activation function, is the weighted matrices for the enhanced part, matrix ; define (i = ) as a weight vector from input layer to the intensification nodes are stochastically generated, in such way, the activation functions in , are not fully saturated; is the input matrix that is structured by input samples, and = [] is used to denote the unresolved input weight matrix. Structure of the TT2-ELM network

Remark 3

In the light of Eq. (6), when , TT2-RVFL will be degenerated to RVFL; when , TT2-RVFL is a mixed model of tensor-based extreme learning machine and RVFL; when , TT2-RVFL will be transformed to tensor-based extreme learning machine.

Tensor based type-2 ELM

The TT2-ELM was first proposed in Huang et al. (2019). The advantage of tensor structure is that the information of quadratic MF can be contained and modelled directly into one high dimensional array. This characteristic can avoid the type reduction operation in the process of type-2 fuzzy reasoning. Therefore, the tensor structure can seamlessly embed type-2 fuzzy sets into ELM. Figure 3 intuitively shows the structure of TT2-ELM. Obviously, TT2-ELM is a single hidden layer feed forward neural network. TT2-ELM has three layers, input layer is the first, hidden layer which is reformed via the triangular type-2 fuzzy sets (Huang et al. 2019) is the second layer, and the triangular type-2 fuzzy sets are used to form the 3-D tensor, output layer is the third layer. For the specified test dataset, which is expressed as , where = (, ) represents inputs, and represents outputs. The mathematical model of TT2-ELM is formed as:where can be reshaped by with a specified size , the is regression tensor’s dimension, N is the training patterns, is the value of output weight, is the output matrix. Obviously, when , Eq. (7) degenerates into matrices case. According to Huang et al. (2019) (Theorem 1), for multi-linear system (7), if there exist any , then the multi-linear system (7) is solvable, and the solution of the Eq. (7) is:where is a solution of . The resultant in formula (8) is the solution of TT2-ELM. For the Eq. (7), if cannot be obtained via the analysis, then the multi-linear system (7) is unsolvable, and the Eq. (7) has a minimum norm solution alternatively. The following tensor equation can be obtained.The gain tensor in Eq. (9) can be considered as ‘square tensor’, and the solution of tensor equation can still be obtained. Based of theory of tensor and kernel, the following equation can be determined.where is a random tensor that can satisfy formula (10). As long as this tensor has a suitable order, it can be established. A generally accepted condition is gratifies . This condition indicates that is included in the null space of . Therefore, a general solution can be obtained as follows:where represents the 1-inverse of . The minimum solution for the tensor equation consists of two parts, one of which comes from the null space of equation . Note that the tensor equation that is formulated by (7) and (11), we have:For Eq. (12), if , is the minimizer. In the light of Corollary 2.14(1) Behera and Mishra (2017), the existence of makes:Subsequently, the least squares solution of Eq. (13) is shown as below:Equation (14) is the minimum norm solution of Eq. (7). Implementation steps of TROP-ELM

Tikhonov regularized OP-ELM

In Miche et al. (2010), Yoan Miche et al. first proposed OP-ELM, it is an improvement of ELM. Figure 4 shows the Tikhonov regularized ELM (TROP-ELM) Miche et al. (2011). First, the OP-ELM network uses three different types of activation functions to form the kernel, and the assembled kernel is able to improve robustness and generality. For the initial condition of the ELM algorithm, Sigmoid kernel is used in its structure, while OP-ELM could use linear kernel, Sigmoid kernel and Gaussian kernel. Second, compared with the original proposed ELM, the multi-response sparse regression algorithm (MRSR) and the verification method leave one out (LOO) are also introduces in OP-ELM. The main role is to prune irrelevant variables by pruning the related neurons of SLFN constructed by ELM. The MRSR algorithm can rank neurons according to the usefulness of neurons, and the actual pruning technique is performed by leave one out validation results.
Fig. 4

Implementation steps of TROP-ELM

The structure of the proposed tensor-based stacked fuzzy neural networks (TSFNN)

Tensor based hybrid single fuzzy neural networks design

In the previous section, three single fuzzy neural networks, that are, the TT2-ELM, TT2-RVFL and TROP-ELM are briefed. In this section, the three networks are stacked into one network, and a regression method that is based on the three regression results are introduced. The architecture is shown in Fig. 5.
Fig. 5

The structure of the proposed tensor-based stacked fuzzy neural networks (TSFNN)

The framework of tensor-based stacked single fuzzy neural network that is design based on three single fuzzy neural networks (TT2-ELM, TT2-RVFL and TROP-ELM), while the three algorithms have been proposed by researchers. The layer 2 is the hidden layer that is constituted by the three algorithms to construct a tensor structure. The tensor structure could be constructed in layer 3, then the tensor is unfolded into three different matrices for each single network. The final regression uses a simple normalized scalar weighting, which is the part to be optimized in the future. The unfolding of tensor could use the definitions from (Yu et al. 2019), which are appended as Definition 1.

Definition 1

()-unfolding Consider a tensor with N-dimensional, and the tensor follows the dimension that is a k-partition of , where and . The (, )-unfolding of , denoted by , is a tensor with the size of such that the -th entry of is the entry of at the position (), where = [], and is the linear index of a multi-dimensional array, see, e.g., Baranyi (2016) and Baranyi et al. (2014). For example, = [] = [], = [], then is a 2-partition of , where = [] = [], = = 3.

Remark 4

By comparing the structures of TT2-RVFL, TT2-ELM and TROP-ELM, we have made a few modifications to TT2-RVFL and TT2-ELM. The original activation function of TT2-RVFL and TT2-ELM is Sigmoid function, while TROP-ELM uses linear activation function and Gaussian activation function in addition to Sigmoid function. In order to ensure that their structures correspondingly and facilitate the composition of tensor structures, linear activation function and Gaussian activation function are also added to TT2-RVFL and TT2-ELM. After adding two distinct activation functions, TT2-ELM and TT2-RVFL obtained better performance than the original proposed ones based on our test. The 3-tensor generates three matrices, and the row number of matrices is just equal to sample number. Three matrices are denoted by and , respectively, and , . The three matrices, and could reconstruct the tensor easily, that is, is the first aspect of A( : ,  : , 1), which is the mapping results from the TT2-RVFL. is the second aspect of A( : ,  : , 2), which is the mapping results from the TT2-ELM. is the three aspect of A( : ,  : , 3), which is the mapping results from the TROP-ELM. To the type-2 fuzzy networks, the LMF, UMF and defuzzification of the secondary membership function are used to solve the consequent parameters’ learning problem. To the regression layer, , and are three matrices that is unfolded from the tensor, regression equation can be denoted as , where , and . To make the network perform best, first, the error of each network is calculated, in addition, root mean square error is used as the measurement standard. Second, the best-performing network is found out, which is used to train the other two network errors. After processing, the network is recorded as . Finally, the output is averaged based on , that iswhich is the average results of the three type-2 fuzzy networks, the MF values are obtained from lower MF and upper MF, since the matrices are the unfolded results from the tensor. And the defuzzificaiton result could be calculated from secondary MFs, it could also be used for the stacked SFNN. The unconstraint regression result of could be denoted as . In statistics, this method is known as ridge regression, it is associated to the Levenberg–Marquardt algorithm and Andrey Tikhonov method to solve the regularization of ill-posed non-linear least-squares problems. Suppose that for a known matrix and vector , a vector is expected to be found, such thatIn most of the cases, ordinary least squares estimation leads to an overdetermined (over-fitted), or more often an underdetermined (under-fitted) system of equations. Therefore, in solving the ill-posed inverse-problem, the inverse mapping operators that has the undesirable tendency of amplifying noise (The eigenvalue is maximum in the reverse mapping and the singular value is minimum in the forward mapping). Moreover, every element of the reconstructed version of is implicitly nullified by ordinary least squares, instead of taking a model to the . For the purpose of the minimize residuals sum of squares, and the particular solution also satisfies some suitable qualities, a regularization term which can be added to this primary minimization problem, it can be succinctly scripted as follows:where is the Euclidean norm, is an appropriately selected Tikhonov weighting matrix. Under many circumstances, the matrix is selected as a multiple of the character matrix , By regularization, solutions with smaller norms can be found (Ng 2004). At other times, in the event that the fundamental vectors are considered primarily uninterrupted, a low-pass operator can be accustomed to strengthen flatness. This canonicalization enhances the conditioning of the problem, which leads to a straightforward numerical solution. An approximated solution is signified through , which is presented by:the individual algorithm of the stacked single fuzzy neural network could use the regularized results for the tensor unfolded structure’s learning method.

Simulation results for the datasets

In this section, the UCI benchmark dataset and the other four actual datasets are tested to evaluate the performance of this method. Among all the simulations, mean squared error (MSE), mean absolute error (MAE) and root mean square error (RMSE) were utilized to assess the performance of the proposed TSFNN and comparison methods. MAE, MSE and RMSE are defined as below:where is the predicted signal, is the target signal, and N is the length of the testing sequence. In order to test performance of the proposed TSFNN, ten different models are used for comparison. They are tensor-based type-2 random vector functional link (TT2-RVFL) network (Zhao and Wu 2019), tensor-based type-2 extreme learning machine (TT2-ELM) (Huang et al. 2019), optimally-pruned extreme learning machine (OP-ELM) (Miche et al. 2010), optimally-pruned extreme learning machine with Tikhonov regularization (TROP-ELM) (TROP-ELM) (Miche et al. 2011), interval type-2 fuzzy neural network (IT2-FNN) (Juang et al. 2010), evolving type-2 quantum fuzzy neural network (eT2QFNN) (Ashfahani et al. 2019), regularized extreme learning machine with biased drop-connect and biased dropout (BD-ELM) (Lai et al. 2020), recurrent neural network with Levenberg–Marquardt (LM) algorithm (RNN-LM) (Moré 1978), recurrent neural network with Broyden-Fletcher-Goldfarb-Shanno (BFGS) algorithm (RNN-BFGS) (Head and Zerner 1985) and long short-term memory (LSTM) (Hochreiter and Schmidhuber 1997). It is noted that RNN-LM and RNN-BFGS are recurrent neural network (RNN) model with different training algorithms. RNN-LM and RNN-BFGS are implemented by Pyrenn Toolbox (Atabay 2016). The key parameters settings of comparison algorithms are listed below. The eleven algorithms uses the same datasets which have the same normalization process. 70% of each dataset is used for training and 30% is used for testing. The hidden nodes of all comparison algorithms are set as 20. The input weight and bias of each comparison algorithm are the same which randomly generated. For TT2-RVFL, the balance weighting factor between enhancement node part and hidden layer is set as 0.9. The Tikhonov regularization is used in matrix regression and regularization tuning parameter is set as 0.01. For TT2-ELM, Tikhonov regularization tuning parameter is set as 0.01. For OP-ELM, the parameters design is the same as the original paper (Miche et al. 2010). For TROP-ELM, the parameters is the same as OP-ELM and Tikhonov regularization tuning parameter is set as 0.01. For IT2-FNN, the parameters setting is similar in Juang et al. (2010) and Tikhonov regularization tuning parameter is set as 0.01. For eTQ2ELM, the learning rate is set as 0.01, and the construction of interval type-2 quantum fuzzy set is similar in Ashfahani et al. (2019). For BD-ELM, the parameters of the ELM part are the same as above. BD-ELM uses the biased dropconnect regularization and biased dropout regularization. The higher and lower group drop possibility of the biased dropconnect regularization (Cao et al. 2015) are set as 0.9 and 0.7, respectively. The parameter settings of biased dropout regularization (Poernomo and Kang 2018) are the same as biased dropconnect regularization. For the RNN-LM and RNN-BFGS, RNN structure is constructed via Pyrenn Toolbox (Poernomo and Kang 2018). RNN structure has 20 neurons and 5 hidden layers. The maximum number of iterations is set as 100. The stopping condition is set as , which is the residual between the original value and the predicted value. The LM algorithm and BFGS algorithm are invoked from Pyrenn Toolbox, directly. For LSTM, the learning rate is set as 0.01, the maximum number of iterations is set as 100, and the stopping condition is set as . For the proposed TSFNN, the Tikhonov regularization tuning parameter is set as 0.01. The parameters of its members is the same as the settings of TT2-RVFL, TT2-ELM and TROP-ELM. All experiments are performed on a computer with AMD Ryzen 7 4800U with Radeon Graphics 1.80 GHz and 16 GB RAM. The result of total 5000 times were performed on each data set. Attributes of the testing datasets The Abalone, Airfoil self noise, Auto-Mpg, Bank, Concrete slump, Diabetes, Delta aileron, Delta elevators, Energy efficiency and Wine quality white datasets could be download via the following URL: https://archive.ics.uci.edu/ml/datasets.php

Regression problems

In this section, ten realistic world regression problems are used for testing. Abalone is a dataset that is used to predict abalone’s age by physical measurements, which includes the whole weight, shucked weight and viscera weight of abalone in Tasmania, and 4447 samples with 9 attributes is included in the dataset. Airfoil Self-Noise dataset is obtained from a series of aerodynamic and acoustic tests on two-dimensional, and three-dimensional airfoil blade profiles in an anechoic wind tunnel, it contains 1503 samples and 6 attributes. The data set of Auto-MPG assembles miles each gallon data with dissimilar car brands, it contains 392 samples with 8 attributes. The bank dataset simulates the customer’s patience who select their favored services in the bank according to 8 factors, for example residential area, distance, virtual temperature regulating bank option and so on, it contains 8192 samples with 8 attributes. Concrete Slump dataset contains information about the factors that affect slump flow of concrete, it includes 103 samples with 11 attributes. Diabetes is a dataset that investigates the reliance of the grade of serum C-peptide on various factors, it can be used to measure residual insulin secretion patterns, 768 samples with 4 attributes are included in the dataset. Delta ailerons and Delta elevators are recorded ailerons’ data and elevators’ data for delta, and they have 7129 samples with 6 attributes, and 9247 with 7 attributes, respectively. Energy efficiency is a dataset that is obtained by energy analysis of 12 various architectural shapes, which is simulated in Ecotect, and there are 768 samples and 8 features in it, the regression problem is for forecast 2 authentic valued responses that are cooling load and heating load. Wine quality white is a dataset associated with red and white wine samples, it contains 4898 samples with 12 attributes. Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1)
Table 1

Attributes of the testing datasets

Dataset#Attributes#Train set#Test set
Abalone929231524
Airfoil self noise61052451
Auto-Mpg7274118
Bank957342458
Concrete slump119112
Diabetes2538230
Delta aileron61052451
Delta elevators766612856
Energy efficiency9537231
Wine quality white1234291469
Electrical_detect784003601
Electrical_No fault61656709
Electrical_LG fault6790339
Electrical_LLG fault6794340
Electrical_LLL fault6767329
Electrical_LLLG fault6793340
Asteroid71750750
Covid19_Beijing3345149
Covid19_Shanghai3345149
Covid19_Tianjin3345149
Covid19_Chongqing3345149
Covid19_Arizona3316136
Covid19_Washington3325140
Covid19_California3316136
Covid19_Illinois3317136

The Abalone, Airfoil self noise, Auto-Mpg, Bank, Concrete slump, Diabetes, Delta aileron, Delta elevators, Energy efficiency and Wine quality white datasets could be download via the following URL: https://archive.ics.uci.edu/ml/datasets.php

Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1) Mathematical expression of different regularization methods which are applied in TSFNN The information of the dataset is presented in Table 1, these datasets involve four small-scale datasets and six moderate-scale datasets. The mean and standard of 5000 experimental results of Abalone, Airfoil self-noise, Auto-Mpg, Bank, Concrete slump, Diabetes, Delta ailerons, Delta elevators, Energy efficiency and Wine quality white are showed in Tables 2 and 3. According to the comparison results, the top three algorithms with the descending order performance rank are TSFNN, TT2-RVFL and TT2-ELM on the whole. The most competitive algorithm of these three algorithms is RNN-LM. Performance of OP-ELM and TROP-ELM is at medium level. TT2-RVFL, TT2-ELM and TROP-ELM are the members of TSFNN, and they are all single-layer feedforward network. In the light of the excellent performance of TSFNN, the proposed TSFNN can be proved that can concentrate the strengths of its members. Meanwhile, the disadvantage of weaker members is offset. Under this stacked mechanism, the performance of TSFNN can be rivalled or surpassed to the RNN.
Table 2

Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1)

DatasetMethodTraining (MAE)Testing (MAE)Training (MSE)Testing (MSE)Training (RMSE)Testing (RMSE)
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
AbaloneTSFNN1.51e+001.85e024.72e+006.26e+014.38e+001.08e018.97e+033.27e+052.09e+002.58e027.09e+009.45e+01
TT2-RVFL1.52e+001.96e−021.54e+003.56e024.42e+001.14e−014.66e+003.40e−012.10e+002.77e−022.16e+007.62e−02
TT2-ELM1.51e+001.94e−021.54e+003.57e−024.38e+001.13e−014.64e+003.70e−012.09e+002.70e−022.15e+008.23e−02
OP-ELM1.66e+004.11e−022.51e+009.58e+005.24e+002.56e−011.16e+053.22e+062.29e+005.56e−023.17e+013.39e+02
TROP-ELM1.66e+004.20e−023.74e+006.92e+015.25e+002.60e−016.02e+064.07e+082.29e+005.63e−027.55e+012.45e+03
IT2-FNN1.62e+003.47e−021.63e+004.63e−024.96e+002.11e−015.05e+003.59e−012.23e+004.71e−022.25e+007.93e−02
eT2QFNN1.51e+007.76e−021.58e+007.85e−024.61e+004.20e−015.11e+005.92e−012.15e+009.52e−022.26e+001.25e−01
BD-ELM1.66e+009.53e−021.66e+009.89e−025.21e+006.21e−015.29e+006.83e−012.28e+001.27e−012.30e+001.40e−01
RNN-LM1.51e+003.47e−021.53e+004.36e−024.39e+001.59e−014.58e+002.81e012.09e+003.78e−022.14e+006.55e02
RNN-BFGS1.66e+009.95e−021.67e+001.02e−015.14e+006.21e−015.21e+006.54e−012.26e+001.12e−012.28e+001.22e−01
LSTM3.67e+002.13e+004.40e+003.13e+003.78e+013.39e+013.49e+014.88e+015.53e+002.67e+005.03e+003.09e+00
Airfoil self noiseTSFNN2.97e+008.96e023.03e+001.30e011.47e+018.73e011.54e+011.33e+003.83e+001.13e013.93e+001.69e01
TT2-RVFL3.17e+001.46e−013.29e+001.76e−011.66e+011.38e+001.79e+011.86e+004.06e+001.79e−014.23e+002.18e−01
TT2-ELM3.06e+001.26e−013.19e+001.58e−011.55e+011.16e+001.69e+011.65e+003.93e+001.46e−014.10e+002.00e−01
OP-ELM3.92e+001.58e−013.96e+001.98e−012.50e+011.71e+002.55e+012.33e+005.00e+001.70e−015.05e+002.28e−01
TROP-ELM3.92e+001.63e−013.96e+001.98e−012.50e+011.75e+002.55e+012.31e+005.00e+001.74e−015.05e+002.27e−01
IT2-FNN3.69e+001.75e−013.72e+002.07e−012.22e+011.78e+002.26e+012.26e+004.70e+001.87e−014.75e+002.37e−01
eT2QFNN3.33e+002.85e−017.19e+003.32e+002.04e+011.22e+011.03e+021.72e+024.45e+007.76e−019.15e+004.43e+00
BD-ELM4.17e+005.89e−014.20e+005.97e−012.80e+017.28e+002.85e+017.43e+005.26e+006.45e−015.29e+006.56e−01
RNN-LM2.58e+001.95e−012.68e+002.08e−011.19e+011.58e+001.28e+011.81e+003.44e+002.29e−013.57e+002.53e−01
RNN-BFGS3.58e+009.70e−023.62e+001.40e−012.11e+011.10e+002.16e+011.68e+004.60e+001.19e−014.64e+001.81e−01
LSTM2.94e+002.14e−013.21e+004.00e−011.47e+012.59e+001.75e+017.02e+003.82e+002.60e−014.16e+004.51e−01
AutoMPGTSFNN2.57e−011.88e022.71e−012.86e021.25e−011.38e021.34e012.46e023.53e−011.95e023.65e−013.37e02
TT2-RVFL2.87e−012.56e−023.34e−013.59e−021.56e−012.36e−022.21e−015.13e−023.92e−013.14e−024.67e−015.28e−02
TT2-ELM2.66e−012.45e−023.17e−013.53e−021.37e−012.04e−022.05e−014.78e−023.69e−012.73e−024.50e−015.15e−02
OP-ELM3.70e−012.77e−023.85e−013.74e−022.30e−012.73e−022.70e−019.01e−014.79e−012.83e−025.02e−011.33e−01
TROP-ELM3.70e−012.77e−023.86e−016.86e−022.30e−012.64e−026.67e−012.51e+014.79e−012.73e−025.11e−016.37e−01
IT2-FNN4.16e−014.08e−024.32e−014.71e−022.85e−015.23e−023.07e−016.46e−025.31e−014.71e−025.52e−015.65e−02
eT2QFNN4.16e−012.74e−024.59e−011.73e−013.44e−014.03e−023.91e−014.54e−015.86e−013.30e−025.94e−011.96e−01
BD-ELM5.08e−011.77e−015.21e−011.79e−014.48e−013.74e−014.70e−013.82e−016.37e−012.06e−016.53e−012.08e−01
RNN-LM1.52e−013.57e−022.07e013.77e−026.40e022.18e−021.34e−013.97e−022.49e014.32e−023.62e015.33e−02
RNN-BFGS2.83e−012.43e−023.02e−012.98e−021.60e−011.79e−021.86e−013.62e−024.00e−012.22e−024.29e−014.16e−02
LSTM2.26e013.97e−022.87e−015.81e−021.15e−016.57e−021.97e−011.40e−013.35e−014.84e−024.38e−017.30e−02
BankTSFNN2.00e−026.06e042.01e−026.96e047.92e−044.71e057.97e−046.20e052.81e−028.46e042.82e−021.09e03
TT2-RVFL3.20e−022.83e−033.23e−022.89e−032.05e−033.53e−042.10e−033.70e−044.65e−024.22e−034.53e−023.96e−03
TT2-ELM3.05e−023.08e−033.09e−023.15e−031.84e−033.62e−041.90e−033.79e−044.27e−024.21e−034.33e−024.35e−03
OP-ELM2.20e−021.70e−032.21e−021.70e−031.00e−032.20e−041.01e−032.20e−043.15e−022.46e−033.16e−022.53e−03
TROP-ELM2.20e−021.69e−032.21e−021.70e−031.00e−032.18e−041.01e−032.26e−043.15e−022.44e−033.16e−022.58e−03
IT2-FNN3.02e−023.77e−033.03e−023.79e−031.88e−034.87e−041.89e−034.93e−044.30e−025.51e−034.31e−025.58e−03
eT2QFNN2.26e−021.35e−032.13e−022.37e−031.11e−031.03e−048.40e−041.66e−043.32e−021.53e−032.89e−022.63e−03
BD-ELM5.14e−021.36e−025.14e−021.36e−024.99e−032.65e−035.00e−032.67e−036.90e−021.49e−026.91e−021.50e−02
RNN-LM1.21e021.33e−031.24e021.34e−033.59e041.24e−043.86e041.27e−041.89e021.66e−031.96e021.71e−03
RNN-BFGS2.21e−023.40e−032.22e−023.42e−031.02e−032.62e−041.03e−032.68e−043.17e−023.77e−033.19e−023.85e−03
LSTM2.37e−012.21e−021.79e−011.04e−018.04e−028.06e−035.15e−025.74e−022.83e−011.47e−022.02e−011.05e−01
Concrete slumpTSFNN9.76e−011.41e011.15e+001.97e−011.59e+004.48e−012.23e+008.47e011.25e+001.78e−011.47e+002.48e−01
TT2-RVFL1.84e+003.42e−013.73e+007.93e−015.70e+002.20e+002.67e+011.29e+012.41e+004.79e−014.97e+001.13e+00
TT2-ELM1.48e+003.17e−013.50e+008.17e−013.63e+001.60e+002.33e+011.22e+011.86e+004.00e−014.68e+001.16e+00
OP-ELM1.78e+001.56e−012.23e+003.51e−015.15e+008.63e−018.81e+003.18e+012.26e+001.89e−012.86e+007.80e−01
TROP-ELM1.77e+001.55e−012.22e+003.33e−015.13e+008.49e−018.46e+001.39e+012.26e+001.87e−012.85e+005.87e−01
IT2-FNN2.49e+005.30e−012.88e+006.89e−011.07e+014.77e+001.44e+017.31e+003.20e+006.65e−013.69e+008.76e−01
eT2QFNN3.17e+001.66e−023.68e+001.18e012.09e+011.13e011.99e+011.40e+004.57e+001.23e024.46e+001.59e01
BD-ELM4.12e+001.25e+004.44e+001.34e+002.92e+011.59e+013.35e+011.85e+015.18e+001.52e+005.56e+001.61e+00
RNN-LM2.41e011.93e−011.02e+005.80e−011.56e012.56e−012.50e+002.92e+003.09e−012.46e−011.37e+007.82e−01
RNN-BFGS1.71e+002.14e−012.21e+003.78e−014.75e+001.29e+008.21e+003.08e+002.16e+002.75e−012.82e+005.04e−01
LSTM1.04e+005.20e−012.29e+007.09e−012.23e+004.35e+009.48e+007.04e+001.33e+006.84e−012.94e+009.03e−01
Table 3

Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1)

DatasetMethodTraining (MAE)Testing (MAE)Training (MSE)Testing (MSE)Training (RMSE)Testing (RMSE)
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
Delta aileronTSFNN2.85e−034.10e052.88e−035.33e−051.28e053.08e071.30e−054.39e073.58e−034.32e053.61e−036.08e05
TT2-RVFL2.88e−034.22e−052.90e−035.46e−051.30e−053.26e−071.33e−054.62e−073.60e−034.75e−053.65e−036.33e−05
TT2-ELM2.85e−034.68e−052.88e−035.78e−051.28e053.53e−071.31e−054.89e−073.58e−034.93e−053.62e−036.75e−05
OP-ELM3.14e−032.27e−043.15e−032.31e−041.55e−052.26e−061.57e−054.98e−063.93e−032.61e−043.95e−033.61e−04
TROP-ELM3.14e−032.23e−043.14e−032.28e−041.55e−052.23e−061.56e−052.40e−063.92e−032.56e−043.94e−032.72e−04
IT2-FNN3.08e−033.74e−053.08e−035.20e051.48e−053.53e−071.48e−055.16e−073.84e−034.57e−053.85e−036.69e−05
eT2QFNN3.12e−035.47e−052.91e−037.33e−051.56e−056.27e−071.35e−057.39e−073.95e−037.93e−053.68e−039.74e−05
BD-ELM3.50e−036.79e−043.50e−036.79e−041.89e−056.57e−061.90e−056.57e−064.29e−036.86e−044.30e−036.86e−04
RNN-LM2.37e039.92e−052.39e031.03e−049.90e−065.95e−071.01e056.59e−073.15e039.12e−053.17e031.01e−04
RNN-BFGS3.02e−037.77e−053.03e−038.74e−051.43e−057.83e−071.43e−058.81e−073.78e−039.56e−053.79e−031.08e−04
LSTM4.75e−031.05e−031.55e−021.18e−024.87e−051.86e−053.90e−046.03e−046.82e−031.46e−031.60e−021.15e−02
Delta elevtorsTSFNN2.72e+002.35e−022.73e+003.77e−021.18e+011.81e−011.19e+013.09e−013.44e+002.63e−023.45e+004.48e−02
TT2-RVFL2.75e+002.92e−022.77e+004.22e−021.21e+012.29e−011.23e+013.47e−013.47e+003.45e−023.50e+004.96e−02
TT2-ELM2.73e+002.94e−022.75e+004.23e−021.20e+012.26e−011.22e+013.44e−013.46e+003.26e−023.49e+004.94e−02
OP-ELM2.85e+003.95e−022.85e+005.04e−021.29e+013.52e−011.30e+014.53e−013.59e+004.74e−023.60e+006.17e−02
TROP-ELM2.85e+003.68e−022.85e+004.92e−021.29e+013.26e−011.30e+014.36e−013.59e+004.42e−023.60e+005.96e−02
IT2-FNN2.85e+002.92e−022.85e+004.26e−021.29e+012.70e−011.29e+013.83e−013.59e+003.70e−023.60e+005.29e−02
eT2QFNN2.81e+003.56e−022.94e+004.36e−021.28e+013.37e−011.37e+013.74e−013.57e+004.69e−023.70e+004.96e−02
BD-ELM3.78e+003.46e−013.78e+003.48e−012.14e+012.99e+002.14e+013.02e+004.61e+003.38e−014.61e+003.41e−01
RNN-LM2.52e+003.08e−022.59e+006.07e−021.06e+012.67e−011.12e+014.95e−013.25e+004.11e−023.35e+007.38e−02
RNN-BFGS2.82e+001.59e022.87e+002.95e021.27e+019.62e021.31e+011.35e013.56e+001.35e023.62e+001.85e02
LSTM
DiabetesTSFNN1.01e013.18e031.56e−011.22e+002.00e021.15e036.59e+002.44e+021.41e−014.07e032.50e−012.55e+00
TT2-RVFL1.04e−013.41e−031.13e−015.85e−032.08e−021.23e−032.47e−022.95e031.44e−014.36e−031.57e−019.36e03
TT2-ELM1.02e−013.41e−031.13e−015.94e−032.02e−021.22e−032.49e−023.04e−031.42e−014.30e−031.58e−019.60e−03
OP-ELM1.11e−013.75e−032.09e−012.94e+002.39e−021.47e−032.00e+031.11e+051.55e−014.76e−031.61e+004.47e+01
TROP-ELM1.11e−013.74e−031.46e−018.91e−012.39e−021.47e−031.83e+029.05e+031.55e−014.76e−036.40e−011.35e+01
IT2-FNN1.13e−014.75e−031.16e−016.52e−032.44e−021.66e−032.55e−023.34e−031.56e−015.30e−031.59e−011.04e−02
eT2QFNN1.40e−011.40e−021.25e−012.99e−023.69e−027.90e−032.81e−022.37e−021.91e−011.96e−021.64e−013.37e−02
BD-ELM1.21e−011.37e−021.23e−011.40e−022.67e−024.56e−032.76e−025.22e−031.63e−011.30e−021.65e−011.51e−02
RNN-LM9.65e−024.70e−031.12e−016.82e−031.82e−021.55e−032.50e−023.31e−031.35e015.79e−031.58e−011.04e−02
RNN-BFGS1.08e−013.73e−031.11e015.74e032.27e−021.31e−032.41e023.01e−031.51e−014.36e−031.55e019.69e−03
LSTM1.03e−015.13e−021.17e−012.58e−022.28e−023.80e−022.71e−021.21e−021.42e−015.23e−021.63e−012.51e−02
Energy efficiencyTSFNN1.71e+007.49e021.71e+001.13e015.44e+005.30e−015.63e+007.02e012.33e+001.15e−012.37e+001.48e01
TT2-RVFL2.28e+002.24e−012.48e+002.66e−019.32e+001.69e+001.11e+012.28e+003.09e+002.92e−013.30e+003.28e−01
TT2-ELM2.09e+001.87e−012.31e+002.25e−017.91e+001.25e+009.69e+001.78e+002.80e+002.17e−013.10e+002.79e−01
OP-ELM2.07e+008.56e−022.12e+001.30e−018.16e+005.50e−018.57e+009.77e−012.86e+009.48e−022.92e+001.64e−01
TROP-ELM2.07e+008.21e−022.12e+001.29e−018.15e+005.19e−018.56e+009.40e−012.85e+009.02e022.92e+001.60e−01
IT2-FNN2.25e+002.26e−012.29e+002.53e−019.57e+001.70e+009.93e+002.00e+003.08e+002.55e−013.14e+002.97e−01
eT2QFNN2.50e+002.30e−013.70e+001.86e+001.68e+013.86e+002.56e+013.64e+014.07e+004.41e−014.58e+002.15e+00
BD-ELM5.16e+002.16e+005.17e+002.16e+004.55e+013.11e+014.57e+013.11e+016.35e+002.28e+006.37e+002.27e+00
RNN-LM1.20e+003.12e−011.29e+003.38e−012.84e+001.39e+003.28e+001.63e+001.62e+004.55e−011.74e+004.92e−01
RNN-BFGS1.95e+009.37e−022.01e+001.35e−017.62e+005.12e018.10e+009.07e−012.76e+009.19e−022.84e+001.59e−01
LSTM2.03e+005.24e−012.53e+007.85e−017.88e+004.70e+001.16e+017.17e+002.72e+006.97e−013.29e+008.56e−01
Wine quality whiteTSFNN4.81e−017.80e034.90e013.95e−013.81e−011.15e021.23e+005.05e+016.17e019.30e−036.36e019.07e−01
TT2-RVFL4.93e−019.20e−035.13e−011.69e023.96e−011.35e−024.34e−012.97e026.29e−011.11e−026.58e−012.25e02
TT2-ELM4.90e−019.54e−035.13e−011.73e−023.91e−011.38e−024.35e−013.14e−026.25e−011.10e−026.59e−012.37e−02
OP-ELM5.01e−017.80e035.18e−013.06e−014.12e−011.18e−024.52e+011.92e+036.42e−019.18e038.41e−016.67e+00
TROP-ELM5.01e−017.87e−035.10e−012.13e−024.12e−011.19e−025.16e−013.53e+006.42e−019.30e−036.64e−012.75e−01
IT2-FNN5.23e−011.73e−025.29e−012.24e−024.45e−012.41e−024.56e−013.65e−026.67e−011.79e−026.75e−012.68e−02
eT2QFNN5.53e−012.79e−025.41e−015.24e−025.28e−016.95e−025.05e−011.11e−017.25e−014.62e−027.08e−016.51e−02
BD-ELM5.47e−013.64e−025.51e−013.92e−024.79e−015.07e−024.88e−015.78e−026.91e−013.59e−026.97e−014.07e−02
RNN-LM4.65e011.19e−025.03e−011.71e−023.55e011.59e−024.23e013.10e−025.96e−011.33e−026.50e−012.36e−02
RNN-BFGS5.01e−019.59e−035.11e−011.76e−024.15e−011.44e−024.33e−012.99e−026.44e−011.11e−026.58e−012.27e−02
LSTM4.86e−011.25e−025.09e−011.87e−023.85e−011.72e−024.27e−013.11e−026.20e−011.33e−026.53e−012.36e−02

Comparison with different regularization methods

In this paper, the Tikhonov regularization is used to optimize the performance of proposed TSFNN. The regularization method can be used to alleviate the over-fitting problem. In this section, three classical regularization methods are used to verify the effectiveness of the Tikhonov regularization in the proposed TSFNN. These regularization methods are regularization, regularization and dropout regularization (Srivastava et al. 2014). Their mathematical expressions are shown in Table 4. Four regularization methods are applied in the proposed TSFNN via Eq. (17), respectively.
Table 4

Mathematical expression of different regularization methods which are applied in TSFNN

IDRegularizationFormulationSolutionParameter
1classical \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\Vert \varvec{A}_k\varvec{\beta _k} -\varvec{y}_t \Vert ^{2}+\Vert \varvec{\beta _k} \Vert ^{1}$$\end{document}Akβk-yt2+βk1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\beta _k}}=(\varvec{A}_k^{T}\varvec{A}_k+0.5)^{-1}\varvec{A}_k^{T}\varvec{y}_t$$\end{document}βk^=(AkTAk+0.5)-1AkTyt
2classical \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\Vert \varvec{A}_k\varvec{\beta _k} -\varvec{y}_t \Vert ^{2}+\Vert \varvec{\beta _k} \Vert ^{2}$$\end{document}Akβk-yt2+βk2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\beta _k}}=(\varvec{A}_k^{T}\varvec{A}_k+I )^{-1}\varvec{A}_k^{T}\varvec{y}_t$$\end{document}βk^=(AkTAk+I)-1AkTytI: Dentity matrix
3Tikhonov\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\Vert \varvec{A}_k\varvec{\beta _k} -\varvec{y}_t \Vert ^{2}+\Vert \Gamma _k \varvec{\beta _k} \Vert ^{2}$$\end{document}Akβk-yt2+Γkβk2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\beta _k}}=(\varvec{A}_k^{T}\varvec{A}_k+\Gamma ^{T}\Gamma )^{-1}\varvec{A}_k^{T}\varvec{y}_t$$\end{document}βk^=(AkTAk+ΓTΓ)-1AkTyt\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\Gamma $$\end{document}Γ: Tikhonov matrix
4Dropout\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\Vert R*\varvec{(}A_k)\varvec{\beta _k} -\varvec{y}_t\Vert ^{2}$$\end{document}R(Ak)βk-yt2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\beta _k}}=\big ((R*\varvec{A}_k)^{T}(R*\varvec{A}_k)\big )^{-1} (R*\varvec{A}_k)^{T}\varvec{y}_t$$\end{document}βk^=((RAk)T(RAk))-1(RAk)Tyt\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$R\sim Bernoulli (p)$$\end{document}RBernoulli(p)
Comparison results for TSFNN with different regularization methods. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1) Four variants of TSFNN can be constructed via four different regularization methods. For classical regularization and classical regularization, their regularization parameter is set as 1. For dropout regularization, the matrix R satisfies the condition of and the probability p is set as 0.5 in this work. Among the comparisons, the classical regularization, classical regularization and dropout regularization are applied by the suggested tensor-based stacked fuzzy networks model to create variants of original TSFNN. Table 5 expresses comparison results for TSFNN with different regularization methods. On the whole, the descending order performance rank of variants of TSFNN with four regularization methods is regularization, Tikhonov regularization, regularization and dropout regularization. In the training and testing stage, TSFNN with regularization obtains the smallest error in four regularization methods in ten different scales datasets. But the stability of TSFNN with Tikhonov regularization outperforms the TSFNN with regularization. Taken together, the Tikhonov regularization can effectively deal with the uncertain information in the suggested model, and suppress over-fitting risk.
Table 5

Comparison results for TSFNN with different regularization methods. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1)

DatasetMethodTraining (MAE)Testing (MAE)Training (MSE)Testing (MSE)Training (RMSE)Testing (RMSE)
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
AbaloneTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L11.48e+001.83e021.71e+005.65e+004.23e+001.08e−017.89e+013.78e+032.06e+002.63e−022.81e+008.43e+00
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L21.54e+001.89e−023.84e+004.13e+014.54e+001.11e−014.50e+031.91e+052.13e+002.61e−025.77e+006.68e+01
TSFNN-Tikhonov1.52e+001.85e−024.14e+003.29e+014.43e+001.10e−012.51e+035.45e+042.10e+002.60e026.12e+004.97e+01
TSFNN-Dropout1.67e+002.24e−014.45e+004.32e+012.97e+011.00e+036.36e+043.53e+062.52e+004.84e+001.43e+012.52e+02
Airfoil self noiseTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L12.74e+001.01e−012.92e+002.74e−011.26e+018.75e−011.45e+017.18e+003.55e+001.23e−013.78e+004.83e−01
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L23.22e+009.99e−023.11e+001.33e−011.75e+019.98e−011.64e+011.39e+004.18e+001.19e−014.04e+001.71e−01
TSFNN-Tikhonov3.03e+009.04e023.02e+001.29e011.55e+018.86e011.54e+011.33e+003.93e+001.12e013.92e+001.68e01
TSFNN-Dropout3.61e+002.60e+005.07e+001.78e+002.02e+039.87e+043.07e+025.64e+036.29e+004.45e+018.03e+001.56e+01
AutoMPGTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L12.17e011.58e022.25e014.65e−029.18e−021.10e021.48e−012.70e+003.03e011.82e023.14e012.22e−01
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L23.11e−012.41e−022.88e−013.01e−021.67e−011.84e−021.45e−012.59e−024.09e−012.25e−023.80e−013.42e−02
TSFNN-Tikhonov2.70e−011.95e−022.71e−012.82e021.35e011.42e−021.34e012.43e023.67e−011.94e−023.65e−013.33e02
TSFNN-Dropout4.00e−014.92e−019.65e−012.07e+002.71e+011.25e+032.33e+021.07e+047.85e−015.15e+001.76e+001.52e+01
BankTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L11.98e026.25e−042.06e−021.05e−037.75e044.78e−059.46e−046.71e−032.78e028.69e−042.91e−021.01e02
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L22.04e−025.25e042.02e−026.50e048.22e−044.24e058.06e−045.74e052.87e−027.47e−042.84e−021.02e−03
TSFNN-Tikhonov2.01e−025.73e−042.01e026.95e−048.00e−044.51e−057.98e046.39e−052.83e−028.06e042.82e021.10e−03
TSFNN-Dropout2.30e−021.55e−026.29e−023.54e−021.19e+008.39e+011.35e−021.33e−015.84e−021.09e+008.10e−028.35e−02
Concrete slumpTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L12.87e+013.87e+019.55e+011.61e+022.32e+031.25e+043.53e+041.64e+052.87e+013.87e+019.62e+011.61e+02
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L21.45e+001.22e011.16e+001.74e013.40e+005.42e−012.18e+006.38e011.84e+001.47e011.46e+002.14e01
TSFNN-Tikhonov1.12e+001.29e−011.15e+001.89e−012.08e+004.64e012.23e+007.32e−011.43e+001.61e−011.47e+002.36e−01
TSFNN-Dropout1.75e+013.34e+011.74e+011.80e+015.42e+031.57e+051.19e+032.00e+042.41e+016.96e+012.10e+012.73e+01
Delta aileronTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L12.79e034.67e−052.80e038.90e−051.23e053.39e−071.48e−051.01e−043.51e034.83e−053.60e031.36e−03
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L22.93e−033.74e052.91e−035.40e051.35e−053.10e−071.32e−054.71e073.67e−034.23e−053.64e−036.42e05
TSFNN-Tikhonov2.87e−033.86e−052.88e−035.47e−051.30e−053.01e071.30e054.92e−073.60e−034.19e053.61e−036.68e−05
TSFNN-Dropout3.04e−032.14e−043.61e−036.84e−041.05e−042.74e−034.43e−055.05e−044.49e−039.19e−034.97e−034.43e−03
Delta elevtorsTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L12.69e+002.41e−022.70e+004.70e−021.17e+011.82e−011.19e+016.15e+003.42e+002.66e−023.44e+003.24e01
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L22.76e+002.25e022.74e+003.77e021.21e+011.77e011.20e+013.10e013.48e+002.54e023.46e+004.48e−02
TSFNN-Tikhonov2.73e+002.27e−022.73e+003.80e−021.19e+011.77e011.19e+013.11e−013.45e+002.57e−023.45e+004.51e−02
TSFNN-Dropout2.83e+003.68e−023.32e+004.02e−011.69e+011.37e+023.27e+014.59e+023.69e+001.83e+004.40e+003.65e+00
DiabetesTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L19.76e023.16e031.13e013.89e011.86e021.10e035.72e011.94e+011.36e014.05e031.97e−017.30e01
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L21.05e−013.27e−031.21e−016.30e−012.16e−021.21e−031.50e+009.13e+011.47e−014.13e−031.78e011.21e+00
TSFNN-Tikhonov1.02e−013.21e−032.15e−015.32e+002.05e−021.16e−031.06e+026.84e+031.43e−014.08e−033.62e−011.03e+01
TSFNN-Dropout1.14e−015.50e−022.71e−013.76e+009.59e−014.82e+011.21e+027.12e+032.38e−019.50e−016.78e−011.10e+01
Energy efficiency1TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L11.62e+008.35e−021.78e+003.22e−014.67e+005.58e−015.40e+009.08e+002.16e+001.31e−012.29e+004.22e−01
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L21.80e+006.83e021.73e+001.14e−016.52e+004.10e015.93e+006.88e012.55e+008.05e022.43e+001.42e01
TSFNN-Tikhonov1.72e+007.19e−021.71e+001.12e015.75e+004.81e−015.62e+006.90e−012.40e+001.01e−012.37e+001.46e−01
TSFNN-Dropout2.02e+001.95e−014.04e+001.48e+001.81e+016.29e+026.50e+011.25e+032.90e+003.11e+005.36e+006.02e+00
Wine quality whiteTSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_1$$\end{document}L14.69e018.04e−035.05e−011.86e013.63e011.14e027.76e+002.10e+026.02e019.51e−038.64e−012.65e+00
TSFNN-\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_2$$\end{document}L24.91e−017.41e036.24e−019.76e+003.96e−011.15e−023.83e+022.70e+046.29e−019.15e039.01e−011.95e+01
TSFNN-Tikhonov4.84e−017.57e−034.87e011.95e−013.85e−011.14e025.91e011.02e+016.20e−019.21e−036.29e014.43e01
TSFNN-Dropout5.22e−011.23e−018.80e−015.10e+001.59e+013.99e+021.23e+034.43e+041.09e+003.83e+003.16e+003.50e+01

Remark 5

The Tikhonov regularization and dropout regularization are all belong to regularization method. The difference lies in the selection and determination of regularization parameters which can be called the tuning parameters of regularization. For the used dropout regularization in this section, the election of key parameter p is based on the original paper of dropout regularization, that is Srivastava et al. (2014), among it, the model obtains the excellent performance when . Thus, the parameter p is set as 0.5 in this section.

Simulations for other datasets

In this section, three datasets which are Electrical Fault detection and classification datasets, Asteroid Dataset and Novel Corona Virus 2019 Dataset are used for testing model performance.

Electrical fault detection and classification dataset

Power systems consist of many complex dynamic and interactive elements that are always vulnerable to interference or electrical failures. Transmission lines are the most critical part of the power system, and the prominent role of transmission lines is to transmit electricity from the source area to the distribution destination in the network. The faults of power system transmission lines should be first correctly detected and classified, and should be eliminated in the shortest possible time. Electrical Fault detection and classification dataset contains the current and voltage of the line under different fault conditions (Jamil et al. 2015). The dataset contains the detection of power system faults and classifying fault types for the power system faults. The dataset of power system fault detection contains 12001 sampling data with six inputs (, , , , , V, c) and two outputs (0 and 1), no fault is denoted by 0, and fault is denoted by 1. The dataset of classifying fault types contains 7861 group data with six inputs (, , , , , V, c) and four outputs (G, C, B and A), they represents 4 generators of V, respectively. No fault occurs is denoted by 0, and fault occurs is denoted by 1. The combination of G, C, B and A represents various failures, and the failures faults are shown in Table 7.
Table 7

Faults represented by G, C, B and A

[G, C, B, A]Faults
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${[}0,\ 0,\ 0,\ 0]$$\end{document}[0,0,0,0]No fault
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${[}1,\ 0,\ 0,\ 1]$$\end{document}[1,0,0,1]LG fault (between phase A and phase G)
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${[}0,\ 0,\ 1,\ 1]$$\end{document}[0,0,1,1]LL fault (between phase A and phase B)
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${[}1,\ 0,\ 1,\ 1]$$\end{document}[1,0,1,1]LLG fault (between Phases A, B and ground)
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${[}0,\ 1,\ 1,\ 1]$$\end{document}[0,1,1,1]LLL fault (between all three phases)
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${[}1,\ 1,\ 1,\ 1]$$\end{document}[1,1,1,1]LLLG fault (three phase symmetrical fault)
The faults of the system are judged according to the current and voltage of the power system. The dataset consists of two outputs that represents whether the system faults. Figure 6 shows that no faults in Power Electrical detection dataset. Similarly, Fig. 7 shows that faults occurs in Power Electrical detection dataset. The transverse axis delegates samples, and the longitudinal axis delegates the values of , , , , and . Compared with Figs. 6 and 7, if there is no fault in the electric power system, the values of current and voltage are generally stable, and its change trend is similar to the function, which is also consistent with the characteristics of AC in the electric power system. However, once a fault occurs in the power system, the values of current and voltage are abnormal, which represents different fault locations. This anomaly can be clearly seen from Fig. 7. The comparison result of detection of power system faults dataset that is observed to determine whether the power system is faulty is shown in Table 6.
Fig. 6

The data of no faults in Electrical detect dataset

Fig. 7

The data has faults in Electrical detect dataset

Table 6

Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for Electrical-class, Electrical-detect. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1)

DatasetMethodTraining (MAE)Testing (MAE)Training (MSE)Testing (MSE)Training (RMSE)Testing (RMSE)
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
Electrical_detectTSFNN3.04e041.45e−035.18e041.79e−033.76e062.09e−055.80e−063.16e−053.99e041.90e−036.85e−042.31e−03
TT2-RVFL9.71e−033.79e−039.78e−033.82e−032.29e−042.02e−042.33e−042.06e−041.49e−025.71e−031.35e−025.19e−03
TT2-ELM6.57e−032.64e−036.62e−032.66e−039.70e−059.07e−059.87e−059.18e−059.17e−033.59e−039.25e−033.62e−03
OP-ELM2.22e−031.31e−022.22e−031.31e−022.97e−041.82e−032.98e−041.82e−032.91e−031.70e−022.91e−031.70e−02
TROP-ELM2.95e−031.48e−022.95e−031.49e−023.86e−042.03e−033.87e−042.04e−033.84e−031.93e−023.84e−031.93e−02
IT2-FNN2.85e−021.64e−022.85e−021.64e−021.82e−032.49e−031.82e−032.48e−033.72e−022.08e−023.72e−022.08e−02
eT2QFNN1.36e−036.49e055.52e−047.43e055.99e−051.22e064.37e071.31e077.74e−037.85e056.56e048.10e05
BD-ELM1.30e−018.93e−021.30e−018.93e−023.56e−023.74e−023.56e−023.74e−021.58e−011.03e−011.58e−011.03e−01
RNN-LM1.76e−035.59e−031.78e−035.71e−034.98e−056.84e−045.11e−057.01e−042.16e−036.71e−032.19e−036.80e−03
RNN-BFGS3.16e−021.46e−023.17e−021.49e−021.92e−032.38e−031.92e−032.48e−034.01e−021.76e−023.99e−021.81e−02
LSTM3.40e−012.08e−024.64e−013.14e−012.05e−013.24e−023.23e−014.07e−014.52e−012.76e−024.76e−013.10e−01
The data of no faults in Electrical detect dataset The data has faults in Electrical detect dataset Comparison results of TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS, LSTM and TSFNN on the Electrical Fault detection dataset are used to test the performance of the eleven algorithms. Results of TSFNN in the Table 6 show that the generalization ability of TSFNN is better than the other four algorithms. Moreover, data indicates power system failure can be considered as no power system failure added noise by comparing with Figs. 6 and 7. Therefore, Table 6 also shows that the disturbance reject ability of TSFNN is better than the other four algorithms. In Table 6, the proposed TSFNN obtains the smallest error, followed by RNN-LM. The errors of TSFNN’s three members, TT2-RVFL, TT2ELM and TROP-ELM are larger than RNN-LM. The descending ranking of the standard deviation index is TSFNN, TT2-ELM, TT2-RVFL and RNN-LM. TT2-ELM and TT2-RVFL have type-2 fuzzy set. The type-2 fuzzy set is fused into the tensor structure. Type-2 fuzzy sets have great performance in dealing with uncertain information, which also has excellent anti-interference ability. Tensor structure has the similar characteristic as the type-2 fuzzy set. Although RNN-LM is less stable than TSFNN’s three members structure, its generalization performance is better. Therefore, the proposed tensor-based stacked neural network strategy can inherit the advantages of its members, such as stability. It also can integrate the different capabilities of its members to achieve the excellent performance of the entire model. For Electrical Fault classification case, we decompose Electrical Fault classification dataset into five parts according to the different locations of faults. It can be seen from Table 7 that the dataset provides six kinds of faults, but we do not find LL fault ([G, C, B, A] = [0, 0, 1, 1]) in the dataset, which represents the fault between Phase A and Phase B. Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for Electrical-class, Electrical-detect. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1) Faults represented by G, C, B and A Through the above analysis, we know that in the power system, the faults data can be regarded as the no faults data with added noise. Thus, in the five extracted datasets, the dataset of LG fault, LLG fault, LLL fault, and LLLG fault can be treated as imposing different noise on the dataset with no fault status. Moreover, the extracted dataset represents only one power system fault, so its data is more pure and has more obvious characteristics and trends. Through the above-mentioned analysis, the anti-interference performance and generalization performance of the algorithm can be further verified. Results of TSFNN in the Table 8 fully proof that the excellent disturbance rejection and generalization performance of TSFNN. The comparison results of the proposed TSFNN and ten other models are shown in Table 8. TSFNN obtains the smallest error and standard deviation on Electrical No fault dataset and Electrical LG fault dataset in training and testing part. On Electrical LLG fault dataset, RNN-LM obtains the most excellent performance. eT2QFNN is RNN-LM’s the most competitor. The eT2QFNN outperforms RNN-LM, TSFNN and TSFNN on Electrical LLL fault dataset and Electrical LLLG fault dataset.
Table 8

Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for various faults in Electrical dataset (a brief introduction is listed in Table 1)

DatasetMethodTraining (MAE)Testing (MAE)Training (MSE)Testing (MSE)Training (RMSE)Testing (RMSE)
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
Electrical No faultTSFNN1.44e−031.37e−032.07e−031.66e−037.01e061.45e−051.39e−056.21e−051.96e−031.78e−032.91e−032.33e−03
TT2-RVFL2.62e−031.08e−032.70e−031.11e−031.82e−051.67e−052.44e−053.54e−054.16e−031.69e−034.23e−032.08e−03
TT2-ELM2.00e−037.57e042.07e−037.88e049.68e−068.05e061.45e−052.57e052.92e−031.08e033.37e−031.76e−03
OP-ELM1.92e−027.73e−031.94e−027.79e−031.50e−037.24e−042.21e−034.25e−023.74e−029.69e−033.83e−022.72e−02
TROP-ELM1.93e−027.74e−031.94e−027.68e−031.51e−037.48e−041.60e−031.17e−033.76e−029.67e−033.81e−021.22e−02
IT2-FNN7.38e−033.61e−037.47e−033.65e−031.63e−041.74e−041.82e−042.19e−041.15e−025.49e−031.20e−026.15e03
eT2QFNN1.51e−024.80e−034.23e−023.91e−028.26e−046.50e−045.37e−031.68e−022.75e−028.47e−035.35e−025.01e−02
BD-ELM2.16e−011.56e−012.15e−011.56e−018.97e−027.99e−028.96e−027.98e−022.48e−011.67e−012.48e−011.67e−01
RNN-LM6.30e042.45e−036.34e042.41e−038.73e−061.72e−048.67e061.63e−047.50e042.86e−037.59e042.85e−03
RNN-BFGS1.89e−021.11e−021.90e−021.09e−021.04e−031.59e−031.07e−031.54e−032.96e−021.29e−023.00e−021.31e−02
LSTM1.61e−017.55e−029.85e−014.87e−019.77e−026.36e−021.22e+009.33e−012.92e−011.13e−019.90e−014.85e−01
Electrical LG faultTSFNN8.03e−048.03e−041.31e−031.01e−032.21e−064.74e−066.03e−061.26ee051.08e−031.03e−032.12e−031.24e03
TT2-RVFL1.31e−034.90e−041.51e−035.85e−046.20e−064.95e−062.91e−055.59e−052.47e−038.94e−044.08e−033.17e−03
TT2-ELM1.12e−033.91e041.32e−034.95e044.16e−063.12e062.43e−054.94e−051.93e−036.47e043.79e−033.15e−03
OP-ELM9.32e−035.30e−039.50e−035.36e−033.81e−042.79e−044.33e−041.20e−031.86e−026.00e−031.91e−028.22e−03
TROP-ELM9.37e−035.46e−039.53e−035.49e−033.83e−042.90e−044.48e−041.73e−031.85e−026.34e−031.91e−029.22e−03
IT2-FNN5.25e−032.27e−035.40e−032.32e−031.00e−048.23e−051.21e−041.23e−049.33e−033.63e−039.95e−034.66e−03
eT2QFNN4.52e−031.50e−035.17e−031.03e−026.60e−056.02e−051.98e−049.80e−047.81e−032.23e−036.28e−031.26e−02
BD-ELM2.22e−023.91e−022.24e−023.92e−022.94e−039.52e−032.97e−039.56e−033.09e−024.45e−023.14e−024.46e−02
RNN-LM3.51e041.22e−033.62e041.16e−032.16e064.23e−052.04e062.91e−054.37e041.40e−034.93e041.34e−03
RNN-BFGS1.19e−024.43e−031.22e−024.61e−034.17e−042.67e−044.55e−043.13e−041.98e−025.16e−032.04e−026.29e−03
LSTM1.98e−021.55e−024.72e−026.36e−022.15e−035.93e−038.75e−032.63e−023.22e−023.46e−025.63e−027.73e−02
Electrical LLG faultTSFNN1.30e−031.07e−031.92e−031.31e−034.91e−067.29e−061.03e−051.67e−051.74e−031.37e−032.78e−031.61e−03
TT2-RVFL1.73e−036.12e−042.01e−037.48e−041.04e−057.81e−065.58e−051.04e−043.21e−031.12e−035.60e−034.42e−03
TT2-ELM1.38e−034.54e−041.67e−036.17e−046.06e−064.13e−064.82e−059.29e−052.35e−037.45e−045.26e−034.52e−03
OP-ELM1.49e−026.99e−031.51e−027.07e−037.32e−044.27e−048.07e−042.15e−032.60e−027.28e−032.65e−021.02e−02
TROP-ELM1.50e−027.04e−031.52e−027.12e−037.38e−044.38e−047.82e−045.33e−042.61e−027.43e−032.65e−028.84e−03
IT2-FNN6.10e−032.81e−036.22e−032.84e−031.11e−049.91e−051.31e−041.40e−049.72e−034.09e−031.03e−025.01e−03
eT2QFNN1.80e−036.86e052.87e−024.57e043.08e−057.37e076.19e−032.21e−045.55e−035.68e057.87e−021.08e03
BD-ELM2.67e−024.09e−022.68e−024.09e−023.48e−031.02e−023.51e−031.02e−023.63e−024.66e−023.66e−024.67e−02
RNN-LM3.62e048.92e−043.84e048.74e−041.32e061.72e−051.54e061.41e054.73e041.05e−036.10e041.08e03
RNN-BFGS1.15e−024.49e−031.20e−024.86e−033.40e−042.68e−043.83e−043.10e−041.76e−025.58e−031.84e−026.58e−03
LSTM1.85e−021.52e−023.68e−026.19e−023.14e−033.52e−025.87e−034.60e−022.92e−024.79e−024.22e−026.40e−02
Electrical LLL faultTSFNN8.87e−044.25e−041.38e−037.85e−042.32e−062.30e−068.91e−061.39e−041.36e−036.89e−042.20e−032.02e−03
TT2-RVFL1.08e−033.59e−041.22e−034.13e−043.80e−062.56e−061.76e−053.58e−051.96e−036.03e−043.15e−032.47e−03
TT2-ELM8.92e−042.65e−041.03e−033.38e−042.39e−061.41e−061.45e−053.24e−051.49e−034.10e−042.89e−032.47e−03
OP-ELM1.22e−023.34e−031.25e−023.50e−036.07e−042.27e−041.01e−037.52e−032.41e−025.29e−032.61e−021.82e−02
TROP-ELM1.20e−023.41e−031.24e−023.84e−036.04e−042.29e−041.58e−032.38e−022.39e−025.53e−032.67e−022.94e−02
IT2-FNN4.11e−032.36e−034.18e−032.40e−034.41e−056.04e−054.84e−057.52e−055.78e−033.26e−035.98e−033.56e−03
eT2QFNN5.65e−041.15e061.09e−012.61e051.99e−052.37e075.18e−022.46e−054.46e−032.67e052.28e−015.41e05
BD-ELM1.26e−021.58e−021.28e−021.60e−029.44e−042.40e−039.66e−042.46e−031.85e−022.46e−021.87e−022.48e−02
RNN-LM2.41e043.29e−043.15e043.61e−043.24e072.34e−061.18e064.42e063.89e044.16e−047.65e047.69e−04
RNN-BFGS8.47e−033.89e−038.68e−033.81e−032.25e−042.32e−042.71e−042.81e−041.37e−026.07e−031.49e−026.92e−03
LSTM6.82e−024.56e−023.66e−013.49e−012.18e−022.55e−022.63e−015.41e−011.26e−017.69e−023.75e−013.50e−01
Electrical LLLG faultTSFNN1.34e−037.57e−042.02e−031.02e−034.57e−065.23e−061.06e−051.66e−051.90e−039.81e−042.92e−031.44e−03
TT2-RVFL1.13e−033.89e−041.26e−034.35e−044.31e−063.10e−061.64e−052.88e−052.08e−036.77e−043.12e−032.28e−03
TT2-ELM9.19e−042.87e−041.05e−033.50e−042.68e−061.67e−061.34e−052.56e−051.57e−034.51e−042.85e−032.29e−03
OP-ELM1.90e−026.36e−031.92e−026.46e−031.04e−035.81e−041.10e−038.30e−043.11e−028.24e−033.17e−029.55e−03
TROP-ELM1.89e−026.47e−031.92e−026.55e−031.03e−035.89e−041.13e−031.38e−033.11e−028.24e−033.19e−021.07e−02
IT2-FNN4.66e−032.56e−034.74e−032.61e−036.19e−057.90e−056.70e−059.07e−056.99e−033.61e−037.21e−033.87e−03
eT2QFNN4.51e−042.10e061.06e−016.45e051.56e−051.23e073.88e−024.32e−053.94e−031.56e051.97e−011.10e04
BD-ELM1.49e−021.39e−021.51e−021.40e−029.33e−041.67e−039.62e−041.71e−032.20e−022.12e−022.23e−022.15e−02
RNN-LM2.06e042.67e−042.51e042.85e−042.51e071.17e−067.57e072.39e063.36e043.72e−046.13e046.18e−04
RNN-BFGS8.65e−033.04e−038.66e−033.27e−032.07e−041.49e−042.32e−041.94e−041.37e−024.45e−031.43e−025.39e−03
LSTM2.04e−021.04e−021.26e−012.05e−011.94e−032.43e−036.02e−021.52e−013.85e−022.19e−021.37e−012.08e−01
Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for various faults in Electrical dataset (a brief introduction is listed in Table 1) It can be seen from the results that the disturbance rejects ability of the proposed TSFNN is worse than eT2QFNN and RNN-LM. However, TSFNN outperforms the OP-ELM, IT2-FNN, BD-ELM and its three constituent members. RNN-LM is a deep network. eT2QFNN has an interval type-2 quantum fuzzy set with uncertain jump positions. The quantum fuzzy set possesses a graded membership degree. The process of generating the graded membership degree make eT2QFNN has a deep structure. Therefore, eT2QFNN can be considered as a deep network that is different from RNN-LM. IT2 fuzzy set is one of the reasons why eT2QFNN performs better than RNN-LM. The deep structure also is the reason why eT2QFNN and RNN-LM outperform the rest of the comparison algorithm. Although the performance of the deep network is better than TSFNN, the proposed TSFNN achieves comparable performance to deep networks. The one of reason is the tensor-based stacked neural network structure. According to the above results, the effectiveness of the tensor-based network stacked strategy can be proved powerfully.

Asteroid dataset

The Asteroid Dataset is officially maintained by Jet Propulsion Laboratory of California Institute of Technology, which is an organization under NASA. The data set is publicly available in JPL Small-Body Database Search Engine. This dataset could also be obtained from kaggle1. Table 9 shows basic column definition for Asteroid dataset.
Table 9

Basic column definition for Asteroid dataset

AttributesDescription
SPK-IDObject primary SPK-ID
Object IDObject internal database ID
Object fullnameObject full name/designation
PdesObject primary designation
NameObject IAU name
NEONear-earth object (NEO) flag
PHAPotentially Hazardous Asteroid (PHA) flag
HAbsolute magnitude parameter
DiameterObject diameter (from equivalent sphere) km unit
AlbedoGeometric albedo
Diameter_sigma1-sigma uncertainty in object diameter km unit
Orbit_idOrbit solution ID
EpochEpoch of osculation in modified Julian day form
EquinoxEquinox of reference frame
eEccentricity
aSemi-major axis au unit
qPerihelion distance au unit
iInclination; angle with respect to \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$x-y$$\end{document}x-y ecliptic plane
tpTime of perihelion passage TDB Unit
moid_ldEarth minimum orbit intersection distance au unit
A portion of the data is extract by us as a comparison test dataset when using Asteroid dataset. The 2500 samples in the dataset presents 7 attributes, these samples are applied to validate the proposed algorithm. These properties are Geometric albedo, Eccentricity, Semi-major axis, inclination angle about the x-y elliptic plane, Earth Minimum Orbit Intersection Distance and RMS for the Asteroid, respectively. The comparison results with eleven methods are demonstrated in Table 9. The results demonstrate that the TSFNN performs best with respect to training error, while TT2-RVFL and TT2-ELM performs best in testing error. Meanwhile, the performance of OP-ELM and TROP-ELM is bad. Because the approach of proposed in this paper is stacked of TT2-RVFL, TT2-ELM and TROP-ELM. The main reason why TSFNN, OP-ELM and TROP-ELM perform well in training and poor in testing is that these three methods all use multi-response sparse regression (MRSR), it is a variable sorting technology that is extended from the least angle regression algorithm (Similä and Tikka 2005; Efron et al. 2004). According to the usefulness of neurons, the MRSR algorithm can obtain a ranking of the neurons in OP-ELM (Miche et al. 2010). TROP-ELM is an improvement of OP-ELM, the MRSR method is also used for the input data of TROP-ELM. Due to the proposed TSFNN includes TROP-ELM, thus the TSFNN is affected by MRSR method. To the MRSR, an important feature is that the obtained ordering is exact in the case of linear problems. The Asteroid dataset collects the attributes of asteroid. A part of its data is used, it is also nonlinear. And the OP-ELM and the TROP-ELM have one detail that the neural networks they constructed are linear between the hidden layer and the output layer, the role of MRSR algorithm is that will get an exact ranking of the neurons. The sequence obtained by sorting can be used to sort the kernels of model. When the whole dataset is nonliner, so the exact ranking of neurons cannot be obtained by OP-ELM. Similarly, TROP-ELM and TSFNN are also affected by this flaw. Therefore, TSFNN performs well in the training part, and in the test part, due to MRSR method, the extracted data set features cannot be well applied to the testing set, resulting in poor performance of TSFNN in the testing phase. Basic column definition for Asteroid dataset Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for Asteroid. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1) According to the data in Table 10, the performance of TT2-RVFL and TT2-ELM is the best. The TT2-RVFL and the TT2-ELM are constructed by tensor structure and interval type-2 fuzzy sets. The membership degree of type-2 fuzzy set is characterized by type-1 fuzzy set. Since the type-1 fuzzy set has a strong ability to deal with uncertainty in the system, so type-2 fuzzy set greatly strengthens the processing ability of fuzzy system for uncertainty and nonlinearity, and it has good performance in nonlinear systems with high uncertainty. Therefore, type-2 fuzzy systems have strong generalization ability. And the tensor structure is also good at dealing with uncertain systems, which can improve the generalization performance of the system.
Table 10

Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for Asteroid. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1)

DatasetMethodTraining (MAE)Testing (MAE)Training (MSE)Testing (MSE)Training (RMSE)Testing (RMSE)
MeanStdMeanStdMeanStdMeanStdMeanStdMeanStd
AsteroidTSFNN2.30e−023.02e046.06e+012.37e+031.03e033.30e051.92e+079.09e+083.22e025.13e041.14e+024.38e+03
TT2-RVFL2.31e−023.03e−042.36e−027.23e−041.04e−033.36e−051.12e−039.40e−053.22e025.23e−043.34e−021.39e−03
TT2-ELM2.30e−023.03e−042.36e−027.28e−041.03e033.33e−051.12e−039.58e−053.22e025.19e−043.35e−021.42e−03
OP-ELM2.38e−026.04e−042.77e+021.33e+041.11e−035.60e−051.32e+117.64e+123.33e−028.33e−047.58e+033.63e+05
TROP-ELM2.38e−026.10e−042.36e+021.45e+041.11e−035.67e−051.58e+111.12e+133.33e−028.42e−046.47e+033.98e+05
IT2-FNN2.34e−023.08e−042.35e027.00e041.08e−033.63e−051.10e−038.64e053.29e−025.53e−043.32e021.30e03
eT2QFNN2.35e−025.84e−042.44e−021.31e−031.15e−035.47e−051.16e−031.17e−043.39e−027.87e−043.40e−021.56e−03
BD-ELM2.35e−024.82e−042.36e−027.94e−041.08e−034.73e−051.11e−039.42e−053.29e−027.09e−043.33e−021.40e−03
RNN-LM2.32e−023.17e−042.36e−027.30e−041.05e−033.66e−051.10e039.21e−053.24e−025.65e−043.32e021.37e−03
RNN-BFGS2.36e−026.89e−042.37e−029.24e−041.10e−036.19e−051.11e−039.83e−053.31e−028.65e−043.33e−021.44e−03
LSTM2.87e021.12e−024.09e−024.01e−022.14e−032.36e−034.01e−038.51e−034.24e−021.84e−025.01e−023.87e−02
The merits of type-2 fuzzy sets, and the tensor structure are inherited by the tensor-based type-2 fuzzy system. On the basis of the above analysis, TT2-RVFL and TT2-ELM have good performance on Asteroid dataset, their training error and testing error are minimal in Table 9. Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for Covid19_Beijing, Covid19_Shanghai, Covid19_Tianjin, Covid19_Chongqing. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1) Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for Covid19_Arizona, Covid19_Washington, Covid19_California, Covid19_Illinois. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1) From the test performance of TSFNN, since TSFNN contains TT2-RVFL and TT2-ELM, it makes up for the defect of insufficient generalization ability of TROP-ELM in nonlinear system. This also proves the excellent generalization ability of type-2 fuzzy systems, the advantages of the stacked tensor-based hybrid single fuzzy neural networks indicate that the stacked way of networks designing can inherit the merits of the used algorithms, and the stacked structure of the three algorithms are complementary with each other.

Novel corona virus 2019 dataset

COVID-19 affected cases in the Corona Virus 2019 dataset contains date information label. This dataset contains daily level information about the number of affected cases, the number of deaths and the rehabilitation of the new coronavirus in 2019. It is worth noting that this is a time series data, so any number of cases given a fixed date is cumulative. the data from national centers for disease control and prevention are collected in github. The data is updated daily. Eight cities in Beijing, Shanghai, Chongqing Tianjin, Arizona, Washington, California and Illinois are used for testing, and the time stamp ranges from 22 Jan, 2020 to 29 May, 2021. The data for Novel Corona Virus 2019 Dataset in Beijing, Shanghai, Tianjin and Chongqing The data for Novel Corona Virus 2019 Dataset in Arizona, Washington, California and Illinois There is no doubt that the extracted data of eight cities is composed of time series dataset, and is a small-scale dataset with three attributes. Four of the eight cities that are selected were from China, four were from the United States, the outbreaks in both regions were predicted. Although the data set itself is a small-scale one, the attribute of the data set is only three. Obviously, attribute is not sufficient. The performance of the proposed network is tested in the case that the feature attributes of the dataset are insufficient. The comparison results are shown in Tables 11 and 12. Figure 8 is the results that is carried out based on samples and attributes from four datasets, Beijing, Shanghai, Tianjin and Chongqing. From the results, TSFNN has the best performance compared with the other four algorithms. By comparing the data results in Tables 11 and 12, on the whole, the data results of the five comparison methods are similar. Moreover, from the data results, the results of Beijing, Shanghai, Tianjin and Chongqing are significantly worse than those of Arizona, Washington, California and Illinois.
Table 11

Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for Covid19_Beijing, Covid19_Shanghai, Covid19_Tianjin, Covid19_Chongqing. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1)

DatasetMethodTraining (MAE)Testing (MAE)Training (MSE)Testing (MSE)Training (RMSE)Testing (RMSE)
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
Covid19_BeijingTSFNN3.56e+012.61e+003.65e+014.26e+003.20e+033.86e+023.41e+037.93e+025.64e+013.47e+005.80e+016.91e+00
TT2-RVFL3.88e+012.31e+003.95e+012.69e+003.91e+033.54e+024.01e+038.51e+026.23e+012.84e+006.30e+016.70e+00
TT2-ELM3.85e+012.29e+003.92e+012.70e+003.87e+033.48e+023.97e+038.36e+026.21e+012.83e+006.26e+016.61e+00
OP-ELM4.80e+017.15e+004.88e+017.88e+005.85e+031.88e+036.01e+032.25e+037.57e+011.13e+017.63e+011.33e+01
TROP-ELM4.79e+016.84e+004.85e+017.37e+005.80e+031.82e+035.93e+032.11e+037.53e+011.10e+017.59e+011.28e+01
IT2-FNN4.20e+012.71e+004.25e+013.17e+004.41e+034.40e+024.49e+039.78e+026.63e+013.34e+006.66e+017.33e+00
eT2QFNN2.58e+015.76e+003.20e+012.33e+012.02e+039.59e+021.85e+038.03e+034.39e+019.70e+003.58e+012.39e+01
BD-ELM3.91e+016.39e+004.94e+012.12e+023.90e+031.87e+035.00e+042.01e+066.20e+017.06e+007.23e+012.12e+02
RNN-LM3.98e+012.91e+004.12e+013.37e+004.00e+034.81e+024.15e+039.26e+026.32e+013.24e+006.40e+016.58e+00
RNN-BFGS4.33e+015.89e+004.45e+016.73e+004.78e+031.03e+034.92e+031.51e+036.89e+015.78e+006.96e+018.95e+00
LSTM
Covid19_ShanghaiTSFNN1.83e+011.51e+002.04e+012.34e+005.79e+028.75e+017.16e+021.62e+022.40e+011.76e+002.66e+012.93e+00
TT2-RVFL2.31e+011.26e+002.38e+011.67e+009.00e+028.93e+019.71e+021.72e+022.90e+011.61e+003.09e+012.72e+00
TT2-ELM2.22e+011.16e+002.29e+011.57e+008.24e+027.93e+018.93e+021.63e+022.87e+011.37e+002.98e+012.66e+00
OP-ELM4.62e+012.43e+014.68e+012.47e+014.44e+034.67e+034.57e+034.84e+036.00e+012.90e+016.09e+012.95e+01
TROP-ELM4.71e+012.52e+014.76e+012.55e+014.63e+034.94e+034.74e+035.05e+036.10e+013.00e+016.18e+013.04e+01
IT2-FNN2.91e+011.67e+002.96e+012.27e+001.54e+031.25e+021.59e+032.32e+023.92e+011.59e+003.98e+012.90e+00
eT2QFNN2.66e+013.04e+004.87e+011.91e+011.69e+034.57e+023.57e+033.77e+034.09e+014.84e+005.59e+012.10e+01
BD-ELM2.46e+011.12e+012.71e+014.49e+011.21e+035.95e+033.33e+038.52e+043.18e+011.39e+013.50e+014.59e+01
RNN-LM2.59e+012.36e+002.76e+012.83e+001.20e+032.30e+021.37e+033.16e+023.44e+013.40e+003.68e+014.28e+00
RNN-BFGS3.79e+011.05e+014.00e+011.10e+012.52e+031.98e+032.78e+032.20e+034.86e+011.25e+015.10e+011.33e+01
LSTM4.65e+019.63e+005.02e+011.09e+013.73e+031.44e+033.70e+031.50e+036.01e+011.17e+015.98e+011.19e+01
Covid19_TianjinTSFNN5.88e+003.26e016.28e+006.66e−019.29e+011.11e+011.00e+022.32e+019.62e+005.92e−019.95e+001.17e+00
TT2-RVFL6.83e+003.61e−017.02e+006.03e−011.10e+021.01e+011.19e+022.79e+011.03e+014.81e−011.08e+011.26e+00
TT2-ELM6.54e+003.34e−016.73e+006.00e011.05e+029.45e+001.13e+022.66e+011.02e+014.64e011.05e+011.23e+00
OP-ELM1.05e+014.01e+001.07e+014.07e+002.45e+021.60e+022.65e+026.54e+021.51e+014.15e+001.54e+015.21e+00
TROP-ELM1.07e+014.08e+001.09e+014.20e+002.49e+021.63e+022.66e+022.64e+021.52e+014.22e+001.56e+014.90e+00
IT2-FNN8.84e+005.11e−018.97e+006.34e−011.88e+022.13e+011.96e+025.03e+011.37e+017.83e−011.39e+011.78e+00
eT2QFNN6.63e+008.16e−011.51e+011.31e+011.28e+025.28e+014.57e+021.56e+031.12e+011.65e+001.63e+011.38e+01
BD-ELM6.83e+002.92e+001.14e+015.02e+011.27e+023.20e+021.53e+043.00e+051.07e+013.58e+002.29e+011.22e+02
RNN-LM7.65e+009.43e−018.19e+001.10e+001.22e+022.01e+011.46e+025.65e+011.10e+019.16e−011.20e+011.71e+00
RNN-BFGS1.01e+013.15e+001.04e+013.71e+002.10e+029.38e+022.36e+021.19e+031.40e+013.95e+001.47e+014.60e+00
LSTM9.20e+001.30e+009.52e+006.94e−011.72e+024.70e+012.13e+025.73e+011.30e+011.87e+001.45e+011.96e+00
Covid19_ChongqingTSFNN1.29e+011.93e+001.28e+012.82e+007.86e+021.44e+028.28e+023.74e+022.79e+012.65e+002.80e+016.50e+00
TT2-RVFL1.34e+011.77e+001.39e+011.25e+009.13e+021.52e+029.83e+023.77e+022.96e+012.60e+003.07e+016.06e+00
TT2-ELM1.33e+011.79e+001.38e+011.18e+008.68e+021.49e+029.30e+023.62e+022.94e+012.61e+002.99e+015.97e+00
OP-ELM1.91e+014.56e+002.01e+014.89e+001.51e+034.10e+022.08e+037.20e+033.84e+015.45e+004.15e+011.88e+01
TROP-ELM1.92e+014.55e+002.02e+017.58e+001.51e+034.17e+024.31e+031.50e+053.85e+015.54e+004.23e+015.02e+01
IT2-FNN1.83e+012.09e+001.86e+012.30e+001.78e+032.73e+021.83e+036.25e+024.20e+013.28e+004.21e+017.45e+00
eT2QFNN1.18e+014.80e+003.67e+003.93e+008.91e+025.57e+022.96e+013.99e+022.92e+016.06e+003.76e+003.93e+00
BD-ELM1.37e+012.26e+002.64e+012.09e+028.62e+023.15e+024.63e+041.56e+062.91e+014.01e+004.74e+012.10e+02
RNN-LM1.52e+013.86e+001.94e+016.74e+001.06e+033.35e+022.38e+032.84e+033.22e+014.94e+004.44e+012.02e+01
RNN-BFGS1.82e+015.50e+001.99e+015.72e+001.55e+039.06e+022.12e+031.03e+033.90e+015.26e+004.51e+019.41e+00
LSTM
Table 12

Comparison results with TSFNN, TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS and LSTM for Covid19_Arizona, Covid19_Washington, Covid19_California, Covid19_Illinois. The bold parts represent the best performance of eleven algorithms on each dataset (a brief introduction is listed in Table 1)

DatasetMethodTraining (MAE)Testing (MAE)Training (MSE)Testing (MSE)Training (RMSE)Testing (RMSE)
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
Covid19_ArizonaTSFNN3.68e−027.53e033.65e−021.28e−021.49e023.00e031.50e027.36e031.22e011.28e021.17e013.52e02
TT2-RVFL3.88e−028.15e−033.95e−025.16e−031.56e−023.22e−031.61e−027.83e−031.24e−011.34e−021.23e−013.20e−02
TT2-ELM3.86e−028.12e−033.93e−025.19e−031.55e−023.21e−031.61e−027.83e−031.24e−011.35e−021.23e−013.20e−02
OP-ELM4.06e−028.78e−034.11e−025.34e−031.59e−023.37e−031.64e−027.88e−031.25e−011.40e−021.24e−013.19e−02
TROP-ELM4.06e−028.78e−034.12e−025.34e−031.59e−023.37e−031.64e−027.87e−031.25e−011.40e−021.24e−013.18e−02
IT2-FNN3.73e−028.05e−033.77e−025.07e031.63e−023.49e−031.67e−028.34e−031.27e−011.44e−021.25e−013.41e−02
eT2QFNN2.97e029.11e−031.10e−011.49e−012.26e−024.95e−023.88e−021.76e−011.43e−014.64e−021.17e−011.59e−01
BD-ELM4.12e−028.65e−035.65e−022.01e−011.51e−023.08e−035.98e−021.01e+001.22e−011.32e−021.36e−012.03e−01
RNN-LM4.00e−028.62e−037.95e−021.47e−011.55e−023.34e−036.26e−022.19e−011.24e−011.39e−021.75e−011.79e−01
RNN-BFGS4.11e−029.49e−034.43e023.31e−021.63e−023.45e−031.86e−024.00e−021.27e−011.41e−021.28e−014.67e−02
LSTM4.88e−021.49e−024.88e−021.49e−021.72e−027.37e−031.72e−027.37e−031.30e−012.02e−021.30e−012.02e−02
Covid19_WashingtonTSFNN3.26e−026.40e033.27e021.16e−021.38e022.65e031.35e026.29e031.17e011.17e021.12e013.21e02
TT2-RVFL3.47e−026.81e−033.56e−026.17e−031.41e−022.74e−031.48e−026.65e−031.18e−011.20e−021.19e−012.76e−02
TT2-ELM3.39e−026.65e−033.48e−026.30e−031.40e−022.72e−031.48e−026.62e−031.18e−011.20e−021.19e−012.75e−02
OP-ELM3.76e−028.53e−033.85e−027.14e−031.43e−022.87e−031.51e−026.75e−031.19e−011.24e−021.20e−012.79e−02
TROP-ELM3.76e−028.63e−033.85e−027.17e−031.43e−022.88e−031.50e−026.71e−031.19e−011.25e−021.19e−012.77e−02
IT2-FNN4.25e−028.77e−034.34e−024.69e031.47e−023.00e−031.54e−027.29e−031.21e−011.30e−021.21e−012.97e−02
eT2QFNN3.05e025.08e−031.66e−011.73e−012.40e−028.71e−037.27e−022.03e−011.53e−012.50e−021.88e−011.93e−01
BD-ELM3.49e−027.49e−034.70e−021.65e−011.38e−022.70e−034.38e−028.11e−011.17e011.21e−021.29e−011.65e−01
RNN-LM3.86e−028.18e−039.24e−022.03e−011.38e−022.80e−039.49e−023.70e−011.17e011.24e−021.96e−012.38e−01
RNN-BFGS4.41e−029.54e−034.68e−022.82e−021.46e−023.00e−031.74e−023.57e−021.20e−011.28e−021.25e−014.20e−02
LSTM4.93e−021.49e−025.02e−022.46e−021.56e−027.28e−031.71e−021.16e−021.23e−011.91e−021.26e−013.48e−02
Covid19_CaliforniaTSFNN1.84e013.99e021.84e016.83e−024.17e−019.64e024.20e012.32e016.41e018.00e026.13e012.10e−01
TT2-RVFL1.94e−014.33e−021.98e−012.79e−024.31e−011.02e−014.52e−012.49e−016.49e−018.25e−026.44e−011.92e−01
TT2-ELM1.93e−014.31e−021.97e−012.83e−024.29e−011.01e−014.50e−012.48e−016.50e−018.27e−026.43e−011.92e01
OP-ELM2.11e−014.88e−022.15e−012.69e−024.44e−011.08e−014.63e−012.53e−016.60e−018.64e−026.52e−011.94e−01
TROP-ELM2.11e−014.85e−022.15e−012.68e−024.43e−011.08e−014.62e−012.52e−016.60e−018.67e−026.52e−011.93e−01
IT2-FNN1.91e−014.42e−021.93e−012.60e024.58e−011.11e−014.67e−012.68e−016.71e−018.85e−026.50e−012.10e−01
eT2QFNN2.54e−018.36e−021.42e+001.81e+001.39e+009.94e−016.15e+002.21e+011.13e+003.41e−011.53e+001.95e+00
BD-ELM2.03e−014.51e−023.89e−012.95e+004.20e019.75e−029.41e+003.00e+026.43e−018.10e−028.15e−012.96e+00
RNN-LM2.12e−014.79e−024.77e−011.06e+004.35e−011.05e−012.46e+009.98e+006.55e−018.26e−029.85e−011.22e+00
RNN-BFGS2.17e−015.45e−022.32e−011.70e−014.61e−011.10e−015.16e−011.08e+006.73e−018.44e−026.68e−012.65e−01
LSTM2.50e−015.82e−022.41e−019.68e−024.82e−011.49e−014.99e−012.81e−016.88e−019.41e−026.72e−012.18e−01
Covid19_IllinoisTSFNN6.93e021.34e026.68e022.24e−025.53e−021.05e025.52e022.55e022.34e−012.34e022.26e−016.29e−02
TT2-RVFL7.08e−021.39e−027.29e−021.23e025.73e−021.11e−026.09e−022.77e−022.37e−012.42e−022.40e−015.62e02
TT2-ELM6.97e−021.37e−027.18e−021.25e−025.71e−021.11e−026.08e−022.77e−022.38e−012.43e−022.40e−015.62e02
OP-ELM7.75e−021.63e−027.95e−021.30e−025.83e−021.17e−026.20e−022.79e−022.40e−012.54e−022.43e−015.60e−02
TROP-ELM7.74e−021.61e−027.94e−021.29e−025.83e−021.18e−026.20e−022.79e−022.40e−012.54e−022.43e−015.59e−02
IT2-FNN7.97e−021.63e−028.12e−021.05e−026.08e−021.23e−026.36e−022.98e−022.45e−012.61e−022.45e−016.03e−02
eT2QFNN8.15e−024.95e−021.53e−011.97e−012.95e−019.15e−017.23e−022.52e−014.17e−013.48e−011.67e012.11e−01
BD-ELM8.28e−021.66e−022.02e−011.77e+005.51e021.07e−023.31e+009.39e+012.33e012.38e−023.53e−011.78e+00
RNN-LM7.79e−021.60e−021.89e−013.78e−015.74e−021.20e−023.74e−011.27e+002.38e−012.59e−024.01e−014.62e−01
RNN-BFGS8.55e−021.87e−029.18e−025.08e−026.09e−021.31e−027.14e−021.07e−012.45e−012.64e−022.53e−018.62e−02
LSTM9.86e−022.45e−021.07e−018.17e−026.42e−022.33e−027.77e−028.39e−022.51e−013.70e−022.63e−019.13e−02
Fig. 8

The data for Novel Corona Virus 2019 Dataset in Beijing, Shanghai, Tianjin and Chongqing

Similarly, the same operation is performed on the four datasets of Arizona, Washington, California and Illinois, and the results are shown in Fig. 9.
Fig. 9

The data for Novel Corona Virus 2019 Dataset in Arizona, Washington, California and Illinois

It can be regarded from Figs. 8 and 9 that, due to the characteristics of the COVID-19 dataset itself, the overall trend of the data is rising, reflecting the characteristics of its time series, and the growth rate of the curve reflects the situation of the new coronavirus. From the figures, these eight datasets are more suitable for regression problems, and the data is relatively stable. Therefore, there is little difference between TSFNN and four comparison methods in Tables 11 and 12. Compared with Figs. 8 and 9, the curve in Fig. 9 is smoother and the overall trend is more obvious. Although the curve in Fig. 8 shows an overall upward trend, the data soon go back to the stable state. Therefore, the data in Fig. 9 is more suitable for forecasting regression problems than the data in Fig. 8, which is also the reason that in Tables 11 and 12 the overall performance of the data sets corresponding to four cities in China on five algorithms is not as good as that of the data sets corresponding to four cities in the United States. Through the above analysis, the data characteristics of the data set itself lead to data differences in Tables 11 and 12. Because of this difference, four data sets in China are comparable to four data sets in the United States. On the premise that the eight data sets are small-scale time series, the data characteristics of the four data sets in China are quite different, and the performance in the regression problem is poor, while the data in the United States are more obvious and more suitable for regression analysis. In addition, four data sets such as Beijing can be regarded as data sets with more complex data structures than four data sets such as Arizona, and the features are more diverse, not limited to reflect the upward trend of data. Obviously, combined with the actual current situation. The four data sets related to the United States can better predict the future epidemic situation in the United States. As far as the data in the data set itself is concerned, the four data sets such as Arizona are more “pure” and suitable for regression problems. Therefore, the performance of the five methods on these four data sets is better, while the performance of Beijing and other four and data sets is relatively poor. According to the overall data of Tables 11 and 12, TSFNN still performs excellent and satisfactory compared with the other ten comparison methods. In four datasets of Beijing, Shanghai, Tianjin and Chongqing, although the overall data is relatively poor due to the problems with the data itself, the performance of TSFNN is still the best in these four datasets. TSFNN also performs best in Arizona, Washington, California and Illinois datasets which has the better data than 4 datasets for China. This shows in the excellent feature extraction and generalization ability of TSFNN. It is demonstrated that the scheme of tensor stacked neural network can integrate the advantages of its members and enhance the ability of data feature extraction. Results of Friedman test on ten datasets. The bold parts represent the best performance of eleven algorithms on each dataset (The testing results are listed in bracket)

Discussion of the eleven algorithms

The effectiveness and performance of the proposed TSFNN are proved via the simulation experiment in Sects. 4.1–4.3. On the whole, the comparison results on a total of 30 different data sets show that the performance of TSFNN is excellent. The parameter settings of the three comparison algorithms follow three principles: (1) The common parameters, such as the number of hidden nodes, are set as the same value. (2) The remaining parameters are the same as the settings of original literature. (3) The division of testing set and training set is the same. Based on the above principles, the performance evaluation of the model is completed. Comprehensively considering of the comparison results are listed in Tables 2, 3, 5, 6, 8, 9, 11 and 12, the performance of TSFNN is slightly better than four deep network models that are RNN-LM, RNN-BFGS, LSTM and eT2QFNN. The descending order performance rank of four deep network models is RNN-LM, eT2QFNN, RNN-BFGS and LSTM. TSFNN’s constituent members which are TT2-RVFL, TT2-ELM and TROP-ELM, which can compete with deep network models, but overall slightly worse. However, the performance of TSFNN is comparable to four deep networks. Although the four deep network models did not reach their extreme performance, their norm performance is shown at least. Therefore, the performance of TT2-RVFL, TT2-ELM and TROP-ELM is enhanced via the suggested stacked neural network scheme. The performance of TSFNN is the combined performance of its constituent members. Results of Friedman test on twenty-five data sets for eleven methods (TT2-RVFL, TT2-ELM, OP-ELM, TROP-ELM, IT2-FNN, eT2QFNN, BD-ELM, RNN-LM, RNN-BFGS, LSTM and TSFNN) are listed in Table 13. TSFNN obtains the best performance with respect to the training error and test error are all smallest. RNN-LM’s performance is second position compared with TSFNN.
Table 13

Results of Friedman test on ten datasets. The bold parts represent the best performance of eleven algorithms on each dataset (The testing results are listed in bracket)

AlgorithmMean rank\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\chi ^2$$\end{document}χ2p Value
TSFNN3.22 (3.39)254698.983 (228535.999)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$<0.05$$\end{document}<0.05 (\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$<0.05$$\end{document}<0.05)
TT2-RVFL5.95 (5.87)
TT2-ELM5.02 (5.24)
OP-ELM7.06 (6.79)
TROP-ELM7.05 (6.78)
TT2-FNN7.46 (6.78)
eT2QFNN6.22 (7.96)
BD-ELM7.79 (7.41)
RNN-LM3.25 (3.43)
RNN-BFGS6.29 (5.88)
LSTM6.69 (6.46)
Based on performance analysis of all simulation experiments, the regression results on different complexity datasets show that the main reasons for the excellent performance of the proposed TSFNN are as below: (1) TSFNN stacks the hidden layers of its constituent members into a tensor structure and the type-2 fuzzy sets are fused into the tensor structure in this process. The stacked structure enhances the model’s feature extraction ability, uncertainty modeling ability and generalization ability. (2) TSFNN is able to inherit the advantages of its constituent members, and amplify them through superimposed structures. (3) Although the shortcomings of TSFNN’s constituent members are also inherited, these shortcomings are compensated by the model itself and other members via the suggested stacked strategy. (4) TSFNN uses the tensor regression algorithm which unfolds the 3-D tensor into three matrices through the tensor unfolded method, and the learning problem is solved via the matrix regression. (5) The regression stage uses Tikhonov regularization to constrain the model. In summary, the proposed TSFNN is an effective fuzzy system modeling and a neural network fusion method, which enriches the content of fuzzy system modeling, especially the construction of high-order fuzzy systems.

Conclusions

In this paper, a tensor-based fuzzy stacked neural network (TSFNN) model was proposed, which is a stacked neural network with several type-2 fuzzy neural network models. To the TSFNN network, the TT2-RVFL, TT2-ELM and TROP-ELM are used to form this stacked structure. In the work, TT2-RVFL and TT2-ELM are optimized by using the kernel space method to enhance their performance. Moreover, TSFNN fuses the hidden layer output networks of its member networks by tensor, while the tensor-based stacked system inherits the advantages of the type-2 fuzzy sets and the fuzzy inference ability of the fuzzy system via tensor structure. Simultaneously, the structure also inherits the merit that is generated by MRSR method and pruning method in TROP-ELM, thus, the advantages of this TROP-ELM are also extracted by tensor structure in fusing the proposed TSFNN. Because tensor structure can concentrate the advantages of member sub-networks, TSFNN obtains great generalization ability and anti-noise ability due to its stacked strategy. TSFNN also inherits the defects of its member networks. In general, the ability of TSFNN to concentrate the advantages of member networks will make up for the shortcomings of some member networks. When the data is too complex, there will still be underfitting. By and large, the proposed TSFNN algorithm has excellent generalization ability, anti-noise ability and feature extraction ability. Under effects of the five reasons or principles which are listed in Sect. 4.4, the proposed TSFNN can compete with normal deep network, such as RNN-LM, RNN-BFGS and LSTM. For the purpose of obtaining a fast model of the data set, a tensor unfolded method is used. Then, the regression results are obtained by matrix regression that is unfolded from a tensor. Tensor regression and tensor equation can also be used in this direction, which is a future optimization direction. These capabilities are demonstrated and validated on ten UCI standard datasets and three real world datasets. The TSFNN algorithm supplements the tensor-based model optimization and model combination methods, indicating that the tensor structure stacked neural network is a feasible neural networks combination method. Moreover, the proposed fuzzy network stacked strategy can be considered as an effective method for constructing higher order fuzzy systems.
  13 in total

1.  OP-ELM: optimally pruned extreme learning machine.

Authors:  Yoan Miche; Antti Sorjamaa; Patrick Bas; Olli Simula; Christian Jutten; Amaury Lendasse
Journal:  IEEE Trans Neural Netw       Date:  2009-12-08

2.  Long short-term memory.

Authors:  S Hochreiter; J Schmidhuber
Journal:  Neural Comput       Date:  1997-11-15       Impact factor: 2.026

3.  A Maximally Split and Relaxed ADMM for Regularized Extreme Learning Machines.

Authors:  Xiaoping Lai; Jiuwen Cao; Xiaofeng Huang; Tianlei Wang; Zhiping Lin
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2019-08-06       Impact factor: 10.451

4.  A Novel Concept Drift Detection Method for Incremental Learning in Nonstationary Environments.

Authors:  Zhe Yang; Sameer Al-Dahidi; Piero Baraldi; Enrico Zio; Lorenzo Montelatici
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2019-03-26       Impact factor: 10.451

5.  Cascaded Multi-Column RVFL+ Classifier for Single-Modal Neuroimaging-Based Diagnosis of Parkinson's Disease.

Authors:  Jun Shi; Zeyu Xue; Yakang Dai; Bo Peng; Yun Dong; Qi Zhang; Yingchun Zhang
Journal:  IEEE Trans Biomed Eng       Date:  2018-12-24       Impact factor: 4.538

6.  Biased Dropout and Crossmap Dropout: Learning towards effective Dropout regularization in convolutional neural network.

Authors:  Alvin Poernomo; Dae-Ki Kang
Journal:  Neural Netw       Date:  2018-04-09

7.  Regularized correntropy criterion based semi-supervised ELM.

Authors:  Jie Yang; Jiuwen Cao; Tianlei Wang; Anke Xue; Badong Chen
Journal:  Neural Netw       Date:  2019-10-09

8.  Transcranial Sonography Based Diagnosis Of Parkinson's Disease Via Cascaded Kernel RVFL.

Authors:  Zeyu Xue; Jun Shi; Yakang Dai; Yun Dong; Qi Zhang; Yingchun Zhang
Journal:  Annu Int Conf IEEE Eng Med Biol Soc       Date:  2018-07

9.  ELM embedded discriminative dictionary learning for image classification.

Authors:  Yijie Zeng; Yue Li; Jichao Chen; Xiaofan Jia; Guang-Bin Huang
Journal:  Neural Netw       Date:  2019-12-20

10.  Fault detection and classification in electrical power transmission system using artificial neural network.

Authors:  Majid Jamil; Sanjeev Kumar Sharma; Rajveer Singh
Journal:  Springerplus       Date:  2015-07-09
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.