Literature DB >> 35694521

Fault Detection and Isolation of Non-Gaussian and Nonlinear Processes Based on Statistics Pattern Analysis and the k-Nearest Neighbor Method.

Zhe Zhou1, Jian Wang1, Chunjie Yang2, Chenglin Wen3, Zuxin Li4.   

Abstract

Only low-order information of process data (i.e., mean, variance, and covariance) was considered in the principal component analysis (PCA)-based process monitoring method. Consequently, it cannot deal with continuous processes with strong dynamics, nonlinearity, and non-Gaussianity. To this aim, the statistics pattern analysis (SPA)-based process monitoring method achieves better monitoring results by extracting higher-order statistics (HOS) of the process variables. However, the extracted statistics do not strictly follow a Gaussian distribution, making the estimated control limits in Hotelling-T 2 and squared prediction error (SPE) charts inaccurate, resulting in unsatisfactory monitoring performance. In order to solve this problem, this paper presents a novel process monitoring method using SPA and the k-nearest neighbor algorithm. In the proposed method, first, the statistics of process variables are calculated through SPA. Then, the k-nearest neighbor (kNN) method is used to monitor the extracted statistics. The kNN method only uses the paired distance of samples to perform fault detection. It has no strict requirements for data distribution. Hence, the proposed method can overcome the problems caused by the non-Gaussianity and nonlinearity of statistics. In addition, the potential of the proposed method in early fault detection or safety alarm and fault isolation is explored. The proposed method can isolate which variable or its statistic is faulty. Finally, the numerical examples and Tennessee Eastman benchmark process illustrate the effectiveness of the proposed method.
© 2022 The Authors. Published by American Chemical Society.

Entities:  

Year:  2022        PMID: 35694521      PMCID: PMC9178650          DOI: 10.1021/acsomega.2c01279

Source DB:  PubMed          Journal:  ACS Omega        ISSN: 2470-1343


Introduction

With the rapid development of industrial processes and the increase in process complexity, a large amount of process operation data has been generated. Fully mining and using the valuable information contained in the process data will be beneficial to early fault detection, thereby minimizing downtime and improving the safety of process operation.[1] In this context, data-driven multivariate statistical process monitoring (MSPM) methods have been developed by leaps and bounds, where principal component analysis (PCA) methods are the most widely used. However, only low-order information of process data (i.e., mean, variance, and covariance) was considered, and high-order statistics (HOS) were ignored in the PCA-based process monitoring method. In addition, the threshold of Hotelling’s T2 and SPE are calculated based on the premise that process variables satisfy a Gaussian distribution.[2] Due to nonlinearity, non-Gaussianity, dynamic, and multimodality in industrial processes, it is difficult to satisfy this assumption in practice. Therefore, the traditional PCA-based process monitoring method has poor monitoring performance when facing the above problems.[3−7] Recently, there have been many studies on nonlinear process monitoring problems. To capture the nonlinear correlation structure, Yu et al.(8) proposed a robust, nonlinear and sparse PCA method. Aiming at the excessive modeling redundancy problem caused by infinite-order mapping, a monitoring method based on constructing polynomial mapping is proposed in ref (9), which significantly improves the fault detection performance of nonlinear processes. Jiang and Yan[10] proposed a parallel PCA-KPCA monitoring method for a process with linearly correlated and nonlinearly related variables. In order to detect the incipient faults of nonlinear industrial processes effectively, an enhanced KPCA method is proposed in ref (11). Mansouri et al.(12) introduced a generalized likelihood ratio test into the KPCA method for fault detection of nonlinear processes. However, these methods above do not consider the high-order statistics of the process data, and the calculation is complicated. In addition, there is a lot of work to solve the problem of fault diagnosis of complex industrial processes using hybrid methods including Bayesian and data-driven approaches. Yu et al.(13) proposed a two-stage fault diagnosis method that combines independent component analysis (ICA) with the Bayesian network (BN) to solve the problem in that conventional MSPM methods cannot isolate fault from unmonitored process variables. In order to reduce the cost of monitoring and alarm flooding, Wang et al.(14) proposed a fault diagnosis technique that combines semi-parametric PCA and BN, which also achieves location of the root cause of faults in process variables. A PCA-BN hybrid method with multiple likelihood evidence for fault diagnosis is proposed in ref (15). This method enables BN to update more information about faults and improves process monitoring performance. Reference (16) uses KPCA and BN for fault diagnosis, which not only diagnoses the root cause of the fault but also shows the propagation path of the fault. He and Wang[17] proposed a statistics pattern analysis (SPA) monitoring framework to overcome these problems. This method extracts HOS from process data, such as skewness and kurtosis. Compared with the traditional MSPM methods that only use low-order statistics information, it has advantages in characterizing the non-Gaussianity and nonlinearity of the process data. They used the SPA to monitor the batch process because it calculates the statistics of batch process data, and data preprocessing is avoided. In ref (2), the sliding window is used to design the SPA to monitor the continuous process and obtain better monitoring performance on the Tennessee Eastman (TE) platform. In addition, the SPA is identified as a novel generation of the SPM method and can handle the 4 V challenges of big data.[18−21] Although the SPA framework has many advantages in monitoring processes of nonlinearity or multimodality,[22−25] it is worth noting that the statistics extracted by the SPA is not a Gaussian distribution, in addition, there is a nonlinear relationship between the extracted statistics. Therefore, it is not suitable to use PCA combined with T2 and SPE charts to monitor the extracted statistics. Ma et al.(26) considered the nonlinearity of statistics: they used the advantages of kernel PCA in processing nonlinear data and proposed a process monitoring method based on statistics KPCA (SKPCA). However, SKPCA only considers the nonlinear problem of statistics, and the non-Gaussianity of statistics is ignored. In this paper, a new process monitoring method, which combines the superiority of SPA in extracting HOS information with kNN in processing the non-Gaussian and nonlinearity of data samples, is proposed to deal with the defects of those methods above. The fault detection method using the kNN rule (FD-kNN) only uses the distance between neighbors to perform fault detection; there is no restriction on the data distribution.[27,28] The qualitative and quantitative analysis for the non-Gaussianity and nonlinearity of statistics are also presented. Meanwhile, the effect of window width on the non-Gaussianity of statistics is investigated. In addition, we explore the potential of the proposed method in fault diagnosis. Specifically, the variable contribution by kNN (VCkNN) is used to determine the contribution of different statistics to the detection index. kNN is a distribution-free method. The proposed method uses the kNN method to monitor process statistics to achieve fault detection and fault isolation. In terms of fault detection, compared with the conventional statistical techniques, it can overcome the non-Gaussian and nonlinear problems of process variables (statistics); in terms of fault isolation, compared to traditional contribution analysis methods, the proposed kNN-based variable contribution method is not affected by fault smearing. The experiments on numerical examples and TE processes illustrate the effectiveness of the proposed process monitoring method. The contributions of the proposed method are as follows: The qualitative and quantitative analysis for the extracted statistics of process data is conducted and verifies that they are non-Gaussian and nonlinear. The statistics extracted from the process data are monitored by using the k-nearest neighbor method, which avoids the problems caused by non-linear, non-Gaussian characteristics of statistics. The employed variable contribution by kNN can indicate the statistic of the faulty variable, which may be useful for root cause analysis and has potential to perform incipient fault detection.

Related Work

The core foundation of the proposed method is related to the SPA and FD-kNN. This section will briefly review these two methods.

SPA

We use Y to denote the data in a window of the original samples:where l is the window width, d is the last sampling time, and m represents the number of variables. The sampling data of each window can be used to calculate a statistics pattern (SP), which is composed of first-order statistics (FOS), second-order statistics (SOS), and HOS. As shown in eq where μ denotes the FOS, i.e., variable means (μ); Σ denotes the SOS, which include variance (v), correlation (r), autocorrelation (r), and cross-correlation (r); Ξ denotes the HOS, which is composed of skewness (γ) and kurtosis (κ). These statistics can be calculated as followswhere c represents the maximum lag time. An SP vector can be obtained using the above equations for each window of process data, and all the SP vectors of training data can be stacked into a matrix, denoted as SPs. The number of columns and rows in SPs represents the number of statistics and windows, respectively, thereby realizing statistics extraction from training data. In ref (2), the SP matrix is then decomposed by PCA, and T2 and SPE charts are established. For online monitoring, statistics are also extracted based on the data obtained at current and the previous (l – 1) sampling times. After that, T2 and SPE of extracted statistics are calculated to perform fault detection.

FD-kNN

The kNN rule was initially widely used in pattern classification. In December 2006, it was selected as one of the top ten classic algorithms in data mining. FD-kNN was first proposed by He and Wang.[29] The main idea is to measure the difference between samples by distance; that is, normal samples and training samples are similar, but fault samples and training samples are significantly different. Training phase (determine the detection control limit): Use the Euclidean distance to get the kNNs of each training sample. Calculate the distance statistic D2.where D2 represents the average squared distance between the pth sample and its k neighbors, d2 denotes the squared Euclidean distance between the pth sample and its qth nearest neighbor. Establish the control limit Dα2 for fault detection. There are many ways to estimate Dα2, such as the estimation using a noncentral chi-square distribution,[29] kernel density estimation (KDE). The (1 – α)-empirical quartile[30] of D2 is used as the threshold in our proposed method.where D(2, i = 1, ..., n is the result of D2 obtained by eq in ascending order. (⌊n(1 – α)⌋) means discard the decimal part of n(1 – α) and keep the integer part. Detection phase: For a sample x to be tested, find its kNNs from the training set. Calculate D2 between x and its k neighbors using eq . Compare D2 with the threshold Dα2. If D2 > Dα2, the process is considered abnormal. Otherwise, it is normal.

Analysis for the Non-Gaussianity and Nonlinearity of Statistics

Although the PCA fault detection method based on SPA (SPCA) improves the monitoring performance by using HOS information to describe the complex characteristics of process data, the statistics extracted by SPA is not satisfied by a Gaussian distribution. Also, the relationship between different statistics is nonlinear. Faced with the above problems, the SPCA is theoretically flawed. In this section, the qualitative and quantitative analysis for the non-Gaussianity and nonlinearity of statistics are presented. In addition, the effect of window width on the non-Gaussianity of statistics is investigated.

Non-Gaussianity of Statistics

Qualitative analysis: The central limit theorem (CLT) says that under rather general circumstances, the sum of independent normalized random variables is a normal distribution at the limit.[2] It needs to be emphasized that CLT assumes that the samples are drawn independently and the number of samples (in a window) is sufficiently large. In practice, however, the wider the window’s width, the greater the delay in giving the monitoring results. Moreover, the quantity of samples is limited by the specific process. In addition, if the overlap rate of two adjacent windows is high, the basic assumption of independent samples will be violated. Therefore, the extracted statistics are difficult to meet the assumption of CLT, which makes them non-Gaussian. Quantitative analysis: The non-Gaussianity of each variable can be evaluated by negative entropy (NE).[31] A kind of robust estimation of NE[32] is employed to calculate the non-Gaussianity of statistics. NE is non-negative, and it is 0 only when a variable is the standard normal distribution. The larger the NE of a random variable, the more it deviates from the standard normal distribution. For variable x, its NE can be expressed as follows:where G(v) is the entropy of the standard normal distributed variable v, G(x*) is the entropy of x*, while x* is the normalized result of x. The G(v) can be calculated as a constant: As for G(x*), it can be calculated as follows:[33]where x*(1), x*(2), ..., x*( denote M measurements of x* arranged in the ascending order. Substituting eqs and 15 into eq yields the NE of x. It is hard to achieve zero for NE even if x has a Gaussian distribution because NE is affected by the measurements of x and the number of measurements. According to ref (34), the value 10–3 is taken as the threshold.

The Nonlinearity of Statistics

Qualitative analysis: If process variables have a nonlinear relationship, the statistics of these variables cannot eliminate the nonlinear relationship of original variables.[26] In addition, it can be seen from the calculation formula of the statistics that there is also a nonlinear relationship between the statistics of different orders. Quantitative analysis: We use the surrogate data (SD) method[35−37] to detect the nonlinearity of statistics. This method first designates a given linear process as the null hypothesis, then generates SD sets consistent with the null hypothesis, and finally calculates a discriminating statistic for the original data and each SD set. If the value calculated for the original data is significantly different from the set of values calculated from the SD, then the null hypothesis is rejected, and nonlinearity is determined. The null hypothesis H0: the data come from a linear Gaussian process. Generate surrogate data sets for original data using Fourier transform (FT). Fourier transform: For variable x with N samples, its FT can be expressed as follows: Randomize the phase:where φ is a random variable and obeys a uniform distribution in the range of [0, 2π]. Inverse Fourier transform: After the inverse Fourier transform, the substitute data x(n) of the original data can be obtained. Calculate the statistic T of a significance test.where R denotes the redundancy of the original data and R̅ and σ denote the mean and standard deviation of redundancy of the surrogate data set, respectively. Redundancy can be calculated using eq .[38] The covariance matrix of variables x and y is denoted as C. c and λ are diagonal elements and eigenvalues of the covariance matrix C, respectively. Determine whether the null hypothesis is true. Under the condition that the significance level is 0.05, the threshold of significance testing is 1.96. If T ⩾ 1.96, the null hypothesis H0 is rejected and original data is nonlinear; otherwise, accept the null hypothesis and original data is linear.

Illustrative Example

A simple static nonlinear example is used to illustrate the non-Gaussianity and nonlinearity of statistics. The process model is as follows:where t is a random variable that obeys a uniform distribution in the range of [−0.5, 0.5] and e is the Gaussian noise with 0 mean and variance of 0.01. Ten thousand normal samples are generated by eq for extracting statistics by SPA. The window width is 10 (i.e., l = 10), and the sliding step is 1. Each SP consists of 8 statistics, which include the FOS, SOS, and HOS of x and y. Figure shows that the multivariate distribution of different statistics is non-Gaussian. The non-Gaussianity of statistics is quantified by estimating the NE. It can be concluded from Table that the NE of the statistics is almost always smaller than that of the original variable, but it is still greater than 10–3. Therefore, the statistics extracted by SPA are indeed non-Gaussian.
Figure 1

Density histogram intensity map for statistics: (a) mean of x and y; (b) variance of x and y; (c) skewness of x and y; (d) kurtosis of x and y; (e) original variables x and y; (f) Gaussian random variables p and q.

Table 1

Negative Entropy (NE) of Variables

variableNEvariableNE
x0.1281y0.4838
mean of x0.0404mean of y0.0456
variance of x0.0436variance of y0.0483
skewness of x0.0458skewness of y0.0454
kurtosis of x0.2357kurtosis of y0.3140
Density histogram intensity map for statistics: (a) mean of x and y; (b) variance of x and y; (c) skewness of x and y; (d) kurtosis of x and y; (e) original variables x and y; (f) Gaussian random variables p and q. The NE of the statistics can be affected by the window width and window shifting step. For a given number of statistics, the larger the window width, the smaller the NE of the statistics, that is, the weaker the non-Gaussianity, especially for HOS because the NE of the statistics is related to the number of the statistics determined by the window width and the window shifting step; therefore, as shown in Figure , there is an approximately monotonous decreasing trend between the NE (or non-Gaussianity) of the statistics and the window width. Figure a,b shows the nonlinearity detection results of the statistics (mean of x and mean of y; variance of x and kurtosis of x) using the surrogate data method. It can be seen that T is always greater than 1.96. Therefore, the null hypothesis should be rejected with a probability of 95%; the statistics are nonlinear.
Figure 2

Effect of window width on the NE of the statistics.

Figure 3

Nonlinearity detection results of the statistics. (a) Mean of and ; (b) variance of and kurtosis of ; (c) original variables and ; (d) Gaussian random variables and .

Effect of window width on the NE of the statistics. Nonlinearity detection results of the statistics. (a) Mean of and ; (b) variance of and kurtosis of ; (c) original variables and ; (d) Gaussian random variables and .

Process Monitoring Method Using the SPA and kNN Algorithm (SKNN)

Fault Detection Based on SKNN

The SKNN integrates the advantages of SPA in extracting HOS information and kNN in dealing with the non-Gaussian and nonlinearity of data samples. Training step: Extract the statistics matrix Y (SPs) from the training sets generated under the normal operation condition (NOC). Find kNNs for each sample (SP) in Y using eq . Calculate the distance statistic using eq . Determine the threshold Dα2 for fault detection using eq . Online monitoring: For a sample x, a statistic sample s can be extracted by SPA. Find kNNs of s from Y. Calculate D2 using eq . Compare D2 with the threshold Dα2. If D2 > Dα2, it is considered as a faulty sample. Otherwise, it is a normal sample.

Fault Diagnosis Based on SKNN

When a fault is detected, finding the root cause of the failure is the next more important thing. VCkNN[27] is used to determine the contribution of different statistics to the detection index. The contribution of each statistic is not affected by other statistics; that is, there is no correlation. Therefore, the diagnosis result is not affected by the fault smearing. Determine the contributions of statistics: Equation is the form of decomposing D2 into the sum of the contributions of m statistics.where cknn is the contribution from the ith statistic of s to D2. Determine the faulty statistics:where F contains all the faulty statistics diagnosed by SKNN and T represents the threshold for the ith statistic contribution. The threshold can be determined empirically; for example, the maximum contribution of the statistic in the training samples is used as the isolation threshold of the corresponding statistic.[27] The statistics can be flexibly selected for different purposes or process characteristics. For example, considering that the dynamics of the batch process are relatively weak, the only covariance is selected as the second-order statistics.[17] For the design of the window width and sliding step, the data set should be fully used. Otherwise, samples at the end of the data set will not be used. For example, if the size of the data set is 100 when the window width is 10 and the sliding step is 5, 19 SP are obtained exactly, and the data set is fully used; when the window width is 16 and the sliding step is 5, the last four samples of the data set will not be used resulting in the loss of key information of the process, making the model established in the training phase insufficient to describe the process accurately, thereby degrading the monitoring performance. The number of nearest neighbors is selected, and k is determined using cross-validation.

Case Studies

In this section, we use numerical examples and TEP to explore the effectiveness of the proposed method SKNN in fault detection and diagnosis. The kernel-independent component analysis (KICA)[39] method is also used to solve the non-Gaussian and nonlinear problems of process data. Therefore, in addition to the SPCA and SKPCA methods, the monitoring performance of the fault detection method using SPA and KICA (SKICA) has also been verified.

Numerical Simulations

The NOC data are produced according to the following process model adopted from ref (4) uo(k) denotes the correlated input, (k) denotes the noise with 0 mean and variance 0.1.where w(k) is a noise with 0 mean and a standard deviation of 1, v(k) represents a noise with 0 mean and a variance of 0.1. There are four variables (i.e., u1, u2, y1, and y2) used for monitoring. Since the process is dynamic, autocorrelation and cross-correlation are added to the second-order statistics. Therefore, a total of 54 statistics are extracted by SPA. For all four methods, the numbers of generated training samples and validation samples are 10,000 and 5000, respectively. There are 10,000 testing samples, of which the first 500 samples are normal, and the rest are faulty. The fault setting is that the mean value of the first element w1 in w changes from 0 to 4. The specific parameter settings of SPCA, SKPCA, SKICA, and SKNN are shown in Table . For SPCA, the number of principal components (PCs) is determined according to cumulative percent variance (CPV)[40] more than 90%. For SKPCA, the size of the kernel parameter is selected as 5m according to ref (5), and m is the number of variables. For SKICA, the number of independent components (ICs) is the same as that of the PCs. For SKNN, the number of nearest neighbors is 3. For SPCA, SKPCA, SKICA, and SKNN, the control limits of different indices are calculated at a confidence level of 99%.
Table 2

Parameter Settings of SPCA, SKPCA, SKICA, and SKNN

methodsstatisticsPCsICswindow widthwindow sliding stepkernel parameterk
SPCA5413 2515  
SKPCA543 2515270 
SKICA54442515270 
SKNN54  2515 3
Fault Detection: Figure shows the NE of statistics, and we can observe that the NE of different statistics is greater than 10–3. Therefore, the statistics extracted from this dynamic process are non-Gaussian. The FDR and FAR of the four methods are shown in Table (note that the FAR is obtained based on the validation samples). SKNN has the best detection performance among the three methods because it can overcome the problems caused by the nonlinearity and non-Gaussianity of statistics. Figure is the comparison result of the detection performance of the three methods. For SPCA, due to the non-Gaussian and nonlinearity of the statistics, the detection performance of SPCA is not good. The performance of SKPCA is better than SPCA because the KPCA has the advantage in dealing with nonlinearity. Although the SKICA method can also deal with non-Gaussian and nonlinear problems of process data, the fault detection performance of SKICA on experiments is not as good as that of SKNN. This may be due to the following reasons: to reduce the computational load of the SKICA method, ICA is performed after the kernel matrix is dimensionally reduced by PCA, which may cause the information used to extract essential ICs to be ignored.
Figure 4

Measures of non-Gaussianity of statistics for the numerical example.

Table 3

FDR (%) and FAR (%) of SPCA, SKPCA, SKICA, and SKNN for the Numerical Example

 SPCA
SKPCA
SKICA
SKNN
indexT2SPET2SPEI2SPEkNN
FDR20.1910.5713.5697.9557.1097.7999.37
FAR0.90.91.51.80.00.30.9
Figure 5

Fault detection results of SPCA, SKPCA, SKICA, and SKNN for the numerical example. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.

Fault Diagnosis: The change of w has different effects on the four monitored variables. Figure shows the difference between each variable of the training samples and the corresponding variable in the fault samples (only partial results are shown due to limited space). The fourth variable (y2) has the maximum difference. Because the kernel method is difficult or even impossible to find the inverse mapping function from the feature space to the original space, the fault diagnosis of SKPCA and SKICA is not considered in this paper. Figure is the comparison result of the diagnosis performance of SPCA and SKNN. SPCA cannot identify the variable y2 that contributes the most to the fault due to the smearing effect. For SKNN, it does not have a smearing effect because VCkNN is defined in the original variable space (here means statistics space). In Figure c, the variable y2 that contributes the most to the fault is successfully isolated. It is worth noting that the diagnosis result of SKNN indicates that the mean of y2 has the maximum contribution as can be seen from Figure , the mean of y2 increases significantly after the fault is introduced from the 33rd window. Therefore, it is effective to perform early fault detection or safety alarm by observing the mean of y2.
Figure 6

Difference between each variable of the training samples and the corresponding variable in the fault samples.

Figure 7

Fault diagnosis results of SPCA-T2, SPCA-SPE and SKNN for the numerical example. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN.

Figure 8

Man of .

Measures of non-Gaussianity of statistics for the numerical example. Fault detection results of SPCA, SKPCA, SKICA, and SKNN for the numerical example. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN. Difference between each variable of the training samples and the corresponding variable in the fault samples. Fault diagnosis results of SPCA-T2, SPCA-SPE and SKNN for the numerical example. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN. Man of .

TE Benchmark Process

When comparing the performance or effectiveness of process monitoring methods, the TEP is a benchmark choice. In ref (41), Downs and Vogel proposed the simulation platform. There are five major operating units in the TE process, namely, a reactor, a product condenser, a vapor–liquid separator, a recycle compressor, and a product stripper. The process has four kinds of reactants (A, C, D, E), two products (G, H), and contains a catalyst (B) and byproducts (F). The flowchart of the process is given in Figure . There are 11 manipulated variables (No.42–No.52), 22 process measurements (No.1–No.22), and 19 composition variables (No.23–No.41). For detailed information on the 52 monitoring variables and 21 fault patterns, see ref (42).
Figure 9

Flowchart of the Tennessee Eastman process.

Flowchart of the Tennessee Eastman process. The parameter settings of SPCA, SKPCA, and SKNN are shown in Table . For all three methods, the number of training samples and the number of validation samples are 960 and 480, respectively. In addition, there are 960 testing samples where the fault is introduced from the 161st sample. The thresholds of different methods are all calculated at a confidence level of 99%. For second-order statistics, only the significant statistics will be selected. The specific selection method is the same as in ref (2). In order to facilitate the establishment of the model, for a large number of second-order statistics, we only select the numerically significant ones. The specific selection rules are as follows:
Table 4

Parameter Settings of SPCA, SKPCA, SKICA, and SKNN for TEP

methodsstatisticsPCsICswindow widthwindow sliding stepkernel parameterk
SPCA22374 126  
SKPCA22335 1261115 
SKICA22338381261115 
SKNN223  126 3
r is selected only if |r| > 0.5 for more than 70% of the training SPs; ris selected only if |r| > 0.5 for more than 90% of the training SPs; ris selected only if |r| > 0.5 for more than 90% of the training SPs. After statistics selection, a total of 223 statistics of the TE process are used for process monitoring.

Results and Discussion

This section analyzes and discusses the experimental results of the TE process. Fault Detection: From Figure , the statistics extracted from the TE process are non-Gaussian because the NE of the statistics is greater than the threshold of 10–3. The FDR of the four methods are listed in Table . Note that faults 3, 9, and 15 are not considered here because they are difficult to be detected.[1,43,44] For faults 5, 10, 11, 16, 18, 20, and 21, the FDR of the SKNN method is higher than those of SPCA, SKPCA, and SKICA. For faults 4, 5, 10, 16, 19, and 21, SPCA, SKPCA, and SKICA have difficulties detecting these faults (the FDR of T2 is less than 30%). It can be concluded from the above experimental results that SKNN is superior to SPCA, SKPCA, and SKICA.
Figure 10

Measures of non-Gaussianity of statistics for TEP.

Table 5

FDR (%) of SPCA, SKPCA, SKICA, and SKNN for the TEP

 SPCA
SKPCA
SKICA
 
faultT2SPET2SPEI2SPESKNN
199.2510099.2599.2599.2544.03100
296.2798.5196.2797.7697.0196.2798.51
41.491000.7591.790.758.21100
521.6450.7523.8830.6024.6322.3955.97
6100100100100100100100
770.1510041.0410087.3135.82100
896.2797.7696.2797.7697.7693.2897.76
1016.4264.9326.8750.7528.3618.6672.39
1167.1697.0168.6694.0379.1067.1698.51
1298.5199.2599.2599.2599.2599.2599.25
1390.3095.5291.7994.0392.5485.0795.52
14100100100100100100100
1622.3973.1325.3754.4830.6014.1876.87
1788.8197.0173.1394.0374.6373.1397.01
1888.8190.3018.6689.5518.6618.6691.04
198.2198.515.2298.519.701.4998.51
2050.0077.6149.2570.1555.9749.2580.60
21044.03030.600046.27
average61.9888.0256.4382.9260.8651.4989.35
The monitoring results of SPCA, SKPCA, SKICA, and SKNN for faults 10, 11, and 20 are shown in Figures –13, respectively. The reason why the performance SKNN is better than those of SPCA and SKPCA is that it can simultaneously overcome the nonlinearity and non-Gaussianity of statistics. For SPCA, the nonlinearity and non-Gaussianity of statistics destroy the premise of the threshold calculation in PCA. SKPCA only considers the nonlinear problem of statistics, and the non-Gaussianity of statistics is ignored. Although the SKICA method can deal with non-Gaussian and nonlinear problems of process variables (statistics), in the process of reducing the dimensionality of the kernel matrix and extracting ICs, some key information of the process may be lost, resulting in poor detection performance.
Figure 11

Fault detection results of SPCA, SKPCA, SKICA, and SKNN for fault 10 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.

Figure 13

Fault detection results of SPCA, SKPCA, SKICA, and SKNN for fault 20 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.

Fault Diagnosis: In the stage of fault diagnosis, fault 1, fault 7, and fault 10 are used for comparison. The faulty variables of fault 1 are the 1st variable, fourth variable, 18th variable, 19th variable, 38th variable, 44th variable, 45th variable, and 50th variable. In Figure , the SPCA-T2 and SPCA-SPE identify some nonfaulty statistics of fault 1 as fault variables due to the smearing effect. SKNN does not have a smearing effect. Therefore, the diagnosis result of SKNN is similar to the actual situation. The cause of fault 7 is the pressure loss of the C header. To maintain the reactor level, the controller adjusts the flow of stream 4 by changing the total feed flow rate (i.e., 45th variable). Therefore, it is the 45th variable that actually generates the exception. The cause of fault 10 is the random variation in the C feed temperature (stream 4). Specifically, it is the increase in variability of the 18th variable. Therefore, there is only one fault variable for fault 7 and fault 10, which is the 45th variable and the 18th variable, respectively. The diagnosis results of fault 7 and fault 10 are shown in Figures and 16, respectively. Both SPCA-SPE and SKNN correctly isolate the fault statistics. However, it is worth noting that the maximum contribution of the 45th statistic (i.e., the mean of the 45th variable) in Figure b,c is about 0.6 and 0.8, respectively; while the maximum contribution of the 70th statistic (i.e., the variance of the 18th variable) is about 0.7 and 0.9, respectively. In addition, the SPCA-T2 identifies many nonfaulty statistics as faulty variables due to the smearing effect. In summary, the SKNN method is also superior to the SPCA method in fault diagnosis.
Figure 14

Fault diagnosis results of SPCA and SKNN for fault 1 of TEP. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN.

Figure 15

Fault diagnosis results of SPCA and SKNN for fault 7 of TEP. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN.

Figure 16

Fault diagnosis results of SPCA and SKNN for fault 10 of TEP. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN.

Measures of non-Gaussianity of statistics for TEP. Fault detection results of SPCA, SKPCA, SKICA, and SKNN for fault 10 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN. Fault detection results of SPCA, SKPCA, SKICA, and SKNN for fault 11 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN. Fault detection results of SPCA, SKPCA, SKICA, and SKNN for fault 20 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN. Fault diagnosis results of SPCA and SKNN for fault 1 of TEP. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN. Fault diagnosis results of SPCA and SKNN for fault 7 of TEP. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN. Fault diagnosis results of SPCA and SKNN for fault 10 of TEP. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN. For fault 7, it can be seen from Figure that the faulty variable isolated by SKNN is the mean of the 45th variable, not the variable itself. For fault 10, in Figure , the fault statistic isolated by SKNN is the variance of the 18th variable. As can be seen from Figure a,b, the mean of the 45th variable and the variance of the 18th variable increase significantly after the fault is introduced from the 26th window. Therefore, for fault 7 and fault 10, it is effective to perform early fault detection by observing the mean of the 45th variable and the variance of the 18th variable.
Figure 17

(a) Mean of the 45th variable in fault 7; (b) variance of the 18th variable in fault 10.

(a) Mean of the 45th variable in fault 7; (b) variance of the 18th variable in fault 10.

Conclusions

The SPA method can extract different statistics to capture the complex characteristics of the process. Therefore, it is a feasible framework for complex industrial process monitoring. The non-Gaussianity and non-linearity of extracted statistics under the SPA framework cause the failure of the traditional MSPM methods. In order to overcome these problems, a new process monitoring method based on SPA and the kNN algorithm is proposed in this paper. In the premise of the kNN method, the data samples are not required to satisfy a certain distribution. Therefore, the proposed method can avoid the problems caused by the non-Gaussianity and nonlinearity of extracted statistics while inheriting the advantages of SPA in extracting HOS information. In addition, in terms of fault diagnosis, SKNN also has better performance because it is not affected by fault smearing. The experiments on numerical examples and TE processes verify the effectiveness of the proposed method. The strength of the proposed method is that it can overcome the non-Gaussian and nonlinear problems in the SPA framework as well as alleviating the fault smearing phenomenon in the fault isolation of statistics. However, the proposed method still has limitations when dealing with multimode processes because the statistics extracted by SPA still retain the multimodality characteristics in the original sample space, which will make the calculation of detection thresholds seriously deviate from the normal level.
  2 in total

1.  Improved Surrogate Data for Nonlinearity Tests.

Authors: 
Journal:  Phys Rev Lett       Date:  1996-07-22       Impact factor: 9.161

2.  Incipient fault detection for nonlinear processes based on dynamic multi-block probability related kernel principal component analysis.

Authors:  Peipei Cai; Xiaogang Deng
Journal:  ISA Trans       Date:  2020-05-20       Impact factor: 5.468

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.