Zhe Zhou1, Jian Wang1, Chunjie Yang2, Chenglin Wen3, Zuxin Li4. 1. School of Engineering, Huzhou University, Huzhou 313000, China. 2. State Key Laboratory of Industrial Control Technology, Zhejiang University, Hangzhou 310027, China. 3. School of Automation, Guangdong University of Petrochemical Technology, Maoming 525000, China. 4. School of Science and Engineering, Huzhou College, Huzhou 313000, China.
Abstract
Only low-order information of process data (i.e., mean, variance, and covariance) was considered in the principal component analysis (PCA)-based process monitoring method. Consequently, it cannot deal with continuous processes with strong dynamics, nonlinearity, and non-Gaussianity. To this aim, the statistics pattern analysis (SPA)-based process monitoring method achieves better monitoring results by extracting higher-order statistics (HOS) of the process variables. However, the extracted statistics do not strictly follow a Gaussian distribution, making the estimated control limits in Hotelling-T 2 and squared prediction error (SPE) charts inaccurate, resulting in unsatisfactory monitoring performance. In order to solve this problem, this paper presents a novel process monitoring method using SPA and the k-nearest neighbor algorithm. In the proposed method, first, the statistics of process variables are calculated through SPA. Then, the k-nearest neighbor (kNN) method is used to monitor the extracted statistics. The kNN method only uses the paired distance of samples to perform fault detection. It has no strict requirements for data distribution. Hence, the proposed method can overcome the problems caused by the non-Gaussianity and nonlinearity of statistics. In addition, the potential of the proposed method in early fault detection or safety alarm and fault isolation is explored. The proposed method can isolate which variable or its statistic is faulty. Finally, the numerical examples and Tennessee Eastman benchmark process illustrate the effectiveness of the proposed method.
Only low-order information of process data (i.e., mean, variance, and covariance) was considered in the principal component analysis (PCA)-based process monitoring method. Consequently, it cannot deal with continuous processes with strong dynamics, nonlinearity, and non-Gaussianity. To this aim, the statistics pattern analysis (SPA)-based process monitoring method achieves better monitoring results by extracting higher-order statistics (HOS) of the process variables. However, the extracted statistics do not strictly follow a Gaussian distribution, making the estimated control limits in Hotelling-T 2 and squared prediction error (SPE) charts inaccurate, resulting in unsatisfactory monitoring performance. In order to solve this problem, this paper presents a novel process monitoring method using SPA and the k-nearest neighbor algorithm. In the proposed method, first, the statistics of process variables are calculated through SPA. Then, the k-nearest neighbor (kNN) method is used to monitor the extracted statistics. The kNN method only uses the paired distance of samples to perform fault detection. It has no strict requirements for data distribution. Hence, the proposed method can overcome the problems caused by the non-Gaussianity and nonlinearity of statistics. In addition, the potential of the proposed method in early fault detection or safety alarm and fault isolation is explored. The proposed method can isolate which variable or its statistic is faulty. Finally, the numerical examples and Tennessee Eastman benchmark process illustrate the effectiveness of the proposed method.
With
the rapid development of industrial processes and the increase
in process complexity, a large amount of process operation data has
been generated. Fully mining and using the valuable information contained
in the process data will be beneficial to early fault detection, thereby
minimizing downtime and improving the safety of process operation.[1] In this context, data-driven multivariate statistical
process monitoring (MSPM) methods have been developed by leaps and
bounds, where principal component analysis (PCA) methods are the most
widely used. However, only low-order information of process data (i.e.,
mean, variance, and covariance) was considered, and high-order statistics
(HOS) were ignored in the PCA-based process monitoring method. In
addition, the threshold of Hotelling’s T2 and SPE are calculated based on the premise that process
variables satisfy a Gaussian distribution.[2] Due to nonlinearity, non-Gaussianity, dynamic, and multimodality
in industrial processes, it is difficult to satisfy this assumption
in practice. Therefore, the traditional PCA-based process monitoring
method has poor monitoring performance when facing the above problems.[3−7]Recently, there have been many studies on nonlinear process
monitoring
problems. To capture the nonlinear correlation structure, Yu et al.(8) proposed a robust, nonlinear
and sparse PCA method. Aiming at the excessive modeling redundancy
problem caused by infinite-order mapping, a monitoring method based
on constructing polynomial mapping is proposed in ref (9), which significantly improves
the fault detection performance of nonlinear processes. Jiang and
Yan[10] proposed a parallel PCA-KPCA monitoring
method for a process with linearly correlated and nonlinearly related
variables. In order to detect the incipient faults of nonlinear industrial
processes effectively, an enhanced KPCA method is proposed in ref (11). Mansouri et al.(12) introduced a generalized likelihood
ratio test into the KPCA method for fault detection of nonlinear processes.
However, these methods above do not consider the high-order statistics
of the process data, and the calculation is complicated.In
addition, there is a lot of work to solve the problem of fault
diagnosis of complex industrial processes using hybrid methods including
Bayesian and data-driven approaches. Yu et al.(13) proposed a two-stage fault diagnosis method
that combines independent component analysis (ICA) with the Bayesian
network (BN) to solve the problem in that conventional MSPM methods
cannot isolate fault from unmonitored process variables. In order
to reduce the cost of monitoring and alarm flooding, Wang et al.(14) proposed a fault diagnosis
technique that combines semi-parametric PCA and BN, which also achieves
location of the root cause of faults in process variables. A PCA-BN
hybrid method with multiple likelihood evidence for fault diagnosis
is proposed in ref (15). This method enables BN to update more information about faults
and improves process monitoring performance. Reference (16) uses KPCA and BN for fault
diagnosis, which not only diagnoses the root cause of the fault but
also shows the propagation path of the fault.He and Wang[17] proposed a statistics
pattern analysis (SPA) monitoring framework to overcome these problems.
This method extracts HOS from process data, such as skewness and kurtosis.
Compared with the traditional MSPM methods that only use low-order
statistics information, it has advantages in characterizing the non-Gaussianity
and nonlinearity of the process data. They used the SPA to monitor
the batch process because it calculates the statistics of batch process
data, and data preprocessing is avoided. In ref (2), the sliding window is
used to design the SPA to monitor the continuous process and obtain
better monitoring performance on the Tennessee Eastman (TE) platform.
In addition, the SPA is identified as a novel generation of the SPM
method and can handle the 4 V challenges of big data.[18−21]Although the SPA framework has many advantages in monitoring
processes
of nonlinearity or multimodality,[22−25] it is worth noting that the statistics
extracted by the SPA is not a Gaussian distribution, in addition,
there is a nonlinear relationship between the extracted statistics.
Therefore, it is not suitable to use PCA combined with T2 and SPE charts to monitor the extracted statistics.
Ma et al.(26) considered
the nonlinearity of statistics: they used the advantages of kernel
PCA in processing nonlinear data and proposed a process monitoring
method based on statistics KPCA (SKPCA). However, SKPCA only considers
the nonlinear problem of statistics, and the non-Gaussianity of statistics
is ignored.In this paper, a new process monitoring method,
which combines
the superiority of SPA in extracting HOS information with kNN in processing
the non-Gaussian and nonlinearity of data samples, is proposed to
deal with the defects of those methods above. The fault detection
method using the kNN rule (FD-kNN) only uses the distance between
neighbors to perform fault detection; there is no restriction on the
data distribution.[27,28] The qualitative and quantitative
analysis for the non-Gaussianity and nonlinearity of statistics are
also presented. Meanwhile, the effect of window width on the non-Gaussianity
of statistics is investigated. In addition, we explore the potential
of the proposed method in fault diagnosis. Specifically, the variable
contribution by kNN (VCkNN) is used to determine the contribution
of different statistics to the detection index. kNN is a distribution-free
method. The proposed method uses the kNN method to monitor process
statistics to achieve fault detection and fault isolation. In terms
of fault detection, compared with the conventional statistical techniques,
it can overcome the non-Gaussian and nonlinear problems of process
variables (statistics); in terms of fault isolation, compared to traditional
contribution analysis methods, the proposed kNN-based variable contribution
method is not affected by fault smearing. The experiments on numerical
examples and TE processes illustrate the effectiveness of the proposed
process monitoring method.The contributions of the proposed
method are as follows:The qualitative
and quantitative analysis for the extracted
statistics of process data is conducted and verifies that they are
non-Gaussian and nonlinear.The statistics
extracted from the process data are monitored
by using the k-nearest neighbor method, which avoids
the problems caused by non-linear, non-Gaussian characteristics of
statistics.The employed variable contribution
by kNN can indicate
the statistic of the faulty variable, which may be useful for root
cause analysis and has potential to perform incipient fault detection.
Related Work
The core foundation
of the proposed method is related to the SPA
and FD-kNN. This section will briefly review these two methods.
SPA
We use Y to denote the
data in a window of the original samples:where l is
the window width, d is the last sampling time, and m represents the number of variables.The sampling
data of each window can be used to calculate a statistics pattern
(SP), which is composed of first-order statistics (FOS), second-order
statistics (SOS), and HOS. As shown in eq where μ denotes
the
FOS, i.e., variable means (μ);
Σ denotes the SOS, which include variance (v), correlation (r), autocorrelation
(r), and cross-correlation (r); Ξ denotes
the HOS, which is composed of skewness (γ) and kurtosis (κ). These
statistics can be calculated as followswhere c represents
the maximum lag time.An SP vector can be obtained using the
above equations for each
window of process data, and all the SP vectors of training data can
be stacked into a matrix, denoted as SPs. The number of columns and
rows in SPs represents the number of statistics and windows, respectively,
thereby realizing statistics extraction from training data. In ref (2), the SP matrix is then
decomposed by PCA, and T2 and SPE charts
are established. For online monitoring, statistics are also extracted
based on the data obtained at current and the previous (l
– 1) sampling times. After that, T2 and SPE of extracted statistics are calculated to perform
fault detection.
FD-kNN
The kNN rule was initially
widely used in pattern
classification. In December 2006, it was selected as one of the top
ten classic algorithms in data mining. FD-kNN was first proposed by
He and Wang.[29] The main idea is to measure
the difference between samples by distance; that is, normal samples
and training samples are similar, but fault samples and training samples
are significantly different.Training phase (determine the detection control limit):Use the Euclidean
distance to get the
kNNs of each training sample.Calculate the distance statistic D2.where D2 represents the average squared distance
between
the pth sample and its k neighbors, d2 denotes the squared Euclidean distance between the pth sample and its qth nearest neighbor.Establish the control limit Dα2 for fault detection. There
are many ways to estimate Dα2, such as the estimation using a noncentral chi-square distribution,[29] kernel density estimation (KDE). The (1 –
α)-empirical quartile[30] of D2 is used as the threshold in our
proposed method.where D(2, i = 1, ..., n is the result of D2 obtained by eq in
ascending order. (⌊n(1 – α)⌋)
means discard the decimal part of n(1 – α)
and keep the integer part.Detection phase:For a sample x to be tested,
find its kNNs from the training set.Calculate D2 between x and its k neighbors
using eq .Compare D2 with the threshold Dα2. If D2 > Dα2, the process is considered abnormal. Otherwise, it is
normal.
Analysis
for the Non-Gaussianity and Nonlinearity of Statistics
Although
the PCA fault detection method based on SPA (SPCA) improves
the monitoring performance by using HOS information to describe the
complex characteristics of process data, the statistics extracted
by SPA is not satisfied by a Gaussian distribution. Also, the relationship
between different statistics is nonlinear. Faced with the above problems,
the SPCA is theoretically flawed.In this section, the qualitative
and quantitative analysis for
the non-Gaussianity and nonlinearity of statistics are presented.
In addition, the effect of window width on the non-Gaussianity of
statistics is investigated.
Non-Gaussianity of Statistics
Qualitative analysis:
The central limit
theorem (CLT) says that under rather general circumstances, the sum
of independent normalized random variables is a normal distribution
at the limit.[2] It needs to be emphasized
that CLT assumes that the samples are drawn independently and the
number of samples (in a window) is sufficiently large. In practice,
however, the wider the window’s width, the greater the delay
in giving the monitoring results. Moreover, the quantity of samples
is limited by the specific process. In addition, if the overlap rate
of two adjacent windows is high, the basic assumption of independent
samples will be violated. Therefore, the extracted statistics are
difficult to meet the assumption of CLT, which makes them non-Gaussian.Quantitative analysis:
The non-Gaussianity
of each variable can be evaluated by negative entropy (NE).[31] A kind of robust estimation of NE[32] is employed to calculate the non-Gaussianity
of statistics. NE is non-negative, and it is 0 only when a variable
is the standard normal distribution. The larger the NE of a random
variable, the more it deviates from the standard normal distribution.
For variable x, its NE can be expressed as follows:where G(v) is the entropy of the standard normal distributed variable v, G(x*) is the entropy
of x*, while x* is the normalized
result of x. The G(v) can be calculated as a constant:As for G(x*), it can be
calculated as follows:[33]where x*(1), x*(2), ..., x*( denote M measurements
of x* arranged in the ascending order. Substituting eqs and 15 into eq yields
the NE of x. It is hard to achieve zero for NE even if x has a Gaussian
distribution because NE is affected by
the measurements of x and the number of measurements.
According to ref (34), the value 10–3 is taken as the threshold.
The Nonlinearity of Statistics
Qualitative analysis:
If process variables
have a nonlinear relationship, the statistics of these variables cannot
eliminate the nonlinear relationship of original variables.[26] In addition, it can be seen from the calculation
formula of the statistics that there is also a nonlinear relationship
between the statistics of different orders.Quantitative analysis: We use the surrogate
data (SD) method[35−37] to detect the nonlinearity of statistics. This method
first designates a given linear process as the null hypothesis, then
generates SD sets consistent with the null hypothesis, and finally
calculates a discriminating statistic for the original data and each
SD set. If the value calculated for the original data is significantly
different from the set of values calculated from the SD, then the
null hypothesis is rejected, and nonlinearity is determined.The null hypothesis H0: the data come from a linear Gaussian process.Generate surrogate data
sets for original
data using Fourier transform (FT).Fourier transform: For variable x with N samples, its FT can be expressed
as follows:Randomize the phase:where φ is a random variable and
obeys a uniform distribution
in the range of [0, 2π].Inverse Fourier transform: After the
inverse Fourier transform, the substitute data x(n) of the original data can be obtained.Calculate the statistic T of a significance test.where R denotes the redundancy of the
original data and R̅ and σ denote the mean and
standard deviation of redundancy
of the surrogate data set, respectively. Redundancy can be calculated
using eq .[38]The covariance matrix
of variables x and y is denoted
as C. c and λ are diagonal elements and eigenvalues of
the covariance matrix C, respectively.Determine whether the null hypothesis
is true. Under the condition that the significance level is 0.05,
the threshold of significance testing is 1.96. If T ⩾ 1.96, the null hypothesis H0 is rejected and original data is nonlinear; otherwise, accept
the null hypothesis and original data is linear.
Illustrative Example
A simple static nonlinear example
is used to illustrate the non-Gaussianity and nonlinearity of statistics.
The process model is as follows:where t is
a random variable that obeys a uniform distribution in the range of
[−0.5, 0.5] and e is the Gaussian noise with
0 mean and variance of 0.01. Ten thousand normal samples are generated
by eq for extracting
statistics by SPA. The window width is 10 (i.e., l = 10), and the sliding step is 1. Each SP consists of 8 statistics,
which include the FOS, SOS, and HOS of x and y.Figure shows that the multivariate distribution of different statistics
is non-Gaussian. The non-Gaussianity of statistics is quantified by
estimating the NE. It can be concluded from Table that the NE of the statistics is almost
always smaller than that of the original variable, but it is still
greater than 10–3. Therefore, the statistics extracted
by SPA are indeed non-Gaussian.
Figure 1
Density histogram intensity map for statistics:
(a) mean of x and y; (b) variance
of x and y; (c) skewness of x and y; (d) kurtosis of x and y; (e) original variables x and y; (f) Gaussian random variables p and q.
Table 1
Negative Entropy (NE) of Variables
variable
NE
variable
NE
x
0.1281
y
0.4838
mean of x
0.0404
mean of y
0.0456
variance of x
0.0436
variance of y
0.0483
skewness of x
0.0458
skewness of y
0.0454
kurtosis of x
0.2357
kurtosis of y
0.3140
Density histogram intensity map for statistics:
(a) mean of x and y; (b) variance
of x and y; (c) skewness of x and y; (d) kurtosis of x and y; (e) original variables x and y; (f) Gaussian random variables p and q.The NE of the statistics
can be affected by the window width and
window shifting step. For a given number of statistics, the larger
the window width, the smaller the NE of the statistics, that is, the
weaker the non-Gaussianity, especially for HOS because the NE of the
statistics is related to the number of the statistics determined by
the window width and the window shifting step; therefore, as shown
in Figure , there
is an approximately monotonous decreasing trend between the NE (or
non-Gaussianity) of the statistics and the window width. Figure a,b shows the nonlinearity
detection results of the statistics (mean of x and
mean of y; variance of x and kurtosis
of x) using the surrogate data method. It can be
seen that T is always greater than 1.96.
Therefore, the null hypothesis should be rejected with a probability
of 95%; the statistics are nonlinear.
Figure 2
Effect of window width on the NE of the
statistics.
Figure 3
Nonlinearity detection results of the statistics.
(a) Mean of and ; (b)
variance of and kurtosis of ; (c) original variables and ; (d) Gaussian random
variables and .
Effect of window width on the NE of the
statistics.Nonlinearity detection results of the statistics.
(a) Mean of and ; (b)
variance of and kurtosis of ; (c) original variables and ; (d) Gaussian random
variables and .
Process Monitoring Method
Using the SPA and kNN Algorithm (SKNN)
Fault Detection Based on
SKNN
The SKNN integrates the
advantages of SPA in extracting HOS information and kNN in dealing
with the non-Gaussian and nonlinearity of data samples.Training step:Extract the statistics matrix Y (SPs) from the training sets generated under the normal
operation condition (NOC).Find kNNs for each sample (SP) in Y using eq .Calculate the distance statistic using eq .Determine the threshold Dα2 for fault detection using eq .Online monitoring:For a sample x, a statistic
sample s can be extracted by SPA.Find kNNs of s from Y.Calculate D2 using eq .Compare D2 with the threshold Dα2. If D2 > Dα2, it is considered as a faulty sample. Otherwise, it
is a normal
sample.
Fault Diagnosis
Based on SKNN
When a fault is detected,
finding the root cause of the failure is the next more important thing.
VCkNN[27] is used to determine the contribution
of different statistics to the detection index. The contribution of
each statistic is not affected by other statistics; that is, there
is no correlation. Therefore, the diagnosis result is not affected
by the fault smearing.Determine
the contributions of statistics:Equation is
the form of decomposing D2 into the sum of the contributions
of m statistics.where cknn is the contribution
from the ith statistic of s to D2.Determine the faulty statistics:where F contains
all the faulty statistics diagnosed by SKNN and T represents the threshold for the ith statistic
contribution. The threshold can be determined empirically; for example,
the maximum contribution of the statistic in the training samples
is used as the isolation threshold of the corresponding statistic.[27]The statistics can
be flexibly selected
for different purposes or process characteristics. For example, considering
that the dynamics of the batch process are relatively weak, the only
covariance is selected as the second-order statistics.[17]For the design of the window width
and sliding step, the data set should be fully used. Otherwise, samples
at the end of the data set will not be used. For example, if the size
of the data set is 100 when the window width is 10 and the sliding
step is 5, 19 SP are obtained exactly, and the data set is fully used;
when the window width is 16 and the sliding step is 5, the last four
samples of the data set will not be used resulting in the loss of
key information of the process, making the model established in the
training phase insufficient to describe the process accurately, thereby
degrading the monitoring performance.The number of nearest neighbors is
selected, and k is determined using cross-validation.
Case Studies
In this section, we
use numerical examples and TEP to explore the
effectiveness of the proposed method SKNN in fault detection and diagnosis.
The kernel-independent component analysis (KICA)[39] method is also used to solve the non-Gaussian and nonlinear
problems of process data. Therefore, in addition to the SPCA and SKPCA
methods, the monitoring performance of the fault detection method
using SPA and KICA (SKICA) has also been verified.
Numerical Simulations
The NOC data are produced according
to the following process model adopted from ref (4)uo(k) denotes the correlated input, (k) denotes the noise with
0 mean
and variance 0.1.where w(k) is a noise with 0 mean and a standard
deviation of 1, v(k) represents a noise
with 0 mean and a
variance of 0.1.There are four variables (i.e., u1, u2, y1, and y2) used for monitoring.
Since the process is
dynamic, autocorrelation and cross-correlation are added to the second-order
statistics. Therefore, a total of 54 statistics are extracted by SPA.
For all four methods, the numbers of generated training samples and
validation samples are 10,000 and 5000, respectively. There are 10,000
testing samples, of which the first 500 samples are normal, and the
rest are faulty.The fault setting is that the mean value of
the first element w1 in w changes from 0 to 4.The specific parameter settings of SPCA,
SKPCA, SKICA, and SKNN
are shown in Table . For SPCA, the number of principal components (PCs) is determined
according to cumulative percent variance (CPV)[40] more than 90%. For SKPCA, the size of the kernel parameter
is selected as 5m according to ref (5), and m is the number of variables. For SKICA, the number of independent
components (ICs) is the same as that of the PCs. For SKNN, the number
of nearest neighbors is 3. For SPCA, SKPCA, SKICA, and SKNN, the control
limits of different indices are calculated at a confidence level of
99%.
Table 2
Parameter Settings of SPCA, SKPCA,
SKICA, and SKNN
methods
statistics
PCs
ICs
window width
window sliding
step
kernel parameter
k
SPCA
54
13
25
15
SKPCA
54
3
25
15
270
SKICA
54
4
4
25
15
270
SKNN
54
25
15
3
Fault Detection:Figure shows the NE of statistics, and we can observe
that the NE of different
statistics is greater than 10–3. Therefore, the
statistics extracted from this dynamic process are non-Gaussian. The
FDR and FAR of the four methods are shown in Table (note that the FAR is obtained based on
the validation samples). SKNN has the best detection performance among
the three methods because it can overcome the problems caused by the
nonlinearity and non-Gaussianity of statistics. Figure is the comparison result of the detection
performance of the three methods. For SPCA, due to the non-Gaussian
and nonlinearity of the statistics, the detection performance of SPCA
is not good. The performance of SKPCA is better than SPCA because
the KPCA has the advantage in dealing with nonlinearity. Although
the SKICA method can also deal with non-Gaussian and nonlinear problems
of process data, the fault detection performance of SKICA on experiments
is not as good as that of SKNN. This may be due to the following reasons:
to reduce the computational load of the SKICA method, ICA is performed
after the kernel matrix is dimensionally reduced by PCA, which may
cause the information used to extract essential ICs to be ignored.
Figure 4
Measures of non-Gaussianity
of statistics for the numerical example.
Table 3
FDR (%) and FAR (%) of SPCA, SKPCA,
SKICA, and SKNN for the Numerical Example
SPCA
SKPCA
SKICA
SKNN
index
T2
SPE
T2
SPE
I2
SPE
kNN
FDR
20.19
10.57
13.56
97.95
57.10
97.79
99.37
FAR
0.9
0.9
1.5
1.8
0.0
0.3
0.9
Figure 5
Fault
detection results of SPCA, SKPCA, SKICA, and SKNN for the
numerical example. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.
Fault Diagnosis:The change of w has
different effects on
the four monitored variables. Figure shows the difference between each variable of the
training samples and the corresponding variable in the fault samples
(only partial results are shown due to limited space). The fourth
variable (y2) has the maximum difference.
Because the kernel method is difficult or even impossible to find
the inverse mapping function from the feature space to the original
space, the fault diagnosis of SKPCA and SKICA is not considered in
this paper. Figure is the comparison result of the diagnosis performance of SPCA and
SKNN. SPCA cannot identify the variable y2 that contributes the most to the fault due to the smearing effect.
For SKNN, it does not have a smearing effect because VCkNN is defined
in the original variable space (here means statistics space). In Figure c, the variable y2 that contributes the most to the fault is
successfully isolated. It is worth noting that the diagnosis result
of SKNN indicates that the mean of y2 has
the maximum contribution as can be seen from Figure , the mean of y2 increases significantly after the fault is introduced from the 33rd
window. Therefore, it is effective to perform early fault detection
or safety alarm by observing the mean of y2.
Figure 6
Difference between each variable of the training samples
and the
corresponding variable in the fault samples.
Figure 7
Fault
diagnosis results of SPCA-T2,
SPCA-SPE and SKNN for the numerical example. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN.
Figure 8
Man of .
Measures of non-Gaussianity
of statistics for the numerical example.Fault
detection results of SPCA, SKPCA, SKICA, and SKNN for the
numerical example. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.Difference between each variable of the training samples
and the
corresponding variable in the fault samples.Fault
diagnosis results of SPCA-T2,
SPCA-SPE and SKNN for the numerical example. (a) SPCA-T2; (b) SPCA-SPE; (c) SKNN.Man of .
TE Benchmark Process
When comparing the performance
or effectiveness of process monitoring methods, the TEP is a benchmark
choice. In ref (41), Downs and Vogel proposed the simulation platform. There are five
major operating units in the TE process, namely, a reactor, a product
condenser, a vapor–liquid separator, a recycle compressor,
and a product stripper. The process has four kinds of reactants (A,
C, D, E), two products (G, H), and contains a catalyst (B) and byproducts
(F). The flowchart of the process is given in Figure . There are 11 manipulated variables (No.42–No.52),
22 process measurements (No.1–No.22), and 19 composition variables
(No.23–No.41). For detailed information on the 52 monitoring
variables and 21 fault patterns, see ref (42).
Figure 9
Flowchart of the Tennessee Eastman process.
Flowchart of the Tennessee Eastman process.The parameter settings of SPCA, SKPCA, and SKNN are shown
in Table . For all
three methods,
the number of training samples and the number of validation samples
are 960 and 480, respectively. In addition, there are 960 testing
samples where the fault is introduced from the 161st sample. The thresholds
of different methods are all calculated at a confidence level of 99%.
For second-order statistics, only the significant statistics will
be selected. The specific selection method is the same as in ref (2). In order to facilitate
the establishment of the model, for a large number of second-order
statistics, we only select the numerically significant ones. The specific
selection rules are as follows:
Table 4
Parameter Settings of SPCA, SKPCA,
SKICA, and SKNN for TEP
methods
statistics
PCs
ICs
window width
window sliding
step
kernel parameter
k
SPCA
223
74
12
6
SKPCA
223
35
12
6
1115
SKICA
223
38
38
12
6
1115
SKNN
223
12
6
3
r is selected only if |r| > 0.5 for
more than 70% of the training SPs;ris selected only
if |r| >
0.5 for more than 90% of the training SPs;ris selected only if |r| > 0.5 for more than 90% of the training SPs.After statistics
selection, a total of 223 statistics of the TE
process are used for process monitoring.
Results and Discussion
This section analyzes and discusses the experimental results of
the TE process.Fault Detection:From Figure , the statistics extracted from the TE process
are non-Gaussian
because the NE of the statistics is greater than the threshold of
10–3. The FDR of the four methods are listed in Table . Note that faults
3, 9, and 15 are not considered here because they are difficult to
be detected.[1,43,44] For faults 5, 10, 11, 16, 18, 20, and 21, the FDR of the SKNN method
is higher than those of SPCA, SKPCA, and SKICA. For faults 4, 5, 10,
16, 19, and 21, SPCA, SKPCA, and SKICA have difficulties detecting
these faults (the FDR of T2 is less than
30%). It can be concluded from the above experimental results that
SKNN is superior to SPCA, SKPCA, and SKICA.
Figure 10
Measures of non-Gaussianity
of statistics for TEP.
Table 5
FDR (%) of SPCA, SKPCA, SKICA, and
SKNN for the TEP
SPCA
SKPCA
SKICA
fault
T2
SPE
T2
SPE
I2
SPE
SKNN
1
99.25
100
99.25
99.25
99.25
44.03
100
2
96.27
98.51
96.27
97.76
97.01
96.27
98.51
4
1.49
100
0.75
91.79
0.75
8.21
100
5
21.64
50.75
23.88
30.60
24.63
22.39
55.97
6
100
100
100
100
100
100
100
7
70.15
100
41.04
100
87.31
35.82
100
8
96.27
97.76
96.27
97.76
97.76
93.28
97.76
10
16.42
64.93
26.87
50.75
28.36
18.66
72.39
11
67.16
97.01
68.66
94.03
79.10
67.16
98.51
12
98.51
99.25
99.25
99.25
99.25
99.25
99.25
13
90.30
95.52
91.79
94.03
92.54
85.07
95.52
14
100
100
100
100
100
100
100
16
22.39
73.13
25.37
54.48
30.60
14.18
76.87
17
88.81
97.01
73.13
94.03
74.63
73.13
97.01
18
88.81
90.30
18.66
89.55
18.66
18.66
91.04
19
8.21
98.51
5.22
98.51
9.70
1.49
98.51
20
50.00
77.61
49.25
70.15
55.97
49.25
80.60
21
0
44.03
0
30.60
0
0
46.27
average
61.98
88.02
56.43
82.92
60.86
51.49
89.35
The monitoring results
of SPCA, SKPCA, SKICA, and SKNN for faults 10, 11, and 20 are shown
in Figures –13, respectively. The reason why
the performance SKNN is better than those of SPCA and SKPCA is that
it can simultaneously overcome the nonlinearity and non-Gaussianity
of statistics. For SPCA, the nonlinearity and non-Gaussianity of statistics
destroy the premise of the threshold calculation in PCA. SKPCA only
considers the nonlinear problem of statistics, and the non-Gaussianity
of statistics is ignored. Although the SKICA method can deal with
non-Gaussian and nonlinear problems of process variables (statistics),
in the process of reducing the dimensionality of the kernel matrix
and extracting ICs, some key information of the process may be lost,
resulting in poor detection performance.
Figure 11
Fault detection results
of SPCA, SKPCA, SKICA, and SKNN for fault
10 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.
Figure 13
Fault detection results of SPCA, SKPCA, SKICA, and SKNN for fault
20 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.
Fault Diagnosis:In the stage of fault diagnosis,
fault 1, fault 7, and
fault 10 are used for comparison. The faulty variables of fault 1
are the 1st variable, fourth variable, 18th variable, 19th variable,
38th variable, 44th variable, 45th variable, and 50th variable. In Figure , the SPCA-T2 and SPCA-SPE identify some
nonfaulty statistics of fault 1 as fault variables due to the smearing
effect. SKNN does not have a smearing effect. Therefore, the diagnosis
result of SKNN is similar to the actual situation. The cause of fault
7 is the pressure loss of the C header. To maintain the reactor level,
the controller adjusts the flow of stream 4 by changing the total
feed flow rate (i.e., 45th variable). Therefore, it is the 45th variable
that actually generates the exception. The cause of fault 10 is the
random variation in the C feed temperature (stream 4). Specifically,
it is the increase in variability of the 18th variable. Therefore,
there is only one fault variable for fault 7 and fault 10, which is
the 45th variable and the 18th variable, respectively. The diagnosis
results of fault 7 and fault 10 are shown in Figures and 16, respectively.
Both SPCA-SPE and SKNN correctly isolate the fault statistics. However,
it is worth noting that the maximum contribution of the 45th statistic
(i.e., the mean of the 45th variable) in Figure b,c is about 0.6 and 0.8, respectively;
while the maximum contribution of the 70th statistic (i.e., the variance
of the 18th variable) is about 0.7 and 0.9, respectively. In addition,
the SPCA-T2 identifies many nonfaulty
statistics as faulty variables due to the smearing effect. In summary,
the SKNN method is also superior to the SPCA method in fault diagnosis.
Figure 14
Fault diagnosis results of SPCA and SKNN for fault 1 of TEP. (a)
SPCA-T2; (b) SPCA-SPE; (c) SKNN.
Figure 15
Fault diagnosis results of SPCA and SKNN for fault 7 of
TEP. (a)
SPCA-T2; (b) SPCA-SPE; (c) SKNN.
Figure 16
Fault diagnosis results of SPCA and SKNN for fault 10
of TEP. (a)
SPCA-T2; (b) SPCA-SPE; (c) SKNN.
Measures of non-Gaussianity
of statistics for TEP.Fault detection results
of SPCA, SKPCA, SKICA, and SKNN for fault
10 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.Fault detection results of SPCA, SKPCA, SKICA, and SKNN for fault
11 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.Fault detection results of SPCA, SKPCA, SKICA, and SKNN for fault
20 of TEP. (a) SPCA; (b) SKPCA; (c) SKICA; (d) SKNN.Fault diagnosis results of SPCA and SKNN for fault 1 of TEP. (a)
SPCA-T2; (b) SPCA-SPE; (c) SKNN.Fault diagnosis results of SPCA and SKNN for fault 7 of
TEP. (a)
SPCA-T2; (b) SPCA-SPE; (c) SKNN.Fault diagnosis results of SPCA and SKNN for fault 10
of TEP. (a)
SPCA-T2; (b) SPCA-SPE; (c) SKNN.For
fault 7, it can be seen from Figure that the faulty variable isolated by SKNN
is the mean of the 45th variable, not the variable itself. For fault
10, in Figure ,
the fault statistic isolated by SKNN is the variance of the 18th variable.
As can be seen from Figure a,b, the mean of the 45th variable and the variance of the
18th variable increase significantly after the fault is introduced
from the 26th window. Therefore, for fault 7 and fault 10, it is effective
to perform early fault detection by observing the mean of the 45th
variable and the variance of the 18th variable.
Figure 17
(a) Mean of the 45th
variable in fault 7; (b) variance of the 18th
variable in fault 10.
(a) Mean of the 45th
variable in fault 7; (b) variance of the 18th
variable in fault 10.
Conclusions
The
SPA method can extract different statistics to capture the
complex characteristics of the process. Therefore, it is a feasible
framework for complex industrial process monitoring. The non-Gaussianity
and non-linearity of extracted statistics under the SPA framework
cause the failure of the traditional MSPM methods. In order to overcome
these problems, a new process monitoring method based on SPA and the
kNN algorithm is proposed in this paper. In the premise of the kNN
method, the data samples are not required to satisfy a certain distribution.
Therefore, the proposed method can avoid the problems caused by the
non-Gaussianity and nonlinearity of extracted statistics while inheriting
the advantages of SPA in extracting HOS information. In addition,
in terms of fault diagnosis, SKNN also has better performance because
it is not affected by fault smearing. The experiments on numerical
examples and TE processes verify the effectiveness of the proposed
method.The strength of the proposed method is that it can overcome
the
non-Gaussian and nonlinear problems in the SPA framework as well as
alleviating the fault smearing phenomenon in the fault isolation of
statistics. However, the proposed method still has limitations when
dealing with multimode processes because the statistics extracted
by SPA still retain the multimodality characteristics in the original
sample space, which will make the calculation of detection thresholds
seriously deviate from the normal level.