Literature DB >> 34876575

Minimalist module analysis for fault detection and localization.

Zhijiang Lou1, Youqing Wang2, Shan Lu3, Pei Sun1.   

Abstract

Traditional multivariate statistical-based process monitoring (MSPM) methods are effective data-driven approaches for monitoring large-scale industrial processes, but have a shortcoming in handling the redundant correlations between process variables. To address this shortcoming, this study proposes a new MSPM method called minimalist module analysis (MMA). MMA divides process data into several different minimalist modules and one more independent module. All variables in the minimalist module are strongly correlated, and no redundant variables exist; therefore, the extracted feature components in one minimalist module will not be disturbed by noise from the other modules. This study also proposes new monitoring indices and a fault localization strategy for MMA, and simulation tests demonstrate that MMA achieves superior performance in fault detection and localization.
© 2021. The Author(s).

Entities:  

Year:  2021        PMID: 34876575      PMCID: PMC8651725          DOI: 10.1038/s41598-021-02676-3

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Multivariate statistical-based process monitoring (MSPM) methods[1-4], e.g., principal component analysis (PCA)[5, 6], partial least squares (PLS)[7, 8], and canonical correlation analysis (CCA)[9, 10], are effective data-driven approaches for monitoring large-scale industrial processes. The main idea of MSPM is analyzing the correlation between process variables and extracting the feature components for the construction of statistical indices. MSPM has been a research hotspot for many years, and a large number of relevant studies are published each year. In recent years, studies have focused on improving the existing methods to deal with process characteristics such as nonlinear, non-Gaussian, and dynamic features. For example, Ge et al.[11] combines the multivariate linear Gaussian state-space model with MSPM for handling the dynamic feature during a process; Du et al.[12] proposed the Gaussian distribution transformation (GDT)-based monitoring method for handling the non-Gaussian feature; and Lou et al.[13] combined artificial neural networks with PCA, and proposed a new neural component analysis for handling nonlinear features. Meanwhile, Zhou et al.[14] proposed a nonlinear key performance indicator (KPI) strategy for the PLS algorithm. Because MSPM can compress the high-dimensional data into two or three statistical indices, it is a convenient tool for detecting the abnormal condition in the whole process object. To address the fault localization problem, the contribution plot method[15, 16] was proposed for MSPM, which calculates the contribution of each variable of the original data set and picks the variables with high contributions as fault sources. Most studies on MSPM use the contribution plot as a basic algorithm tool[17, 18], and a few studies have proposed improved versions of the MSPM method that cannot use the traditional contribution plot directly (examples include the kernel PCA[19] and robust PCA[20]). However, according to actual simulation test results, MSPM is insensitive to specific faults, and the contribution plot method may mistakenly diagnose normal variables as a fault source. The reason for this phenomenon is that the traditional MSPM methods are based on the correlations between all process variables, and some correlations can be deduced by others, which means that these correlations are redundant. As such, the feature components extracted by traditional MSPM methods contain information from many process variables, and hence, are also disturbed by noises from these variables; therefore, traditional MSPM methods are insensitive to specific faults. In addition, the redundant correlations may mislead the contribution plot method, which results in incorrect localization of faults. For handling these problems, multiblock MSPM methods, such as consensus PCA (CPCA)[21], multiblock PLS (MBPLS)[18], and hierarchical PLS (HPLS)[22], are proposed for reducing the number of variables and improving the interpretability of multivariate models. The main idea of multiblock MSPM methods is dividing the process variables into several blocks and combining the monitoring result of each block. However, block division is still an open problem in academic and engineering fields. Though Slama had given a general guideline “blocks should correspond as closely as possible to distinct units of the process where all the variables within a block or process unit may be highly coupled, but where there is minimal coupling among variables in different blocks”[18], this rule is inappropriate for large-scale industrial processes, because (a) in large-scale industrial processes, variables in different process units are still highly coupled; (b) variables in the same unit may be unrelated. In addition, for multiblock MSPM methods, one variable only belongs to one block, as such, the rest blocks may lose key input variables, which causes large model error. For example, for model in Fig. 1, it’s hard to divide the process variables into two or more blocks: when is allocated to block 2, then blocks 1 loses information of . Besides, it’s difficult to divide the blocks with traditional data-driven method, and hence many multiblock MSPM methods demand the process prior knowledge for block division[23].
Figure 1

The traditional multivariate statistical-based process monitoring (MSPM) methods, multiblock MSPM, and minimalist module analysis.

The traditional multivariate statistical-based process monitoring (MSPM) methods, multiblock MSPM, and minimalist module analysis. To eliminate the influence of the redundant correlations among process data, this paper proposes a novel MSPM method called minimalist module analysis (MMA). All variables in the minimalist module are strongly correlated, and no redundant variables exist. As shown in Fig. 1, MMA just analyzes the correlations between variables in the same module, and hence the extracted feature components are not disturbed by the noise from the other modules. In addition, the modularization analysis results can provide more useful information for fault localization. The difference between MMA and the multiblock MSPM methods are as follows: first, for MMA, each variable may belong to more than one modules ( belongs to two modules in Fig. 1), so each module represents one complete correlation without information loss; second, for MMA, module division is based on statistics analysis rather than the process prior knowledge, which is consistent with the data-driven feature of MSPM; third, each module only contain one correlation in MMA, and each block in the multiblock MSPM methods may contain more than one correlations. The main innovations of this study are as follows. First, we propose a modularization method based on singular value decomposition (SVD)[24] and particle swarm optimization (PSO)[25], which can divide the process variables into different minimalist modules and an independent module. Then, we propose new monitoring indices for each module. In addition, we propose a new fault localization strategy for MMA. According to a survey paper[1], PCA is the most commonly used MSPM method. As such, this paper focuses on the comparison of MMA and PCA; our conclusion is also applicable to other algorithms, such as PLS and CCA. The simulation tests in a mathematical model and the Tennessee Eastman (TE) process[26] show that MMA can successfully obtain the minimalist modules; moreover, it achieves much better performance than the traditional MSPM methods in fault detection and fault localization. The remainder of this paper is organized as follows. In “Methods” section, we briefly review some concepts of classical PCA and the contribution plot method, and assess the defects of these methods. “Minimalist module analysis (MMA)” section then proposes MMA for process monitoring, and introduces some details. “Simulation study of MMA” section analyzes the characteristics of MMA, and compares this method with PCA by conducting tests on a mathematical model. “Fault detection in the Tennessee Eastman process” section compares MMA with other improved MSPM methods in the TE process. Lastly, “Conclusions” section summarizes the contributions of this paper, and discusses some directions for future studies.

Methods

Principal component analysis (PCA)

PCA decomposes the data matrix (where n is the number of samples, and s is the number of variables) into a transformed k subspace of reduced dimensions as follows:where refers to the score matrix, which is an orthogonal matrix; refers to the loading matrix, and it is orthonormal; and is the residual matrix. To obtain the loading matrix , one should firstly calculate the covariance matrix: Then, can be presented by singular value decomposition (SVD) as follows:where () is a diagonal matrix. Matrix is actually columns of P0 associated with the k largest eigenvalues, and k is determined by cumulative percent variance (CPV)[27] as follows:where is a parameter usually set to 85%. When CPV is larger than , we take k as the number of the principal components (PCs). Then, two statistics are constructed to monitor the new process data sample as follows:where and . The thresholds for the two indices, and , can be found in reference[28].

Contribution plot

The contributions to SPE are calculated as follows:where and are the jth columns of x and , respectively. The contributions to are calculated as follows:where P is the ith column of P, and P is the element in the jth column and ith row. The role of the contribution plots to fault isolation is to indicate which of the variables are related to the fault rather than to reveal the actual cause of it. In general, variables with a higher contribution have a closer relationship with the fault source. The thresholds of and can be obtained by kernel density estimation[29].

Drawback of PCA and contribution plot method

Theorem

The redundant variables introduce extra noise into the principal components (PCs).

Assume are the variables belonging to a minimalist module, which can be full-rank decomposed aswhere and Matrix are the redundant variables that can be presented as the linear combination of X1 as follows:where . is the linear transformation matrix, and is noise belonging to X2. In this paper, we assume that each measurement variable contains independent sensor noise, and hence, rank(W) = s′. Taking , one obtains Part can be full-rank singular value decomposed aswhere , rank(T1) = rank(T0), and . Hence, one obtains Taking , Because part is non-orthogonal in most situations, we introduce another orthonormal matrix , which makes It should be noted that when is orthogonal, then Q = I. PCA picks the k largest components of T2 as PCs, and we denote them as . Then,where is the corresponding k columns of . Taking , and because and are parts of orthonormal matrices and , one obtains () unless the exceptionally rare situation that all columns of Q belong to the column set of . As , one obtains . As such, T is influenced by W, and the redundant variables X2 introduce extra noise W into the principal components (PCs). This finishes the proof. Based on the Theorem, one finds that PCA is not good at handling process data with redundant variables. As for the contribution plot method, according to Eqs. (6) and (7), it is based on the difference between x and . As shown in Fig. 2, when a fault occurs in a specific variable , (a) according to equation , the relevant principal components are faulty; (b) according to equation , most reconstructed variables are faulty. As such, in a practical engineering application, it is hard to locate the source fault by the contribution plot method because too many variables’ contribution indices alarm the fault.
Figure 2

Fault propagation from original data to reconstructed data.

Fault propagation from original data to reconstructed data.

Section summary

In sum, to eliminate the noise disturbance in the redundant variables, and to improve the fault localization ability, we develop a new monitoring algorithm based on the minimalist module and propose a corresponding fault localization strategy in “Minimalist module analysis (MMA)” section.

Minimalist module analysis (MMA)

The content of this section is listed in Fig. 3 below.
Figure 3

Content of this “Minimalist module analysis (MMA)” section.

Content of this “Minimalist module analysis (MMA)” section.

Minimalist module division

Traditional PCA approaches focus on the k largest eigenvalues in matrix , and the important information contained in the residual part is not used. When is very small (e.g., 0.05), one obtains . Taking P as the columns of P0 associated with the s-k smallest eigenvalues, one obtains We assume , and . Then, Through the transformation of Eq. (17), one obtains As such, one then obtains Unlike , some elements of are 0, and hence Eq. (19) can describe the relationship between x2 and x3 without considering x1. In Eq. (19), variable set is a minimalist module. The flow of minimalist module division is as follows: Find a transformation matrix that maximizes the number of 0 elements in . This paper addresses this optimization problem by using the particle swarm optimization (PSO)[30] algorithm as described below. Step 1 Set num = 1. Step 2 Take the column of P as and the remaining s − k − 1 columns as . Solve the following optimization function by PSO:where denotes the number of elements in interval ( is close to 0, such as 0.01). Step 3 If , go to step 4; else, num = num + 1 and go to step 2. Step 4 . Calculate , adjust each column of to unit variance, and set all elements in interval to 0. Take the variables corresponding to non-zero element parameters in the ith () column of as the ith minimalist module (MMi). The form of the minimalist module is not unique, e.g. through the transformation of Eq. (17), one also obtains.and hence variable set is also a minimalist module. As such, the result of PSO may be different each time.

Independent module

Each variable in the minimalist module is strongly correlated with other variables. As such, some variables, such as x8 and x9 in Fig. 3, are not included in the minimalist module group. Thus, these variables belong to the independent module.

Monitoring indices construction

Each minimalist module can be monitored by the PCA algorithm independently. We assume that are data belonging to MMi. Then, because each minimalist module represents one independent correlation, and hence the number of PCs for each minimalist module is fixed as . The monitoring indices of each module are calculated asandwhere and , and and are the and SPE indices and the corresponding thresholds for MMi, respectively. Different from the traditional SPE index, SPE divides to eliminate the impact of on SPE. The indices for the whole process areandwhere is a positive value (e.g., ). As such, when some minimalist module detects the fault, then these two indices are much larger than their normal values. The threshold for both indices is . As for the variables in the independent module, they can be monitored by the index, which is denoted as .

Fault localization

For MMA, the fault localization rules are different for , , and indices. For the index, when is normal, then all related variables are normal. For example, in the mathematical model in Fig. 3, when and are faulty, and and are normal, then one gets that: (a) variables related to MM1 and MM2, i.e., , , , , and , may be faulty; (b) all variables related to MM3 and MM4, i.e., , , , , and , are normal; (c) must be faulty because it is the only common variable shared by MM1 and MM2, and x2 may also be faulty because we have no more information for judging it. For the index, when is faulty, then the correlation between all variables in MMi maybe faulty. For example, in the mathematical model in Fig. 3, one obtains ; when the correlation between , , and x3 changes to or , then and alarms the fault. When a fault occurs in variables not belonging to the minimalist module, such as x8 and , then they can only be handled with the detection result of the independent module, i.e., the contribution .

Simulation study of MMA

This section aims to study the performance of MMA through simulation tests, and compare it with PCA and mutual information–multiblock PCA (MI-MBPCA)[31]. MI-MBPCA employs mutual information to divide the block automatically and hence it does not need the process prior knowledge for block division. The test model is shown below: Random variables N and follow the standard Gaussian distribution, and indicates the process noise. Approximately 10,000 normal observations are produced for offline modeling. After data normalization, the training data are adjusted to zero-mean and unit-variance. Then the normalized data are processed by MMA. The matrix is obtained as follows: . Thus, MMA successfully obtains four minimalist modules: , , , and . Then, the independent module is . And MI-MBPCA divides the process variables into the following 5 blocks: , , , , , which is not consistent with the process model because is correlated with both and but they do not belong to same block. To compare the monitoring performance between MMA, PCA and MI-MBPCA, five test data sets are generated. Each data set contains 960 samples, and the fault occurs at the 160th sample point. The occurred faults are of the following five types: Fault 1: a step change with amplitude of 5 in x1; Fault 2: term N2 in the expression of x2 changes to ; Fault 3: a step change with amplitude of 0.2 in x3; Fault 4: term in the expression of changes to ; Fault 5: a step change with amplitude of 5 in . The detection results are listed in Table 1. The false alarm rate is calculated as and the detection rate is calculated as . In this study, all control limits are based on a probability of 99% and the best result is marked in bold.
Table 1

False alarm rates (%) and detection rates (%) of the principal component analysis (PCA) method, the mutual information–multiblock PCA (MI-MBPCA), and the minimalist module analysis (MMA) method.

MethodPCAMI-MBPCAMMA
Index\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$T^{2}$$\end{document}T2SPE\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$DR$$\end{document}DR\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$T_{I}^{2}$$\end{document}TI2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$T_{M}^{2}$$\end{document}TM2SPEM
False alarm rate1.93.10.61.91.90.0
Detection rate
 Fault 195.85.389.30.899.00.4
 Fault 229.81.312.50.838.40.4
 Fault 31.30.80.10.81.993.8
 Fault 41.14.75.251.41.590.3
 Fault 533.994.695.897.81.00.4
False alarm rates (%) and detection rates (%) of the principal component analysis (PCA) method, the mutual information–multiblock PCA (MI-MBPCA), and the minimalist module analysis (MMA) method. As shown in Table 1, the performance of MMA is better than that of PCA and MI-MBPCA for all five faults. Because MMA divides the whole process data into several minimalist modules and an independent module, and the noise in each variable will not disturb the unrelated modules, MMA is more robust to process noise than PCA. For MI-MBPCA, because each variable only belongs to one block and the rest blocks may lose key information, the models of blocks maybe biased. One interesting finding in Table 1 is that MMA can successfully detect faults 3 and 4 while PCA fails. The reason for this phenomenon is that PCA monitors the complex correlations between all variables together while MMA monitors each strong correlation (one minimalist module) independently; therefore, MMA is very sensitive to changes in specific correlations. The fault localization results of the two algorithms for faults 3 and 5 are shown in Figs. 4 and 5, respectively. In Fig. 4, for PCA, , , and alarm the fault, and we cannot locate the fault source. For MI-MBPCA, because is influenced by , both variables alarm the fault and we cannot locate the fault source. For MMA, all and indices are normal, which means that all variables in the independent module are normal and all variables in the minimalist modules fluctuate within the normal range; because signals a fault alarm, one finds that the correlations between , , and are changed.
Figure 4

Fault localization for fault 3.

Figure 5

Fault localization for fault 5.

Fault localization for fault 3. Fault localization for fault 5. In Fig. 5, although a fault occurs in , most indices in PCA signal a fault alarm, and we cannot locate the fault source. For MI-MBPCA, it can successfully locate the fault source. However, because MI-MBPCA fails in detecting fault 5, and hence the fault localization step is skipped, as such, MI-MBPCA also fails in locating the fault source. For MMA, all and are normal, and hence one finds that the fault is not in the minimalist modules; only signals a fault alarm, and hence MMA successfully locates the faulty variable .

Fault detection in the Tennessee Eastman process

The Tennessee Eastman (TE) process[32] simulation is the most widely used simulation model to test the MSPM methods, which is outlined in Fig. 6. The TE process uses 12 manipulated variables, 22 continuous process measurements, and 19 composition measurements sampled less frequently to simulate a classical chemical process. Because the 19 composition measurements are difficult to measure in real time and one manipulated variable, i.e., the agitation speed, is not manipulated, this study only monitors the other 22 measurements and 11 manipulated variables, as listed in Table 2. Twenty-one programmed faults that are introduced in the TE process are listed in Table 3. In this study, 960 normal samples are adopted as training data to construct the monitoring models. Each testing data set contains 960 samples, and fault occurs at the 161st sample.
Figure 6

Schematic of the Tennessee Eastman process[33].

Table 2

Monitored variables in the Tennessee Eastman process[33].

Variable
1 A feed (stream 1)18 Stripper temperature
2 D feed (stream 2)19 Stripper steam flow
3 E feed (stream 3)20 Compressor work
4 Total feed (stream 4)21 Reactor cooling water outlet temperature
5 Recycle flow (stream 8)22 Separator cooling water outlet temperature
6 Reactor feed rate (stream 6)23 D feed flow valve (stream 2)
7 Reactor pressure24 E feed flow valve (stream 3)
8 Reactor level25 A feed flow valve (stream 1)
9 Reactor temperature26 Total feed flow valve (stream 4)
10 Purge rate (stream 9)27 Compressor recycle valve
11 Product separator temperature28 Purge valve (stream 9)
12 Product separator level29 Separator pot liquid flow valve (stream 10)
13 Product separator pressure30 Stripper liquid product flow valve (stream 11)
14 Product separator under flow (stream 10)31 Stripper steam valve
15 Stripper level32 Reactor cooling water flow
16 Stripper pressure33 Condenser cooling water flow
17 Stripper underflow (stream 11)
Table 3

Descriptions of faults in the Tennessee Eastman process[33].

No.DescriptionType
1Feed ratio of A/C, composition constant of B (stream 4)Step
2Composition of B, ratio constant of A/C (stream 4)Step
3Feed temperature of D (stream 2)Step
4Inlet temperature of reactor cooling waterStep
5Inlet temperature of condenser cooling waterStep
6Feed loss of A (stream 1)Step
7Header pressure loss of C—reduced availability (stream 4)Step
8Feed composite of A, B, and C (stream 4)Random variation
9Feed temperature of D (stream 2)Random variation
10Feed temperature of C (stream 4)Random variation
11Inlet temperature of reactor cooling waterRandom variation
12Inlet temperature of condenser cooling waterRandom variation
13Reaction kineticsSlow drift
14Valve of reactor cooling waterSticking
15Valve of condenser cooling waterSticking
16–20UnknownUnknown
21The valve for stream 4 was fixed at the steady-state positionConstant position
Schematic of the Tennessee Eastman process[33]. Monitored variables in the Tennessee Eastman process[33]. Descriptions of faults in the Tennessee Eastman process[33]. In this section, we compare MMA with PCA, MI-MBPCA, Deep principal component analysis (DePCA)[34], and kernel dynamic PCA (KDPCA)[35]; the latter two methods are improved versions of PCA. The detection results of the four methods are listed in Table 4. The false alarm rate is calculated as the , and the detection rate is calculated as . In this study, all control limits are based on a probability of 99% and the best result is marked in bold.
Table 4

False alarm rates (%) and detection rates (%) of the four fault detection methods.

MethodPCADePCAKDPCAMI-MBPCAMMA
Index\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$T^{2}$$\end{document}T2SPEET2ESPE\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$T^{2}$$\end{document}T2SPE\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$DR$$\end{document}DR\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$T_{I}^{2}$$\end{document}TI2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$T_{M}^{2}$$\end{document}TM2SPEM
False alarm rate0.51.46.111.511.214.051.250.81.30.2
Detection rate
 Fault 199.199.999.1100.099.099.699.944.6100.00.0
 Fault 298.495.898.598.098.396.698.074.498.60.0
 Fault 30.92.617.617.40.93.10.81.96.61.8
 Fault 420.9100.078.3100.020.299.9100.00.3100.00.9
 Fault 524.320.938.845.024.024.823.514.033.0100.0
 Fault 699.1100.099.4100.098.999.9100.093.9100.0100.0
 Fault 7100.0100.0100.0100.099.999.9100.0100.045.63.6
 Fault 896.983.697.598.396.893.097.860.898.43.0
 Fault 91.81.816.914.01.53.12.51.56.01.3
 Fault 1029.925.857.158.129.527.641.86.888.50.1
 Fault 1140.674.986.385.040.574.982.50.989.41.1
 Fault 1298.489.599.699.398.393.499.066.199.652.0
 Fault 1393.695.394.495.193.595.095.466.595.622.9
 Fault 1499.3100.0100.0100.099.199.999.90.3100.00.0
 Fault 151.43.017.819.61.33.42.51.611.62.0
 Fault 1613.527.443.557.413.727.827.13.691.974.1
 Fault 1776.495.491.694.476.594.893.51.097.10.1
 Fault 1889.390.192.192.089.390.389.688.191.083.9
 Fault 1911.012.568.868.98.721.013.81.690.448.3
 Fault 2031.849.863.561.831.250.857.42.483.981.0
 Fault 2139.347.354.661.835.350.147.439.866.30.6
False alarm rates (%) and detection rates (%) of the four fault detection methods. As shown in Table 4, we find that MMA, MI-MBPCA, and PCA achieve similar false alarm rates, and their values are much lower than those of the two improved PCA methods (over 10%). For fault detection rates, MMA achieves the best results in 17 of the 21 faults; as for the remaining 4 faults, MMA’s detection rates are not as high as those of DePCA only because DePCA sacrifices the false alarm rate. An eye-catching result is obtained in the case of fault 5: the detection rates of the compared methods are generally below 50%, whereas MMA achieves a 100.0% detection rate, which indicates the superiority of MMA. In addition, the performance of MMA in faults 10, 16, 19, and 20 is much better than that of the other four methods. As the papers that proposed DePCA and KDPCA did not give a description of the contribution plot construction, we only compare the fault localization ability between PCA, MI-MBPCA, and MMA. The matrix of MMA is shown in Table 5.
Table 5

Matrix for the Tennessee Eastman process.

VariableMM
1234567891011121314
1− 0.10.00.3− 0.30.10.00.00.30.0− 0.2− 0.20.30.00.0
20.10.00.10.10.30.00.00.0− 0.20.00.00.2− 0.10.1
30.00.00.00.00.00.00.00.00.00.00.00.00.00.0
40.00.00.00.00.00.00.00.00.00.00.00.00.00.0
50.00.00.10.00.00.00.00.00.00.00.00.00.00.1
60.00.00.00.00.00.00.00.00.00.00.00.00.00.0
7− 0.3− 0.60.2− 0.30.00.40.00.1− 0.10.00.00.00.00.4
80.00.00.00.00.00.00.00.00.00.00.00.00.00.0
90.50.00.00.00.00.30.0− 0.30.00.00.00.00.40.0
10− 0.10.0− 0.20.00.10.00.00.20.1− 0.20.00.0− 0.20.1
110.00.10.20.00.10.00.40.20.0− 0.40.10.30.00.0
120.00.00.00.10.00.00.00.20.00.00.00.00.00.0
130.00.50.30.00.30.00.0− 0.3− 0.20.0− 0.1− 0.50.4− 0.1
140.00.00.00.00.00.00.00.00.00.00.00.00.00.0
15− 0.10.00.00.00.00.00.00.00.00.00.10.00.00.4
160.50.5− 0.1− 0.1− 0.40.1− 0.30.40.50.40.00.2− 0.20.3
170.00.00.00.00.00.0− 0.40.0− 0.50.00.00.00.00.0
180.00.20.0− 0.20.20.6− 0.4− 0.10.0− 0.10.10.00.10.0
190.30.00.60.0− 0.4− 0.1− 0.10.00.00.40.50.0− 0.20.5
200.00.0− 0.10.60.1− 0.50.30.00.0− 0.60.00.10.4− 0.4
210.30.10.20.20.50.00.00.0− 0.30.0− 0.20.50.00.0
220.10.00.00.10.10.0− 0.4− 0.10.00.3− 0.10.0− 0.10.0
230.00.00.00.20.3− 0.10.00.0− 0.10.0− 0.10.2− 0.1− 0.1
240.00.00.00.10.00.00.00.00.0− 0.10.10.00.00.0
250.0− 0.2− 0.40.40.0− 0.10.0− 0.4− 0.10.00.4− 0.2− 0.10.0
260.00.00.00.00.00.00.00.00.00.00.00.00.00.0
270.0− 0.2− 0.2− 0.20.00.20.20.00.00.00.20.3− 0.50.0
280.00.00.10.0− 0.2− 0.10.0− 0.10.00.10.00.00.2− 0.1
290.00.00.0− 0.10.00.00.0− 0.20.00.00.00.00.00.0
300.10.00.00.00.00.00.00.00.00.0− 0.10.00.0− 0.4
310.00.0− 0.2− 0.20.1− 0.10.00.20.00.2− 0.60.00.00.0
32− 0.50.00.00.00.0− 0.30.00.30.00.00.00.0− 0.40.0
330.00.00.00.00.00.0− 0.40.0− 0.50.00.00.00.00.0

Significant values are in [bold].

Matrix for the Tennessee Eastman process. Significant values are in [bold]. Figure 7 shows the fault localization results of fault 4. According to Table 3, fault 4 is a step change in inlet temperature of reactor cooling water. As depicted in Fig. 6, the reactor temperature (variable 9 in Table 2) changes, and hence the reactor cooling water flow (variable 32 in Table 2) also changes to compensate for the temperature change. For PCA, , , and signal a fault alarm; for MI-MBPCA, about 14 variables alarm the fault and it fails in locating the fault source; for MMA, , , , and signal a fault alarm based on the fault localization rules presented in “Monitoring indices construction” section, and then one finds that variables 9 and 32 are faulty. Both PCA and MMA can locate this fault. Different from the contribution plot method of PCA, all of MMA are normal, which tells the engineers that the correlation between variables have not changed, and hence the fault source is the change in amplitude of some variables. Thus, it can be seen that, compared with PCA, MMA can provide more useful information for fault localization.
Figure 7

Fault localization for fault 4.

Fault localization for fault 4.

Conclusions

In this study, a new MSPM called MMA was proposed to overcome the shortcoming of the traditional MSPM method in handling the redundant correlations among process variables. The superiority of MMA was verified by several simulation tests. It achieved much better detection performance for five different types of faults on a mathematical model test, and two of which could not be detected by PCA and MI-MBPCA. MMA also had a better performance than other improved MSPM algorithms for 17 of the 21 faults in the Tennessee Eastman process. MMA is a completely new method, and hence much work can be done based on it. First, we can combine it with the traditional nonlinear, dynamic, robust strategy to improve its fault detection ability. We can also combine it with the traditional contribution plot method to improve its fault localization ability. Moreover, we can combine it with the key performance indicator[14] monitoring strategy. All of these investigations will be part of our future work.
  2 in total

1.  In-Depth Evaluation of Data Collected During a Continuous Pharmaceutical Manufacturing Process: A Multivariate Statistical Process Monitoring Approach.

Authors:  Ana F Silva; Jurgen Vercruysse; Chris Vervaet; Jean P Remon; João A Lopes; Thomas De Beer; Mafalda C Sarraguça
Journal:  J Pharm Sci       Date:  2018-08-27       Impact factor: 3.534

2.  Plant-wide process monitoring based on mutual information-multiblock principal component analysis.

Authors:  Qingchao Jiang; Xuefeng Yan
Journal:  ISA Trans       Date:  2014-06-20       Impact factor: 5.468

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.