| Literature DB >> 36188261 |
Honghua Chen1,2, Jian Cen1,2, Zhuohong Yang1,2, Weiwei Si1,2, Hongchao Cheng1,2.
Abstract
Deep learning provides new ideas for chemical process fault diagnosis, reducing potential risks and ensuring safe process operation in recent years. To address the problem that existing methods have difficulty extracting the dynamic fault features of a chemical process, a fusion model (CS-IMLSTM) based on a convolutional neural network (CNN), squeeze-and-excitation (SE) attention mechanism, and improved long short-term memory network (IMLSTM) is developed for chemical process fault diagnosis in this paper. First, an extended sliding window is utilized to transform data into augmented dynamic data to enhance the dynamic features. Second, the SE is utilized to optimize the key fault features of augmented dynamic data extracted by CNN. Then, IMLSTM is used to balance fault information and further mine the dynamic features of time series data. Finally, the feasibility of the proposed method is verified in the Tennessee-Eastman process (TEP). The average accuracies of this method in two subdata sets of TEP are 98.29% and 97.74%, respectively. Compared with the traditional CNN-LSTM model, the proposed method improves the average accuracies by 5.18% and 2.10%, respectively. Experimental results confirm that the method developed in this paper is suitable for chemical process fault diagnosis.Entities:
Year: 2022 PMID: 36188261 PMCID: PMC9521029 DOI: 10.1021/acsomega.2c04017
Source DB: PubMed Journal: ACS Omega ISSN: 2470-1343
Figure 1Basic structure of the CNN.
Figure 2Basic structure of the SE block.
Figure 3Structures of the (a) LSTM and (b) IMLSTM.
Figure 4Extended sliding window mechanism schematic.
Figure 5Fault diagnosis flow based on the CS-IMLSTM model.
Figure 6Flowchart of TEP.[36] Reprinted with permission from ref (36). Copyright 2019 Elsevier.
Fault Modes of Case 1 and Case 2
| case | fault | fault cause | fault type |
|---|---|---|---|
| Case 1 | 1 | A/C feed ratio fluctuates, B feed is stable | Step |
| 2 | B feed fluctuates, A/C feed ratio is stable | Step | |
| 6 | A material leak | Step | |
| 7 | Feed C inlet pressure loss: availability reduction | Step | |
| 8 | A, B, C feed composition fluctuation | Random variable | |
| Case 2 | 4 | Temperature disturbance at reactor cooling water inlet | Step |
| 5 | Temperature disturbance at reactor cooling water inlet | Step | |
| 10 | C feed temperature disturbance | Random variable | |
| 11 | Inlet temperature fluctuation of reactor cooling water | Random variable | |
| 12 | Inlet temperature fluctuation of condenser cooling water | Random variable |
Sample Size of Raw Data and Augmented Dynamic Data
| fault | data set | sample size of raw data | sample size of augmented dynamic data |
|---|---|---|---|
| Each fault | Train | 480 | 461 |
| Test | 800 | 781 |
Model Structure and Parameter Settings
| model | structure |
|---|---|
| CS-IMLSTM | CONV(32)-SE(32)-CONV(64)-SE(64)-CONV(64)-SE(64)-FC*(512)-IMLSTM(1024)-FC(5) |
| CNN-IMLSTM | CONV(32)-CONV(64)-CONV(64)-FC*(512)-IMLSTM(1024)-FC(5) |
| CS-LSTM | CONV(32)-SE(32)-CONV(64)-SE(64)-CONV(64)-SE(64)-FC*(512)-LSTM(1024)-FC(5) |
| CNN-LSTM | CONV(32)-CONV(64)-CONV(64)-FC*(512)-LSTM(1024)-FC(5) |
| LSTM | Lstm1(1024)-lstm2(1024)-lstm3(1024)-lstm4(1024)-FC(5) |
| CS-CNN | CONV(32)-SE(32)-CONV(64)-SE(64)-CONV(64)-SE(64)-FC*(512)-CNN(512)-FC(5) |
For convenience, the CONV(@) module is used to denote Conv1d(@)-BN(@)-LeakyReLU-maxpooling(@), where @ denotes the output channel. * indicates FC layer with dropout rate of p = 0.5.
Figure 7The 10 times average training loss curves of (a) case 1 and (b) case 2 on different models.
Figure 8The 10 times average validation loss curves of (a) case 1 and (b) case 2 on different models.
Classification Accuracy of Each Fault in Each Model
| case | fault | proposed model (%) | CS-LSTM (%) | CNN-IMLSTM (%) | CNN-LSTM (%) | LSTM (%) | CS-CNN (%) |
|---|---|---|---|---|---|---|---|
| Case 1 | 1 | 99.36 | 99.35 | 98.85 | 99.36 | 97.57 | |
| 2 | 99.74 | 99.87 | 99.62 | 99.62 | 99.49 | ||
| 6 | 99.23 | 99.87 | 99.87 | 98.98 | 99.74 | ||
| 7 | 90.39 | 97.95 | 86.94 | 94.88 | 84.76 | ||
| 8 | 94.24 | 88.35 | 88.99 | 85.53 | 92.70 | ||
| Case 2 | 4 | 99.74 | 99.87 | 95.26 | 97.95 | ||
| 5 | 97.70 | 98.98 | 98.72 | 91.93 | 97.18 | ||
| 10 | 97.95 | 96.41 | 92.70 | 73.24 | 83.99 | ||
| 11 | 90.01 | 91.17 | 90.78 | 47.25 | 89.88 | ||
| 12 | 96.67 | 96.93 | 97.18 | 70.04 | 97.06 |
Figure 9Classification results of (a) case 1 and (b) case 2 under each model.
Figure 10Confusion matrix for worst case prediction in case 1.
Analytical Results of the Worst Confusion Matrix in Case 1
| indicator | fault 1 | fault 2 | fault 6 | fault 7 | fault 8 |
|---|---|---|---|---|---|
| PPV (%) | 100 | 95.13 | 100 | 98.96 | 96.74 |
| TPR (%) | 99.49 | 100 | 98.98 | 97.31 | 94.88 |
| F1_Score (%) | 99.74 | 97.50 | 99.49 | 98.13 | 95.80 |
| MacroF1_Score (%) | 98.13 | ||||
Figure 11Confusion matrix for worst case prediction in case 2.
Analytical Results of the Worst Confusion Matrix in Case 2
| indicator | fault 4 | fault 5 | fault 10 | fault 11 | fault 12 |
|---|---|---|---|---|---|
| PPV (%) | 94.53 | 98.10 | 95.71 | 100 | 99.87 |
| TPR (%) | 99.62 | 99.49 | 100 | 90.78 | 97.82 |
| F1_Score (%) | 97.01 | 98.79 | 97.81 | 95.17 | 98.84 |
| MacroF1_Score (%) | 98.13 | ||||
Compare with Existing Advanced Methods
| method | case 1 | case 2 | average accuracy |
|---|---|---|---|
| DPCA-SVM | 89.73% | 60.83% | 75.28% |
| transformer | 97.23% | 78.51% | 87.87% |
| proposed method |