| Literature DB >> 30177670 |
Huihui Qiao1,2, Taiyong Wang3,4, Peng Wang5,6, Shibin Qiao7, Lan Zhang8,9.
Abstract
Data-driven methods with multi-sensor time series data are the most promising approaches for monitoring machine health. Extracting fault-sensitive features from multi-sensor time series is a daunting task for both traditional data-driven methods and current deep learning models. A novel hybrid end-to-end deep learning framework named Time-distributed ConvLSTM model (TDConvLSTM) is proposed in the paper for machine health monitoring, which works directly on raw multi-sensor time series. In TDConvLSTM, the normalized multi-sensor data is first segmented into a collection of subsequences by a sliding window along the temporal dimension. Time-distributed local feature extractors are simultaneously applied to each subsequence to extract local spatiotemporal features. Then a holistic ConvLSTM layer is designed to extract holistic spatiotemporal features between subsequences. At last, a fully-connected layer and a supervised learning layer are stacked on the top of the model to obtain the target. TDConvLSTM can extract spatiotemporal features on different time scales without any handcrafted feature engineering. The proposed model can achieve better performance in both time series classification tasks and regression prediction tasks than some state-of-the-art models, which has been verified in the gearbox fault diagnosis experiment and the tool wear prediction experiment.Entities:
Keywords: deep learning; machine health monitoring; multi-sensor time series; spatiotemporal feature learning; time-distributed ConvLSTM model
Mesh:
Year: 2018 PMID: 30177670 PMCID: PMC6164508 DOI: 10.3390/s18092932
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1The structure of three gates in the Long Short-Term Memory (LSTM) cell.
Figure 2The framework of the basic ConvLSTM model.
Figure 3The framework of the proposed TDConvLSTM model (TDConvLSTM).
Figure 4Local spatiotemporal feature extractor: (a) Structure of the local spatiotemporal feature extractor; (b) Diagram of the recurrent cell in ConvLSTM.
Figure 5Gearbox test rig: (a) Main units: (1) Motor (2) Parallel gearbox (3) Magnetic powder brake (b) Locations of sensors.
Description of four gearbox health conditions.
| Label | Condition | Description | Speed (rpm) |
|---|---|---|---|
| 0 | FT | A root fracture tooth in the big gear | 280, 860 and 1450 |
| 1 | CT | A root crack tooth in the big gear | 280, 860 and 1450 |
| 2 | CTFT | A root crack tooth in the big gear and a half fracture tooth in the small gear | 280, 860 and 1450 |
| 3 | CIB | A crack on the inner race of the bearing | 280, 860 and 1450 |
Parameters of the proposed model used in gearbox fault diagnosis experiments.
| No. | Layer Type | Kernel | Stride | Channel | BN Axis | Activation |
|---|---|---|---|---|---|---|
| 1 | Local Convolution | (4,1) | (4,1) | 4 | 4 | sigmoid |
| 2 | Local ConvLSTM | (1,4) | (1,1) | 4 | 5 | tanh |
| 3 | Holistic ConvLSTM | (2,2) | (1,1) | 4 | 4 | tanh |
| 4 | FC layer | 100 | - | 1 | −1 | sigmoid |
| 5 | Supervised learning layer | 4 | - | 1 | - | softmax |
Figure 6Accuracy and training time under different length of the subsequence.
Testing accuracy of comparative methods.
| Model | Constant Rotation Speed | Nonstationary Rotation Speed | ||
|---|---|---|---|---|
| D1 | D2 | D3 | D4 | |
| TDConvLSTM |
|
|
|
|
| TDConvLSTM | 100% | 99.5% | 99.78% | 93.11% |
| CNN-LSTM | 98.67% | 97% | 98.33% | 91.89% |
| CNN | 96.83% | 99.5% | 98.17% | 86.78% |
| LSTM | 96.67% | 99.83% | 100% | 80.94% |
| EMD-SVM | 90.67% | 89.67% | 91.33% | 75.67% |
Figure 7The effect of batch normalization (BN) on model training.
Figure 8Feature visualization: (a) Raw data; (b) Local Convolutional layer; (c) Local ConvLSTM layer; (d) Holistic ConvLSTM layer; (e) Fully-connected (FC) layer.
Figure 9The experiment setup for tool wear monitoring
Parameters of the proposed model used in tool wear monitoring experiments
| No. | LAYER TYPE | Kernel | Stride | Channel | BN Axis | Activation |
|---|---|---|---|---|---|---|
| 1 | Local Convolution | (10,3) | (5,3) | 4 | 4 | ReLu |
| 2 | Local ConvLSTM | (2,2) | (1,1) | 4 | 5 | tanh |
| 3 | Holistic ConvLSTM | (4,4) | (1,1) | 1 | 4 | tanh |
| 4 | FC layer | 10 | - | 1 | −1 | ReLu |
| 5 | Supervised learning layer | 1 | - | 1 | - | linear |
Mean absolute error (MAE) and root mean square error (RMSE) of models.
| Model | MAE | RMSE | ||||
|---|---|---|---|---|---|---|
| C4,C6/C1 1 | C1,C6/C4 2 | C1,C4/C6 3 | C4,C6/C1 | C1,C6/C4 | C1,C4/C6 | |
| TDConvLSTM |
|
|
|
|
|
|
| CNN-LSTM | 11.18 | 9.39 | 11.34 | 13.77 | 11.85 | 14.33 |
| CNN | 15.32 | 14.34 | 17.36 | 18.50 | 18.80 | 21.85 |
| LSTM | 19.09 | 16.00 | 22.61 | 21.42 | 17.78 | 25.81 |
1 “C4,C6/C1” denote that C4 and C6 are the training datasets, C1 is the testing dataset. 2 “C1,C6/C4” denote that C1 and C6 are the training datasets, C4 is the testing dataset. 3 “C1,C4/C6” denote that C1 and C4 are the training datasets, C6 is the testing dataset.
Figure 10Regression performances of TDConvLSTM for three different testing cases: (a) C1; (b) C4; (c) C6.