| Literature DB >> 35042894 |
Abstract
Supervised machine learning is a traditionally remaining useful life (RUL) estimation tool, which requires a lot of prior knowledge. For the situation lacking labeled data, supervised methods are invalid for the issue of domain shift in data distribution. In this paper, a adversarial-based domain adaptation (ADA) architecture with convolution neural networks (CNN) for RUL estimation of bearings under different conditions and platforms, referred to as ADACNN, is proposed. Specifically, ADACNN is trained in source labeled data and fine-tunes to similar target unlabeled data via an adversarial training and parameters shared mechanism. Besides a feature extractor and source domain regressive predictor, ADACNN also includes a domain classifier that tries to guide feature extractor find some domain-invariant features, which differents with traditional methods and belongs to a unsupervised learning in target domain, which has potential application value and far-reaching significance in academia. In addition, according to different first predictive time (FPT) detection mechanisms, we also explores the impact of different FPT detection mechanisms on RUL estimation performance. Finally, according to extensive experiments, the results of RUL estimation of bearing in cross-condition and cross-platform prove that ADACNN architecture has satisfactory generalization performance and great practical value in industry.Entities:
Year: 2022 PMID: 35042894 PMCID: PMC8766616 DOI: 10.1038/s41598-021-03835-2
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1The architecture of DANN.
Figure 2The architecture of ADACNN.
Figure 3The schematic diagram of the process of the proposed model.
The information of FEMTO dataset.
| Operating conditions | Training data | Test data |
|---|---|---|
| A1 (1800 rpm and 4000 N) | A1-1 | A1-3 |
| A2 (1650 rpm and 4200 N) | A2-1 | A2-3 |
| A3 (1500 rpm and 5000 N) | A3-1 | A3-3 |
The information of XJTU-SY dataset.
| Operating conditions | Training data | Test data |
|---|---|---|
| B1 (2100 rpm and 12 kN) | B1-1 | B1-3 |
| B2 (2250 rpm and 11 kN) | B2-1 | B2-3 |
Hyperparameter evaluated in the proposed method.
| Hyper-parameter | Range |
|---|---|
| Layers: ( | |
| Units: ( | |
| Learning rate: ( | |
| Dropout rate | |
| Batch size | |
| Threshold of patience M | |
| Max iteration N |
Selected hyperparemeter for each source-target experiment pair.
| No. | From to | CNN: Layer, (units), [Dropout] | Source regression: Layers, (units), [Dropout] | Domain classification: Layers, (units), [Dropout] | Batch size | ( | ||
|---|---|---|---|---|---|---|---|---|
| E1 | A1 | 2, (128, 64), 0.9 | 512 | 1, (64), 0.1 | 2, (256, 128), 0.9 | 2 | 256 | 0.01, 0.01 |
| E2 | A1 | 2, (128, 32), 0.5 | 64 | 2, (32, 16), 0.3 | 2, (32, 16), 0.3 | 0.8 | 256 | 0.0001, 0.0001 |
| E3 | A2 | 2, (64, 32), 0.1 | 64 | 2, (32, 32), 0.1 | 2, (16, 16), 0.1 | 1 | 256 | 0.001, 0.001 |
| E4 | A2 | 2, (128, 64), 0.1 | 512 | 2, (64, 32), 0.1 | 2, (64,32), 0.1 | 2 | 256 | 0.001,0.001 |
| E5 | A3 | 2, (64,32), 0.3 | 128 | 2, (64, 64), 0.1 | 2, (64, 64), 0.1 | 2 | 256 | 0.001, 0.001 |
| E6 | A3 | 2, (128, 32), 0.9 | 512 | 2, (128, 128), 0.1 | 2, (128, 64), 0.1 | 2 | 256 | 0.001, 0.001 |
| E7 | B1 | 2, (128, 32), 0.1 | 64 | 2, (256, 128), 0.1 | 2, (128, 64), 0.5 | 0.8 | 128 | 0.001, 0.001 |
| E8 | B2 | 2, (128, 32), 0.9 | 64 | 2, (256, 128), 0.1 | 2, (128, 64), 0.9 | 0.8 | 128 | 0.001, 0.001 |
| E9 | A1 | 2, (32, 32), 0.1 | 64 | 1, (64), 0.1 | 1, (64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
| E10 | A1 | 2, (128, 32), 0.1 | 64 | 1, (64), 0.1 | 2, (64, 64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
| E11 | A2 | 2, (32, 32), 0.1 | 64 | 1, (64), 0.1 | 1, (64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
| E12 | A2 | 2, (32, 32), 0.1 | 64 | 1, (64), 0.1 | 1, (64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
| E13 | A3 | 2, (32, 32), 0.1 | 64 | 1, (64), 0.1 | 1, (64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
| E14 | A3 | 2, (32, 32), 0.1 | 64 | 1, (64), 0.1 | 1, (64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
| E15 | B1 | 2, (32,32), 0.9 | 64 | 1, (64), 0.9 | 2, (64,64), 0.5 | 0.8 | 64 | 0.001,0.001 |
| E16 | B1 | 2, (128, 32), 0.9 | 64 | 1, (64), 0.9 | 2, (64, 64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
| E17 | B1 | 2, (128, 32), 0.9 | 64 | 1, (64), 0.9 | 2, (64, 64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
| E18 | B2 | 2, (64,64), 0.1 | 64 | 1, (32), 0.9 | 2, (64,64), 0.5 | 1 | 64 | 0.001,0.001 |
| E19 | B2 | 2, (128, 32), 0.9 | 64 | 1, (64), 0.9 | 2, (64, 64), 0.5 | 2 | 64 | 0.001, 0.001 |
| E20 | B2 | 2, (128, 32), 0.9 | 64 | 1, (64), 0.1 | 2, (64, 64), 0.5 | 0.8 | 64 | 0.001, 0.001 |
Figure 4RUL estimation comparisons with different FPT detection mechanisms on FEMTO and XJTU-SY datasets: (a) A3 A1, (b) A3 A2, (c) B1 B2, (d) B2 B1.
Figure 5RUL estimation comparisons with source-only and target-only methods on FEMTO and XJTU-SY datasets. (a) E5: A3 A1, (b) E6: A3 A2, (c) E7: B1 B2, (d) E8: B2 B1.
RMSE/score ± standard deviation comparison between source-only, target-only, NoFPT-ADACNN, Kur-ADACNN and MD-ADACNN on FEMTO dataset and XJTU-SY dataset.
| No. | Source-only | Target-only | MD-ADACNN | NoFPT-ADACNN | Kur-ADACNN | |||||
|---|---|---|---|---|---|---|---|---|---|---|
| RMSE | Score | RMSE | Score | RMSE | Score | RMSE | Score | RMSE | Score | |
| E1 | 53.2 | 101.3 | 32.6 | 64.6 | 36.2 ± 1.6 | 74.0 ± 46.4 | 41.5 ± 2.7 | 91.9 ± 66.7 | 36.5 ± 2.5 | 88.3 ± 88.2 |
| E2 | 26.0 | 12.5 | 8.7 | 1.0 | 8.5 ± 14.4 | 5.1 ± 14.0 | 3.4 ± 4.0 | 0.3 ± 0.4 | 2.9 ± 0.5 | 0.3 ± 0.1 |
| E3 | 39.4 | 239.8 | 22.5 | 15.3 | 26.8 ± 5.9 | 14.4 ± 10.9 | 44.2 ± 7.4 | 328.0 ± 700.0 | 64.3 ± 8.9 | 8542.5 ± 9330.2 |
| E4 | 30.5 | 20.2 | 24.6 | 10.7 | 19.8 ± 18.8 | 10.3 ± 4.5 | 9.9 ± 9.1 | 1.9 ± 2.8 | 3.8 ± 1.3 | 0.4 ± 0.1 |
| E5 | 67.0 | 2084.7 | 33.5 | 930.4 | 31.4 ± 3.2 | 7.1 ± 4.7 | 40.0 ± 7.4 | 136.6 ± 215.9 | 52.0 ± 4.0 | 544.4 ± 322.3 |
| E6 | 48.8 | 1752.6 | 30.7 | 35.9 | 35.4 ± 0.8 | 37.2 ± 9.1 | 43.6 ± 14.6 | 1088.6 ± 2835.6 | 37.3 ± 2.0 | 89.4 ± 43.4 |
| E7 | 18.3 | 10.1 | 13.0 | 1.0 | 13.3 ± 0.3 | 1.2 ± 0.6 | 14.0 ± 2.9 | 2.1 ± 1.1 | 17.6 ± 3.4 | 9.6 ± 1.0 |
| E8 | 5.6 | 0.7 | 3.1 | 0.3 | 3.6 ± 0.7 | 0.4 ± 0.4 | 5.3 ± 2.2 | 0.6 ± 0.3 | 4.1 ± 0.8 | 0.7 ± 0.1 |
RMSE/score ± standard deviation comparison under cross-platform on FEMTO and XJTU-SY.
| No. | From to | MD-ADACNN | NoFPT-ADACNNs | Kur-ADACNN | |||
|---|---|---|---|---|---|---|---|
| RMSE | Score | RMSE | Score | RMSE | Score | ||
| E9 | A1 | 6.1 ± 2.6 | 0.9 ± 0.5 | 6.6 ± 2.1 | 0.8 ± 0.4 | 2.3 ± 2.0 | 0.2 ± 0.1 |
| E10 | A1 | 60.9 ± 4.6 | 4027.6 ± 1904.8 | 54.1 ± 3.4 | 2531.6 ± 1341.1 | 65.2 ± 3.9 | 22933.6 ± 10923.1 |
| E11 | A2 | 1.7 ± 1.6 | 0.1 ± 0.2 | 4.5 ± 3.7 | 0.4 ± 0.5 | 4.6 ± 3.5 | 0.7 ± 0.9 |
| E12 | A2 | 8.5 ± 12.7 | 10.4 ± 30.1 | 18.7 ± 15.4 | 18.3 ± 29.3 | 63.1 ± 15.5 | 97874.2 ± 125079.4 |
| E13 | A3 | 4.4 ± 2.1 | 0.5 ± 0.3 | 6.9 ± 2.6 | 0.9 ± 0.6 | 4.4 ± 5.2 | 0.5 ± 0.7 |
| E14 | A3 | 13.7 ± 9.0 | 7.8 ± 11.6 | 23.9 ± 10.3 | 71.1 ± 141.4 | 44.2 ± 6.4 | 117.3 ± 85.6 |
| E15 | B1 | 32.7 ± 5.1 | 59.3 ± 86.9 | 41.5 ± 8.4 | 81.6 ± 70.7 | 45.8 ± 13.3 | 101.3 ± 86.4 |
| E16 | B1 | 32.6 ± 1.7 | 26.5 ± 7.8 | 36.3 ± 3.3 | 44.8 ± 23.0 | 36.2 ± 2.0 | 64.1 ± 21.5 |
| E17 | B1 | 21.6 ± 6.8 | 10.2 ± 8.6 | 7.7 ± 14.0 | 3.0 ± 6.2 | 5.2 ± 1.8 | 0.7 ± 0.2 |
| E18 | B2 | 30.2 ± 5.8 | 25.1 ± 16.0 | 31.9 ± 7.1 | 14.9 ± 6.0 | 34.8 ± 10.0 | 39.6 ± 18.9 |
| E19 | B2 | 35.5 ± 1.9 | 43.4 ± 13.4 | 35.8 ± 3.2 | 122.1 ± 179.7 | 47.0 ± 9.0 | 1510.6 ± 2003.2 |
| E20 | B2 | 1.5 ± 2.6 | 0.1 ± 0.3 | 10.7 ± 14.9 | 4.4 ± 8.0 | 6.1 ± 0.5 | 0.8 ± 0.1 |
Figure 6RUL estimation comparisons with different FPT detection mechanisms between two platforms: (a) E11: A2 B1, (b) E12: A2 B2, (c) E16: B1 A2, (d) E19: B2 A2.
Figure 7The feature visualization for cross-condition.
Figure 8The feature visualization for cross-platform.