| Literature DB >> 35746322 |
Ko-Chieh Chao1, Chuan-Bi Chou1, Ching-Hung Lee1.
Abstract
Traditional machine learning methods rely on the training data and target data having the same feature space and data distribution. The performance may be unacceptable if there is a difference in data distribution between the training and target data, which is called cross-domain learning problem. In recent years, many domain adaptation methods have been proposed to solve this kind of problems and make much progress. However, existing domain adaptation approaches have a common assumption that the number of the data in source domain (labeled data) and target domain (unlabeled data) is matched. In this paper, the scenarios in real manufacturing site are considered, that the target domain data is much less than source domain data at the beginning, but the number of target domain data will increase as time goes by. A novel method is proposed for fault diagnosis of rolling bearing with online imbalanced cross-domain data. Finally, the proposed method which is tested on bearing dataset (CWRU) has achieved prediction accuracy of 95.89% with only 40 target samples. The results have been compared with other traditional methods. The comparisons show that the proposed online domain adaptation fault diagnosis method has achieved significant improvements. In addition, the deep transfer learning model by adaptive- network-based fuzzy inference system (ANFIS) is introduced to interpretation the results.Entities:
Keywords: ANFIS; domain adaptation; domain transfer; imbalanced cross-domain data
Mesh:
Year: 2022 PMID: 35746322 PMCID: PMC9228669 DOI: 10.3390/s22124540
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1CNN structure [13].
Figure 2ANFIS structure [17].
Figure 3Illustration of the proposed cross-domain fault diagnosis model structure.
Architecture of the proposed method.
| Layer Type | Parameters |
|---|---|
| 2D Convolutional Layer (1st) | Filter Number = 6 Filter Size = (9, 9) |
| Batch Normalization layer (1st) | Activation Function = Leaky ReLU |
| Max Pooling Layer (1st) | Filter Size = (2, 2) |
| 2D Convolutional Layer (2nd) | Filter Number = 12 Filter Size = (3, 3) |
| Batch Normalization layer (2nd) | Activation Function = Leaky ReLU |
| Max Pooling Layer (2nd) | Filter Size = (2, 2) |
| Flatten | - |
| Fully Connected Layer | 100 neurons |
| Softmax Output Layer | 10 neurons |
Figure 4Results of STFT.
Hyperparameters of the proposed method.
| Hyperparameters | Value |
|---|---|
| Epochs | 40 |
| Learning rate | 0.0001 |
| Optimizer | Adam |
| Batch size | 64 |
| Sample length | 500 |
| Source domain training sample | 2000 |
| Target domain training sample | 0~2000 |
| Target domain testing sample | 400 |
Results of the proposed method with imbalanced cross-domain data.
| Number of Source Domain Data | Number of Target Domain Data | Imbalanced Ratio | Average Accuracy (%) | Standard Deviation (10 Trials) |
|---|---|---|---|---|
| 2000 | 0 (without DA) | ∞ (without DA) | 74.54 | 4.590 |
| 2000 | 10 | 200 | 77.17 | 4.326 |
| 2000 | 20 | 100 | 82.61 | 4.595 |
| 2000 | 24 | 83.33 | 87.19 | 8.518 |
| 2000 | 26 | 76.92 | 91.79 | 2.994 |
| 2000 | 30 | 66.67 | 92.36 | 4.976 |
| 2000 | 40 | 50 | 95.89 | 1.895 |
| 2000 | 50 | 40 | 96.13 | 2.417 |
| 2000 | 60 | 33.33 | 97.71 | 2.194 |
| 2000 | 70 | 28.57 | 97.92 | 0.982 |
| 2000 | 80 | 25 | 97.19 | 0.718 |
| 2000 | 100 | 20 | 98.13 | 0.570 |
| 2000 | 120 | 16.67 | 98.56 | 0.464 |
| 2000 | 150 | 13.33 | 99.03 | 0.714 |
| 2000 | 200 | 10 | 99.06 | 0.561 |
| 2000 | 400 | 5 | 99.39 | 0.480 |
| 2000 | 500 | 4 | 99.44 | 0.644 |
| 2000 | 1000 | 2 | 99.71 | 0.325 |
| 2000 | 2000 | 1 | 99.78 | 0.245 |
Results of traditional method with imbalanced cross-domain data.
| Number of Source Domain Data | Number of Target Domain Data | Imbalanced Ratio | Average Accuracy (%) | Standard Deviation (10 Times) |
|---|---|---|---|---|
| 2000 | 0 (without DA) | ∞ (without DA) | 67.84 | 3.760 |
| 2000 | 10 | 200 | 70.38 | 7.043 |
| 2000 | 20 | 100 | 78.81 | 6.139 |
| 2000 | 30 | 83.33 | 85.03 | 8.919 |
| 2000 | 40 | 50 | 90.66 | 4.913 |
| 2000 | 50 | 40 | 90.09 | 2.353 |
| 2000 | 60 | 33.33 | 92.63 | 3.493 |
| 2000 | 70 | 28.57 | 95.84 | 0.970 |
| 2000 | 80 | 25 | 96.09 | 2.611 |
| 2000 | 100 | 20 | 96.91 | 2.124 |
| 2000 | 150 | 16.67 | 96.00 | 2.113 |
| 2000 | 200 | 13.33 | 97.19 | 1.304 |
| 2000 | 400 | 10 | 97.47 | 2.496 |
| 2000 | 800 | 5 | 98.93 | 0.768 |
| 2000 | 1000 | 4 | 99.36 | 0.274 |
| 2000 | 1200 | 2 | 99.39 | 0.375 |
| 2000 | 2000 | 1 | 99.58 | 0.274 |
Figure 5Comparison results of the proposed method (STFT + CNN) with FFT + NN.
Figure 6Accuracy of the two models with limited target domain data: blue line is the proposed method (STFT + 2D CNN + MMD loss) and orange line is the traditional method (FFT + NN + MMD loss).
Classification accuracy for six different quantity of target domain data.
| Source: 2 hp Target: 0 hp | ||||||
| Methods | The number of target domain training data | |||||
| 20 | 50 | 100 | 500 | 1000 | 2000 | |
| DADA | 87.42% | 88.08% | 88.17% | 88.58% | 88.92% | 91.25% |
| FFT + NN | 85.92% | 93.42% | 96.08% | 99.17% | 99.75% | 99.92% |
| FFT + CNN | 93.81% | 96.67% | 98.83% | 99.33% | 99.67% | 99.92% |
| STFT + CNN (Proposed) | 92.50% | 98.25% | 98.92% | 99.42% | 99.50% | 99.92% |
| Source: 0 hp Target: 2 hp | ||||||
| Methods | The number of target domain training data | |||||
| 20 | 50 | 100 | 500 | 1000 | 2000 | |
| DADA | 84.83% | 88.58% | 91.17% | 91.42% | 92.50% | 93.00% |
| FFT + NN | 82.17% | 92.00% | 96.50% | 99.75% | 99.67% | 99.83% |
| FFT + CNN | 90.58% | 97.33% | 98.42% | 99.73% | 99.92% | 99.92% |
| STFT + CNN (Proposed) | 95.75% | 97.42% | 99.17% | 99.83% | 99.83% | 100.00% |
| Source: 1 hp Target: 3 hp | ||||||
| Methods | The number of target domain training data | |||||
| 20 | 50 | 100 | 500 | 1000 | 2000 | |
| DADA | 95.75% | 95.92% | 96.67% | 97.33% | 97.42% | 97.75% |
| FFT + NN | 88.75% | 93.58% | 97.17% | 99.17% | 99.50% | 99.58% |
| FFT + CNN | 94.33% | 97.67% | 99.00% | 99.67% | 99.75% | 99.92% |
| STFT + CNN (Proposed) | 94.83% | 98.50% | 98.87% | 99.42% | 99.83% | 99.83% |
| Source: 3 hp Target: 1 hp | ||||||
| Methods | The number of target domain training data | |||||
| 20 | 50 | 100 | 500 | 1000 | 2000 | |
| DADA | 81.00% | 84.33% | 85.50% | 86.17% | 86.25% | 86.83% |
| FFT + NN | 80.67% | 89.00% | 88.92% | 97.33% | 99.50% | 99.67% |
| FFT + CNN | 87.67% | 96.25% | 97.00% | 99.08% | 99.58% | 99.67% |
| STFT + CNN (Proposed) | 89.92% | 97.00% | 97.17% | 98.92% | 99.58% | 99.75% |
Figure 7The structure of the proposed model combines with ANFIS.