| Literature DB >> 29261689 |
Kar Hoou Hui1, Ching Sheng Ooi1, Meng Hee Lim1, Mohd Salman Leong1, Salah Mahdi Al-Obaidi1.
Abstract
A major issue of machinery fault diagnosis using vibration signals is that it is over-reliant on personnel knowledge and experience in interpreting the signal. Thus, machine learning has been adapted for machinery fault diagnosis. The quantity and quality of the input features, however, influence the fault classification performance. Feature selection plays a vital role in selecting the most representative feature subset for the machine learning algorithm. In contrast, the trade-off relationship between capability when selecting the best feature subset and computational effort is inevitable in the wrapper-based feature selection (WFS) method. This paper proposes an improved WFS technique before integration with a support vector machine (SVM) model classifier as a complete fault diagnosis system for a rolling element bearing case study. The bearing vibration dataset made available by the Case Western Reserve University Bearing Data Centre was executed using the proposed WFS and its performance has been analysed and discussed. The results reveal that the proposed WFS secures the best feature subset with a lower computational effort by eliminating the redundancy of re-evaluation. The proposed WFS has therefore been found to be capable and efficient to carry out feature selection tasks.Entities:
Mesh:
Year: 2017 PMID: 29261689 PMCID: PMC5738058 DOI: 10.1371/journal.pone.0189143
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Number of combinations based on the number of features extracted.
| Number of features | Number of combinations |
|---|---|
| 3 | 7 |
| 6 | 63 |
| 12 | 4,095 |
| 24 | 16,777,215 |
| 48 | 281,474,976,710,655 |
Fig 1Experimental test rig.
Vibration data distribution.
| Bearing condition | Training data | Testing data |
|---|---|---|
| Healthy | 50 | 50 |
| Rolling element fault | 50 | 50 |
| Inner raceway fault | 50 | 50 |
| Outer raceway fault | 50 | 50 |
Statistical features.
| No. | Statistical Feature | Equation |
|---|---|---|
| A | Skewness factor | |
| B | Kurtosis factor | |
| C | Crest factor | |
| D | Shape factor | |
| E | Impulse factor | |
| F | Margin factor |
Fig 2(a) Skewness factor, (b) kurtosis factor, (c) crest factor, (d) shape factor, (e) impulse factor and (f) margin factor of all bearing conditions.
Fig 3The proposed feature selection algorithm (features A, B, C, D, E and F represent skewness factor, kurtosis factor, crest factor, shape factor, impulse factor and margin factor, respectively).
Training accuracy for the key combination of features (features A, B, C, D, E and F represent skewness factor, kurtosis factor, crest factor, shape factor, impulse factor and margin factor, respectively).
| Level 1 | Level 2 | Level 3 | Level 4 | Level 5 | Level 6 | ||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Feature | Accuracy | Feature | Accuracy | Feature | Accuracy | Feature | Accuracy | Feature | Accuracy | Feature | Accuracy |
| A | 28.5% | D,A | 81.0% | D,A,B | 73.0% | D,A,C,B | 73.5% | D,A,C,B,E | 73.5% | D,A,C,B,E,F | 74.0% |
| B | 40.5% | D,B | 50.0% | D,A,C | 73.5% | D,A,C,E | 73.5% | D,A,C,B,F | 73.5% | ||
| C | 2.5% | D,C | 50.0% | D,A,E | 72.5% | D,A,C,F | 73.5% | D,A,C,E,F | 73.5% | ||
| D | 50.0% | D,E | 50.0% | D,A,F | 73.5% | D,A,F,B | 72.0% | ||||
| E | 23.0% | D,F | 50.0% | D,A,F,E | 73.0% | ||||||
| F | 34.0% | ||||||||||
Cyclical assessment for the proposed WFS by 10-fold cross-validation.
| Cycle | Number of Feature Dimension | |||||
|---|---|---|---|---|---|---|
| 1 | 2 | 3 | 4 | 5 | 6 | |
| 1 | 0.500 | 0.850 | 0.880 | 0.815 | 0.845 | 0.850 |
| 2 | 0.500 | 0.870 | 0.880 | 0.835 | 0.880 | 0.860 |
| 3 | 0.500 | 0.840 | 0.870 | 0.875 | 0.875 | 0.865 |
| 4 | 0.500 | 0.825 | 0.845 | 0.860 | 0.805 | 0.805 |
| 5 | 0.500 | 0.885 | 0.900 | 0.865 | 0.745 | 0.850 |
| 6 | 0.500 | 0.845 | 0.845 | 0.845 | 0.860 | 0.845 |
| 7 | 0.500 | 0.860 | 0.840 | 0.885 | 0.870 | 0.870 |
| 8 | 0.500 | 0.860 | 0.865 | 0.875 | 0.860 | 0.825 |
| 9 | 0.500 | 0.880 | 0.860 | 0.820 | 0.850 | 0.835 |
| 10 | 0.500 | 0.860 | 0.880 | 0.865 | 0.865 | 0.800 |
| Mean | 0.500 | 0.858 | 0.867 | 0.854 | 0.846 | 0.841 |
| ± SD | ± 0 | ± 0.018 | ± 0.019 | ± 0.024 | ± 0.041 | ± 0.024 |
Cyclical assessment for the MRMD by 10-fold cross-validation.
| Cycle | Number of Feature Dimension | |||||
|---|---|---|---|---|---|---|
| 1 | 2 | 3 | 4 | 5 | 6 | |
| 1 | 0.500 | 0.500 | 0.505 | 0.500 | 0.500 | 0.850 |
| 2 | 0.500 | 0.500 | 0.515 | 0.550 | 0.500 | 0.860 |
| 3 | 0.500 | 0.500 | 0.500 | 0.520 | 0.500 | 0.865 |
| 4 | 0.500 | 0.500 | 0.500 | 0.500 | 0.500 | 0.805 |
| 5 | 0.500 | 0.500 | 0.500 | 0.500 | 0.520 | 0.850 |
| 6 | 0.500 | 0.500 | 0.500 | 0.500 | 0.525 | 0.845 |
| 7 | 0.500 | 0.505 | 0.535 | 0.500 | 0.570 | 0.870 |
| 8 | 0.500 | 0.510 | 0.520 | 0.500 | 0.500 | 0.825 |
| 9 | 0.500 | 0.500 | 0.545 | 0.500 | 0.550 | 0.835 |
| 10 | 0.500 | 0.500 | 0.500 | 0.530 | 0.510 | 0.800 |
| Mean | 0.500 | 0.502 | 0.512 | 0.510 | 0.518 | 0.841 |
| ± SD | ± 0 | ± 0.003 | ± 0.017 | ± 0.018 | ± 0.025 | ± 0.024 |
Fig 4Comparison of the testing accuracy (average of 10-fold cross-validation).