| Literature DB >> 28830438 |
Shirin Najdi1,2, Ali Abdollahi Gharbali3,4, José Manuel Fonseca3,4.
Abstract
BACKGROUND: Nowadays, sleep quality is one of the most important measures of healthy life, especially considering the huge number of sleep-related disorders. Identifying sleep stages using polysomnographic (PSG) signals is the traditional way of assessing sleep quality. However, the manual process of sleep stage classification is time-consuming, subjective and costly. Therefore, in order to improve the accuracy and efficiency of the sleep stage classification, researchers have been trying to develop automatic classification algorithms. Automatic sleep stage classification mainly consists of three steps: pre-processing, feature extraction and classification. Since classification accuracy is deeply affected by the extracted features, a poor feature vector will adversely affect the classifier and eventually lead to low classification accuracy. Therefore, special attention should be given to the feature extraction and selection process.Entities:
Keywords: Biomedical signal processing; Feature ranking; Feature selection; Neural network; Polysomnography; Rank aggregation; Sleep stage classification; k-NN
Mesh:
Year: 2017 PMID: 28830438 PMCID: PMC5568624 DOI: 10.1186/s12938-017-0358-3
Source DB: PubMed Journal: Biomed Eng Online ISSN: 1475-925X Impact factor: 2.819
Fig. 1Block diagram for comparison of feature selection methods
Summary of the data provided by six selected subjects
| Wake | REM | S1 | S2 | SWS | |
|---|---|---|---|---|---|
| Subject #1 | 146 | 122 | 101 | 527 | 136 |
| Subject #2 | 41 | 159 | 71 | 351 | 284 |
| Subject #3 | 85 | 226 | 120 | 392 | 180 |
| Subject #4 | 40 | 143 | 47 | 266 | 152 |
| Subject #5 | 149 | 80 | 102 | 428 | 218 |
| Subject #6 | 131 | 142 | 135 | 378 | 198 |
Summary of the features extracted from PSG recordings
| Signal | Category | Feature name |
|---|---|---|
| EEG | Time domain (F1 to F12) | Statistical features (minimum value, maximum value, arithmetic mean, standard deviation, variance, skewness, kurtosis, median), zero-crossing rate, Hjorth parameters (activity, mobility and complexity) [ |
| Time–frequency domain (F13 to F26) | Features extracted from wavelet packet coefficients including energy of | |
| Entropy (F27 to F30) | Spectral entropy, Rényi entropy, approximate entropy, permutation entropy [ | |
| Non-linear (F31 to F36) | Petrosian fractal dimension, teager energy, energy, mean curve length, hurst exponent [ | |
| EOG | Time domain (F37 to F41) | Mean, maximum, standard deviation, skewness, kurtosis [ |
| Non-linear (F42) | Energy [ | |
| EMG | Frequency domain (F43 to F46) | Total power in the EMG frequency spectrum, statistical features of EMG frequency spectrum (maximum, mean, standard deviation) [ |
| Non-linear (F47 to F49) | Energy, ratio of the EMG Signal energy for the current epoch and previous epoch, ratio of the EMG signal energy for the current epoch and next epoch [ |
Hjorth parameters
| Feature name | Formula |
|---|---|
| Hjorth activity |
|
| Hjorth mobility |
|
| Hjorth Complexity |
|
Fig. 2Stability measure of each feature selection method
Mean stability for selected features
| ReliefF | Fisher | Chi2 | IG | CMIM | MRMR-MID | MRMR-MIQ | Borda | RRA | |
|---|---|---|---|---|---|---|---|---|---|
| Mean stability up to 5th feature | 0.50 | 0.80 | 0.79 | 0.73 | 0.20 | 0.72 |
| 0.39 | 0.65 |
| Mean stability up to 13th feature | 0.66 |
| 0.95 | 0.92 | 0.21 | 0.79 | 0.82 | 0.68 | 0.78 |
| Mean stability up to 29th feature | 0.69 | 0.86 | 0.86 |
| 0.24 | 0.75 | 0.77 | 0.70 | 0.70 |
Italic values indicate the maximum of each row
Similarity of the feature selection techniques
| ReliefF | Fisher | Chi2 | IG | CMIM | MRMR-MID | MRMR-MIQ | Borda | RRA | |
|---|---|---|---|---|---|---|---|---|---|
| ReliefF | 1 | 0.26 | 0.18 | 0.18 | 0.35 | 0.40 | 0.40 | 0.31 | 0.31 |
| Fisher | 1 | 0.58 | 0.52 |
| 0.58 | 0.65 | 0.72 | 0.65 | |
| Chi2 | 1 |
| 0.15 | 0.35 | 0.35 | 0.52 | 0.52 | ||
| IG | 1 | 0.18 | 0.35 | 0.35 | 0.46 | 0.46 | |||
| CMIM | 1 | 0.22 | 0.22 | 0.22 | 0.22 | ||||
| MRMR-MID | 1 |
| 0.72 | 0.65 | |||||
| MRMR-MIQ | 1 | 0.72 | 0.65 | ||||||
| Borda | 1 | 0.72 | |||||||
| RRA | 1 |
Italic values indicate the maximum and minimum similarity
Fig. 3Classification accuracy for different feature selection methods. a k-NN classifier, b MLFN classifier
Top 10 features selected by each method and the optimum number selected by Kneedle algorithm (corresponding accuracy)
| ReliefF | Fisher | Chi2 | IG | CMIM | MRMR-MID | MRMR-MIQ | Borda | RRA | |
|---|---|---|---|---|---|---|---|---|---|
| Top 10 features | F28 |
| F35 | F9 | F15 | F35 | F35 |
|
|
|
| F35 | F9 | F35 |
| F39 | F42 | F35 | F35 | |
| F7 | F31 | F11 | F11 | F9 |
| F15 | F9 | F9 | |
| F49 | F9 | F31 | F31 | F8 | F22 |
| F31 | F31 | |
| F41 | F29 |
|
| F1 | F15 | F22 | F22 | F27 | |
| F27 | F11 | F27 | F4 | F34 | F31 | F23 | F27 | F22 | |
| F20 | F25 | F26 | F27 | F35 | F29 | F31 | F29 | F17 | |
| F23 | F27 | F4 | F26 | F28 | F23 | F38 | F11 | F29 | |
| F6 | F12 | F25 | F25 | F6 | F9 | F29 | F15 | F11 | |
| F22 | F22 | F14 | F29 | F48 | F38 | F9 | F20 | F20 | |
| MLFN | 7 (0.75) | 5 (0.76) | 7 (0.76) | 7 (0.76) |
| 5 (0.76) | 5 (0.76) | 5 (0.76) | 7 (0.77) |
|
| 7 (0.69) | 5 (0.71) | 9 (0.73) | 9 (0.73) |
| 7 (0.75) | 11 (0.75) | 9 (0.74) | 7 (0.73) |
Italic value indicates ISD feature