| Literature DB >> 35270548 |
Shanguang Zhao1, Fangfang Long2, Xin Wei3, Xiaoli Ni3, Hui Wang4, Bokun Wei5.
Abstract
Sleep staging is the basis of sleep assessment and plays a crucial role in the early diagnosis and intervention of sleep disorders. Manual sleep staging by a specialist is time-consuming and is influenced by subjective factors. Moreover, some automatic sleep staging algorithms are complex and inaccurate. The paper proposes a single-channel EEG-based sleep staging method that provides reliable technical support for diagnosing sleep problems. In this study, 59 features were extracted from three aspects: time domain, frequency domain, and nonlinear indexes based on single-channel EEG data. Support vector machine, neural network, decision tree, and random forest classifier were used to classify sleep stages automatically. The results reveal that the random forest classifier has the best sleep staging performance among the four algorithms. The recognition rate of the Wake phase was the highest, at 92.13%, and that of the N1 phase was the lowest, at 73.46%, with an average accuracy of 83.61%. The embedded method was adopted for feature filtering. The results of sleep staging of the 11-dimensional features after filtering show that the random forest model achieved 83.51% staging accuracy under the condition of reduced feature dimensions, and the coincidence rate with the use of all features for sleep staging was 94.85%. Our study confirms the robustness of the random forest model in sleep staging, which also represents a high classification accuracy with appropriate classifier algorithms, even using single-channel EEG data. This study provides a new direction for the portability of clinical EEG monitoring.Entities:
Keywords: EEG; back propagation neural network; decision tree; random forest; sleep staging; support vector machine
Mesh:
Year: 2022 PMID: 35270548 PMCID: PMC8910622 DOI: 10.3390/ijerph19052845
Source DB: PubMed Journal: Int J Environ Res Public Health ISSN: 1660-4601 Impact factor: 3.390
Sample distribution by stage.
| Sleep Stages | Sample Number |
|---|---|
| W stage | 2029 |
| N1 stage | 2029 |
| N2 stage | 2029 |
| N3 stage | 1671 |
| REM stage | 1938 |
| Total | 9696 |
EEG sleep staging characteristics.
| Feature Symbol | Computational Method | Feature Symbol | Computational Method | Feature Symbol | Computational Method |
|---|---|---|---|---|---|
| T1 | Amplitude | F5 | E6 + E7 | F24 | (E2 + E3)/E4 |
| T2 | Mean Value | F6 | E8 | F25 | (E2 + E6)/E9 |
| T3 | Variance | F7 | E2/E1 | F26 | MPF |
| T4 | SD | F8 | E3/E1 | F27 | MPF-low- |
| T5 | Median | F9 | E4/E1 | F28 | MPF-high- |
| T6 | Skewness | F10 | E5/E1 | F29 | MPF- |
| T7 | Kurtosis | F11 | (E6 + E7)/E1 | F30 | MPF- |
| T8 | Maximum | F12 | E8/E1 | F31 | MPF- |
| T9 | Minimum | F13 | (E4 + E5)/E1 | F32 | MPF- |
| T10 | ZCR | F14 | E5/(E6 + E7) | F33 | FV |
| T11 | AFDN | F15 | (E4 + E5)/(E5 + E6 + E7) | F34 | FV-low- |
| T12 | ASDN | F16 | E4/(E6 + E7) | F35 | FV-high- |
| T13 | Activity | F17 | E3/(E4 + E5) | F36 | FV- |
| T14 | Mobility | F18 | E4/(E3 + E5) | F37 | FV- |
| T15 | Complexity | F19 | E5/(E3 + E4) | F38 | FV- |
| F1 | E1 | F20 | E2/(E3 + E9) | F39 | FV- |
| F2 | E2 + E3 | F21 | E5/E9 | N1 | FD |
| F3 | E4 | F22 | (E6 + E7)/E9 | N2 | NSI |
| F4 | E5 | F23 | E5/E4 | N3 | E |
T, time domain features; F, frequency domain features; N, non-stationary features. FD, fractal dimension; SE, sample entropy; ZCR, zero crossing rate; SD, standard deviation; E1, the total band power; E2, the low-frequency δ-band (0.5–2 Hz) power; E3, the high-frequency δ-band (1.2–4 Hz) power; E4, the θ-band (4–8 Hz) power; E5, the α-band (8–13 Hz) power; E6, the low β-band (13–20 Hz) power; E7, the high-frequency β-band (20–30 Hz) power; E8, the low-frequency γ-band (30–45 Hz) power; E9, the δ + θ + α + β + γ + δ band power.
Weight coefficients of EEG sleep staging features.
| Feature Symbol | Weight Coefficient | Feature Symbol | Weight Coefficient | Feature Symbol | Weight Coefficient |
|---|---|---|---|---|---|
| T1 | 0.0030 | F5 | 0.0394 | F24 | 0.0012 |
| T2 | 0.0006 | F6 | 0.1459 | F25 | 0.0053 |
| T3 | 0.0051 | F7 | 0.0038 | F26 | 0.0011 |
| T4 | 0.0122 | F8 | 0.0281 | F27 | 0.0028 |
| T5 | 0.0111 | F9 | 0.0457 | F28 | 0.0007 |
| T6 | 0.0513 | F10 | 0.0038 | F29 | 0.0011 |
| T7 | 0.0201 | F11 | 0.0106 | F30 | 0.0013 |
| T8 | 0.0015 | F12 | 0.2043 | F31 | 0.0007 |
| T9 | 0.0066 | F13 | 0.0024 | F32 | 0.0006 |
| T10 | 0.0086 | F14 | 0.0142 | F33 | 0.0007 |
| T11 | 0.0013 | F15 | 0.0124 | F34 | 0.0037 |
| T12 | 0.0049 | F16 | 0.0110 | F35 | 0.0014 |
| T13 | 0.0101 | F17 | 0.0105 | F36 | 0.0006 |
| T14 | 0.0048 | F18 | 0.0104 | F37 | 0.0010 |
| T15 | 0.0050 | F19 | 0.0540 | F38 | 0.0005 |
| F1 | 0.0148 | F20 | 0.0008 | F39 | 0.0006 |
| F2 | 0.1049 | F21 | 0.0031 | N1 | 0.0045 |
| F3 | 0.0056 | F22 | 0.0301 | N2 | 0.0470 |
| F4 | 0.0125 | F23 | 0.0077 | N3 | 0.0032 |
Comparison of SVM model staging results for all features with expert manual staging results.
| Sleep Stages | Training Samples | Test Samples | Correct Samples | Precision | Recall | f1-Score |
|---|---|---|---|---|---|---|
| N3 | 1326 | 345 | 321 | 0.8992 | 0.9304 | 0.9145 |
| N2 | 1601 | 428 | 336 | 0.8276 | 0.7850 | 0.8058 |
| N1 | 1639 | 390 | 255 | 0.7143 | 0.6538 | 0.6827 |
| REM | 1547 | 391 | 316 | 0.7215 | 0.8082 | 0.7624 |
| W | 1643 | 386 | 360 | 0.9424 | 0.9326 | 0.9375 |
Accuracy = 81.86%.
Figure 1Confusion matrix of SVM model staging results.
Figure 2Comparison of expert manual staging results with SVM model staging results. The real label represents the result of manual staging by experts, while the prediction label is the result of the SVM model.
Comparison of BP model staging results for all features with expert manual staging results.
| Sleep Stages | Training Samples | Test Samples | Correct Samples | Precision | Recall | f1-Score |
|---|---|---|---|---|---|---|
| N3 | 1326 | 345 | 321 | 0.7685 | 0.9623 | 0.8546 |
| N2 | 1601 | 428 | 336 | 0.8397 | 0.6729 | 0.7471 |
| N1 | 1639 | 390 | 255 | 0.6555 | 0.6538 | 0.6547 |
| REM | 1547 | 391 | 316 | 0.7521 | 0.6982 | 0.7241 |
| W | 1643 | 386 | 360 | 0.9007 | 0.9637 | 0.9312 |
Accuracy = 78.35%.
Figure 3Confusion matrix of BPNN model staging results.
Figure 4Comparison of expert manual staging results and BPNN model staging results. The real label represents the result of manual staging by experts, while the prediction label is the result of the BPNN model.
Comparison of decision tree model staging results for all features with expert manual staging results.
| Sleep Stages | Training Samples | Test Samples | Correct Samples | Precision | Recall | f1-Score |
|---|---|---|---|---|---|---|
| N3 | 1326 | 345 | 321 | 0.8773 | 0.8609 | 0.8787 |
| N2 | 1601 | 428 | 336 | 0.7301 | 0.7079 | 0.7189 |
| N1 | 1639 | 390 | 255 | 0.6247 | 0.6103 | 0.6174 |
| REM | 1547 | 391 | 316 | 0.6804 | 0.7187 | 0.6990 |
| W | 1643 | 386 | 360 | 0.8800 | 0.9119 | 0.8957 |
Accuracy = 75.82%.
Figure 5Confusion matrix of DT model staging results.
Figure 6Comparison of expert manual staging results with DT model staging results. The real label represents the result of manual staging by experts, while the prediction label is the result of the DT model.
Comparison of random forest model staging results for all features with expert manual staging results.
| Sleep Stages | Training Samples | Test Samples | Correct Samples | Precision | Recall | f1-Score |
|---|---|---|---|---|---|---|
| N3 | 1326 | 345 | 321 | 0.9171 | 0.9304 | 0.9237 |
| N2 | 1601 | 428 | 336 | 0.8199 | 0.8294 | 0.8246 |
| N1 | 1639 | 390 | 255 | 0.7346 | 0.7032 | 0.7032 |
| REM | 1547 | 391 | 316 | 0.7877 | 0.8015 | 0.8015 |
| W | 1643 | 386 | 360 | 0.9213 | 0.9404 | 0.9308 |
Accuracy = 83.56%.
Figure 7Confusion matrix of RF model staging results.
Figure 8Comparison of expert manual staging results with RF model staging results. The real label represents the result of manual staging by experts, while the prediction label is the result of the RF model.
Comparison of sleep staging accuracy of different models before and after feature screening (%).
| Sleep Stages | SVM | BP | DT | RF | ||||
|---|---|---|---|---|---|---|---|---|
| 57 | 11 | 57 | 11 | 57 | 11 | 57 | 11 | |
| N3 | 89.92 | 88.67 | 76.85 | 89.89 | 89.73 | 88.48 | 91.71 | 92.11 |
| N2 | 82.76 | 76.58 | 83.97 | 80.05 | 73.01 | 70.91 | 81.99 | 80.61 |
| N1 | 71.43 | 70.50 | 65.55 | 63.27 | 62.47 | 61.42 | 73.46 | 72.85 |
| REM | 72.15 | 64.44 | 75.21 | 72.95 | 68.04 | 69.57 | 78.77 | 76.09 |
| W | 94.24 | 91.75 | 90.07 | 95.79 | 88.00 | 90.84 | 92.13 | 91.90 |
| Total accuracy | 81.86 | 77.99 | 78.35 | 79.59 | 75.82 | 75.88 | 83.56 | 82.53 |
The accuracy of sleep staging using single-channel EEG information.
| Author/Year | Sleep Stages | Number of Features | Classifier | Channel of EEG | ACC (%) or KC |
|---|---|---|---|---|---|
| [ | Five stages | 39 | SVM | C3-A2 | ACC = 85.7% |
| [ | Five stages | Multiscale entropy and autoregressive models | LDA | C3-A2 | KC = 0.81 |
| [ | Five stages | 9 | SVM | Pz-Oz | ACC = 87.5% KC = 0.81 |
| [ | Five stages | The IMFs factor was set to 7 to obtain the optimal number of features | LDA, BPNN, SVM, k-NN, LS-SVM, Bagging, AdaBoost and Naïve Bayes | Pz–Oz | ACC (44.80–88.62%), AdaBoost algorithm has the highest accuracy of 88.62% |
| [ | Five stages | 10 | K-NN, DT, RF, Multilayer perceptron and Naïve Bayes | Fpz-Cz (highest ACC), Cz-A1, C3-A2, Pz-Cz | ACC (71.80–89.74%) |
| [ | Five stages | 136 | RF | Fpz-Cz (highest ACC), Cz-A1, C3-A2, Pz-Cz | ACC = 87.82% |
| The present study | Five stages | 57 and 11 (Embedded method feature optimization) | SVM, DT, RF and BPNN | Fpz-Cz | The ACC of 57 features (75.82–83.56%), RF with the highest |
Note: KC, kappa coefficient; ACC, accuracy; SVM, support vector machine; LS-SVM, Least Squares-support vector machine; DT, decision trees; RF, random forest; LDA, linear discriminant analysis; BPNN, backpropagation neural network; k-NN, k-nearest neighbor.