| Literature DB >> 35591246 |
Sunil Kumar Prabhakar1, Harikumar Rajaguru2, Semin Ryu1, In Cheol Jeong1, Dong-Ok Won1.
Abstract
Manual sleep stage scoring is usually implemented with the help of sleep specialists by means of visual inspection of the neurophysiological signals of the patient. As it is a very hectic task to perform, automated sleep stage classification systems were developed in the past, and advancements are being made consistently by researchers. The various stages of sleep are identified by these automated sleep stage classification systems, and it is quite an important step to assist doctors for the diagnosis of sleep-related disorders. In this work, a holistic strategy named as clustering and dimensionality reduction with feature extraction cum selection for classification along with deep learning (CDFCD) is proposed for the classification of sleep stages with EEG signals. Though the methodology follows a similar structural flow as proposed in the past works, many advanced and novel techniques are proposed under each category in this work flow. Initially, clustering is applied with the help of hierarchical clustering, spectral clustering, and the proposed principal component analysis (PCA)-based subspace clustering. Then the dimensionality of it is reduced with the help of the proposed singular value decomposition (SVD)-based spectral algorithm and the standard variational Bayesian matrix factorization (VBMF) technique. Then the features are extracted and selected with the two novel proposed techniques, such as the sparse group lasso technique with dual-level implementation (SGL-DLI) and the ridge regression technique with limiting weight scheme (RR-LWS). Finally, the classification happens with the less explored multiclass Gaussian process classification (MGC), the proposed random arbitrary collective classification (RACC), and the deep learning technique using long short-term memory (LSTM) along with other conventional machine learning techniques. This methodology is validated on the sleep EDF database, and the results obtained with this methodology have surpassed the results of the previous studies in terms of the obtained classification accuracy reporting a high accuracy of 93.51% even for the six-classes classification problem.Entities:
Keywords: classification; clustering; dimensionality reduction; feature extraction; selection
Mesh:
Year: 2022 PMID: 35591246 PMCID: PMC9103466 DOI: 10.3390/s22093557
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1Pictorial representation of the work.
Figure 2An LSTM representation.
Figure 3Illustration of the LSTM implementation.
Total number of samples in sleep-EDF dataset (R and K criteria).
| Number of Classes | AWA | REM | S1 | S2 | S3 | S4 |
|---|---|---|---|---|---|---|
| 6 | 74,827 | 11,848 | 4848 | 27,292 | 5070 | 3773 |
| 5 | 74,827 | 11,848 | 4848 | 27,292 | 8843 | |
| 4 | 74,827 | 11,848 | 32,140 | 8843 | ||
| 3 | 74,827 | 11,848 | 40,983 | |||
| 2 | 74,827 | 52,831 | ||||
Hierarchical clustering with SVD-based spectral algorithm.
| SGL-DLI | RR-LWS | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | |
| LDA | 93.24 | 92.34 | 90.62 | 88.24 | 86.34 | 92.67 | 93.67 | 92.45 | 89.67 | 88.35 |
| KNN | 91.32 | 91.26 | 90.67 | 89.32 | 88.36 | 92.32 | 90.83 | 91.78 | 91.13 | 86.16 |
| NBC | 89.92 | 88.11 | 87.13 | 85.92 | 84.65 | 88.98 | 87.56 | 89.43 | 87.84 | 86.56 |
| DT | 89.27 | 88.69 | 87.57 | 86.27 | 83.68 | 88.27 | 88.31 | 88.21 | 87.79 | 81.26 |
| RF | 87.57 | 85.73 | 84.29 | 82.57 | 81.82 | 86.76 | 84.69 | 85.44 | 84.54 | 84.68 |
| Adaboost | 88.32 | 85.25 | 83.41 | 83.32 | 82.78 | 89.15 | 86.45 | 84.67 | 83.63 | 83.16 |
| SVM | 96.57 | 93.22 | 92.62 | 91.57 | 90.91 | 95.93 | 94.32 | 94.86 | 92.85 | 90.83 |
| MGC | 97.73 | 94.43 | 93.73 | 92.73 | 92.16 | 96.68 | 92.84 | 92.31 | 92.67 | 91.45 |
| RACC | 97.96 | 94.56 | 92.99 | 92.96 | 92.72 | 97.55 | 91.78 | 93.56 | 91.34 | 92.12 |
Spectral clustering with SVD-based spectral algorithm.
| SGL-DLI | RR-LWS | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | |
| LDA | 92.46 | 90.21 | 88.67 | 89.45 | 85.09 | 94.08 | 92.22 | 91.01 | 87.11 | 86.22 |
| KNN | 93.78 | 89.32 | 91.53 | 91.21 | 89.86 | 93.87 | 88.81 | 89.92 | 86.58 | 84.79 |
| NBC | 87.45 | 87.91 | 85.59 | 87.48 | 87.82 | 87.61 | 86.39 | 88.61 | 85.31 | 83.53 |
| DT | 87.32 | 89.72 | 89.87 | 88.98 | 8.34 | 88.25 | 87.41 | 87.78 | 88.69 | 80.16 |
| RF | 86.12 | 86.78 | 87.51 | 84.32 | 84.57 | 87.68 | 85.37 | 86.92 | 82.83 | 80.81 |
| Adaboost | 87.35 | 85.16 | 85.34 | 85.57 | 83.81 | 88.93 | 87.28 | 83.53 | 81.47 | 81.21 |
| SVM | 95.69 | 94.83 | 91.69 | 92.89 | 89.24 | 93.26 | 91.61 | 92.80 | 90.84 | 89.34 |
| MGC | 95.87 | 92.56 | 93.81 | 93.07 | 93.51 | 95.74 | 90.80 | 90.01 | 89.12 | 88.75 |
| RACC | 94.34 | 91.34 | 93.24 | 90.19 | 91.20 | 95.68 | 91.01 | 90.35 | 88.38 | 86.49 |
Subspace clustering with SVD-based spectral algorithm.
| SGL-DLI | RR-LWS | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | |
| LDA | 93.34 | 91.01 | 90.11 | 89.09 | 88.21 | 95.11 | 94.21 | 92.34 | 91.11 | 89.89 |
| KNN | 95.57 | 92.43 | 91.24 | 90.01 | 87.65 | 96.87 | 95.58 | 92.52 | 91.36 | 88.36 |
| NBC | 92.97 | 91.99 | 91.67 | 88.81 | 86.79 | 93.62 | 92.97 | 91.67 | 90.87 | 89.72 |
| DT | 91.42 | 90.51 | 90.84 | 89.61 | 88.81 | 94.41 | 94.41 | 93.94 | 92.42 | 86.38 |
| RF | 92.56 | 92.12 | 91.57 | 88.54 | 88.27 | 94.64 | 93.46 | 92.28 | 92.19 | 88.67 |
| Adaboost | 94.89 | 93.95 | 92.82 | 89.34 | 88.75 | 93.98 | 93.87 | 92.41 | 91.82 | 89.93 |
| SVM | 96.42 | 95.58 | 95.43 | 92.72 | 90.43 | 94.73 | 92.54 | 91.86 | 91.34 | 88.67 |
| MGC | 97.57 | 96.64 | 96.61 | 91.76 | 90.49 | 92.26 | 92.25 | 91.47 | 91.07 | 89.23 |
| RACC | 98.41 | 97.56 | 97.21 | 94.78 | 93.12 | 96.78 | 95.97 | 94.32 | 93.89 | 89.11 |
Hierarchical clustering with VBMF.
| SGL-DLI | RR-LWS | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | |
| LDA | 92.01 | 90.04 | 88.66 | 82.46 | 81.44 | 90.52 | 89.44 | 90.17 | 87.12 | 86.48 |
| KNN | 90.97 | 90.29 | 89.67 | 87.32 | 83.59 | 91.15 | 90.24 | 90.43 | 88.46 | 84.23 |
| NBC | 87.53 | 88.11 | 86.42 | 85.89 | 83.86 | 87.05 | 86.56 | 88.92 | 85.87 | 85.73 |
| DT | 87.73 | 86.75 | 86.31 | 83.26 | 83.31 | 86.53 | 85.14 | 87.36 | 85.63 | 80.56 |
| RF | 85.15 | 84.58 | 82.57 | 84.51 | 84.13 | 84.78 | 83.84 | 84.82 | 82.22 | 82.98 |
| Adaboost | 89.68 | 86.32 | 81.86 | 81.87 | 85.56 | 88.15 | 85.14 | 82.57 | 81.59 | 80.78 |
| SVM | 94.75 | 92.87 | 90.32 | 89.56 | 88.98 | 94.84 | 92.62 | 92.23 | 90.98 | 89.56 |
| MGC | 95.23 | 91.35 | 91.45 | 89.31 | 90.34 | 95.35 | 90.29 | 90.01 | 90.03 | 90.01 |
| RACC | 95.11 | 92.22 | 90.89 | 89.21 | 92.24 | 96.11 | 90.83 | 91.09 | 89.01 | 90.12 |
Spectral clustering with VBMF.
| SGL-DLI | RR-LWS | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | |
| LDA | 90.09 | 88.87 | 86.11 | 87.02 | 83.22 | 92.25 | 91.04 | 91.01 | 85.12 | 81.47 |
| KNN | 91.03 | 87.31 | 88.36 | 89.49 | 87.45 | 91.04 | 89.56 | 89.24 | 85.45 | 82.32 |
| NBC | 86.23 | 85.65 | 84.89 | 83.23 | 85.66 | 85.09 | 85.78 | 8431 | 83.78 | 81.65 |
| DT | 85.56 | 86.98 | 85.03 | 85.98 | 83.78 | 85.30 | 84.92 | 84.89 | 81.74 | 79.87 |
| RF | 84.74 | 83.23 | 86.03 | 82.28 | 82.98 | 84.12 | 83.34 | 83.67 | 80.52 | 80.47 |
| Adaboost | 84.13 | 81.87 | 80.56 | 83.51 | 81.92 | 85.51 | 84.78 | 83.05 | 80.14 | 80.23 |
| SVM | 94.87 | 91.65 | 90.87 | 90.78 | 86.34 | 91.67 | 90.94 | 90.36 | 87.67 | 85.11 |
| MGC | 93.66 | 90.23 | 90.15 | 90.34 | 89.65 | 92.98 | 90.55 | 90.02 | 88.98 | 87.67 |
| RACC | 92.54 | 89.11 | 88.77 | 88.02 | 87.18 | 93.09 | 89.67 | 88.11 | 86.71 | 85.45 |
Subspace clustering with VBMF.
| SGL-DLI | RR-LWS | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | |
| LDA | 92.46 | 90.23 | 90.14 | 88.11 | 87.57 | 94.46 | 93.54 | 91.98 | 90.67 | 88.11 |
| KNN | 94.86 | 91.67 | 90.78 | 89.24 | 87.14 | 95.98 | 94.81 | 91.76 | 90.33 | 87.23 |
| NBC | 91.45 | 90.09 | 90.05 | 88.67 | 87.52 | 92.72 | 91.79 | 90.24 | 90.21 | 87.56 |
| DT | 90.02 | 89.35 | 88.41 | 88.89 | 86.89 | 93.14 | 93.46 | 92.57 | 91.98 | 85.78 |
| RF | 91.64 | 91.78 | 9059 | 87.43 | 86.57 | 93.69 | 92.93 | 91.89 | 91.56 | 87.33 |
| Adaboost | 93.05 | 92.06 | 90.23 | 88.60 | 87.23 | 92.03 | 92.57 | 91.35 | 90.12 | 88.24 |
| SVM | 95.89 | 94.81 | 93.01 | 90.02 | 89.95 | 93.26 | 91.34 | 90.25 | 90.75 | 87.78 |
| MGC | 96.45 | 95.26 | 94.08 | 90.45 | 90.08 | 92.89 | 91.02 | 90.74 | 90.56 | 88.42 |
| RACC | 97.33 | 96.98 | 95.85 | 93.84 | 92.03 | 95.01 | 94.05 | 93.13 | 92.43 | 89.85 |
Results of the clustering methodology with deep learning LSTM.
| 2 Classes | 3 Classes | 4 Classes | 5 Classes | 6 Classes | |
|---|---|---|---|---|---|
| Hierarchical clustering | 96.35 | 95.11 | 94.71 | 90.65 | 90.42 |
| Spectral clustering | 97.85 | 95.99 | 95.37 | 93.35 | 92.31 |
| Subspace clustering | 97.38 | 96.78 | 96.42 | 93.22 | 92.47 |
Comparison with previous works for two channels (Pz-Oz and Fpz-Cz).
| Reference | Methodology | Number of Classes | Accuracy (%) |
|---|---|---|---|
| [ | High-dimensional FFT features with SVM classifier | 2 | 97.88 |
| 3 | 94.41 | ||
| 4 | 92.82 | ||
| 5 | 91.73 | ||
| 6 | 90.77 | ||
| Proposed method | Hierarchical clustering with SVD-based spectral algorithm | 2 | 97.96 |
| 3 | 94.56 | ||
| 4 | 93.73 | ||
| 5 | 92.96 | ||
| 6 | 92.72 | ||
| Proposed method | Spectral clustering with SVD-based spectral algorithm | 2 | 95.87 |
| 3 | 92.56 | ||
| 4 | 93.81 | ||
| 5 | 93.07 | ||
| 6 | 93.51 | ||
| Proposed method | Subspace clustering with SVD-based spectral algorithm | 2 | 98.41 |
| 3 | 97.56 | ||
| 4 | 96.61 | ||
| 5 | 94.78 | ||
| 6 | 93.12 | ||
| Proposed method | Hierarchical clustering with VBMF | 2 | 96.11 |
| 3 | 92.22 | ||
| 4 | 91.45 | ||
| 5 | 90.03 | ||
| 6 | 92.24 | ||
| Proposed method | Spectral clustering with VBMF | 2 | 93.66 |
| 3 | 90.55 | ||
| 4 | 90.15 | ||
| 5 | 90.34 | ||
| 6 | 89.65 | ||
| Proposed method | Subspace clustering with VBMF | 2 | 97.33 |
| 3 | 96.98 | ||
| 4 | 95.85 | ||
| 5 | 93.84 | ||
| 6 | 92.03 | ||
| Proposed method | Clustering methodology with deep learning LSTM | 2 | 97.85 |
| 3 | 96.78 | ||
| 4 | 96.42 | ||
| 5 | 93.35 | ||
| 6 | 92.47 |