| Literature DB >> 28991172 |
David Lee1, Sang-Hoon Park2, Sang-Goog Lee3.
Abstract
In this paper, we propose a set of wavelet-based combined feature vectors and a Gaussian mixture model (GMM)-supervector to enhance training speed and classification accuracy in motor imagery brain-computer interfaces. The proposed method is configured as follows: first, wavelet transforms are applied to extract the feature vectors for identification of motor imagery electroencephalography (EEG) and principal component analyses are used to reduce the dimensionality of the feature vectors and linearly combine them. Subsequently, the GMM universal background model is trained by the expectation-maximization (EM) algorithm to purify the training data and reduce its size. Finally, a purified and reduced GMM-supervector is used to train the support vector machine classifier. The performance of the proposed method was evaluated for three different motor imagery datasets in terms of accuracy, kappa, mutual information, and computation time, and compared with the state-of-the-art algorithms. The results from the study indicate that the proposed method achieves high accuracy with a small amount of training data compared with the state-of-the-art algorithms in motor imagery EEG classification.Entities:
Keywords: brain–computer interface (BCI); electroencephalogram (EEG); motor imagery; support vector machine; training data reduction; wavelet transform
Mesh:
Year: 2017 PMID: 28991172 PMCID: PMC5677306 DOI: 10.3390/s17102282
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Block diagram of motor imagery brain-computer interface (BCI) using the wavelet-based combined feature vector and gaussian mixture model (GMM)-supervector.
Summary information of the data for 12 subjects from BCI Competition II, III, and IV.
| Subject | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 140 | 540 | 540 | 400 | 400 | 400 | 420 | 420 | 400 | 400 | 440 | 400 | |
| 140 | 540 | 540 | 320 | 280 | 320 | 320 | 320 | 320 | 320 | 320 | 320 |
Figure 2GMM-supervector for two features for Subject 7 in dataset IIb of the BCI competition IV.
Comparative results of the feature extraction methods in terms of the average classification accuracy (%).
| Subject | DWT | CWT | Combined Feature Vectors | |||
|---|---|---|---|---|---|---|
| without PCA | with PCA | |||||
| Accuracy | Number of | Accuracy | Number of | |||
| 92.9 | 94.1 | 96.4 | 96 | 97.5 | 33 | |
| 73.3 | 81.4 | 83.1 | 96 | 83.1 | 40 | |
| 62.5 | 80.7 | 80.5 | 96 | 83.3 | 39 | |
| 75.2 | 77.7 | 79.1 | 96 | 76.7 | 35 | |
| 52.6 | 60.1 | 60.0 | 96 | 61.4 | 34 | |
| 55.4 | 55.6 | 51.8 | 96 | 56.2 | 32 | |
| 95.5 | 96.2 | 96.2 | 96 | 96.1 | 30 | |
| 88.3 | 87.4 | 94.0 | 96 | 94.1 | 32 | |
| 79.2 | 87.8 | 90.2 | 96 | 88.1 | 34 | |
| 74.8 | 72.6 | 81.4 | 96 | 80.7 | 36 | |
| 90.6 | 87.7 | 89.3 | 96 | 90.0 | 32 | |
| 78.7 | 83.3 | 84.2 | 96 | 84.8 | 34 | |
| 76.6 | 80.4 | 82.2 | 96 | 82.7 | 34.3 | |
| - | - | - | ||||
DWT: discrete wavelet transform, CWT: continuous wavelet transform, PCA: principal component analysis.
Mutual information of the proposed combined feature vectors (100% and 30% of training data), methods in previous literature [21,65], and the winning methods of dataset I of BCI Competition II.
| Ranking | Methods | Subject 1 | |
|---|---|---|---|
| Maximal MI (bit) | Accuracy (%) | ||
| ALL-SVM | 0.84 | 97.50 | |
| 30%-SVM | 0.67 | 93.79 | |
| FSVM in [ | 0.66 | 87.86 | |
| SVM in [ | 0.65 | 89.83 | |
| NN in [ | 0.64 | 90.00 | |
| LDA in [ | 0.63 | 89.29 | |
| 1st winner | 0.61 | 89.29 | |
| SVM in [ | 0.58 | 90.00 | |
| 2nd winner | 0.46 | 84.29 | |
SVM: support vector machine, FSVM fuzzy SVM, NN: neural network, LDA: linear discriminant analysis.
Mutual information of the proposed combined feature vectors (100% and 30% of training data), methods in [21] and the winning methods of dataset II.
| Ranking | Methods | Maximal MI(bit) | ||
|---|---|---|---|---|
| Subject 2 | Subject 3 | Mean | ||
| 1st winner | 0.4382 | 0.3489 | 0.3936 | |
| ALL-SVM | 0.3447 | 0.3562 | 0.3505 | |
| 30%-SVM | 0.3105 | 0.3216 | 0.3161 | |
| 2nd winner | 0.4174 | 0.1719 | 0.2947 | |
| FSVM in [ | 0.0718 | 0.0863 | 0.0791 | |
| SVM in [ | 0.0718 | 0.0809 | 0.0764 | |
Maximum Kappa value of the proposed combined feature vectors (100% and 30% of training data), and winning methods of the dataset III.
| Ranking | Methods | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | Mean |
|---|---|---|---|---|---|---|---|---|---|---|---|
| ALL-SVM | 0.54 | 0.24 | 0.12 | 0.92 | 0.88 | 0.76 | 0.61 | 0.80 | 0.70 | 0.62 | |
| 1st winner | 0.40 | 0.21 | 0.22 | 0.95 | 0.86 | 0.61 | 0.56 | 0.85 | 0.74 | 0.60 | |
| 30%-SVM | 0.51 | 0.17 | 0.12 | 0.92 | 0.83 | 0.76 | 0.55 | 0.79 | 0.67 | 0.59 | |
| 2nd winner | 0.42 | 0.21 | 0.14 | 0.94 | 0.71 | 0.62 | 0.61 | 0.84 | 0.78 | 0.59 | |
| 3rd winner | 0.19 | 0.12 | 0.12 | 0.77 | 0.57 | 0.49 | 0.37 | 0.85 | 0.61 | 0.45 |
Figure 3Classification accuracy based on reduction rate of whole training data on individual subjects.
Figure 4Computation time for training procedure based on reduction rate of whole training data on individual subjects.
Figure 5Mean classification accuracy based on selected training data on all subjects.
Figure 6Computation time for training procedure based on selected training data on all subjects.