| Literature DB >> 35223807 |
Yimin Hou1, Shuyue Jia2, Xiangmin Lun1,3, Shu Zhang4, Tao Chen1, Fang Wang1, Jinglei Lv5.
Abstract
Recognition accuracy and response time are both critically essential ahead of building the practical electroencephalography (EEG)-based brain-computer interface (BCI). However, recent approaches have compromised either the classification accuracy or the responding time. This paper presents a novel deep learning approach designed toward both remarkably accurate and responsive motor imagery (MI) recognition based on scalp EEG. Bidirectional long short-term memory (BiLSTM) with the attention mechanism is employed, and the graph convolutional neural network (GCN) promotes the decoding performance by cooperating with the topological structure of features, which are estimated from the overall data. Particularly, this method is trained and tested on the short EEG recording with only 0.4 s in length, and the result has shown effective and efficient prediction based on individual and groupwise training, with 98.81% and 94.64% accuracy, respectively, which outperformed all the state-of-the-art studies. The introduced deep feature mining approach can precisely recognize human motion intents from raw and almost-instant EEG signals, which paves the road to translate the EEG-based MI recognition to practical BCI systems.Entities:
Keywords: bidirectional long short-term memory (BiLSTM); brain–computer interface (BCI); electroencephalography (EEG); graph convolutional neural network (GCN); motor imagery (MI)
Year: 2022 PMID: 35223807 PMCID: PMC8873790 DOI: 10.3389/fbioe.2021.706229
Source DB: PubMed Journal: Front Bioeng Biotechnol ISSN: 2296-4185
FIGURE 1The schematical overview consisted of the 64-channel raw electroencephalography (EEG) signal acquisition, the bidirectional long short-term memory (BiLSTM) with the attention model for feature extraction, and the graph convolutional neural network (GCN) model for classification.
FIGURE 2Presented BiLSTM with the attention mechanism for feature extraction.
FIGURE 3The Pearson, absolute Pearson, adjacency, and Laplacian matrices for subject nine. (A) Pearson matrix for subject nine. (B) Absolute Pearson matrix for subject nine. (C) Adjacency matrix for subject nine. (D) Laplacian matrix for subject nine.
FIGURE 4Comparison of models and hyperparameters w.r.t. the recurrent neural network (RNN)-based methods for feature extraction. (A) Global average accuracy (GAA) w.r.t. RNN-based models. (B) GAA w.r.t. BiLSTM cell size. (C) GAA w.r.t. attention size of the BiLSTM. (D) GAA w.r.t. the number of the extracted features. (E) Loss w.r.t. RNN-based models. (F) Loss w.r.t. BiLSTM cell size. (G) Loss w.r.t. attention size of the BiLSTM. (H) Loss w.r.t. the number of the extracted features.
FIGURE 5Box plot and confusion matrix for 10-fold cross-validation. (A) Box plot for repetitive experiments. (B) Confusion matrix for test one.
FIGURE 6GAA and receiver operating characteristic curve (ROC curve) for 20 and 50 subjects, separately. (A) GAA w.r.t. groupwise prediction. (B) ROC curve w.r.t. groupwise prediction.
Comparison on groupwise evaluation.
| Related work | Max. global average accuracy (GAA) (%) | Approach | Number of subjects | Database |
|---|---|---|---|---|
|
| 68.20 | Recurrent neural networks (RNNs) | 12 | PhysioNet database |
|
| 94.50 | ESI + convolutional neural networks (CNNs) | 10 | |
| 92.50 | 14 | |||
| This work | 94.64 | Attention-based bidirectional long short-term memory (BiLSTM)–graph convolutional neural network (GCN) | 20 |
Subject-level evaluation.
| No. of subject | GAA (%) | Kappa (%) | Precision (%) | Recall (%) | F1 score (%) |
|---|---|---|---|---|---|
| 1 | 94.05 | 92.06 | 94.20 | 94.32 | 94.16 |
| 2 | 96.43 | 95.19 | 96.06 | 96.06 | 96.06 |
| 3 | 97.62 | 96.79 | 97.33 | 97.08 | 97.18 |
| 4 | 90.48 | 87.34 | 91.30 | 91.11 | 90.42 |
| 5 | 95.24 | 93.61 | 95.96 | 95.06 | 95.38 |
| 6 | 94.05 | 92.02 | 93.40 | 94.96 | 93.66 |
| 7 | 98.81 | 98.40 | 98.81 | 99.07 | 98.92 |
| 8 | 95.24 | 93.60 | 95.39 | 95.04 | 95.19 |
| 9 | 98.81 | 98.39 | 99.11 | 98.68 | 98.87 |
| 10 | 94.05 | 91.98 | 93.39 | 94.70 | 93.61 |
| Average | 95.48 | 93.94 | 95.50 | 95.61 | 95.35 |
Comparison of current studies on subject-level prediction.
| Related work | Max. GAA (%) | Approach | Database |
|---|---|---|---|
|
| 94.66 | Sorted-fast ICA-CWT + CNNs | Brain–computer interface (BCI) Competition IV-a dataset |
|
| 95.20 | EWT + LS-SVM | |
|
| 96.89 | TQWT + LS-SVM | |
|
| 83.00 | CNNs–long short-term memory (LSTM) | BCI Competition IV-2a dataset |
|
| 95.10 | SVM | |
|
| 95.40 | MCNNs | |
|
| 68.51 | CNNs | PhysioNet database |
| Hou et al. (2019) | 96.00 | ESI + CNNs | |
| This work | 98.81 | Attention-based BiLSTM–GCN |
FIGURE 7Loss and ROC curve for subject-level evaluation. (A) Loss w.r.t. subject-level validation. (B) ROC curve w.r.t. subject-level validation.