| Literature DB >> 31059799 |
Ivan Zubarev1, Rasmus Zetter2, Hanna-Leena Halme2, Lauri Parkkonen3.
Abstract
We introduce two Convolutional Neural Network (CNN) classifiers optimized for inferring brain states from magnetoencephalographic (MEG) measurements. Network design follows a generative model of the electromagnetic (EEG and MEG) brain signals allowing explorative analysis of neural sources informing classification. The proposed networks outperform traditional classifiers as well as more complex neural networks when decoding evoked and induced responses to different stimuli across subjects. Importantly, these models can successfully generalize to new subjects in real-time classification enabling more efficient brain-computer interfaces (BCI).Entities:
Keywords: Brain–computer interface; Convolutional neural network; Magnetoencephalography
Mesh:
Year: 2019 PMID: 31059799 PMCID: PMC6609925 DOI: 10.1016/j.neuroimage.2019.04.068
Source DB: PubMed Journal: Neuroimage ISSN: 1053-8119 Impact factor: 6.556
Fig. 1Architecture of the two variants of the CNN.
Fig. 3Interpretation of informative LF-CNN model parameters in a representative subject from Experiment 1. A. Components having the maximum contribution to the decoding of each class were extracted from the model and interpreted in terms of their spatial topographies (top), source estimates (middle) and latency estimates (bottom). B. Spatial topographies(top), source estimates (middle) and peak latencies (bottom) of the corresponding evoked responses. Source estimates visualization is thresholded to 95% of the peak source activity.
Fig. 4Identification of informative latent components in a single representative subject from Experiment 1. The weights of the contributions of the final layer to each class are represented as a raster plot (top) with rows corresponding to the index of latent component and columns corresponding to (pooled) time points. Informative components (middle) are found by identifying single features (indicated by red boxes) with maximum positive weight for each class. Non-informative components are defined as having the minimal absolute sum of weights across all classes. Component topographies are scaled to interval [0,1] for comparability.
Tested and optimal hyperparameter values.
| Parameter | Tested | Optimal |
|---|---|---|
| Number of latent sources | 16,32,64 | 32 |
| Temporal filter length | 3,5,7,9,11 | 7 |
| Learning rate | 1⋅10−3, 3⋅10−4, 1⋅10−4 | 3⋅10−4 |
| 1⋅10−3, 3⋅10−4, 1⋅10−4 | 3⋅10−4 | |
| Pooling | ||
| Pooling factor | 2,3,5 | 2 |
| Drop-out coefficient | 0.25, 0.50, 0.75, 0.90 | 0.50 |
| Input layer link function | ||
| Hidden layer link function | ||
| Output nonlinearity | ||
| Number of dense hidden layers | 1,2 | 1 |
Across-subject performance in a 5-class sensory stimulation task. Grand-average accuracy scores (mean SD) from leave-one-subject-out cross-validation. Highest-performing model in each test is indicated in bold.
| Model | Validation (%) | Initial test (%) | Pseudo-real-time (%) |
|---|---|---|---|
| LF-CNN | 95.0 ± 0.8 | 83.1 ± 8.4 | 93.3 ± 3.6 |
| VAR-CNN | |||
| Linear SVM | 93.3 ± 1.2 | 80.2 ± 9.7 | 87.0 ± 5.4 |
| RBF-SVM | 93.6 ± 1.7 | 82.7 ± 8.3 | 83.9 ± 8.4 |
| ShallowFBCSP-CNN | 85.3 ± 2.4 | 60.1 ± 11.7 | n.a. |
| EEGNet-8 | 88.7 ± 2.0 | 76.8 ± 11.7 | 89.2 ± 5.0 |
| VGG19 | 80.5 ± 3.3 | 70.1 ± 12.8 | 73.9 ± 10.5 |
Across-subject performance in a 3-class motor imagery task. Grand-average accuracy scores (mean SD) from leave-one-subject-out cross-validation. Highest-performing model in each test is indicated in bold.
| Model | Validation (%) | Initial test (%) | Pseudo-real-time (%) |
|---|---|---|---|
| LF-CNN | 84.3 ± 2.7 | 74.2 ± 6.5 | 78.0 ± 6.5 |
| VAR-CNN | |||
| Linear SVM | 76.9 ± 3.0 | 68.2 ± 7.2 | 71.4 ± 7.3 |
| RBF-SVM | 80.3 ± 2.6 | 74.1 ± 8.4 | 73.6 ± 8,8 |
| ShallowFBCSP-CNN | 70.2 ± 4.1 | 60.2 ± 10.3 | n.a. |
| EEGNet-8 | 80.8 ± 2.4 | 72.1 ± 5.8 | 80.9 ± 6.7 |
| VGG19 | 71.4 ± 9.6 | 60.2 ± 6.8 | 57.4 ± 9.2 |
VAR-CNN classification accuracy in the real-time motor imagery BCI experiment.
| Subject | Run 1, no updates (%) | Run 2, online updates (%) | Run 3, online updates (%) |
|---|---|---|---|
| s01 | 80.0 | 88.0 | 92.0 |
| s02 | 62.0 | 90.0 | 82.0 |
Across-subject performance on 2-class Cam-CAN dataset. Grand-average accuracy scores (mean SD) estimated on a test set comprising 50 held-out subjects. Highest-performing model in each test is indicated in bold.
| Model | Validation (%) | Initial test (%) | Pseudo-real-time (%) |
|---|---|---|---|
| LF-CNN | 94.9 | 95.1 ± 4.2 | |
| VAR-CNN | 96.0 ± 2.8 | ||
| Linear SVM | 91.6 | 92.1 ± 5.5 | 92.7 ± 7.6 |
| RBF-SVM | 93.1 | 93.9 ± 5.7 | 94.4 ± 7.3 |
| ShallowFBCSP-CNN | n.a. | n.a. | n.a. |
| EEGNet-8 | 93.8 | 93.0 ± 4.3 | 94.3 ± 3.9 |
| VGG19 | 94.7 | 92.3 ± 5.0 | 93.2 ± 5.2 |
Fig. 5Interpretation of informative LF-CNN model parameters in a representative subject from Experiment 2. Latent components having the maximum positive (red) and negative (blue) sum of weights over all time points for each class were extracted from the model and interpreted in terms of their spatial topographies, and spectral estimates.
Fig. 2Interpretation of model parameters (LF-CNN only).