| Literature DB >> 36188460 |
Abstract
Emotion recognition from electroencephalogram (EEG) signals requires accurate and efficient signal processing and feature extraction. Deep learning technology has enabled the automatic extraction of raw EEG signal features that contribute to classifying emotions more accurately. Despite such advances, classification of emotions from EEG signals, especially recorded during recalling specific memories or imagining emotional situations has not yet been investigated. In addition, high-density EEG signal classification using deep neural networks faces challenges, such as high computational complexity, redundant channels, and low accuracy. To address these problems, we evaluate the effects of using a simple channel selection method for classifying self-induced emotions based on deep learning. The experiments demonstrate that selecting key channels based on signal statistics can reduce the computational complexity by 89% without decreasing the classification accuracy. The channel selection method with the highest accuracy was the kurtosis-based method, which achieved accuracies of 79.03% and 79.36% for the valence and arousal scales, respectively. The experimental results show that the proposed framework outperforms conventional methods, even though it uses fewer channels. Our proposed method can be beneficial for the effective use of EEG signals in practical applications.Entities:
Keywords: channel selection; convolutional neural network; deep learning; high-density EEG; self-induced emotion recognition
Year: 2022 PMID: 36188460 PMCID: PMC9523358 DOI: 10.3389/fnins.2022.985709
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 5.152
Figure 1Flow diagram of the proposed system for recognition of self-induced emotions.
Number of samples for each class of emotion.
|
|
|
|
|---|---|---|
| Valence | Low (Negative) | 498 |
| High (Positive) | 636 | |
| Arousal | Low (Calm) | 489 |
| High (Active) | 645 |
EEG signal statistics used for channel selection.
|
|
|
|---|---|
| Mean |
|
| Variance |
|
| Root mean square |
|
| Skewness |
|
| Kurtosis |
|
In the equations, x(i) is the i-th data point of the EEG signal for channel c and N denotes the total number of data points.
Architecture of ShallowConvNet.
|
|
|
|---|---|
| L1 | 40 × Conv(3 × 1), stride(1 × 1) |
| 40 × Conv(1 × | |
| BatchNorm | |
| Activation(Square) | |
| AvgPool(30 × 1), stride(4 × 1) | |
| Activation(Log) | |
| Dropout(0.5) | |
| Output | Dense |
| Softmax classification |
Hyperparameter values of ShallowConvNet.
|
|
|
|---|---|
| Optimizer | Adam |
| Learning rate | 0.000625 |
| Batch size | 8 |
| Epochs | 150 [valence] |
| 50 [arousal] | |
| Loss function | Negative log likelihood |
Average classification performance for different frequency bands using all channels.
|
|
|
| ||
|---|---|---|---|---|
|
|
|
|
| |
| δ band | 62.07 | 56.33 | 60.81 | 56.45 |
| θ band | 62.80 | 57.70 | 60.32 | 56.43 |
| α band | 64.67 | 59.62 | 65.30 | 61.61 |
| β band | 73.39 | 70.60 | 71.76 | 69.16 |
|
|
|
|
| |
| All (δ, θ, α, β, γ) | 72.37 | 68.93 | 71.24 | 68.87 |
The best results are in bold.
Figure 2Comparison of valence classification accuracies for different EEG channel selection methods.
Figure 3Comparison of arousal classification accuracies for different EEG channel selection methods.
Comparison of the accuracy (%) of different channel selection methods for the γ band.
|
|
|
|
|
|---|---|---|---|
|
|
| ||
| Valence | Mean | 77.13 | 74.23 |
| (79) | |||
| Variance | 77.28 | 75.06 | |
| (108) | |||
| RMS | 78.15 | 75.22 | |
| (87) | |||
| Skewness | 76.97 | 75.33 | |
| (78) | |||
|
|
|
| |
|
| |||
| Arousal | Mean | 79.01 | 75.86 |
| (122) | |||
| Variance | 78.52 | 76.38 | |
| (70) | |||
| RMS | 77.78 | 75.33 | |
| (114) | |||
|
|
|
| |
| (119) | |||
|
|
|
| |
|
|
K is the number of selected channels. The best results are in bold.
Figure 4Average classification accuracies of different frequency bands for the kurtosis-based channel selection as a function of the number of selected channels. (A) Valence scale classification accuracy. (B) Arousal scale classification accuracy.
Performance of the proposed framework in terms of average accuracy (%) and execution time.
|
|
|
|
| ||
|---|---|---|---|---|---|
|
|
|
|
| ||
|
|
|
|
| ||
| ShallowConvNet (Schirrmeister et al., | Baseline (1-50 Hz) | 72.37 ± 15.40 | 6 m 60 s | 71.24 ± 16.11 | 2 m 20 s |
| (246) | (246) | ||||
| BPF (30–50 Hz) | 75.97 ± 16.24 | 6 m 50 s | 77.68 ± 13.38 | 2 m 18 s | |
| (246) | (246) | ||||
|
|
| ||||
|
|
| ||||
| DeepConvNet (Schirrmeister et al., | Baseline (1-50 Hz) | 69.67 ± 16.66 | 30 m 31 s | 65.93 ± 15.15 | 10 m 17 s |
| (246) | (246) | ||||
| BPF (30–50 Hz) | 73.27 ± 17.63 | 30 m 02 s | 72.89 ± 14.59 | 10 m 11 s | |
| (246) | (246) | ||||
|
|
|
| |||
|
|
| ||||
BPF stands for “band-pass filter”, which indicates the frequency band selection process. The best results are in bold.
Performance of subject-independent classification using ShallowConvNet.
|
|
|
| ||
|---|---|---|---|---|
|
|
|
|
| |
|
|
| |||
| Baseline (1-50 Hz) | 59.95 ± 8.96 | 4 m 10 s | 57.71 ± 8.40 | 4 m 11 s |
| BPF (30-50 Hz) | 62.67 ± 8.57 | 4 m 08 s | 60.29 ± 8.33 | 4 m 10 s |
|
|
|
| ||
The best results are in bold.