| Literature DB >> 34335165 |
Jing Chen1, Haifeng Li1, Lin Ma1, Hongjian Bo2, Frank Soong3, Yaohui Shi4.
Abstract
Recently, emotion classification from electroencephalogram (EEG) data has attracted much attention. As EEG is an unsteady and rapidly changing voltage signal, the features extracted from EEG usually change dramatically, whereas emotion states change gradually. Most existing feature extraction approaches do not consider these differences between EEG and emotion. Microstate analysis could capture important spatio-temporal properties of EEG signals. At the same time, it could reduce the fast-changing EEG signals to a sequence of prototypical topographical maps. While microstate analysis has been widely used to study brain function, few studies have used this method to analyze how brain responds to emotional auditory stimuli. In this study, the authors proposed a novel feature extraction method based on EEG microstates for emotion recognition. Determining the optimal number of microstates automatically is a challenge for applying microstate analysis to emotion. This research proposed dual-threshold-based atomize and agglomerate hierarchical clustering (DTAAHC) to determine the optimal number of microstate classes automatically. By using the proposed method to model the temporal dynamics of auditory emotion process, we extracted microstate characteristics as novel temporospatial features to improve the performance of emotion recognition from EEG signals. We evaluated the proposed method on two datasets. For public music-evoked EEG Dataset for Emotion Analysis using Physiological signals, the microstate analysis identified 10 microstates which together explained around 86% of the data in global field power peaks. The accuracy of emotion recognition achieved 75.8% in valence and 77.1% in arousal using microstate sequence characteristics as features. Compared to previous studies, the proposed method outperformed the current feature sets. For the speech-evoked EEG dataset, the microstate analysis identified nine microstates which together explained around 85% of the data. The accuracy of emotion recognition achieved 74.2% in valence and 72.3% in arousal using microstate sequence characteristics as features. The experimental results indicated that microstate characteristics can effectively improve the performance of emotion recognition from EEG signals.Entities:
Keywords: EEG; auditory emotion process; dual-threshold-based AAHC; emotion recognition; microstate characteristics
Year: 2021 PMID: 34335165 PMCID: PMC8318040 DOI: 10.3389/fnins.2021.689791
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 4.677
FIGURE 1The schema of the methodology. The main six steps are: (A) The auditory emotional experimental design. (B) The pre-processing for EEG signals. (C) The proposed microstate analysis to identify the microstates. (D) Back-fitting to obtain the microstate sequences. (E) Microstate characteristics extraction as features. (F) Multivariate pattern analysis for emotion recognition.
FIGURE 2The distribution of ratings on arousal–valence plane.
FIGURE 3The process of speech-evoked emotion cognitive experiment.
Database summary.
| Valence | Arousal | |||
| Condition | High | Low | High | Low |
| Number of trials | 790 | 583 | 815 | 558 |
| Rating | 5.9 ± 0.8 | 3.3 ± 0.5 | 6.4 ± 0.7 | 3.7 ± 0.3 |
Pseudocode for dual-threshold-based atomize and agglomerate hierarchical clustering (DTAAHC).
| the spatial correlation |
| Th |
| Th |
| 1: |
| 2: |
| 3: |
| 4: |
| 5: |
| 6: |
| 7: |
| 8: |
| 9: |
| 10: merge two clusters |
| 11: define the centroid (mathematical average) as the template map for that cluster |
| 12: calculate the GEV |
| 13: the cluster with the min (GEV |
| 14: calculate the GMD for each pair of template map |
| 15: merge clusters if the GMD is lower than Th |
| 16: |
FIGURE 4The topographical maps of the microstates across subjects. (A) Microstates from speech-evoked emotion cognitive experiment. (B) Microstates from music-evoked datasets. (C) The global explained variance (GEV) of each microstate for two datasets.
The global map dissimilarity (GMD) between different microstates of dataset 1.
| GMD | Microstates from dataset 1 | |||||||||
| #1 | #2 | #3 | #4 | #5 | #6 | #7 | #8 | #9 | ||
| #1 | 0 | 0.11 | 0.23 | 0.25 | 0.12 | 0.23 | 0.15 | 0.23 | 0.10 | |
| #2 | 0.11 | 0 | 0.23 | 0.23 | 0.14 | 0.21 | 0.17 | 0.21 | 0.11 | |
| #3 | 0.23 | 0.23 | 0 | 0.11 | 0.24 | 0.11 | 0.23 | 0.12 | 0.23 | |
| #4 | 0.25 | 0.23 | 0.11 | 0 | 0.25 | 0.11 | 0.22 | 0.10 | 0.24 | |
| #5 | 0.12 | 0.14 | 0.24 | 0.25 | 0 | 0.23 | 0.10 | 0.24 | 0.10 | |
| #6 | 0.23 | 0.21 | 0.11 | 0.11 | 0.22 | 0 | 0.23 | 0.10 | 0.23 | |
| #7 | 0.15 | 0.17 | 0.23 | 0.22 | 0.10 | 0.23 | 0 | 0.24 | 0.10 | |
| #8 | 0.23 | 0.21 | 0.12 | 0.10 | 0.24 | 0.10 | 0.24 | 0 | 0.23 | |
| #9 | 0.10 | 0.11 | 0.23 | 0.24 | 0.10 | 0.23 | 0.10 | 0.23 | 0 | |
The GMD between different microstates of dataset 2 (Dataset for Emotion Analysis using Physiological signals, DEAP).
| GMD | Microstates from DEAP | ||||||||||
| #1 | #2 | #3 | #4 | #5 | #6 | #7 | #8 | #9 | #10 | ||
| #1 | 0 | 0.25 | 0.29 | 0.30 | 0.18 | 0.24 | 0.28 | 0.14 | 0.31 | 0.33 | |
| #2 | 0.25 | 0 | 0.33 | 0.19 | 0.34 | 0.22 | 0.11 | 0.33 | 0.16 | 0.16 | |
| #3 | 0.29 | 0.33 | 0 | 0.30 | 0.14 | 0.32 | 0.28 | 0.20 | 0.23 | 0.28 | |
| #4 | 0.30 | 0.19 | 0.30 | 0 | 0.33 | 0.11 | 0.24 | 0.31 | 0.25 | 0.11 | |
| #5 | 0.18 | 0.34 | 0.14 | 0.33 | 0 | 0.30 | 0.32 | 0.10 | 0.30 | 0.34 | |
| #6 | 0.24 | 0.22 | 0.32 | 0.11 | 0.30 | 0 | 0.29 | 0.26 | 0.31 | 0.17 | |
| #7 | 0.28 | 0.11 | 0.28 | 0.24 | 0.32 | 0.29 | 0 | 0.33 | 0.11 | 0.19 | |
| #8 | 0.14 | 0.33 | 0.20 | 0.31 | 0.10 | 0.26 | 0.33 | 0 | 0.33 | 0.34 | |
| #9 | 0.31 | 0.16 | 0.23 | 0.25 | 0.30 | 0.31 | 0.11 | 0.33 | 0 | 0.194 | |
| #10 | 0.33 | 0.16 | 0.28 | 0.11 | 0.34 | 0.17 | 0.19 | 0.34 | 0.19 | 0 | |
Means for all microstate parameters of speech-evoked EEG signals.
| Microstate classes | ||||||||||
| Temporal parameters | #1 | #2 | #3 | #4 | #5 | #6 | #7 | #8 | #9 | |
| Valence | High | 97.58 (17.4) | 106.33 (21.0) | 110.73 (22.5) | 74.12 (15.6) | 87.19 (19.3) | 107.06 (23.9) | 104.96 (20.8) | 83.89 (16.8) | 69.03 (28.1) |
| Low | 98.98 (19.6) | 107.26 (19.5) | 102.80 (21.8) | 73.85 (17.5) | 82.46 (17.0) | 105.71 (22.5) | 103.66 (19.5) | 80.56 (18.6) | 67.27 (227.5) | |
| 0.53 | 0.22 | 0.83 | 0.86 | 0.77 | 0.51 | 0.79 | 0.79 | |||
| Arousal | High | 98.32 (19.0) | 106.77 (20.1) | 105.21 (22.8) | 73.82 (16.8) | 83.18 (17.5) | 106.64 (23.3) | 103.86 (19.7) | 81.60 (18.4) | 67.66 (27.8) |
| Low | 102.72 (23.9) | 110.36 (17.2) | 98.90 (18.9) | 75.19 (22.8) | 89.18 (24.3) | 98.55 (20.1) | 105.45 (18.3) | 78.48 (13.1) | 68.32 (30.5) | |
| 0.11 | 0.09 | 0.64 | 0.52 | 0.49 | 0.79 | 0.73 | ||||
| Valence | High | 1.20 (0.6) | 1.51 (0.6) | 1.30 (0.6) | 0.46 (0.3) | 0.67 (0.4) | 1.30 (0.6) | 1.44 (0.5) | 0.78 (0.5) | 0.47 (0.2) |
| Low | 1.23 (0.6) | 1.55 (0.5) | 1.35 (0.7) | 0.45 (0.3) | 0.68 (0.4) | 1.31 (0.6) | 1.45 (0.5) | 0.74 (0.5) | 0.51 (0.2) | |
| 0.52 | 0.50 | 0.49 | 0.93 | 0.86 | 0.56 | 0.84 | 0.64 | 0.27 | ||
| Arousal | High | 1.22 (0.6) | 1.54 (0.6) | 1.34 (0.7) | 0.45 (0.3) | 0.68 (0.4) | 1.31 (0.6) | 1.43 (0.5) | 0.75 (0.5) | 0.51 (0.2) |
| Low | 1.31 (0.6) | 1.57 (0.5) | 1.25 (0.6) | 0.43 (0.3) | 0.65 (0.3) | 1.31 (0.8) | 1.62 (0.6) | 0.71 (0.4) | 0.44 (0.2) | |
| 0.20 | 0.62 | 0.39 | 0.55 | 0.72 | 0.76 | 0.70 | 0.74 | |||
| Valence | High | 12.11 (6.4) | 16.57 (8.0) | 15.84 (9.3) | 4.13 (2.7) | 6.48 (3.5) | 14.29 (6.9) | 15.53 (6.6) | 7.26 (4.8) | 7.74 (1.6) |
| Low | 12.79 (6.6) | 17.02 (7.0) | 15.22 (9.3) | 4.02 (2.5) | 6.37 (3.8) | 14.59 (8.2) | 15.16 (6.1) | 7.12 (4.7) | 7.67 (1.8) | |
| 0.45 | 0.37 | 0.78 | 0.97 | 0.92 | 0.65 | 0.82 | 0.64 | 0.58 | ||
| Arousal | High | 12.52 (6.5) | 16.85 (7.3) | 15.51 (9.4) | 4.06 (2.5) | 6.36 (3.7) | 14.61 (8.0) | 15.10 (6.2) | 7.21 (4.8) | 7.73 (1.8) |
| Low | 14.02 (6.8) | 17.64 (5.4) | 13.59 (7.4) | 3.88 (2.6) | 6.88 (3.9) | 13.27 (7.8) | 17.13 (6.6) | 6.49 (3.7) | 7.07 (1.7) | |
| 0.09 | 0.32 | 0.20 | 0.71 | 0.87 | 0.25 | 0.58 | 0.68 | |||
| Valence | High | 6.35 (4.3) | 6.38 (3.8) | 9.22 (6.8) | 3.65 (2.9) | 3.82 (2.3) | 6.01 (4.3) | 4.87 (2.7) | 4.84 (3.9) | 3.06 (0.6) |
| Low | 6.53 (4.5) | 6.79 (4.0) | 9.08 (7.2) | 3.27 (2.2) | 3.64 (2.2) | 6.16 (4.0) | 4.82 (2.5) | 4.58 (3.5) | 2.90 (0.5) | |
| 0.78 | 0.21 | 0.91 | 0.67 | 0.93 | 0.69 | 0.93 | 0.68 | 0.47 | ||
| Arousal | High | 6.42 (4.4) | 6.62 (4.0) | 9.21 (7.1) | 3.38 (2.4) | 3.65 (2.3) | 6.20 (4.1) | 4.76 (2.5) | 4.71 (3.7) | 2.97 (0.5) |
| Low | 7.37 (4.0) | 7.59 (2.5) | 7.97 (6.4) | 3.19 (2.7) | 4.13 (1.6) | 5.16 (3.6) | 5.65 (2.5) | 3.78 (2.8) | 2.67 (0.5) | |
| 0.12 | 0.08 | 0.39 | 0.71 | 0.72 | 0.32 | 0.54 | 0.92 | |||
The differences (p-value) of transition probabilities between high and low valence or arousal.
| → | Dimensions | |||||||||
| Valence | – | 0.99 | 0.09 | 0.47 | 0.003 | 0.97 | 0.47 | 0.87 | 0.78 | |
| Arousal | – | 0.54 | 0.56 | 0.48 | 0.77 | 0.42 | 0.65 | 0.61 | 0.94 | |
| Valence | 0.29 | – | 0.48 | 0.49 | 0.10 | 0.80 | 0.29 | 0.75 | 0.88 | |
| Arousal | 0.76 | – | 0.98 | 0.86 | 0.20 | 0.98 | 0.12 | 0.64 | 0.86 | |
| Valence | 0.66 | 0.83 | – | 0.79 | 0.66 | 0.12 | 0.20 | 0.93 | 0.57 | |
| Arousal | 0.91 | 0.43 | – | 0.18 | 0.26 | 0.94 | 0.45 | 0.61 | 0.19 | |
| Valence | 0.72 | 0.41 | 0.75 | – | 0.31 | 0.28 | 0.66 | 0.48 | 0.08 | |
| Arousal | 0.44 | 0.53 | 0.91 | – | 0.46 | 0.12 | 0.41 | 0.74 | 0.38 | |
| Valence | 0.10 | 0.70 | 0.72 | 0.76 | – | 0.07 | 0.67 | 0.98 | 0.83 | |
| Arousal | 0.29 | 0.08 | 0.21 | 0.64 | – | 0.51 | 0.13 | 0.37 | 0.26 | |
| Valence | 0.29 | 0.19 | 0.43 | 0.12 | 0.20 | – | 0.23 | 0.24 | 0.24 | |
| Arousal | 0.42 | 0.48 | 0.23 | 0.09 | 0.40 | – | 0.37 | 0.36 | 0.76 | |
| Valence | 0.99 | 0.34 | 0.06 | 0.28 | 0.32 | 0.89 | – | 0.32 | 0.48 | |
| Arousal | 0.12 | 0.87 | 0.68 | 0.43 | 0.92 | 0.52 | – | 0.43 | 0.70 | |
| Valence | 0.04 | 0.72 | 0.06 | 0.51 | 0.68 | 0.22 | 0.96 | – | 0.26 | |
| Arousal | 0.21 | 0.46 | 0.04 | 0.77 | 0.60 | 0.98 | 0.0002 | – | 0.71 | |
| Valence | 0.74 | 0.84 | 0.65 | 0.93 | 0.97 | 0.98 | 0.91 | 0.92 | – | |
| Arousal | 0.67 | 0.65 | 0.12 | 0.01 | 0.08 | 0.23 | 0.78 | 0.21 | – |
FIGURE 5Connections with the statistically significant difference between groups. The blue arrows represent p < 0.05. The red arrows represent p < 0.10 for (A) high vs. low valence groups and for (B) high vs. low arousal groups.
The classification accuracies of different feature sets on dataset 2 (Dataset for Emotion Analysis using Physiological signals, DEAP).
| References | Feature set | Classifier | Accuracy | |
| Valence (%) | Arousal (%) | |||
| MEMD, PSD, | k-NN | 67.0 | 51.0 | |
| EMD, TSD, PD, NE | SVM | 69.1 | 71.9 | |
| DT-CWPT | SVM | 65.3 | 66.9 | |
| Temporal parameters | SVM | 72.5 | 72.1 | |
| Transition probabilities | SVM | 74.4 | 73.9 | |
| This study | Temporal parameters + | |||
| transition probabilities | SVM | |||
The classification accuracies of different feature sets on speech-evoked EEG signals.
| Dataset | Feature set | Classifier | Accuracy | |
| Valence (%) | Arousal (%) | |||
| Support vector | 71.8 | 68.8 | ||
| parameters | Random forest (RF) | 72.0 | 67.9 | |
| Artificial neural network (ANN) | 72.3 | 69.5 | ||
| SVM | 69.9 | 70.5 | ||
| This study | Transition probabilities | RF | 68.5 | 68.3 |
| ANN | 70.4 | 69.8 | ||
| Temporal parameters + transition probabilities | SVM | 71.9 | ||
| RF | 73.1 | 70.7 | ||
| ANN | 73.9 | |||