| Literature DB >> 29615853 |
Xiang Li1, Dawei Song2,3, Peng Zhang1, Yazhou Zhang1, Yuexian Hou1, Bin Hu4.
Abstract
Recognizing cross-subject emotions based on brain imaging data, e.g., EEG, has always been difficult due to the poor generalizability of features across subjects. Thus, systematically exploring the ability of different EEG features to identify emotional information across subjects is crucial. Prior related work has explored this question based only on one or two kinds of features, and different findings and conclusions have been presented. In this work, we aim at a more comprehensive investigation on this question with a wider range of feature types, including 18 kinds of linear and non-linear EEG features. The effectiveness of these features was examined on two publicly accessible datasets, namely, the dataset for emotion analysis using physiological signals (DEAP) and the SJTU emotion EEG dataset (SEED). We adopted the support vector machine (SVM) approach and the "leave-one-subject-out" verification strategy to evaluate recognition performance. Using automatic feature selection methods, the highest mean recognition accuracy of 59.06% (AUC = 0.605) on the DEAP dataset and of 83.33% (AUC = 0.904) on the SEED dataset were reached. Furthermore, using manually operated feature selection on the SEED dataset, we explored the importance of different EEG features in cross-subject emotion recognition from multiple perspectives, including different channels, brain regions, rhythms, and feature types. For example, we found that the Hjorth parameter of mobility in the beta rhythm achieved the best mean recognition accuracy compared to the other features. Through a pilot correlation analysis, we further examined the highly correlated features, for a better understanding of the implications hidden in those features that allow for differentiating cross-subject emotions. Various remarkable observations have been made. The results of this paper validate the possibility of exploring robust EEG features in cross-subject emotion recognition.Entities:
Keywords: DEAP dataset; EEG; SEED dataset; emotion recognition; feature engineering
Year: 2018 PMID: 29615853 PMCID: PMC5867345 DOI: 10.3389/fnins.2018.00162
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 4.677
Figure 1The feature engineering-based method and the procedure for verifying the performance of cross-subject emotion recognition.
Figure 2(A) The data normalization method for one subject's multi-channel signals. (B) The sliding window-based feature extraction method for one EEG signal (taking one 12-s signal as an example). The mean of the calculated values in all sliding windows was adopted as the feature.
This table lists the two main categories of EEG features that we extracted.
| Time-frequency domain features | 1. Peak-Peak Mean. 2. Mean Square Value. 3. Variance. |
| 4. Hjorth Parameter: Activity. 5. Hjorth Parameter: Mobility. | |
| 6. Hjorth Parameter: Complexity. | |
| 7. Maximum Power Spectral Frequency. | |
| 8. Maximum Power Spectral Density. 9. Power Sum. | |
| Non-linear dynamical system features | 10. Approximate Entropy. 11. C0 Complexity. |
| 12. Correlation Dimension. 13. Kolmogorov Entropy. | |
| 14. Lyapunov Exponent. 15. Permutation Entropy. | |
| 16. Singular Entropy. 17. Shannon Entropy. 18. Spectral Entropy. |
The features were extracted for four rhythms. For the DEAP dataset, the total number of features extracted for one trial is 2304. For the SEED dataset, the total number of features extracted for one trial is 4464.
Figure 3Mean cross-subject recognition performance with different methods and settings on DEAP. (A) The filter-based methods. (B) The RFE-based method. (C) The L1-based method. (D) The ROC curves of different methods with their best settings.
Figure 4Mean cross-subject recognition performance with different methods and settings on the SEED dataset. (A) The filter-based methods. (B) The RFE-based method. (C) The L1-based method. (D) The ROC curves of different methods with their best settings.
The performance upper bound of the proposed features using different automatic feature selection methods.
| Step No.: 56 | Step No.: 29 | Step No.: 22 | Step No.: 181 | |
| Mean: 0.5773 | Mean: 0.5617 | Mean: 0.5789 | Mean: 0.5594 | |
| Step No.: 8 | ||||
| Mean: | ||||
| Step No.: 20 | Step No.: 2 | Step No.: 2 | Step No.: 13 | |
| Mean: 0.8244 | Mean: 0.8133 | Mean: 0.8111 | Mean: 0.8289 | |
| Step No.: 12 | ||||
| Mean: | ||||
Meanwhile, the p-values calculated through one-way ANOVA between the method with best performance and other methods are also exhibited. The highest mean recognition accuracy is shown in bold type.
Figure 5The cross-subject recognition performance based on features from different channels (A), different regions (B), different rhythms (C), different features (D), and different feature types (E).
Figure 6The Pearson correlation between 18 different features (linear features: f1, f2, f3, f4, f5, f6, f7, f8, f9; non-linear features: f10, f11, f12, f13, f14, f15, f16, f17, f18) in theta rhythm (A), alpha rhythm (B), beta rhythm (C), and gamma rhythm (D), respectively.
Figure 7The constructed correlation matrices of the negative emotion group and the positive emotion group when the 18 different features in different rhythms are adopted.
Figure 8The comparison of the mean global correlation between the groups of negative emotion and positive emotion when the 18 different features in theta rhythm (A), alpha rhythm (B), beta rhythm (C), and gamma rhythm (D) are adopted. (***p < 0.001, **p < 0.01, *p < 0.05).
The threshold scope that can significantly differentiate the clustering coefficients in groups of positive emotion and negative emotion (p < 0.05).
| Theta | No.1 | 0.34~0.65, 0.92~0.99 | No.10 | 0.01~0.36, 0.62~0.63, 0.69~0.71, 0.89~0.92 |
| Theta | No.2 | 0.01~0.66, 0.92~0.99 | No.11 | 0.14~0.47, 0.71~0.74, 0.84~0.87, 0.97~0.98 |
| Theta | No.3 | 0.01~0.66, 0.92~0.99 | No.12 | 0.01~0.35, 0.59~0.62, 0.77~0.78, 0.94~0.96 |
| Theta | No.4 | 0.01~0.66, 0.92~0.99 | No.13 | 0.01~0.20, 0.49~0.50 |
| Theta | No.5 | 0.01~0.39 | No.14 | 0.21~0.31 |
| Theta | No.6 | 0.01~0.24 | No.15 | 0.07~0.32, 0.40~0.49, 0.70~0.72, 0.95~0.96 |
| Theta | No.7 | 0.01~0.15 | No.16 | 0.01~0.34, 0.54~0.58 |
| Theta | No.8 | 0.01~0.65, 0.93~0.99 | No.17 | 0.01~0.42 |
| Theta | No.9 | 0.01~0.66, 0.93~0.99 | No.18 | 0.49~0.50, 0.90~0.98 |
| Alpha | No.1 | 0.01~0.58, 0.66~0.95 | No.10 | 0.01~0.08, 0.53~0.56, 0.67~0.72, 0.96~0.99 |
| Alpha | No.2 | 0.01~0.27, 0.49~0.60, 0.75~0.91 | No.11 | 0.01~0.32, 0.64~0.76, 0.95~0.99 |
| Alpha | No.3 | 0.01~0.41, 0.56~0.66, 0.83~0.95 | No.12 | 0.01~0.08, 0.23~0.32, 0.67~0.76, 0.95~0.99 |
| Alpha | No.4 | 0.01~0.40, 0.47~0.59, 0.90~0.93 | No.13 | 0.01~0.05, 0.28~0.32, 0.67~0.78, 0.95~0.99 |
| Alpha | No.5 | 0.01~0.29, 0.50~0.52 | No.14 | 0.01~0.25 |
| Alpha | No.6 | 0.01~0.25, 0.48~0.73 | No.15 | 0.01~0.17, 0.24~0.28, 0.38~0.40, 0.89~0.92 |
| Alpha | No.7 | 0.01~0.35, 0.46~0.48, 0.95~0.97 | No.16 | 0.01~0.14, 0.36~0.39 |
| Alpha | No.8 | 0.01~0.36, 0.49~0.70, 0.79~0.99 | No.17 | 0.01~0.05, 0.64~0.66, 0.90~0.99 |
| Alpha | No.9 | 0.01~0.35, 0.82~0.85, 0.95~0.98 | No.18 | 0.01~0.06, 0.65~0.78, 0.95~0.99 |
| Beta | No.1 | 0.01~0.44, 0.53~0.58, 0.90~0.92 | No.10 | 0.01~0.13, 0.68~0.69 |
| Beta | No.2 | 0.01~0.47 | No.11 | 0.01~0.32, 0.42~0.43, 0.46~0.54, 0.67~0.80 |
| Beta | No.3 | 0.01~0.49, 0.85~0.86 | No.12 | 0.01~0.26, 0.91~0.98 |
| Beta | No.4 | 0.01~0.49, 0.53~0.56, 0.85~0.86 | No.13 | 0.01~0.34, 0.48~0.54, 0.87~0.92 |
| Beta | No.5 | 0.01~0.16, 0.20~0.23, 0.48~0.61 | No.14 | 0.01~0.26, 0.38~0.43, 0.62~0.74, 0.81~0.87 |
| Beta | No.6 | 0.01~0.41, 0.55~0.64, 0.85~0.91 | No.15 | 0.01~0.21 |
| Beta | No.7 | 0.01~0.31, 0.67~0.71, 0.91~0.99 | No.16 | 0.01~0.21, 0.57~0.62, 0.81~0.83, 0.86~0.88 |
| Beta | No.8 | 0.01~0.40, 0.86~0.89, 0.96~0.99 | No.17 | 0.01~0.12, 0.67~0.69, 0.86~0.88, 0.95~0.99 |
| Beta | No.9 | 0.01~0.47 | No.18 | 0.01~0.29, 0.39~0.46, 0.74~0.88 |
| Gamma | No.1 | 0.14~0.27, 0.66~0.70, 0.76~0.81 | No.10 | 0.15~0.33, 0.40~0.64, 0.83~0.88 |
| Gamma | No.2 | 0.01~0.33, 0.85~0.98 | No.11 | 0.17~0.22, 0.28~0.29 |
| Gamma | No.3 | 0.01~0.17, 0.50~0.51 | No.12 | 0.26~0.49 |
| Gamma | No.4 | 0.01~0.47, 0.97~0.98 | No.13 | 0.23~0.54 |
| Gamma | No.5 | 0.01~0.31, 0.73~0.98 | No.14 | 0.01~0.35, 0.97~0.99 |
| Gamma | No.6 | 0.01~0.30, 0.80~0.99 | No.15 | 0.01~0.22, 0.42~0.47, 0.72~0.74, 0.83~0.88 |
| Gamma | No.7 | 0.01~0.35, 0.93~0.99 | No.16 | 0.01~0.05, 0.81~0.84 |
| Gamma | No.8 | 0.48~0.61, 0.98~0.99 | No.17 | 0.01~0.49, 0.97~0.99 |
| Gamma | No.9 | 0.01~0.32, 0.90~0.91, 0.95~0.99 | No.18 | 0.01~0.17, 0.21~0.32, 0.40~0.42 |
Figure 9(A,B) The topographic plot of the clustering coefficient of the groups of negative emotion and positive emotion when feature No. 5 (Hjorth parameter: mobility) and feature No. 6 (Hjorth parameter: complexity) in beta rhythm were utilized. Conditions with different thresholds (T) are illustrated.
Pseudo Code for Recursive Feature Elimination (RFE) Algorithm
| Training set: |
| Feature set: |
| Ranking method: |
| Desired feature number: |
| Number of feature to eliminate in each step: |
| Final ranking feature set: |
| Final selected feature set: |
| 1 |
| 2 Steps: S = (p-q)/k; |
| 3 |
| 8 |