| Literature DB >> 33286283 |
Lizheng Pan1,2, Zeming Yin1, Shigang She1, Aiguo Song2.
Abstract
Emotion recognition realizing human inner perception has a very important application prospect in human-computer interaction. In order to improve the accuracy of emotion recognition, a novel method combining fused nonlinear features and team-collaboration identification strategy was proposed for emotion recognition using physiological signals. Four nonlinear features, namely approximate entropy (ApEn), sample entropy (SaEn), fuzzy entropy (FuEn) and wavelet packet entropy (WpEn) are employed to reflect emotional states deeply with each type of physiological signal. Then the features of different physiological signals are fused to represent the emotional states from multiple perspectives. Each classifier has its own advantages and disadvantages. In order to make full use of the advantages of other classifiers and avoid the limitation of single classifier, the team-collaboration model is built and the team-collaboration decision-making mechanism is designed according to the proposed team-collaboration identification strategy which is based on the fusion of support vector machine (SVM), decision tree (DT) and extreme learning machine (ELM). Through analysis, SVM is selected as the main classifier with DT and ELM as auxiliary classifiers. According to the designed decision-making mechanism, the proposed team-collaboration identification strategy can effectively employ different classification methods to make decision based on the characteristics of the samples through SVM classification. For samples which are easy to be identified by SVM, SVM directly determines the identification results, whereas SVM-DT-ELM collaboratively determines the identification results, which can effectively utilize the characteristics of each classifier and improve the classification accuracy. The effectiveness and universality of the proposed method are verified by Augsburg database and database for emotion analysis using physiological (DEAP) signals. The experimental results uniformly indicated that the proposed method combining fused nonlinear features and team-collaboration identification strategy presents better performance than the existing methods.Entities:
Keywords: emotion recognition; identification strategy; nonlinear features; physiological signals
Year: 2020 PMID: 33286283 PMCID: PMC7517002 DOI: 10.3390/e22050511
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
The content summary of Augsburg Database.
| Elicitation Material | Music |
|---|---|
| Emotional states | Joy, anger, sadness, pleasure |
| Number of subjects | 1 |
| Collected signals | ECG, EMG, RSP, SC |
| Length | 120 seconds |
| Sampling frequency | ECG: 256 Hz; EMG, RSP and SC: 32 Hz; |
| Collected days | 25 |
The preprocessed DEAP database content summary.
| Elicitation Material | Videos |
|---|---|
| Emotion labels | Arousal, valence |
| Number of subjects | 32 |
| Collected signals | EEG, EOG, GSR, BVP, RSP, EMG, SKT |
| Length | 60 seconds |
| Sampling frequency | 128 Hz; |
| Rating values | Arousal: 1–9 |
The number of samples for four dimensional emotional states of each subject.
| Subject | HVHA | HVLA | LVLA | LVHA | Total |
|---|---|---|---|---|---|
| s01 | 130 | 60 | 100 | 110 | 400 |
| s02 | 160 | 60 | 100 | 80 | 400 |
| s03 | 10 | 210 | 110 | 70 | 400 |
| s04 | 120 | 40 | 200 | 40 | 400 |
| s05 | 130 | 110 | 100 | 60 | 400 |
| Total | 550 | 480 | 610 | 360 | 2000 |
Figure 1Emotion model classification based on the scale of valence and arousal.
Figure 2The structure diagram of decision tree (DT).
Figure 3The network structure of Extreme Learning Machine (ELM).
Figure 4The flow chart of SVM-DT-ELM algorithm.
The hardware and framework specifications.
| CPU | Intel Core i7-8750H |
|---|---|
| GPU | NVIDIA GeForce GTX1050Ti 4GB |
| OS | Windows 10 |
| RAM | DDR4 16GB |
| Frameworks | MATLAB (R2015b) |
Figure 5The architecture of emotion recognition with fused features and team-collaboration identification strategy.
Comparison of the performances between single signal features and feature fusion of multitype signals.
| Physiological Sensor | Acc*(%) | |||
|---|---|---|---|---|
| SVM | DT | ELM | ||
| Single sensor | ECG | 65.7 ± 1.55 | 62.1 ± 1.92 | 58.4 ± 2.72 |
| EMG | 72.1 ± 0.97 | 60.1 ± 1.84 | 62.3 ± 2.47 | |
| RSP | 66.4 ± 1.22 | 66.2 ± 1.75 | 59.7 ± 2.62 | |
| SC | 70.9 ± 1.43 | 64.5 ± 1.54 | 60.6 ± 2.33 | |
| Multi sensors | ECG + EMG + RSP + SC | 95.5 ± 0.85 | 90.5 ± 1.27 | 89.4 ± 1.78 |
An overview of the comparison of the classification accuracy using different classifiers.
| Number of Experiments | Classification Methods (Acc*/%) | |||
|---|---|---|---|---|
| SVM | DT | ELM | SVM-DT-ELM | |
| 1 | 96 | 91 | 86 | 98 |
| 2 | 96 | 91 | 89 | 98 |
| 3 | 95 | 92 | 89 | 98 |
| 4 | 95 | 89 | 91 | 99 |
| 5 | 94 | 90 | 88 | 98 |
| 6 | 95 | 90 | 90 | 98 |
| 7 | 95 | 93 | 92 | 99 |
| 8 | 96 | 89 | 90 | 100 |
| 9 | 96 | 90 | 88 | 99 |
| 10 | 97 | 90 | 91 | 99 |
| Acc* (%) | 95.5 ± 0.85 | 90.5 ± 1.27 | 89.4 ± 1.78 | 98.6 ± 0.70 |
Accuracy comparison of various studies.
| Classification Method | Feature Dimension | Acc* (%) |
|---|---|---|
| LDF [ | 32 | 92.05 |
| SVM [ | 64 | 95 |
| PSO-SNC [ | 32 | 86 |
| SVM [ | 28 | 76 |
| C4.5 DT [ | 155 | 93 |
| This paper | 16 | 98.6 |
The classification accuracy for each subject using single signal and multimodal signals with SVM.
| Subject | Physiological Sensors | ||||
|---|---|---|---|---|---|
| Single Sensor (Acc*/%) | Multi Sensors (Acc*/%) | ||||
| GSR | RSP | BVP | EMG | GSR + RSP + EMG + BVP | |
| s01 | 38.0 ± 1.46 | 43.3 ± 2.73 | 53.5 ± 3.06 | 50.6 ± 1.56 | 73.5 ± 2.07 |
| s02 | 34.0 ± 2.25 | 53.1 ± 1.73 | 43.2 ± 1.67 | 52.1 ± 3.03 | 65.1 ± 2.69 |
| s03 | 54.3 ± 1.51 | 60.8 ± 2.72 | 65.6 ± 1.83 | 63.2 ± 2.33 | 81.5 ± 1.35 |
| s04 | 48.6 ± 4.13 | 52.3 ± 3.67 | 56.5 ± 3.99 | 56.1 ± 2.17 | 62.7 ± 2.21 |
| s05 | 32.8 ± 2.56 | 43.1 ± 3.88 | 47.6 ± 3.69 | 42.5 ± 3.04 | 69.2 ± 2.70 |
The classification accuracy for each subject using single signal and multimodal signals with DT.
| Subject | Physiological Sensors | ||||
|---|---|---|---|---|---|
| Single Sensor (Acc*/%) | Multi Sensors (Acc*/%) | ||||
| GSR | RSP | BVP | EMG | GSR + RSP + EMG + BVP | |
| s01 | 32.8 ± 2.95 | 39.8 ± 3.11 | 41.6 ± 4.34 | 46.4 ± 2.41 | 60.3 ± 1.95 |
| s02 | 35.6 ± 1.82 | 33.0 ± 2.23 | 50.0 ± 2.35 | 44.2 ± 3.70 | 53.3 ± 2.36 |
| s03 | 40.8 ± 1.30 | 49.4 ± 2.70 | 55.2 ± 0.84 | 57.4 ± 1.82 | 60.3 ± 2.00 |
| s04 | 40.8 ± 2.17 | 40.2 ± 3.03 | 40.4 ± 4.16 | 45.4 ± 1.95 | 59.4 ± 2.01 |
| s05 | 28.6 ± 1.52 | 41.4 ± 2.40 | 37.2 ± 1.48 | 41.6 ± 2.30 | 52.7 ± 2.83 |
The classification accuracy for each subject using single signal and multimodal signals with ELM.
| Subject | Physiological Sensors | ||||
|---|---|---|---|---|---|
| Single Sensor (Acc*/%) | Multi Sensors (Acc*/%) | ||||
| GSR | RSP | BVP | EMG | GSR + RSP + EMG + BVP | |
| s01 | 29.2 ± 1.64 | 47.0 ± 2.55 | 46.8 ± 1.92 | 43.4 ± 3.65 | 61.5 ± 2.37 |
| s02 | 30.4 ± 1.67 | 34.4 ± 2.70 | 49.6 ± 3.50 | 42.6 ± 1.82 | 55.4 ± 2.37 |
| s03 | 40.0 ± 3.08 | 54.6 ± 3.36 | 54.4 ± 2.97 | 54.6 ± 2.30 | 62.9 ± 1.29 |
| s04 | 39.6 ± 2.70 | 47.8 ± 2.56 | 42.6 ± 3.21 | 45.8 ± 3.56 | 50.1 ± 2.28 |
| s05 | 34.8 ± 3.35 | 44.4 ± 2.97 | 39.8 ± 3.83 | 43.2 ± 3.11 | 53.9 ± 2.73 |
Comparison of the results using SVM, DT, ELM and proposed classification strategy for each subject.
| Subject | Method | The Identification Accuracy of Each Experiment (%) | Average (%) | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | |||
| s01 | DT | 64 | 61 | 58 | 60 | 58 | 59 | 59 | 60 | 62 | 62 | 60.3 ± 1.95 |
| ELM | 62 | 60 | 64 | 58 | 62 | 58 | 64 | 60 | 64 | 63 | 61.5 ± 2.37 | |
| SVM | 75 | 72 | 72 | 76 | 75 | 77 | 72 | 73 | 71 | 72 | 73.5 ± 2.07 | |
| Proposed | 80 | 80 | 80 | 81 | 80 | 82 | 79 | 79 | 76 | 78 | 79.5 ± 1.65 | |
| s02 | DT | 55 | 51 | 53 | 53 | 50 | 56 | 51 | 57 | 52 | 55 | 53.3 ± 2.36 |
| ELM | 58 | 54 | 59 | 52 | 53 | 56 | 53 | 57 | 57 | 55 | 55.4 ± 2.37 | |
| SVM | 68 | 66 | 71 | 63 | 63 | 64 | 64 | 66 | 63 | 63 | 65.1 ± 2.69 | |
| Proposed | 72 | 70 | 76 | 70 | 69 | 70 | 68 | 70 | 68 | 70 | 70.3 ± 2.31 | |
| s03 | DT | 62 | 58 | 59 | 63 | 62 | 61 | 57 | 59 | 60 | 62 | 60.3 ± 2.00 |
| ELM | 65 | 63 | 61 | 63 | 62 | 62 | 63 | 63 | 65 | 62 | 62.9 ± 1.29 | |
| SVM | 84 | 82 | 80 | 82 | 83 | 81 | 80 | 81 | 80 | 82 | 81.5 ± 1.35 | |
| Proposed | 88 | 88 | 87 | 89 | 86 | 86 | 87 | 86 | 87 | 86 | 87 ± 1.05 | |
| s04 | DT | 58 | 62 | 60 | 57 | 61 | 58 | 59 | 58 | 63 | 58 | 59.4 ± 2.01 |
| ELM | 48 | 51 | 49 | 47 | 53 | 50 | 47 | 52 | 53 | 51 | 50.1 ± 2.28 | |
| SVM | 63 | 66 | 65 | 64 | 65 | 60 | 61 | 60 | 62 | 61 | 62.7 ± 2.21 | |
| Proposed | 70 | 70 | 72 | 72 | 73 | 68 | 70 | 70 | 68 | 67 | 70 ± 1.94 | |
| s05 | DT | 50 | 52 | 52 | 55 | 50 | 51 | 59 | 51 | 52 | 55 | 52.7 ± 2.83 |
| ELM | 57 | 58 | 52 | 50 | 56 | 53 | 54 | 51 | 52 | 56 | 53.9 ± 2.73 | |
| SVM | 69 | 67 | 68 | 66 | 76 | 69 | 70 | 70 | 69 | 68 | 69.2 ± 2.70 | |
| Proposed | 75 | 73 | 78 | 73 | 82 | 75 | 75 | 74 | 75 | 75 | 75.5 ± 2.68 | |
| Overall average | DT | 57.2 ± 2.23 | ||||||||||
| Proposed | 76.46 ± 1.93 | |||||||||||
Figure 6Average accuracy of each subject using SVM, DT, ELM and proposed method for four types of emotional identification.
Accuracy comparison of various studies in two-dimensional classification.
| Method | Acc* (%) | |
|---|---|---|
| Arousal | Valence | |
| Chen et al. [ | 69.09 | 67.89 |
| Zhuang et al. [ | 71.9 | 69.1 |
| Yin et al. [ | 77.1 | 76.1 |
| Alazrai et al. [ | 86.6 | 85.8 |
| Kwon et al [ | 76.56 | 80.46 |
Accuracy comparison of various studies.
| Menthod | Emotions | Acc*(%) |
|---|---|---|
| M Zubair and C Yoon [ | HVHA, HVLA, | 49.7 |
| Alazrai et al [ | HVHA, HVLA, | 75.1 |
| Zheng et al [ | HVHA, HVLA, | 69.67 |
| This paper | HVHA, HVLA, | 76.46 |