| Literature DB >> 29056963 |
Yongrui Huang1, Jianhao Yang1, Pengkai Liao1, Jiahui Pan1.
Abstract
This paper proposes two multimodal fusion methods between brain and peripheral signals for emotion recognition. The input signals are electroencephalogram and facial expression. The stimuli are based on a subset of movie clips that correspond to four specific areas of valance-arousal emotional space (happiness, neutral, sadness, and fear). For facial expression detection, four basic emotion states (happiness, neutral, sadness, and fear) are detected by a neural network classifier. For EEG detection, four basic emotion states and three emotion intensity levels (strong, ordinary, and weak) are detected by two support vector machines (SVM) classifiers, respectively. Emotion recognition is based on two decision-level fusion methods of both EEG and facial expression detections by using a sum rule or a production rule. Twenty healthy subjects attended two experiments. The results show that the accuracies of two multimodal fusion detections are 81.25% and 82.75%, respectively, which are both higher than that of facial expression (74.38%) or EEG detection (66.88%). The combination of facial expressions and EEG information for emotion recognition compensates for their defects as single information sources.Entities:
Mesh:
Year: 2017 PMID: 29056963 PMCID: PMC5625811 DOI: 10.1155/2017/2107451
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1Data processing procedure of the multimodal emotion recognition.
Figure 2The architecture of the proposed system for face expression classification: the network has one hidden layer with 200 neurons. The input of this network is 169 image features we get from dimensionality reduction, while the output is the scores of four emotion states (happiness, neutral, sadness, and fear). The learning rate of this network is 0.1. We use sigmoid function as the activation function of this network.
The production rules of combining the emotion state and intensity level.
|
|
|
|
|---|---|---|
|
| (Happiness, strong) | Happiness |
|
| (Happiness, moderate) | Happiness |
|
| (Happiness, weak) | Neutral |
|
| (Neutral, strong) | Happiness |
|
| (Neutral, moderate) | Neutral |
|
| (Neutral, weak) | Neutral |
|
| (Sadness, strong) | Fear |
|
| (Sadness, moderate) | Sadness |
|
| (Sadness, weak) | Sadness |
|
| (Fear, strong) | Fear |
|
| (Fear, moderate) | Fear |
|
| (Fear, weak) | Sadness |
Figure 3Example screenshots of face videos from Experiment 2.
The accuracies for the detections of face expression, EEG, and two fusion methods.
| Subject | Face expression | EEG | The first fusion method (online) | The second fusion method (offline) |
|---|---|---|---|---|
| 1 | 92.5 | 67.5 | 92.5 | 87.5 |
| 2 | 57.5 | 70.0 | 82.5 | 87.5 |
| 3 | 50.0 | 72.5 | 87.5 | 87.5 |
| 4 | 62.5 | 75.0 | 92.5 | 87.5 |
| 5 | 60.0 | 60.0 | 75.0 | 75.0 |
| 6 | 87.5 | 75.0 | 95.0 | 92.5 |
| 7 | 72.5 | 72.5 | 72.5 | 80.0 |
| 8 | 70.0 | 70.0 | 80.0 | 87.5 |
| 9 | 92.5 | 60.0 | 75.0 | 80.0 |
| 10 | 85.0 | 62.5 | 72.5 | 80.0 |
| 11 | 67.5 | 72.5 | 80.0 | 80.0 |
| 12 | 80.0 | 75.0 | 85.0 | 85.0 |
| 13 | 92.5 | 57.5 | 92.5 | 87.5 |
| 14 | 72.5 | 55.0 | 77.5 | 80.0 |
| 15 | 70.0 | 52.5 | 75.0 | 77.5 |
| 16 | 92.5 | 62.5 | 77.5 | 77.5 |
| 17 | 77.5 | 57.5 | 90.0 | 87.5 |
| 18 | 92.5 | 80.0 | 92.5 | 85.0 |
| 19 | 50.0 | 62.5 | 60.0 | 75.0 |
| 20 | 62.5 | 77.5 | 70.0 | 75.0 |
|
| ||||
|
|
|
|
|
|