| Literature DB >> 29891829 |
Mouhannad Ali1, Fadi Al Machot2, Ahmad Haj Mosa3, Midhat Jdeed4, Elyan Al Machot5, Kyandoghere Kyamakya6.
Abstract
Machine learning approaches for human emotion recognition have recently demonstrated high performance. However, only/mostly for subject-dependent approaches, in a variety of applications like advanced driver assisted systems, smart homes and medical environments. Therefore, now the focus is shifted more towards subject-independent approaches, which are more universal and where the emotion recognition system is trained using a specific group of subjects and then tested on totally new persons and thereby possibly while using other sensors of same physiological signals in order to recognize their emotions. In this paper, we explore a novel robust subject-independent human emotion recognition system, which consists of two major models. The first one is an automatic feature calibration model and the second one is a classification model based on Cellular Neural Networks (CNN). The proposed system produces state-of-the-art results with an accuracy rate between 80% and 89% when using the same elicitation materials and physiological sensors brands for both training and testing and an accuracy rate of 71.05% when the elicitation materials and physiological sensors brands used in training are different from those used in training. Here, the following physiological signals are involved: ECG (Electrocardiogram), EDA (Electrodermal activity) and ST (Skin-Temperature).Entities:
Keywords: cellular neural networks (CNN); classification; dynamic calibration; emotion recognition; physiological signals
Mesh:
Year: 2018 PMID: 29891829 PMCID: PMC6021954 DOI: 10.3390/s18061905
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Literature review on emotion recognition using physiological and speech signals.
| Ref. No. | Signals | Features | Classifiers | Emotion Parameters | Stimuli | No. of Subjects | Accuracy in % |
|---|---|---|---|---|---|---|---|
| [ | EMG | Statistical, Energy, Sub band Spectrum, Entropy | Linear Discriminant Analysis | Joy, Anger, Sad, Pleasure | Music | 3 , MITdatabase | 95 (Subject-Dependent) |
| [ | EDA | No specific features stated | KNN, Discriminant Function Analysis, Marquardt backpropagation | Sadness, Anger, Fear, Surprise, Frustration, Amusement | Movies | 14 | 91.7 (Subject-Dependent) |
| [ | EMG | Running mean Running standard deviation Slope | NN | Arousal, Valance | IAPS (Visual Affective Picture System) | 1 | 96.58 Arousal |
| [ | ECG | Fast Fourier | Tabu Search | Joy, Sadness | Movies | 154 | 86 (Subject-Independent) |
| [ | EDA | No specific features stated | fuzzy logic | Stress | Hyperventilation | 80 | 99.5 (Subject-Independent) |
| [ | BVP | Statistical Features | SVM, Fisher LDA | Amusement, Contentment, Disgust, Fear, Sad, Neutral | IAPS | 10 | 90 (Subject-Dependent) |
| [ | EMG | Statistical Features, BRV, Zero-crossing, MFCCs | KNN | Arousal, Valance | Quiz dataset | 3 | 92 (Subject-Dependent) |
| [ | EDA | No specific features stated | HMM | Arousal, Valance | Robot Actions | 36 | 81 (Subject-Dependent) |
| [ | EDA | Statistical Features average power SCL SCR | CNN | Arousal, Valance | Movies | 10 | 82.35 (Subject-Independent) |
EMG: Electromyography; ECG: Electrocardiography; EDA: Electrodermal Activity; RSP: Respiration; ST: Skin Temperature; EEG: Electroencephalogram; BVP: Blood Volume Pulse; HR: Heart Rate; KNN: k-nearest neighbors algorithm; SVM: Support vector machine; HMM: Hidden Markov Model; ANN: Artificial Neural Network; CNN: Cellular Neural Network.
Figure 1The general architecture of the proposed emotion recognition system.
Figure 2The Cellular Neural Network (CNN) classification model (SimulinkModel A).
Figure 3The multi-CNN modal emotional state estimation (SimulinkModel B).
The configuration parameters of the involved classifiers.
| Classifier | Type | Parameters |
|---|---|---|
| RBSVM [ | C-SVC | KernelType= radial basis function, eps= 0.001, gamma= 0.0001 |
| NB [ | NaiveBayes -k | UseKernelEstimator= True |
| KNN [ | Default |
|
| ANN [ | Multilayer Perceptron |
|
| CNN | Echo State |
|
The recognition accuracy in % with respect to different signals combinations.
| Physiological Sensor | KNN | NB | ANN | SVM | CNN | |
|---|---|---|---|---|---|---|
| Single sensor | ECG | 61.2 | 53.58 | 53.92 | 62.91 | 56.41 |
| EDA | 63.73 | 53.1 | 60.32 | 68.4 | 75.34 | |
| ST | 33.12 | 35.7 | 42.64 | 41.8 | 42.6 | |
| Multi sensors | EDA + ECG | 71.12 | 55.4 | 60.78 | 72.64 | 83.43 |
| ST + ECG | 68.45 | 54.53 | 55.86 | 70 | 68.63 | |
| ST + EDA | 69.13 | 55.34 | 58.43 | 69.64 | 78.5 | |
| ST + EDA + ECG | 76.88 | 56.88 | 62.5 | 77.5 | 89.38 |
Performance measures in percentage while using the reference database MAHNOB for both training and testing data (subject-independent evaluation).
| Measure | KNN | NB | ANN | SVM | CNN |
|---|---|---|---|---|---|
| Accuracy | 76.88% | 56.88% | 62.5% | 77.5% | 89.38% |
| Specificity | 95% | 86.67% | 85.84% | 95% | 97.5% |
| Precision | 81.82% | 57.9% | 57.5% | 82.86% | 92.11% |
| Recall | 67.5% | 55% | 57.5% | 72.5% | 87.5% |
Performance measures in percentage while using our experiment for both training and testing data (subject-independent evaluation).
| Measure | KNN | NB | ANN | SVM | CNN |
|---|---|---|---|---|---|
| Accuracy | 56.25% | 25.63% | 45.63% | 71.88% | 81.88% |
| Specificity | 80% | 72.5% | 83.34% | 89.17% | 95% |
| Precision | 50% | 25% | 42.86% | 67.5% | 82.86% |
| Recall | 60% | 27% | 37.5% | 67.5% | 72.5% |
The performance accuracy in percentage while using the MAHNOB reference database for training and data from our experiment for testing (without calibration model).
| Subject | KNN | NB | ANN | SVM | CNN |
|---|---|---|---|---|---|
| Subject1 | 30.54% | 26.43% | 31.21% | 14.31% | 58.76% |
| Subject2 | 44.15% | 22.23% | 30.76% | 20.87% | 60% |
| Subject3 | 32.25% | 13.15% | 34.89% | 25.90% | 54.38% |
| Subject4 | 29.64% | 19.22% | 28.33% | 20.33% | 57.5% |
| Subject5 | 29.90% | 25.21% | 12.85% | 22.58% | 59.38% |
| Subject6 | 27.45% | 10.89% | 25.07% | 17.32% | 53.13% |
|
| 32.33% | 19.53% | 27.18% | 20.22% |
|
The performance accuracy in percentage while using the MAHNOB reference database for training and data from our experiment for testing (using calibration model).
| Subject | KNN | NB | ANN | SVM | CNN |
|---|---|---|---|---|---|
| Subject1 | 44.58% | 31.88% | 55.63% | 48.54% | 70.63% |
| Subject2 | 53.93% | 29.12% | 58.38% | 50.82% | 81.26% |
| Subject3 | 48.62% | 40.19% | 60.45% | 42.73% | 71.88% |
| Subject4 | 35.55% | 27.45% | 33.77% | 44.21% | 72.5% |
| Subject5 | 24.29% | 30.83% | 22.36% | 43.16% | 63.13% |
| Subject6 | 25.23% | 23.28% | 30.83% | 26.87% | 66.88% |
|
| 38.7% | 30.46% | 43.57% | 42.73% |
|
The performance measures of the CNN model in percentage before and after involving the calibration model.
| Measure | CNN without Calibration Model | CNN with Calibration Model |
|---|---|---|
| Accuracy | 57.19% | |
| Specificity | 84.31% | |
| Precision | 55.86% | |
| Recall | 59.59% |