| Literature DB >> 35735474 |
Yuqi Wang1,2, Lijun Zhang1,2, Pan Xia2,3, Peng Wang2,3, Xianxiang Chen2,3, Lidong Du2,3, Zhen Fang2,3,4, Mingyan Du5.
Abstract
Emotion recognition is receiving significant attention in research on health care and Human-Computer Interaction (HCI). Due to the high correlation with emotion and the capability to affect deceptive external expressions such as voices and faces, Electroencephalogram (EEG) based emotion recognition methods have been globally accepted and widely applied. Recently, great improvements have been made in the development of machine learning for EEG-based emotion detection. However, there are still some major disadvantages in previous studies. Firstly, traditional machine learning methods require extracting features manually which is time-consuming and rely heavily on human experts. Secondly, to improve the model accuracies, many researchers used user-dependent models that lack generalization and universality. Moreover, there is still room for improvement in the recognition accuracies in most studies. Therefore, to overcome these shortcomings, an EEG-based novel deep neural network is proposed for emotion classification in this article. The proposed 2D CNN uses two convolutional kernels of different sizes to extract emotion-related features along both the time direction and the spatial direction. To verify the feasibility of the proposed model, the pubic emotion dataset DEAP is used in experiments. The results show accuracies of up to 99.99% and 99.98 for arousal and valence binary classification, respectively, which are encouraging for research and applications in the emotion recognition field.Entities:
Keywords: convolutional neural network; electroencephalogram; emotion recognition; machine learning
Year: 2022 PMID: 35735474 PMCID: PMC9219701 DOI: 10.3390/bioengineering9060231
Source DB: PubMed Journal: Bioengineering (Basel) ISSN: 2306-5354
Figure 1Plutchik emotion wheel.
Figure 2The Arousal-Valence model.
Figure 3Experiment process.
Figure 4The self–assessment scales.
Figure 5Process of emotion recognition using EEG.
Figure 6Dropout layer diagram.
Figure 7The process of Max pooling.
The value or types of the proposed model’s hyper-parameters.
| Hyper-Parameter of the Proposed Model | Value/Type |
|---|---|
| Batch size | 128 |
| Learning rate | 0.0001 |
| Momentum | 0.9 |
| Dropout | 0.25 |
| Number of epochs | 200 |
| Pooling layer | Max pooling |
| Activation function | LeakyReLU |
Figure 8The architecture of the proposed network.
The shapes of the proposed model.
| Number of Layers | Layer Type | Numbers of |
|---|---|---|
| 1 | Input (shape:1, 384, 32) | |
| 2 | conv_1 (Conv2d) | 1/25 (kernel size: 5 × 1) |
| 3 | droputout1 (Dropout=0.25) | 1/25 |
| 4 | conv_2 (Conv2d) | 25/25 (kernel size: 1 × 3, stride = (1,2)) |
| 5 | bn1 (BatchNorm2d) | 25 |
| 6 | pool1 (MaxPool2d (2,1)) | 25/25 |
| 7 | conv_3 (Conv2d) | 25/50 (kernel size: 5 × 1) |
Figure 9Model accuracy and loss curves (a) accuracy in Valence, (b) loss in Valence.
Figure 10Model accuracy and loss curves (a) accuracy in Arousal, (b) loss in Arousal.
Figure 11Model accuracy and loss curves during the first 25 epochs. (a) histogram, (b) line chart.
Comparison with other studies that used the DEAP dataset.
| Author | Accuracies (%) |
|---|---|
| Chao et al. [ | Val:68.28, Aro:66.73 (deep learning) |
| Pandey and Seeja [ | Val:62.5, Aro:61.25 (deep learning) |
| Islam and Ahmad [ | Val:81.51, Aro:79.42 (deep learning) |
| Alazrai et al. [ | Val:75.1, Aro:73.8 (traditional machine learn) |
| Alhagry et al. [ | Val:85.45, Aro:85.65 (deep learning) |
Figure 12(a–c) Comparison in the form of histograms.