| Literature DB >> 35280845 |
Guangcheng Bao1, Kai Yang1, Li Tong1, Jun Shu1, Rongkai Zhang1, Linyuan Wang1, Bin Yan1, Ying Zeng1,2.
Abstract
Electroencephalography (EEG)-based emotion computing has become one of the research hotspots of human-computer interaction (HCI). However, it is difficult to effectively learn the interactions between brain regions in emotional states by using traditional convolutional neural networks because there is information transmission between neurons, which constitutes the brain network structure. In this paper, we proposed a novel model combining graph convolutional network and convolutional neural network, namely MDGCN-SRCNN, aiming to fully extract features of channel connectivity in different receptive fields and deep layer abstract features to distinguish different emotions. Particularly, we add style-based recalibration module to CNN to extract deep layer features, which can better select features that are highly related to emotion. We conducted two individual experiments on SEED data set and SEED-IV data set, respectively, and the experiments proved the effectiveness of MDGCN-SRCNN model. The recognition accuracy on SEED and SEED-IV is 95.08 and 85.52%, respectively. Our model has better performance than other state-of-art methods. In addition, by visualizing the distribution of different layers features, we prove that the combination of shallow layer and deep layer features can effectively improve the recognition performance. Finally, we verified the important brain regions and the connection relationships between channels for emotion generation by analyzing the connection weights between channels after model learning.Entities:
Keywords: convolutional neural networks (CNN); electroencephalography (EEG); emotion recognition; graph convolutional neural networks (GCNN); style-based recalibration module (SRM)
Year: 2022 PMID: 35280845 PMCID: PMC8907537 DOI: 10.3389/fnbot.2022.834952
Source DB: PubMed Journal: Front Neurorobot ISSN: 1662-5218 Impact factor: 2.650
Figure 1The overall architecture of the MDGCN-SRCNN model consists of four blocks: graph construction block, graph convolutional block, SRM-based convolutional block, and classification block. The output of the model is a predicted label with probability.
Figure 2SRM module. This module is mainly composed of two parts: style pooling and style integration. AvgPool refers to global average pooling, StdPool refers to global standard deviation pooling; CFC refers to the channel fully connected layer; BN refers to batch standardization.
The training process of MDGCN-SRCNN.
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
MDGCN-SRCNN architecture.
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|
| Graph convolution | Input | ( | ||||
| GCN1 | ( | ( | Leaky_ReLU | |||
| Global_add_pool | ( | 16 | ||||
| GCN2 | ( | ( | Leaky_ReLU | |||
| Global_add_pool | ( | 64 | ||||
| SMR-based convolution | Reshape | 64 | (8,8,1) | |||
| Conv1 | (2,2) | 2 | (8,8,1) | (7,7,16) | Leaky_ReLU | |
| SMR1 | (7,7,16) | (7,7,16) | Sigmoid | |||
| Conv2 | (2,2) | 2 | (7,7,16) | (6,6,32) | Leaky_ReLU | |
| SMR1 | (6,6,32) | (6,6,32) | Sigmoid | |||
| Max_pool | (2,2) | (6,6,32) | (3,3,32) | |||
| Classifier | Reshape | (3,3,32) | 3*3*32 | |||
| FC1 | 16+64+3*3*32 | 256 | Leaky_ReLU | |||
| FC2 | 256 | 128 | Leaky_ReLU | |||
| FC3 | 128 | C | Softmax |
Compare the accuracy rate (mean/std) with different existing methods on the SEED data set.
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|
| SVM (Zheng and Lu, | 60.50/14.14 | 60.95/10.20 | 66.64/14.41 | 80.76/115.6 | 79.56/11.38 | 83.99/9.72 |
| GSCCA (Zheng, | 63.92/11.16 | 64.64/10.33 | 70.10/14.76 | 76.93/11.00 | 77.98/10.72 | 82.96/9.95 |
| DBN (Zheng and Lu, | 64.32/12.45 | 60.77/10.42 | 64.01/15.97 | 78.92/12.48 | 79.19/14.58 | 86.08/8.34 |
| STRNN (Zhang et al., |
|
|
| 83.41/10.16 | 69.61/15.65 | 89.50/7.63 |
| GCNN (Song et al., | 72.75/10.85 | 74.40/8.23 | 73.46/12.17 | 83.24/9.93 | 83.36/9.43 | 87.40/9.20 |
| DGCNN (Song et al., | 74.25/11.42 | 71.52/5.99 | 74.43/12.16 | 83.65/10.17 | 85.73/10.64 | 90.40/8.49 |
| BiDANN (Li Y. et al., | 76.97/10.95 | 75.56/7.88 | 81.03/11.74 |
| 88.64/9.46 | 92.38/7.04 |
| GCB-net (Zhang T. et al., | 80.38/10.04 | 76.09/7.54 | 81.36/11.44 | 88.05/9.84 | 88.45/9.67 | 92.30/7.40 |
| GCB-net+BLS (Zhang T. et al., | 79.98/8.93 | 76.51/9.56 | 81.97/11.05 | 89.06/8.69 | 89.10/9.55 | 94.24/6.70 |
| RGNN (Zhong et al., | 76.17/7.91 | 72.26/7.25 | 75.33/8.85 | 84.25/12.54 |
| 94.24/ |
| MDGCN-SRCNN | 77.73/10.23 | 77.27/9.38 | 80.47/13.22 | 87.59/12.13 | 89.02/9.13 |
Bold represents the best result.
The accuracy of the proposed method is compared with the existing methods on the SEED-IV dataset.
|
|
|
|
|---|---|---|
| SVM (Zhong et al., | 56.61 | 20.05 |
| DBN (Zhong et al., | 66.77 | 7.38 |
| GSCCA (Zheng, | 69.08 | 16.66 |
| DGCNN(Zhong et al., | 69.88 | 16.29 |
| BiDANN (Li Y. et al., | 70.29 | 12.63 |
| EmotionMeter (Zheng et al., | 70.58 | 17.01 |
| BiHDM (Li et al., | 74.35 | 14.09 |
| RGNN (Zhong et al., | 79.37 | 10.54 |
| SST-EmotionNet (Jia et al., | 84.92 |
|
| MDGCN-SRCNN |
| 11.58 |
Bold represents the best result.
Figure 3Confusion matrix of different data sets. (A) is the confusion matrix of the SEED data set; (B) is the confusion matrix of the SEED-IV data set.
Figure 4Visualization of t-SNE output from different layers. (A,F) are the original data; (B,G) are the feature distributions output by the first layer of GCN; (C,H) are the feature distributions output by the second layer of GCN; (D,I) is the feature distribution of the output of the convolutional neural network; (E,J) are the feature distributions after connecting the two layers of GCN and SRCNN. Different colors represent different emotions.
Figure 5The connection weights between the first 10 channels are selected from the initial adjacency matrix and the learned adjacency matrix. (A) is the initial adjacency matrix of the SEED data set, (B) is the adjacency matrix learned from SEED dataset. (C) is the initial adjacency matrix of the SEED-IV data set, (D) is the adjacency matrix learned from SEED-IV dataset.
The SEED data set and SEED-IV data set are compared by using different adjacency matrix A initialization methods.
|
|
|
| ||
|---|---|---|---|---|
|
|
|
|
| |
| PCC |
|
|
| 11.58 |
| RGNN | 91.98 | 7.21 | 84.16 |
|
| PLV | 92.04 | 7.56 | 80.92 | 13.48 |
| Random | 91.83 | 8.47 | 82.39 | 11.74 |
Bold represents the best result.
The results of ablation experiments on SEED and SEED-IV (mean/std), “~” represents the module is removed.
|
|
|
|
|---|---|---|
| MDGCN-SRCNN | ||
| ~SRM | 93.36/6.49 | 83.63/10.20 |
| ~SRCNN | 91.38/7.74 | 81.15/10.89 |
| One-layer GCN | 89.72/6.52 | 79.73/9.61 |
Bold represents the best result.