| Literature DB >> 35890933 |
Xiaoliang Zhu1, Wenting Rong1, Liang Zhao1, Zili He1, Qiaolai Yang1, Junyi Sun1, Gendong Liu1.
Abstract
Understanding learners' emotions can help optimize instruction sand further conduct effective learning interventions. Most existing studies on student emotion recognition are based on multiple manifestations of external behavior, which do not fully use physiological signals. In this context, on the one hand, a learning emotion EEG dataset (LE-EEG) is constructed, which captures physiological signals reflecting the emotions of boredom, neutrality, and engagement during learning; on the other hand, an EEG emotion classification network based on attention fusion (ECN-AF) is proposed. To be specific, on the basis of key frequency bands and channels selection, multi-channel band features are first extracted (using a multi-channel backbone network) and then fused (using attention units). In order to verify the performance, the proposed model is tested on an open-access dataset SEED (N = 15) and the self-collected dataset LE-EEG (N = 45), respectively. The experimental results using five-fold cross validation show the following: (i) on the SEED dataset, the highest accuracy of 96.45% is achieved by the proposed model, demonstrating a slight increase of 1.37% compared to the baseline models; and (ii) on the LE-EEG dataset, the highest accuracy of 95.87% is achieved, demonstrating a 21.49% increase compared to the baseline models.Entities:
Keywords: EEG; attention; convolutional neural network; emotion recognition; learning emotions; multi-channel band features
Mesh:
Year: 2022 PMID: 35890933 PMCID: PMC9318779 DOI: 10.3390/s22145252
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1ECN-AF framework diagram.
Multi-channel convolutional backbone network construction.
| Stage | Stage Setting | Output |
|---|---|---|
| Conv-1 | 32, strides = 2, activation = “relu” | (1000,32) |
| Conv-2 | 64, strides = 2, activation = “relu” | (498,64) |
| Pool_1 | 2, AvgPool | (249,64) |
| Batch_norm1 | BatchNormalization | (249,64) |
| Drop_1 | Dropout1D | (249,64) |
Figure 2Band attention fusion unit.
Classification network construction.
| Stage | Stage Setting | Output |
|---|---|---|
| Conv-1 | 128, strides = 2, activation = “relu” | (245,128) |
| Conv-2 | 128, strides = 2, activation = “relu” | (245,128) |
| Pool_1 | 2, AvgPool | (122,128) |
| Batch_norm1 | BatchNormalization | (122,128) |
| Drop_1 | Dropout | (122,128) |
| Conv-3 | 256, strides = 2, activation = “relu” | (118,256) |
| Conv-4 | 256, strides = 2, activation = “relu” | (118,256) |
| Pool_2 | GlobalAvgPool | (256) |
| Drop_2 | Dropout | (256) |
| Dense | Activation = “softmax” | (3) |
Figure 35-point scale score of the subjects.
Figure 4Description statistics of the 28 target videos, with 0–4 ratings.
Figure 5Experimental flow of the LE-EEG dataset.
Accuracy comparison (i.e., ACC/STD) of different frequency bands (average 5-fold cross validation results).
| Frequency Band | SEED | LE-EEG | ||
|---|---|---|---|---|
| C1 | C2 | C3 | All_Band | |
| δ | 83.18/2.42 | 84.23/2.85 | 93.69/0.40 | 95.22/0.49 |
| θ | 67.05/7.71 | 69.88/7.52 | 93.06/0.45 | 94.64/1.15 |
| α | 77.55/6.82 | 82.68/5.58 | 93.09/1.11 | 94.64/0.63 |
| β | 81.46/7.27 | 87.09/4.17 | 93.56/0.44 | 94.97/0.51 |
| γ | 83.60/4.91 | 90.90/4.38 | 93.83/0.48 | 95.52/0.62 |
| β + γ | 84.14/6.12 | 92.10/4.02 | - | - |
| β × γ | 91.30/4.56 | 93.39/2.42 | - | - |
| Attention (β, γ) | 90.03/3.40 | 94.20/2.38 | - | - |
Figure 6Channel selection maps: (a) C2 on the SEED dataset; (b) C3 on the LE-EEG dataset.
Accuracy comparison (i.e., ACC/STD) of various fusion methods validated on SEED dataset (average 5-fold cross validation results).
| Method | C1 | C2 | ||||
|---|---|---|---|---|---|---|
| Add | Mult | Attention | Add | Mult | Attention | |
| α, β | 72.34/10.70 | 72.54/11.50 |
| 83.16/4.84 | 87.63/7.67 |
|
| α, γ | 69.48/12.10 | 78.84/10.22 |
| 80.56/8.80 |
| 90.77/4.59 |
| δ, β |
| 77.62/11.56 | 93.77/2.27 | 94.68/3.45 |
| 87.40/4.41 |
| δ, γ | 95.03/2.45 | 82.41/8.30 |
| 92.00/2.26 | 95.60/2.75 |
|
| β, γ | 84.14/6.12 |
| 90.03/3.40 | 92.10/4.02 | 93.39/2.42 |
|
| δ, α, β | 94.79/3.22 |
| 94.95/2.73 | 94.24/3.32 |
| 95.87/4.17 |
| θ, β, γ |
| 92.23/4.99 | 92.46/6.92 | 95.44/2.35 |
| 94.89/4.06 |
| α, β, γ | 92.70/5.52 |
| 93.84/3.63 | 95.31/3.21 | 94.66/5.43 |
|
| δ, β, γ | 95.17/2.17 | 95.13/3.67 |
| 95.78/3.45 | 96.15/2.13 |
|
| δ, α, β, γ |
| 87.07/12.96 | 77.0/16.81 |
| 80.99/14.82 | 86.49/17.90 |
Notably, Add means to directly add and fuse the features; Mult means that the features are multiplied and fused; Attention means that the attention fusion unit is used for feature-level fusion, and Bold indicates the best accuracy achieved using different fusion methods (for a given channel combination, C1 or C2).
Accuracy comparison (i.e., ACC/STD) versus baseline models (average 5-fold cross validation results).
| Method | SEED | LE-EEG |
|---|---|---|
| SVM [ | 83.30/--- | --- |
| DBN [ | 86.08/--- | --- |
| SOGNN [ | 86.81/5.79 | 74.38/1.50 |
| LDA [ | 90.93/--- | --- |
| DGCNN [ | 90.40/8.48 | --- |
| BiHDM [ | 93.12/6.06 | --- |
| TANN [ | 93.34/6.64 | --- |
| 3DCNN-BiLSTM [ | 93.38/2.66 | --- |
| 4D_CRNN [ | 94.08/2.55 | 67.48/0.39 |
| RGNN [ | 94.24/5.95 | --- |
| DE-CNN-BiLSTM [ | 94.82/--- | --- |
| DCCA [ | 95.08/6.42 | --- |
| ECN-AF (C1) | 95.32/3.53 | --- |
| ECN-AF (C2) |
| --- |
| ECN-AF (C3) | --- | 94.80/0.57 |
| ECN-AF (All_band) | 95.7/4.71 |
|
Dotted line (i.e., “---”) indicates that data was not provided; and bold indicates the best accuracy achieved for a given dataset.
Figure 7Accuracy of the model’s validation set.
Summary of methods of careless/insufficient effort (C/IE) detection.
| Index | Method | Type | Description |
|---|---|---|---|
| 1 | bogus or infrequency [ | check items | Odd items placed in scale to solicit particular responses. |
| 2 | long-string analysis [ | invariance | Length of longest sequential string of the same response |
| 3 | self-report data [ | self-report | Items which ask the participant how much effort they applied or how they judge the quality of their data |
| 4 | semantic antonyms/synonyms [ | consistency | Within-person correlations on sets of semantically matched pairs of items with opposite or similar meaning |
| 5 | instructional manipulation checks [ | check items | Items with extended instructions which include instructing participant to answer in unique manner |
| 6 | polytomous guttman errors [ | consistency | Count of the number of instances where a respondent broke the pattern of monotonically increasing response on the set of survey items ordered by difficulty. |