Literature DB >> 31316901

Providing a Four-layer Method Based on Deep Belief Network to Improve Emotion Recognition in Electroencephalography in Brain Signals.

Seyed Mohammad Reza Mousavinasr1, Ali Pourmohammad2, Mohammad Sadegh Moayed Saffari1.   

Abstract

BACKGROUND: One of the fields of research in recent years that has been under focused is emotion recognition in electroencephalography (EEG) signals. This study provides a four-layer method to improve people's emotion recognition through these signals and deep belief neural networks.
METHODS: In this study, using DEAP dataset, a four-layer method is established, which includes (1) preprocessing, (2) extracting features, (3) dimension reduction, and (4) emotion identification and estimation. To find the optimal choice in some of the steps of these layers, three different tests have been conducted. The first is finding the perfect window in feature extraction section that resulted in superiority of Hamming window to the other windows. The second is choosing the most appropriate number of filter bank and the best result was 26. The third test was also emotion recognition that its accuracy was 92.93 for arousal dimension, 92.64 for valence dimension, 93.14 for dominance dimension in two-class experiment and 76.28 for the arousal, 74.83 for the valence, and 75.64 for dominance in three-class experiment.
RESULTS: The results of this method show an improvement of 12.34% and 7.74% in two- and three-class levels in the arousal dimension. This improvement in the valence is 12.77 and 8.52, respectively.
CONCLUSION: The results show that the proposed method can be used to improve the accuracy of emotion recognition.

Entities:  

Keywords:  Deep belief neural network; deep neural network; electroencephalography; emotion recognition; independent component analysis

Year:  2019        PMID: 31316901      PMCID: PMC6601226          DOI: 10.4103/jmss.JMSS_34_17

Source DB:  PubMed          Journal:  J Med Signals Sens        ISSN: 2228-7477


Introduction

In today's world that various waves and signals surround humans, the use of these signals and waves in different cases is inevitable. In many cases, they are the solution of problems in different areas. As a result of activities, thoughts, ideas, and feelings of human beings, signals are generated in the brain. These signals can be analyzed and used in various ways. There are various types of human brain signals, of which electroencephalography (EEG) signals are the most widely used. This signal is an electrical signal that is generated as a result of brain activity. The signal also contains various information that can be used in different contexts. Among these contexts is individuals' emotion recognition against various sensory stimuli that are usually induced through observation and sometimes by ear and through human hearing. During these stimulations, different feelings of individuals emerge in EEG brain signals. These signals that contain useful information about people's feelings can be analyzed and emotions can be extracted. The importance of this task becomes clear when emotions, while only are imagined, appear in brain signals without moving a muscle. It can be easily inferred that this is so useful for people who lack mobility and suffer muscle diseases such as sclerosis, or hardening of the brain tissue, and muscle paralysis. In this study, important papers presented between 2010 and 2015 are reviewed. These articles are about emotion recognition through brain signals using different methods. Majority of them have used two-dimensional (2D) arousal-valence model proposed by Russell[1] or 3D arousal-valence-dominance model which has dominance for controlling the emotions. Considering to previous works, deep learning is a way to improve the accuracy of emotion recognition in EEG brain signals. In this method, according to previous research conducted at Malek Ashtar University, it is required to remove noise from signals to increase the accuracy of diagnosis in different applications. Imani et al.[2] conducted a study as “investigating independent-component analysis (ICA) algorithm to distinguish between two conceptual groups of risk and informing words using traffic signs” and proposed ICA algorithm as one of the best methods to eliminate noise from EEG brain signals. This method has also been used by Imani et al.[2] in their study entitled as “evaluation of the performance of the separator and the extracted features to differentiate patterns of brain related to mental activities associated with the four cardinal directions,” who presented more appropriate and acceptable results compared to previous research; accordingly, it is decided to benefit from this method in the current study.

Research Significance and Objective

The main objective of this study is to improve emotion recognition through aroused brain signals using deep learning, which is one of the identification and classification methods. Given the application of this research, the significance of this study can be studied from various aspects. One of these aspects is helping solve some problems for amyotrophic lateral sclerosis patients. These patients lack muscle mobility and this method can be used to help them express their feelings. Another aspect is this method's military use for the diagnosis of mental states such as distress and anxiety in the soldiers. On the other hand, it is possible to detect a liar, which is one of the most used one in the detection of emotions. Further, given that this research field is new, this research can be suitable for research and development in other fields related to this issue.

10–20 standard

In 1949, 10–20 electrode positioning method was known as an international standard that could cover almost all the areas of head by the electrodes. Electrodes are chosen based on particular parts of the skull. The electrodes are placed in the areas of intersection of the skull, and other central electrodes are arranged based on the 10% and 20% of the whole distance [Figure 1].[2]
Figure 1

Electrodes positions in 10–20 international standards

Electrodes positions in 10–20 international standards

Emotion recognition

Happiness, sorrow, fear, sadness, and more! All these words are derived from a concept called emotion. It can be said that this concept is in all creatures on earth. These feelings are stimulated and aroused by different situations. To stimulate and arouse emotions, it is needed to stimulate person by a stimulus so as to arouse feelings in the person. The stimulus can arouse emotions through all five senses of sight, taste, hearing, smell, and touch. Emotion recognition can be done in different ways on different grounds. Emotion recognition can also be used in various areas and affect them. One of the applications of emotion recognition is helping those who are not able to mobilize and are paralysis, or it can be helpful in the treatment of depression. Emotion recognition, as noted, can be done in several ways. One approach is using brain signals for emotion recognition after recording brain signals. Given that all activities occurring in the body are processed in the brain, electrical brain signals that are created as a result of brain nerve cells' activities can be used for emotion recognition. The most important model for emotion recognition is arousal-valence model proposed by Russell.[1] This model is a 2D model that includes all the feelings. In addition, another model that was more comprehensive was proposed by those who proposed the 2D model, in which emotion control aspect was also added. This model is used as a reference model for emotion recognition either through brain signals or through other methods.

Arousal-valence model

The model was presented in an article published in 1980 by Russell.[1] In this model, all the feelings of people are placed on a page that has 2Ds of valence and arousal. This page is divided into four parts by these dimensions. People's different emotions are placed in these four areas and given the amount of valence and arousal of each emotion; the feeling is placed in a level of this model. Vertical dimension is arousal and horizontal dimension is valence [Figure 2].[1]
Figure 2

Two-dimensional arousal-valence model

Two-dimensional arousal-valence model For example, since excitement is an inner sense with high valence and great arousal, it is located in the top right quadrant of the model. The opposite of excitement is depression which is located in the bottom left quadrant of the model. This model has been used in most of the researches carried out in the field of emotion recognition. It can be divided into various levels according to valence and arousal of different emotions. Given the complexities of emotion recognition, emotions are usually recognized in two or three levels in different studies. For instance, Hatamikia et al.[3] in 2014 recognized emotions in both two and three levels. There is another dimension which called dominance in this model too. This dimension indicates user's control over his/her emotions that affect performance and his/her manifestation in the brain signals. This model is used less than the 2D model.

Challenges

Despite the various studies conducted on emotion recognition, there are still major challenges. Among all challenges in this field, how to remove unwanted noise in the recorded signal signals is one of the important issues. To remove unwanted noises, some methods have been used in different studies, but it has been proven that using combined methods such as the combination of wavelet decomposition and ICA has superior performance compared to the native wavelet or ICA separately.[2] Combination of two methods such as empirical mode decomposition and canonical correlation analysis has a better performance in removing noises compared to using canonical correlation analysis solo. Moreover, combination of empirical mode decomposition and ICA has had better performance than ICA.[2] The main challenge in these systems is improvement of performance, especially accuracy of emotion recognition. This is related to different categories and just one factor cannot be effective in improving the performance of emotion recognition in brain signals. Among the factors that contribute to this improvement, selection of the optimal feature as well as the selection of appropriate category to classify the feelings extracted from brain signals is more important to others. These two factors are somehow linked together and we can say that proper selection of features will improve the classification performance by sure. Another factor that affects the accuracy of the results is the number of levels in the arousal-valence model of emotions. As was mentioned earlier, the reference model used for emotion recognition is 2D arousal-valence model proposed by Russell in 1980.[1] The number of levels studied in this model can affect the performance and accuracy. In this regard, the study of Hatamikia et al.[3] in 2014 on DEAP datasets well represents this issue. The accuracy of recognition of two levels in the dimension of valence in their study was equal to 72.23 and was 74.20 in the dimension of arousal; however, when they were attempted to recognize at three levels, the accuracy values were reduced to 61.1 and 65.16, respectively.

The most important classification methods used

To recognize emotions through brain signals, there is a need to methods that can be used to identify and classify signals related to emotions arouse by stimuli and displayed in brain signals. Machine learning methods can be used to detect emotions in the brain signals. One of the most important elements in this task is the classifier. The most widely used classifiers in different studies are described in [Table 1].
Table 1

Major studies on emotion recognition in electroencephalography signals

FeaturesNumber of emotionsParticipantsAccuracy (%)DatasetAuthors
SVMHigh-order crossings61683.3Online sampling with stimulus images[17]
PSD30, DAMS12, RAMS12 and PSD244Not mentioned82.2916 pieces of music tracks produced by Oscar film festival[7]
PCA and PSDArousal and valence32Arousal 67Some emotional videos from YouTube[8]
Valence 76
Magnitude mutual information and squared coherence estimateArousal and valence26Arousal 75IAPS[8]
Valence 81
Quality dimensions, approximate entropy, and wavelet entropyNot mentioned1573.25IAPS[9]
High-order crossing and cross correlationNot mentioned16Independent participants 65.8Online sampling[6]
Dependent participants 94.40
Fourier transform log band energyNot mentionedNot mentioned87.534-min movie parts prepared from Oscar film festival[10]
PSDArousal and valence30Arousal 52.430 subjects prepared by film and photos[4]
Valence 57.7
Chebyshev II IIRArousal and valenceNot mentionedArousal 82.03IIa and IIb data of BCI Competition data set[11]
Valence 65.39
Minimum, maximum, mean, standard deviationNot mentioned2077.75IAPS[12]
Fractal dimension, high-order crossing and 6 statistical features8324-electrod 53.7IADS, IAPS, and DEAP[13]
44-electrod 87.02
Differential entropy of each 5 alpha, beta, gamma, delta and theta bands from 62 channelsNot mentioned6 84.08Displaying a series of pieces from famous films[14]
PSD and Fisher Kernel - PCA and LDA - SVM with imbalanced Quasiconformal kernel and imbalanced SVMArousal and valence100 college studentsArousal 84.79IAPS[5]
Valence 82.68
fractal dimension, high-order crossing, inter-class correlation coefficient, 6 statistical features2100 college students71.75IADS[15]
449.63
Hjorth parameters, i.e., activity, mobility and complexity for each 5 bands of five lobes of the brain2100 college students67IAPS[16]
534
K-NNHigh-order crossing61673.94Online Sampling with stimulus images[17]
Magnitude mutual information, squared coherence estimate4 emotions in arousal and valence26Arousal 79IAPS database and unique pieces of music from bernard bouchard[18]
Valence 82
PSD, combined Gaussian model of EEG spectrumsArousal and valence modelNot mentionedArousal Minimum=55.07, Maximum=67.0Some specific clips for emotion stimulation and signal recording[19]
Valence Minimum=58.8 Maximum=76
Chebyshev II IIRArousal and valenceNot mentionedArousal 82.25IIa and IIb data of BCI competition data set[11]
Valence 66.51
Minimum, maximum, mean, and standard deviation and ICA for noise removalNot mentioned2070.37IAPS[12]
Auto regression and KNN2 classes arousal and valence32Arousal 74.2DEAP[3]
Valence 72.33
3-classes arousal and valenceArousal 65.156
Valence 61.1
Differential entropy of each 5 bands obtained 62 channelsNegative and positive poles669.22Some clips of famous films[14]
Artificial neural networksPSD30, RAMS12, DAMS12, PSD24Not mentioned2651.5216 pieces of music tracks produced by Oscar Film Festival[7]
Differential entropy of each 5 bands obtained 62 channelsNegative and positive poles6DBN 86.91Some clips of famous films[14]
DBN-HMM 87.62
PSD of 32 EEG signal channels and PCAWithout compliance in arousal and valence32Arousal 74.2IADS, IAPS[20]
Valence 72.33
Compliance based in arousal and valenceArousal 65.156
Valence 61.1
LDAFrequency of EEG signal bands, brain asymmetry and dependenceNot mentioned110Frequency 38.8Some Movies[21]
Asymmetry 57.9
Dependency 55.3
Auto regression, SFS, and Davies-Bouldin index2 classes arousal and valence32Arousal 65.54DEAP[3]
Valence 63.22
3-classes arousal and valenceArousal 52.36
Valence 51.20
QDAHigh-order crossing61662.03Online Sampling with stimulus images[17]
Auto regression, SFS, and Davies-Bouldin index2 classes arousal and valence32Arousal 69.26DEAP[3]
Valence 70.35
3-classes arousal and valenceArousal 57.42
Valence 57.18
Naïve BaysFeatures extracted from IIR, i.e., CSP, ASP, and AF features and fisher linear discriminant for dimension reductionArousal and valenceNot mentionedArousal 66.24IIa and IIb data of BCI Competition data[11]
Valence 83.10
Minimum, maximum, mean, and standard deviation and ICA for noise removalNot mentioned2059.26IAPS[12]
Fast Fourier transform analysis for feature extraction and features based on Pearson correlation coefficients2 Classes arousal and valenceNot mentionedArousal 70.1IAPS, IADS[22]
Valence 70.9
3 Classes arousal and valenceArousal 55.2
Valence 55.4
Mean, standard deviation and linear fisher discriminantArousal and valence32Arousal 59Some emotional videos from YouTube[8]
Valence 57

SVM – Support vector machine; PSD – Power spectral density; PCA – Principal component analysis; ECG – Electroencephalography; IAPS – International affective picture system; IADS – International Affective Digitized Sounds; IIR – Infinite impulse response; BCI – Brain-computer interface; DEAP – Dataset for emotion analysis using EEG, physiological and video signals; K-NN – K-nearest neighbor; LDA – Linear discriminant analyzer; SFS – Sequential forward selection; ICA – Independent-component analysis; AF – Atrial fibrillation; QDA – Quadratic discriminant analysis; ASP – Asymmetric spatial pattern; CSP – Common spatial pattern; DBN-HMM – Deep belief network – Hidden markov model

Major studies on emotion recognition in electroencephalography signals SVM – Support vector machine; PSD – Power spectral density; PCA – Principal component analysis; ECG – Electroencephalography; IAPS – International affective picture system; IADS – International Affective Digitized Sounds; IIR – Infinite impulse response; BCI – Brain-computer interface; DEAP – Dataset for emotion analysis using EEG, physiological and video signals; K-NN – K-nearest neighbor; LDA – Linear discriminant analyzer; SFS – Sequential forward selection; ICA – Independent-component analysis; AF – Atrial fibrillation; QDA – Quadratic discriminant analysis; ASP – Asymmetric spatial pattern; CSP – Common spatial pattern; DBN-HMM – Deep belief network – Hidden markov model

Materials and Methods

Based on previous works in this fields such as Imani et al[2], it decided to use ICA for removing the noises from the raw EEG signals in this study. According to other studies conducted on emotion recognition that mentioned before and also employment of deep learning in different studies for sound detection, a four-step process is proposed. The four steps include: Preprocessing Extracting features Dimension reduction Emotion identification and estimation. Each step also has several substeps. To increase accuracy of emotion recognition, it is necessary to carry out each of these steps and their substeps in a way that the highest possible accuracy achieved. To find the optimal choice in some of the steps of these layers, different tests have been conducted. The first step, preprocessing, includes the following four substeps: Noise reduction Selecting the appropriate signals Filtering for extracting the five principal bands of EEG signal Selecting the appropriate bands to recognize emotions. The second step also includes the following steps for desired feature extraction, i.e., linear frequency cepstral coefficient (LFCC). The steps also include: Framing Windowing Fast Fourier transform (FFT) Apply filter bank Logarithm Discrete cosine transform (DCT). After this, step 3 starts, i.e., dimension reduction using Kernel Fisher discriminant analysis (KFDA) algorithm. Finally, in step 4, emotion detection and estimation, after testing both deep neural network and deep belief neural network (DBN), attempted to estimate and recognize emotions through EEG signals. Then, recognition accuracy of each deep learning methods is calculated using mean square error (MSE). It should be noted that this study has been carried out at two two-class and three-class levels of Russell reference model.[1] In previous studies, which were described in detail, accuracy of three-class recognition was lower than two-class recognition, and it is one of the challenges of emotion recognition. In this study, both levels were tested by the proposed method [Figure 3].
Figure 3

Proposed method for emotion recognition

Proposed method for emotion recognition

DEAP dataset

This dataset prepared by Koelstra[8] in collaboration with four universities of EPFL, Genève, Twente, and Queens Marry for Emotion Analysis using EEG, physiological, and video signals. DEAP contains some video clips to stimulate emotions and recorded brain signals of the experimented individuals. It was the first standard dataset which attempted to incite people by video clips and induces emotions in the brain signals. The data has been prepared from 32 participants during the 40-min movie watching. Face of the 22 out of 32 participants was recorded as videos. This dataset is publicly available via a dedicated website at Queens Marry datasets.

Testing the best window for extracting linear frequency cepstral coefficient feature

In the step of feature extraction from DEAP dataset, i.e., LFCC feature extraction, there are some steps there, one of them called windowing. In this part, four known and most widely used windows were tested to find the best window. The four windows are: Hamming Hanning Black-Man Rectangle. The process of testing the best window is such that for each window, for all the individuals and tests conducted in DEAP dataset, LFCC feature extraction is done based on that window, and then, 3Ds of emotion, i.e., arousal, valence, and dominance, were estimated. The input data were first preprocessed to remove their noise. Then, 32 EEG channels were selected and applied filter to extract the main five EEG bands. After that, two bands appropriate for emotion recognition were selected according to reference.[7] Then, feature extraction starts. At this point, the data are divided into 1-s frames, and on every frame, one window of above-mentioned four windows is applied every time. According to the research of Imani et al.,[2] window size was considered 512. After that, FFT is taken from results obtained and 21 filter banks with 50% overlapping are applied on it. In this experiment, the number of filter banks is fixed and considered to 21 to simplify calculations and carry out experiments in the same environment. Here, according to the range of 8–30 Hz, which is related to the alpha and beta bands, by applying a triangular filter with size 2 and an overlay of 50%, there are exactly 21 filters in this range. After extracting the data of this step, logarithm and DCT is calculated for them. After that, according to paper,[5] KFDA with Gaussian Kernel is used for dimension reduction and emotion estimation for three dimensions of arousal, valence, and dominance is done. It should be noted that to obtain better features with fewer dependency to the other dimensions; dimension reduction is conducted for three times. Three DBNs have been used for estimation so that each of them estimates one dimension and give better results as much as possible. For selecting data for estimation, cross-validation method is used in this study. In this way, 25% of data are allocated for testing, 25% for validating, and the remaining 50% for training. Data selection process is done randomly. After applying these settings, DBNs are trained and then tested. Finally, their results were compared based on different windows. This comparison was done as person-to-person and also as average of all patients. The accuracy of the results is calculated using the MSE. After that, considering the results of the comparison, the best window to estimate emotions is selected. It should be noted that this experiment was conducted in a two-class level. The mean results of experiment on four tested windows are shown in Table 2.
Table 2

Average of accuracy for suitable window test

HammingHanningBlack-ManRectangle
Arousal92.7989.73288.2970.71
Valence91.0290.00685.1170.02
Dominance92.6688.4585.2571.45
Average of accuracy for suitable window test

Testing the number of filter banks

In this experiment, according to previous research on the use of filter banks whether with the use of Mel-frequency cepstral coefficients or other methods such as LFCC, a range of 10-number widely used filter banks are selected to choose the most appropriate number for emotion recognition and estimation. The range is from 20 to 29. The process of testing the most appropriate number of filter banks for extracting features to recognize emotions in brain signals is the same as testing the most appropriate window with the exception that the window in this experiment is hamming and filter banks presented in Table 3 are tested for all the experiments of individuals in DEAP dataset. This means that after framing and applying Hamming on data and taking FFT, a different number of filter bank is applied based on Table 2 and then their logarithm and DCT are calculated. The experiment is done for all elements of the row of filter bank in Table 3.
Table 3

Tests numbers and correspond number of filter banks

Tests numberNumber of filter banks
120
221
322
423
524
625
726
827
928
1029
Tests numbers and correspond number of filter banks Similar to window testing, KFDA with Gaussian Kernel is used for dimension reduction for all 3Ds, three times, and three DBNs are used for emotion recognition. The structure of neural networks used in this test is similar to the test of the most appropriate window. Finally, the accuracy of the results obtained using MSE is calculated and results of each filter bank are compared to other results. It should be noted that experiment of this section is also two-class.

Results and figures related to mean recognition accuracy

According to tests performed by different studies, the mean recognition accuracy using different filter banks is required to make a decision to choose the most appropriate number of filter banks to increasing recognition rate. The mean results of recognition accuracy based on different number of filter banks are shown in Table 4.
Table 4

Mean recognition accuracy based on the number of different filter banks

ArousalValenceDominance
The number of used filter banks
2091.97590.34791.51
2192.21291.76792.596
2291.8390.88492.334
2391.51491.69591.147
2491.52291.96791.469
2592.97491.39992.369
2693.92592.64493.142
2792.08991.92692.284
2892.54690.80291.02
2992.12591.64190.753
Mean recognition accuracy based on the number of different filter banks

Test of emotion recognition in brain signals

At this step, after setting the previous steps for emotion recognition, the test of the most appropriate window and the appropriate number of filter banks with the results of Kaiser with the parameters of Hamming window and 26 filter banks, deep neural networks and DBN were tested to recognize emotions. The test is done in two two-class and three-class levels that cover two levels of the reference model of Russell and three levels of it. The process and results of these tests at two levels are as follows.

Test conducted by deep neural network

In this test, LFCC features, extracted from the proposed method based on Kaiser Window with parameters of Hamming and 26 filter banks and after dimension reduction by KFDA algorithm with Gaussian Kernel, were entered to the deep neural network with the following structure to estimate and recognize emotions. As seen in Figure 4, the network is composed of layers including an input layer with 32 neurons, three hidden layer with 70, 70, and 200 neurons, respectively, and an output layer with one neuron. This type of network is fitnet and its learning function is selected as trainscg.
Figure 4

The structure of used deep neural network in emotion recognition tests

The structure of used deep neural network in emotion recognition tests To obtain appropriate results of each emotion dimensions, obtained LFCC features based on the dimension are trained, validated, and tested each time so as to obtain better accuracy for each dimension. Dividing of input data or LFCC features extracted from data is done using cross-validation. By this method, 50% of extracted LFCC features are used for training, 25% for validation, and 25% for testing. Finally, after receiving outputs of mentioned deep neural network, detection accuracy is measured using the MSE.

Test conducted by deep belief neural network

Like previous test, LFCC features, extracted from the proposed method based on Kaiser Window with parameters of Hamming and 26 filter banks and after dimension reduction by KFDA algorithm with Gaussian Kernel, were entered to the DBN with the following structure to estimate and recognize emotions. As seen in Figure 5, the network is composed of three restricted boltzmann machine (RBMs) of size 50, 50, and 100.
Figure 5

The structure of used deep belief neural network in emotion recognition tests

The structure of used deep belief neural network in emotion recognition tests To obtain appropriate results of each emotion dimension, obtained LFCC features based on the dimension are trained, validated, and tested each time so as to obtain better accuracy for each dimension. Dividing of input data or LFCC features extracted from data is done using cross-validation. By this method, 50% of extracted LFCC features are used for training, 25% for validation, and 25% for testing. For carrying out this test, three DBNs have been used and the structure of each of them is as follows: All three DBNs used are function approximation Sampling method for all of them is Free Energy in Persistent Contrastive Divergence They are composed of three RBMs The network units are probability type with actual quantities between [0, 1] The first RBM is 50 The second RBM is 50 The third RBM is 100 All RBMs are of generative type The maximum number of steps taken in the training process executed on the entire training data for each three RBMs is considered to be 50 DeebNet toolbox has been used for DBNs implementation.[23] Finally, after receiving outputs of mentioned DBN, detection accuracy is measured using the mean square error.

Results

The results of this experiment are as follows:

Results of emotion recognition test using deep neural network

Based on the details given in previous sections, the results of conducting emotion recognition test at two-class level in Russell's model using the proposed method by deep neural network estimator is shown in Table 5.
Table 5

The results of conducting emotion recognition test at two-class level using the method proposed by deep neural network

ArousalValenceDominance
81.583979.875080.3595
The results of conducting emotion recognition test at two-class level using the method proposed by deep neural network The results of conducting emotion recognition test at three-class level in Russel's model using the proposed method by deep neural network estimator is shown in Table 6.
Table 6

The results of conducting emotion recognition test at three-class level using the method proposed by deep neural network

ArousalValenceDominance
68.5466.3166.92
The results of conducting emotion recognition test at three-class level using the method proposed by deep neural network

Results of emotion recognition test using deep belief neural network

The results of conducting emotion recognition test at two-class level in Russell's model using the proposed method by DBN estimator is shown in Table 7.
Table 7

The results of conducting emotion recognition test at two-class level using the method proposed by deep belief neural network

ArousalValenceDominance
93.924892.644493.1416
The results of conducting emotion recognition test at two-class level using the method proposed by deep belief neural network The results of conducting emotion recognition test at three-class level in Russell's model using the proposed method by DBN estimator is shown in Table 8.
Table 8

The results of conducting emotion recognition test at three-class level using the method proposed by deep belief neural network

ArousalValenceDominance
76.2874.8375.64
The results of conducting emotion recognition test at three-class level using the method proposed by deep belief neural network Finally, the recognition accuracy obtained from experiments conducted with the proposed method compared to most important studies using the same dataset is shown in Table 9.
Table 9

Compared table of obtained result in two-class and three-class with other outstanding studies

LevelDimensionHatamikia et al.[3]Yoon and Chung[22]Koelstra[8]Jirayucharoensak et al.[20]Proposed method by DNN (Me-DNN)Proposed method by DBN (Me-DBN)
2-classArousal74.270.1--81.5893.9248
3-class65.1655.26252.0368.5476.28
2-classValence72.3370.9--79.8792.6444
3-class61.155.457.653.4266.3174.83
2-classOther dimensions----80.3593.1416
3-class--55.4-66.9275.64
Details--Interest±9.74DominanceDominance

DBN – Deep belief neural network; DNN – Deep neural network

Compared table of obtained result in two-class and three-class with other outstanding studies DBN – Deep belief neural network; DNN – Deep neural network

Conclusion

The aim of this study was to improve emotion recognition in EEG signals through deep learning. In this regard, as stated earlier, a four-step method was proposed. Among these four steps, there are some steps which needed further examination for more accurate selection to improve emotion recognition. One of these steps was feature extraction which LFCC features were extracted. Another one was windowing. According to tests conducted to determine the best window, among four windows, Hamming, Hanning, Black-Man, and Rectangle, Hamming was the best window for this section. According to the results of this experiment, Hamming window resulted in more accuracy in testing for the most appropriate two-class window. Mean accuracy achieved for this window in the dimension of arousal was 92.79, 91.02 in valence, and 92.66 in dominance. Likewise, for the Hanning window, these values were equal to 89.73, 90.006, and 88.45, respectively. Further, for Black-Man window, the stated values were 88.29 for arousal, 85.11 for valence, and 85.25 for dominance. Finally, for the fourth window, Rectangle, mean accuracy of the arousal was 70.71, 70.02 for valence, and 71.45 for dominance. In another part of the feature extraction procedure, it was needed to determine how many filter banks are required for emotion recognition in EEG signals. Therefore, a test was designed to evaluate different values of the filter banks. Based on the results of the test described in detail in the previous sections, the most appropriate number of filter banks in LFCC procedure for feature extraction was 26. The mean results of the number of filter banks in two-class test were 93.92, 92.64 and 93.14 for arousal, valence, and dominance, respectively. After that, the second best choice for the number of filter banks was 25. Mean accuracy obtained for this number is 92.97, 91.39, and 92.36 for arousal, valence, and dominance, respectively. According to the result of the tests, the window which used for LFCC was Hamming. Finally, to carry out the final test, i.e., emotion recognition in EEG brain signals using deep learning, two classification methods of deep neural network with five layers and DBN composed of three RBMs have been used. The mean results of emotion recognition test accuracy using deep neural network was 81.85, 79.87, and 80.35 for arousal, valence, and dominance in two-class experiment. The accuracy values in three-class experiment were 68.54, 66.31, and 66.92 for arousal, valence, and dominance. The results for DBN in two-class experiment were 93.92, 92.64, and 93.14 for arousal, valence, and dominance and in three-class experiment were 76.28, 74.83, and 75.64, respectively. As can be seen, the results of this method compared with previous works, both at two-class and three-class levels, show better accuracy. These results show that the proposed method could solve this problem, with higher accuracy. In the proposed method using deep neural network, the accuracy value obtained had an improvement of 7.38% in two-class level and 3.38% in three-class level for arousal compared to the best accuracy obtained so far, i.e., the study of Hatamikia et al.[3] This improvement is 7.54% and 5.21% for valence. In the proposed method using DBN, the accuracy value obtained had an improvement of 12.34% in two-class level and 7.74% in three-class level for arousal. This improvement was 12.77 and 8.52% for valence. Generally, the accuracy improvement can be the result of change in three different parts employed in this experiment compared to previous studies. These changes are: Using all signals without downsampling and noise removal that although increased computation time, affected accuracy, and increased it to the same extent Determining the most appropriate selections in feature extraction section that resulted in increased accuracy compared to other choices Selecting classification that has had very good results in sound recognition section.

Financial support and sponsorship

None.

Conflicts of interest

There are no conflicts of interest.

BIOGRAPHIES

Seyed Mohammad Reza Mousavinasr has received his BSc in Software Engineering from Mazandaran University of Science and Technology, Mazandaran, Iran in 2011. He also received his MSc from Malek-Ashtar University of Technology in 2015 in artificial intelligence. Research interest of him are EEG Signal processing, Machine Learning Information Security and Deep Leaning. Email: Reza.Mousavinasr@gmail.com Ali Pourmohammad was born in Azerbaijan (Tabriz City), IRAN. He has a Ph.D. in Electrical & Electronics Engineering (Applied Signal Processing) from the Electrical Engineering Department, Amirkabir University of Technology, Tehran, IRAN, Jan. 2012. He is an invited lecturer at this department now and teaches several courses as C++ Programming, Multimedia Systems, Microprocessor Systems, and Digital Signal Processing. His research interests include digital signal processing and applications (audio & speech, image & video, and biomedical & brain signals processing and applications). Email: pourmohammad@aut.ac.ir Mohammad Sadegh Moayed Saffari has received his BSc in Software Engineering from University of Science and Culture, Tehran, Iran in 2011. He also received his MSc from Malek-Ashtar University of Technology in 2015 in artificial intelligence. His research interests are Signal processing, Machine Learning Information Security and Deep Leaning. Email: ms.moayedsaffari@gmail.com
  7 in total

1.  EEG-based emotion estimation using Bayesian weighted-log-posterior function and perceptron convergence algorithm.

Authors:  Hyun Joong Yoon; Seong Youb Chung
Journal:  Comput Biol Med       Date:  2013-10-26       Impact factor: 4.589

2.  Emotion recognition from EEG using higher order crossings.

Authors:  Panagiotis C Petrantonakis; Leontios J Hadjileontiadis
Journal:  IEEE Trans Inf Technol Biomed       Date:  2009-10-23

3.  A novel emotion elicitation index using frontal brain asymmetry for enhanced EEG-based emotion recognition.

Authors:  Panagiotis C Petrantonakis; Leontios J Hadjileontiadis
Journal:  IEEE Trans Inf Technol Biomed       Date:  2011-05-27

4.  Emotion recognition from single-trial EEG based on kernel Fisher's emotion pattern and imbalanced quasiconformal kernel support vector machine.

Authors:  Yi-Hung Liu; Chien-Te Wu; Wei-Teng Cheng; Yu-Tsung Hsiao; Po-Ming Chen; Jyh-Tong Teng
Journal:  Sensors (Basel)       Date:  2014-07-24       Impact factor: 3.576

5.  The emotion recognition system based on autoregressive model and sequential forward feature selection of electroencephalogram signals.

Authors:  Sepideh Hatamikia; Keivan Maghooli; Ali Motie Nasrabadi
Journal:  J Med Signals Sens       Date:  2014-07

6.  EEG-based emotion recognition using deep learning network with principal component based covariate shift adaptation.

Authors:  Suwicha Jirayucharoensak; Setha Pan-Ngum; Pasin Israsena
Journal:  ScientificWorldJournal       Date:  2014-09-01

7.  ICA-Based Imagined Conceptual Words Classification on EEG Signals.

Authors:  Ehsan Imani; Ali Pourmohammad; Mahsa Bagheri; Vida Mobasheri
Journal:  J Med Signals Sens       Date:  2017 Jul-Sep
  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.