| Literature DB >> 35578619 |
Yusra Ghafoor1, Shi Jinping2, Fernando H Calderon1, Yen-Hao Huang2, Kuan-Ta Chen3, Yi-Shin Chen4.
Abstract
Microblogs generate a vast amount of data in which users express their emotions regarding almost all aspects of everyday life. Capturing affective content from these context-dependent and subjective texts is a challenging task. We propose an intelligent probabilistic model for textual emotion recognition in multidimensional space (TERMS) that captures the subjective emotional boundaries and contextual information embedded in a text for robust emotion recognition. It is implausible with discrete label assignment;therefore, the model employs a soft assignment by mapping varying emotional perceptions in a multidimensional space and generates them as distributions via the Gaussian mixture model (GMM). To strengthen emotion distributions, TERMS integrates a probabilistic emotion classifier that captures the contextual and linguistic information from texts. The integration of these aspects, the context-aware emotion classifier and the learned GMM parameters provide a complete coverage for accurate emotion recognition. The large-scale experimentation shows that compared to baseline and state-of-the-art models, TERMS achieved better performance in terms of distinguishability, prediction, and classification performance. In addition, TERMS provide insights on emotion classes, the annotation patterns, and the models application in different scenarios.Entities:
Keywords: Emotion distribution; Emotion recognition; Gaussian mixture model; Subjectivity; Text classification; Valence-Arousal
Year: 2022 PMID: 35578619 PMCID: PMC9094737 DOI: 10.1007/s10489-022-03567-4
Source DB: PubMed Journal: Appl Intell (Dordr) ISSN: 0924-669X Impact factor: 5.086
Notations’ Table
| Notation | Description |
|---|---|
| Feature vector of all texts | |
| Feature vector of text | |
| Emotion distribution for emotion | |
| Mixing coefficient for emotion | |
| Mean of | |
| Covariance matrix of | |
| Gaussian distribution | |
| Labelled valence and arousal dataset | |
| The | |
| Number of annotators for text | |
| Labelled dataset of texts and VA-ratings | |
| Predicted mean and covariance for a text | |
| Number of emotional classes |
Fig. 1An illustration of the TERMS probabilistic process. EmoClass is a textual emotion classification module that outputs emotion probabilities for each text into specified affective classes. EmoGMM is an emotion GMM modeling that takes in the probabilities and combines them with VA-ratings to parameterize emotion distributions in a VA-space. The prediction module employs a single affective Gaussian on weighted GMMs to predict an emotion distribution for each unseen text
Emotion distribution statistics
| Emotions | Anger | Anti. | Disgust | Fear | Joy | Sad. | Surprise | Trust | Total |
|---|---|---|---|---|---|---|---|---|---|
| No. of texts | 535 | 482 | 481 | 539 | 495 | 511 | 470 | 487 | 4000 |
Fig. 2Valence and arousal rating interface. Top: arousal. Bottom: valence
Fig. 3Distinguishability results
Fig. 4AEmoD for each model to determine distinguishability; the larger the value, the better the clarity in the emotion distributions on VA-space
Overall performance of prediction
| Method | AKL | AED | ||
|---|---|---|---|---|
| GBM | 5.97 | 1.51 | 0.34 | 0.26 |
| SVM | 4.88 | 1.36 | 0.53 | 0.25 |
| NB | 5.45 | 1.42 | 0.52 | 0.23 |
| CNN | 5.07 | 1.35 | 0.58 | 0.24 |
| DeepMoji | 4.81 | 1.35 | 0.54 | 0.23 |
| NTUA-SLP_NBOW | NA | NA | 0.56 | 0.39 |
| NTUA-SLP_LSTM | NA | NA | 0.59 | |
| SRV-SLSTM | NA | NA | 0.53 | 0.26 |
| C-LSTM-CNN | NA | NA | 0.56 | 0.28 |
| TERMS | 0.30 |
Fig. 5Classification evaluation metrics for TERMS and all the comparative models. TERMS performs better by demonstrating higher precision, recall, F1-score, and Jaccard
Fig. 6Classification evaluation metrics with macro and micro-averaging scores
Bayesian analysis comparative results
| Proposed | Others | t-value | p-value | Mean diff. | Lower | Upper |
|---|---|---|---|---|---|---|
| TERMS | C-LSTM-CNN | 3.12 | 0.00 | 0.32 | 0.18 | 0.53 |
| NTUA-SLP_NBOW | 57.10 | 0.00 | 2.46 | 2.41 | 2.64 | |
| NTUA-SLP_LSTM | 30.46 | 8E-18 | 0.23 | 0.21 | 0.23 | |
| CNN | 5.79 | 8E-09 | 0.29 | 0.18 | 0.38 | |
| DeepMoji | 15.87 | 7E-56 | 0.78 | 0.70 | 0.88 | |
| GBM | 7.64 | 2E-14 | 0.37 | 0.28 | 0.46 | |
| SVM | 42.92 | 0.00 | 1.88 | 1.81 | 1.97 | |
| NB | 40.19 | 0.00 | 1.76 | 1.99 | 2.16 |
Recall by emotion class
| Emotions | Anger | Anti. | Disgust | Fear | Joy | Sad. | Surprise | Trust |
|---|---|---|---|---|---|---|---|---|
| Recall | 0.64 | 0.70 | 0.61 | 0.69 | 0.73 | 0.82 |
Misclassified texts (Sarcasm & satire)
| Texts | Actual | Predicted |
|---|---|---|
| I love when i can’t sleep. | anger | trust |
| seriously?! we had to turn around because my mom forgot the | anger | sadness |
| chicken in the freezer. | ||
| ummmm grow up? please. thank you! | anger | joy |
| sorry sweetheart you downgraded | anger | joy |
| lol oh really? is that what its all about?!! hahahaha | anger | sadness |
Misclassified texts (Lack of explicitness)
| Texts | Actual | Predicted |
|---|---|---|
| royal mail... why you loose my parcel? | anger | trust |
| do some girls really think its attractive to look like prostitutes | anger | disgust |
| on a daily basis... | ||
| that’s fucked up.. | anger | joy |
| i don’t even know you anymore. | sadness | trust |
| guess i’m not good enough for you... | sadness | trust |
Misclassified texts (Word sense disambiguation)
| Texts | Actual | Predicted |
|---|---|---|
| this | sadness | fear |
| there are so many disrespectful and | anger | disgust |
| had to say it....because this generation is going straight down | sadness | disgust |
| the | ||
| i | sadness | disgust |
| i was in such a | anger | joy |
Fig. 7Predicted values of valence and arousal by TERMS
Fig. 8PCC curves of valence and arousal at varied number of annotators.