| Literature DB >> 34977357 |
Heng-Yang Lu1,2, Jun Yang3, Cong Hu1, Wei Fang1.
Abstract
BACKGROUND: Fine-grained sentiment analysis is used to interpret consumers' sentiments, from their written comments, towards specific entities on specific aspects. Previous researchers have introduced three main tasks in this field (ABSA, TABSA, MEABSA), covering all kinds of social media data (e.g., review specific, questions and answers, and community-based). In this paper, we identify and address two common challenges encountered in these three tasks, including the low-resource problem and the sentiment polarity bias.Entities:
Keywords: Data augmentation; Fine-grained; Low-resource; Sentiment analysis
Year: 2021 PMID: 34977357 PMCID: PMC8670363 DOI: 10.7717/peerj-cs.816
Source DB: PubMed Journal: PeerJ Comput Sci ISSN: 2376-5992
The comparison between three tasks of sentiment prediction towards entities and aspects.
|
|
|
| |||||
|---|---|---|---|---|---|---|---|
|
|
| text and aspects mentioned | text, entity mentioned, and all kinds of aspects | text, entity mentioned, and aspect mentioned | |||
|
| predict sentiment towards mentioned aspects | predict sentiment towards the combination of mentioned entity and all kinds of aspects | predict sentiment towards mentioned entity aspect combination | ||||
|
|
| < | < | < | |||
|
| (convenience level, positive) | (MacBookPro, price, none) (MacBookPro, convenience level, positive) (MacBookPro, battery, none) ... | (MacBookPro, convenience level, positive) | ||||
|
|
| < | < | < | |||
|
| (battry, positive) | (ThinkPad, price, none) (ThinkPad, convenience level, none) (ThinkPad, battery, positive) ... | (ThinkPad, battery, positive) | ||||
Figure 1The graphical abstract of the PEA model.
An example of entity replacement.
The replacement maintains the same sentiment polarity and correct grammar.
|
|
| I’ve used |
|
| MacBookPro | |
|
|
| I’ve used |
|
| Thinkpad |
Figure 2An example of dual noise injection.
Figure 3General structure of CEA with noise-injected vectors.
Statistics of used datasets.
|
|
|
|
|
|
| ||
|---|---|---|---|---|---|---|---|
|
|
|
| |||||
| Restaurant | English | 3,608 | – | 1,120 |
| ||
| Laptop | English | 2,328 | – | 638 |
| ||
| SentiHood | English | 3,650 | 522 | 1,043 |
| ||
| BabyCare | Chinese | 29,354 | 3,682 | 3,677 |
| ||
Top 10 low-resource entities in the BabyCare dataset, with the number of instances that belong to every polarity category for both the original training set and entity-replaced dataset.
| The original training set | The entity-replaced dataset | |||||
|---|---|---|---|---|---|---|
| Entity | Negative | Neutral | Positive | Negative | Neutral | Positive |
| 佳贝艾特 (Kabrita) | 0 | 0 | 26 | 69 | 243 | 361 |
| 可瑞康 (Karicare) | 0 | 0 | 17 | 108 | 382 | 527 |
| 君乐宝 (JunLeBao) | 0 | 0 | 15 | 121 | 409 | 571 |
| 咔哇熊 (Cowala) | 0 | 0 | 14 | 144 | 446 | 642 |
| 多美滋 (Dumex) | 1 | 0 | 64 | 22 | 84 | 219 |
| 太子乐 (Happy Prince) | 1 | 0 | 19 | 102 | 306 | 485 |
| 奶粉 (milk powder) | 0 | 0 | 19 | 102 | 304 | 479 |
| 欧贝嘉 (OuBecca) | 1 | 0 | 19 | 86 | 305 | 459 |
| 百立乐 (Natrapure) | 4 | 0 | 73 | 37 | 71 | 183 |
| 诺优能 (Nutrilon) | 2 | 0 | 42 | 44 | 146 | 227 |
Figure 4Macro-F1 performance on four datasets with different values of σ in noise infection.
Performance (%) on two datasets for the ABSA task, Accuracy, Marco-Precision, Macro-Recall and Marco-F1 are reported.
|
|
|
| ||||||
|---|---|---|---|---|---|---|---|---|
| Accuracy | Precision | Recall | F1 | Accuracy | Precision | Recall | F1 | |
| TD-LSTM | 75.18 | 70.60 | 56.57 | 58.51 | 64.26 | 57.67 | 56.67 | 54.10 |
| MemNet | 77.32 | 69.87 | 64.38 | 64.61 | 68.65 | 63.58 | 63.62 | 62.69 |
| ATAE-LSTM | 74.38 | 67.43 | 57.28 | 58.32 | 66.14 | 61.22 | 58.97 | 56.91 |
| IAN | 76.16 | 67.43 | 59.31 | 60.56 | 65.20 | 61.64 | 58.54 | 54.08 |
| RAM | 76.07 | 72.07 | 58.65 | 59.59 | 68.03 | 64.03 | 63.86 | 60.82 |
| TransCap | 79.20 | 70.76 | 70.81 | 70.78 | 74.76 | 71.77 | 71.99 | 70.08 |
| ASGCN | 74.29 | 71.95 | 56.74 | 56.45 | 69.75 | 66.21 | 63.75 | 62.29 |
| IACapsNet | 81.79 | – | – | 73.40 | 76.80 | – | – | 73.29 |
| PEA(Our) |
|
|
|
|
|
|
|
|
Performance (%) on the SentiHood dataset for the TABSA task, Accuracy, Marco-Precision, Macro-Recall, Marco-F1 and AUC are reported.
|
|
|
|
|
|
|
|---|---|---|---|---|---|
| LR | 87.5 | – | – | – | 90.5 |
| LSTM+TA+SA | 86.8 | – | – | – | – |
| SenticLSTM | 89.3 | – | – | – | – |
| Dmu-Entnet | 90.2 | 74.8 | 76.3 | 75.5 | 94.8 |
| RE+Delayed-memory | 92.8 | – | – | – | 96.2 |
| BERT-pair-QA-B | 93.3 | – | – | – | 97.0 |
| BERT-pair-QA-M | 93.8 | 83.4 |
| 84.5 | 97.1 |
| PEA(Our) |
|
| 84.5 |
|
|
Performance (%) on the BabyCare dataset for the MEABSA task, Accuracy, Marco-Precision, Macro-Recall and Marco-F1 are reported.
|
|
|
|
|
|
|---|---|---|---|---|
| MemNet | 62.74 | 59.81 | 48.84 | 46.13 |
| ATAE-LSTM | 66.09 | 58.47 | 49.68 | 47.75 |
| IAN | 61.93 | 41.71 | 47.04 | 43.73 |
| MemNet+ | 65.32 | 59.93 | 50.55 | 47.93 |
| ATAE-LSTM+ | 66.25 | 56.01 | 51.93 | 51.87 |
| IAN+ | 65.81 | 44.42 | 50.06 | 46.50 |
| CEA | 80.20 | 77.68 | 75.23 | 76.29 |
| DT-CEA | 81.74 | – | – | 78.23 |
| CADMN | 81.45 | – | – | 78.37 |
| PEA(Our) |
|
|
|
|
p-value between PEA and other baselines on ABSA, TABSA and MEABSA tasks.
|
| ||
|---|---|---|
|
|
|
|
| TD-LSTM | 1.4379e−14 | 1.4331e−14 |
| MemNet | 1.6116e−11 | 1.7323e−09 |
| ATAE-LSTM | 2.0494e−16 | 1.1331e−12 |
| IAN | 5.8819e−13 | 1.2102e−12 |
| RAM | 9.4895e−13 | 2.1595e−09 |
| TransCap | 1.1872e−06 | 0.0138 |
| ASGCN | 7.3462e−16 | 3.7338e−07 |
|
| ||
|
|
| |
| Dmu-Entnet | 6.7790e−41 | |
| BERT-pair-NLI-M | 0.0174 | |
|
| ||
|
|
| |
| MemNet | 6.6475e−140 | |
| ATAE-LSTM | 8.3216e−113 | |
| IAN | 7.1802e−148 | |
| MemNet+ | 3.4485e−120 | |
| ATAE-LSTM+ | 1.1143e−114 | |
| IAN+ | 1.6552e−114 | |
| CEA | 2.2666e−20 | |
Performance (%) of ablation study on four datasets.
|
|
|
|
|
| ||||
|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
| |
| CEA | 80.20 | 76.29 | 90.3 | 93.2 | 78.13 | 67.96 | 71.41 | 67.05 |
| CEA+ER | 80.78 | 77.33 | 91.1 | 93.3 | – | – | – | – |
| CEA+ER+NI | 81.06 | 77.55 | 91.3 | 94.0 | 79.45 | 70.31 | 71.83 | 67.22 |
| BERT-based model | 84.12 | 81.62 | 93.8 | 97.1 | 83.52 | 76.11 | 76.99 | 72.40 |
| PEA (our) |
|
|
|
|
|
|
|
|
Figure 5Box plot of the number of instances for every entity based on the original training and the entity-replaced dataset, respectively.
Case study on misclassifications of BERT-based and data augmented CEA model.
The straight underlined words are entity terms and the wavy underlined words are aspect terms.
|
|
|
| |
|---|---|---|---|
|
| I tried |
| I am too poor to afford |
|
| Pampers | BobDog | Kao |
|
| Anti-leakage | Thickness | Cost |
|
| Negative | Neutral | Neutral |
|
| Neutral | Positive | Neutral |
|
| Neutral | Neutral | Neutral |
|
| Neutral | Neutral | Negative |
Figure 6Performance on extreme low-resource conditions.
Macro-F1 and standard deviation of Macro-F1 (in the brackets) on evident polarity biased (EPB) test set and original test set.
“DA” is short for Data Augmentation.
|
|
|
|
|
|---|---|---|---|
| CEA | 0.7542 (0.0123) | 0.7714 (0.0040) | 1.72% |
| CEA+DA | 0.7753 (0.0069) | 0.7768 (0.0036) | 0.15% |
| BERT-based model | 0.8068 (0.0070) | 0.8162 (0.0040) | 0.94% |
| PEA | 0.8153 (0.0069) | 0.8234 (0.0061) | 0.81% |
|
| [CLS] I’ve used MacBookPro, it’s convenient. [SEP] What is the sentiment towards the convenience of MacBookPro? [SEP] |
|
| Positive |