| Literature DB >> 33204245 |
Yinglin Zhu1, Wenbin Zheng1,2, Hong Tang3.
Abstract
Text sentiment classification is an essential research field of natural language processing. Recently, numerous deep learning-based methods for sentiment classification have been proposed and achieved better performances compared with conventional machine learning methods. However, most of the proposed methods ignore the interactive relationship between contextual semantics and sentimental tendency while modeling their text representation. In this paper, we propose a novel Interactive Dual Attention Network (IDAN) model that aims to interactively learn the representation between contextual semantics and sentimental tendency information. Firstly, we design an algorithm that utilizes linguistic resources to obtain sentimental tendency information from text and then extract word embeddings from the BERT (Bidirectional Encoder Representations from Transformers) pretraining model as the embedding layer of IDAN. Next, we use two Bidirectional LSTM (BiLSTM) networks to learn the long-range dependencies of contextual semantics and sentimental tendency information, respectively. Finally, two types of attention mechanisms are implemented in IDAN. One is multihead attention, which is the next layer of BiLSTM and is used to learn the interactive relationship between contextual semantics and sentimental tendency information. The other is global attention that aims to make the model focus on the important parts of the sequence and generate the final representation for classification. These two attention mechanisms enable IDAN to interactively learn the relationship between semantics and sentimental tendency information and improve the classification performance. A large number of experiments on four benchmark datasets show that our IDAN model is superior to competitive methods. Moreover, both the result analysis and the attention weight visualization further demonstrate the effectiveness of our proposed method.Entities:
Mesh:
Year: 2020 PMID: 33204245 PMCID: PMC7657682 DOI: 10.1155/2020/8858717
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1The architecture of IDAN.
Algorithm 1Sentimental tendency information extraction.
Figure 2The structure of scaled dot-product attention (a) and multihead attention (b).
Figure 3The structure of interactive learning.
Summary of the datasets after tokenization.
| Dataset |
| | | | | ||
|---|---|---|---|---|---|
| Positive | Negative | Positive | Negative | ||
| ChnSentiCorp | 136 | 2400 | 2400 | 600 | 600 |
| NLPCC-CN | 64 | 5000 | 5000 | 1250 | 1250 |
| NLPCC-EN | 130 | 4987 | 4998 | 1250 | 1250 |
| MR | 20 | 4264 | 4264 | 1067 | 1067 |
Summary of the lexicons used in the experiments.
| Language | Lexicon types | Words count | Examples |
|---|---|---|---|
| English | Positive sentiment | 4363 | Applause, satisfied |
| Negative sentiment | 4572 | Abuse, get sick of | |
| Intensity words | 171 | Absolutely, ultra | |
|
| |||
| Chinese | Positive sentiment | 10191 |
|
| Negative sentiment | 13712 |
| |
| Intensity words | 79 |
| |
| Negative words | 71 |
| |
Performance comparison with baseline methods.
| Approach | ChnSentiCorp | NLPCC-CN | NLPCC-EN | MR | ||||
|---|---|---|---|---|---|---|---|---|
| Accuracy | Macro- | Accuracy | Macro- | Accuracy | Macro- | Accuracy | Macro- | |
| SVM | 0.8618 | 0.8528 | 0.7479 | 0.7441 | 0.8226 | 0.8143 | 0.7914 | 0.7852 |
| LSTM | 0.8681 | 0.8570 | 0.7572 | 0.7557 | 0.8381 | 0.8379 | 0.7844 | 0.7705 |
| BiLSTM | 0.8831 | 0.8693 | 0.7603 | 0.7573 | 0.8488 | 0.8477 | 0.7941 | 0.7877 |
| ATT-BiLSTM | 0.8945 | 0.8892 | 0.7665 | 0.7585 | 0.8503 | 0.8491 | 0.7952 | 0.7909 |
| H-RNN-CNN | 0.8940 | 0.9030 | 0.7550 | 0.7790 | — | — | 0.8190 | — |
| CRNN | 0.9108 | 0.9082 | 0.7702 | 0.7648 | 0.8579 | 0.8456 | 0.8228 | — |
| fastText | 0.9203 | 0.9170 | 0.7706 | 0.7624 | 0.8670 | 0.8615 | 0.8181 | 0.8121 |
| LR-BiLSTM | — | — | — | — | — | — | 0.8210 | — |
| IDAN |
|
|
|
|
|
|
|
|
Bold values indicate the best performances.
Results for the ablation experiments.
| Approach | ChnSentiCorp | NLPCC-CN | NLPCC-EN | MR | ||||
|---|---|---|---|---|---|---|---|---|
| Accuracy | Macro- | Accuracy | Macro- | Accuracy | Macro- | Accuracy | Macro- | |
| IDAN-W2V | 0.9145 | 0.9078 | 0.7667 | 0.7657 | 0.8621 | 0.8515 | 0.7986 | 0.7870 |
| IDAN-NIL | 0.9262 | 0.9141 | 0.8002 | 0.7866 | 0.9155 | 0.9069 | 0.8214 | 0.8100 |
| IDAN-NSTI | 0.9233 | 0.9099 | 0.8045 | 0.7911 | 0.9128 | 0.9062 | 0.8225 | 0.8134 |
| IDAN-NGA | 0.9184 | 0.9133 |
|
| 0.9164 | 0.9068 | 0.8254 | 0.8130 |
| IDAN |
|
| 0.8005 | 0.7875 |
|
|
|
|
Bold values indicate the best performances.
Figure 4Visualization of attention weights for two test cases: (a) case 1 and (b) case 2.