| Literature DB >> 35161459 |
Lin Wang1, Zuqiang Meng1.
Abstract
In Chinese sentiment analysis tasks, many existing methods tend to use recurrent neural networks (e.g., long short-term memory networks and gated recurrent units) and standard one-dimensional convolutional neural networks (1D-CNN) to extract features. This is because a recurrent neural network can deal with the order dependence of the data to a certain extent and the one-dimensional convolution can extract local features. Although these methods have good performance in sentiment analysis tasks, recurrent neural networks (RNNs) cannot be parallelized, resulting in time-inefficiency, and the standard 1D-CNN can only extract a single sample feature, with the result that the feature information cannot be fully utilized. To this end, in this paper, we propose a multichannel two-dimensional convolutional neural network based on interactive features and group strategy (MCNN-IFGS) for Chinese sentiment analysis. Firstly, we no longer use word encoding technology but use character-based integer encoding to retain more fine-grained information. Besides, in character-level vectors, the interactive features of different elements are introduced to improve the dimensionality of feature vectors and supplement semantic information so that the input matches the model network. In order to ensure that more sentiment features are learned, group strategies are used to form several feature mapping groups, so the learning object is converted from the traditional single sample to the learning of the feature mapping group, so as to achieve the purpose of learning more features. Finally, multichannel two-dimensional convolutional neural networks with different sizes of convolution kernels are used to extract sentiment features of different scales. The experimental results on the Chinese dataset show that our proposed method outperforms other baseline and state-of-the-art methods.Entities:
Keywords: feature mapping group; group strategy; interactive features; multichannel; two-dimensional convolutional neural network
Mesh:
Year: 2022 PMID: 35161459 PMCID: PMC8840113 DOI: 10.3390/s22030714
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1The framework of MCNN-IFGS. Here, 2D Conv-1 represents 2D-CNN with a size of 1*1 convolution kernel and so on.
Figure 2Multichannel two-dimensional convolutional neural network.
General statistics of Chinese review dataset.
| Sentiment Polarity | Number | Avg_len | |
|---|---|---|---|
| Train | Test | ||
| positive | 36,450 | 12,150 | 32.58 |
| negative | 36,450 | 12,150 | |
Distribution of text length in Chinese review dataset.
| Length | 10 | 20 | 30 | 40 | 50 | >50 |
| Number | 30,910 | 22,973 | 13,820 | 8984 | 5492 | 17,821 |
| Probability | 0.31 | 0.23 | 0.14 | 0.09 | 0.05 | 0.18 |
Figure 3The changes in the number of samples corresponding to different text lengths.
Figure 4The probability distribution corresponding to different text length ranges.
The results of different methods on Chinese review dataset. Here, each method was run five times. The value before “±” was the mean value, followed by the standard deviation.
| Methods | Accuracy | Precision | Recall | F1 |
|---|---|---|---|---|
| LSTM | 0.932 ± 0.001 | 0.931 ± 0.002 | 0.934 ± 0.003 | 0.933 ± 0.001 |
| 2-layer LSTM | 0.935 ± 0.001 | 0.932 ± 0.002 | 0.933 ± 0.002 | 0.932 ± 0.001 |
| BiLSTM | 0.931 ± 0.000 | 0.932 ± 0.004 | 0.932 ± 0.005 | 0.932 ± 0.000 |
| 2-layer BiLSTM | 0.930 ± 0.001 | 0.932 ± 0.003 | 0.929 ± 0.003 | 0.931 ± 0.001 |
| GRU | 0.867 ± 0.002 | 0.871 ± 0.003 | 0.864 ± 0.006 | 0.868 ± 0.002 |
| BiGRU | 0.930 ± 0.001 | 0.928 ± 0.002 | 0.933 ± 0.003 | 0.931 ± 0.001 |
| Character-level ConvNets | 0.928 ± 0.000 | 0.929 ± 0.004 | 0.930 ± 0.006 | 0.929 ± 0.001 |
| SLCABG | 0.934 ± 0.000 | 0.931 ± 0.005 | 0.937 ± 0.006 | 0.934 ± 0.000 |
| MDMLSM | 0.930 ± 0.001 | 0.931 ± 0.002 | 0.929 ± 0.003 | 0.930 ± 0.001 |
| MCNN-IFGS (word-based) | 0.970 ± 0.004 | 0.978 ± 0.004 | 0.966 ± 0.006 | 0.972 ± 0.004 |
| MCNN-IFGS (Ours) |
|
|
|
|
Performance of MCNN-IFGS on the Chinese review dataset. Here, different learning rates were used to carry out experiments. The value before “ ± ” was the mean value, followed by the standard deviation.
| Learning Rate | Accuracy | Precision | Recall | F1 |
|---|---|---|---|---|
| 0.001 |
|
|
|
|
| 0.002 | 0.965 ± 0.006 | 0.971 ± 0.007 | 0.966 ± 0.009 | 0.968 ± 0.005 |
| 0.003 | 0.969 ± 0.003 | 0.971 ± 0.002 | 0.973 ± 0.005 | 0.972 ± 0.002 |
| 0.004 | 0.955 ± 0.025 | 0.961 ± 0.012 | 0.957 ± 0.040 | 0.959 ± 0.024 |
| 0.005 | 0.933 ± 0.029 | 0.923 ± 0.031 | 0.957 ± 0.024 | 0.940 ± 0.026 |
| 0.006 | 0.913 ± 0.066 | 0.923 ± 0.055 | 0.917 ± 0.071 | 0.920 ± 0.062 |
| 0.007 | 0.890 ± 0.053 | 0.911 ± 0.047 | 0.886 ± 0.060 | 0.898 ± 0.050 |
| 0.008 | 0.818 ± 0.162 | 0.848 ± 0.121 | 0.777 ± 0.251 | 0.934 ± 0.209 |
Figure 5Model performance comparison of different learning rates on the Chinese review dataset. The value at the top of the violet dotted line represents the average of the standard deviations of all evaluation metrics scores corresponding to different dropout values.
MCNN-IFGS performance comparison of different dropout values on the Chinese review dataset. The value before “ ± ” was the mean value, followed by the standard deviation.
| Dropout | Accuracy | Precision | Recall | F1 |
|---|---|---|---|---|
| 0.2 | 0.964 ± 0.002 | 0.962 ± 0.002 | 0.973 ± 0.004 | 0.968 ± 0.002 |
| 0.3 | 0.962 ± 0.002 | 0.967 ± 0.007 | 0.965 ± 0.011 | 0.966 ± 0.002 |
| 0.4 | 0.966 ± 0.002 | 0.963 ± 0.006 | 0.975 ± 0.005 | 0.969 ± 0.002 |
| 0.5 | 0.967 ± 0.004 | 0.971 ± 0.010 | 0.968 ± 0.007 | 0.970 ± 0.003 |
| 0.6 | 0.968 ± 0.005 | 0.971 ± 0.004 | 0.971 ± 0.013 | 0.971 ± 0.005 |
| 0.7 | 0.967 ± 0.002 | 0.969 ± 0.005 | 0.971 ± 0.007 | 0.970 ± 0.002 |
| 0.8 |
|
|
|
|
Figure 6The test accuracy and test loss value change with the learning epochs. The value at the top of the spring green dashed line represents the learning epoch with the maximum test accuracy and its corresponding accuracy, and the value at the top of the lawn green dashed line represents the learning epoch with the minimum test loss and its corresponding test loss value.