| Literature DB >> 33967391 |
Ishaani Priyadarshini1, Chase Cotton1.
Abstract
As the number of users getting acquainted with the Internet is escalating rapidly, there is more user-generated content on the web. Comprehending hidden opinions, sentiments, and emotions in emails, tweets, reviews, and comments is a challenge and equally crucial for social media monitoring, brand monitoring, customer services, and market research. Sentiment analysis determines the emotional tone behind a series of words may essentially be used to understand the attitude, opinions, and emotions of users. We propose a novel long short-term memory (LSTM)-convolutional neural networks (CNN)-grid search-based deep neural network model for sentiment analysis. The study considers baseline algorithms like convolutional neural networks, K-nearest neighbor, LSTM, neural networks, LSTM-CNN, and CNN-LSTM which have been evaluated using accuracy, precision, sensitivity, specificity, and F-1 score, on multiple datasets. Our results show that the proposed model based on hyperparameter optimization outperforms other baseline models with an overall accuracy greater than 96%.Entities:
Keywords: Convolutional neural networks (CNN); Deep neural network; Grid search; Long short-term memory (LSTM); Sentiment analysis
Year: 2021 PMID: 33967391 PMCID: PMC8097246 DOI: 10.1007/s11227-021-03838-w
Source DB: PubMed Journal: J Supercomput ISSN: 0920-8542 Impact factor: 2.474
Sentiment analysis methodologies based on past research works
| References | Methodology |
|---|---|
| Basiri et al. [ | Attention-based bidirectional CNN–RNN deep model |
| Jin et al. [ | Heterogeneous graph network embedding |
| Pota et al. [ | BERT-based pipeline |
| Lu et al. [ | Aspect-gated graph convolutional networks |
| Nemes and Kiss [ | Recurrent neural networks |
| Tubishat et al. [ | Explicit aspects extraction using optimal rules combination |
| Kandasamy et al. [ | Refined neutrosophic sets |
| Huang et al. [ | Attention–emotion-enhanced convolutional LSTM |
| Zhao et al. [ | Combination of convolutional neural network and gated recurrent unit |
| Srividya and Sowjanya [ | Neural attention-based model |
Fig. 1CNN–LSTM architecture
Fig. 2LSTM–CNN architecture
Fig. 3Architecture of the proposed model
Grid search hyperparameters
| Hyperparameters | Values |
|---|---|
| Batch size | 10, 20, 40, 60, 80, 100 |
| Epcohes | 10, 50, 100 |
| Optimization | SGD, RMSprop, Adagrad, Adadelta, Adam, Adamax, Nadam |
| Drop regularization | 0.0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9 |
| Neurons in hidden layer | 1, 5, 10, 15, 20, 25, 30 |
Fig. 4Overall methodology of the proposed work
Fig. 5Confusion matrix
Performance evaluation of Dataset 1
| AI models | Accuracy | Precision | Sensitivity | Specificity | F-1 score |
|---|---|---|---|---|---|
| 0.779 | 0.835 | 0.867 | 0.928 | 0.85 | |
| NN | 0.897 | 0.90 | 0.90 | 0.915 | 0.90 |
| CNN | 0.927 | 0.922 | 0.918 | 0.913 | 0.928 |
| LSTM | 0.905 | 0.912 | 0.897 | 0.929 | 0.915 |
| CNN–LSTM | 0.945 | 0.939 | 0.925 | 0.925 | 0.931 |
| LSTM–CNN | 0.958 | 0.943 | 0.933 | 0.949 | 0.939 |
Performance evaluation of Dataset 2
| AI models | Accuracy | Precision | Sensitivity | Specificity | F-1 score |
|---|---|---|---|---|---|
| 0.889 | 0.896 | 0.891 | 0.918 | 0.889 | |
| NN | 0.892 | 0.918 | 0.911 | 0.92 | 0.911 |
| CNN | 0.927 | 0.918 | 0.92 | 0.927 | 0.92 |
| LSTM | 0.915 | 0.909 | 0.92 | 0.919 | 0.91 |
| CNN–LSTM | 0.938 | 0.929 | 0.936 | 0.92 | 0.92 |
| LSTM–CNN | 0.947 | 0.931 | 0.922 | 0.94 | 0.959 |
Fig. 6a Model accuracy across datasets, b model precision across datasets, c model sensitivity across datasets, d model specificity across datasets, e model F-1 score across datasets
Comparative analysis of our proposed work with previous works
| References | Study/research | Methodology/parameters | Results |
|---|---|---|---|
| Das and Chakraborty [ | Sentiment analysis (Amazon, IMDB) | TF-IDF and next word negation | Accuracy is 88.58% and 89.91% |
| Güner et al. [ | Sentiment analysis (Amazon) | Machine learning techniques | LSTM performs the best with an accuracy of 90% |
| Bajeh et al. [ | Performance analysis of particle swarm optimization | C4.5 decision tree, | Accuracy < 71.23%, recall < 0.713 |
| Iqbal et al. [ | Hybrid framework for sentiment analysis | Genetic algorithm-based feature reduction | Accuracy for (IMDB < 76.7%, Amazon < 77.9%, Yelp < 75.3%) |
| Shreshtha and Nasoz [ | Sentiment analysis (Amazon dataset) | Deep learning | Accuracy ~ 0.81, precision ~ 0.58, recall ~ 0.40 |
| Rehman et al. [ | Sentiment analysis (IMDB Movie Reviews) | Hybrid CNN–LSTM model | Accuracy is 91% |
| Tammina [ | Sentiment classification in Telugu language | Hybrid learning (Lexicon based and machine learning) | Accuracy is 85% |
| Wang et al. [ | Negative sentiment analysis during COVID-19 pandemic | Fine-tuned BERT (Bidirectional Encoder Representations from Transformers) | Accuracy is 75.65% |
| Dong et al. [ | Sentiment analysis | Capsule network model with BiLSTM | Accuracy is 81.47%, 91.96%, and 48.34% for three datasets, respectively |