| Literature DB >> 35539003 |
Abstract
The whole world has been experiencing the COVID-19 pandemic since December 2019. During the pandemic, a new life has been started by necessity where people have extensively used social media to express their feelings, and find information. Twitter was used as the source of what people have shared regarding the COVID-19 pandemic. Sentiment analysis deals with the extraction of the sentiment of a given text. Most of the related works deal with sentiment analysis in English, while studies for Turkish sentiment analysis lack in the research field. To this end, a novel sentiment analysis model based on the combination of convolutional neural network and bidirectional long short-term memory was proposed in this study. The proposed deep neural network model was trained on the constructed Twitter dataset, which consists of 15 k Turkish tweets regarding the COVID-19 pandemic, to classify a given tweet into three sentiment classes, namely, (i) positive , (ii) negative , and (iii) neutral . A set of experiments were conducted for the evaluation of the proposed model. According to the experimental result, the proposed model obtained an accuracy as high as 97.895 % , which outperformed the state-of-the-art baseline models for sentiment analysis of tweets in Turkish.Entities:
Keywords: COVID‐19; Twitter; deep neural network; sentiment analysis; text classification
Year: 2022 PMID: 35539003 PMCID: PMC9074424 DOI: 10.1002/cpe.6883
Source DB: PubMed Journal: Concurr Comput ISSN: 1532-0626 Impact factor: 1.831
The comparison of the related work in terms of (i) employed technique(s), (ii) covered emotions, and (iii) drawback(s)
| Related work | Employed technique(s) | Covered emotions | Drawback(s) |
|---|---|---|---|
| Coban et al. | Traditional ML algorithms | ( | Labeling tweets through the predefined set of emoticons |
| Akgun et al. | Lexicon‐based techniques | ( | Removal of emotional expressions during preprocessing |
| Assigning the thresholds for the emotion classes in a not‐formulaic way | |||
| Demirci | Traditional ML algorithms | ( | Not employing DNNs alongside the traditional ML algorithms |
| Ileri | User‐centric techniques | ( | Removal of emotional expressions during preprocessing |
| Ignorance of weighted edges as all edge weights were assumed to equal to | |||
| Tocoglu et al. | Lexicon‐based techniques and DNNs | ( | Not employing the cross‐validation technique |
| Demirci et al. | Traditional ML algorithms, and DNNs | ( | Removal of emotional expressions during preprocessing |
| The constructed dataset is too small to effectively train a DNN | |||
| Ucan et al. | Pre‐trained language models, traditional ML algorithms, and DNNs | ( | Absence of the employment of the features that target the emotional contexts of tweets |
| Onan | Pre‐trained language models, and DNNs | ( | Absence of the hyperparameter optimization task |
| Vo et al. | DNNs | ( | Absence of the employment of text preprocessing techniques |
| Ombabi et al. | Pre‐trained language models, traditional ML algorithms, and DNNs | ( | Hyperparameters of the proposed DNNs were empirically determined |
FIGURE 1The sequence length distribution of the tweets in the constructed dataset
The number of samples that each sentiment class of the final dataset consisted of
| Sentiment class | Number of samples |
|---|---|
|
|
|
|
|
|
|
|
|
FIGURE 2An overview of the architecture of the proposed DNN model, which is a combination of CNN and BiLSTM
The evaluated values for the hyper‐parameters of the proposed model
| Hyper‐parameter | Evaluated values |
|---|---|
| Number of filters of the |
|
| Kernel size of the |
|
| Number of units of the |
|
| Optimization algorithm |
|
| Learning rate of the optimization algorithm |
|
| Size of vector space of the |
|
| Activation function (except for the last |
|
| Batch size |
|
| Dropout rate |
|
The layers of the proposed model, including the employed hyper‐parameters
| Layer # | Layer type | Hyper‐parameters with their values |
|---|---|---|
|
|
|
Vocabulary size: |
|
Dimension of embedding: | ||
|
Length of input sequences: | ||
|
|
|
Number of filters: |
|
Kernel size: | ||
|
Activation function: | ||
|
|
|
Pool size: |
|
|
|
Number of filters: |
|
Kernel size: | ||
|
Activation function: | ||
|
|
|
Pool size: |
| 6 |
|
Number of units: |
|
|
|
Dropout rate: |
|
|
|
Number of units: |
|
|
|
Dropout rate: |
|
|
|
|
|
|
|
Number of units: |
|
Activation function: | ||
|
|
|
Dropout rate: |
|
|
|
Number of units: |
|
Activation function: | ||
|
|
Learning rate: | |
FIGURE 3The plots of the calculated accuracy values for the training and validation sets over the epochs
FIGURE 4The confusion matrix of the evaluation of the proposed model on the test set
The experimental modification with their effects on the accuracy obtained by the proposed model
| Modification on the proposed model | Change on accuracy (%) |
|---|---|
| Adding a |
|
| Adding another |
|
| Increasing the number of units in |
|
| Removing the |
|
FIGURE 5The generated word clouds of (A) all tweets, (B) the tweets labeled , and (C) the tweets labeled as
The comparison of the proposed model with the baseline models in terms of their efficiencies of classifying the given tweets
| Related work | Accuracy (%) | F1‐score (%) | Precision (%) | Recall (%) |
|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|