| Literature DB >> 35729983 |
Serge Nyawa1, Dieudonné Tchuente1, Samuel Fosso-Wamba1.
Abstract
Hesitant attitudes have been a significant issue since the development of the first vaccines-the WHO sees them as one of the most critical global health threats. The increasing use of social media to spread questionable information about vaccination strongly impacts the population's decision to get vaccinated. Developing text classification methods that can identify hesitant messages on social media could be useful for health campaigns in their efforts to address negative influences from social media platforms and provide reliable information to support their strategies against hesitant-vaccination sentiments. This study aims to evaluate the performance of different machine learning models and deep learning methods in identifying vaccine-hesitant tweets that are being published during the COVID-19 pandemic. Our concluding remarks are that Long Short-Term Memory and Recurrent Neural Network models have outperformed traditional machine learning models on detecting vaccine-hesitant messages in social media, with an accuracy rate of 86% against 83%.Entities:
Keywords: COVID-19; Deep learning; LSTM; Neural network; Text classification; Twitter; Vaccine hesitancy
Year: 2022 PMID: 35729983 PMCID: PMC9202977 DOI: 10.1007/s10479-022-04792-3
Source DB: PubMed Journal: Ann Oper Res ISSN: 0254-5330 Impact factor: 4.820
Synthesis of existing studies
| Paper | Disease(s) | Goal | SA | TM | Type SA | Method SA | Method TM | Other analysis | Data source | Period | Size |
|---|---|---|---|---|---|---|---|---|---|---|---|
| Tavoschi et al. ( | General (mainly meales) | Monitoring the public opinion on vaccination in Italy | x | Positive Negative Neutral | SVM | 13 months (sept 2016–Aug 2017) | 180, 620 tweets | ||||
| Zhou et al. ( | HPV | Examining if social connection information from tweets about human papillomavirus (HPV) vaccines could be used to train classifiers that identify antivaccine opinions | x | Antivaccine Other | SVM | Social network analysis | 7 months (Oct. 2013–March 2014) | 42,533 tweets | |||
| Hussain et al. ( | COVID-19 | Analyzing public sentiments on social media in the UK and the USA | x | Positive Negative Neutral | BERT | 9 months (March 2020–Nov 20,020) | 300,000 posts or tweets | ||||
| To et al. ( | COVID-19 | Evaluating the performance of different natural language processing models to identify anti-vaccination tweets | x | Antivaccine Other | BERT Bi-LSTM SVM Naïve Bayes | 8 months (Jan 2020–Aug 2020) | 1,651,687 tweets | ||||
| Ma et al. ( | COVID-19 | Comparing two different topic models to identify topics related to vaccine hesitancy in the USA | x | Top2Vec LDA | 3 months (Jan 2021–March 2021) | 3,403,166 tweets | |||||
| Rodríguez-González et al. ( | General (e.g., HPV, measles, Influenza, Hepatitis, Chickenpox, Varicela) | Identifying sentiment in tweets by using different machine-learning techniques and methods, and dealing with the unbalanced data problem in Spain | x | Negative Non-negative | C5.0 Logit Boost Bayesian GLM Neural Networks Random Forest SVM | 3 years (2015–2018) | 1,028,742 tweets | ||||
| Bar-Lev et al. ( | General (e.g. Hepatitis B, Diphtheria, Tetanus, Whooping cough, Polio, Pneumococcal, Rotavirus, MMR) | Using machine-learning strategies to assess how online content regarding vaccination affects vaccine hesitancy in Israel | x | Positive Negative Neutral | Logistic regression Random Forest Neural Networks Linear Regression | Tapuz | 5 years (2013–2018) | 9,596 posts on Facebook groups or Tapuz platform | |||
| Piedrahita-Valdés et al. ( | General | Evaluating public perceptions regarding vaccination and comparison among several countries | x | Positive Negative Neutral | Lexicon Analysis SVM | Trend analysis | 8 years (2011–2019) | 1,499,227 tweets | |||
| Yuan et al. ( | MMR | Examining emergent communities and social bots within the polarized online vaccination debate in Twitter | x | Pro-vaccine antivaccine Neutral | Logistic regression SVM kNN Nearest Centroid Naïve Bayes | Social network analysis | 2 months (Feb 2015–March 2015) | 669, 136 tweets Retweets relations | |||
| Furini ( | General (e.g. Measles, Autism, Rubella, Meningitis, Meningococcus, Polio, Tetanus) | Identifying psycho-linguistics signals of distrust toward vaccines to help health authorities to restore the trust toward vaccines in Italy | Psycho-linguistics and time-domain analyses | 2 years (Oct 2015–Oct 2017) | 172, 799 posts from ProVax and NoVax users | ||||||
| Jiang et al. ( | COVID-19 | Understanding how vaccine favorability and specific vaccine-related concerns were articulated and transmitted by Twitter users from opposing ideological camps and with different follower scopes in USA | x | x | Favorable to vax, Unfavorable to vax, Side effect, Distrust of medical professionals, conspiracy theory | BERT | Structural Topic Modeling (STM) | 4 months (March 2020–June 2020) | 16,959 tweets | ||
| Argyris et al. ( | General | Comparing discursive topics chosen by pro- and antivaccine advocates in their attempts to influence the public to accept or reject immunization in the engagement-persuasion spectrum | x | x | Pro-vaccine antivaccine Neutral | Logistic Regression | K-Means | 1 month (Nov 2019) | 39,962 tweets | ||
| Wang et al. ( | General | Developing an automatic detector for antivaccine messages to counteract the negative impact that antivaccine messages have on the public health using images, texts, and hashtags | x | antivaccine Other | SVM LSTM VGG, ResNet, RNN, EAN, MVAE | OCR | 3 years, 10 months (Jan 2016- Oct 2019) | 30,000 samples | |||
| Sear et al. ( | COVID-19 | Using machine learning to quantify COVID-19 content among online pro-vaccines and anti-vaccines | x | LDA | Trend analysis | 2 months (Jan2020–Feb 2020) | 8277 posts on Facebook pages | ||||
| Cotfas et al. ( | COVID-19 | Analyzing the dynamics of public opinion on Twitter in the first month after the start of the vaccination process in the UK, with a focus on COVID-19 vaccine hesitancy messages in connection with the major events in the analyzed period | x | x | In favor Neutral Against | Random Forest SVM Multinomial Naïve Bayes BERT RoBERTa | LDA | Trend analysis | 2 months (Dec 2020–Jan 2021) | 5,030,866 tweets | |
| Karami et al. ( | COVID-19 | Identifying the sentiment of tweets using a machine learning rule-based approach, discovers major topics, explores temporal trend, and compares topics of negative and non-negative tweets using statistical tests, and discloses top topics of tweets having negative and non-negative sentiment (in the USA) | x | x | Negative Non-negative | LIWC VADER BrandWatch | LDA | Trend analysis | 3 months (Nov 2020–Feb 2021) | 200,000 tweets | |
| Abd Rahim and Rafie ( | General (e.g. Measles) | Developing a model that uses SVM classifier to classify the polarity of sentiments: positive, negative and neutral | x | Positive Negative Neutral | SVM | 6 months (Oct 2019–March 2020) | 105,965 tweets | ||||
| Tomaszewski et al. ( | HPV | Developing a systematic and generalizable approach to identifying false HPV vaccine information on social media | x | x | True Information False Information | SVM Naïve Bayes CNN BiLSTM | DBSCAN | 4 years (2013–2017) | 705,858 tweets |
Tweets and labels
| Text Vaccine_acceptence | |
|---|---|
| News afternoon digest approval covid vaccine c… | Hesitant |
| Talking things missed mom revealed nothing mis… | Hesitant |
| One right mind would injected vax using totall… | Hesitant |
| Israelis found covid vaccine | Hesitant |
| One might agree accept cfr actually covid deat… | Hesitant |
| … | … |
| Opinion Nigerian vaccinated Europe let | Non Hesitant |
| Get second vaccination covid | Non Hesitant |
| Fully vaccinated zero complaints | Non Hesitant |
| Never thought would ever see world vaccination… | Hesitant |
| Wait covid got vaccinated covid need buy lotte… | Hesitant |
Additional numerical tweet features
| 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | |
|---|---|---|---|---|---|---|---|---|
| 0 | 9.8 | 56.97 | 28.0 | 1.5555 | 111.0 | 111.0 | 18.0 | 18.0 |
| 1 | 10.5 | 55.41 | 31.0 | 1.5500 | 135.0 | 135.0 | 20.0 | 18.0 |
| 2 | 17.4 | 25.27 | 55.0 | 1.7742 | 214.0 | 214.0 | 31.0 | 30.0 |
| 3 | 3.7 | 75.88 | 6.0 | 1.4999 | 28.0 | 28.0 | 4.0 | 4.0 |
| 4 | 18.0 | 6.94 | 48.0 | 2.0869 | 175.0 | 175.0 | 23.0 | 20.0 |
| … | … | … | … | … | … | … | … | … |
Hyperparameters of machine and deep learning algorithms
| ML algorithm | Hyperparameters chosen | Different values used |
|---|---|---|
| Logistic regression | solver penalty max_iter | ‘newton-cg’, ‘lbfgs’, ‘liblinear’, ‘sag’, ‘saga’ ’l1’,’l2’, ‘elasticnet’ 5000 |
| Random forest | n_estimators criterion max_features | 100, 200 “gini”, “entropy” ”auto” |
| Decision tree | max_features criterion min_samples_split | ”auto” “gini”, “entropy” 2 |
| Ada boosting | Type of estimator (base_estimator) Decision tree max depth (max_depth) Number of estimators (n_estimators) Learning rate (learning_rate) | DecisionTreeRegressor 8, 32 100, 200, 250 0.001, 0.05, 0.1 |
| Gradient boosting | Decision tree max depth (max_depth) Number of estimators (n_estimators) Learning rate (learning_rate) Loss function (loss) | 8,32 100, 200, 250 0.001, 0.05, 0.1 ‘deviance’, ‘exponential’ |
| K-nearest neighbors | Number of neighbors (n_neighbors) Neighbor’s weight function (weights) Neighbour’s algorithm (algorithm) | 5, 30, 100 uniform, distance ball_tree, kd_tree, brute, auto |
| Support vector classifier | Intercept fitting (fit_intercept) Regularization parameter (C) Max number of iterations (max_iter) Loss function (loss) | True, False 1.0, 2.0, 3.0 1000, 2000 ‘hinge’, ‘squared_hinge’ |
| Artificial neural networks | Network architecture (hidden_layer_sizes) Activation function (activation) Learning rate (learning_rate_init) Optimizer (solver) | 150, (150,50), (50, 20) relu, logistic 0.001, 0.005, 0.1 adam, lbfgs |
| Ligth LSTM | Embedding size (embedding_size) Batch size (batch_size) Epoch (epoch) Optimizer (optimizer) Dropout (dropout) Units (units) | 128, 300 16, 32, 64 10, 20, 30, 40, 50, 60, 70, 80, 90, 100 ‘adam’, ‘rmsprop’, ‘Adadelta’ 0.1, 0.2, 0.3 20, 30, 40 |
| LSTM | Embedding size (embedding_size) Batch size (batch_size) Epoch (epoch) Optimizer (optimizer) Dropout (dropout) Units of the RNN layer Units of dense layers | 128, 300 16, 32, 64 10, 20, 30, 40, 50, 60, 70, 80, 90, 100 ‘adam’, ‘rmsprop’, ‘Adadelta’ 0.1, 0.2, 0.3 25, 50 25, 50 |
| Recurrent neural network | Embedding size (embedding_size) Batch size (batch_size) Epoch (epoch) Optimizer (optimizer) Units of dense layers | 128, 300 16, 32, 64 10, 20, 30, 40, 50, 60, 70, 80, 90, 100 ‘adam’, ‘rmsprop’, ‘Adadelta’ 25, 50 |
Fig. 1Random forest.
Source: Image courtesy
Fig. 2Boosting.
Source: Image courtesy
Fig. 3K-nearest neighbors.
Source: Image courtesy
Fig. 4Support vector classifier.
Source: Image courtesy
Fig. 5Artificial neural networks.
Source: Image courtesy
Fig. 6Recurrent neural network. Reprinted from Graves et al. (2013)
Fig. 7Long short-term memory. Reprinted from Graves et al. (2013)
Fig. 8Experiment process
Fig. 9World cloud of hesitant comments
Performance of machine learning models on the testing sets
| Accuracy | Precision | Recall | Micro F1 | Macro F1 | Weighted F1 | F1 score | |
|---|---|---|---|---|---|---|---|
| LR | 0.825 | 0.607 | 0.604 | 0.797 | 0.725 | 0.798 | 0.605 |
| RF | 0.830 | 0.471 | 0.615 | 0.830 | 0.727 | 0.815 | 0.532 |
| SVC | 0.746 | 0.736 | 0.510 | 0.746 | 0.713 | 0.764 | 0.599 |
| KNN | 0.655 | 0.566 | 0.402 | 0.655 | 0.612 | 0.677 | 0.466 |
| GD | 0.686 | 0.674 | 0.448 | 0.686 | 0.654 | 0.707 | 0.534 |
| DT | 0.777 | 0.538 | 0.520 | 0.777 | 0.702 | 0.779 | 0.528 |
| Gboost | 0.812 | 0.581 | 0.620 | 0.812 | 0.749 | 0.816 | 0.600 |
| AdaBoost | 0.786 | 0.612 | 0.541 | 0.786 | 0.725 | 0.793 | 0.574 |
| ANN | 0.813 | 0.265 | 0.547 | 0.813 | 0.632 | 0.762 | 0.342 |
Performance of LSTM models on the validation sets
| Learning rate | Epoch | Training loss | Training accuracy | Validation loss | Validation accuracy |
|---|---|---|---|---|---|
| 0.0001 | 10 | 0.2077 | 0.9402 | 0.7576 | 0.8215 |
| 20 | 0.0858 | 0.9755 | 0.9843 | 0.8004 | |
| 30 | 0.0815 | 0.9774 | 1.0438 | 0.8089 | |
| 40 | 0.1290 | 0.9638 | 0.8218 | 0.8004 | |
| 50 | 0.0948 | 0.9718 | 1.0258 | 0.7709 | |
| 60 | 0.0770 | 0.9784 | 0.9674 | 0.8173 | |
| 70 | 0.0733 | 0.9807 | 0.9745 | 0.8046 | |
| 80 | 0.0871 | 0.9722 | 1.0035 | 0.8173 | |
| 90 | 0.0759 | 0.9760 | 1.0670 | 0.8089 | |
| 100 | 0.0832 | 0.9774 | 0.9803 | 0.8004 | |
| 0.0005 | 10 | 0.0364 | 0.9901 | 1.7323 | 0.8089 |
| 20 | 0.0315 | 0.9906 | 1.6416 | 0.7667 | |
| 30 | 0.0328 | 0.9882 | 1.6626 | 0.7793 | |
| 40 | 0.0396 | 0.9835 | 1.4843 | 0.8089 | |
| 50 | 0.0178 | 0.9939 | 2.0509 | 0.7878 | |
| 60 | 0.0293 | 0.9901 | 1.7441 | 0.7835 | |
| 70 | 0.0371 | 0.9868 | 1.6550 | 0.7371 | |
| 80 | 0.0175 | 0.9939 | 1.8112 | 0.7751 | |
| 90 | 0.0308 | 0.9911 | 1.6600 | 0.7920 | |
| 100 | 0.0426 | 0.9878 | 1.4149 | 0.7920 | |
| 0.001 | 10 | 0.0247 | 0.9915 | 1.9479 | 0.8089 |
| 20 | 0.0308 | 0.9906 | 1.7341 | 0.7709 | |
| 30 | 0.0295 | 0.9929 | 1.9028 | 0.8173 | |
| 40 | 0.0144 | 0.9953 | 2.1552 | 0.7751 | |
| 50 | 0.0235 | 0.9911 | 2.1632 | 0.8004 | |
| 60 | 0.0240 | 0.9920 | 1.8590 | 0.7751 | |
| 70 | 0.0208 | 0.9920 | 1.8483 | 0.8215 | |
| 80 | 0.0322 | 0.9896 | 1.7984 | 0.8173 | |
| 90 | 0.0324 | 0.9882 | 1.9341 | 0.7709 | |
| 100 | 0.0130 | 0.9948 | 2.3347 | 0.8004 | |
| 0.01 | 10 | 0.0603 | 0.9760 | 2.2411 | 0.8173 |
| 20 | 0.0539 | 0.9835 | 1.7254 | 0.7751 | |
| 30 | 0.0566 | 0.9831 | 2.0825 | 0.7835 | |
| 40 | 0.0483 | 0.9849 | 1.3854 | 0.7962 | |
| 50 | 0.0512 | 0.9807 | 2.0150 | 0.7751 | |
| 60 | 0.0432 | 0.9868 | 3.1149 | 0.8215 | |
| 70 | 0.0550 | 0.9807 | 1.4100 | 0.8131 | |
| 80 | 0.0346 | 0.9859 | 2.6846 | 0.7709 | |
| 90 | 0.0410 | 0.9864 | 1.9356 | 0.7835 | |
| 100 | 0.0525 | 0.9845 | 1.7164 | 0.8215 |
Performance of RNN models on the validation sets
| Learning rate | Epoch | Training loss | Training accuracy | Validation loss | Validation accuracy |
|---|---|---|---|---|---|
| 10 | 0.3816 | 0.8900 | 0.5605 | 0.8225 | |
| 20 | 0.2641 | 0.9217 | 0.5599 | 0.8242 | |
| 30 | 0.1821 | 0.9932 | 0.8249 | 0.8225 | |
| 50 | 0.1560 | 0.9992 | 1.0189 | 0.8225 | |
| 60 | 0.1472 | 0.9992 | 0.9325 | 0.8225 | |
| 70 | 0.1387 | 0.9996 | 1.0169 | 0.8208 | |
| 80 | 0.1337 | 0.9996 | 1.0942 | 0.8225 | |
| 90 | 0.1245 | 0.9996 | 1.1369 | 0.8191 | |
| 100 | 0.1171 | 1.0000 | 1.2516 | 0.8242 | |
| 0.0005 | 10 | 0.1683 | 0.9966 | 0.7601 | 0.8286 |
| 20 | 0.1222 | 0.9996 | 0.9602 | 0.8242 | |
| 30 | 8.6605e−04 | 1.0000 | 0.9204 | 0.8184 | |
| 40 | 3.4382e−04 | 1.0000 | 1.0501 | 0.8158 | |
| 50 | 0.0580 | 1.0000 | 1.0768 | 0.8140 | |
| 60 | 0.0486 | 1.0000 | 1.1318 | 0.8090 | |
| 70 | 4.9768e−05 | 1.0000 | 1.2876 | 0.6904 | |
| 80 | 4.7368e−05 | 1.0000 | 1.2700 | 0.8108 | |
| 90 | 2.4053e−05 | 1.0000 | 1.3996 | 0.8039 | |
| 100 | 1.9185e−05 | 1.0000 | 1.4046 | 0.6818 | |
| 0.001 | 10 | 0.0046 | 0.9992 | 0.7631 | 0.8286 |
| 20 | 0.0798 | 0.9996 | 0.8916 | 0.8140 | |
| 30 | 4.3638e−04 | 1.0000 | 1.1188 | 0.8225 | |
| 40 | 0.0338 | 1.0000 | 1.1959 | 0.8184 | |
| 50 | 6.2307e−05 | 1.0000 | 1.2890 | 0.8184 | |
| 60 | 0.0163 | 1.0000 | 1.2956 | 0.8090 | |
| 70 | 0.0115 | 1.0000 | 1.3688 | 0.8124 | |
| 80 | 0.0100 | 1.0000 | 1.1483 | 0.6988 | |
| 90 | 8.6857e−06 | 1.0000 | 1.5743 | 0.8022 | |
| 100 | 0.0045 | 1.0000 | 1.3327 | 0.8005 | |
| 0.01 | 10 | 0.5895 | 0.7240 | 0.5906 | 0.8225 |
| 20 | 0.5894 | 0.7240 | 0.5906 | 0.8225 | |
| 30 | 0.5894 | 0.7240 | 0.5907 | 0.8225 | |
| 40 | 4.8075e−06 | 1.0000 | 2.1718 | 0.6853 | |
| 50 | 0.5895 | 0.7240 | 0.5906 | 0.8225 | |
| 60 | 0.5896 | 0.7240 | 0.5906 | 0.8225 | |
| 70 | 0.5896 | 0.7240 | 0.5906 | 0.8225 | |
| 80 | 2.0296e−06 | 1.0000 | 2.0767 | 0.6954 | |
| 90 | 0.5894 | 0.7240 | 0.5907 | 0.8225 | |
| 100 | 6.6937e−07 | 1.0000 | 2.1622 | 0.6880 |
Performance of light LSTM models on the validation sets
| Learning rate | Epoch | Training loss | Training accuracy | Validation loss | Validation accuracy |
|---|---|---|---|---|---|
| 10 | 0.2925 | 0.8993 | 0.4976 | 0.8637 | |
| 20 | 0.1657 | 0.9647 | 0.5961 | 0.8637 | |
| 30 | 0.1579 | 0.9581 | 0.5990 | 0.8553 | |
| 40 | 0.1684 | 0.9638 | 0.5661 | 0.8553 | |
| 60 | 0.1657 | 0.9605 | 0.6230 | 0.8637 | |
| 70 | 0.1597 | 0.9624 | 0.5922 | 0.8553 | |
| 80 | 0.1549 | 0.9605 | 0.6110 | 0.8637 | |
| 90 | 0.1843 | 0.9544 | 0.5814 | 0.8553 | |
| 100 | 0.1527 | 0.9675 | 0.5930 | 0.8637 | |
| 0.0005 | 10 | 0.0257 | 0.9958 | 1.0331 | 0.7962 |
| 20 | 0.0290 | 0.9962 | 1.0413 | 0.8173 | |
| 30 | 0.0323 | 0.9939 | 0.9663 | 0.8131 | |
| 40 | 0.0248 | 0.9962 | 1.1099 | 0.8131 | |
| 50 | 0.0297 | 0.9958 | 1.0660 | 0.8173 | |
| 60 | 0.0274 | 0.9962 | 1.0400 | 0.8004 | |
| 70 | 0.0384 | 0.9920 | 0.9841 | 0.8215 | |
| 80 | 0.0391 | 0.9929 | 0.9981 | 0.8468 | |
| 90 | 0.0259 | 0.9953 | 1.0317 | 0.8215 | |
| 100 | 0.0339 | 0.9929 | 1.0758 | 0.8215 | |
| 0.001 | 10 | 0.0108 | 0.9972 | 1.2537 | 0.8257 |
| 20 | 0.0167 | 0.9962 | 1.3142 | 0.7962 | |
| 30 | 0.0122 | 0.9962 | 1.4101 | 0.7793 | |
| 40 | 0.0096 | 0.9981 | 1.3247 | 0.8046 | |
| 50 | 0.0136 | 0.9962 | 1.2704 | 0.7962 | |
| 60 | 0.0189 | 0.9967 | 1.2359 | 0.7667 | |
| 70 | 0.0099 | 0.9976 | 1.2931 | 0.8046 | |
| 80 | 0.0142 | 0.9958 | 1.2112 | 0.8215 | |
| 90 | 0.0124 | 0.9972 | 1.2930 | 0.7920 | |
| 100 | 0.0101 | 0.9976 | 1.2897 | 0.7920 | |
| 0.01 | 10 | 0.0025 | 0.9995 | 1.8656 | 0.8004 |
| 20 | 0.0039 | 0.9991 | 1.7533 | 0.8300 | |
| 30 | 0.0080 | 0.9976 | 1.5326 | 0.8215 | |
| 40 | 0.0033 | 0.9981 | 1.4588 | 0.8342 | |
| 50 | 0.0028 | 0.9991 | 1.7429 | 0.8257 | |
| 60 | 0.0031 | 0.9995 | 1.7227 | 0.7878 | |
| 70 | 0.0037 | 0.9995 | 1.6904 | 0.7793 | |
| 80 | 0.0039 | 0.9986 | 1.6116 | 0.7962 | |
| 90 | 0.0059 | 0.9991 | 1.4797 | 0.8342 | |
| 100 | 0.0048 | 0.9981 | 1.6606 | 0.8173 |
Performance of LSTM models on the testing sets
| Learning rate | Epoch | Accuracy | Precision | Recall | F1 score |
|---|---|---|---|---|---|
| 10 | 0.83 | 0.80 | 0.83 | 0.69 | |
| 20 | 0.85 | 0.83 | 0.85 | 0.83 | |
| 40 | 0.85 | 0.83 | 0.85 | 0.83 | |
| 50 | 0.81 | 0.81 | 0.81 | 0.81 | |
| 60 | 0.85 | 0.82 | 0.85 | 0.82 | |
| 70 | 0.82 | 0.81 | 0.82 | 0.81 | |
| 80 | 0.83 | 0.81 | 0.83 | 0.81 | |
| 90 | 0.83 | 0.81 | 0.83 | 0.82 | |
| 100 | 0.84 | 0.82 | 0.84 | 0.82 | |
| 0.0005 | 10 | 0.83 | 0.80 | 0.83 | 0.81 |
| 20 | 0.80 | 0.80 | 0.80 | 0.80 | |
| 30 | 0.82 | 0.81 | 0.82 | 0.81 | |
| 40 | 0.82 | 0.82 | 0.82 | 0.82 | |
| 50 | 0.83 | 0.80 | 0.83 | 0.81 | |
| 60 | 0.83 | 0.83 | 0.83 | 0.83 | |
| 70 | 0.82 | 0.82 | 0.82 | 0.82 | |
| 80 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 90 | 0.82 | 0.80 | 0.82 | 0.81 | |
| 100 | 0.83 | 0.81 | 0.83 | 0.81 | |
| 0.001 | 10 | 0.83 | 0.81 | 0.83 | 0.81 |
| 20 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 30 | 0.82 | 0.80 | 0.82 | 0.81 | |
| 40 | 0.82 | 0.80 | 0.82 | 0.81 | |
| 50 | 0.82 | 0.80 | 0.82 | 0.80 | |
| 60 | 0.82 | 0.82 | 0.82 | 0.82 | |
| 70 | 0.83 | 0.80 | 0.83 | 0.81 | |
| 80 | 0.84 | 0.82 | 0.84 | 0.83 | |
| 90 | 0.81 | 0.81 | 0.81 | 0.81 | |
| 100 | 0.81 | 0.69 | 0.81 | 0.80 | |
| 0.01 | 10 | 0.82 | 0.80 | 0.82 | 0.80 |
| 20 | 0.82 | 0.81 | 0.82 | 0.81 | |
| 30 | 0.69 | 0.69 | 0.69 | 0.69 | |
| 40 | 0.81 | 0.69 | 0.81 | 0.80 | |
| 50 | 0.68 | 0.69 | 0.68 | 0.69 | |
| 60 | 0.84 | 0.82 | 0.84 | 0.82 | |
| 70 | 0.82 | 0.69 | 0.82 | 0.69 | |
| 80 | 0.82 | 0.82 | 0.82 | 0.82 | |
| 90 | 0.80 | 0.68 | 0.80 | 0.69 | |
| 100 | 0.82 | 0.80 | 0.82 | 0.81 |
Performance of RNN models on the testing sets
| Learning rate | Epoch | Accuracy | Precision | Recall | F1 score |
|---|---|---|---|---|---|
| 0.0001 | 10 | 0.82 | 0.62 | 0.82 | 0.71 |
| 20 | 0.82 | 0.80 | 0.82 | 0.71 | |
| 30 | 0.82 | 0.62 | 0.82 | 0.71 | |
| 40 | 0.83 | 0.80 | 0.83 | 0.79 | |
| 50 | 0.82 | 0.52 | 0.82 | 0.71 | |
| 60 | 0.82 | 0.77 | 0.82 | 0.71 | |
| 70 | 0.82 | 0.71 | 0.82 | 0.71 | |
| 80 | 0.82 | 0.77 | 0.82 | 0.71 | |
| 90 | 0.82 | 0.59 | 0.82 | 0.71 | |
| 100 | 0.82 | 0.79 | 0.82 | 0.72 | |
| 20 | 0.82 | 0.78 | 0.82 | 0.73 | |
| 30 | 0.82 | 0.79 | 0.82 | 0.80 | |
| 40 | 0.82 | 0.79 | 0.82 | 0.80 | |
| 50 | 0.81 | 0.75 | 0.81 | 0.75 | |
| 60 | 0.81 | 0.74 | 0.81 | 0.74 | |
| 70 | 0.79 | 0.78 | 0.79 | 0.78 | |
| 80 | 0.81 | 0.79 | 0.81 | 0.79 | |
| 90 | 0.80 | 0.78 | 0.80 | 0.78 | |
| 100 | 0.78 | 0.78 | 0.78 | 0.78 | |
| 0.001 | 10 | 0.83 | 0.80 | 0.83 | 0.80 |
| 20 | 0.81 | 0.77 | 0.81 | 0.77 | |
| 30 | 0.82 | 0.79 | 0.82 | 0.79 | |
| 40 | 0.82 | 0.78 | 0.82 | 0.78 | |
| 50 | 0.82 | 0.79 | 0.82 | 0.80 | |
| 60 | 0.81 | 0.78 | 0.81 | 0.78 | |
| 70 | 0.81 | 0.78 | 0.81 | 0.78 | |
| 80 | 0.80 | 0.78 | 0.80 | 0.78 | |
| 90 | 0.80 | 0.78 | 0.80 | 0.79 | |
| 100 | 0.80 | 0.78 | 0.80 | 0.78 | |
| 0.01 | 10 | 0.82 | 0.52 | 0.82 | 0.71 |
| 20 | 0.82 | 0.52 | 0.82 | 0.71 | |
| 30 | 0.82 | 0.52 | 0.82 | 0.71 | |
| 40 | 0.79 | 0.78 | 0.79 | 0.78 | |
| 50 | 0.82 | 0.52 | 0.82 | 0.71 | |
| 60 | 0.82 | 0.52 | 0.82 | 0.71 | |
| 70 | 0.82 | 0.52 | 0.82 | 0.71 | |
| 80 | 0.80 | 0.77 | 0.80 | 0.78 | |
| 90 | 0.82 | 0.52 | 0.82 | 0.71 | |
| 100 | 0.79 | 0.75 | 0.79 | 0.77 |
Performance of light LSTM models on the testing sets
| Learning rate | Epoch | Accuracy | Precision | Recall | F1 score |
|---|---|---|---|---|---|
| 0.0001 | 10 | 0.84 | 0.81 | 0.84 | 0.79 |
| 20 | 0.84 | 0.81 | 0.84 | 0.81 | |
| 30 | 0.83 | 0.81 | 0.83 | 0.82 | |
| 40 | 0.84 | 0.82 | 0.84 | 0.82 | |
| 50 | 0.85 | 0.83 | 0.85 | 0.83 | |
| 60 | 0.83 | 0.80 | 0.83 | 0.81 | |
| 70 | 0.84 | 0.82 | 0.84 | 0.82 | |
| 80 | 0.83 | 0.81 | 0.83 | 0.81 | |
| 90 | 0.83 | 0.81 | 0.83 | 0.82 | |
| 100 | 0.85 | 0.82 | 0.85 | 0.81 | |
| 0.0005 | 10 | 0.83 | 0.82 | 0.83 | 0.82 |
| 20 | 0.83 | 0.82 | 0.83 | 0.83 | |
| 30 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 40 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 50 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 60 | 0.81 | 0.81 | 0.81 | 0.81 | |
| 70 | 0.81 | 0.80 | 0.81 | 0.81 | |
| 80 | 0.84 | 0.82 | 0.84 | 0.83 | |
| 90 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 100 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 10 | 0.83 | 0.81 | 0.83 | 0.82 | |
| 20 | 0.82 | 0.81 | 0.82 | 0.82 | |
| 30 | 0.81 | 0.81 | 0.81 | 0.81 | |
| 40 | 0.82 | 0.82 | 0.82 | 0.82 | |
| 50 | 0.82 | 0.82 | 0.82 | 0.82 | |
| 60 | 0.82 | 0.82 | 0.82 | 0.82 | |
| 70 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 90 | 0.81 | 0.80 | 0.81 | 0.81 | |
| 100 | 0.82 | 0.82 | 0.82 | 0.82 | |
| 0.01 | 10 | 0.80 | 0.80 | 0.80 | 0.80 |
| 20 | 0.82 | 0.80 | 0.82 | 0.80 | |
| 30 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 40 | 0.83 | 0.80 | 0.83 | 0.81 | |
| 50 | 0.82 | 0.81 | 0.82 | 0.81 | |
| 60 | 0.81 | 0.82 | 0.81 | 0.81 | |
| 70 | 0.83 | 0.82 | 0.83 | 0.82 | |
| 80 | 0.81 | 0.80 | 0.81 | 0.80 | |
| 90 | 0.83 | 0.81 | 0.83 | 0.82 | |
| 100 | 0.80 | 0.69 | 0.80 | 0.69 |