| Literature DB >> 33921539 |
Quyen G To1, Kien G To2, Van-Anh N Huynh2, Nhung T Q Nguyen3, Diep T N Ngo2, Stephanie J Alley1, Anh N Q Tran2, Anh N P Tran2, Ngan T T Pham2, Thanh X Bui2, Corneel Vandelanotte1.
Abstract
Anti-vaccination attitudes have been an issue since the development of the first vaccines. The increasing use of social media as a source of health information may contribute to vaccine hesitancy due to anti-vaccination content widely available on social media, including Twitter. Being able to identify anti-vaccination tweets could provide useful information for formulating strategies to reduce anti-vaccination sentiments among different groups. This study aims to evaluate the performance of different natural language processing models to identify anti-vaccination tweets that were published during the COVID-19 pandemic. We compared the performance of the bidirectional encoder representations from transformers (BERT) and the bidirectional long short-term memory networks with pre-trained GLoVe embeddings (Bi-LSTM) with classic machine learning methods including support vector machine (SVM) and naïve Bayes (NB). The results show that performance on the test set of the BERT model was: accuracy = 91.6%, precision = 93.4%, recall = 97.6%, F1 score = 95.5%, and AUC = 84.7%. Bi-LSTM model performance showed: accuracy = 89.8%, precision = 44.0%, recall = 47.2%, F1 score = 45.5%, and AUC = 85.8%. SVM with linear kernel performed at: accuracy = 92.3%, Precision = 19.5%, Recall = 78.6%, F1 score = 31.2%, and AUC = 85.6%. Complement NB demonstrated: accuracy = 88.8%, precision = 23.0%, recall = 32.8%, F1 score = 27.1%, and AUC = 62.7%. In conclusion, the BERT models outperformed the Bi-LSTM, SVM, and NB models in this task. Moreover, the BERT model achieved excellent performance and can be used to identify anti-vaccination tweets in future studies.Entities:
Keywords: BERT; LSTM; deep learning; neural network; stance analysis; transformer; vaccine
Year: 2021 PMID: 33921539 PMCID: PMC8069687 DOI: 10.3390/ijerph18084069
Source DB: PubMed Journal: Int J Environ Res Public Health ISSN: 1660-4601 Impact factor: 3.390
Performance of the Bi-LSTM-128 models on the development set.
| Learning Rate | Epoch | Accuracy | Precision | Recall | F1 Score | AUC |
|---|---|---|---|---|---|---|
| 0.00003 | 10 | 77.2% | 26.2% | 74.9% | 38.9% | 84.2% |
| 20 | 78.7% | 28.1% | 76.9% | 41.1% | 85.1% | |
| 30 | 79.8% | 28.7% | 73.3% | 41.2% | 86.3% | |
| 40 | 82.3% | 31.2% | 68.6% | 42.9% | 87.1% | |
| 50 | 82.3% | 31.3% | 69.3% | 43.2% | 87.6% | |
| 60 | 82.7% | 31.8% | 68.6% | 43.4% | 87.7% | |
| 70 | 81.3% | 30.1% | 70.6% | 42.2% | 87.6% | |
| 80 | 82.5% | 31.9% | 71.3% | 44.0% | 88.0% | |
| 0.0001 | 10 | 80.7% | 29.9% | 73.9% | 42.6% | 87.2% |
| 20 | 80.5% | 29.9% | 75.6% | 42.8% | 88.3% | |
| 30 | 80.8% | 30.1% | 74.9% | 43.0% | 88.0% | |
| 40 | 85.8% | 37.0% | 66.3% | 47.5% | 88.2% | |
| 50 | 89.7% | 47.2% | 53.1% | 50.0% | 86.2% | |
|
|
|
|
|
|
| |
| 70 | 86.1% | 37.4% | 64.0% | 47.2% | 87.3% | |
| 80 | 88.8% | 43.9% | 54.1% | 48.4% | 85.0% | |
| 0.001 | 10 | 84.5% | 35.2% | 71.3% | 47.1% | 88.1% |
| 20 | 84.7% | 35.1% | 68.0% | 46.3% | 86.9% | |
| 30 | 90.0% | 47.9% | 41.6% | 44.5% | 83.8% | |
| 40 | 89.1% | 44.6% | 52.1% | 48.1% | 80.4% | |
| 50 | 90.7% | 53.3% | 34.3% | 41.8% | 74.5% | |
| 60 | 89.6% | 46.2% | 42.6% | 44.3% | 78.3% | |
| 70 | 88.6% | 42.1% | 47.2% | 44.5% | 79.4% | |
| 80 | 88.7% | 42.5% | 46.5% | 44.4% | 77.5% |
Performance of the BERT models on the development set.
| Learning Rate | Epoch | Accuracy | Precision | Recall | F1 Score | AUC |
|---|---|---|---|---|---|---|
| 0.00003 | 1 | 91.7% | 92.5% | 98.8% | 95.5% | 90.7% |
| 2 | 91.8% | 94.3% | 96.8% | 95.5% | 91.5% | |
| 3 | 92.2% | 94.6% | 96.8% | 95.7% | 86.5% | |
| 4 | 92.1% | 94.5% | 96.9% | 95.7% | 83.7% | |
| 5 | 91.7% | 94.5% | 96.4% | 95.4% | 79.8% | |
| 0.0001 | 1 | 92.1% | 93.5% | 98.0% | 95.7% | 91.0% |
| 2 | 92.0% | 94.1% | 97.2% | 95.6% | 91.4% | |
|
|
|
|
|
|
| |
| 4 | 92.1% | 94.5% | 96.9% | 95.7% | 84.6% | |
| 5 | 92.0% | 94.4% | 96.9% | 95.6% | 82.1% |
Performance of the SVM and NB models on the development set.
| Accuracy | Precision | Recall | F1 Score | AUC | |
|---|---|---|---|---|---|
| SVM-linear | 91.7% | 20.5% | 75.6% | 32.2% | 83.9% |
| SVM-poly | 90.8% | 9.2% | 66.7% | 16.2% | 78.9% |
| SVM-rbf | 91.1% | 12.5% | 74.5% | 21.5% | 83.0% |
| SVM-sigmoid | 91.5% | 17.5% | 75.7% | 28.4% | 83.8% |
| Complement NB | 88.8% | 25.4% | 38.1% | 30.5% | 65.2% |
| Multinomial NB | 90.6% | 5.6% | 68.0% | 10.4% | 79.4% |
Performance among Bi-LSTM, BERT, SVM, and NB on the test set.
| Accuracy | Precision | Recall | F1 Score | AUC | |
|---|---|---|---|---|---|
| Bi-LSTM-128, learning rate = 0.0001, epoch = 50 | 89.8% | 44.0% | 47.2% | 45.5% | 85.8% |
| BERT, learning rate = 0.0001, epoch = 3 | 91.6% | 93.4% | 97.6% | 95.5% | 84.7% |
| SVM-linear | 92.3% | 19.5% | 78.6% | 31.2% | 85.6% |
| Complement NB | 88.8% | 23.0% | 32.8% | 27.1% | 62.7% |