| Literature DB >> 36091222 |
Abstract
COVID-19 has emerged as the greatest threat in recent times, causing extensive mortality and morbidity in the entire world. India is among the highly affected countries suffering severe disruptions due this pandemic. To overcome the adverse effects of COVID-19, vaccination has been identified as the most effective preventive measure globally. However, a growing amount of hesitancy has been observed among the general public regarding the efficacy and possible side-effects of vaccination. Such hesitancy may proved to be the greatest hindrance towards combating this deadly pandemic. This paper introduces a multimodal deep learning method for Indian Twitter user classification, leveraging both content-based and network-based features. To explore the fundamental features of different modalities, improvisations of transformer models, BERT and GraphBERT are utilized to encode the textual and network structure information. The proposed approach thus integrates multiple data representations, utilizing the advances in both transformer based deep learning as well as multimodal learning. Experimental results demonstrates the efficacy of proposed approach over state of the art approaches. Aggregated feature representations from multiple modalities embed additional information that improves the classification results. The findings of the proposed model has been further utilized to perform a study on the dynamics of COVID-19 vaccine hesitancy in India.Entities:
Keywords: COVID-19; Feature learning; Multimodal learning; Social media; Social network analysis
Year: 2022 PMID: 36091222 PMCID: PMC9449921 DOI: 10.1007/s10844-022-00745-1
Source DB: PubMed Journal: J Intell Inf Syst ISSN: 0925-9902 Impact factor: 2.504
Fig. 1Number of confirmed cases and doses in India in 2021
Fig. 2Proposed Transformer based Multimodal Architecture
Fig. 3Example Twitter Retweet Graph
Graph metrics for retweet graph of the users
| Metric | Value |
|---|---|
| No. of Nodes | 35,432 |
| No. of Edges | 42,763 |
| Average Degree | 2.41 |
| Average Clustering Coefficient | 0.023 |
| Network Density | 2.4E-03 |
| Transitivity | 0.084 |
Hyper-parameters considered for pre-trained models
| Model | Parameter | Value |
|---|---|---|
| GraphBERT | Hidden Layer Number | 2 |
| Subgraph Size | 7 | |
| Learning Rate | 5e-4 | |
| Hidden Size | 32 | |
| Hidden Dropout Rate | 0.5 | |
| Attention Head Number | 3 | |
| Attention Dropout Rate | 0.4 | |
| Weight Decay | 0.01 | |
| BERT | Number of Transformer Blocks | 12 |
| Training Batch Size | 32 | |
| Learning Rate | 5e-4 | |
| Weight Decay | 0.01 |
Comparative analysis with unimodal and multimodal approaches
| Modality | Model | Accuracy(%) | Precision(%) | Recall(%) | F1-score(%) | |
|---|---|---|---|---|---|---|
| Unimodal | Text | LSTM | 77.24 | 78.36 | 81.52 | 79.92 |
| Text | BiLSTM | 79.74 | 83.32 | 81.21 | 82.23 | |
| Text | BERT | 83.45 | 81.67 | 82.64 | 82.15 | |
| Graph | GCN | 86.42 | 87.34 | 85.57 | 86.44 | |
| Graph | GraphBERT | 87.36 | 86.56 | 87.79 | 87.19 | |
| Multimodal | Text+Graph | Doc2Vec+Node2Vec | 90.17 | 89.64 | 88.75 | 89.19 |
| Text+Graph | TFIDF+Doc2Vec+Node2Vec | 92.34 | 91.23 | 90.34 | 90.74 | |
| Text+Graph | Proposed | 95.24 | 94.67 | 92.34 | 93.48 |
Results on different modalities
| Modality | Accuracy(%) | Precision(%) | Recall(%) | F1-score(%) |
|---|---|---|---|---|
| Text | 86.35 | 86.43 | 83.12 | 84.72 |
| Graph | 87.36 | 86.56 | 87.79 | 87.19 |
| Proposed | 95.24 | 94.67 | 92.34 | 93.48 |
Fig. 4Confusion matrix of proposed approach using different modalities
Fig. 5Hesitancy Dynamics