| Literature DB >> 35615261 |
Shouzhi Xu1, Xiaodi Liu1, Kai Ma1, Fangmin Dong1, Basheer Riskhan1, Shunzhi Xiang1, Changsong Bing1.
Abstract
In the era of the Internet and big data, online social media platforms have been developing rapidly, which accelerate rumors circulation. Rumor detection on social media is a worldwide challenging task due to rumor's feature of high speed, fragmental information and extensive range. Most existing approaches identify rumors based on single-layered hybrid features like word features, sentiment features and user characteristics, or multimodal features like the combination of text features and image features. Some researchers adopted the hierarchical structure, but they neither used rumor propagation nor made full use of its retweet posts. In this paper, we propose a novel model for rumor detection based on Graph Neural Networks (GNN), named Hierarchically Aggregated Graph Neural Networks (HAGNN). This task focuses on capturing different granularities of high-level representations of text content and fusing the rumor propagation structure. It applies a Graph Convolutional Network (GCN) with a graph of rumor propagation to learn the text-granularity representations with the spreading of events. A GNN model with a document graph is employed to update aggregated features of both word and text granularity, it helps to form final representations of events to detect rumors. Experiments on two real-world datasets demonstrate the superiority of the proposed method over the baseline methods. Our model achieves the accuracy of 95.7% and 88.2% on the Weibo dataset Ma et al. 2017 and the CED dataset Song et al. IEEE Trans Knowl Data Eng 33(8):3035-3047, 2019respectively.Entities:
Keywords: Graph neural networks; Hierarchical aggregation; Rumor detection; Rumor propagation
Year: 2022 PMID: 35615261 PMCID: PMC9122810 DOI: 10.1007/s10489-022-03592-3
Source DB: PubMed Journal: Appl Intell (Dordr) ISSN: 0924-669X Impact factor: 5.019
Fig. 1Our HAGNN rumor detection model
Statistics of the datasets
| Statistic | CED | |
|---|---|---|
| # of events | 4664 | 3387 |
| # of Rumors | 2351 | 1538 |
| # of Non-rumors | 2313 | 1849 |
| # of Posts | 3,805,656 | 1,217,212 |
| # of Users | 2,746,818 | 771,960 |
Confusion Matrix
| Actual | |||
|---|---|---|---|
| Positive | Negative | ||
| Predicted | Positive | TP | FP |
| Negative | FN | TN | |
Fig. 2Comparison of different methods on Weibo dataset
Results on the Weibo dataset (F:False Rumor,T:True Rumor)
| Method | Class | Accuracy | Precision | Recall | F1 |
|---|---|---|---|---|---|
| SVM-RBF∗ | F | 0.818 | 0.822 | 0.812 | 0.817 |
| T | 0.815 | 0.824 | 0.819 | ||
| TextGCN | F | 0.837 | 0.809 | 0.840 | 0.824 |
| T | 0.862 | 0.835 | 0.848 | ||
| TextING | F | 0.842 | 0.851 | 0.844 | 0.848 |
| T | 0.832 | 0.839 | 0.836 | ||
| RvNN∗ | F | 0.908 | 0.912 | 0.897 | 0.905 |
| T | 0.904 | 0.918 | 0.911 | ||
| PPC_RNN+CNN∗ | F | 0.916 | 0.884 | 0.957 | 0.919 |
| T | 0.955 | 0.876 | 0.913 | ||
| HAGNN | F | ||||
| T |
The bold entries are used to highlight the results of our proposed model
Results on the CED dataset (F:False Rumor,T:True Rumor)
| Method | Class | Accuracy | Precision | Recall | F1 |
|---|---|---|---|---|---|
| TextING | F | 0.842 | 0.851 | 0.844 | 0.848 |
| T | 0.832 | 0.839 | 0.836 | ||
| TextGCN | F | 0.873 | 0.913 | 0.846 | 0.878 |
| T | 0.833 | 0.906 | 0.868 | ||
| F | |||||
| T |
The bold entries are used to highlight the results of our proposed model
Fig. 3Results on the comparison of GNN layer number
Fig. 4Results on the comparison of GNN unit number
Fig. 5Results on the comparison of GNN window size
Fig. 6The comparison of our model, GGNN and GCN
Fig. 7The comparison of our model and GGNN