| Literature DB >> 32565619 |
Zhiyuan Wu1, Dechang Pi1, Junfu Chen1, Meng Xie1, Jianjun Cao2.
Abstract
Rumors on social media have always been an important issue that seriously endangers social security. Researches on timely and effective detection of rumors have aroused lots of interest in both academia and industry. At present, most existing methods identify rumors based solely on the linguistic information without considering the temporal dynamics and propagation patterns. In this work, we aim to solve rumor detection task under the framework of representation learning. We first propose a novel way to construct the propagation graph by following the propagation structure (who replies to whom) of posts on Twitter. Then we propose a gated graph neural network based algorithm called PGNN, which can generate powerful representations for each node in the propagation graph. The proposed PGNN algorithm repeatedly updates node representations by exchanging information between the neighbor nodes via relation paths within a limited time steps. On this basis, we propose two models, namely GLO-PGNN (rumor detection model based on the global embedding with propagation graph neural network) and ENS-PGNN (rumor detection model based on the ensemble learning with propagation graph neural network). They respectively adopt different classification strategies for rumor detection task, and further improve the performance by including attention mechanism to dynamically adjust the weight of each node in the propagation graph. Experiments on a real-world Twitter dataset demonstrate that our proposed models achieve much better performance than state-of-the-art methods both on the rumor detection task and early detection task.Entities:
Keywords: Graph neural network; Representation learning; Rumor detection; Social network; Social security
Year: 2020 PMID: 32565619 PMCID: PMC7274137 DOI: 10.1016/j.eswa.2020.113595
Source DB: PubMed Journal: Expert Syst Appl ISSN: 0957-4174 Impact factor: 6.954
Fig. 1The way of constructing the propagation graph.
Fig. 2Node Update Process.
Begin |
For |
|
End for |
For |
For |
If |
For |
Use the representation of node |
End for |
Else |
For |
For |
Use attention mechanism to adjust the weights of different incoming edges of node v with the same type of relation path, the attention score is calculated according to Eq. |
End for |
Aggregate the information from the neighbors of node |
End for |
End if |
End for |
End for |
End |
The statistics of PHEME dataset.
| PHEME | |
|---|---|
| # Total Post Sets | 6425 |
| # Total Tweets | 105,354 |
| # Avg/Post Set | 16 |
| # Non-rumors | 4023 |
| # False-rumors | 638 |
| # True-rumors | 1067 |
| # Unverified-rumors | 697 |
Fig. 3The F1-score vs the update time when
Fig. 4The F1-score vs the number of layers when
Fig. 5The selection of the optimal model
Best performance comparison for rumor classification.
| Method | Micro-F1 | Macro-F1 | NR | FR | TR | UR |
|---|---|---|---|---|---|---|
| F1 | F1 | F1 | F1 | |||
| SVM-BOW | 0.543 | 0.535 | 0.607 | 0.479 | 0.537 | 0.518 |
| RFC | 0.550 | 0.539 | 0.586 | 0.468 | 0.594 | 0.508 |
| SVM-TS | 0.452 | 0.437 | 0.484 | 0.336 | 0.471 | 0.449 |
| GRU-RNN | 0.695 | 0.680 | 0.846 | 0.526 | 0.724 | 0.624 |
| BU-RvNN | 0.702 | 0.700 | 0.865 | 0.544 | 0.706 | 0.659 |
| TD-RvNN | 0.727 | 0.702 | 0.879 | 0.570 | 0.705 | 0.656 |
| GLO-PGNN(basic) | 0.743 | 0.739 | 0.884 | 0.650 | 0.747 | 0.674 |
| ENS-PGNN(basic) | 0.731 | 0.720 | 0.861 | 0.614 | 0.732 | 0.673 |
| GLO-PGNN | 0.879 | |||||
| ENS-PGNN | 0.748 | 0.738 | 0.638 | 0.745 | 0.684 |
Fig. 6The prediction time of the compared models
Fig. 7The statistics of post sets in the testing set
Fig. 8Early stop performance at different checkpoints in terms of elapsed time and tweets count
Fig. 9The partial propagation structure of the source tweet whose ID is 544,307,028,815,253,504
Fig. 10The similarity measured by cosine distance