| Literature DB >> 35469150 |
Li Tan1, Ge Wang1, Feiyang Jia1, Xiaofeng Lian2.
Abstract
To manage the rumors in social media to reduce the harm of rumors in society. Many studies used methods of deep learning to detect rumors in open networks. To comprehensively sort out the research status of rumor detection from multiple perspectives, this paper analyzes the highly focused work from three perspectives: Feature Selection, Model Structure, and Research Methods. From the perspective of feature selection, we divide methods into content feature, social feature, and propagation structure feature of the rumors. Then, this work divides deep learning models of rumor detection into CNN, RNN, GNN, Transformer based on the model structure, which is convenient for comparison. Besides, this work summarizes 30 works into 7 rumor detection methods such as propagation trees, adversarial learning, cross-domain methods, multi-task learning, unsupervised and semi-supervised methods, based knowledge graph, and other methods for the first time. And compare the advantages of different methods to detect rumors. In addition, this review enumerate datasets available and discusses the potential issues and future work to help researchers advance the development of field.Entities:
Keywords: Deep learning; Research status; Rumor detection; Social media
Year: 2022 PMID: 35469150 PMCID: PMC9022167 DOI: 10.1007/s11042-022-12800-8
Source DB: PubMed Journal: Multimed Tools Appl ISSN: 1380-7501 Impact factor: 2.577
Fig. 1Development history of rumor detection
Questions considered in our systematic review
| Question | Description |
|---|---|
| RQ1: What features of data mining rumors did these studies use and how to use? | This question lets researchers more quickly understand how to mine the features of rumors in massive data. |
| RQ2: Which deep learning structures are used for rumor detection? | Researchers can understand more clearly which deep learning structures are more suitable for rumor detection problems and how to use them by this question. |
| RQ3: What deep learning methods are used in these studies? | This problem allows researchers to master the deep learning method of popular rumors detection faster. |
| RQ4: Which datasets are used mostly for conducting rumors analysis? | Identify available datasets helps researchers to use them as benchmarks as well as to compare with their work. |
| RQ5: What are the main challenges within the rumor detection field? | The answer to this question helps new researchers recognize the open research challenges in this field. |
Fig. 2Distribution of publications
Fig. 3Example of Twitter
Fig. 4Example of non-rumor spreading (left) and a rumor spreading example (right)
Detail of rumor detection model based on CNN
| Paper | DL Model | Features in experiment | Dataset | Description | Characteristic | |||
|---|---|---|---|---|---|---|---|---|
| T | S | V | P S | |||||
| [ | CNN |
|
|
|
| CNN extract text and visual features | Combining displayed features and hidden features to jointly judge fake news | |
| [ | CNN |
|
|
|
| Combination of residual unit and attention module | Attention residual network can capture long-distance information and mine deep correlation | |
| [ | CNN |
|
|
|
| PolitiFact GossipCop | Extracting text and visual feature using Text-CNN | The semantic consistency of text and image is detected. Image embedding using image2sensence. |
| [ | CNN |
|
|
|
| Twitter15 Twitter16 | Extracting text features and propagation structure features from 1D CNN | The propagation structure is embedded into a vector. Connecting text features and communication structure features |
| [ | CNN, RNN |
|
|
|
| MutiSource-Fake | CNN and Bi-GRU are combined to extract text and emotional feature respectively | CNN and Bi-GRU are combined to extract text |
| [ | CNN, RNN |
|
|
|
| Twitter Weibo | TextCNN extracts text features and VGG extracts visual feature | Multi-modal feature fusion, using adversarial learning to obtain cross-domain features |
DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure
Detail of rumor detection model based on RNN
| Paper | DL Model | Features in experiment | Dataset | Description | Characteristic | |||
|---|---|---|---|---|---|---|---|---|
| T | S | V | P S | |||||
| [ | RNN |
|
|
|
| Twitter15 Twitter16 | Bi-LSTM extracts text features | Encoder-decoder structure,end-to-end framework |
| [ | RNN |
|
|
|
| Twitter Weibo | LSTM combined with Attention layer | Fusion of visual features, social features, and text features |
| [ | RNN |
|
|
|
| PHEME | Bi-LSTM extracts text features | Divided into blog post level and event level modules, event level modules increase attenuation factor |
| [ | RNN |
|
|
|
| Twitter15 Twitter16 | GRU as a hidden unit | Extract features along the propagation tree and preserve the propagation structure |
| [ | RNN |
|
|
|
| PHEME | LSTM extracts text content and comment content features | LSTM-Tree structure, keeping the relationship between source posts and replies |
| [ | RNN |
|
|
|
| LSTM extracts propagation structure features | Keep sequence and get farer dependencies | |
| [ | RNN |
|
|
|
| RumorEval PHEME | LSTM to extract user comments and fuse two task features | user characteristics to position detection tasks |
| [ | RNN |
|
|
|
| RumorEval | GRU extracts features after fusion of text and visual information | Better integration of two different types of meta-tasks |
| [ | RNN |
|
|
|
| FakeNewsNet | Bi-GRU extracts the contextual semantics of text content and user comments | Explore interpretable information from user reviews |
| [ | RNN |
|
|
|
| Twitter PHEME | GRU storage hidden repress-entation | Generative adversarial learning, divided into GRU encoder and GRU decoder |
| [ | RNN |
|
|
|
| RumorEval Weibo16 Weibo20 | Bi-GRU extracts original content features and emotional features in comments | Combine publisher sentiment and social comment sentiment |
| [ | RNN |
|
|
|
| Twitter15 Twitter16 | GRU extracts the propagation process and original text features | GCN is used as graph perception, CNN and GRU jointly extract propagation features |
DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure
Detail of rumor detection model based on GNN
| Paper | DL Model | Features in experiment | Dataset | Description | Characteristic | |||
|---|---|---|---|---|---|---|---|---|
| T | S | V | P S | |||||
| [ | GNN |
|
|
|
| unnamed | Four layers of GCN and two convolutional layers | Combine user profiles, activities, networks and communications, and content |
| [ | GNN |
|
|
|
| Twitter15 Twitter16 | GCN extracts propagation structure features | Bi-GCN extract propagateion and diffusion features from top-down and bottom-up respectively |
| [ | GNN CNN RNN |
|
|
|
| Twitter15 Twitter16 | GCN extracts propagation feature | Graph structure adversarial learning |
| [ | GNN |
|
|
|
| [ | Attention Graph Neural Network to extract graph features compose of text vectors | Semi-supervised learning method |
| [ | GNN |
|
|
|
| Twitter15 Twitter16 | GCN extracts propagation structure features in the encoder | Graph Auto-encoder Extract both dissemination structure features and text structure features at the same time |
| [ | GNN CNN |
|
|
|
| Twitter15 Twitter16 Weibo | GCN extracts graph structure features | Graph attention network to obtain structural semantics |
| [ | GNN |
|
|
|
| Twitter15 Twitter16 | GCN extracts graph structure features | Use the characteristics of transmission structure to judge the accuracy of the source of rumors |
| [ | GNN RNN |
|
|
|
| PHEME RumorEval | GCN extracts nonlinear diffusion feature | Combination of dual propagation feature ,GCN extracts nonlinear features, LSTM extracts time series linear features |
DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure
Detail of rumor detection model based on Transformer
| Paper | DL Model | Features in experiment | Dataset | Description | Characteristic | |||
|---|---|---|---|---|---|---|---|---|
| T | S | V | P S | |||||
| [ | Transformer |
|
|
|
| Twitter15 Twitter16 | Transformer captures the reliance between tweets and retweets | Breaking the propagation tree structure, the retweet information in different conversation threads can be calculated for correlation |
| [ | Transformer |
|
|
|
| SemEval2017 PHEME | Transformer extracts position feature | Using tags for position detection makes the results of rumor detection more accurate |
| [ | Transformer |
|
|
|
| PolitiFact GossipCop PHEME | Transformer’s coding layer term news coding | External knowledge using knowledge graph |
| [ | Transformer |
|
|
|
| CLEF-EN CLEF-Ar MediaEval LESA | Transformer extracts text features | Explore the importance of images in rumor detection |
| [ | Transformer |
|
|
|
| Twitter PHEME | Transformer extract position and attitude feature and communication feature | Tree Transformer to preserve the user interaction of the dialog |
DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure
Fig. 5Classification and related work of the latest methods of rumor detection
Detail of deep learning research studies in Section 5
| Study | DL Method | Performance metrics | Tool/ Library | Desctiption of Architecture |
|---|---|---|---|---|
| [ | RvNN | Acc.:0.723 NR:F1:0.682 FR:F1:0.753 TR:F1:0.821 UR:F1:0.654 | theano | They input each propagation node into the RNN and propagate the features along the path of the propagation tree to the next RNN Unit. |
| [ | RNN GNN | Acc.:0.925 Pre.:0.9245 Recall:0.925 F1:0.9247 | - | Graph convolutional neural(GCN) network to extract non-Linear structure features, LSTM extract linear sequence feature. |
| [ | RNN Transformer | Acc.:0.85 NR:Acc:0.844 FR:Acc.:0.857 TR:Acc.:0.883 UR:Acc.:0.814 | pytorch tqdm sklearn | Use RNN to learn context and use Transformer to find tightly related conversation threads |
| [ | RNN | Acc.:0.781 | theano | Gated Recurrent Unit used for representing hidden units |
| [ | CNN RNN | Acc.:0.827 Pre.:0.847 Recall:0.812 F1:0.829 | sklearn pytorch scipy jieba Word2Vec | TextCNN extracts text features and VGG extracts visual feature |
| [ | RNN CNN | Acc.:0.824 | keras sklearn Word2Vec nltk | Bi-LSTM as both encoder and decoder to extracts text features and VGG-19 with double fully connect layers extracts images features. |
| [ | RNN | Acc.:0.752 Pre.:0.755 Recall:0.752 F1:0.752 | - | Bidirectional LSTM Unit as the hidden units of encoder-decoder.And learn multi-task features by Bi-LSTM. |
| [ | Dense | Acc.:0.877 Pre.:0.840 Recall:0.832 F1:0.836 | tensorflow keras sklearn nltk | Dense Net as the domain-specific discriminator,the domain-shared discriminator and generator. |
| [ | RNN CNN | Acc:0.763 | - | Use GRU as the hidden unit of decoder-encoder; Residual Blocks as content encoder and decoder to Extract visual features. |
| [ | RNN CNN | Acc:0.866 | - | VGG-19 extract image feature and Bi-GRU extract text feature. The rumor characteristics between different posts are transmitted through GRU,and Separating domain specific features with event memory network. |
| [ | RNN | F1:0.405 Acc.:0.405 | keras | LSTM is used to learn text features, features of rumor detection task, stance detection features and rumor verification features. |
| [ | RNN | Acc.:0.638 F1:0.606 | - | LSTM unit is used to Rumor Verification and Stance detection. |
| [ | RNN | F1:0.464 NR:F1:0.876 FR:F1:0.543 TR:F1:0.114 UR:F1:0.333 | theano | Use GRU to merge the characteristics of the two tasks. |
| [ | RNN BERT CNN | Acc.:0.819 Pre.:0.75 Recall:0.8667 F1:0.8041 | pretrained BistilBERT | word embedding and image embedding by pretrained BistilBERT and VGG19. Use GRU as the hidden layer unit in Meta Multi-task layer. |
| [ | Transformer | Pre.:0.8687 Recall:0.8499 F1:0.8539 Acc.:0.8586 AUC:0.9197 | Wikipedia TagMe Word2Vec | Transformer as encoder to encode entities, news, and entity contexts sequence. |
| [ | GNN RNN | F1:0.8912 Prec:0.8982 Recall:0.8917 | pytorch Word2Vec functools TagME | GCN learns the structural characteristics of heterogeneous graphs with attention mechanism and use LSTM to embed information describing entities. |
| [ | CNN XLNet | Acc.:0.846 | keras pandas sklearn | Use Pretrained XLNeT to embed the title and content to get the text features.Use VGG-19 to embed the image to get visual feature. |
| [ | CNN | Acc.:0.824 AUC:0.873 | pytorch | TextCNN to get text embedding. |
| [ | RNN | F1:0.807 True:Pre.:0.637 Recall:0.665 F1:0.651 False:Pre.:0.874 Recall:0.860 F1:0.867 | theano | GRU with attention mechanism for claim verification. |
DL:Deep Learning; NR: nonrumor; FR: false rumor; TR: true rumor; UR: unverified rumor; Pre.:Precision; Acc.:Accuracy
Fig. 6Architecture of Propagation Trees Method.Top-down RvNN is used as an example to extract the linear sequence features of Twitter propagation. Some studies have taken the nonlinear structural features of the propagation tree and fused with the linear features to classify rumors
Fig. 7Architecture of Adversarial Learning(below) and Automatic Coding Structure(above) Method
Fig. 8To improve the performance of cross-domain rumor detection, the primary research is to remove the domain-specific features in the rumor feature. After the features are extracted, the domain discriminator is used to improve the performance of the feature extractor to remove the features of the specific domain. Regardless of how the latest research assists in domain elimination tasks, the primary method used remains in adversarial learning
Fig. 9Architecture of Multi-task learning for Rumor Detection. Including two tasks: Stance Detection and Rumor Verification. These two tasks are also the choice of most multi-task rumor detection. Helping rumor detection tasks by detecting the stance of comments
Fig. 10Architecture of Knowledge Graph(KG) Methods for Rumor Detection. The steps of current general method for rumor detection based on knowledge graph : first identify the entity of the content of the rumor, then entity link to obtain external knowledge from knowledge graph , then represent the knowledge to vector, last classify the rumors after fusion or comparison with the original features
Detail of dataset
| Dataset | Reference | Size | Label | Number for labels | Data type | Detail of dataset |
|---|---|---|---|---|---|---|
| PHEME | [ | 5802 tweets | rumor /none-rumor | 1792/3830 | text | It contains five events: Charlie Hebdo, Ferguson, Germanwings Crash, Ottawa Shooting, Sydney Siege. It’s imbalanced dataset. |
| Twitter15 | [ | 1380 tweet-tree | non-rumor /false-rumor /true rumor /unverified | 374/370/ 372/374 | Tree-structure text | It contains 276663 users, 1490tweets, 331612 threads. It’s balance dataset. |
| Twitter16 | [ | 1181 tweet-tree | non-rumor /false-rumor /true rumor /unverified | 205/205/ 205/203 | Tree-structure text | It contains 173486 users, 818 tweets, 204820 threads. It’s balance dataset. |
| MutiSource-Fake | [ | 11397 news | fake/real | 5403/5994 | text | It contains news from OpenSources.co, MediaBiasFactCheck.com, PolitiFact news websites’ lists. It’s balance dataset. |
| Weiibo | [ | 146 events, 50287 tweets | rumor /none-rumor | 23456/26257 | text, image, users | Real world multimedia dataset from Sina Weibo contains 50287 tweets and 25953 images and 42310 users. It’s balance dataset. |
| MediaEval | [ | 15821 tweets | rumor /none-rumor | 9596/6225 | text, image | It contains tweets related to the 11 events, comprising in total 193 cases of real and 220 cases of misused images and videos, associated with 6225 real and 9596 fake tweets posted by 5,895 and 9,216 unique users respectively. It’s imbalanced dataset. |
| FakeNewNet | [ | 23196 news | fake/real | 5755/17441 | text, image, users, network, response | This dataset contains 23196 news, 19200 images from PolitiFact and GossiCop. It’s imbalanced dataset. |
| Buzzfeed Election | [ | 71 news | fake/real | 35/36 | text | It collected the news stories found in Buzzfeed’s 2016 article on fake election news on Facebook (Silverman 2016). It’s balance dataset |
| LIAR | [ | 12836 shot texts | pants-fire/ false /barely-true /half-true /mostly-true /true | - | text | It collected a decade-long, 12.8K manually labeled short statements in various contexts from POLITIFACT.COM, which provides detailed analysis report and links to source documents for each case. |
| [ | 13924 news | fake/real | 7898/6026 | text, image | This dataset contains 13924 news and 514 imgaes from MediaEval. It’s imbalanced dataset . | |
| WeChat Datasets | [ | 27161 articles | fake/real | 2090/2090 | text | This dataset collected from WeChat’s Official Accounts, dated from March 2018 to October, 2018.It contains 27161with 37971 reports,and 22981 articles is unlabeled. It’s balanced dataset. |
| [ | 4 million Tweets | fake/real | - | text | This dataset is unpublished and contains 4 million tweets, 3 million users, 28893 hashtags, and 305115 linked articles, revolving around 1022 rumours from 01/05/2017 to 01/11/2017. | |
| [ | 126000 rumor cascades | rumor/non-rumor | - | text | 126000 rumor cascades spread by 3 million people more than 4.5 million times from Twitter inception in 2006 to 2017. | |
| Weibo-20 | [ | 6362 articles | fake/real | 3161/3201 | text | This dataset is constructed on thr basis of Weibo-16([ |
| FEVER | [ | 185,446 statements | support /refuted /not enough info | 93367/43107/ 48971 | text | It contains judgment and evidence of the statement.And the statements are from Wikipedia. It’s imbalanced dataset. |
| Fakeddit | [ | 825100 news | fake/real; completely true /fake news with true text /fake news with false text ; true /Satire(Parody) /Misleading Content /Imposter Content /False Connection | - | text, image, user | Dataset for 2-way , 3-way, and 5-way classification. It sourced from Reddit. It contains 628,501 Fake samples and 527,049 True samples. It also contains 682,996 Multimodal samples and 358,504 users. It’s imbalanced dataset. |
| NewsBag | [ | 215000 news | fake/real | 200000 /15000 | text, image | It contains 200,000 real news , 15,000 fake news and 15000 images. It’s imbalanced dataset. |
| Fake News Evolution | [ | 950 paired data | real/fake /Turned into fake news | - | text | This dataset includes 950 pieces of data, each of them contains three articles representing the three phases of the evolution process, and they are the truth, the fake news and the evolved fake news. |
The list of abbreviations
| Abbreviation | Full Name |
|---|---|
| RQ | Research Questions |
| DL | Deep Learning |
| T | Text |
| S | Social |
| V | Visual |
| PS | Propagation Structure |
| Acc. | Accuracy |
| Pre. | Precision |
| UR | Unverified rumor |
| TR | True rumor |
| FR | False rumor |
| NR | Nonrumor |
| CNN | Convolutional Neural Networks |
| TextCNN | Text Convolutional Neural Networks |
| VGG | Visual Geometry Group |
| RNN | Recurrent Neural Networks |
| GRU | Gated Recurrent Unit |
| Bi-GRU | Bi-directional Gated Recurrent Unit |
| LSTM | Long Short-Term Memory |
| Bi-LSTM | Bi-directional Long Short-Term Memory |
| GNN | Graph Neural Networks |
| GCN | Graph Convolutional Networks |
| Bi-GCN | Bi-directional Graph Convolutional Networks |
| BERT | Bi-directional Encoder Representation from Transformers |
| TF-IDF | Term Frequency–Inverse Document Frequency |
| LDA | Latent Dirichlet Allocation |