Literature DB >> 35469150

Research status of deep learning methods for rumor detection.

Li Tan1, Ge Wang1, Feiyang Jia1, Xiaofeng Lian2.   

Abstract

To manage the rumors in social media to reduce the harm of rumors in society. Many studies used methods of deep learning to detect rumors in open networks. To comprehensively sort out the research status of rumor detection from multiple perspectives, this paper analyzes the highly focused work from three perspectives: Feature Selection, Model Structure, and Research Methods. From the perspective of feature selection, we divide methods into content feature, social feature, and propagation structure feature of the rumors. Then, this work divides deep learning models of rumor detection into CNN, RNN, GNN, Transformer based on the model structure, which is convenient for comparison. Besides, this work summarizes 30 works into 7 rumor detection methods such as propagation trees, adversarial learning, cross-domain methods, multi-task learning, unsupervised and semi-supervised methods, based knowledge graph, and other methods for the first time. And compare the advantages of different methods to detect rumors. In addition, this review enumerate datasets available and discusses the potential issues and future work to help researchers advance the development of field.
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022.

Entities:  

Keywords:  Deep learning; Research status; Rumor detection; Social media

Year:  2022        PMID: 35469150      PMCID: PMC9022167          DOI: 10.1007/s11042-022-12800-8

Source DB:  PubMed          Journal:  Multimed Tools Appl        ISSN: 1380-7501            Impact factor:   2.577


Introduction

The Internet and social media have become comprehensive and large-scale platforms for disseminating real-time information. It is worth emphasizing that rumors can change the perceptions of billions of people during the rumor spread in social media. According to the 2018 Internet trend report([69]) , more than a third of social media news events contain false information. The characteristic of rumors is spread widely and quickly. The latest research shows that the spread of rumors is 6-20 faster than non-rumors [54]. And the rumors get more people’s attention by their emotional which cause spread widely [124]. In addition, the spread of rumors also showed a lot of incredible harm. For example: in 2013, the Twitter account of the Associated Press was hacked. Then it post a claim that two explosions had occurred in the White House, and president to be injured. Although this rumor was debunked soon, it still spread to millions of users, causing serious social panic and leading to a rapid stock market crash [20, 21]. Even some rumors about the COVID-19 in 2020 pose a threat to life safety and increase the pressure on medical personnel, such as false statements suggesting drinking bleach to cure diseases [108]. Therefore, since 2011, researchers have made many efforts to establish a prioritized and automatic rumor detection method so that alleviate the pressure of rumors. Figure 1 are representative works with the development of rumor detection. Among them, [10] conducted rumor detection related work on Twitter for the first time in 2011. They manually extracting message-based features and topic-based features to be credible for platform-specific topic-related information degree to be evaluated. In 2012, [107] is the first research on rumor detection in Chinese social media. In the same year, [34] used communication methods to solve the problem of rumors by constructing a network composed of users, messages, and events. In 2013, [51] successfully extracted the temporal features of rumors for the first time. In 2015, [104] modeled the propagation mode of rumors in a propagation tree for the first time. In 2016, [81] proposed the first rumor detection platform named Hoaxy, and for the first time mentioned the difference from propagation of rumors. However, there are hand-craft approaches above. In the same year, [64] learn the characteristics of the content of rumors by deep learning for the first time, opening a new page of rumor detection. Then in 2017, [114] applied CNN to rumor detection for the first time; [42] first incorporated visual features to classify rumors. Their work leads to more multimodal fusion rumor detection work. In 2018, [101] tried event-level cross-domain learning for the first time to solve the imbalance of rumors in different domains. [66] first proposed a multi-task rumor detection method based on stance detection task. In 2019, [68]. tried first to use the features of generative confrontation networks to simulate rumors. In 2020, the reinforcement learning method was first introduced in the rumor detection task by [102]. In 2021, [122] proposed dual emotional features, which is ahead of the existing emotional feature methods for rumor detection. The above methods are only part of the representative work of rumor detection. Since deep learning has made incredible achievements in rumor detection, many researchers have extracted visual feature ([13, 42]), propagation feature ([65, 67]) and user social background feature ([16]) of rumors from rich data. Based on the fusion of multiple features, more novel and excellent methods are proposed, such as multi-task methods [56], adversarial learning [101], semi-supervised methods [30], weakly supervised methods [37, 109] and so on.
Fig. 1

Development history of rumor detection

Development history of rumor detection Due to the numerous researches related to rumor detection, it is difficult for new researchers to enter this field and grasp its current research status. Therefore, some surveys and reviews have conducted in-depth research on detecting rumors. For example, [126] proposes an overview of how to develop a rumor detection system, which consists of four steps: rumor detection, rumor tracking, rumor stance classification, and rumor veracity classification. As their work pays less attention to feature extraction and neural network-based algorithms, it is impossible to summarize and compare different methods from an algorithmic point of view. The survey by [8] described many methods based on manual features. However, due to the early publication, there is a lack of the latest deep learning methods. The survey by [70] introduced the phased issues of the origin, dissemination, and detection of false information but also focused on machine learning methods, ignoring most of the deep learning methods. Overall, most of the published surveys focus on machine learning methods and only involve a small number of deep learning methods, which inspired our work. We aim to provide a comprehensive introduction to the rumor detection algorithm based on deep learning compared with the above survey. Based on deep learning, we focus on the three aspects of feature engineering, model structure, and algorithm perspective in rumor detection. In addition, we also analyzed the shortcomings of each algorithm. This analysis will help the future development of the algorithm. The main contributions of the work are as under: We discussed how to select and mine the feature of rumor from the massive data in the Internet, and how to use such feature in recent rumor detection research. As far as we know ,it is the first time for listing the rumor detection models based on deep learning and classify them into methods based on CNN, RNN, GNN, Transformer and discuss how to use each basic model and improved models in popular research to analyze the advantages and disadvantages. As far as we know, this article is the first to categorize popular research based on rumor detection methods, sort out and introduce the latest research ideas and methods. Summarize and describe the publicly available rumor detection datasets with size, label, number for each label, data type and remark. We will summarize the latest research and the most representative methods in rumor detection from three dimensions: characteristics, structure, and methods. The chapter structure of this article is as follows: Section 2 introduced the review methodology of this article. Section 3 introduces how to use data features to solve the problem of rumor detection in the latest research on rumor detection; Section 4 introduces the latest research models by model structure; Section 5 categorizes according to research methods Introduce each research content and method; in Section 6, we present the datasets available in rumor detection. Finally, in Section 7, we proposes potential challenges and issues to existing research. Section 8 summarizes the work.

Research methodology

Development of the review protocol

To systematically investigate and sort out the research status of rumor detection, we conducted a systematic review of rumor detection based on deep learning. The first step of the review is to search for relevant research in multiple digital libraries and databases. Then the selection criteria are used to reduce the number of selected studies to further improve the quality of the papers involved and involve as many different deep learning methods as possible. After that, a set of research questions was formulated to thoroughly solve the research on the current situation of rumor detection.

Source of information

To conduct our systematic literature review, the following digital libraries and datasets have been selected : IEEE Explore (www.ieeexplore.ieee.org) Springer Link (www.springerlink.com) Sciencez Direct (www.sciencedirect.com) ACM Digital Library (https://dl.acm.org) Google Scholar (www.scholar.google.com) The search keywords used to find related researchs are “Rumor detection” OR “Fake information” AND “Deep Learning” OR “Deep Neural Networks” OR “[the name of Deep Learning Methods]”. The searched literature results involve rumor detection and deep learning. To further improve the quality of the papers involved, we further screened the papers through multiple factors of the selection criteria.

Research questions

To sort out the comprehensive research status of rumor detection, we raised five research questions in Table 1. We sorted out and analyzed these five research questions in our recent research work and answered these five questions in turn through the subsequent chapters of this article.
Table 1

Questions considered in our systematic review

QuestionDescription
RQ1: What features of data mining rumors did these studies use and how to use?This question lets researchers more quickly understand how to mine the features of rumors in massive data.
RQ2: Which deep learning structures are used for rumor detection?Researchers can understand more clearly which deep learning structures are more suitable for rumor detection problems and how to use them by this question.
RQ3: What deep learning methods are used in these studies?This problem allows researchers to master the deep learning method of popular rumors detection faster.
RQ4: Which datasets are used mostly for conducting rumors analysis?Identify available datasets helps researchers to use them as benchmarks as well as to compare with their work.
RQ5: What are the main challenges within the rumor detection field?The answer to this question helps new researchers recognize the open research challenges in this field.
Questions considered in our systematic review

Selection Criteria

The selection criteria are as follows: This article’s rumor detection method for deep learning mainly covers the research work during 2017-2021. After deep learning has made fantastic progress in many fields, the iterative update speed of deep learning has dramatically increased. To enable researchers to grasp the current research status in this field more quickly, we focused on deep learning method research into nearly five years of work. To further improve the quality of the articles involved in this article, this article only screens high-quality conferences or journals (such as conferences and journals recommended by China Computer Federation ), or studies with high citation rates, or research groups that have made outstanding contributions to the field of rumor detection. This article only covers research papers written in English. In terms of screening, we eliminated researches irrelevant to the subject of the indexed literature and finally screened out 98 works with great attention by selection criteria and Fig. 2 is the distribution of the publication we selected.
Fig. 2

Distribution of publications

Distribution of publications

Feature extraction

Rumor detection will be regarded as a binary classification problem in most cases, and a few studies will treat it as a multi-classification problem. Most of the literature follows the general learning rules of supervised classification in machine learning: first, extract the characteristics of the representative samples from the two types of samples; second, train the extracted samples to a suitable model; finally, test or test the data sets with unknown labels. It is to evaluate its label. In the latest literature, most of the research focuses on how to extract features and where to extract features. How to deal with the characteristics after obtaining the characteristics is also a critical issue of the research. To answer RQ1, in this chapter, we review the content features, social features and propagation structure features used in rumor detection, and introduce the methods of feature extraction and how to use features. In social media, general news or blog content contains a wealth of multimodal resources, such as pictures, videos, texts, etc., as well as some links, comments, likes, and other information associated with the communication process. As shown in Fig. 3. In addition, in the latest literature, it is found that there is a big difference in propagation structure between rumors and real information. Therefore, there is a lot of work to extract features from the propagation structure.
Fig. 3

Example of Twitter

Example of Twitter

Content feature

Text features

Compared with real information, rumors are used to confuse and mislead the public. The text makes people more emotional and can attract more people’s attention. Traditional text features are mainly divided into three aspects: lexical features, syntactic features and topic features. Lexical features refer to language features extracted from a single word or a single word level. In the early detection of rumors with manual features, [10] calculated the “number of words”, “number of words”, “number of different words”, and other indicators based on the words contained in a message. Syntactic feature refers to a type of feature that is sobbing from the sentence level, such as part-of-speech tagging, word frequency, etc. The topic feature refers to the text feature extracted from the entire message collection, and the purpose is to understand the message and its implicit semantics. Unlike manually extracting features from the text in the past, most latest studies([27, 28, 64, 68, 73, 97, 100, 111, 114]) no longer only use text feature, but the fusion of multiple text features solves the problem of rumor detection. For example: [64] introduced a recurrent neural network to learn hidden representations from the text content of related articles, opening a precedent for using deep learning to automatically extract the features of rumor texts. Later, [114] used convolutional neural networks to obtain key features and hidden semantic features from text content. However, these studies take the full text as input only and lose the article’s original chapter structure. So, [111] adopted the hierarchical structure of “word-sentence-article” to understand a text and extracted the word level and sentence hierarchy as the feature of understanding the text. [73] learned the word-level and sentence-level representation of statement texts and news articles as text features. However, the above research did not consider the connection between the article and terminology, [37, 101] believe the terminology of different domains will affect the effect of rumor detection, so they used terms to distinguish fake news and capture the potential relationship between articles and terms. Studies have shown that most fake news and rumors have much emotional information, and the topic of rumors generally contains words that attract attention, which is used to drive readers’ emotions and attract readers to read to promote spread. Therefore, the latest research uses the emotional information in the article as one of the characteristics of judging rumors. [28] manually extracted words with emotional characteristics such as emotion, morality, and exaggeration words in the news, combines them with theme characteristics to obtain fake news and real information in emotions and topics. [122] used the emotional information of news content in the experiment and extracted the emotional information from news reviews and used dual emotional features to detect fake news. [27] used LDA (Latent Dirichlet Allocation) to extract topic features combined with text features extracted by XLNet[112] as critical features.

Visual features

Even if the method using text features in rumor detection has achieved good results, in general, rumors have multi-modal information, such as images, videos, etc. Many studies combine visual feature to the features of rumor detection. During the hand-craft feature period, visual features can be divided into visual statistical features and visual content features. Among them, visual statistical feature is the statistics of rumor image, such as the number of images or the propagation time of the false image, and so on. The visual content feature refers to the content in the visual image, such as the clarity and diversity of the visual image. The early research of [33] was the first attempt to manually extract information such as the propagation time of fake images on Twitter and proposed a classification model to identify fake images on Twitter. In the latest literature, many studies prove the importance of visual information ([9, 11]). [11]pointed out that images are more influential than texts and often appear in rumors. In recent years, most of the research directly extracts high-dimensional representations of images from visual information (such as images and videos) using deep learning pre-training models for calculation. Such as research ([46, 101, 111, 118]) extract fake news, rumor and tweets image information by pre-trained convolutional neural network(CNN), mining deep-level visual features, and extract visual information High-dimensional representation is combined with text and other modal features: [120] used residual networks to learn visual features in fake news, and for the first time proposed visual separation representation learning, which can remove features in specific areas of visual features, so that model learns cross-domain features to realize cross-domain event fake news detection. In addition, there are studies using image embedding methods to convert images into matrices. [125] embed visual information into a matrix by image2sentence ([45]) and then used TextCNN(Text Convolutional Neural Network,[17] ) to extract vision features. Compared with the pre-trained model using convolutional neural networks such as VGG(Visual Geometry Group, [87]), it can calculate the similarity between different modal data and increase the receptive field. There are also studies using deep learning methods to extract the content information contained in images. For example, [11] used the research results of [95] to extract image emotional information in pictures.

Social feature

In social media, a series of social interactions occur during rumors propagate. Under normal circumstances, interactions in social media are divided into three categories: follow, link, and forward comments. Social background information will be generated during the interaction process, and research has found that users with special characteristics in the user group participating in rumor interaction, such as low-credit users such as “marketing account”, will forward and comment on rumors that are sufficiently eye-catching. To increase their exposure to the public. This survey divides the social background information into user feature and propagation feature.

User feature

User feature are derived from the user’s social network. Rumors are created by a few users and spread by many users due to profit factors. The analysis of user feature can provide critical clues for rumor detection. User feature include individual features and group features. Among them, individual features are extracted from a single user, such as: “registration time”, “age”, “identity authentication”, etc. In the early research, the reliability of users was evaluated from the perspective of reporters, and information such as “user credibility” and “user location” were used as user characteristics. Features of the user group are the characteristics extracted from the user group, such as “verified user ratio” and so on. [31] modelled user participation by selecting social concerns of critical user comments and set 22 social characteristics, including “proportion of users with avatars”, “proportion of users verified by verification”, and “average number of fans”. Among them, there are eight user characteristics. [43] extracted user credit features to represent users’ social reactions. [56] combined the user’s credit information with the user’s comment information on the statement to solve the task of rumor detection. [61] extracted user personal information in the retweeted user sequence as user features.

Propagation feature

In the process of dissemination, rumors will be forwarded, commented, and liked by most people. Propagation features include features extracted from some information that appears in the process of communication, such as: “user comments”, “number of reposts”, “number of likes”, “number of clicks”, and so on. Some studies extract features from user comments to determine users’ attitudes toward rumors. For example, [43] proposed mining users’ conflicting views from social media posts and estimating their credibility values to detect fake news. There is also much work ([56, 91, 116]) extracting position information based on users’ replies to posts and statements. Most of these works use users’ replies to statements as propagation information for position detection tasks as the key to completing auxiliary rumor detection tasks. In some studies, user comments collected during the dissemination process are extracted and combined with text features to solve rumors. [116] build heterogeneous graphs based on posts on social networks, using comments and related users. [26] uses social contextual content (such as comment replies, forwarding replies) and social context metadata (such as the number of reposts, the number of likes, and the number of favorites) for modeling. In research of [19],they extract semantic information from users’ comments on rumors, find consistent comments, and capture the implicit relationship between comments and rumors to enhance the model’s semantic reasoning ability further. However, the above research only considers the connection between users and rumors and ignores the relationship between users. Therefore, [47] correlated the conversation threads of comments in pairs to extract the hidden dependencies between user interactions. The user-to-user mutual attention method simulates the potential influence relationship between users in the propagation path to stably capture the mode and connotation related to the spread of rumors and non-rumors.([72])

Propagation structure feature

The propagation structure feature records and reproduces the propagation process of rumors and uses the user’s reposting as a node to construct a reposting graph. Studies have pointed out that rumors and non-rumors are significantly different in the propagation structure ([40]). Figure 4 shows the comparison of the dissemination structure of non-rumors and rumors. The communication structure features include the relationship or similarity between articles, the relationship between users, the feature of the communication path, etc. The key communication features for judging rumors are extracted through the communication path and the communication graph. The earliest method of using propagation structure features to apply to rumor detection ([34]) formed a network of users, messages, and events, and unilaterally assumed that trusted users would not provide credibility for rumor events and that the links to trusted messages were better than rumor .The weight of information is more significant. However, some trusted users might be confused by disguised rumors, resulting in unsatisfactory results.
Fig. 4

Example of non-rumor spreading (left) and a rumor spreading example (right)

Example of non-rumor spreading (left) and a rumor spreading example (right) Some studies have proposed the use of graph kernels to capture communication features. For example, [104] proposed a hybrid SVM (Support Vector Machine) classifier, combined with RBF (Radial Basis Function), based on the random walk graph kernel to capture the feature of propagation. Some studies model the propagation path of rumours as a propagation tree. For example, [65] first proposed the concept of propagation tree core in the rumor detection on Twitter, starting with the propagation mode and using core learning. The tree models Twitter and compares the nuclear learning tree of non-rumors and rumors to analyze whether it is a rumor. In addition, some studies regard the spread of rumors as a sequence with time-series feature. In the study of [105], media users and social network structures are embedded, and then LSTM (Long Short Time Memory) is used. To represent and classify the propagation path of the message, however, the study did not consider the time sequence change of the diffusion structure path. In the latest research, [52] proposed to participate in changes along the path of the diffusion structure and the time of response to integrate the feature of the propagation structure. The above research still focuses on extracting the propagation structure’s feature through the propagation sequence, but the propagation sequence may lose key structural information ([94]). Therefore [94] uses Node2Vec[29] method, embeds the nodes in the propagation graph as a matrix as the communication information, and uses the convolutional layer to extract the propagation features. However, Node2Vec does not consider the weights in the figure, and in social media networks, accounts with different reputations have different propagation path weights. So more work began to join the Graph Neural Network (GNN) method to extract the propagation structure features from the adjacency matrix. In the latest research, many works have proved that graph neural networks can cope with the characteristics of the spreading structure of rumors well([4, 5, 22, 106]). For example: [4] used graph neural networks to capture the similarities between disseminated fake news. In the same year, [22] used graph convolutional networks(GCN) to locate multiple rumor sources without prior knowledge of the propagation model, and used the multi-level neighborhood information of nodes to establish node representations to improve the accuracy of the source prediction. [5] used bottom-up and top-down GCN to extract the spread and spread patterns of rumors respectively.

Model structure

The second chapter introduces the features and extraction methods currently used in rumor detection according to feature dimensions.To answer RQ2, we will introduce the rumor detection method in the dimension of the structure.

Rumor detection based on CNN

The convolutional neural network(CNN) is a feedforward neural network that includes convolutional calculations and has a deep structure. It is one of the representative algorithms of deep learning.It and its improved models ([14]) have been proven to perform well in computer vision ([1]), natural language processing ([78]) and other fields. Table 2 shows some representative examples of using CNN for rumor detection tasks in the latest research and details the latest method.
Table 2

Detail of rumor detection model based on CNN

PaperDL ModelFeatures in experimentDatasetDescriptionCharacteristic
TSVP S
[111]CNN TwitterCNN extract text and visual featuresCombining displayed features and hidden features to jointly judge fake news
[18]CNN TwitterCombination of residual unit and attention moduleAttention residual network can capture long-distance information and mine deep correlation
[125]CNN PolitiFact GossipCopExtracting text and visual feature using Text-CNNThe semantic consistency of text and image is detected. Image embedding using image2sensence.
[94]CNN Twitter15 Twitter16Extracting text features and propagation structure features from 1D CNNThe propagation structure is embedded into a vector. Connecting text features and communication structure features
[28]CNN, RNN MutiSource-FakeCNN and Bi-GRU are combined to extract text and emotional feature respectivelyCNN and Bi-GRU are combined to extract text
[101]CNN, RNN Twitter WeiboTextCNN extracts text features and VGG extracts visual featureMulti-modal feature fusion, using adversarial learning to obtain cross-domain features

DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure

Detail of rumor detection model based on CNN DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure In recent years, most studies ([18, 60, 111, 114]) use convolutional neural networks to extract hidden features of text content and visual information. For example: [114] first used convolutional neural networks for rumor detection tasks and used convolutional neural networks to extract key features from the text content of related posts for rumor detection. [18] proposed an attention residual network, which can obtain important information in local and global content features by learning text content information in rumors. Use this to understand the classification of rumors. In addition, some research ([101]) uses CNN to extract the visual features of fake news combined with text features. In the latest research, CNN has been used to extract the features of the embedding matrix in many researches. For example, [125] used fake news text and visual content to embed it and then used a convolutional neural network to extract it as a feature map to calculate cross-modal similarity to judge the inconsistency between the text and the image description. Similarly, [94] used convolutional layers in the proposed Rumor2vec framework for feature extraction after text embedding and node embedding. They combined the two features to make the framework learn a joint representation of text and propagation features to detect rumors. Many studies have proved the effectiveness of convolutional neural networks in extracting rumor features, which can obtain local key features well. However, in convolution, the timing information of the input vector is ignored and disrupted. However, whether it is the text data of the rumor or the spread sequence of the rumor, the timing information can play a vital role in rumor detection.

Rumor detection based on RNN

Recurrent Neural Network (RNN) with a sequence of data as input is recursive in the sequence’s evolution direction, and all nodes are connected in a chain. The most significant difference between it and the convolutional neural network is that it retains the last state during recursion and retains timing information. Therefore, most existing researches are keen on using recurrent neural networks for natural language processing ([59]). [64] first cited RNN for the rumor detection task, automatically learning Twitter content based on time series, using TF-IDF to model words, and then using RNN to learn the potential content of rumors. Since then, recurrent neural networks and their variants have been used more in rumor detection, and many of them are the latest research results. Most studies ([2, 28, 42, 46, 90]) use RNN to learn text features, image features, and multimodal fusion features. For example: [46] used multiple stacked two-way LSTMs to learn text features and used LSTMs to fuse the text and social background features. [82] used RNN to obtain word-level representations of news content, sentence-level representations of news content, and user comment representations to describe the modeling from news language features to potential feature spaces. [90] proposed a BiLSTM with multiple loss levels based on attenuation factors to deal with rumors detection on Twitter. These two directions can extract deep context information from a limited amount of text. In addition, there are also studies using the characteristics of RNN to retain timing to capture the timing characteristics of rumors in propagation. [67] modeled the propagation structure of rumors as a tree structure and used GRU(Gate Recurrent Unit) to calculate each branch of the tree sequence. [49] proposed an LSTM tree based on a convolutional unit to predict the authenticity of positions and rumors in social media conversations, using the source blog post as the root node and the response as the child node. In each node, LSTM is used. To learn the position of evidence or comment, to carry out position detection. In addition, [105] used LSTM to represent and classify the propagation path of messages, which was used to solve malicious spreaders disguising fake news as real news. In the study, the news propagation tree structure in research of [67] was converted into time. At the same time, LSTM can be used to use the dependency between items separated by a long distance in the sequence. Some studies ([56, 119]) use LSTM as a shared unit in multi-task learning. They use LSTM to share the hidden vectors calculated by multiple meta-tasks to assist in rumor detection. For example, [56] and [119] use LSTM to fuse position detection and rumor detection as two meta-tasks. Although RNN can extract timing information, the input of RNN is generally a sequence, which leads to the destruction of the propagated structural information, and the effect of extracting the propagated structural features is not good. Detail of rumor detection model based on RNN in Table 3.
Table 3

Detail of rumor detection model based on RNN

PaperDL ModelFeatures in experimentDatasetDescriptionCharacteristic
TSVP S
[46]RNN Twitter15 Twitter16Bi-LSTM extracts text featuresEncoder-decoder structure,end-to-end framework
[2]RNN Twitter WeiboLSTM combined with Attention layerFusion of visual features, social features, and text features
[90]RNN PHEMEBi-LSTM extracts text featuresDivided into blog post level and event level modules, event level modules increase attenuation factor
[67]RNN Twitter15 Twitter16GRU as a hidden unitExtract features along the propagation tree and preserve the propagation structure
[49]RNN PHEMELSTM extracts text content and comment content featuresLSTM-Tree structure, keeping the relationship between source posts and replies
[105]RNN TwitterLSTM extracts propagation structure featuresKeep sequence and get farer dependencies
[56]RNN RumorEval PHEMELSTM to extract user comments and fuse two task featuresuser characteristics to position detection tasks
[119]RNN RumorEvalGRU extracts features after fusion of text and visual informationBetter integration of two different types of meta-tasks
[82]RNN FakeNewsNetBi-GRU extracts the contextual semantics of text content and user commentsExplore interpretable information from user reviews
[68]RNN Twitter PHEMEGRU storage hidden repress-entationGenerative adversarial learning, divided into GRU encoder and GRU decoder
[122]RNN RumorEval Weibo16 Weibo20Bi-GRU extracts original content features and emotional features in commentsCombine publisher sentiment and social comment sentiment
[61]RNN Twitter15 Twitter16GRU extracts the propagation process and original text featuresGCN is used as graph perception, CNN and GRU jointly extract propagation features

DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure

Detail of rumor detection model based on RNN DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure

Rumor detection method based on GNN

In recent years, graph neural networks have received a lot of attention. The social network is a common type of graph data representing the social relationship between various individuals or organizations. At the same time, the graph neural network uses non-Euclidean graphs as input, so compared to CNN and RNN, it can Keep the structure of the spread of rumors. The latest research shows that the dissemination structure of rumors and true information is different ([40]). In the latest work, GNN is mainly used to extract the structural features of propagation or the structural features of user interaction. [74] used stacked graph convolutional neural network layers to extract the structural feature of dissemination from heterogeneous graphs composed of various data such as users, user comments, news dissemination, etc., to detect rumors automatically. [5] proposed using a two-way graph convolutional network in their latest work, using top-down and bottom-up propagation directions to simulate the spread and spread of rumors. In addition, [4] also used a graph neural network to propose a semi-supervised fake news detection method to solve training in a limited number of labeled articles. [58] proposed a rumor detection method based on a graph autoencoder, which uses an encoder to use an efficient graph convolutional network to treat the initial text and the propagation graph as input and update the representation vector through propagation to learn the text and dissemination of information. [61] proposed a social media rumor detection method based on graph structure adversarial learning to deal with rumors in social networks disguised in various ways to avoid rumor detectors. This study established a heterogeneous information network to simulate the rich information between users, posts, and comments using a graph confrontation learning framework. The encoder tries to dynamically add intentional disturbances to the graph structure to trick the detector, containing disguised heterogeneous graphs. Input to the graph convolutional neural network to obtain the perturbation structure representation of each blog post. The detector will learn more unique structural features to resist this perturbation, thereby enhancing the detector’s camouflage way to debunk rumors, and learn the characteristics of diversified patterns. [116] mainly focused on integrating complex semantic information by learning forwarding sequences and how to model the heterogeneous graph structure of all microblogs and participants globally for rumor detection. Research uses graph attention network to capture global semantic information in heterogeneous graphs constructed by posts, comments and related users on social networks. The research uses a graph convolutional neural network to generate the latent representation of each node, uses the attention mechanism to obtain the weight of each node and iterates, and finally captures the global representation. Table 4 is the detailed information of the latest GNN-based rumor detection work. From these studies, it can be found that the GNN-based multi-modal rumor detection data is not well used. This may also be one of the research directions of future work.
Table 4

Detail of rumor detection model based on GNN

PaperDL ModelFeatures in experimentDatasetDescriptionCharacteristic
TSVP S
[74]GNN unnamedFour layers of GCN and two convolutional layersCombine user profiles, activities, networks and communications, and content
[5]GNN Twitter15 Twitter16GCN extracts propagation structure featuresBi-GCN extract propagateion and diffusion features from top-down and bottom-up respectively
[61]GNN CNN RNN Twitter15 Twitter16GCN extracts propagation featureGraph structure adversarial learning
[4]GNN [105]Attention Graph Neural Network to extract graph features compose of text vectorsSemi-supervised learning method
[58]GNN Twitter15 Twitter16GCN extracts propagation structure features in the encoderGraph Auto-encoder Extract both dissemination structure features and text structure features at the same time
[116]GNN CNN Twitter15 Twitter16 WeiboGCN extracts graph structure featuresGraph attention network to obtain structural semantics
[22]GNN Twitter15 Twitter16GCN extracts graph structure featuresUse the characteristics of transmission structure to judge the accuracy of the source of rumors
[52]GNN RNN PHEME RumorEvalGCN extracts nonlinear diffusion featureCombination of dual propagation feature ,GCN extracts nonlinear features, LSTM extracts time series linear features

DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure

Detail of rumor detection model based on GNN DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure

Rumor detection based on transformer

Since CNN is easy to parallelize, it is not suitable for capturing dependencies within the field sequence. RNN can capture the dependence of long-distance sequences. However, it is challenging to realize parallel processing sequences. In order to integrate the advantages of CNN and RNN, [96] combined the attention mechanism to design Transformer. This model uses the attention mechanism to achieve parallel capture of sequence dependencies and can process tokens at each position of the sequence at the same time. In recent years, Transformer has been proven to perform well in natural language processing tasks such as machine translation ([24]). Therefore, there are studies to migrate the Transformer structure to the rumor detection work. Since Transformer can capture the dependence of long-distance field sequences, [47] used the multi-attention mechanism in Transformer to model the long-distance interaction between tweets. The tree structure information is disturbed in the research, and the dependence is obtained from user comments between different conversation threads. Use Transformer’s multi-attention mechanism to obtain the characteristics of each original blog post and retweet comment, calculate the correlation between them and give the correlation weight so that the user’s comment characteristics can be used as the basis for judging rumors more accurately. [115] used the establishment of a transformer specific to position detection and a cross-task Transformer. They used the multi-head attention mechanism to guide the model to pay more attention to the particular characteristics of position rumors, to obtain the dependence of position and rumors, and respectively For rumor detection, use predicted position tags as the basis for judgment. Transformer is based on the encoder-decoder architecture. The Transformer’s encoder can encode word embedding vectors into high-dimensional representations containing text features. Therefore, some studies ([23, 48]) use Transformer’s encoder to extract features in word vectors. For example, [23] and [48] encoded all news through the Transformer encoder to generate the representation of the news and extract text features for subsequent calculations. [62] proposed a Transformer tree-based method to utilize user interaction in conversations. The study models the propagation of each statement as a tree structure. Transformer is established as a bottom-up Transformer, a top-down Transformer, and a hybrid Transformer model in this work. The former compares the response tweets of each book borrowing point in pairs to capture the consistent attitude towards each tree node. The latter describes how information flows from the source blog post to the current node. Finally, a hybrid Transformer is used to fuse position features and structural features for rumor detection. Table 5 is the detailed information of the latest work based on Transformer’s rumor detection. The existing Transformer-based rumor detection work does not make good use of the dissemination of structural information.
Table 5

Detail of rumor detection model based on Transformer

PaperDL ModelFeatures in experimentDatasetDescriptionCharacteristic
TSVP S
[26]Transformer Twitter15 Twitter16Transformer captures the reliance between tweets and retweetsBreaking the propagation tree structure, the retweet information in different conversation threads can be calculated for correlation
[115]Transformer SemEval2017 PHEMETransformer extracts position featureUsing tags for position detection makes the results of rumor detection more accurate
[23]Transformer PolitiFact GossipCop PHEMETransformer’s coding layer term news codingExternal knowledge using knowledge graph
[48]Transformer CLEF-EN CLEF-Ar MediaEval LESATransformer extracts text featuresExplore the importance of images in rumor detection
[62]Transformer Twitter PHEMETransformer extract position and attitude feature and communication featureTree Transformer to preserve the user interaction of the dialog

DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure

Detail of rumor detection model based on Transformer DL:Deep Learning; T:Text; S:Social; V:Vision; P S:Propagation Structure

Rumor Detection Based on Other Structures

There are much other deep learning structures[12, 89]. In addition to the above methods, there are studies to build their algorithms through other model structures[86, 88] and obtained good experimental results. For example, in the work of [88], XLNet is used to map the contextual content of rumors to a high-dimensional to learn the content features of rumors and combined with the rumor topic distribution of LDA(Latent Dirichlet Allocation) to solve the problem of rumor detection for COVID-19. Since this method focuses too much on the content itself, it cannot guarantee the outbreak of new rumors in the new crown era. However, in order to solve the problem of difficulty in the communication of rumor features between different domains, [86] separated the domain-specific features and non-domain-specific features in the rumor text content through a fully-connected network. Weakly supervised classification on unlabeled datasets of range domains. Nevertheless, they ignored the full convergence speed and high computational complexity of the fully connected network.

Rumor detection methodology

To answer RQ3, in this chapter, we classify rumors detection methods based on the newest work and introduce various novel methods used in the newest research to detect rumors. After experiments, these methods are better, more efficient, and more automated than traditional manual features methods. The proportion is higher. The structure of this article is shown in the Fig. 5. And summary about the deep learning architectures, tools/libraries and performance matrices used in research studies in Table 6.
Fig. 5

Classification and related work of the latest methods of rumor detection

Table 6

Detail of deep learning research studies in Section 5

StudyDL MethodPerformance metricsTool/ LibraryDesctiption of Architecture
[67]RvNNAcc.:0.723 NR:F1:0.682 FR:F1:0.753 TR:F1:0.821 UR:F1:0.654theanoThey input each propagation node into the RNN and propagate the features along the path of the propagation tree to the next RNN Unit.
[52]RNN GNNAcc.:0.925 Pre.:0.9245 Recall:0.925 F1:0.9247-Graph convolutional neural(GCN) network to extract non-Linear structure features, LSTM extract linear sequence feature.
[47]RNN TransformerAcc.:0.85 NR:Acc:0.844 FR:Acc.:0.857 TR:Acc.:0.883 UR:Acc.:0.814pytorch tqdm sklearnUse RNN to learn context and use Transformer to find tightly related conversation threads
[68]RNNAcc.:0.781 Rumor: Pre.:0.796 Rec.:0.796 F1:0.784 None-rumor: Pre:0.791 Rec.:0.766 F1:0.778theanoGated Recurrent Unit used for representing hidden units
[101]CNN RNNAcc.:0.827 Pre.:0.847 Recall:0.812 F1:0.829sklearn pytorch scipy jieba Word2VecTextCNN extracts text features and VGG extracts visual feature
[46]RNN CNNAcc.:0.824 Fake: Pre.:0.854 Recall:0.769 F1:0.809 Real: Pre.:0.802 Recall:0.875 F1:0837keras sklearn Word2Vec nltkBi-LSTM as both encoder and decoder to extracts text features and VGG-19 with double fully connect layers extracts images features.
[19]RNNAcc.:0.752 Pre.:0.755 Recall:0.752 F1:0.752-Bidirectional LSTM Unit as the hidden units of encoder-decoder.And learn multi-task features by Bi-LSTM.
[86]DenseAcc.:0.877 Pre.:0.840 Recall:0.832 F1:0.836tensorflow keras sklearn nltkDense Net as the domain-specific discriminator,the domain-shared discriminator and generator.
[120]RNN CNNAcc:0.763-Use GRU as the hidden unit of decoder-encoder; Residual Blocks as content encoder and decoder to Extract visual features.
[118]RNN CNNAcc:0.866 Rumor: Pre.:0.893 Recall:0.847 F1:0.870 Non-Rumor: Pre.:0.866 Recall:0.867 F1:0.866-VGG-19 extract image feature and Bi-GRU extract text feature. The rumor characteristics between different posts are transmitted through GRU,and Separating domain specific features with event memory network.
[19]RNNF1:0.405 Acc.:0.405kerasLSTM is used to learn text features, features of rumor detection task, stance detection features and rumor verification features.
[119]RNNAcc.:0.638 F1:0.606-LSTM unit is used to Rumor Verification and Stance detection.
[66]RNNF1:0.464 NR:F1:0.876 FR:F1:0.543 TR:F1:0.114 UR:F1:0.333theanoUse GRU to merge the characteristics of the two tasks.
[48]RNN BERT CNNAcc.:0.819 Pre.:0.75 Recall:0.8667 F1:0.8041pretrained BistilBERTword embedding and image embedding by pretrained BistilBERT and VGG19. Use GRU as the hidden layer unit in Meta Multi-task layer.
[23]TransformerPre.:0.8687 Recall:0.8499 F1:0.8539 Acc.:0.8586 AUC:0.9197Wikipedia TagMe Word2VecTransformer as encoder to encode entities, news, and entity contexts sequence.
[38]GNN RNNF1:0.8912 Prec:0.8982 Recall:0.8917pytorch Word2Vec functools TagMEGCN learns the structural characteristics of heterogeneous graphs with attention mechanism and use LSTM to embed information describing entities.
[88]CNN XLNetAcc.:0.846keras pandas sklearnUse Pretrained XLNeT to embed the title and content to get the text features.Use VGG-19 to embed the image to get visual feature.
[102]CNNAcc.:0.824 AUC:0.873pytorchTextCNN to get text embedding.
[63]RNNF1:0.807 True:Pre.:0.637 Recall:0.665 F1:0.651 False:Pre.:0.874 Recall:0.860 F1:0.867theanoGRU with attention mechanism for claim verification.

DL:Deep Learning; NR: nonrumor; FR: false rumor; TR: true rumor; UR: unverified rumor; Pre.:Precision; Acc.:Accuracy

Classification and related work of the latest methods of rumor detection Detail of deep learning research studies in Section 5 DL:Deep Learning; NR: nonrumor; FR: false rumor; TR: true rumor; UR: unverified rumor; Pre.:Precision; Acc.:Accuracy

Propagation trees method

Since tweets in social media will lead to many comments, and users can further comment on the comment information, thereby forming a variety of branches. In this way, [104] modeled the spread of rumors in the form of a tree, as Fig. 6 [65] identified rumors by capturing the substructure of the propagation tree to evaluate the similarity between the propagation trees. However, this study only compares the structural similarity between propagation trees. Afterwards, [67] improved [65] and established a tree-based recurrent neural network model, taking a propagation tree rooted at the source post as input and passing a bottom-up and top-down tree structure recurrent neural network (Fig. 6 uses a top-down RvNN as an example) jointly captures the linear sequence feature . Compared to [65], this method does not require comparison the structure of propagation tree.
Fig. 6

Architecture of Propagation Trees Method.Top-down RvNN is used as an example to extract the linear sequence features of Twitter propagation. Some studies have taken the nonlinear structural features of the propagation tree and fused with the linear features to classify rumors

Architecture of Propagation Trees Method.Top-down RvNN is used as an example to extract the linear sequence features of Twitter propagation. Some studies have taken the nonlinear structural features of the propagation tree and fused with the linear features to classify rumors [52] mentioned a combination of nonlinear structure learning and linear sequence learning to learn the propagation characteristics in the propagation tree. This study uses nonlinear structure learning to learn diffusion propagation features along the diffusion path of the propagation tree and uses the linear sequence to learn to aggregate the features of context nodes and retains temporal features to represent sequential propagation. The graph convolutional neural network is used to extract the nonlinear propagation structure features, and LSTM is used as the linear feature learning that retains the timing. Finally, combining two features to classify rumors and explore the relevance. However, different from the methods mentioned above, [47] flattened the structure of the propagation tree, allowing all possible pairwise interactions between tweets, and used the self-attention mechanism to obtain two significantly related tweets from a large number of tweets. Because they believe that although the tree model can model the structural information that exists in the dialog thread, and the information is passed from parent to child in the tree model, each user can usually observe all the replies in different branches of the dialog. The content that debunks fake news may be tweets used in other tree branches. However, although their approach can link the branch information of different propagation trees, it destroys the structure of the propagation tree and cannot use structural features. Although they ([65],[67], [52], [47]) use structure and content information, not every propagation path is essential. Some users are easily affected by rumors, and their comments may affect the judgment of rumors. Therefore, they lack information on users. The use of attributes to assign the weights of propagation paths to different users.In addition, there is a lack of work on the changing trend of public opinion in the use of the communication path. Because many comments believe that rumors will affect subsequent comment users, it will cause the user’s position in the communication path to change, which is also the trend of public opinion.

Adversarial learning and automatic coding structure method

Figure 7 shows the rumor detection framework of Adversarial Learning and the rumor detection framework of Automatic Coding Structure. Among them, Adversarial learning generally uses generators to generate real information and rumors from rumors and real information to train the discriminator to judge the features of rumors in all directions. After the final training, the discriminator is used to distinguish between rumors and real information. The rumor detection method of the self-encoder usually uses an encoder to encode the rumor data set into a feature vector, and tries to restore the feature vector to the original data through the decoder structure opposite to the previous encoder, and feeds back the generated data to the encoder by comparing the generated data with the original data. Finally, the feature vector generated by the well-trained encoder is classified into rumors through the fully connected layer and activation function.
Fig. 7

Architecture of Adversarial Learning(below) and Automatic Coding Structure(above) Method

Architecture of Adversarial Learning(below) and Automatic Coding Structure(above) Method [68] pointed out that rumor makers expand the spread of rumors, which poses greater challenges to such data-driven methods. Therefore, using the characteristics of generative adversarial networks, generators are used to simulate rumors and propaganda to output more challenging examples, so as to promote the discriminator to strengthen the feature learning of such difficult examples to capture more discriminative patterns. The research uses GRU as generator to get representation of rumors. In the model, the generator is encouraged to generate motion-like instances to fool the discriminator, and the discriminator focuses on learning more discriminative features. The discriminator can master more discriminative features by using the data features combined by the generator, especially the information of non-trivial patterns learned from low-frequency patterns. Later, [101] proposed a cross-domain rumor detection framework, EANN(Event Adversarial Neural Networks), in which a multi-modal feature extractor was used as a generator to generate rumor feature vectors to deceive the event discriminator in the framework. Expect to abandon the characteristics of events in a particular field. On the other hand, the event discriminator tries to find the information related to the specific event contained in the feature representation to identify the event. [46] got inspiration from EANN and built an end-to-end network with variational auto-encoder called Multimodal Variational Autoencoder (MVAE), using a dual-modal autoencoder and a classifier to complete Rumor detection task. It mainly includes three parts: encoder, decoder and detector. Encoders and decoders use the idea of adversarial learning. The encoder learns features from both text and image, and then encodes them as vectors. The decoder takes the output vector of the encoder as input and decodes it into text and image features. And through the decoded features and the input features of the encoder to calculate the loss to train the encoder and the decoder, and finally complete the rumor detection judgment through the detector. [19] took inspiration from MVAE and created a text-based multi-task auxiliary variational autoencoder. In this work, a LSTM-based variable autoencoder model was proposed to pass the text The RNN encoder with LSTM unit performs encoding, and then uses the same parameter RNN network for decoding. The current self-encoding method or adversarial learning method for rumor detection only rests on restoring the image and text features but lacks the generation of structured data sets, such as propagation trees and propagation graphs. Although [110] has used graph structure adversarial learning for the learning task of propagating graphs, they only consider the detection of abnormal propagation points to help the rumor detection task.

Cross-domain method

Figure 8 shows the rumor detection framework of cross-domain method. Rumor detection methods were based on specific events or specific areas of learning, resulting in failure to identify fake news in real-world news streams. For example, specific nouns in the fields of politics, entertainment, medical care, etc. will cause the training results to perform poorly in other fields. However, research hopes that a rumor detection method can be truly applied in the real world. [35] also showed in their research that most fake news detection techniques are not good at identifying fake news from rare fields and recording their characteristics during training. Therefore, there are many studies aimed at solving how to retain specific domain and cross-domain knowledge in news and rumor records to detect fake news in cross-domain news data sets.
Fig. 8

To improve the performance of cross-domain rumor detection, the primary research is to remove the domain-specific features in the rumor feature. After the features are extracted, the domain discriminator is used to improve the performance of the feature extractor to remove the features of the specific domain. Regardless of how the latest research assists in domain elimination tasks, the primary method used remains in adversarial learning

To improve the performance of cross-domain rumor detection, the primary research is to remove the domain-specific features in the rumor feature. After the features are extracted, the domain discriminator is used to improve the performance of the feature extractor to remove the features of the specific domain. Regardless of how the latest research assists in domain elimination tasks, the primary method used remains in adversarial learning Wang et al. [101] pointed out that most early models have poor monitoring performance for emergencies and cannot capture the feature of different events that are not shared in specific events. To solve this problem, we propose to extract cross-domain features from multi-modal data of fake news. Use the adversarial learning method mentioned to remove the specific event features in the multi-modality and keep the feature representation of the event unchanged. The effectiveness of cross-domain rumor detection has been improved. Silva et al. [86] proposed a framework to save news features in specific fields jointly and cross-fields to detect false news from different fields. The framework integrates an unsupervised domain embedding learning module and a regulated and domain-independent news classification module. The unsupervised domain embedding learning module uses multi-modal tasks to express news as low-dimensional representation and give labels to unlabeled news through learning. The classification module learn specific domain and cross-domain knowledge in news while identifying fake news. The research maps the representation of news to two parts for learning specific domain knowledge and cross-domain knowledge respectively. In 2021, [120] were inspired by [101] to conduct cross-domain event fake news detection work for multi-modal data and introduced the multi-modal separation representation learning in study of [101] to social media rumors detection to explore the commonalities and characteristics of different modalities across domains. A multi-modal unsupervised domain-adaptive method is proposed, which can derive event-invariant features, conducive to detecting rumors about emerging social media events. In the research, the rumor style feature classifier and content style classifier are trained on text and images respectively to extract rumor style and content features from visual information and content information. The unsupervised domain adaptation module based on adversarial learning learns the characteristics of the transferable rumor style in multimedia posts and transfers the knowledge obtained from historical events to new events. Unlike the above work, [118] extracted entity knowledge from the outside and record it in shared storage space. When a high-level representation of an emerging event is given, the event memory network will retrieve it from the external knowledge base and output the features of entities existing in events in the same field or features of rumors with similar entities. The work mentioned above is from removing specific domain features or extracting external knowledge to retain the general domain features of rumors as cross-domain rumor detection methods. Although these methods solve the interference problem of domain nouns in some memorable domains, in the research process Ignoring the different ways of dissemination of rumors in different fields because many users have limited understanding of specific fields, the spread of rumors has accelerated, and users who question rumors are even rarer.

Multi-task learning

Multi-task Learning is a machine learning method opposite to Single-task Learning. In traditional machine learning, the standard algorithm theory is to learn one task when the output of the system is a real number. Multi-task learning is a kind of joint learning, where multiple meta-tasks are learned in parallel so that the learning results influence each other. That means multi-task learning is to solve multiple meta-problems at the same time. In the latest research, studies are using multi-task learning methods to solve rumor detection ([19, 48, 56, 66, 119]). In the latest work, a large amount of literature uses user comments and evidence to extract the user’s position on rumors. This type of work is called position detection. Joint stance detection task and rumor classification task, and then share the features extracted from the two tasks for rumor detection, as Fig. 9 Many experiments have proved that position detection has a huge positive impact on rumor detection. [66] and [48] used GRU to optimize stance detection jointly, and rumor classification tasks in their work, and both achieved significant improvements. However, they assign the same weight to each user’s comment, ignoring that some untrusted users’ comment characteristics have little or even a negative effect on rumor detection..Since not all users have high credit, so [119] added user confidence features and text feature embeddings in their work. Nevertheless, it still ignores the structure of the comment information [49] proposed a tree-structured LSTM multi-task model with convolutional units and used the tree structure to propagate effective position signals upwards to classify rumors at the root node. However, the above research did not correlate the results calculated by position detection with the specific output layer of rumor detection, so [115] added a Transformer layer to the multi-task framework of joint position detection and rumor classification to capture one of the blog posts in the entire conversation thread. By combining two transformer components to solve the problem of not explicitly modeling the interaction between the position-specific layer and the rumor-specific layer in previous work.
Fig. 9

Architecture of Multi-task learning for Rumor Detection. Including two tasks: Stance Detection and Rumor Verification. These two tasks are also the choice of most multi-task rumor detection. Helping rumor detection tasks by detecting the stance of comments

Architecture of Multi-task learning for Rumor Detection. Including two tasks: Stance Detection and Rumor Verification. These two tasks are also the choice of most multi-task rumor detection. Helping rumor detection tasks by detecting the stance of comments For the multi-task learning method that uses stance detection to help rumor detection, although users’ feedback on rumors can be used to make preliminary judgments on rumors, we believe that this method negates the role of public opinion trends. The dissemination harm of rumors is the pressure of public opinion trends. Some users who are not firm will shake their stances when observing the positions of the public. Stance detection will have a counterproductive effect on this phenomenon, which will mislead the task of rumor detection. Although some studies have used user credibility information, they still have not made any countermeasures against changes in public opinion trends. Therefore, this is also a challenge to the multi-task learning method using stance detection. The framework of [19] is different from that shown in Fig. 9. The reason is that they were inspired by [126] that they classified rumors into four parts: rumor detection, rumors Tracking, position classification, and accuracy classification. So, [19] took these four parts as four meta-tasks and established Rumor Detector, Rumor Tracker, Stance Classifier, and Veracity Classifier to target the above four tasks respectively and combine the results of the four classifiers to assist the rumor detection task.

Knowledge graph method

Figure 10 shows the architecture of Knowledge Graph (KG) methods for rumor detection. The knowledge graph is a semantic network that reveals the relationships between entities. There are millions of items describing real-world entities, such as people, places, and organizations. In the knowledge graph, entities are represented as nodes, and the relationships between these nodes are described as edges. Knowledge graphs are widely used through entity links, such as movie recommendation ([117]), machine reading ([113]), text classification ([99]), etc. In the latest work, some documents extract the entity nouns in the rumors and then obtain knowledge other than the rumors through the knowledge graph to assist the rumor detection task.
Fig. 10

Architecture of Knowledge Graph(KG) Methods for Rumor Detection. The steps of current general method for rumor detection based on knowledge graph : first identify the entity of the content of the rumor, then entity link to obtain external knowledge from knowledge graph , then represent the knowledge to vector, last classify the rumors after fusion or comparison with the original features

Architecture of Knowledge Graph(KG) Methods for Rumor Detection. The steps of current general method for rumor detection based on knowledge graph : first identify the entity of the content of the rumor, then entity link to obtain external knowledge from knowledge graph , then represent the knowledge to vector, last classify the rumors after fusion or comparison with the original features [77] first proposed a content-based fake news detection method using knowledge graphs and created three types of knowledge graphs: based on fake news databases, open knowledge graphs, and reliable news organizations. The actual news library. And use the triple information extracted from the fake news to retrieve the created knowledge graph, represent the triple in the vector space, and judge the truth of the news article through the vector.However, this method can only record news events that have occurred, and its ability to detect new rumors is minimal. [23] integrated the external knowledge language news detection from the knowledge graph, identified the entities in the news content, and matched them with the entities in the knowledge graph. The context in the knowledge graph is used as external knowledge for supplementary information, and each entity is given a weight to express its importance.However, they embed the context in the knowledge graph together. Since new knowledge may be missing in the knowledge graph in new outbreaks, the impact on the rumor detection task is not stable. [38] proposed a CompareNet, which uses the knowledge in the news knowledge graph to align with the entities of the news text to find the corresponding entity in the knowledge graph, that is, context information. The structural information and text knowledge of the joint knowledge map is compared with the description of the entity in the original news, and similar features are calculated, and the high-dimensional features of the original news are combined to classify news. Compared with [23] ,their method compares the embedded information and structural features of the original text to get better results in entity comparison. However, they lack the consideration of multiple feature fusion weights in the feature fusion process.

Unsupervised and semi-supervised learning methods

Unsupervised learning and semi-supervised learning are two typical methods in machine learning. The unsupervised method is due to a lack of sufficient prior knowledge and has to solve the problem in pattern recognition based on the training samples of the location category. Semi-supervised learning is a learning method that combines supervised learning and unsupervised learning. Semi-supervised learning uses a large amount of unlabeled data while using labeled data for pattern recognition. Because the relevant data sets for rumor detection are less labeled and involve a wide range of fields, the professional requirements of the labeled personnel are higher, making it challenging to label data sets for rumor detection. Therefore, there are works for unsupervised learning methods and semi-supervised learning methods for rumor detection. Because there are fewer tagged articles in the current popular data set, and the scope of inclusion is relatively narrow. However, there is the possibility of inaccuracy in marking news in a crowdsourced manner. Guacho et al. [30] mapped the text to the latent representation in the Euclidean space and proposed a semi-supervised learning method combining the K-nearest neighbor algorithm with the graph neural network and the graph attention neural network to classify the text into rumors and non-rumors. Yang et al. [109] proposed an unsupervised generation method for social media fake news detection, using auxiliary information of users participating in news tweets on social media, extracting users’ opinions on news, and collect opinions in an unsupervised way and generate estimates result. The authenticity of news and the credibility of users are regarded as potential random variables, and users’ participation in social media is used to determine their views on the authenticity of the news. And use the Bayesian network model to capture the conditional dependence between news authenticity, user opinions, and user credibility. Since rumors in social media are designed in many fields and require a large number of domain experts to classify rumors, the data set labeling of rumors detection is a very difficult and requires a lot of money. In theory, unsupervised learning methods And semi-supervised learning methods can solve this kind of problems very well ([30],[109],[41]). However, although there have been some studies, there are still no outstanding applications.

Other methods

In addition to the rumor monitoring methods mentioned in 5.1-5.6, there is also much work using new technologies for rumor monitoring tasks. For example: transfer learning ([88]), reinforcement learning ([102]), etc. Singhal et al. [88] uses a multi-modal method that uses transfer learning to obtain semantic and contextual information from news articles and related images to improve fake news detection accuracy. The research uses pre-trained language embedding and image network models to extract features. These feature vectors are input into fully connected layers for classification. Wang et al. [102] proposed a weakly-supervised framework based on reinforcement learning. The framework can use user reports as weak supervision to expand the training data used for fake news detection. The framework contains three components: an annotator, a fake news detector, and an enhanced selector. Using a set of labeled fake news samples and user feedback on these news articles, the framework can train annotators based on the feedback and automatically assign weak tags for user feedback to unlabeled news articles based on the content. Using reinforcement learning technology, the reinforcement selector selects high-quality samples from the weakly labeled samples as the input to the fake news detector. The fake news detector finally assigns a label to each input article based on its content. Rosenfeld et al. [79] studied the effect of separate diffusion mode on rumor detection, used graph kernel to extract complex topological information from Twitter cascade structure, and trained a predictive model that ignores language, user identity and time, and proved the removal for the first time. The diffusion model of complex information has high accuracy. The results show that through proper gathering, even in the early stages of dissemination, the spread pattern of rumors in the crowd may reveal a powerful signal of the truth and falsehood of the rumors, and at the same time prove that the spread pattern can indeed provide the accuracy of the rumors. Ma et al. [63] proposed a sentence-level embedding method for rumor detection based on hierarchical attention networks, which integrates the implicated relationship between each sentence of the rumor and the evidence and ensures the consistency of the evidence. The end-to-end hierarchical attention network designed in the work The end-to-end hierarchical attention network is used for sentence-level evidence embedding. The purpose is to focus on important sentences/evidence by considering the topic coherence and semantic reasoning strength. The model can more reasonably determine the rumor’s verdict and embed the evidence sentence into the learned statement. At the same time, with the help of attention, key evidence can be highlighted and cited to explain the judgment better.

Dataset

To answer the RQ4, the types of datasets available for rumor detection research are introduced in detail with this section. The detailed information of the data set is summarized and plotted in Table 7, in which only the relevant data set information given by the original provider of the data set is provided. If the number is deleted and modified by other work, no statistics will be made.
Table 7

Detail of dataset

DatasetReferenceSizeLabelNumber for labelsData typeDetail of dataset
PHEME[127]5802 tweetsrumor /none-rumor1792/3830textIt contains five events: Charlie Hebdo, Ferguson, Germanwings Crash, Ottawa Shooting, Sydney Siege. It’s imbalanced dataset.
Twitter15[65]1380 tweet-treenon-rumor /false-rumor /true rumor /unverified374/370/ 372/374Tree-structure textIt contains 276663 users, 1490tweets, 331612 threads. It’s balance dataset.
Twitter16[65]1181 tweet-treenon-rumor /false-rumor /true rumor /unverified205/205/ 205/203Tree-structure textIt contains 173486 users, 818 tweets, 204820 threads. It’s balance dataset.
MutiSource-Fake[28]11397 newsfake/real5403/5994textIt contains news from OpenSources.co, MediaBiasFactCheck.com, PolitiFact news websites’ lists. It’s balance dataset.
Weiibo[42]146 events, 50287 tweetsrumor /none-rumor23456/26257text, image, usersReal world multimedia dataset from Sina Weibo contains 50287 tweets and 25953 images and 42310 users. It’s balance dataset.
MediaEval[6]15821 tweetsrumor /none-rumor9596/6225text, imageIt contains tweets related to the 11 events, comprising in total 193 cases of real and 220 cases of misused images and videos, associated with 6225 real and 9596 fake tweets posted by 5,895 and 9,216 unique users respectively. It’s imbalanced dataset.
FakeNewNet[83]23196 newsfake/real5755/17441text, image, users, network, responseThis dataset contains 23196 news, 19200 images from PolitiFact and GossiCop. It’s imbalanced dataset.
Buzzfeed Election[36]71 newsfake/real35/36textIt collected the news stories found in Buzzfeed’s 2016 article on fake election news on Facebook (Silverman 2016). It’s balance dataset
LIAR[100]12836 shot textspants-fire/ false /barely-true /half-true /mostly-true /true-textIt collected a decade-long, 12.8K manually labeled short statements in various contexts from POLITIFACT.COM, which provides detailed analysis report and links to source documents for each case.
Twitter[101]13924 newsfake/real7898/6026text, imageThis dataset contains 13924 news and 514 imgaes from MediaEval. It’s imbalanced dataset .
WeChat Datasets[102]27161 articlesfake/real2090/2090textThis dataset collected from WeChat’s Official Accounts, dated from March 2018 to October, 2018.It contains 27161with 37971 reports,and 22981 articles is unlabeled. It’s balanced dataset.
Twitter[92]4 million Tweetsfake/real-textThis dataset is unpublished and contains 4 million tweets, 3 million users, 28893 hashtags, and 305115 linked articles, revolving around 1022 rumours from 01/05/2017 to 01/11/2017.
Twitter[98]126000 rumor cascadesrumor/non-rumor-text126000 rumor cascades spread by 3 million people more than 4.5 million times from Twitter inception in 2006 to 2017.
Weibo-20[122]6362 articlesfake/real3161/3201textThis dataset is constructed on thr basis of Weibo-16([42]), it contains 6362 articles whit 1983440 comments. It’s balanced dataset.
FEVER[93]185,446 statementssupport /refuted /not enough info93367/43107/ 48971textIt contains judgment and evidence of the statement.And the statements are from Wikipedia. It’s imbalanced dataset.
Fakeddit[75]825100 newsfake/real; completely true /fake news with true text /fake news with false text ; true /Satire(Parody) /Misleading Content /Imposter Content /False Connection-text, image, userDataset for 2-way , 3-way, and 5-way classification. It sourced from Reddit. It contains 628,501 Fake samples and 527,049 True samples. It also contains 682,996 Multimodal samples and 358,504 users. It’s imbalanced dataset.
NewsBag[44]215000 newsfake/real200000 /15000text, imageIt contains 200,000 real news , 15,000 fake news and 15000 images. It’s imbalanced dataset.
Fake News Evolution[32]950 paired datareal/fake /Turned into fake news-textThis dataset includes 950 pieces of data, each of them contains three articles representing the three phases of the evolution process, and they are the truth, the fake news and the evolved fake news.
Detail of dataset Most of the work uses data sets from Twitter, Weibo, and news in the latest research. Some of these data sets were obtained by researchers based on the public APIs of social media. For example, [65] used their published APIs to collect data from Twitter to establish the Twitter15 and Twitter16 data sets, which contain 1381 and 1181 propagation trees, respectively. And use four labels to annotate each tree: non-rumors, false rumors, true rumors and unverified rumors. In addition, jobs are using publicly available news event data sets. For example: MediaEval([6]), Buzzfeed Election([36]), PHEME([127]), LIAR([100]), etc. Among them, PHEME uses manually marked rumors and non-rumors, including original blog posts and replies about nine breaking news events. The Buzzfeed Election dataset includes complete Facebook news from September 19 to September 27 in the 2016 U.S. election. LIAR contains fake news from the PolitiFact collection, including short statements, in the form of press releases, TV news, etc. However, in the latest research, researchers have found that most of the existing data sets are small in scale or contain only a few areas of fake news. For this reason, many works have proposed new large-scale, multi-domain data sets. For example, recent work proposes a large-scale, multi-modal data set NewsBag([44]), containing 200,000 real news and 15,000 fake news. The real news comes from The Wall Street Journal, and the fake news comes from The Onion. [32] believe that some fake news is maliciously modified or distorted in the process of multiple dissemination of real news. If this situation is ignored, the model will be deceived. Therefore, a new data set to track the evolution of fake news is released, called the Fake News Evolution (FNE) data set. The data set comprises 950 paired data, and each data set is composed of articles representing three important stages of the evolution process, namely, truth, fake news, and evolved fake news. There are also studies focusing on claim extraction and verification. For example, [93] disclosed a data set called FEVER, which consists of 185,445 statements generated by modifying sentences extracted from Wikipedia, and then without knowing the source of these sentences authenticating. These statements were classified as supporting, refuting, and insufficient information by the commenter. And provide the annotator’s judgment basis in the data set to judge the first two types of information.

Potential issues and future work

In the past few years, to make the information contained in the network more reliable, researchers worldwide have done much work and made considerable improvements. However, some key areas have not been solved well. In order to answer the RQ5, this chapter focuses on the shortcomings of current research and gives the challenges faced by rumor detection and potential directions for future research.

Data ethics issues

Although the research on rumors detection is vibrant, it has been applied to some scenarios, such as Weibo, rumors have been improved. However, in the process of exploration in this field, some ethical issues have been faced. In the research process, most social science researchers collect Twitter data, such as text content, pictures, videos, and even private information such as the gender of the account owner, the age of the user, and ethnicity. Twitter is particularly challenging in ethics because its data is partially accessible. Although the terms of service stipulate that users’ public posts will be provided to third parties, and by accepting these terms, users will legally agree to this, but surveys ([103]) show that less than two-thirds of people have read these terms in whole or in part. Only 76% of this group know that when they accept the terms of service, they agree to a third party to access some of their information. Although some studies have noticed this problem, they did not disclose their information or hide private information such as user ID for confidentiality or privacy protection, but there is still a problem of privacy leakage. For example, some public rumor data sets (such as Fakeddit([75]), Twitter Dataset([101]) have compassionate information related to politics and social figures, but no effective measures have been taken to protect privacy. Furthermore, most of them have not obtained the informed consent of users. In addition, survey data shows that 49% of respondents from the government and 51% of commercial companies have concerns about the use of the Twitter data set for research, and respondents expressed a high degree of agreement with the Twitter research on consent to anonymity, and hope to get their consent before publishing academic results using their Twitter posts. Therefore, the issue of data ethics is one of the issues that need to be paid attention to in future research of rumor detection.

Algorithmic ethics issues

[126] stated that the rumor detection algorithm performed well on social media data sets surrounding specific events. However, due to changes in language usage, their accuracy will drop beyond the events they developed. Therefore, as to how researchers should develop, use and reuse algorithms, they usually use sensitive labels to classify content and users without their knowledge, which creates ethical challenges. According to the scale and speed of the data, researchers should ensure that the algorithm performs well on rumor detection to establish standards for rumor classification and ensure that the data labels used are accurate. In addition, if research intends to use rumor classification algorithms for data outside of the designed algorithm data set, the researcher should be responsible for ensuring the continued effectiveness of the classification algorithm because the failure of big data will cause the masses to question the power and longevity of algorithms ([53]). Therefore, rumor detection algorithms need to be publicly released and reproduced transparently so that their effectiveness can be tested regularly to avoid mislabeling content and users.

Legal issues

The data extracted from the Twitter API contains personal information, which is subject to relevant data protection legislation. Therefore, researchers should establish a fair and legal basis for collecting personal information if the user’s informed consent cannot be obtained. Researchers can accept the terms of service of social media networks, and the terms of service provide sufficient conditions to cover relevant data protection legislation. At present, many countries have issued data protection laws, such as the British Data Protection Act (DPA), the Personal Information Protection Law of the People’s Republic of China, and other documents ([103]). Therefore, in the future research process, the collection of relevant personal factor data in social media needs to be collected, used, and disclosed under the relevant data protection law conditions.

Real-time Visualization and learning for rumor detection

Although the research of rumor detection is gradually mature, the current rumor detection task is still based on binary classification (Real/Fake) or ternary classification (Real/Fake/Unverified). However, due to the difficulty and complexity of rumor detection tasks in real scenarios, binary or ternary classification models are far from sufficient to identify the characteristics of abnormal online information. Due to the real-time and heterogeneous nature of social communication data, data visualization is a powerful tool to illustrate different aspects and distribution patterns of online social media information. The interactive visualization system can facilitate human supervision and understanding according to the different latitudes and maps of the data, interpret the time-based patterns and behaviors of the data, and summarize important features more clearly. Nowadays, there have been many studies on social media visualization extensively ([76], [50], [80],[3]). Except for the work of ([123]), few studies have focused on the interactive exploration of visualizing rumor information on social media. Such a visualization platform can indicate the nature of information dissemination and be regarded as an information resource about the relationship between online users. Through the real-time rumor visualization system, the rumor information or abnormal user behavior can be detected at the moment when the rumor information or user’s abnormal behavior occurs. Then an appropriate rumor rejection mechanism can be used to limit the negative impact of the rumor. An online system is a real-time protection measure that attempts to protect online readers immediately. In addition, in combination with real-time rumor detection learning, progressive ideas such as reinforcement learning are used to strengthen the ability of rumor detection and real-time learning of the content and rules of newly erupted rumors. Moreover, there is currently a lack of extensions that deploy research technologies into Web-based real-time verification applications. Such applications should use reinforcement learning methods to learn the characteristics of the latest outbreak of rumors in an emergency and provide the ability to detect rumors. Combined with the real-time rumor visualization system, Internet users or network experts can be one step ahead of the full spread of online rumors, thereby reducing the impact of such information attacks. Usually combined with real-time learning and visualization technology, real-time systems can keep up with the new trend of rumors and can be applied to more scenarios. Therefore, an essential part of online rumor detection and monitoring in real-time visualization systems is an area worth exploring.

Dataset issues and unsupervised learning for rumor analysis

Although there are many public datasets in the field of rumor detection, and social media such as Weibo and Twitter have opened APIs for scientific researchers to obtain data. However, labeling the rumor detection data set is a tricky thing. Rumors involve many fields, and the work of labeling rumors requires professionals in different fields to complete, and the training of such professionals is challenging and costly. Much money. In addition, even for professionals in the field, labeling news as true or false is a very challenging task for them ([41]). Some works ([102]) suggest using crowdsourcing to label the rumor detection data set, but the accuracy of the labeling cannot be guaranteed. Some works ([23], [86], [84], [125]) believe that data from some Fact-checking platforms can be used as a rumor detection data set. However, such platforms provide The accuracy of the data is to be investigated, and it is mostly limited to major areas such as politics, medical treatment, and society, and lacks data in proprietary or small domains. Therefore, dataset annotation are difficult, and there is a lack of publicly available large-scale data sets is also one of the challenges to be solved in the future. The unsupervised method can be applied to the actual analysis of real-world data sets. We mentioned the current rumor detection research using unsupervised and semi-supervised learning methods in 5.6. However, we think that in addition to the use of distance measurement ([30]) and methods based on outlier analysis ([109] introduced in 5.6 ), the challenge of the data set can also be solved from the following two directions. Semantic similarity analysis: It is used to detect almost repeated news content. Due to the lack of relevant knowledge and imagination of rumor publishers, they often reuse existing rumor content ([57]). For example, a rumor reviewer only needs to modify some essential parts to modify a correct message to be published as a rumor, thereby misleading users. Therefore, semantic similarity analysis to detect tampered rumors can provide a suitable method and can be used for potential rumors detection. Unsupervised rumor embedding: Due to the textual nature of rumors, semantic similarity analysis, sentiment analysis, and other related tasks are essential components of rumor detection. Embedding is an essential step in natural language processing. It refers to the process of extracting high-dimensional representations of original text data. In rumor detection, high-dimensional representations can be used as input for further analysis. Different embedding technologies can capture the characteristics of data from different angles. Choosing a good embedding method plays an essential role in obtaining the underlying nature of news and successfully detecting false information on the Internet. Some popular unsupervised embedding techniques include Word2vec([71]), FastText([7]) and Doc2vec([55]).

Early rumor detection

The current research on rumor detection can only detect rumors when they have been created and have spread to a specific scale through the Internet. However, although some Fact-checking platforms can warn users that the information may contain misinformation, they cannot organize the spread of such rumors in the early stages of the spread of rumors in social media. Furthermore, this kind of problem also needs the attention of future researchers. Therefore, we believe that the early prediction of rumors as future work is significant for rumors detection. Because most of the current rumor detection work is trying to study the true and false detection of the information spread on a large scale, it is essential to detect any trends or potential rumors as early as possible. By learning from historical data, early detection of rumors hopes to find them when the rumors break out. However, there has been work ([15]) hoping to find the characteristics of rumors in social media in the early stage through the content of the article or the emotional characteristics ([26]). However, there is no guarantee that their generalization capabilities can cope with early rumors detection for newly emerging rumors or new areas of time. Moreover, we believe that early rumors can be detected with more effort, such as potential rumors subject analysis, use of rumor publishers, analysis of data sources, characteristic transfer of different rumors, Etc.

Rumor intervention and refutation mechanism

Although rumor prediction can remind users of any potential false information during the existence of fake news, the rumor has already caused a certain degree of panic in the process of spreading. However, as far as the current research is concerned, there is still a lack of research on the intervention of rumors and the mechanism of dispelling rumors. ([25]) The influence of rumors in social media has been reduced by constructing a network model for the spread of rumors, and some work ([85]) mentioned some other methods to suppress the spread of rumors, such as deleting the accounts of some high-risk users or making it easier for some people to be vulnerable. Confused users provide immune space. Therefore, the work related to the intervention of rumors is one of the potential research directions to reduce the mass panic and social harm of users caused by the spread of rumors. In addition, the research on the rumor refutation mechanism can understand the mechanism of rumor refutation and rumor intervention ,then disseminate accurate information through a wide range of crucial dissemination nodes. Thereby further reducing the harm of rumors. [121] established a dissemination model to study the mechanism of dispelling rumors through the dissemination dynamics of 8 kinds of rumors. [39] established a rumor rejection mechanism model based on the principles of game theory. Therefore, the rumor rejection mechanism should also be combined with the intervention of rumors as one of the research directions to reduce the harm of rumors.

Conclusion

Rumors and fake news have become by-products of the digital communication ecosystem, and facts have proven very dangerous. Rumor detection work can analyze, characterize, compare and comprehensively evaluate the current scene and classify rumors for the problem. This article attempts to introduce a summary of the latest work in rumor detection. Focus on three aspects: feature extraction, model structure, and methodology. In feature extraction, about 80% of the research uses machine learning and deep learning to implicitly and explicitly extract rumor features. In the model structure, this article classifies the latest research based on popular structures. The latest work is classified, compared, and introduced according to their respective research directions in the methodology. Finally, the deficiencies of existing research are proposed in the future research for potential research directions for researchers’ reference. To some extent, this work helps new researchers understand the latest developments in the direction of rumor detection and helps newcomers adapt to the field more quickly and sort out their ideas. Fields such as real-time visualization and learning, unsuprevised learning, and early detection of rumors are challenging work that have not yet been resolved and require further research. There is still a long way to improve the accuracy of rumor detection and apply the research content. Overcoming the above challenges will provide further technical support for rumor management and network security. Table 8 shows the list of abbreviations for this paper.
Table 8

The list of abbreviations

AbbreviationFull Name
RQResearch Questions
DLDeep Learning
TText
SSocial
VVisual
PSPropagation Structure
Acc.Accuracy
Pre.Precision
URUnverified rumor
TRTrue rumor
FRFalse rumor
NRNonrumor
CNNConvolutional Neural Networks
TextCNNText Convolutional Neural Networks
VGGVisual Geometry Group
RNNRecurrent Neural Networks
GRUGated Recurrent Unit
Bi-GRUBi-directional Gated Recurrent Unit
LSTMLong Short-Term Memory
Bi-LSTMBi-directional Long Short-Term Memory
GNNGraph Neural Networks
GCNGraph Convolutional Networks
Bi-GCNBi-directional Graph Convolutional Networks
BERTBi-directional Encoder Representation from Transformers
TF-IDFTerm Frequency–Inverse Document Frequency
LDALatent Dirichlet Allocation
The list of abbreviations
  10 in total

1.  #FluxFlow: Visual Analysis of Anomalous Information Spreading on Social Media.

Authors:  Jian Zhao; Nan Cao; Zhen Wen; Yale Song; Yu-Ru Lin; Christopher Collins
Journal:  IEEE Trans Vis Comput Graph       Date:  2014-12       Impact factor: 4.579

2.  Big data. The parable of Google Flu: traps in big data analysis.

Authors:  David Lazer; Ryan Kennedy; Gary King; Alessandro Vespignani
Journal:  Science       Date:  2014-03-14       Impact factor: 47.728

3.  node2vec: Scalable Feature Learning for Networks.

Authors:  Aditya Grover; Jure Leskovec
Journal:  KDD       Date:  2016-08

4.  The science of fake news.

Authors:  David M J Lazer; Matthew A Baum; Yochai Benkler; Adam J Berinsky; Kelly M Greenhill; Filippo Menczer; Miriam J Metzger; Brendan Nyhan; Gordon Pennycook; David Rothschild; Michael Schudson; Steven A Sloman; Cass R Sunstein; Emily A Thorson; Duncan J Watts; Jonathan L Zittrain
Journal:  Science       Date:  2018-03-08       Impact factor: 47.728

5.  The spread of true and false news online.

Authors:  Soroush Vosoughi; Deb Roy; Sinan Aral
Journal:  Science       Date:  2018-03-09       Impact factor: 47.728

6.  FakeNewsNet: A Data Repository with News Content, Social Context, and Spatiotemporal Information for Studying Fake News on Social Media.

Authors:  Kai Shu; Deepak Mahudeswaran; Suhang Wang; Dongwon Lee; Huan Liu
Journal:  Big Data       Date:  2020-06       Impact factor: 2.128

7.  A Novel Supertwisting Zeroing Neural Network With Application to Mobile Robot Manipulators.

Authors:  Dechao Chen; Shuai Li; Qing Wu
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2021-04-02       Impact factor: 10.451

8.  A New Approach to Descriptors Generation for Image Retrieval by Analyzing Activations of Deep Neural Network Layers.

Authors:  Pawel Staszewski; Maciej Jaworski; Jinde Cao; Leszek Rutkowski
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2021-06-10       Impact factor: 14.255

9.  Towards an Ethical Framework for Publishing Twitter Data in Social Research: Taking into Account Users' Views, Online Context and Algorithmic Estimation.

Authors:  Matthew L Williams; Pete Burnap; Luke Sloan
Journal:  Sociology       Date:  2017-05-26

10.  Rumor Propagation is Amplified by Echo Chambers in Social Media.

Authors:  Daejin Choi; Selin Chun; Hyunchul Oh; Jinyoung Han; Ted Taekyoung Kwon
Journal:  Sci Rep       Date:  2020-01-15       Impact factor: 4.379

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.