Literature DB >> 36034676

A survey of uncover misleading and cyberbullying on social media for public health.

Omar Darwish1, Yahya Tashtoush2, Amjad Bashayreh2, Alaa Alomar2, Shahed Alkhaza'leh2, Dirar Darweesh2.   

Abstract

Misleading health information is a critical phenomenon in our modern life due to advance in technology. In fact, social media facilitated the dissemination of information, and as a result, misinformation spread rapidly, cheaply, and successfully. Fake health information can have a significant effect on human behavior and attitudes. This survey presents the current works developed for misleading information detection (MLID) in health fields based on machine learning and deep learning techniques and introduces a detailed discussion of the main phases of the generic adopted approach for MLID. In addition, we highlight the benchmarking datasets and the most used metrics to evaluate the performance of MLID algorithms are discussed and finally, a deep investigation of the limitations and drawbacks of the current progressing technologies in various research directions is provided to help the researchers to use the most proper methods in this emerging task of MLID.
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022, Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Entities:  

Keywords:  BERT; COVID-19; Deep learning; Disinformation; Machine learning; Misinformative; Misleading information

Year:  2022        PMID: 36034676      PMCID: PMC9396598          DOI: 10.1007/s10586-022-03706-z

Source DB:  PubMed          Journal:  Cluster Comput        ISSN: 1386-7857            Impact factor:   2.303


Introduction

Misleading information is a critical problem that affects controversial topics such as human health. Misinformation lacks scientific evidence since the information exchange is primarily happened by a clash of disparate and emotionally charged narratives. Social media networks such as Facebook, Twitter, and Instagram are widely regarded as one of the most essential ways for people to communicate [1]. In the last decade, social media has evolved into a significant instrument for gathering information and developing solutions in a variety of disciplines, including business, entertainment, and crisis management in health care, research, and politics [2]. Social media is regarded as one of the most important sources of information for health monitoring [3, 4]. Open source to the health information on the Internet and other easy-to-access online databases has made up new opportunities for medical professionals, but most of the information is susceptible to manipulation because contextual conventions and the veracity of provenance are constantly questioned. False health information is based on opinions, mythologies, and oral histories. Any post, tweet, or shared resource that misinterprets the knowledge of medicine accepted by experts is considered misleading health information. This contains fake news articles, memes, and posts about things that are false. The problem of obtaining health information from social media is determining what is reliable. That was one of the main topics of “Freedom of Information,” a BioMed Central-sponsored conference held on the 6th and 7th of July at New York’s Academy of Medicine. One of the attendees mentioned a widely publicized study published in the professional journal Cancer. J Sybil Biermann and her colleagues at the University of Michigan discovered that one webpage documented the mortality rate for a particular type of bone cancer as 5%, when in fact it was closer to 75%. According to the attendee, such misinformation could be disastrous. The term “misleading Information” is defined in the literature as fake news, misinformation, disinformation, false news, or non-informative news which is published to mislead the users and make them realize that it is truthful and reliable. According to the Federal Trade Commission, more than 25 million individuals will use the Internet to search for health information. The number of medically related websites on the Internet is estimated to be at least 100,000 doctors only review about half of these sites’ content [5]. There are two main types of misleading information on social media: “Disinformation,” which is purposefully spread by nefarious people. The other type is “Misinformation,” which spreads unintentionally but is still false. Also if we look at the psychological side of the problem, we can say that the so-called online cyberbullying, abusive language and hate speech is one of the most important aspects of the problem. According to the above mentioned, the urgency of detecting this type of information was first felt by researchers. Therefore, they started using various types of machine learning (ML) and deep learning (DL) algorithms to help people distinguish between informative and non-informative information by collecting datasets from public posts on social media sites, first and foremost from Twitter, and conducting their scientific experiences on them. The main contributions of this comprehensive study are: The paper is organized as follows: Sect. 2 introduces the research scope and some statistics for the previous works, Sect. 3 describes the standard approach of MLID, and Sect. 4 presents and summarizes the latest methods applied in MLID, Sect. 5 discusses the benchmarking datasets used during MLID, Sect. 6 shows the most used metrics to evaluate the performance of MLID algorithms, and Sect. 7 highlight the main limitations and drawbacks in MLID. Section 8 concludes this research study. Presenting some statistics for the previous work that demonstrate the interest of researchers in the MLID field and provide recommendations for future works. Introducing a detailed discussion of the main phases of this framework to show the generic adopted approach that is used by the plurality of proposed MLID techniques. Presenting the main ML and DL methods utilized in the literature to review the recent MLID approaches. Discussing the benchmarking datasets and the most used metrics to evaluate the performance of MLID algorithms. Providing a deep investigation of the limitations and drawbacks of the current progressing technologies in various research directions to help the researchers to use the best methods in this emerging task of MLID.

Research scope

Social media websites permeate our culture by evolving into ideal platforms for information access. As a result, these platforms have an important role in persuading people’s decisions and opinions. Everywhere now, most people spend hours on social websites to communicate with the rest of the world. Also, they use social platforms to read news and gather information rather than conventional media because information dissemination on social networks takes a shorter time and lower cost. Misleading news is a major problem and a global issue since it has become a daily expression and an essential component of any media discourse over the last 4 years. It has huge popularity among stakeholders such as non-governmental organizations, journalists, politicians, civil society and researchers. Furthermore, misinformation has significant repercussions, such as influencing people’s decisions, opinions, and attitudes. As a result, the detection of fake information is critical. Fake information is typically associated with time-sensitive and evolving events that cannot be accurately verified by previous studies. On the flip side, data generated by misinformation are noisy, unorganized, and incomplete [6]. There are a variety of human-based websites that introduce pre-checked data for misleading information detection (MLID). These data are analyzed manually by expert analysts knowledgeable about this research topic. The most popular websites are Snopes.com, PolitiFact.com, and Factcheck.org. Due to the huge amount of data on social media, the human approach becomes lifeless, costly, subjective, and unworkable [7-9]. On the other hand, the technical approach classifies the data automatically using numeric machines that make the process easier, less expensive, and more efficient. The Internet is a massive repository of information. With the advent of technology, every user has become a self-publisher, with no proofreading, no fact-checking, and no accountability. People have the freedom to publish anything they want, when they want, and wherever they want. Due to the lack of verifiability, users can spread their opinions across social media platforms. These opinions represent many types of misleading information that have been listed in Table 1 [10].
Table 1

Types of misleading information

TypeDescription
Fabricated contentTotally fake content
Manipulated contentReal information or imagery is distorted. For example, a headline is created to provoke public interest, which is often promoted by ‘clickbait’
Imposter contentImpersonation of authentic sources, such as using the branding of a well-known news organization
Misleading contentMisleading information use, such as presenting a statement as a reality
False context of connectionContent that is factually correct but is attended by fake contextual information, such as when an article’s headline does not accurately reflect the content, such as when an article’s headline does not accurately reflect the content
Satire and parodyPresenting amusing but fake information as if it were real even though not typically classified as disinformation, this may unintentionally mislead readers
Types of misleading information We generated a list of the most recent research studies on the misleading information task from (2015 to 2022) as shown in Fig. 1. We extracted the results from the Association for Computing Machinery (ACM) explorer using “misleading information” “fake news” “misinformation” and “disinformation” keywords.
Fig. 1

A representation of recent studies on misleading information topic from 2015 to 2022

A representation of recent studies on misleading information topic from 2015 to 2022 Also, we evaluated the publishers’ interest in this research topic and got the top three names of journals and conferences that are famous in this field. Concerning journals and magazines, the rapid evaluation of misleading information topics is highlighted in Proceedings of the ACM on Human–Computer Interaction with 90 articles, Communications of the ACM with 69 articles, and ACM Computing Surveys with 26 articles on this topic. As for Proceedings (Conferences), Proceedings of the ACM Web Conference 2022 published 28 papers, Companion Proceedings of The Web Conference 2018 had 26 papers, and Companion Proceedings of The 2019 World Wide Web Conference published 26 papers. Our research study differs from previous surveys in that it addresses a comparison between the latest studies of MLID and provides a comprehensive overview of previous developments and algorithms that will introduce a deep understanding to inspire the researchers to use the appropriate methods that would improve their contributions in this field. It concentrates on the MLID approach’s ML and DL techniques, architectures, and models. In addition, benchmarking datasets and evaluation metrics are discussed. Many limitations and thought-provoking research directions are presented in detail. This research study will motivate the researchers to add more improvements to the task of MLID.

The MLID approach

This section demonstrates the MLID phases, beginning with the stages through which all training data should pass before being used to create misleading-information detection models. The following stages make it easier to handle the large amount of data required to build a detection model. As shown in Fig. 1, the generic methodology is primarily based on ML and DL models, which are widely used to learn distinguishing characteristics of misleading information. As seen in the pipeline, several critical steps are typically implemented in the development of the final detection system. The summary of MLID approaches are also explained in Table 2.
Table 2

A summary of MLID approaches

ReferencesModelMethodSpecificationsDatasetSize (text)Topic
[82]SVM, KNN, RF, NB, DistilBERT, DistilRoBERTaCost-weighting settingsNVIDIA Tesla V100S 32 GB GPU, 240 GB of storage, 32 GB of RAM, an Intel(R) Core(TM) i7-9750H CPUCLEF 2018 Consumer Health Search5,535,120Public Health
[83]KG-Miner TransE, text-CNN, CSI, dEFEND, GUpdater, HGAT, DETERRENTKnowledge guided graph attention networkAn attention mechanism, the knowledge guided article embeddingsDiabetes Cancer2269 6099Public Health
[92]RFRecursive feature selectionLinguistic features, LIWC featureCollected dataset2225Public Health
[95]MLP, NN, SVM, RT, MAda and RFThe graph theory and the social influence modelsIdentifies two structural levels (named as user level and network level in the following) that lead to the extraction of representative featureCollected dataset709Public Health
[96]GB, LR, NB, RF, Bi-LSTM, CNNL2 regularizationTextual representation, linguistic–stylistic, linguistic emotional, linguistic–medical, propagation-network, and user-profile featuresCoAID, ReCOVery, FakeHealth (release) FakeHealth (Story)3555 2029 606 1690Public Health
[100]DT, kNN, MNB, NN, BNB, LSVM, LR, ERF, XGBoostVoting ensembleFeature Engineering, TF-IDF, N-gramCollected dataset7486COVID-19
[101]SBERT, BiLSTM, SBERT (DA)BiLSTM and SBERT are jointly trained with the linear classifierBi-encoder, forward/backward LSTM, concatenate flattenCOVIDLIES, SNLI, MultiNLI MedNLI6761 570,000 433,000 14,049COVID-19
[106]Random Forest, DTTwo RF and combination of multiple decision treesTF-IDF, Count vectorizer, Bag of wordCOVID19FN2800COVID-19
[109]DT, kNN, SVM, RF, NB, LSTM, LR, GRUGrid search with cross validation, Keras_tunerTF-IDF, N-gram, word embeddingCoAID PolitiFact Disasters Gossip cop926 1050 7613 10,650COVID-19
[108]SVM, PAC, MLP, LSTM with FastText, CNN with FastText, LSTM + CNN, BiLSTM + Attention, Ensemble Model (BERT, ALBERT, XLNet)Transformer model using HuggingFace libraryEnglish Glove word embeddings, PyTorch transformer library, Transformer-XLCollected dataset10,700COVID-19
[112]BERTCluster analysisSpecial character removal, stemming and lemmatization, TF-IDFCollected dataset6731COVID-19
[116]kNN–BSSA, kNN–BGA with feature selection, kNN–BPSO, and kNNWrapper feature selection methodsReduce the number of features, TF, TF-IDF, and bag-of-wordsKoirala3002COVID-19
[118]Hybrid model (CNN and LSTM)Hybrid modelTF-IDF, word embedding and hyperparameter optimizationDataset1 Dataset2 Dataset31100 10,202 3001COVID-19
[119]DT (C4.5), RF, NB, SVM, kNN, Bayes Net + kNNStacking methodData annotation and feature extractioncollected dataset409,484COVID-19
[120]LSTM, BiLSTM, CNN, hybrid of LSTM-CNNGloVe pre-trained word embedding featuresPython regular expressions and NLTKCOVID-19 Fake News21,379COVID-19
[122]Attention-based BiLSTM-CRF modelConditional Random Field (CRF)PubMed pretrained embeddingsCollected dataset20,137Cancer
[124]SVMLinearSVC class based on the LibLinear libraryUser engagement featuresCollected dataset250Cancer
[125]LSTMGloVe, Word2Vec (CBOW and skip-gram), FastText (CBOW and skip-gram)Word embeddingsCollected dataset140,000Influenza
[129]SVM, RF, RUSBoost, XGBoost, CNNSequence Alignment-free MethodsWord encoding, word embeddingsGI-SAID60,087Influenza
[131]DNN, SVM, J48, NBCFS reduction methodWEKA, Sklearn libraryCollected dataset7000Heart Attack
[133]RF, NB, J48, NNSession-based modelUser-based features, text-based features, and network-based featuresCollected dataset1.6MCyberbullying (Psychological Health)
[134]RF, NB, J48SentiStrength, Indico APIBig Five, Dark Triad models, psychological features such as personalities, sentiment, and emotionCollected dataset9484Cyberbullying (Psychological Health)
[135]RF + Big Five and Dark Triad ModelsEnsemble techniqueBig Five, Dark Triad models to determine user personalityCollected dataset9484Cyberbullying (Psychological Health)
[138]LR, SVMSVM with two new hypothesis for feature extractionN-gram, Counting, TF-IDF Score and hypotheses for feature extraction (Capturing pronouns, Skip-grams)Collected dataset6547Cyberbullying (Psychological Health)
[139]kNN, SVM, NB, DT, RFSMOTE techniqueNetwork-Based Features, Activity Features, User Features, Content-Based Features, Personality Features, Pointwise Mutual Information-Semantic Orientation (PMI-SO)Collected dataset14,495Cyberbullying (Psychological Health)
[141]NB, LibSVM, RF, kNNSMOTE techniqueNetwork features, Activity features, User features, Content features, Pearson correlation, chi-square test, and information gainCollected data set10,007Cyberbullying (Psychological Health)
[142]CapsNet–ConvNetHybrid deep technique of Capsule network (CapsNet) and convolution neural network (ConvNet)Google Lens of Google Photos AppCollected dataset10,000Cyberbullying (Psychological Health)
[143]NBNB Lexicon Based featuresBag of Word features and Lexicon Based featuresCollected dataset350Cyberbullying (Psychological Health)
[144]BoW, sBoW, LSA, LDA, EBoWBoW and latent semantic featuresBoW features, latent semantic features and bullying featuresTwitter dataset1762Cyberbullying (Psychological Health)
[149]Bi-LSTM , CNN, Bi-LSTM with attention, CNN–LSTM combinedAutomatically identifying abusive language on ArabicFor Bayesian hyperparameter optimization, the Structured Parzen Estimator (TPE) algorithm is usedCollected dataset15,050Cyberbullying (Psychological Health)
[151]SVM, NBWEKATweet To Senti Strength Feature Vector filterArabic dataset, English dataset35,273 91,431Cyberbullying (Psychological Health)
[154]NB, LR, SVM, XGBoost, CNN, LSTM, BLSTM and GRUYoutube APITF-IDF, Word Embedding, Sklearn, Tensorflow, NLTK, matplotlibDataset1 Dataset2 Dataset35000 7000 12,000Cyberbullying (Psychological Health)
[156]SVM, Logistic Regression, NB, Random ForestEnsembleLabel EncoderDataset1, OLID, Dataset3, Dataset41990 14,100 8817 24,784Cyberbullying (Psychological Health)
[157]GBDT, Random Forest, SVM, XGB_CTDFuzzy C-Means (FCM)Scikit-learn, Keras, FCM and XGBoost Libraries, Intel I7-8500H 3.60 GHz CPU and a laptop with a 12 GB RAMCollected dataset542Cyberbullying (Psychological Health)
[158]FFNN4 Hidden layersHot encodingDataset1 Dataset24913 34,890Cyberbullying (Psychological Health)
[159]NBPredefined listBoWTraining Testing1,600,000 359Cyberbullying (Psychological Health)
[161]SVM, CNN-CBTwitter streaming APISpyder environment, 12 GB of RAMCollected dataset39,000Cyberbullying (Psychological Health)
[162]CNN, LRCNSkip-gramfour NVIDIA GTX 1080 serversCollected dataset8815Cyberbullying (Psychological Health)
[163]SVM, DT (C4.5), NB, and kNNInformation gain and chi-squareTokenizationCollected dataset900Cyberbullying (Psychological Health)
[164]CNN and PCNNTM (threshold moving), CFA (cost function adjusting), and a hybrid solution (TM CFA)Removing non-alphanumeric characters tokensDataset1 Dataset 21313 13,000Cyberbullying (Psychological Health)
[166]SVMn-gramTokenization, NormalizationCollected dataset15,050Cyberbullying (Psychological Health)
A summary of MLID approaches

Data preprocessing and feature engineering

Preprocessing

Text preprocessing is a technique for cleaning up text data to feed it into a model. Text data includes different types of noise such as emotions, emojis, punctuation, characters, and text in different scenarios. Hence, there are numerous ways to express the same concept, and this is only the most serious problem we face because machines don’t understand words; they require numbers, so we must encode and transform the text into numbers to pass it to all subsequent processing stages [11]. This stage is critical for reducing the indexing (or data) file size of the text documents and improving the IR system’s quality and productivity [12]. The most used techniques for data cleaning are lower case, expand contractions, remove punctuation, words, and digits. remove stop-words. Text rephrasing, lemmatization, and stemming, and white spaces removing. Contraction is the shortening of a word; for example, veggie stands for Vegetarian, and limo stands for Limousine. For better analysis, we must broaden this compression in the text data. Lower case is important because a machine can easily interpret the words when the text is in the same case since the machine treats the lower and upper cases differently. Another text processing technique is punctuation removal. There are a total of 32 main punctuations that must be addressed. The string module can be used with a regular expression to replace every punctuation in the text with an empty string. Some people tend to write characters and digits together to form words in the text, such as amjad12 or amjad37bash. This type of word must be eliminated because it causes a problem for machines to understand and process. As a result, the best solution is to remove this type of text or replace it with an empty string. Stop-words such as these are common words in a text that do not provide any helpful information. Therefore, the researchers use the NLTK library to remove these types of words. The process of minimizing a word to its root stem is known as stemming. There are numerous stemming algorithms, such as the porter stemmer and the snowball stemmer. Porter stemmer is a popular tool in the NLTK library. The stemming method is not used in production because it is inefficient, and most of the time it stems from unwanted words. As a result, another technique known as Lemmatization entered the market to solve the problem. Lemmatization is comparable to stemming. since it is used to stem words into root words, but it works differently. Lemmatization is a systematic method of decreasing words to their lemma by comparing them with a language glossary [12]. Also, the researchers tend to use word embedding as an essential part of the preprocessing step to represent the words that will be used in the text analysis process. Word embeddings are working to form real-valued vectors that encode the meaning of the words. As a result, the words that have similar meanings will be closer in the vector space [13]. Our suggested framework is presented on Fig. 2. Framework for detecting misleading information

Feature selection

Even though there are many current classifiers for text categorization, the large dimensionality of the feature space is a serious difficulty [14]. A document typically comprises hundreds or thousands of distinct words that are considered features; however, many of them may be noisy, less useful, or redundant in terms of class labels. This may cause classifiers to be misled and, as a result, decrease their overall performance [15, 16]. As a result, feature selection must be utilized to eliminate noisy, less useful, and redundant features, reducing the feature space to a manageable level and boosting the efficiency and accuracy of the classifiers used. A feature selection approach, in general, consists of four basic steps: feature subset generation, subset evaluation, halting condition, and classification result confirmation [17]. The researchers employ a search approach to select a candidate feature subset in the first phase, which is then evaluated using a goodness criterion in the second step. When stopping requirements are met in the third phase, subset formation and evaluation are terminated, and the best feature subset from all candidates is picked. The feature subset will be validated using a validation set in the last stage. Feature selection approaches are classified into four types based on how they create feature subsets: filter model [18-20], wrapper model [21, 22], embedding model, and hybrid model [23, 24]. Most feature selection approaches for text classification are filter-based, owing to their simplicity and efficiency. [15, 16, 25, 26] provide a detailed analysis and comparison of alternative feature selection techniques for generic data [27].

Feature extraction

In-text categorization, there are two main types of feature extraction methods: n-gram and termset. n-Gram the n-gram extraction procedure requires sliding across an entire corpus with a window of length n [28-31]. Then, in each window, all the sets of consecutive words or characters should be extracted. The goal of the n-gram is to obtain the composite features that emerge repeatedly to reduce the ambiguity of individual words. Bigram and trigram are two often used n-grams. Nonetheless, the impact of text structure, such as punctuation and stop-words, is not considered. Termset a termset is not the same as an n-gram in that composite characteristics are retrieved solely based on their co-occurrence, regardless of the order and position of the individual terms [28, 32, 33]. Specifically, termsets are randomly matched combinations of vocabulary. However, this combination has one drawback: a combination explosion, even for 2-termsets. It indicates that with a vocabulary size of n, there will be 2n different types of pairings.

Data splitting

When starting a modeling project, one of the first decisions to make is how to use the available data. One popular method is to divide the data into training set and testing set. Models and feature sets are created using the training set. The training set serves as the foundation for parameter estimation, model comparison, and all other operations required to arrive at a final version of the model. At the end of these operations, the testing set is used to estimate a final, unbiased evaluation in terms of the model’s effectiveness. Looking at the results of the test sets would skew the results because the testing data would have become part of the outcome. How much data should be kept for testing purposes? It is hard to create a consistent guideline. Many factors can influence data proportion, including the length of the initial sample pool and the actual amount of predictors. The importance of this decision is reduced with a large pool of samples once the training set contains sufficient samples. Alternative options to a simple original data partition may also be a smart idea in this situation [34].

Model selection and building

Plenty of well methods have been proposed to detect misleading information. The researcher used DL methods such as BERT [35], SBERT [36], ALBERT [37], LSTM [38], BI-LSTM [39], BSSA [40], SSA [41] , CNN [42], and ML models such as SVM [43], DT [44], RF [45], NB [46], kNN [47], XGBoost [48], and GA [49].

BERT

The BERT framework consists of two steps: pre-training and fine-tuning. On unlabeled data, the model is trained across many pre-training activities during the pre-training phase. To fine-tune the BERT model, begins with the pre-trained parameters. The parameters are then fine-tuned utilizing labeled data from down-stream jobs. Despite the fact that they are all started with identical pre-trained parameters, each downstream job has its own fine-tuned model. BERT’s model structure as described in the original implementation in [50] is a bidirectional multi-layer transformer encoder available in the tensor2tensor library. The BERT was trained on two activities: language modeling (15% masked tokens predicted from the context by trained BERT) and the following statement prediction (trained BERT was trained used to predict whether a selected following statement was possible or not depending on the first statement). BERT learns word contextual embeddings as a result of the training procedure. Following the computationally expensive pretraining, BERT can be finetuned with fewer dataset resources with small sizes to enhance its performance on detailed tasks [35, 51].

CNN

A convolutional neural network (CNN) is used for image processing [52] but in other circumstances, researchers prefer to work with TextCNN for text classification, which necessitates the addition of a word embedding layer and a one-dimensional convolutional network to the model’s original structure [53]. CNN is made up of three layers: an input layer, hidden layers, and an output layer. A feed-forward neural network’s intermediary layers are referred to as hidden because their model parameters (model’s input and model’s outcomes) are veiled by the final convolution and activation function. The hidden layers of a CNN include convolutional layers. Generally, this comprises a layer that does a dot product of the convolution kernel and the input matrix of the layer. This is typically the inner product of Frobenius and its activation function is typically ReLU. As the convolution kernel slipped along the input matrix for the layer, a feature map is generated, which then plays an important role to the following layer’s input. Following that, additional layers such as fully connected layers, normalization layers, and pooling layers are added. [54].

RNN

A recurrent neural network (RNN) connects the nodes to create a graph that follows a time sequence. As a result, it can show dynamic behavior. RNNs, which are derived from feedforward neural networks, can use their internal memory to perform variable-length input sequences [55-57]. Consequently, they can be used for tasks like unsegmented, handwriting recognition [58], or speech recognition [59, 60]. RNNs are turning complete in theory and may execute arbitrary programs to perform arbitrary input data [61]. Fully RNNs (FRNN) connect all neurons’ outputs to all neurons’ inputs. Because all other topologies may be replicated by reducing the weights of some connections to zero to imitate the lack of connections between particular neurons, this is the most general neural network architecture. Many people may be misled by the figure on the right because practical neural network topologies are typically organized in “layers,” and the illustration conveys that impression. What looks to be layered, however, are various time steps of the same FRNN. The arc labeled ‘v’ on the left-hand side of the illustration depicts the recurring connections. It “unfolds” in time to provide the illusion of layers [62].

LSTM

LSTM is an abbreviation for long-short term memory. In terms of memory, LSTM is a sort of RNN that outperforms traditional RNNs. LSTMs perform far better when it comes to learning specific patterns. LSTM, like any other NN, can have several hidden layers, and as it travels through each layer, the relevant information is retained while the irrelevant information is discarded in each cell [63]. A cell, an input gate, an output gate, and a forget gate comprise the LSTM model. The cell retains values over arbitrary time intervals, and the three gates control data flow in or out of the cell [64].

GRU

GRU is an abbreviation for gated recurrent unit, which is a method of gating in RNNs proposed by [65]. The GRU functions similarly to a LSTM with a forget gate [66], but with fewer parameters because It is missing an output gate. [67]. GRU outperformed LSTM on specific tasks such as natural language processing (NLP), polyphonic music modeling, and speech signal modeling [68, 69]. GRUs have been found to outperform other methods on unduplicated datasets of small size [70, 71].

Machine learning models

Decision tree (DT)

A DT is a data flow structure that contains nodes that represent the test attributes. Each branch in the DT reflects the result of the test, and each leaf node provides a class label (decision reached after calculating all attributes). Classification rules are defined by the pathways from root to leaf. As a visual and analytical decision assistance tool, a DT and its closely related effect diagram are used in decision analysis, where competing alternatives’ expected values are determined [72].

Random forest (RF)

Random decision forests are another name for RFs that are well-suited for coping with high-dimensional noisy data in text classification [73]. RF is an ensemble learning method that can be used for classification, regression, and other issues that work by generating a plethora of DTs during training. For classification problems, the class selected by the most of trees is the random forest output. For regression problems, the mean or average prediction of each tree is returned [74, 75].

Naive Bayes (NB)

NB is a classification algorithm based on Bayes’ Theorem and the assumption of predictor independence. A NB classifier, in simple terms, assumes that the presence of one feature in a class is independent of the values of the other features. NB has three types: Gaussian, Multinomial, and Bernoulli [76].

Support vector machine (SVM)

SVM is a supervised learning system with an accompanying algorithm that is used for classification issues. Each data object is represented graphically in n-dimensional space in this procedure, with the value and coordinate determining the item’s value. Classification is accomplished by determining the hyper-plane that best divides the categories. SVMs may conduct non-linear classification by converting inputs straight into high-dimensional feature spaces [77-79].

K-nearest neighbor (kNN)

kNN is among the most basic and extensively used algorithms that is based on its k value. The k value is used to specify how many neighbors in the algorithm. This technique works by selecting the number k value as the first step. Then it uses Euclidean distance to determine the kNN of a new data point. Following that, it counts how many data points there are in each category among the K-neighbors before assigning the new data point to the category with the most neighbors [80, 81].

Literature review

With the advancement of technology and social media addiction, avoiding misleading information becomes an essential part of our daily life due to the easy dissemination of any information especially when it comes to the health fields. This section presents the latest works proposed for MLID regarding public health fields such as COVID-19, cancer disease, influenza, heart attacks, and psychological heath (cyberbullying). We selected these research papers because they were of the best research in the MLID field from different aspects such as high results, benchmarking datasets usage, obtaining various types of algorithms, and well documentation.

Public health

Fernández-Pichel et al. [82], provided a comprehensive comparison of recent deep NLP models, such as newer BERT-based models (DistilBERT, DistilRoBERTa), and traditional algorithms, such as SVM, RF, NB, and kNN, for detecting health-related misinformation, and identifying low-quality online content (web pages that are unreliable and hard to read). The CLEF 2018 Consumer Health Search task dataset was chosen by the authors. To compare the mentioned models, the researchers examined them in terms of trustworthiness, readability, usefulness (both trustworthiness and readability), and the effect of the training set size. Yet, SVM, RF, and DistilRoBERTa achieved the highest F1-score (93%). Cui et al. [83], proposed the DETERRENT approach, which uses a relational graph attention network to represent various positive and negative relationships in the medical knowledge graph. The authors manually construct two datasets, diabetes and cancer, using a public medical knowledge network KnowLife3 [84], which contains 25,334 entity names and 591,171 triples. The authors collect six positive and four negative relationships. Also, the authors compared DETERRENT with KG-cutting-edge Miner’s misinformation detection algorithms [85], TransE [86], text-CNN [87], CSI [88], dEFEND [89], HGAT [90], GUpdater [91]. After the comparison, DETERRENT got the highest score in the cancer dataset with an accuracy of 0.9652, precision of 0.9469, recall of 0.9153 and F1-score of 0.9309. Kinsora et al. [92], created a medical misinformation-labeled dataset to build ML classifiers for detecting false medical information automatically. This dataset was derived from a dataset created by Vydiswaran et al. [93]. The dataset is a combination of disinformation and correct facts collected from online health forum comments consisting of 2225 comments from MedHelp were labeled as misinformative or non-misinformative. The authors accomplished this by employing information retrieval techniques such as linguistic inquiry and word count (LIWC) psycho-linguistic lexicon feature [94] to obtain a wide variety of natural language features. They later devised a coding technique for labeling and annotating the dataset. The scientists created a classifier that can recognize medical misinformation with an accuracy of 90.1% by using nine designed characteristics in the produced dataset. Sicilia et al. [95], provided a method for detecting rumor in each post in a specific topic area connected to health news. The authors introduced new descriptors influenced by graph theory and social influence models, such as the likelihood of a tweet being retweeted or a URL being shared, conversation size, a fraction of users’ followers of a root, and a fraction of tweets with URLs. The authors gathered the dataset from Twitter using the hashtags #zikavirus and #zikamicrocephaly, which this was one of the major healthcare trends in 2016. On February 1st, 2016, the World Health Organization (WHO) classified Zika virus disease as a Public Health Emergency of International Concern (PHEIC). The authors used multiple classifiers including multi-layer perceptron (MLP), nearest neighbour (NN), SVM, random tree (RT), multiclass Adaboost (MAda), and RF as an ensemble of trees and obtained the following results: the overall accuracy is equal to 73.63%, the average precision per class is 72.80%, the average recall per class is 73.60% and the average AUC per class is 89.00%. Di Sotto and Viviani [96], introduced an analysis of unique characteristics groups and ML algorithms that can be useful in assessing misinformation in online health-related material, in the form of websites or social media content. The authors used three well-known datasets; CoAID (COVID-19 heAlthcare mIsinformation Dataset) [97], ReCOVery [98], and FakeHealth [99]. The authors identify six types of health misinformation textual representation features, linguistic–stylistic features, linguistic emotional features, linguistic–medical features, propagation-network features, and user-profile features. In particular, several ML and DL algorithms have been considered. The results show that DL solutions are efficient when word embedding selected features from appropriate medical vocabulary training are used, without the need for other types of features. When “classical” ML classifiers are used, however, the impact of considering other types of features grows. The best results in the CoAID are obtained by CNN. AUC equals 0.973, and F-measure of 0.953.

COVID-19 misinformation

With the COVID-19 pandemic, Elhadad et al. [100] developed a model for misleading-information detection in the English language. The authors used Google Fact Check Tools API to gather public ground truth from various fact-checking websites. The used dataset is collected from WHO, UNICEF, and UN websites from (February 4 to March 10) 2020. The NN classifier provides the best ACC, ERR, and AUC evaluations, with values ranging from 93.75% to 99.68%, 0.32% to 6.25%, and 89.46% to 99.47%, respectively. Hossain et al. [101] evaluated the performance of COVID-19 misinformation on social media by working on a dataset of 4.8K expert annotated social media posts that contain general misconceptions and their misinformative and informative expressions on Twitter. The dataset contains 86 misconceptions classified into three labels: 465 Misinformative, 164 Informative, and 4161 Irrelevant. The authors performed their analysis on Misinformative and Informative classes and achieved the best results using Avg.Glove Embedding. The authors updated their research and achieved remarkable results compared to the previous one by training the used linear classifiers on three datasets: SNLI [102], MultiNLI [103], and MedNL [104] and applying SBERT (DA) and BiLSTM models on COVIDLIES dataset [105]. Amer and Siddiqui [106] detected COVID-19 Fake News using COVID19FN dataset [107]. The authors applied RF and DT algorithms and got an accuracy score equal to 94.49%. Gundapu and Mamidi [108] presented a framework for analyzing the credibility of information disseminated on social media about the COVID-19 pandemic in this research. The best strategy for identifying false news is based on an ensemble of three transformer models (BERT, ALBERT, and XLNET). This model was developed and tested as part of the ConstraintAI 2021 shared problem “COVID-19 Fake News Detection in English.” The author compared a group of ML models and a group of DL models with transformer models based on the ensemble method. On the test set, the proposed model received a 0.9855 F1-score and finished fifth out of 160 teams. This infodemic has made it more difficult to access and identify trustworthy information, as well as rumors spread more quickly, putting public health at risk by making effective preventative tactics impossible to implement. The authors in [109] employed a new system for COVID-19 Fake News Detection on social media by applying ML and modified deep neural network (DNN) methods. In this study, four datasets were used: CoAID [97], The disaster dataset [110], The PolitiFact dataset [111], The gossip cop dataset [111]. The authors achieved accuracy equals 98.57%. Ng and Carley [112] worked on a study that looks at a corpus of coronavirus-related fact checks gathered from the three main fact-checking organizations (PolitiFact [113], Poynter [114], and Snopes [115]). The authors collected 6731 fact-checked stories, categorized these stories into six clusters, analyzed the temporal trends of story validity and the level of agreement across sites, created a novel pipeline for categorizing tales into more detailed narrative kinds, and applied it to a corpus of COVID-related misleading tweets. The authors utilized the BoW classifier and the BERT model. The BERT model performed well, attaining an average accuracy of 87% in the supervised classification of story validity, accurately detecting an average of 59% and 43% of the tales, respectively. Al-Ahmad et al. [116], suggested a powerful approach for detecting misleading news. This approach employs kNN–BSSA, kNN–BPSO, kNN–BGA with feature selection, and the standard kNN models to reduce the number of symmetrical features. The authors combined three wrapper feature selection methods: Particle Swarm Optimization (PSO), Genetic Algorithm (GA), and Salp Swarm Algorithm (SSA) to get higher accuracy. Koirala dataset [117] was used to create another six datasets by applying various tokenizers with stemming strategies. Based on the prediction findings, the suggested model showed an accuracy of 75.43% and decreased the number of symmetrical characteristics to 303. Alouffi et al. [118] offer a hybrid DL model that detects COVID-19 fake news by combining a CNN with a LSTM. The authors used six ML and two DL models. The proposed model for experimental outcomes is tested using three COVID-19 fake news datasets. To validate the results, four matrices are used: accuracy, precision, recall, and F1-measure. Experimental results demonstrate model outperforms the used ML and DL models. As a result, the suggested method produces the best ACC, PRE, REC, and F1 score evaluations, with values ranging from 97.7 to 97.5%, 97.53%, and 97.7%. Al-Rakhami and Al-Amri [119], designed a novel framework for identifying disinformation on the Twitter network by combining six machine-learning algorithms with ensemble learning, and seven combinations of base models were considered: C4.5 + RF, SVM + RF, C4.5 + kNN, SVM + kNN, SVM + Bayes Net + kNN, C4.5 + Bayes Net + kNN, in the end, all the models combined. The authors used Twitter’s streaming application program interface (API) for tweets to acquire a massive dataset about the COVID-19 outbreak. Human annotators reviewed and labeled the dataset. After that, the author extracted essential COVID-19 traits and used them to develop their system to automatically assess the credibility of tweets. The SVM + RF enabled the ensemble model to produce better results for root relative squared error (RRSE), the area under the receiver operating characteristic (AUROC) curve, accuracy, and Kappa statistic with values ranging from 0.257%, 99.7%, 0.978%, and 97.5%, respectively. In our previous research [120], to automatically identify and classify misinformation content on social media platforms associated with the COVID-19 pandemic The dataset “COVID-19 Fake News” which includes 21,379 instances of actual and fake news for the COVID-19 pandemic and associated vaccines, was used to train and test these DNNs. In terms of accuracy, the CNN model significantly outperformed the other DNNs in terms of accuracy, scoring 94.2%.

Cancer disease misinformation

Bianchini et al. [121] introduced a comprehensive study on sixteen websites to define various types of falsehoods, evaluate the risk of encountering deceptive information, and examine the variations in expert and layperson assessments. The authors gathered a dataset using if MONITOR and divided it into three classes (correct, incorrect, and undefined). Bal et al. [122] discussed the spread of misinformation about diseases and health issues via social media, especially cancer, on Twitter. The authors provided a new dataset and used classifiers to distinguish between medically relevant and non-medically relevant tweets. The dataset was collected using the Twitter Streaming API1. The gathered tweets contain specific keywords in combination with cancer. Also, the authors proposed complex neural techniques to specify the objects/items/techniques suspected of causing cancer or contributing to healing. To distinguish between tweets. The authors obtained an F1 score of 0.7818 for “causes”, 0.8341 for “prevents”, and 0.5106 for “cures” using TF-IDF-weighted PubMed embeddings [123]. The best F1 score of 0.6846 was achieved using the attention-based BiLSRM CRF model. Hou et al. [124], worked on the automatic detection of misinformation on YouTube. The researchers introduced a new collection of 250 prostate cancer movies that had been manually annotated for inaccuracy. To study the use of linguistic, auditory, and user participation characteristics, the researchers built a new dataset of 250 prostate cancer-related videos and manually labeled them. SVM classifiers for misinformative video identification were trained using this dataset. The authors achieved 74% accuracy, 76.5% precision, and 73.2% recall. Dai et al. [99], created the first complete fake health news datasets. The repository’s extensive properties include news content, news reviews, and social contexts. The authors performed exploratory analysis on the datasets to assess their quality and identify their essential aspects. These datasets perform detection on relatively basic baselines because the goal is not to attain high performance but rather to test the quality of datasets for fake news identification and, offer reference findings for future system evaluations; conduct fake health news detection on the two datasets. using Random guess, Unigram, Unigram + NS, Unigram + Tags, SVM, Random Forest, CNN, Bi-GRU, and SAF, the best result SAF 0.760 Accuracy and getting highest F1-score and AUC using second datasets 0.802, 0.809 respectively.

Influenza misinformation

Jang et al. [125], suggested a plan for extracting training data that is used to represent the hidden features and enhance the performance by filtering and choosing only influenza-related phrases before prediction. The authors utilized the Pearson correlation coefficient, PCC [126] to order the retrieved keywords and the root-mean-square error, RMSE [127] to evaluate each model. The quantity of the collected data was around 761 MB, containing nearly 140,000 words. Using the LSTM model, the best PCC equals 89% and the best RMSE equals 90%. Brainard and Hunter [128], created an agent-based model that simulates independent but related circulating contagious diseases as well as the sharing of health advice (classified as useful or harmful). The scientists used three stages of modeling (1 = no misinformation disseminated, 2 = misinformation causing epidemics to worsen, and 3 = ways to limit the influence of disinformation). Xu and Wojtczak [129], used the GI-SAID [130] database to collect viral protein sequences obtained from avian, swine, and human samples. The authors applied five classifiers (SVM, RF, RUSBoost, XGBoost, and CNN) to a 5-gram model. The results reveal that the PSSM-based model achieved an MCC of approximately 95% and an F1 of around 96. The model achieved MCC of 96%, and F1 of 97%.

Heart disease misinformation

Karajeh et al [131], classified over 7000 heart attack tweets as informative and non-informative. The dataset consists of 11% informative tweets and 89% non-informative tweets. The tweets were classified using DNNs, SVM, DT (J48), and NB with k-cross-validation, k values of 2, 5, and 10. The dataset was divided into 66% training and 34% testing, and all algorithms were tested using the CFS method. DNN had the highest accuracy (95.2%) and the highest F1-score (73.6%) for both informative and non-informational. Based on heart failure fake news. O’Connor [132], presented a solid case for medical experts to strongly resist overstated treatments, unverified entities, unproven vaccinations, and nutraceuticals. so that may keep a calm yet resolute attitude to the advancement of new chances for the patients they serve.

Cyberbullying (psychological health)

Chatzakou et al. [133] introduced a scalable system for identifying bullying on Twitter and provided a methodology for gathering text, user, and network-based information, investigating the characteristics of bullies and aggressors, and what qualities distinguish them from other users. The authors assessed the technique using a corpus of 1.6 million tweets produced over three months. Two settings are examined to determine the capability of sensing user behavior: four classifications (bully, aggressive, spam, and typical users) and three classifications (bully, aggressive, and normal users), Three ML methods [RF, NV, and DTs (J48)] are employed for detection. When employing 3-class classification, RF got an accuracy of 91.08%, kappa value of 0.5284, RMSE of 0.2117, precision of 0.899%, recall of 0.917%, and AUC of 0.907%. Balakrishnan et al. [134] provided an automatic cyberbullying detection approach based on Twitter users’ psychological features such as personalities, sentiment, and emotion. In particular, it constructs a detection model based on existing empirical evidence on user personalities and cyberbullying perpetration, and then enhances it with user sentiment and emotion. The dataset contained 5453 tweets from Twitter and was manually annotated by experts. The annotated dataset was obtained from [133] with 9484 annotated tweet IDs. For detection three ML algorithms are used [RF, NV, and DTs (J48)]. The authors obtain high results with J48 algorithms and with an accuracy of 91.88%, a weighted AUC of 0.97%, F-score 0.92, kappa 0.840, and RMSE 0.178. Balakrishnan et al. [135] proposed a cyberbullying detection technique based on user personality as indicated by the Big Five [136] and Dark Triad [137] models. The model seeks to identify bullying patterns within Twitter groups based on correlations between personality factors and cyberbullying. The authors used RF for cyberbullying categorization in addition to a baseline approach comprising seven Twitter variables. The best result was found when using baseline + the key personalities from Big Five and Dark Triad with values equal 96% (precision), 95% (recall), and 0.929% (F-measure). Chavan and Shylaja [138] used supervised ML algorithms like SVM and LR to identify swear words and obnoxiousness in comments on social networks. The datasets used for experiments are collected from the Kaggle website. The best result achieved is logistic regression with the occurrence of pronouns and skip-gram as features of accuracy of 86%, AUC of 86.92%, recall of 0.71%, and precision of 0.769%. Talpur and O’Sullivan [139] developed a feature-based approach that takes characteristics from a tweet’s content to generate a machine learning classification model for determining whether a tweet is low, non-cyberbullied, medium, or high-level cyberbullied. Enabling feature extraction (age, sexual identity, and personal characteristics) and Twitter API capabilities to integrate pointwise semantic orientation as a new input feature to indicate the degree of cyberbullying in a tweet. To test the effectiveness of each feature and classifier performance, feature selection approaches such as information gain, chi-square, and correlation were applied in different combinations of features. For detection, the author used NB, SVM, KNN, DT, and RF models. Tweets were gathered between November and December of 2019 with 14,495 tweets and annotated to become 11,904 tweets classified. The authors employ the SMOTE technique [140] by oversampling the minority classes to balance the data. RF achieved the highest Kappa with a value of 84%, F-Measure of 92%, and Accuracy of 93% when parameters were set to Base Classifier + SMOTE + Cost Adjusted + Predicted Features + PMI. Al-Garadi et al. [141] created a functionality model based on tweet features such as network, user, activity, and tweet content. To assess the performance of the four chosen classification models, specifically, NB, LibSVM, RF, and KNN and to evaluate the most important features, three feature selection methods are used: the information gain, chi-square test, and Pearson correlation. The dataset was gathered from Twitter during January and February of 2015 and contains 10,007 tweets. SMOTE was applied to deal with the data set’s unequal class distribution. RF using SMOTE alone showed the best AUC (0.943), F-measure (0.936), Precision (0.941), Recall (0.939). Kumar and Sachdeva [142] developed a deep neural model for detecting cyberbullying in three different social data modalities (textual, visual, and info-graphic). CapsNet–ConvNet is an architecture that includes a capsule network (CapsNet) DNN with dynamic routing for predicting textual bullying content and a convolution neural network (ConvNet) for predicting visual bullying content. Adikara et al. [143] used NB Classifier with Lexicon Based Features to identify cyberbullying comments on Instagram by integrating Bag of Word features and Lexicon Based features. The data used is drawn at random from Bahasa Indonesia Instagram comments (Indonesian). Zhao et al. [144] proposed a framework for detecting cyberbullying called embeddings enhanced Bag-of-Words model (EBoW), constructed a collection of predefined insulting terms using word embeddings and gave varying weights to obtain bullying features, which are then concatenated using Bag-of-Words and Latent semantic features (LSA) [145] and utilized to generate the final representation before putting it into a linear SVM classifier. The authors used the Twitter dataset [146] to evaluate the proposed model EBoW. Several learning models are used for detection such as BoW Model, Semantic-enhanced BoW Model [147], LSA, LDA [148], and EBoW. The proposed model obtained the best result with a precision equal to 76.8%, recall of 79.4%, and F1-score of 78.0%. The authors in [149] addressed the issue of automatically identifying abusive language on Arabic social media platforms using CNN, Bi-LSTM, Bi-LSTM with the attention mechanism, and a combined CNN–LSTM. The researchers used the authors’ dataset in [150]. The best recall is achieved by the combined CNN–LSTM network with a value equal to 83.46% while the best accuracy, precision, and F1-score are achieved by the CNN with values equal to 87.84%, 86.10%, and 84.05% respectively. Haidar et al. [151] created two customized tools for data acquisition. The author used these tools for scraping data from social networks such as Facebook and Twitter. Twitter Scrapper was written in PHP [152], but Facebook Scrapper [153] was written in Python. Both scrapers were linked to a MongoDB server. The tweets were mostly from Lebanon, Syria, the Gulf region, and Egypt. The total size of the tweet database was 4.93 GB. The data was cleaned and preprocessed using WEKA. The TweetToSentiStrengthFeatureVector filter was used as part of the preprocessing techniques, which converts strings to word vectors and normalizes them. After deduplication, the Arabic dataset comprised 35,273 distinct tweets. There were 91,431 tweets in the English data set. The data were subjected to a NB classifier and SVM analysis. In the bullying instances, 801 tweets were classified correctly. while 1395 tweets were classified as not bullying content. Finally, the SVM model had the highest recall with a value of 94.1%. Ahmed et al. [154], created a model to detect cyberbullying in Bangla and Romanized Bangla writings. The authors manually chose YouTube videos of several well-known Bangladeshi social media personalities and collected comments using YouTube API version 3.0 [155]. The texts were divided into two databases by the writers. The first dataset had 5000 Bangla texts, whereas the second contained 7000 Romanized Bangla texts. Following that, a third dataset was formed by bringing the first two datasets together, which totaled 12,000 texts. Following that, the authors manually labelled all three datasets as either bullying or not bullying. The authors applied NB, SVM, LR, XGBoost, CNN, LSTM, BLSTM, and GRU algorithms on three datasets. The authors got an accuracy of 84%. CNN outperformed all other algorithms in the dataset, including Bangla texts. Ali and Syed [156], used LR, SVM, RF, NB, and the Ensemble approach to detect cyberbullying. Four different datasets were used. Because ML methods require numerical input for training, the text was first transformed into a numerical form using a label encoder. The results showed that SVM, Logistic Regression, and Ensemble outperformed the other classifiers, with an average accuracy of 92%. Süzen and Duman [157], designed a neural network to estimate the type of cyberbullying that young people have been exposed to. It is an XGBoost-based method from the model community learning algorithm XGB-CTD. A survey was created from the cyber security scale that was used as a sample to construct the data collection. To achieve high accuracy, XGB CTD and other methods such as SVM, RF, and gradient boosting DTs (GBDT) were trained by identifying optimal hyperparameters. XGB-CTD has the highest accuracy rating of 91.75%. Haidar et al. [158], introduced a research to detect cyberbullying in Arabic. The authors used two datasets, the first of which had 4913 records. The second dataset had 34,890 records. The dataset was divided into 80% for training and 20% for testing, and the FFNN model was used on these datasets, yielding a 94.56% accuracy rate on the second dataset. Sanchez and Kumar [159], applied the NB model to recognize the power of sentiment analysis in detecting bullying on Twitter. The used dataset comprised 1,600,000 training tweets and 359 testing tweets. The authors used Amazon’s Mechanical Turk (crowdsourcing) [160] to categorize unlabeled data and to verify and validate freshly labeled data as part of the outcome evaluation. The NB model achieved a maximum accuracy of 67.3%. The authors in [161], proposed a CNN cyberbullying detection (CNN-CB) algorithm and used an SVM approach to compare it with the proposed system. The dataset was acquired by the Twitter API and contained 39,000 tweets taken from the Twitter public timeline. Many trials revealed that the CNN-CB algorithm outperforms standard text cyberbullying identification with a 95% accuracy, precision of 93%, and recall of 73%. The authors in [162] suggested a hybrid architecture combining character-level CNN and word-level LRCN, with LRCN implemented with Tensorflow. The two models were applied to a dataset containing 8815 comments that were binary tagged as 0 (neutral) and 1 (positive) (cyberbullying). Using LRCN, the accuracy rate was 87.22%. The authors in [163] created a dataset from Instagram and Twitter posts written in Turkish. It contains 900 messages, 450 of which contain cyberbullying content and 450 of which do not. To detect cyberbullying, the authors used ML approaches such as SVMs, DTs (C4.5), NB, and kNNs classifiers. For the dataset utilized, the NB improved accuracy to 84%. The authors in [164] employed the CNN model and presented a novel pronunciation-based CNN (PCNN), and then they used two datasets. The first dataset was gathered from the Twitter platform, which has only 1313 messages. The second dataset was gathered from the social networking site Formspring.me, included 13,000 messages that were labeled by a web service called Amazon Mechanical Turk. After the datasets were collected, the authors applied CNN and PCNN classifiers, and the final results were an accuracy of 0.968%, a precision of 74%, a recall of 0.453%, and F1-score of 0.562% when using the PCNN classifier. Malmasi and Zampieri [165], used text categorization algorithms to discriminate between hate speech, profanity, and other texts. To construct a baseline for this job, the authors used typical lexical characteristics and a linear SVM classifier. A character 4-gram model yielded the best results, with an accuracy equal to 78%. Alakrot et al. [166], gathered and labeled a large dataset of Arabic YouTube comments that included both offensive and inoffensive comments. This dataset was used to train an SVM classifier, and the authors experimented using a variety of N-gram features, word-level features, and pre-processing approaches. The authors then summarized the pre-processing methods and attributes that allow training a more accurate classifier, with accuracy of 90.05%, than earlier studies on Arabic text classifiers.

Standard datasets

This section presents the most used benchmarking datasets in the literature for evaluating MLID methods. CLEF 2018 Consumer Health Search Task [82] focuses on the efficacy of health-related information offered by search engines. The collection includes 5,535,120 web pages retrieved from CommonCrawl3. The CLEF eHealth dataset was labeled as Trustworthiness (10,405 positive, 3820 negative), Readability (3102 positive, 12,455 negative), Useful (1567 positive, 11,488 negative). The Stanford Natural Language Inference (SNLI) dataset [102] contains 570k sentence pairs that have been labeled as neutral, contradiction, or entailment by hand. It has Flickr30k image captions and hypotheses generated by crowd-sourced annotators. It also has annotators who can judge the relationship between sentences that represent the same event. Each pair is labeled as “entailment,” “neutral,” “contradiction,” or “-,” with “-” indicating that an agreement could not be gained. The Multi-Genre Natural Language Inference (MultiNLI) dataset [103] contains 433K sentence pairs. Its dimensions and collection method are very similar to SNLI. MultiNLI presents ten different written and spoken English features (Face-to-face, 9/11, Telephone, Letters, Travel, Oxford University Press, Verbatim, Government, Slate, and Fiction). There are matches between the development and testing datasets that are obtained from the same sources as the training dataset and mismatches between the datasets that do not closely reflect any seen all through training. The Medical Natural Language Inference (MedNLI) dataset [104] has sentence pairs created from the past medical history section of MIMIC-III clinical notes labeled as Definitely True, Maybe True, and Definitely False by Physicians. The dataset consists of 11,232 training, 1395 developments, and 1422 test samples. COVIDLIES [105] includes 86 misconceptions, along with 6761 annotated tweet–misconception pairs. The dataset labeled as 670 agree, 343 disagree, and 5748 no stance. COVID19FN [107] is a dataset of misinformation news collected from Poynter and other fact-checking websites during the Infodemic. It comprises roughly 2800 news articles that are labeled as real or fake news. CoAID (COVID-19 Healthcare Misinformation Dataset) [97] collected from websites and social media fake news related to COVID-19 healthcare. This dataset includes 296,000 user engagements, 4251 news mentions, 3555 tweets, and a ground-truth label. The disaster dataset [110] contains 7613 tweets labeled as a real disaster (1) or unreal disaster (0). Specifically, it includes 4342 tweets representing real disasters and 3271 representing the opposite. This dataset consists of five features: id, text, keyword, target, and location. In PolitiFact [111] there are two CSV files. The first one holds 432 tweets related to real news and the second file contains 618 tweets about fake news. The gossip cop dataset [111], like the PolitiFact dataset, consists of two CSV files. Both datasets have five features: id, tweet-id, title, URL, and label. The Koirala dataset [117] was collected using the Webhose.io tool. This dataset consists of 3002 tweets which are divided into three labels: fake news, real news, and partially fake news. COVID-19 Fake News dataset [167] was built from the ground up using a subset of the Zenodo dataset [168]. The summary of the used datasets has been demonstrated in Table 3.
Table 3

A summary of the used datasets

Ref.DatasetSize (text)No. of features
[82]CLEFeHealth5,535,1206
[97]CoAID35554
[98]ReCOVery2029
[99]FakeHealth (release)606
[99]FakeHealth (Story)1690
[102]SNLI570,00013
[103]MultiNLI433,00010
[104]MedNLI14,0497
[105]COVIDLIES67614
[107]COVID19FN280011
[110]Disasters76135
[111]PolitiFact10504
[111]Gossip cop10,6505
[117]Koirala30025
[130]GI-SAID60,087
[169]COVID-19 Fake News21,379
A summary of the used datasets The dataset was collected from the official accounts of global health organizations, as well as their vaccines, from the websites of the WHO [169], International Committee of the Red Cross (ICRC) [170], United Nations (UN) [171], and United Nations Children’s Fund (UNICEF) [172], as well as their official Twitter accounts. After preprocessing the dataset, 1379 new data points were available, with 80% of them committed to training and 20% for usage in the testing phase. The training dataset contains 17,103 tweets. 9179 Tweets were labeled as true, and 7924 tweets were labeled as false. The testing dataset contains 4276 tweets. 2186 Tweets were labeled as true, and 2090 were labeled as false. ReCOVery [98] is a set of COVID-19 news written in English language, gathered from various websites. An automated approach is used to evaluate the news without involving the experts’ domain. To evaluate the news’ reliability, two famous fact-checking sites were used. These websites are NewsGuard [173], and Media Bias/Fact Check, MBFC [174]. These websites create trustworthiness ratings: the news is labeled as trustworthy if it exceeds certain thresholds, and as unreliable if not. FakeHealth [99] is a set of reviews written in English by specialists about medical interventions, health, treatments, etc. The dataset published on the HealthNewsReview website [175] which is a web page represents a project supported by the Informed Medical Decisions Foundation [176] that ran from 2005 to 2018. GI-SAID [130] is a global science project and primary source founded in 2008 that gives free access to genomic data of influenza viruses and the COVID-19 coronavirus. It has 60,087 features.

Standard evaluation metrics

Accuracy (ACC): is a popular metric for analyzing MLID problems. Based on the input, this metric is used to decide which model is better at defining patterns and relations between samples in a dataset. It decides if a set of measurements are correct on average [177]. This metric can be calculated as follows: A comparison of the accuracy values reached by MLID approaches have been explained in Table 4.
Table 4

A comparison of the accuracies reached by MLID approaches

ReferencesModelACC (%)
[83]DETERRENT96.00
[92]RF90.10
[95]RF73.63
[100]NN99.68
[106]RF94.49
[109]Modified LSTM98.57
[110]Ensemble Model98.55
[112]BERT87.00
[116]kNN–BGA75.40
[120]CNN94.20
[131]DNN95.20
[133]RF91.08
[134]J4891.88
[138]LR86.00
[139]RF93.00
[142]ConvNet97.05
[149]CNN87.84
[154]CNN84.00
[156]SVM, LR, Ensemble92.00
[157]XGB-CTD91.75
[158]FFNN94.56
[159]NB67.30
[161]CNN-CB95.00
[131]DNN95.20
[133]RF91.08
[134]J4891.88
[138]LR86.00
[139]RF93.00
[142]ConvNet97.05
[149]CNN87.84
[154]CNN84.00
[156]SVM, LR, Ensemble92.00
[157]XGB-CTD91.75
[158]FFNN94.56
[159]NB67.30
[161]CNN-CB95.00
[162]LRCN87.22
[163]NB84.00
[164]PCNN96.00
[165]SVM78.00
[166]SVM90.50
Error Rate (ERR): is calculated by dividing the total number of wrong predictions by the total number of samples in the dataset. The ideal error rate is zero and the worst is one [178]. It can be described as follows: Precision: represents the number of true positives divided by the total number of true positives and false positives [179]. It can be calculated as follows: Recall: This metric is also known as sensitivity. It represents the ratio of a specific class identified in a correct manner using all the instances of that class as the following formula describes [179]: F1-score: It’s among the most important metrics for evaluating ML and DL models. It combines precision and recall to calculate the F Score or F Measure by summarizing the model’s performance predictions as described below [180]: Area under curve (AUC): is another name for c-statistics. It calculates the total two-dimensional area beneath the entire ROC curve. It is computed by adding Concordance Percentage and 0.5 times of Tied Percentage [181]. False-positive rate (FPR): is a reliable and valid instrument used to identify the ratio of incorrectly classified negative samples to the overall number of negative samples [52]. Specificity: is a metric used to assess the accuracy of negative instances [52]. Miss Rate: It is the percentage of positive samples that are misclassified [100]. Root Mean Square Error (RMSE): is the residuals’ standard deviation (prediction errors). It indicates how narrowly the data is centered near the line of best fit [182, 183]. A comparison of the accuracies reached by MLID approaches A comparison of the most popular performance metrics obtained using MLID approaches is shown in Table 5.
Table 5

A comparison of other performance metrics obtained using MLID approaches

ReferencesModelMetric(%)
[82]RF, SVM, DistilRoBERTaF1-score93.00
[83]DETERRENTPrecision94.00
[83]DETERRENTRecall91.00
[83]DETERRENTF1-score93.00
[95]RFPrecision72.80%
[95]RFRecall73.60
[95]RFAUC89.00
[96]CNNAUC97.30
[96]CNNF1-score95.30
[100]NNERR32.00
[100]NNAUC99.47
[101]BERTSCORE (DA) + SBERT (DA)Precision63.30
[101]BiLSTMRecall94.20
[101]BiLSTMF1-score89.50
[106]RFPrecision95.00
[106]RFRecall95.00
[106]RFF1-score95.00
[108]Ensemble ModelPrecision98.55
[108]Ensemble ModelRecall98.55
[108]Ensemble ModelF1-score98.55
[109]Modified LSTMPrecision98.55
[109]Modified LSTMRecall98.60
[109]Modified LSTMF1-score98.50
[116]kNN–BGAPrecision66.22
[116]kNNRecall69.57
[116]kNN–BSSAF1-score61.96
[120]CNNPrecision93.60
[120]CNNRecall93.90
[120]CNNF1-score93.70
[120]CNNSpecificity93.90
[120]CNNError rate5.80
[120]CNNMiss rate5.50
[120]CNNFPR6.00
[131]DNNF1-score73.60
[134]J48AUC97.00
[134]J48F-score92.00
[134]J48Kappa84.00
[134]J48RMSE17.00
[133]RFKappa59.00
[133]RFRMSE14.00
[133]RFAUC81.00
[133]RFPrecision90.00
[133]RFRecall91.00
[135]RF + Big Five and Dark TriadPrecision96.00
[135]RF + Big Five and Dark TriadRecall95.00
[135]RF + Big Five and Dark TriadF1-score92.00
[138]LRAUC86.92
[138]LRRecall71.00
[138]LRPrecision76.90
[139]RFF1-score92.00
[139]RFKappa84.00
[141]RFAUC94.30
[141]RFRecall93.00
[141]RFPrecision94.00
[141]RFF1-score93.00
[142]CapsNet–ConvNetAUC98.00
[142]ConvNeRecall95.08
[142]ConvNePrecision98.60
[149]CNN–LSTMRecall83.46
[149]CNNPrecision86.10
[149]CNNF1-score84.05
[151]SVMPrecision93.40
[151]SVMRecall94.10
[151]SVMF1-score92.70
[154]CNNPrecision84.00
[154]CNNROC84.00
[154]XGBoostRecall91.00
[154]NBF1-score86.00
[161]CNN-CBPrecision93.00
[161]CNN-CBRecall73.00
[144]SVMPrecision76.80
[144]SVMReacll79.40
[144]SVMF1-score78.00
[164]PCNNPrecision74.00
[164]PCNNRecall45.30
[164]PCNNF1-score56.20
[166]SVMPrecision88.00
[166]SVMRecall80.00
[166]SVMF1-score82.00
A comparison of other performance metrics obtained using MLID approaches

Data limitations and challenges

All research has challenges and/or limitations associated to the dataset. So, this chapter considers the most common of them.

Lack of a particular form

Detecting misleading information is difficult manually and automatically because it comes in many different forms in the online environment. For example, clickbait headlines are used to entice users to open potentially biased articles to profit from views. When the title reflects an important subject for the reader while the content is irrelevant and completely contrary to the reader’s expectations. Fake news most often exists on malicious websites that spread it on purpose. They hire malicious users to do this job or use social media bots. As a result, distracted users who fail to verify the article’s source before will help them by sharing it. On the other hand, Fake news can also exist on trustworthy websites. This might happen by accident or in the rush to publish shocking news without checking the source first.

Reliable data

In data collection, fact-checking is a critical aspect to produce a reliable dataset. The researcher in [184] introduced fact-checking techniques used in modeling and identifying fake news. Fact-checking techniques depend on human professionals to evaluate data integrity. This methodology is applied by websites such as FactCheck.org, and Snopes. The main problem with these techniques is time and resources consuming.

Untrusted sources

Despite the advent of numerous professional fact-checking sites and tools, there are still flaws and issues. Because of the absence of consistency and coherence among different fact-checking sites, the sites may not be as dependable or credible as intended. For example, Fact Checker [115] and Politifact rarely verify the same facts, and when they do, they frequently disagree [185]. This may further perplex people and possibly compel them not to reevaluate their preexisting impressions of the truthfulness of a tale. In fact, fact-checking websites are not available to the general public as originally anticipated. Escaping our complex misinformation networks may be more difficult than we believe [186].

Unbalanced data

Data collection is an expensive and difficult process, so gathering unbalanced data consider a big problem for researchers because, after this effort and in most cases, they have to drop part of this collected data. To apply most of the classification algorithms, we are required to balance uneven data by equalizing the size of label classes, which leads in dealing with significantly less data than we may wish or dealing with data redundancy. As a result, this will affect our ability to achieve a large enough and representative sample of examples from the minority class negatively. Karajeh et al. [131], faced this problem by using an unbalanced dataset that contains a higher percentage of non-informational tweets than informational tweets.

Limited data sources

Misleading information can be found anywhere, but the researchers tend to use Twitter data because the tweets are considered open source and their use are legal, while other platforms such as Facebook and LinkedIn are considered private and illegal content. This problem has faced a lot of researchers such as Karajeh et al. [131] since Twitter was the only platform used in this work.

Conclusion

This paper proposed a detailed study of the most recent works developed for MLID in health fields based on ML and DL techniques. It also introduced a detailed discussion of the main phases of the generic adopted approach for MLID. Many critical topics that affect MLID techniques have been discussed, including text preprocessing, feature engineering, data splitting, and model building. Moreover, the benchmarking datasets and the most used metrics to evaluate the performance of MLID algorithms are explored. A deep investigation of the limitations and drawbacks of the current technological advancements in various research directions is provided to help the researchers to choose the most suitable for this emerging task of MLID. Detecting misleading information in the text is a difficult task with numerous applications. More DL transformers should be used to get better results. Further, expanding the annotated datasets by including data from different domains and languages are fertile research areas for future works paths.
Table 6

The complete list of abbreviations used in this survey

AbbreviationsDefinitionAbbreviationsDefinition
MLIDMisleading information detectionERFEnsemble random forest
DTDecision treeXGBoostExtreme gradient boosting
kNNk-Nearest neighborACCAccuracy
MLMachine learningSVMSupport vector machine
PACPassive aggressive classifierBi-LSTMBidirectional long short-term memory
MLPMulti-Layer perceptronBERTBidirectional encoder representations from transformers
ALBERTA Lite BERTNBNaive Bayes
CNNConvolutional neural networkLSTMLong short-term memory
LSVMLinear support vector machinesERRError rate
NLINatural language inferenceSBERTSentence-BERT
PSOParticle swarm optimizationGAGenetic algorithm
BNBBernoulli Naïve BayesAUCArea under The curve
SSASalp swarm algorithmBSSABinary salp swarm algorithm
RRSERoot relative squared errorPREPrecision
EBoWEmbeddings-enhanced Bag-of-WordsRECREC
BPSOBinary particle swarm optimizationBGABinary-coded genetic algorithm
NNNeural networkCAMComplementary and alternative medicine
PMI-SOPointwise semantic orientation of each word and phrasesSMOTESynthetic Minority Over-sampling Technique
LDALatent Dirichlet allocationLSALatent semantic features
LibSVMLibrary for support vector machinesCapsNetCapsule network deep neural network
  20 in total

1.  Learning to forget: continual prediction with LSTM.

Authors:  F A Gers; J Schmidhuber; F Cummins
Journal:  Neural Comput       Date:  2000-10       Impact factor: 2.026

2.  Random forest: a classification and regression tool for compound classification and QSAR modeling.

Authors:  Vladimir Svetnik; Andy Liaw; Christopher Tong; J Christopher Culberson; Robert P Sheridan; Bradley P Feuston
Journal:  J Chem Inf Comput Sci       Date:  2003 Nov-Dec

Review 3.  An introduction to the five-factor model and its applications.

Authors:  R R McCrae; O P John
Journal:  J Pers       Date:  1992-06

4.  Diagnosis of tuberculosis--newer tests.

Authors:  A K Agarwal; S Wadhwa; S Chandra
Journal:  J Assoc Physicians India       Date:  1994-08

5.  Lies Kill, Facts Save: Detecting COVID-19 Misinformation in Twitter.

Authors:  Mabrook S Al-Rakhami; Atif M Al-Amri
Journal:  IEEE Access       Date:  2020-08-26       Impact factor: 3.367

6.  Anatomy of an online misinformation network.

Authors:  Chengcheng Shao; Pik-Mai Hui; Lei Wang; Xinwen Jiang; Alessandro Flammini; Filippo Menczer; Giovanni Luca Ciampaglia
Journal:  PLoS One       Date:  2018-04-27       Impact factor: 3.240

Review 7.  Health Misinformation Detection in the Social Web: An Overview and a Data Science Approach.

Authors:  Stefano Di Sotto; Marco Viviani
Journal:  Int J Environ Res Public Health       Date:  2022-02-15       Impact factor: 3.390

8.  Misinformation making a disease outbreak worse: outcomes compared for influenza, monkeypox, and norovirus.

Authors:  Julii Brainard; Paul R Hunter
Journal:  Simulation       Date:  2020-04       Impact factor: 1.377

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.