| Literature DB >> 35611303 |
Poonam Rani1, Vibha Jain1, Jyoti Shokeen2, Arnav Balyan1.
Abstract
The ubiquity of handheld devices and easy access to the Internet help users get easy and quick updates from social media. Generally, people share information with their friends and groups without inspecting the posts' veracity, which causes false information propagation in the network. Moreover, detecting false news and rumors in such a massive load of unstructured information is a very tedious task. Results, many literature papers explored different machine learning and deep learning approaches to detect the presence of rumors on social media networks. Although detection of misleading news and rumors is not sufficient, therefore, we have proposed a model for the detection and prevention of transmitted rumors in this paper. In this paper, we use blockchain technology to verify the credibility of information and design a framework with four layers: network layer, blockchain layer, machine layer, and device layer, to prevent the propagation of rumors in the network. We also use deep learning techniques to identify the anomalies in the network. The Bi-directional Long Short Term Memory (Bi-LSTM) model is used to prevent the introduction of new rumors by continuously monitoring incoming messages in the network. The experimental results demonstrate that the proposed Bi-LSTM model outperforms state-of-the-art machine learning methods and recent baseline work. Performance is compared over different metrics such as accuracy, precision, recall, f1-score, and specificity. Experiment results show that our Bi-LSTM model outperforms all the other approaches and achieved 99.63 % accuracy. Additionally, the probability of incorrect detection is significantly low with only 0.13% false positive.Entities:
Keywords: Blockchain; COVID-19; LSTM; Rumor
Year: 2022 PMID: 35611303 PMCID: PMC9120809 DOI: 10.1007/s12652-022-03900-2
Source DB: PubMed Journal: J Ambient Intell Humaniz Comput
Summary of related works
| Authors | Main work | Technique | Parameters | Limitations |
|---|---|---|---|---|
|
Jin et al. ( | Online rumor behavior | Text-based | Textual data | No comparison with previous studies |
|
Yang et al. ( | CNN | Image and textual data | Slow | |
|
O’Brien et al. ( | Fake news detection | Deep learning, CNN | Language patterns | No comparison with previous studies |
|
Shae and Tsai ( | Fake news detection and prevention | AI, Blockchain | Smart contracts, News rooms | No implementation performed |
| Gao and Gao ( | Rumor prevention | Voting rule, Blockchain | Private blockchain | No consensus mechanism |
| Alsaeedi and Al-Sarem ( | Rumor detection | CNN-based, deep learning | Layers | Not capable of detecting rumors online |
| Polap et al. ( | Internet of Medical Things | Blockchain, CNN | Federated learning, Medical data | No consensus mechanism provided |
| Polap et al. ( | Internet of Medical Things | Blockchain, Machine learning | Threaded federated learning, Multi-agent systems, Security | Internet connection problem |
|
Dibaei et al. ( | Vehicular networks | Blockchain, Machine learning | Security | – |
|
Tida et al. ( | Fake news detection | BERT, Transfer learning | Encoders, decoders | – |
|
Raza and Ding ( | Fake news detection | Transformer approach | Encoders, decoders, social context, news content | Domain level error analysis, use of ground-truth labels, weak supervision of deep neural networks, and less quantity of user profile data |
Fig. 1Blockchain
Fig. 2Message flow in network
Fig. 3System architecture
Fig. 4Steps for classification
Fig. 5LSTM model
Parameters used for the proposed model
| Parameters | Information |
|---|---|
| Number of layers | 6 |
| Activation function | Relu |
| Dropout rate | 30% |
| Optimizer | Adam |
| Learning rate | 0.001 |
| Loss function | Binary crossentropy |
| Batch size | 16 |
| Number of Epochs | 8 |
| Validation split | 20% |
Fig. 6A simulation of the network in a 100 100 meter square area
Fig. 7Random Forest confusion matrix
Fig. 8Random Forest classification report
Fig. 9Ada Boost confusion matrix
Fig. 10Ada Boost classification report
Fig. 11Logistic Regression confusion matrix
Fig. 12Logistic Regression classification report
Fig. 13KNN confusion matrix
Fig. 14KNN classification report
Fig. 15Bernoulli Naive Bayes confusion matrix
Fig. 16Bernoulli Naive Bayes classification report
Fig. 17Gradient Boost confusion matrix
Fig. 18Gradient Boost classification report
Fig. 19Extra Tree confusion matrix
Fig. 20Extra Tree classification report
Fig. 21Bi-LSTM confusion matrix
Fig. 22Accuracy under LSTM model training vs validation
Fig. 23Loss under LSTM model training vs validation
Fig. 24Block size vs. network traffic
Comparison of proposed approach with different methods
| Measure | Algorithm | |||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Random Forest | Ada Boost | Logistic regression | k-nearest Neighbour | Bernoulli Naive Bayes | Gradient boost | Extra tree | Bi-LSTM | |||||||||
| 60–40% | 70–30% | 60–40% | 70–30% | 60–40% | 70–30% | 60–40% | 70–30% | 60–40% | 70–30% | 60–40% | 70–30% | 60–40% | 70–30% | 60–40% | 70–30% | |
| TP | 10377 | 5827 | 10379 | 5835 | 10417 | 5842 | 8836 | 4868 | 10140 | 5663 | 10449 | 5829 | 10268 | 5759 | 4559 | 4647 |
| TN | 5615 | 5316 | 9661 | 5343 | 9576 | 5345 | 6992 | 3957 | 9351 | 5176 | 9657 | 5338 | 9542 | 5242 | 4234 | 4318 |
| FP | 4066 | 37 | 20 | 10 | 105 | 8 | 2689 | 1396 | 330 | 177 | 24 | 15 | 139 | 71 | 100 | 12 |
| FN | 147 | 45 | 145 | 37 | 107 | 30 | 1688 | 1004 | 384 | 209 | 75 | 43 | 256 | 113 | 87 | 3 |
| Accuracy | 0.79 | 0.98 | 0.99 | 0.99 | 0.98 | 0.97 | 0.78 | 0.78 | 0.96 | 0.96 | 0.99 | 0.99 | 0.98 | 0.98 | 0.97 | 0.99 |
| Precision | 0.71 | 0.99 | 0.99 | 0.99 | 0.99 | 0.99 | 0.76 | 0.77 | 0.96 | 0.96 | 0.99 | 0.99 | 0.98 | 0.98 | 0.97 | 0.99 |
| Recall | 0.98 | 0.99 | 0.98 | 0.99 | 0.98 | 0.99 | 0.83 | 0.82 | 0.96 | 0.96 | 0.99 | 0.99 | 0.97 | 0.98 | 0.98 | 0.99 |
| F1-score | 0.83 | 0.99 | 0.99 | 0.99 | 0.98 | 0.99 | 0.80 | 0.80 | 0.96 | 0.96 | 0.99 | 0.99 | 0.98 | 0.98 | 0.97 | 0.99 |
| Specificity | 0.38 | 0.90 | 0.92 | 0.91 | 0.91 | 0.91 | 0.60 | 0.63 | 0.89 | 0.88 | 0.92 | 0.91 | 0.91 | 0.89 | 0.90 | 0.92 |
Comparison of results with recent works
| Paper | Methodology | Accuracy (%) |
|---|---|---|
|
Yang et al. ( | CNN | 92.10 |
|
O’Brien et al. ( | Deep neural network (black-box approach) | 93.50 |
|
Ghanem et al. ( | Word embeddings and n-gram | 48.80 |
|
Singh et al. ( | Linguistic analysis and word count | 87.00 |
|
Ahmed ( | Uni-gram model | 89.00 |
|
Ruchansky et al. ( | Hybrid model | 89.20 |
|
Tida et al. ( | Hybrid BERT | 97.00 |
| Proposed model | Bi-LSTM | 99.63 |