| Literature DB >> 35937201 |
Kun Ma1, Changhao Tang1, Weijuan Zhang2, Benkuan Cui1, Ke Ji1, Zhenxiang Chen1, Ajith Abraham3.
Abstract
Fake news detection mainly relies on the extraction of article content features with neural networks. However, it has brought some challenges to reduce the noisy data and redundant features, and learn the long-distance dependencies. To solve the above problems, Dual-channel Convolutional Neural Networks with Attention-pooling for Fake News Detection (abbreviated as DC-CNN) is proposed. This model benefits from Skip-Gram and Fasttext. It can effectively reduce noisy data and improve the learning ability of the model for non-derived words. A parallel dual-channel pooling layer was proposed to replace the traditional CNN pooling layer in DC-CNN. The Max-pooling layer, as one of the channels, maintains the advantages in learning local information between adjacent words. The Attention-pooling layer with multi-head attention mechanism serves as another pooling channel to enhance the learning of context semantics and global dependencies. This model benefits from the learning advantages of the two channels and solves the problem that pooling layer is easy to lose local-global feature correlation. This model is tested on two different COVID-19 fake news datasets, and the experimental results show that our model has the optimal performance in dealing with noisy data and balancing the correlation between local features and global features.Entities:
Keywords: COVID-19; Convolutional neural network; Fake news; Noisy data
Year: 2022 PMID: 35937201 PMCID: PMC9340725 DOI: 10.1007/s10489-022-03910-9
Source DB: PubMed Journal: Appl Intell (Dordr) ISSN: 0924-669X Impact factor: 5.019
Fig. 1General architecture of fake news detection
Fig. 2Architecture of DC-CNN
Fig. 3Dynamic word embedding for Chinese text
Fig. 4Dual-channel Pooling Layer
Dataset introduction
| Dataset | Category | Average length | Fake/True number | Training/Validation/Testing number |
|---|---|---|---|---|
| COVID-19 | 2 | 344 | 883/1955 | 1702/568/568 |
| BAAI | 2 | 300 | 19285/19186 | 23082/7694/7695 |
Configuration of DC-CNN and baseline models
| Model | Detailed configuration |
|---|---|
| TextCNN | Convolution kernel = [3,4,5] |
| Feature maps = 100 | |
| Pooling layer = Max-pooling layer | |
| Multi-channel CNN | Pooling layer 1 = Max-pooling layer (updated) |
| Pooling layer 2 = Max-pooling layer (not updated) | |
| RCNN | Pooling layer = Max-pooling layer |
| hidden_size of LSTM = 256 | |
| DPCNN | Pooling layer = Max-pooling layer |
| num_filters = 200 | |
| Transformer | Attention head = 8 |
Environment configuration
| Hardware and software | Configure |
|---|---|
| CPU | I7-8750H |
| GPU | GeForce GTX 1060 |
| RAM | 16 GB |
| Operating System | Windows 10 x64 |
| Environment | Python: 3.6.5 |
| Tensorflow: 1.12.0 | |
| Keras: 2.2.4 |
Experimental results of COVID-19 datasets
| Model | P | R | ACC | F1 |
|---|---|---|---|---|
| textCNN | 0.8926 | 0.8926 | 0.8926 | 0.8926 |
| Multi-channel CNN | 0.9194 | 0.9243 | 0.9217 | 0.9218 |
| RCNN | 0.8116 | 0.7204 | 0.7457 | 0.7457 |
| DPCNN | 0.8690 | 0.7683 | 0.8995 | 0.8155 |
| Transformer | 0.8299 | 0.8908 | 0.8592 | 0.8586 |
| BERT | 0.9281 | 0.8820 | 0.9470 | 0.9045 |
| ERNIE | 0.8771 | 0.9752 | 0.9541 | 0.9235 |
| DC-CNN | 0.9473 | 0.9489 | 0.9481 |
The bold parts are the optimal results of different experiments
Experimental results of BAAI datasets
| Model | P | R | ACC | F1 |
|---|---|---|---|---|
| textCNN | 0.9059 | 0.9059 | 0.9059 | 0.9059 |
| Multi-channel CNN | 0.9774 | 0.9780 | 0.9777 | 0.9777 |
| RCNN | 0.8886 | 0.8651 | 0.8761 | 0.8767 |
| DPCNN | 0.9218 | 0.8923 | 0.9066 | 0.9068 |
| Transformer | 0.8621 | 0.8123 | 0.8882 | 0.8350 |
| BERT | 0.9541 | 0.9771 | 0.9647 | 0.9655 |
| ERNIE | 0.9705 | 0.9645 | 0.9673 | 0.9675 |
| DC-CNN | 0.9812 | 0.9819 | 0.9815 |
The bold parts are the optimal results of different experiments
Experimental results of different word embedding methods (COVID-19 datasets)
| Method | P | R | ACC | F1 |
|---|---|---|---|---|
| Fasttext | 0.9241 | 0.9489 | 0.9225 | 0.9224 |
| Skip-Gram | 0.9299 | 0.9331 | 0.9313 | 0.9315 |
| CBOW | 0.9332 | 0.9331 | 0.9331 | 0.9331 |
| DWtext | 0.9473 | 0.9489 | 0.9481 |
The bold parts are the optimal results of different experiments
Experimental results of different word embedding methods (BAAI datasets)
| Method | P | R | ACC | F1 |
|---|---|---|---|---|
| Fasttext | 0.9775 | 0.9778 | 0.9776 | 0.9777 |
| Skip-Gram | 0.9805 | 0.9801 | 0.9803 | 0.9803 |
| CBOW | 0.9808 | 0.9806 | 0.9807 | 0.9807 |
| DWtext | 0.9812 | 0.9819 | 0.9815 |
The bold parts are the optimal results of different experiments
Fig. 5The ablation experimental results
Fig. 6Data token length distribution
Experimental effects of different interval data subsets (COVID-19 datasets)
| Section | P | R | ACC | F1 | AUROC |
|---|---|---|---|---|---|
| (0,20] | 0.8545 | 0.8545 | 0.8545 | 0.8545 | 0.7445 |
| (20,40] | 0.6614 | 0.6718 | 0.6640 | 0.6665 | 0.7126 |
| (40,60] | 0.8618 | 0.8545 | 0.8590 | 0.8581 | 0.8769 |
| (60,80] | 0.9349 | 0.9349 | 0.9349 | 0.9349 | 0.6105 |
| > 80 | 0.9186 | 0.9186 | 0.9186 | 0.9186 | 0.9409 |
Experimental effects of different interval data subsets (BAAI datasets)
| Section | P | R | ACC | F1 | AUROC |
|---|---|---|---|---|---|
| (0,20] | 0.9279 | 0.927 | 0.9275 | 0.9275 | 0.9731 |
| (20,40] | 0.9446 | 0.9459 | 0.9452 | 0.9452 | 0.9870 |
| (40,60] | 0.9850 | 0.9841 | 0.9846 | 0.9846 | 0.9990 |
| (60,80] | 0.9878 | 0.9878 | 0.9879 | 0.9878 | 0.9983 |
| > 80 | 0.9673 | 0.9673 | 0.9673 | 0.9673 | 0.9946 |
Fig. 7Data token length distribution