| Literature DB >> 35965845 |
Abderrazek Azri1, Cécile Favre1, Nouria Harbi1, Jérôme Darmont1, Camille Noûs2.
Abstract
The proliferation of rumors on social media has become a major concern due to its ability to create a devastating impact. Manually assessing the veracity of social media messages is a very time-consuming task that can be much helped by machine learning. Most message veracity verification methods only exploit textual contents and metadata. Very few take both textual and visual contents, and more particularly images, into account. Moreover, prior works have used many classical machine learning models to detect rumors. However, although recent studies have proven the effectiveness of ensemble machine learning approaches, such models have seldom been applied. Thus, in this paper, we propose a set of advanced image features that are inspired from the field of image quality assessment, and introduce the Multimodal fusiON framework to assess message veracIty in social neTwORks (MONITOR), which exploits all message features by exploring various machine learning models. Moreover, we demonstrate the effectiveness of ensemble learning algorithms for rumor detection by using five metalearning models. Eventually, we conduct extensive experiments on two real-world datasets. Results show that MONITOR outperforms state-of-the-art machine learning baselines and that ensemble models significantly increase MONITOR's performance.Entities:
Keywords: Ensemble learning; Image features; Machine learning; Rumor verification; Social networks
Year: 2022 PMID: 35965845 PMCID: PMC9362091 DOI: 10.1007/s10796-022-10315-z
Source DB: PubMed Journal: Inf Syst Front ISSN: 1387-3326 Impact factor: 5.261
Fig. 1Two sample rumors posted on Twitter
Fig. 2Overview of MONITOR
Content features
| Description |
|---|
| # of chars, words |
| # of (?), (!) mark |
| # of uppercase chars |
| # of positive, negative words |
| # of mentions, hashtags, URLs |
| # of happy, sad mood emoticon |
| # of 1st, 2nd, 3rd order pronoun |
| Readability score |
Social context features
| Description |
|---|
| # of followers, friends, posts |
| Friends/followers ratio, times listed |
| # of retweets, likes |
| The user shares a homepage URL |
| The user has a profile image |
| The user has a verified account |
| # of tweets the user has liked |
Fig. 3BRISQUE score computed for a natural image and its distorted versions
Fig. 4BRISQUE score computed for real and fake GANs images
Description of image features
| Type | Feature | Description |
|---|---|---|
| Visual | BRISQUE | BRISQUE score of a given image |
| PIQE | PIQE score of a given image | |
| features | NIQE | NIQE score of a given image |
| Statistical | Count_Img | Number of all images in a news event |
| Ratio_Img1 | Ratio of the multi-image tweets in all tweets | |
| features | Ratio_Img2 | Ratio of image number to tweet number |
| Ratio_Img3 | Ratio of the most widespread image in all distinct images |
MediaEval and FakeNewsNet statistics
| Dataset | Set | Tweets | Images | |
|---|---|---|---|---|
| Real | Fake | |||
| MediaEval | Training set | 5,008 | 6,841 | 361 |
| Testing set | 1,217 | 717 | 50 | |
| FakeNewsNet | Training set | 25,673 | 19,422 | 47,870 |
| Testing set | 6,466 | 4,808 | 11,968 | |
Features from the literature
| Feature |
|---|
| Fraction of (?), (!) Mark, # of messages |
| Average # of words, char lengths |
| Fraction of 1st, 2nd, 3rd pronouns |
| Fraction of URLs, @, # |
| Count of distinct URLs, @, # |
| Fraction of popular URLs, @, # |
| The tweet includes pictures |
| Average sentiment score |
| Fraction of positive and negative tweets |
| # of distinct people, loc, org |
| Fraction of people, loc, org |
| Fraction of popular people, loc, org |
| # of Users, fraction of popular users |
| # of followers, followees, posted tweets |
| The user has a Facebook link |
| Fraction of verified users, org |
| # of comments on the original message |
| Time between original message and repost |
Best textual features selected
| MediaEval | FakeNewsNet |
|---|---|
| Tweet_Length | Tweet_Length |
| Num_Negwords | Num_Words |
| Num_Mentions | Num_Questmark |
| Num_URLs | Num_Upperchars |
| Num_Words | Num_Exclmark |
| Num_Upperchars | Num_Hashtags |
| Num_Hashtags | Num_Negwords |
| Num_Exclmark | Num_Poswords |
| Num_Thirdpron | Num_Followers |
| Times_Listed | Num_Friends |
| Num_Tweets | Num_Favorites |
| Num_Friends | Times_Listed |
| Num_Retweets | Num_Likes |
| Has_Url | Num_Retweets |
| Num_Followers | Num_Tweets |
Hyper-parameters configuration space
| Model | Main hyper-parameters | Type | Search space |
|---|---|---|---|
| CART | max_depth | Discrete | [1,21] |
| criterion | Categorical | [‘gini’,‘entropy’] | |
| KNN | n_neighbors | Discrete | [1,21] |
| SVM | C | Discrete | [0.1,2.0] |
| Discrete | [0.1,1.0] | ||
| Kernel | Categorical | [‘linear’, ‘poly’, ‘rbf’,‘sigmoid’] | |
| RF | n_estimators | Discrete | [10,500] |
| max_depth | Discrete | [3,20] |
Performance of individual machine learning models. Bold entries indicates the best performance achieved for each evaluation metric
| Model | Features | MediaEval | FakeNewsNet | ||||||
|---|---|---|---|---|---|---|---|---|---|
| Acc | Prec | Rec | Acc | Prec | Rec | ||||
| CART | Textual | 0.673 | 0.672 | 0.771 | 0.718 | 0.699 | 0.652 | 0.65 | |
| Image | 0.632 | 0.701 | 0.639 | 0.668 | 0.647 | 0.595 | 0.533 | 0.563 | |
| MONITOR | 0.623 | ||||||||
| Castillo | 0.643 | 0.711 | 0.648 | 0.678 | 0.683 | 0.674 | 0.491 | 0.569 | |
| Wu | 0.65 | 0.709 | 0.715 | 0.711 | 0.694 | 0.663 | 0.593 | 0.627 | |
| KNN | Textual | 0.707 | 0.704 | 0.777 | 0.739 | 0.698 | 0.67 | 0.599 | 0.633 |
| Image | 0.608 | 0.607 | 0.734 | 0.665 | 0.647 | 0.595 | 0.533 | 0.563 | |
| MONITOR | |||||||||
| Castillo | 0.652 | 0.698 | 0.665 | 0.681 | 0.681 | 0.651 | 0.566 | 0.606 | |
| Wu | 0.668 | 0.71 | 0.678 | 0.693 | 0.694 | 0.663 | 0.593 | 0.627 | |
| SVM | Textual | 0.74 | 0.729 | 0.834 | 0.779 | 0.658 | 0.657 | 0.44 | 0.528 |
| Image | 0.693 | 0.69 | 0.775 | 0.73 | 0.595 | 0.618 | 0.125 | 0.208 | |
| MONITOR | |||||||||
| Castillo | 0.702 | 0.761 | 0.716 | 0.737 | 0.629 | 0.687 | 0.259 | 0.377 | |
| Wu | 0.725 | 0.763 | 0.73 | 0.746 | 0.642 | 0.625 | 0.394 | 0.484 | |
| RF | Textual | 0.747 | 0.717 | 0.879 | 0.789 | 0.778 | 0.726 | 0.768 | 0.747 |
| Image | 0.652 | 0.646 | 0.771 | 0.703 | 0.652 | 0.646 | 0.771 | 0.703 | |
| MONITOR | |||||||||
| Castillo | 0.702 | 0.727 | 0.723 | 0.725 | 0.714 | 0.669 | 0.67 | 0.67 | |
| Wu | 0.728 | 0.752 | 0.748 | 0.75 | 0.736 | 0.699 | 0.682 | 0.691 | |
Fig. 5Random Forest feature importance
Fig. 6Distribution of true and false classes for top-15 important features
Fig. 7Late fusion scheme
Fig. 8Performance of early and late fusion
Fig. 9Stacking ensemble
Fig. 10Dataset splitting
Fig. 11Super learner ensemble data flow Van der Laan et al. (2007)
Performance of MONITOR and stacking ensemble models. Bold entries indicates the best performance achieved for each evaluation metric
| Model | MediaEval | FakeNewsNet | ||||||
|---|---|---|---|---|---|---|---|---|
| Acc | Prec | Rec | Acc | Prec | Rec | |||
| MONITOR | 0.962 | 0.965 | 0.966 | 0.965 | 0.889 | 0.914 | 0.864 | 0.889 |
| MONITOR | 0.966 | 0.955 | 0.976 | 0.965 | 0.897 | 0.911 | 0.873 | 0.892 |
| MONITOR | 0.968 | 0.968 | 0.970 | 0.969 | 0.906 | 0.90 | 0.927 | 0.914 |
| MONITOR | 0.973 | 0.975 | 0.971 | 0.973 | 0.915 | 0.909 | 0.932 | 0.921 |
| MONITOR | 0.970 | 0.980 | 0.959 | 0.969 | 0.921 | 0.915 | 0.937 | 0.926 |
Fig. 12Stacking ensemble model vs. standalone models on MediaEval
Fig. 13Stacking ensemble model vs. standalone models on FakeNewsNet