| Literature DB >> 31430308 |
Sean MacAvaney1, Hao-Ren Yao1, Eugene Yang1, Katina Russell1, Nazli Goharian1, Ophir Frieder1.
Abstract
As online content continues to grow, so does the spread of hate speech. We identify and examine challenges faced by online automatic approaches for hate speech detection in text. Among these difficulties are subtleties in language, differing definitions on what constitutes hate speech, and limitations of data availability for training and testing of these systems. Furthermore, many recent approaches suffer from an interpretability problem-that is, it can be difficult to understand why the systems make the decisions that they do. We propose a multi-view SVM approach that achieves near state-of-the-art performance, while being simpler and producing more easily interpretable decisions than neural methods. We also discuss both technical and practical challenges that remain for this task.Entities:
Mesh:
Year: 2019 PMID: 31430308 PMCID: PMC6701757 DOI: 10.1371/journal.pone.0221152
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Hate-related dataset characteristics.
| Dataset | Labels and percents in dataset | Origin Source | Language |
|---|---|---|---|
| HatebaseTwitter [ | Hate 5% | English | |
| WaseemA [ | Racism 12% | English | |
| WaseemB [ | Racism1 1% | English | |
| Stormfront [ | Hate 11% | Online Forum | English |
| TRAC (Facebook) [ | Non-aggressive 69% | English & Hindi | |
| TRAC (Twitter) [ | Non-aggressive 38% | English & Hindi | |
| HatEval [ | Hate 43% / Not Hate 57% | English & Spanish | |
| Kaggle [ | Insulting 26% | English | |
| GermanTwitter | Hate 23% | German |
Hate speech classification results.
The top two approaches on each dataset are reported.
| Dataset | Model | Accuracy | Macro |
|---|---|---|---|
| Stormfront | BERT | 0.8201 | 0.8201 |
| mSVM (ours) | 0.8033 | 0.8031 | |
| TRAC (Facebook) | mSVM (ours) | 0.6121 | 0.5368 |
| BERT | 0.5809 | 0.5234 | |
| HatebaseTwitter | Neural Ensemble | 0.9217 | 0.9118 |
| BERT | 0.9209 | 0.8917 | |
| HatEval | BERT | 0.7480 | 0.7452 |
| Neural Ensemble | 0.7470 | 0.7481 |
Fig 1Self-attention weights for the classification token of the trained BERT model for a sample post.
Each color represents a different attention head, and the lightness of the color represents the amount of attention. For instance, the figure indicates that nearly all attention heads focus heavily on the term ‘we’.