| Literature DB >> 32355598 |
Pete Burnap1, Matthew L Williams2.
Abstract
Hateful and antagonistic content published and propagated via the World Wide Web has the potential to cause harm and suffering on an individual basis, and lead to social tension and disorder beyond cyber space. Despite new legislation aimed at prosecuting those who misuse new forms of communication to post threatening, harassing, or grossly offensive language - or cyber hate - and the fact large social media companies have committed to protecting their users from harm, it goes largely unpunished due to difficulties in policing online public spaces. To support the automatic detection of cyber hate online, specifically on Twitter, we build multiple individual models to classify cyber hate for a range of protected characteristics including race, disability and sexual orientation. We use text parsing to extract typed dependencies, which represent syntactic and grammatical relationships between words, and are shown to capture 'othering' language - consistently improving machine classification for different types of cyber hate beyond the use of a Bag of Words and known hateful terms. Furthermore, we build a data-driven blended model of cyber hate to improve classification where more than one protected characteristic may be attacked (e.g. race and sexual orientation), contributing to the nascent study of intersectionality in hate crime. © Burnap and Williams 2016.Entities:
Keywords: NLP; Twitter; cyber hate; hate speech; machine learning
Year: 2016 PMID: 32355598 PMCID: PMC7175598 DOI: 10.1140/epjds/s13688-016-0072-6
Source DB: PubMed Journal: EPJ Data Sci ISSN: 2193-1127 Impact factor: 3.184
Figure 1Example of text transformation to typed dependency feature set.
Figure 2Formula for generating performance metrics.
Machine classification performance for cyber hate based on disability, race and sexual orientation (results rounded to 2dp)
|
|
|
|
| |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
| |
|
| 0.80 FP = 38 | 0.69 FN = 69 | 0.74 | 0.969 FP = 1 | 0.608 FN = 20 | 0.73 | 0.72 FP = 15 | 0.54 FN = 32 | 0.62 | 0.53 FP = 67 | 0.42 FN = 107 | 0.47 |
|
| 0.89 FP = 19 | 0.66 FN = 75 | 0.76 | 0.00 | 0.00 | 0.00 | 0.93 FP = 3 | 0.53 FN = 33 | 0.67 | 1.00 FP = 0 | 0.098 FN = 165 | 0.18 |
|
| 0.74 FP = 58 | 0.65 FN = 78 | 0.69 | 0.89 FP = 4 | 0.61 FN = 20 | 0.72 | 0.79 FP = 13 | 0.71 FP = 20 | 0.75 | 0.57 FP = 60 | 0.44 FN = 105 | 0.49 |
|
| 0.53 FP = 48 | 0.24 FP = 168 | 0.33 | 0.97 FP = 1 | 0.61 FP = 20 | 0.75 | 0.87 FP = 3 | 0.29 FN = 50 | 0.43 | 0.95 FP = 2 | 0.22 FN = 142 | 0.36 |
|
| 0.89 FP = 19 | 0.69 FN = 70 | 0.77 | 0.97 FP = 1 | 0.61 FP = 20 | 0.75 | 0.91 FP = 4 | 0.59 FN = 29 | 0.71 | 0.96 FP = 2 | 0.27 FN = 134 | 0.42 |
|
| 0.89 FP = 19 | 0.69 FN = 70 | 0.77 | 0.97 FP = 1 | 0.61 FN = 20 | 0.75 | 0.87 FP = 7 | 0.66 FN = 24 | 0.75 | 0.72 FP = 25 | 0.35 FN = 119 | 0.47 |
Cross validation of different types of cyber hate
|
| ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
| ||||||||
|
|
|
|
|
|
|
|
|
| ||
| Testing Data | Disability | 0.96 FP = 1 | 0.61 FN = 20 | 0.75 | 0.00 FP = 0 | 0.00 FN = 51 | 0.00 | 0.00 FP = 1 | 0.00 FN = 51 | 0.00 |
| Race | 0.00 FP = 1 | 0.00 FN = 70 | 0.00 | 0.87 FP = 7 | 0.64 FN = 25 | 0.74 | 0.95 FP = 1 | 0.29 FN = 50 | 0.44 | |
| Sexual orientation | 0.00 FP = 2 | 0.00 FN = 183 | 0.00 | 1.00 FP = 0 | 0.09 FN = 165 | 0.18 | 0.74 FP = 23 | 0.37 FN = 116 | 0.49 | |
Binary cyber hate classification using a combined dataset of 3 different protected characteristics
|
|
|
| |
|---|---|---|---|
| Non-hate | 0.97 | 0.99 | 0.98 |
| Hate | 0.79 (FP = 62) | 0.59 (FN = 162) | 0.68 |
| Overall | 0.96 | 0.97 | 0.96 |
Multi-class cyber hate classification using a combined dataset of 3 different protected characteristics
|
|
|
| |
|---|---|---|---|
| Non-hate-disability | 0.95 | 0.97 | 0.96 |
| Hate-disability | 0.91 | 0.61 | 0.73 |
| Non-hate-race | 0.95 | 0.96 | 0.95 |
| Hate-race | 0.86 | 0.60 | 0.71 |
| Non-hate-sexual orientation | 0.94 | 0.97 | 0.95 |
| Hate-sexual orientation | 0.66 | 0.41 | 0.51 |
Confusion matrix for multi-class cyber hate classification using a combined dataset of 3 different protected characteristics
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|
| 1,798 | 3 | 61 | 0 | 1 | 0 | a = non-hate-disability |
| 18 | 31 | 2 | 0 | 0 | 0 | b = hate-disability |
| 74 | 0 | 1,724 | 7 | 0 | 1 | c = non-hate-race |
| 0 | 0 | 28 | 42 | 0 | 0 | d = hate-race |
| 3 | 0 | 3 | 0 | 1,577 | 37 | e = non-hate-sexual orientation |
| 0 | 0 | 0 | 0 | 108 | 75 | f = hate-sexual orientation |
Confusion matrix for multi-class cyber hate classification using a combined dataset of 3 different protected characteristics
|
|
|
|---|---|
|
| |
| det(backdoor-7, the-6) | Determiner (a specific reference to a noun phrase) discussing ‘the backdoor’ in a context of homosexual activity |
| dobj(kill-2, yourself-3) | Direct object (an accusatory object of the verb) suggesting homosexual ‘others’ should ‘kill yourself’ |
| det(closet-8, the-7) | Determiner (a specific reference to a noun phrase) discussing ‘the closet’ - most likely referring to where the person should have remained |
| amod(disgrace-6, absolute-5) | Adjectival modifier (a descriptive phrase related to a noun phrase) discussing ‘disgrace’ - and amplifying this accusation with ‘absolute’ |
| det(disgrace-6, an-4) aux(commending-12, him-13) | Determiner (a specific reference to a noun phrase) discussing ‘disgrace’ - plus Auxiliary ‘commending’, branding people commending the person a disgrace |
|
| |
| advcl(won-7, black-11) advcl(won-7, obama-13) | Two adverbial clause modifiers relating ‘won’ and ‘obama’ & ‘won’ and ‘black’ - highlighting the colour of skin as a key related term to the victory |
| aux(destroying-10, is-9) | Auxiliary verb potentially suggesting Obama is having a ‘destroying’ impact |
| amod(people-7, white-6), advmod(won-11, how-9) | Modifiers linking ‘white people’ to the outcome of the election outcomes ‘how…won’ |
| dobj(see-13, you-14) | Direct object (an accusatory object of the verb) referring to ‘you’ and the impact the outcome may have |
|
| |
| amod(athletes-11, olympic-10) advmod(drunk-14, really-13) | Modifiers referring to ‘olympic athletes’ and ‘really drunk’ in mocking manner referring to ‘you’ and the impact the outcome may have |
| det(jokes-10,the-9) | Referring to noun ‘joke’ in relation to paralympic athletes |
| amod(women-12,disabled-11) dobj(falling-13,wish-15) | Modifier of ‘women’ to refer to ‘disabled’ female athletes and ‘wish’ they would be ‘falling’ using direct object |
| amod(bench-11, midget-9) | The key term here being the derogatory term ‘midget’ |