| Literature DB >> 32218569 |
Mohammadreza Rezvan1, Saeedeh Shekarpour2, Faisal Alshargi3, Krishnaprasad Thirunarayan4, Valerie L Shalin4, Amit Sheth5.
Abstract
THIS ARTICLE USES WORDS OR LANGUAGE THAT IS CONSIDERED PROFANE, VULGAR, OR OFFENSIVE BY SOME READERS. The presence of a significant amount of harassment in user-generated content and its negative impact calls for robust automatic detection approaches. This requires the identification of different types of harassment. Earlier work has classified harassing language in terms of hurtfulness, abusiveness, sentiment, and profanity. However, to identify and understand harassment more accurately, it is essential to determine the contextual type that captures the interrelated conditions in which harassing language occurs. In this paper we introduce the notion of contextual type in harassment by distinguishing between five contextual types: (i) sexual, (ii) racial, (iii) appearance-related, (iv) intellectual and (v) political. We utilize an annotated corpus from Twitter distinguishing these types of harassment. We study the context of each kind to shed light on the linguistic meaning, interpretation, and distribution, with results from two lines of investigation: an extensive linguistic analysis, and the statistical distribution of uni-grams. We then build type- aware classifiers to automate the identification of type-specific harassment. Our experiments demonstrate that these classifiers provide competitive accuracy for identifying and analyzing harassment on social media. We present extensive discussion and significant observations about the effectiveness of type-aware classifiers using a detailed comparison setup, providing insight into the role of type-dependent features.Entities:
Mesh:
Year: 2020 PMID: 32218569 PMCID: PMC7100939 DOI: 10.1371/journal.pone.0227330
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1Five contextual types of harassment.
Summery of the related research.
| Paper | Goal | Data | Conclusions |
|---|---|---|---|
| [ | Detecting offensive and hateful speech language | 85.4 Million Tweets Collected from 33458 twitter user using profane words. 25000 tweets are selected. | Collected discriminating terms for hate speech and offensive language |
| [ | Detecting aggressors and their behavior on social media | 1.6 million tweets collected in 3 months, using crowd sourcing for annotation. | Determined that posts of aggressor profiles are more negative |
| [ | Detecting offensive language and identifying its sender. | The data set includes comments from 2,175,474 Youtube users in reaction to the top 18 videos on different Topics. | (i) Conceptualized offensive content, and (ii) enhanced features using lexical, style, structural, and context-specific features. |
| [ | Predicting cyberbullying incidents on Instagram social media | 41K users that are cyber bullied according to the random seed nodes. 3165K tweets collected from 25K public users while 697K Tweets labeled as profane tweets | Classifier designed, trained, and applied for collecting data. Logistic regression classifier |
| [ | Detecting harassment based wrt. content, sentiment, and context | ∼11K tweets used in experiments Fundacio’n Barcelona Media (FBM): Kongregate, Slashdot and MySpace. Totally 10,951 tweets collected and nearly 167 labeled offensive. | Improving accuracy in detecting harassing language using discussion-style and chat-style language |
| [ | Detecting harassers and victims in cyberbullying incidents | Collected twitter data using profane words Twitter data contains 180,355 users and 296,308 tweets. | Accuracy improved wrt. network features. |
| [ | Detecting instigators and victims of bullying | 180K profile on Twitter and ∼300K tweets using profane words as seed | scoring level of cyber bully and victim. |
| [ | Detecting cyber bullying in the Japanese community. | Data from Japanese secondary schools | Automatically extract new vulgarities from the Internet to keep their offensive lexicon up to date. |
| [ | Understanding behavior and actions of individuals using emotion detection | ∼2.5M tweets | tweets dataset using harassment-related and emotion hashtags |
| [ | Detecting bullying incident on social networks | ∼2M tweets collected in 4 weeks | Developed a practical method of text mining, clustering, dimensionality reduction and classification. |
| [ | Classifying cyberbullying activities on social network | Collected data from 18,554 users data from Formspring.Me and MySpace. | predicting cyber bullying using fuzzy logic |
| [ | investigating the correlation of harassment on Facebook | 555 Facebook users in the United States (59% female; Mage = 30.90, SDage = 9.19) | Results show most of the updating posts related to the intellect people, children, and who they are in the romantic relation. |
| [ | Identifying narcissism, activities on Facebook social media | 256 Facebook users from locations around the world. | Text mining for narcissistic using on the comment likes |
| [ | Automatic cyber bulling detection on social media text | English and Dutch corpora from | detecting signals of cyber bulling on social media, about bullies, victims, and bystanders. |
| [ | Decompose the overall detection problem into detection of sensitive topics, lending itself into text classification. | corpus contain 4500 YouTube comments. | Concluded binary classifier for individual labels outperform multiclass classifier. |
| [ | Cyberbulling detection with in multi modal content. | ∼K entries from Instagram and Vine Dataset | proposed cyberbulling detection framwork |
| [ | Identification of fake content in online news. | 980 entries from fakeNewsAMT and celebrity Dataset | Linguistic analysis shows the importance of the lexical, syntactic, and semantic of content. |
Annotation statistics of our categorized corpus.
| Contextual Type | Annotated Tweets | Harassing ✔ | Non-Harassing ✘ |
|---|---|---|---|
| Sexual | 3,855 | 230 | 3,619 |
| Racial | 4,976 | 701 | 4,275 |
| Appearance-related | 4,828 | 678 | 4,150 |
| Intellectual | 4,867 | 811 | 4,056 |
| Political | 5,663 | 699 | 4,964 |
| Combined | 24,189 | 3,119 | 21,070 |
Agreement rate.
| Content Type | Agreement Rate |
|---|---|
| Sexual | 0.70 |
| Racial | 0.84 |
| Appearance-related | 1.00 |
| Intellectual | 0.80 |
| Political | 0.69 |
Statistics for the Golbeck corpus after our annotation wrt. contextual type.
| Contextual Type | #of Tweets |
|---|---|
| Sexual | 380 |
| Racial | 4148 |
| Appearance-related | 145 |
| Intellectual | 381 |
| Political | 163 |
| Non Harassing | 41 |
| Total | 5277 |
Fig 2Significant LIWC features in comparing harassing corpus to non-harassing corpus for six categories.
The extreme red (green) color indicates the significance of a given feature in the harassing corpus (non-harassing corpus). E.g. the negation feature with the value 2.34 in the appearance harassing corpus is significantly higher than non-harassing corpus. The white color indicates a lack of difference for a given feature when comparing two corpora.
Fig 3Top-25 frequent words within each harassing corpora.
Fig 4Top-25 frequent words within each non-harassing corpora.
Percentage of type-dependent of top-15 frequent words within each sub-corpus.
H stands for the harassing corpus and NH stands for the non-harassing corpus.
| Category | Type | Percentage |
|---|---|---|
| H | 66.6% | |
| NH | 93.3% | |
| H | 73.3% | |
| NH | 73.3% | |
| H | 80% | |
| NH | 73.3% | |
| H | 80% | |
| NH | 53.3% | |
| H | 80% | |
| NH | 60% |
Size of the training datasets for each type.
| Category | Number of tweets |
|---|---|
| 1,344 | |
| 1,622 | |
| 1,397 | |
| 1,401 | |
| 461 | |
| 6,225 |
Fig 5Comparative study of the F-score from four major classifiers i.e., SVM stands for support vector machine, KNN = K-Nearest Neighbor, GBM = Gradient Boosting Machine, NB = Naive Bayes, NN = Nueral Network).
Fig 6Comparative study of the various feature settings on the performance of the GBM classifier using measures such as precision, recall, F-score, accuracy, and specificity.
The extreme colors, i.e., purple, yellow, green, olive, and pink show the higher values versus the white color that shows a lower value.
Performance of the GBM binary classifier on the combined corpus.
| Feature | Precision | Recall | F-Score | Accuracy | Specificity |
|---|---|---|---|---|---|
| T | 0.84 | 0.81 | 0.82 | ||
| T+L | 0.9 | 0.87 | 0.88 | ||
| F(S)+L+T | 0.94 | 0.92 | 0.90 | 0.88 | 0.37 |
| F(C)+L+T | 0.94 | 0.88 | 0.86 | 0.84 | 0.44 |
| F(S) | 0.83 | 0.83 | 0.82 | 0.80 | 0.75 |
| F(C) | 0.78 | 0.76 | 0.76 | 0.75 | 0.73 |
| F(S)+L | 0.94 | 0.95 | 0.93 | 0.91 | 0.69 |
| W(S)+L+T | 0.94 | 0.93 | 0.91 | 0.89 | 0.70 |
| W(S)+L | 0.93 | 0.94 | 0.92 | 0.90 | 0.74 |
| F(S)+W(S) | 0.90 | 0.89 | 0.88 | 0.87 | 0.83 |
Performance of our multi-class classifier for predicting type of harassment incident.
| Category | Precision | Recall | F-score |
|---|---|---|---|
| 0.84 | 0.85 | 0.84 | |
| 0.87 | 0.85 | 0.86 | |
| 0.81 | 0.84 | 0.83 | |
| 0.82 | 0.83 | 0.82 | |
| 0.58 | 0.62 | 0.60 | |
| 0.98 | 0.97 | 0.98 | |
| 0.92 | 0.82 | ||
| 0.92 | 0.83 | ||
| 0.92 | 0.82 |
Performance of our classifier for predicting tweets for Golbeck corpus.
| Category | Precision | Recall | F-score | Proportion Rate |
|---|---|---|---|---|
| 0.74 | 0.63 | 0.68 | 2.7% | |
| 0.91 | 0.92 | 0.91 | 7.2% | |
| 0.90 | 0.95 | 0.92 | 3.0% | |
| 0.99 | 0.97 | 0.98 | 78.6% | |
| 0.94 | 0.96 | 0.95 | 7.2% | |
| 0.99 | 0.98 | 0.98 | ||
| 0.97 | 0.91 | |||
| 0.97 | 0.90 | |||
| 0.97 | 0.91 |