Literature DB >> 35281704

Bootstrapping semi-supervised annotation method for potential suicidal messages.

Roberto Wellington Acuña Caicedo1,2, José Manuel Gómez Soriano3, Héctor Andrés Melgar Sasieta2.   

Abstract

The suicide of a person is a tragedy that deeply affects families, communities, and countries. According to the standardized rate of suicides per number of inhabitants worldwide, in 2022 there will be approximately about 903,450 suicides and 18,069,000 unconsummated suicides, affecting people of all ages, countries, races, beliefs, social status, economic status, sex, etc. The publication of suicidal intentions by users of social networks has led to the initiation of research processes in this field, to detect them and encourage them not to commit suicide. This study focused on determining a semi-supervised method to populate the Life Corpus, using a bootstrapping technique, to automatically detect and classify texts extracted from social networks and forums related to suicide and depression based on initial supervised samples. To carry out the experiments we used two different classifiers: Support Vector Machine (SVM) (with Bag of Words (BoW) features with and without Term-Frequency/Inverse Document Frequency (Tf/Idf), as a weighted term, and with or without stopwords) and Rasa (with the default feature extraction system). In addition, we performed the experiments using five data collections: Life, Reddit, Life+Reddit, Life_en, and Life_en + Reddit. Using the semi-supervised method, we managed to increase the size of the Life Corpus from 102 to 273 samples with texts from the social network Reddit, in a combination Life+Reddit+BoW_Embeddings, with the SVM classifier, with which a macro f1 value of 0.80 was achieved. These texts were in turn evaluated by annotators manually with a Cohen's Kappa level of agreement of 0.86.
© 2022 The Authors. Published by Elsevier B.V.

Entities:  

Keywords:  Natural language processing; Social networks; Suicidal behavior; Suicidal ideation; Suicide prevention

Year:  2022        PMID: 35281704      PMCID: PMC8913319          DOI: 10.1016/j.invent.2022.100519

Source DB:  PubMed          Journal:  Internet Interv        ISSN: 2214-7829


Introduction

The world had 7925 billion people (Worldometer, n.d.) in February 2022. From that population, 5.31 billion are unique mobile phone users (67.1%), 4.95 billion are internet users (62.5%), and 4.62 billion are active social media users (58.4%) (We Are Social, 2022). Up to 2020, humanity generated 44 zettabytes of digital data and will generate 175 zettabytes until the year 2025, most of it with information created and consumed by users through digital television, interaction with social networks, sending images and videos from camera phones between devices through the internet (IDC Corporate USA, 2012). From this entire digital universe, 33% of the information (over 13,000 exabytes) is tagged data and 77% is untagged data (IDC Corporate USA, 2012). This data offers great opportunities for data analytics, through different types of machine learning: supervised learning (Akpınar et al., 2019), unsupervised learning, or semi-supervised learning (Raschka and Mirjalili, 2019; Reagan et al., 2017). Today, the explosive growth of online social networking services has changed the way people work and share their opinions, ideas, and views (Liang and Dai, 2013; Weng et al., 2010; Liao et al., 2013), regardless of their geographical location or physical limitations (Al-Garadi et al., 2016). Therefore, social networks can be extremely useful for various real-life applications, such as marketing (Leung and Chung, 2014), applications in an e-learning environment (Choudhury and Pattnaik, 2020), or discovering opinions about a particular product (Mircoli et al., 2017), knowing the geographical location, food preferences (Peschel et al., 2019), hobbies, favorite store, political tendencies (Stefanova and Kiryantsev, 2019; Rodriguez et al., 2012) and even identifying socially dangerous people (Stefanova and Kiryantsev, 2019), or those with suicidal intentions (Desmet and Hoste, 2013; O’Dea et al., 2015; Luxton et al., 2012; Velupillai and Hadlaczky, 2019; Zhang et al., 2014; Braithwaite et al., 2016; Egmond and R. D.-C. T. J. of Crisis, 1990). In the framework of suicide, the World Health Organization (WHO) in their report, “Suicide prevention: A global imperative”, of the Mental Health Action Plan 2013–2020, published in 2014, estimated 804,000 suicide deaths occurred worldwide in 2012, representing an annual global age-standardized suicide rate of 11.4 per 100,000 people (903,450 suicides approximately in 2022). For each person who commits suicide, 20 more try to do so. In consequence, the WHO recognizes suicide as a global public health priority, which affects not only rich countries (3 men to each woman) also poor and middle-income countries (1.5 men to each woman) as well (World Health Organization, 2014). Suicide affects all age groups and is highest in persons aged 70 years or over (Jeong et al., 2020; Chang et al., 2018; Santini et al., 2015). Globally suicide is the second leading cause of violent death in the 15–29 years age group (World Health Organization, 2014; Sweeney et al., 2015). Suicide not only involves the individual in a personal way but deeply affects family members and close friends (Cerel et al., 2008; Silenzio et al., 2009). According to this reality in 2013, in the framework of the 66th World Health Assembly, the member states of the WHO pledged, among other issues, to reduce national rates of suicide by 10% before the year 2020. To accomplish this goal, they agreed to develop and put into practice comprehensive national suicide prevention strategies, strengthening their information systems, scientific data, research, and university collaboration on mental health, and paying particular attention to groups with the highest suicide risk, such as homosexuals, lesbians, bisexuals, transsexuals, young people, refugees, migrants, and any other vulnerable group (World Health Organization, 2014; Silenzio et al., 2009). However, despite the good intentions of these organizations, their objectives have not been reached. In general, a suicide victim goes through a period of deep personal suffering, often in silence, before making the fateful decision to end his or her own life, so predicting that someone will commit suicide has been an impossible task (Goldstein et al., 1991; Hughes, 1995; Large and Ryan, 2014; Large and Nielssen, 2012; Paris, 2006). But it is possible to detect factors that contribute to suicide risk, using standard clinical tools operated by well-trained clinicians (Beck et al., 1975; Beck et al., 1979). In addition, computer science, specifically Natural Language Processing (NLP), offers the opportunity to understand indicators of suicidal thoughts through the interaction between computers and human language (Larsen et al., 2015; Pestian and Grupp-Phelan, 2016) when these are expressed in written and spoken forms. In this framework, social networks have provided researchers with new ways to use automated methods to analyze human language (Lacson and Khorasani, 2011), by analyzing the sentiments expressed therein, in written or spoken form by users (Abbasi et al., 2014; Girju and Moldovan, 2002; Asghar, 2016; Cole et al., 2006). Thus, by using automated methods, is possible to better understand an individual's thoughts, feelings, beliefs, behavior, and personality (Schwartz and Ungar, 2015), and to be able to successfully identify suicide notes in newsgroups and social media (Hernandez and Pontes, 2014; Huang et al., 2007; Matykiewicz et al., 2009). Machine learning algorithms have been shown to distinguish between notes written by people who died by suicide and simulated suicide notes better than mental health professionals (71% vs 70%) (Pestian and Matykiewicz, 2008). In addition to suicide notes, microblogging data have been used to build machine learning models to identify users with suicidal sentiments with 90% accuracy (Zhang et al., 2014). The success of sentiment analysis (Pang and L. L.-F. and T. in Information, 2008; Pang and L. L.-P. of the 42nd annual meeting on, 2004; Lai et al., n.d.; Birjali et al., 2017a) relies heavily on the quality of implicit and unspecified information in the data (Chen and De Tseng, 2011; Fielding et al., 2008) that can be extracted from the large stream of information currently available (Birjali et al., 2017b; Barnes, 2007; Lieberman, 2014), which provides a better understanding of thoughts, feelings, beliefs, behaviors, and personality of individuals (Schwartz and Ungar, 2015). With the information collected, compiled, and correctly annotated from oral or written texts, a corpus can be formed (Llisterri, 1999), which can be supervised with labeled data, which is more accurate, but generally consumes a lot of human and computer resources (Bentivogli and Pianta, 2005; Akpınar et al., 2019), or semi-supervised with unlabeled or unknown structure data, which reduces manual work and processing time (Akpınar et al., 2019; Raschka and Mirjalili, 2019), even though, the lower quality of the collected data and the accuracy of the annotations must be taken into account (Ren and Matsumoto, 2016). Through a process of collecting information from social networks or other sources, evaluating it through agreements (Ben-David, 2008; Vieira et al., 2010; Vioules et al., 2018; Tapia et al., 2018; O’Dea et al., 2015; Hallgren, 2012; Fu et al., 2013; Canales and Strapparava, 2016), performed by a small group of experts (Bontcheva et al., 2013; Alameda-Pineda et al., 2013; Karimzadeh and MacEachren, 2019) or by multiple groups of experts through crowdsourcing (Ling et al., 2016; Karimzadeh and MacEachren, 2019), a gold standard corpus can be generated (Gundlapalli et al., 2013; Scheible et al., 2011; José, 2017; Karimzadeh and MacEachren, 2019) and used as training to add new data (Silveira et al., n.d.). Different research groups have created gold standard corpora in other knowledge areas, such as suicidality (Cremades et al., 2017a), medication abuse (O'Connor et al., 2020), depressive symptom and acquired psychosocial stressors (Mowery, 2017), Philippine species (Nguyen et al., 2019), cyberbullying (Van Hee et al., 2018), agriculture (Amorim et al., 2019) to gain a proper understanding of the relevant NLP technology to take full advantage of its capabilities (Lu, 2014). The annotations that are made for the creation of the different corpora are generally supervised, and later in the data expansion process, these annotations are made semi-supervised, or unsupervised (Mowery, 2017; O'Connor et al., 2020; Van Hee et al., 2018; Halike et al., 2020; Amorim et al., 2019; Du et al., 2017; O’Dea et al., 2015), depending on the type of experiments carried out. The data from the experiments carried out are usually focused on a single data source: from Twitter (Jashinsky et al., 2014; Mowery, 2017; O’Dea et al., 2015; Purver et al., 2012; Wu et al., 2019), Weibo (Huang et al., 2014; Zhang et al., 2015), Netlog (Desmet and Hoste, 2018), other microblogs (Guan et al., 2015), and more than one sources (Ling et al., 2016; Cremades et al., 2017a). This article proposes, a semi-supervised learning approach to classify potential suicide messages on social networks automatically. Our goal is to assign texts to the suicide and non-suicide category in the Life Corpus. It would be a starting point towards the semi-automatic annotation of Corpus data to detect suicide messages. Semi-automatic annotations will ease the annotation process and reduce the workload of the annotation team in terms of time and resources invested. This document is organized as follows: Section 2 reviews the research on the use of automatic corpus annotation, Section 3 explains the methodology and resources used to develop this work. Subsequently, in Section 4 the results are presented, in Section 5 the results are discussed; finally, in Section 6 conclusions are made and opportunities for future works are proposed.

Related research

The research carried out on “semi-supervised corpus annotation” shows some developed works, among them the one created by Gupta et al. (Gupta et al., 2018) who studied the problem of mentions of Adverse-Drug-Reaction (ADR) in social networks. They used deep neural networks in their research process, specifically a class of Recurrent Neural Network (RNN) that is long-term memory, based on which they proposed a new RNN model based on semi-supervised learning, which can take advantage of data without tag present on social media. With the semi-supervised ADR extraction method, they obtained an f-measure of 0.75. Brum et al. (Brum and Nunes, 2018) worked in a framework based on semi-supervised learning to extend CasSUL Corpus with unlabeled data. In the experiments, six characteristics were used: a bag of words, negation words, emoticons, emojis, the lexicon of feelings, and the label of part of the voice. As well as the classification algorithms Support Vector Machines (SVM), Naïve Bayes, Logistic Regression, Random Forest, Decision Trees, and Multilayer Perceptron, with which the best results were obtained with the combination of BoW + negation words + emoticons + emojis and feature selection using 200 estimators, entropy as a criterion and without maximum depth, with an f-measure of 0.62. O'Dea et al. (O’Dea et al., 2015) worked on detecting whether the level of concern for a suicide-related post on Twitter could generate a training corpus for automatic learning models and implement an automated computer classifier that could replicate the accuracy of the human coders. The data for the experiments were collected from the social network Twitter, and the overall agreement rate among the human coders was 0.76. The classifiers used were the Support Vector Machine and the Logistic Regression methods. The algorithm with the best performance was Support Vector Machine with Tf/Idf without filter, obtaining an f-measure of 0.67. Gómez (Gómez, 2014) worked on the creation of Life Corpus which is a bilingual text corpus (English and Spanish) oriented to detecting suicide ideation. This corpus was constructed retrieving texts from several social networks. Its quality was measured using mutual annotation agreement, obtaining a moderate agreement Cohen's Kappa of 0.52 in four categories, three risk classes (Possible, Urgent, and Immediate) and one the not risk. Given the imbalance of that corpus by the four categories and the small number of samples for each category, these were grouped into only two: Risk and No risk, to achieve better results in the development of different experiments carried out. In these experiments, it was decided to determine which default classification algorithm of the Weka machine learning and data mining software (Hall et al., 2009) achieved better performance when training the Life Corpus texts, using features such as Part of Speech, Wordnet Synset, and reading all numbers by a keyword, achieving the best results the KStar algorithm with the ROC area metric 0.81 and the f-measure of 0.70. It should be noted that the quality of the corpus developed by O'Dea (O’Dea et al., 2015) and Gómez (Gómez, 2014) was determined by evaluating their texts through an agreement between annotators using Cohen's Kappa method (Cohen, 1960a), which is given by Equation one. While in their articles Gupta (Gupta et al., 2018) and Brum (Brum and Nunes, 2018) do not mention if they used any measure to evaluate the quality of the texts. In previous works with the Life Corpus (Life Corpus, n.d.; José, 2017), different experiments have been carried out (Caicedo et al., 2020; Parraga-Alava et al., 2019). Access to the Life Corpus is free under a Creative Commons license (Life Corpus, n.d.). Therefore, the experiments carried out with this corpus can be replicated or improved.

Methodology

The objective of this research was to determine a semi-supervised method to populate the Life Corpus, using a bootstrapping technique. With this, we have tried to improve the automatic detection and classification of texts extracted from social networks and forums related to suicide and depression based on the Life Corpus. In previous works (Caicedo et al., 2020; Parraga-Alava et al., 2019) with Life Corpus, authors used machine learning techniques to systematically analyze all possible combinations of textual characteristics. They tested 28 supervised classifier algorithms using different corpus features. The study concluded that it would be interesting to increase the corpus to improve the performance. The Life Corpus originally consisted of 102 suicidal messages, (71 texts in English and 31 texts in Spanish) 70 samples (No risk), and 32 texts (Risk), divided into four classes: No Risk, Urgent, Possible, and Immediate, all of them unbalanced (Table 1).
Table 1

Number of samples for each “Alert Level” type.

Alert levelQuantityENES
No risk70 (68.6%)45 (63.4%)25 (80.6%)
Urgent19 (18.6%)15 (21.1%)4 (12.9%)
Possible8 (7.8%)6 (8.5%)2 (6.5%)
Immediate5 (4.9%)5 (7%)0 (0%)
Number of samples for each “Alert Level” type. As the corpus was very small and there were too many categories to obtain statistically significant data (Caicedo et al., 2020), it was decided to merge the three risk classes (Possible, Urgent, and Immediate) into one, keeping the No-Risk class intact, to reduce the imbalance and therefore improve the quality of the experiments. We make the same decision in this work using the two categories (Table 2).
Table 2

Number of samples for each “Alert Level” type.

Alert LevelQuantityENES
No risk70 (68.63%)4525
Risk32 (31.37%)266
Number of samples for each “Alert Level” type. To increase the number of samples in the corpus, we decided to collect texts from the social platform Reddit (Gilbert, 2013), within which is the subreddit “SuicideWatch”, which had 984 texts in English in the extraction date. These texts were extracted using the PRAW library (an acronym for “Python Reddit API Wrapper”) in Python, which allows access to Reddit through a developer account (Reddit, n.d.). After the 984 texts had been extracted, they were preprocessed, eliminating HTML tags. With the original supervised Life Corpus and the new unsupervised Reddit Corpus, we developed the system shown in Fig, 1 to increase the Life Corpus more samples, especially those tagged as Risk.
Fig. 1

System workflow scheme. The system was evaluated using, the original Life Corpus and translated Life Corpus. The system is composed of three processes: i) translation process, ii) bootstrapping Corpus expansion, iii) reviewing, building, and evaluating the final supervised corpus.

System workflow scheme. The system was evaluated using, the original Life Corpus and translated Life Corpus. The system is composed of three processes: i) translation process, ii) bootstrapping Corpus expansion, iii) reviewing, building, and evaluating the final supervised corpus. After this, using the samples of the Life Corpus (English + Spanish translated to English), a classifier was created based on the SVM algorithm and using the characteristics of BoW (Cao et al., 2014) with Tf/Idf terms weighting. This initial classifier was used to increase the corpus with a Bootstrapping Uncertainty Sampling technique. As the Life Corpus was very unbalanced towards samples without risk (Table 2), we were only interested in choosing the samples that the classifier categorized as Risk. In this way, considering that the classifier would have a significant error rate, samples without risk would also be included. In each iteration, the cutoff threshold was increased exponentially to limit the acceptance of new samples comparing that threshold with the confidence score given by the sklearn SVM classifier for each sample. That means if I have the sample X and Y, and the SVM score for the sample X is higher than the threshold, and the score of the sample Y is lower, sample X was added, and the sample Y was rejected in that iteration. Specifically, the threshold grew in each iteration (1 − 0. 2), where n is the iteration number. We repeated this iteration until any of the SVM confidence scores for the evaluated samples overcame the threshold. In each iteration of the bootstrapping, the unsupervised accepted samples were added to the supervised corpus, training the model again with the new samples and repeating the iteration until, as mentioned, the classifier stopped classifying no previously tagged samples as at risk. Table 3, shows the number of corpus samples of each iteration and the new calculated threshold.
Table 3

Number of samples and threshold by iteration.

IterationNum samplesThreshold
11020.8
22250.96
33020.992
Number of samples and threshold by iteration. The bootstrapping process above left us with a corpus of 302 Reddit samples, 200 of these samples were evaluated by six annotators divided into four groups of 50 samples, such that one of the annotators evaluated the samples of the four groups, while five independent annotators annotated only their group of 50 samples. The mutual agreement of 0.86 was reached, representing 171 agreements in the evaluation of the texts (Cohen, 1960b). The annotation results are shown in Table 4.
Table 4

Agreements between reviewers.

GroupReviewerSuicide textRisk/no riskMutual-agreementTP/TNKappa Cohen
1Reviewer GC38/1238/30.82
Reviewer RA47/3
2Reviewer CC46/446/41.00
Reviewer RA46/4
3Reviewer KM45/541/30.88
Reviewer RA47/3
4Reviewer AR32/1826/100.72
Reviewer JG31/19
Reviewer RA35/15
Totales364/87151/200.86
Agreements between reviewers. To ensure the quality of the data, of these 302 samples we kept the 171 in which all the annotators agreed. That is, there was mutual agreement in classifying such samples. These samples were joined to the Life Corpus samples to build five different corpora of different sizes (Table 5).
Table 5

Corpus used in experiments.

CorpusRiskNot riskTotal
Corpora only in English languages
Life3072102
Reddit15318171
Life+Reddit18390273



Corpus in Spanish and English languages
Life_es_en3072102
Life_es_en + Reddit18390273
Corpus used in experiments. The experiments of the present work were carried out with this data. The reasons why we used five corpora were the following: i) the data sources were very heterogeneous and we had to verify that the results did not deteriorate using the separate corpora; and ii) as we can see in Table 2, the Life Corpus contains a mixture of messages in English and Spanish and we wanted to test how automatic translation affected the performance of the system with a corpus where all messages were in the same language with automatic translation, or keeping the original messages. There is no option for the Reddit Corpus with and without translation because all the messages were already in English in the Reddit Corpus. Therefore, we repeat all the experiments with: Life: Only the translated version of Life. Reddit: Only the Reddit Corpus samples with a mutual agreement between the annotators. Life + Reddit: The combination of samples from the Life Corpus translated into English and the samples from Reddit by mutual agreement. Life_es_en: The original and untranslated Life Corpus, with mixed samples in English and Spanish. Life_es_en + Reddit: The original untranslated Life Corpus plus samples with mutual agreement from Reddit. With this combination, we assessed the performance improvement when we added the Reddit Corpus, or how the automated translation of the texts affected this performance. The automatic translation of the texts in Spanish from the Life Corpus to the English language was carried out through the free and unlimited Python GoogleTrans library (Google, n.d.). Once the different corpora were obtained, we used two different classifiers: Support Vector Machine (SVM) and Rasa intent classifier. The SVM has been widely used to classify texts, giving good results in different research processes (Suthaharan, 2016). For this learning machine, its implementation by sklearn was used (Siglidis et al., 2020). Moreover, we also wanted to conduct experiments with deep learning algorithms. Still, due to the small size of the Corpus, we decided to use the Rasa algorithm of Natural Language Understanding (NLU) (Goyal et al., 2008) that enables classifications using language models together with deep learning techniques. With the Rasa algorithm, we use the characteristics that are defined by default in the Lexical Syntactic Featurizer algorithm: low, title, upper, BOS, EOS, digit, pos. low indicates if the term is lowercase or not, title if the word starts with a capital letter, upper if the word is all capitalized, digit if it is a number, pos is Part of Speech, BOS beginning of a sentence, and EOS end of the sentence. On the other hand, the following characteristics were used for the SVM algorithms: Bag of Words (BoW) with and without Term-Frequency/Inverse Document Frequency (Tf/Idf), as term weighter, and with or without stopwords. To improve the coverage of the results, we use word embeddings to expand the terms of each message. For the use of word embeddings, we use the Polyglot library (Al-Rfou, n.d.). When given a term, this library suggests a number n of terms close to such term whose embedding vector is less than a distance d from the vector of the searched term. After several preliminary tests, we decided to set n to a value of 10 and d to 0.85. We evaluate the results with the following metrics: simple accuracy, balanced accuracy, micro f1, macro f1, weighted f1, micro-precision, macro precision, weighted precision, micro recall, macro recall, weighted recall, micro Jaccard, macro Jaccard, weighted Jaccard. However, as the corpus were unbalanced, either towards no risk samples in the case of the Life Corpus or risk samples for the Reddit or Life+Reddit Corpus, we decided to use macro f1 as our primary metric, since it is the one that best responded to the unbalanced corpus, which calculates the f1 statistic, separated by classes and does not use weights for grouping (Overflow, n.d.). The experiments were performed using the 10-fold cross-validation technique and repeated 30 times, with different cross-validation random divisions, to obtain statistical significance using a t-test for mean differences of the values of macro f1.

Results

As we had two variants of the Life Corpus, one with all the samples translated into English and the other keeping the samples in Spanish, we decided to divide the experiments with all the corpora that had a variant and with those that did not. Therefore, for the first part of this section, we conducted the experiments with the Life, Reddit, and Life+Reddit collections with all the texts translated into English. However, in the second part, we used the Life_es_en and Life_es_en + Reddit Corpus, with the Life Corpus untranslated. We did not use only the Reddit Corpus for the second part because it is entirely in English.

Experiments with Corpus in English

In Fig. 2, we can observe the results using the SVM classifier (using the features BoW, BoW+Embeddings, Tf/Idf, and Tf/Idf + Embeddings) and Rasa (with the default features extraction system) for the three different English data collections: Life, Reddit, and Life+Reddit.
Fig. 2

The vertical pointed line is the original f-measure result with Life Corpus.

The vertical pointed line is the original f-measure result with Life Corpus. The results of macro f1 were better with the SVM classifier and BoW as features (without Tf/Idf weighter). There were no statistically significant differences whether we used BoW with or without word embeddings expansion. Therefore, both systems appear to have similar performance. Nevertheless, for smaller corpus, the Rasa system was the best, confirming that this classifier can improve other processes when the number of samples is small including learned language models. In Table 6 we can see the results with more detail and observe the macro precision and macro recall, achieving a value of 0.79 of macro f1. As we can see in the table, the best system has better performance in the three measures (macro f1, macro precision, and macro recall).
Table 6

Macro f1, macro precision, and macro recall. Corpus in the English language combined with the training features. The confidence interval was calculated with p < .01.

FeaturesMacro f1Macro precisionMacro recall
Life
Rasa0.49 ± 0.020.52 ± 0.030.53 ± 0.02
BoW0.43 ± 0.020.40 ± 0.030.50 ± 0.01
BoW + Embedding0.42 ± 0.020.39 ± 0.030.51 ± 0.02
Tf/Idf0.41 ± 0.010.35 ± 0.010.51 ± 0.01
Tf/Idf + Embeddings0.41 ± 0.010.35 ± 0.010.51 ± 0.01



Reddit
Rasa0.55 ± 0.030.54 ± 0.040.57 ± 0.03
BoW0.51 ± 0.020.48 ± 0.030.55 ± 0.02
BoW + Embedding0.50 ± 0.020.47 ± 0.020.54 ± 0.02
Tf/Idf0.52 ± 0.030.50 ± 0.030.55 ± 0.03
Tf/Idf + Embeddings0.52 ± 0.030.49 ± 0.030.55 ± 0.03



Life + Reddit
Rasa0.65 ± 0.010.76 ± 0.020.66 ± 0.01
BoW0.77 ± 0.010.77 ± 0.010.79 ± 0.01
BoW + Embeddings0.79 ± 0.010.80 ± 0.010.79 ± 0.01
Tf/Idf0.53 ± 0.040.72 ± 0.060.58 ± 0.05
Tf/Idf + Embeddings0.51 ± 0.020.68 ± 0.040.56 ± 0.01
Macro f1, macro precision, and macro recall. Corpus in the English language combined with the training features. The confidence interval was calculated with p < .01.

Experiments with Corpus in Spanish and English

As mentioned above, in the second group of experiments, the original Life Corpus was used alone without translating any samples or in combination with Reddit, entirely in English. The objective of these experiments was to observe whether or not the translation affected the results in the classification of messages with suicidal ideation. As in previous experiments, the SVM and Rasa classifiers were used. For the first, the characteristics of BoW were extracted from each corpus, BoW expanded with word embeddings, and its weight variants with Tf/Idf. The result of these experiments can be seen in Fig. 3.
Fig. 3

The vertical pointed line is the original f-measure result with Life Corpus.

The vertical pointed line is the original f-measure result with Life Corpus. Once again, the best results were obtained using the most extensive corpus (Life_es_en + Reddit) with the classifier SVM and extracting the BoW features, with or without word embedding expansion. This system gives a macro f1 of 0.80 (p < .01). The best result for the Life Corpus without translation was using Rasa (0.48). In Table 7, we present the best system for each corpus with better performance in the three measures (macro f1, macro precision, and macro recall).
Table 7

Macro f1, macro precision, and macro recall. Corpus in English and Spanish language combined with training features. The confidence interval was calculated with p < .01.

FeaturesMacro f1Macro precisionMacro recall
Life_es_en
Rasa0.48 ± 0.030.51 ± 0.030.50 ± 0.02
BoW0.40 ± 0.010.35 ± 0.020.49 ± 0.01
BoW + Embeddings0.43 ± 0.020.40 ± 0.040.51 ± 0.01
Tf/Idf0.42 ± 0.020.37 ± 0.020.51 ± 0.01
Tf/Idf + Embeddings0.43 ± 0.020.37 ± 0.020.52 ± 0.01



Life_es_en + Reddit
Rasa0.67 ± 0.020.78 ± 0.020.67 ± 0.01
BoW0.78 ± 0.010.78 ± 0.010.79 ± 0.01
BoW + Embeddings0.80 ± 0.010.81 ± 0.010.81 ± 0.01
Tf/Idf0.63 ± 0.010.78 ± 0.020.64 ± 0.01
Tf/Idf + Embeddings0.62 ± 0.010.76 ± 0.020.63 ± 0.01
Macro f1, macro precision, and macro recall. Corpus in English and Spanish language combined with training features. The confidence interval was calculated with p < .01. Although the biggest corpus precision and recall match the best results, this does not happen with the smaller corpus. The best recall system differs from the best macro f1 system (p < .01). Moreover, if we compare these results with the translated corpus (Table 5), we can observe a slight non-significant difference (0.80 ± 0.01 vs 0.79 ± 0.01 p < .01). This means that the process of automatic translation with the GoogleTrans library of the samples in Spanish from the original corpus to the English language does not significantly worsen the performance of the detection of suicidal messages.

Discussion

Although O'Dea and Gómez developed corpus whose quality was evaluated through agreements, their development methodology was different: the first, one developed a corpus from data downloaded from Twitter in the same period, of which the 14% selected were randomly divided into two data sets to be evaluated by human coders who classified them into three categories: “Very concerning” (14%), “Possibly concerning” (56%) and “Safe to ignore” (29%), with a Cohen's Kappa agreement of 0.76, while the classifier correctly identified 80% of the tweets in the category “Very concerning”. The second, one initially developed a supervised corpus with annotations from different sources that, because it was evaluated with four categories, had a moderate Cohen's Kappa agreement of 0.52 (average k = 0.55), in the process of increasing the size of the corpus samples, a semi-supervised methodology was used with texts from the “SuicideWatch” subreddit, which allowed expanding the number of samples from 102 to 273 (183 risk and 90 No risk) with an agreement between Cohen's Kappa annotators of 0.86. As we have seen in Section 4, the results are promising, being a semi-supervised learning system capable of achieving a macro f1 of 0.78–0.81, close to the mutual agreement reached by human reviewers (Cohen's Kappa of 0.86). These results came when the Life Corpus increased in size by adding the 171 samples from the Reddit Corpus, in which the annotators reached a mutual agreement. These results also show us that the semi-supervised Bootstrapping Uncertainty Sampling methodology chosen to expand the Life Corpus with new samples is valid and useful to improve the results of the automatic system for detecting messages of depression or suicidal ideation. The Rasa NLU classifier works better than SVM for smaller corpus, maybe because it uses some pre-trained language models. However, BoW and SVM work better than the Rasa deep learning approach. Expanding the text using word embeddings does not affect significative the results (p < .01) or the automatic translation of the texts in Spanish from the Life Corpus to the English language using the GoogleTrans library. On the other hand, before using this methodology we assumed that there would not be so many suicidal messages in the Reddit subgroup and that the initial classifier could not find so many messages of this type. Therefore, the final corpus will be more balanced. However, the initial classifier for the bootstrapping techniques using a preliminary SVM classifier worked better than we expected. This has led us to move from a Life Corpus where most posts had no suicide risk to another Life_es_en + Reddit Corpus, where most posts, according to reviewers, had clues of suicidal ideation or depression (Table 5).

Conclusions and future work

Because they have been annotated through a supervised methodology (Egmond and R. D.-C. T. J. of Crisis, 1990; Barraclough and Hughes, 1987; Huang et al., 2014), the suicide corpus population has been costly (Mircoli et al., 2017; Akpınar et al., 2019; Cremades et al., 2017b; Priyanthan et al., 2012). The objective of this research was to increase the number of samples of the Life Corpus (Liu et al., n.d.) using a semi-supervised method (Komiya et al., 2018; Braithwaite et al., 2016), which allowed to maintain the quality of the added texts, reducing the human effort. In this work we have demonstrated two things: i) Bootstrapping Uncertainty Sampling technique used in the present work can be helpful to increase a corpus suitable for suicide prevention using supervised machine learning approaches, and ii) the expanded Life Corpus can be supportive to build a classifier in which messages of depressive or suicide ideation can be detected almost as well as the human mutual agreement. In future work, we plan to experiment with other classification algorithms and new features, to optimize the semi-supervised annotation methodology of new samples, from microblogging networks, blogs, forums, or other sources for the Life Corpus. To constantly increase the number of samples of this corpus, to carry out more and more exact searches of potential suicidal users in social networks. Moreover, we want to explore more the deep learning algorithms using other Rasa features or parameters, and other NLU classifiers or technologies such as BERT embeddings. On the other hand, we want to change the use of word embeddings to use their vectors directly instead of expanding text terms with neighbor terms or to try the use of sentence embeddings. Likewise, we intend to validate the semi-supervised annotation methodology that in this research, it was used samples from microblogging networks, blogs, forums, or other sources, with samples from actual clinical diagnoses, based on psychiatric epidemiological studies to determine its effectiveness in the real world and to be able to more accurately predict the probability of future thoughts of death, suicidal ideation, suicide plan or suicide attempts. In addition, we plan to test different machine translation methodologies, such as the one provided by Google's translator through the GoogleTrans library, to generate parallel corpus expanded into different languages, from the original texts written primarily in the English language to do experiments for the detection of suicidal users in social networks in other languages, carried out by our research group or by other groups, which will be able to access this parallel corpus of the Life Corpus for free under the Creative Commons license in https://github.com/PlataformaLifeUA.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  34 in total

1.  Social media and suicide: a public health perspective.

Authors:  David D Luxton; Jennifer D June; Jonathan M Fairall
Journal:  Am J Public Health       Date:  2012-03-08       Impact factor: 9.308

2.  Suicidal ideation and later suicide.

Authors:  Matthew M Large; Olav Nielssen
Journal:  Am J Psychiatry       Date:  2012-06       Impact factor: 18.112

3.  The use of technology in Suicide Prevention.

Authors:  Mark E Larsen; Nicholas Cummins; Tjeerd W Boonstra; Bridianne O'Dea; Joe Tighe; Jennifer Nicholas; Fiona Shand; Julien Epps; Helen Christensen
Journal:  Conf Proc IEEE Eng Med Biol Soc       Date:  2015

4.  Classification of suicidal behaviors: I. Quantifying intent and medical lethality.

Authors:  A T Beck; R Beck; M Kovacs
Journal:  Am J Psychiatry       Date:  1975-03       Impact factor: 18.112

5.  Tracking suicide risk factors through Twitter in the US.

Authors:  Jared Jashinsky; Scott H Burton; Carl L Hanson; Josh West; Christophe Giraud-Carrier; Michael D Barnes; Trenton Argyle
Journal:  Crisis       Date:  2014

6.  Suicide risk assessment: myth and reality.

Authors:  M Large; C Ryan
Journal:  Int J Clin Pract       Date:  2014-06       Impact factor: 2.503

7.  Assessment of suicidal intention: the Scale for Suicide Ideation.

Authors:  A T Beck; M Kovacs; A Weissman
Journal:  J Consult Clin Psychol       Date:  1979-04

Review 8.  Predicting and preventing suicide: do we know enough to do either?

Authors:  Joel Paris
Journal:  Harv Rev Psychiatry       Date:  2006 Sep-Oct       Impact factor: 3.732

9.  Computing Inter-Rater Reliability for Observational Data: An Overview and Tutorial.

Authors:  Kevin A Hallgren
Journal:  Tutor Quant Methods Psychol       Date:  2012

10.  Semi-Supervised Recurrent Neural Network for Adverse Drug Reaction mention extraction.

Authors:  Shashank Gupta; Sachin Pawar; Nitin Ramrakhiyani; Girish Keshav Palshikar; Vasudeva Varma
Journal:  BMC Bioinformatics       Date:  2018-06-13       Impact factor: 3.169

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.