| Literature DB >> 28663166 |
Akkapon Wongkoblap1, Miguel A Vadillo2,3, Vasa Curcin1,2.
Abstract
BACKGROUND: Mental illness is quickly becoming one of the most prevalent public health problems worldwide. Social network platforms, where users can express their emotions, feelings, and thoughts, are a valuable source of data for researching mental health, and techniques based on machine learning are increasingly used for this purpose.Entities:
Keywords: anxiety; artificial intelligence; depression; infodemiology; machine learning; mental disorders; mental health; public health informatics; social networking
Mesh:
Year: 2017 PMID: 28663166 PMCID: PMC5509952 DOI: 10.2196/jmir.7215
Source DB: PubMed Journal: J Med Internet Res ISSN: 1438-8871 Impact factor: 5.428
Figure 1Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) flow diagram. CLPsych: Computational Linguistics and Clinical Psychology Workshops.
Summaries of articles reviewed.
| First author, date, reference | Aims | Findings |
| Wang, 2017 [ | To explore and characterize the structure of the community of people with eating disorders using Twitter data and then classify users into those with and without the disorder. | There was assortativity among users with eating disorder. The classifier distinguished 2 groups of people. |
| Volkava, 2016 [ | To explore academic discourse from tweets and build predictive models to analyze the data. | Tweets from students across 44 universities were related to student surveys on satisfaction and happiness. |
| Saravia, 2016 [ | To present a new data collection method and classify individuals with mental illness and nonmental illness. | The proposed method and a classifier were built as an online system, which distinguished 2 groups of individuals and provided mental illness information. |
| Kang, 2016 [ | To propose classification models to detect tweets of users with depression for a long period of time. Classifiers were based on the texts, emoticons, and images they posted. | The models detected users with depression. |
| Schwartz, 2016 [ | To present predictive models to estimate individual well-being through textual content on social networks. | A combination of message- and user-level aggregation of posts performed well. |
| Chancellor, 2016 [ | To explore posts from Instagram to forecast levels of mental illness severity of pro-eating disorder. | Future mental illness severity could be predicted from user-generated messages. |
| Braithwaite, 2016 [ | To explore machine learning algorithms to measure suicide risk in the United States. | Machine learning algorithms successfully classified users with suicidal ideation. |
| Coppersmith, 2016 [ | To explore linguistics and emotional patterns in Twitter users with and without suicide attempt. | There were quantifiable signals of suicide attempt in tweets. |
| Lv, 2015 [ | To build a Chinese suicide dictionary, based on Weibo posts, to detect suicide risk. | The Chinese suicide dictionary detected individuals and tweets at suicide risk. |
| O’Dea, 2015 [ | To explore machine learning models to automatically detect the level of concern for each suicide-related tweet. | Machine learning classifiers estimated the level of concern from suicide-related tweets. |
| Liu, 2015 [ | To investigate and predict users’ subjective well-being based on Facebook posts. | Users’ subjective well-being could be predicted from posts and their time frame. |
| Burnap, 2015 [ | To explore suicide-related tweets to understand users’ communications on social media. | Classification models classified tweets into relevant suicide categories. |
| Park, 2015 [ | To analyze the relationships between Facebook activities and the depression state of users. | Participants with depression had fewer interactions, such as receiving likes and comments. Depressed users posted at a higher rate. |
| Hu, 2015 [ | To present classifiers with different lengths of observation time to detect depressed users. | Behavioral and linguistic features predicted depression. A 2-month period of observation enabled prediction cues of depression half a month in advance. |
| Tsugawa, 2015 [ | To develop a model to recognize individuals with depression from non-English social media posts and activities. | Activities extracted from Twitter were useful to detect depression; 2 months of observation data enabled detection of symptoms of depression. The topics estimated by LDAa were useful. |
| Zhang, 2015 [ | To explore 2 natural language processing algorithms to identify posts predicting the probability of suicide. | LDA automatically detected suicide probability from textual contents on social media. |
| Coppersmith, 2015 [ | To explore tweet content with self-reported health sentences and language differences in 10 mental health conditions. | There were quantifiable signals of 10 mental health conditions in social network messages and relations between them. |
| Preotiuc-Pietro, 2015 [ | To implement linear classifiers to detect users with PTSDc and depression based on user metadata, and several textual and topic features. | The combination of linear classifiers performed better than average classifiers. All unigram features performed well. |
| Mitchell, 2015 [ | To use several natural language processing techniques to explore the language of schizophrenic users on Twitter. | Character ngram features were used to train models to classify users with and without schizophrenia. LDA outperformed linguistic inquiry and word count. |
| Preotiuc-Pietro, 2015 [ | To study differences in language use in tweets about mental health depending on the role of personality, age, and sex of users. | Personality and demographic data extracted from tweets detected users with depression or PTSD. |
| Pedersen, 2015 [ | To explore and study the accuracy of decision lists of ngrams to classify users with depression and PTSD. | Bigram features underperformed ngram 1-6 features. |
| Resnik, 2015 [ | To build classifiers to categorize depressed and nondepressed users, based on supervised topic models. | LDA mined useful information from tweets. Supervised topic models such as supervised LDA and supervised anchor model improved LDA accuracy. |
| Resnik, 2015 [ | To build classifiers with TF-IDFd weighting, using support vector machine with a linear kernel or radial basis function kernel. | TF-IDF showed good performance, and TF-IDF with supervised topic model performed even better. |
| Durahim, 2015 [ | To explore data from social networks to measure the Gross National Happiness of Turkey. | Sentiment analysis estimated Gross National Happiness levels similar to Turkish statistics. |
| Guan, 2015 [ | To explore 2 types of classifiers to detect posts revealing high suicide risk. | Users’ profiles and their generated text were used to classify users with high or low suicide risk. |
| Landeiro Dos Reis, 2015 [ | To explore exercise-related tweets to measure their association with mental health. | Users who posted workouts regularly tended to express lower levels of depression and anxiety. |
| De Choudhury, 2014 [ | To explore several types of Facebook data to detect and predict postpartum depression. | Postpartum depression was predicted from an increase of social isolation and a decrease of social capital. |
| Huang, 2014 [ | To present a framework to detect posts related to suicidal ideation. | The best predictive model was based on support vector machine. |
| Wilson, 2014 [ | To explore the types of mental health information posted and shared on Twitter. | The study distinguished 8 themes of information about depression in Twitter posts, each having different features. |
| Coppersmith, 2014 [ | To present a novel method to collect posts related to PTSD and build a classifier. | The classifier distinguished users with and without self-reported PTSD. |
| Kuang, 2014 [ | To create the Chinese version of the extended PERMAe corpus and use it to measure happiness scores. | The proposed model measured happiness. |
| Hao, 2014 [ | To propose machine learning models to measure subjective well-being of social media users. | The model measured subjective well-being from social media data. |
| Prieto, 2014 [ | To develop a machine learning model to detect and measure the prevalence of health conditions. | The proposed methods identified the presence of health conditions on Twitter. |
| Lin, 2014 [ | To develop a deep neural network model to classify users with or without stress. | The trained model detected stress from user-generated content. |
| Schwartz, 2014 [ | To build predictive models to detect depression based on Facebook text. | Facebook updates enabled distinguishing depressed users. Predictive models offered insights into seasonal affective disorder. |
| Coppersmith, 2014 [ | To analyze tweets related to health and propose a new method to quickly collect public tweets containing statements of mental illnesses. | There were differences in quantifiable linguistic signals of bipolar disorder, depression, PTSD, and seasonal affective disorder in tweets. |
| Homan, 2014 [ | To examine the potential of tweet content to classify suicidal risk factors. | Annotations from novices and experts were used to train classifiers, although expert annotations outperformed novice annotations. |
| Park, 2013 [ | To develop a Web app to detect symptoms of depression from features extracted from Facebook. | Depressed users had fewer Facebook friends, used fewer location tags, and tended to have fewer interactions. |
| Wang, 2013 [ | To build a depression detection model based on sentiment analysis of data from social media. | Sentiment analysis with 10 features detected users with depression, with 80% accuracy. |
| Wang, 2013 [ | To explore a detection model, based on node and linkage features, to recognize the presence of depression in social media users. This was an extended version of their earlier study [ | The node and linkage features model performed better than the model based just on node features. |
| Tsugawa, 2013 [ | To explore the effectiveness of an analytic model to estimate depressive tendencies from users’ activities on a social network. | There was a correlation between the Zung Self-Rating Depression Scale and the model estimations. |
| De Choudhury, 2013 [ | To explore predictive models to classify mothers with a tendency to change behavior after giving birth or to experience postpartum depression. | Tweets during prenatal and early postnatal periods predicted future behavior changes, with an accuracy of 71%. Data over 2-3 weeks after giving birth improved prediction results, with an accuracy of 80%-83%. |
| De Choudhury, 2013 [ | To explore the potential of a machine learning model to measure levels of depression in populations. | The proposed model estimated levels of depression. |
| De Choudhury, 2013 [ | To develop a prediction model to classify individual users with depression. | The predictive model classified users with depression. |
| Schwartz, 2013 [ | To analyze tweets from different US counties to predict well-being of people in those areas. | Topic features provided useful information about life satisfaction. |
| Hao, 2013 [ | To explore the mental state of users through their online behavior. | Online behavior enabled prediction of mental health problems. |
| Jamison-Powell, 2012 [ | To explore the characteristics of tweets that included the #insomnia hashtag. | Tweets about insomnia contained more negative words. People used Twitter to express their symptoms and ideas for coping strategies. |
| Bollen, 2011 [ | To explore an online social network to measure subjective well-being levels of users and calculated assortativity. | There was assortativity among Twitter users. |
aLDA: latent Dirichlet allocation.
bCLPsych: Computational Linguistics and Clinical Psychology Workshops.
cPTSD: posttraumatic stress disorder.
dTF-IDF: term frequency-inverse document frequency.
ePERMA: positive emotions, engagement, relationships, meaning, and accomplishment.
Figure 2Conceptual view of social network-based mental health research. CBT: cognitive behavioral therapy.