| Literature DB >> 34518741 |
Ramin Safa1, Peyman Bayat1, Leila Moghtader2.
Abstract
Depression is the most prevalent mental disorder that can lead to suicide. Due to the tendency of people to share their thoughts on social platforms, social data contain valuable information that can be used to identify user's psychological states. In this paper, we provide an automated approach to collect and evaluate tweets based on self-reported statements and present a novel multimodal framework to predict depression symptoms from user profiles. We used n-gram language models, LIWC dictionaries, automatic image tagging, and bag-of-visual-words. We consider the correlation-based feature selection and nine different classifiers with standard evaluation metrics to assess the effectiveness of the method. Based on the analysis, the tweets and bio-text alone showed 91% and 83% accuracy in predicting depressive symptoms, respectively, which seems to be an acceptable result. We also believe performance improvements can be achieved by limiting the user domain or presence of clinical information.Entities:
Keywords: Depression detection; Mental health; Multimodal framework; Social media; Text mining
Year: 2021 PMID: 34518741 PMCID: PMC8426595 DOI: 10.1007/s11227-021-04040-8
Source DB: PubMed Journal: J Supercomput ISSN: 0920-8542 Impact factor: 2.557
Fig. 1The growing number of US youths with a major depressive episode from 2004 to 2019, by gender [6]
Summaries of some recent important studies
| First author, date, reference | Aims | Collection method | Studied feature(s) | ML model(s) /Approach(es) | Metrics | SM platform |
|---|---|---|---|---|---|---|
| Our work | Presenting a multimodal automated framework for predicting potential depressed users from profile information | Self-report statements | Textual tweets, bio-text, profile picture, and banner image | SVM, LR, DT, Gradient Boosting, RF, RidgeClassifier, AdaBoost, Catboost, and Multilayer Perceptron | Precision, Recall, F1-score, Accuracy, and AUC | |
| Zhou, 2021, [ | Studying community depression dynamics due to COVID-19 pandemic in Australia | Self-report statements according to a specific location | Textual tweets | LR, Linear Discriminant Analysis, and NB | Precision, Recall, F1-score, Accuracy | |
| Ríssola, 2020, [ | Presenting a textual dataset on automatic detection of depression | CLEF eRisk 2018 dataset | Textual posts | LR | Precision, Recall F1-score, and AUC | |
| Kim, 2020, [ | Developing a deep learning model to identify user’s mental state | Manual analysis and characterization of subreddits | Textual posts | XGBoost and Convolutional Neural Network | Precision, Recall, F1-score, Accuracy | |
| Guntuku, 2019, [ | Finding which attributes of profile and posted images are associated with depression and anxiety | Survey-reported depression and anxiety (BDI and BAI) | Profile picture and posted images | Pearson Correlation | N/A | |
| Tadesse, 2019, [ | Examining users’ posts to detect factors that may reveal the depression | Manual analysis and characterization of subreddits | Textual posts | LR, SVM, AdaBoost, RF, Multilayer Perceptron | Precision, Recall, F1-score, Accuracy | |
| Islam, 2018, [ | Aiming to perform depression analysis on Facebook data | Depression/non-depression indicative comments | Textual posts | DT, KNN, SVM, and Ensemble | Precision, Recall F1-score | |
| Ferwerda, 2018, [ | Finding the relationship between the content of the uploaded Instagram pictures and the personality traits of users | Survey-reported personality traits (BFI) | Photos | K-means Clustering and Spearman’s Correlation Analysis | N/A | |
| Chen, 2018, [ | Employing fine-grained emotions as features in the task of identifying users with bipolar, depression, PTSD, and SAD disorder | Self-report statements | Textual tweets | RF, SVM, NB, LR, and DT | Precision, Recall, F1-score, and Accuracy |
RF = Random Forest, SVM = Support Vector Machine, NB = Naïve Bayes, LR = Logistic Regression, DT = Decision Tree, AUC = Area Under the ROC Curve
Fig. 2The high-level architecture of the proposed framework consists of three main modules: 1) data collection and dataset building, 2) cross-analysis, and 3) classification
Fig. 3The confirmed range of the polarity score, after initial filtering
Fig. 4The structure of the term-document in the case study (T = Term, D = Document, W = Weight)
Two samples of header images with their top-10 predicted tags
| Labels | Header images | Top-10 Predicted tags |
|---|---|---|
|
| astronaut, aviator, man, person, male, people, helmet, professional, worker, happy | |
|
| human, person, black, man, art, horror, male, cartoon, figure, skeleton |
Top-20 tags of profile picture and banner image from the diagnosed and control groups, sorted by difference absolute value
| Feature Type | Distinct tags (sorted by difference absolute value) |
|---|---|
| Profile Picture | male, man, hair, pretty, attractive, symbol, sign, design, icon, face, portrait, people, graphic, cartoon, model, fashion, child, happy, eyes |
| Header Image | person, building, city, man, sea, water, architecture, symbol, texture, people, blank, sign, word, structure, ocean, border, sky, icon, happy |
Fig. 5The proportion of distinct LIWC categories
Definition of some frequent LIWC features
| Feature | Definition | Examples |
|---|---|---|
| i | First person singular | I, me, mine |
| prep | Prepositions | to, with, above |
| drives | Drives and needs | Power, Risk focus |
| relativ | Relativity | area, bend, exit |
| social | Social processes | mate, talk, they |
| bio | Biological processes | eat, blood, pain |
| adj | Common adjectives | free, happy, long |
| affect | Affective processes | happy, cried |
| posemo | Positive emotion | love, nice, sweet |
| cogproc | Cognitive processes | cause, know, ought |
| sad | Sadness | crying, grief, sad |
| percept | Perceptual processes | look, heard, feeling |
| ingest | Ingestion | dish, eat, pizza |
| achiev | Achievement | win, success, better |
Fig. 6Unigrams and bigrams word cloud
Different types of approaches to text encoding methods
| Feature Type | Methods | # of Main Features | # of Selected Features via CA | # of Selected Features via SVD |
|---|---|---|---|---|
| LIWC | Tweet | 73 | 22 | 15 |
| Bio-description | 73 | 11 | 31 | |
| Profile Picture | 73 | 12 | 15 | |
| Header Image | 73 | 18 | 17 | |
| Word1, 2 g | Tweet | 3000, 3000 | 2942, 2610 | 590, 664 |
| Bio-description | 3000, 3000 | 411, 1165 | 716, 735 | |
| Char2, 4 g | Tweet | 3000, 3000 | 2750, 2400 | 445, 497 |
| Bio-description | 1398, 3000 | 875, 497 | 339, 618 | |
| Tagger | Profile Picture | 84 | 23 | 44 |
| Header Image | 111 | 30 | 62 |
Char = Character, CA. = Correlation analysis
Fig. 7The Pearson correlation heatmap among the LIWC dictionaries for both and
Performance metrics
| Relevant | Non-Relevant | |
|---|---|---|
| Retrieved | TP (True Positive) | FN (False Negative) |
| Not Retrieved | FP (False Positive) | TN (True Negative) |
Top-5 comparison of different methods using different features
| Classifier | Feature | Methods | Prec | Recall | Acc | F1 | AUC |
|---|---|---|---|---|---|---|---|
| Logistic Regression | T | LIWC | 0.65 | 0.72 | 0.67 | 0.68 | 0.67 |
| Char2 | 0.66 | 0.62 | 0.65 | 0.64 | 0.65 | ||
| Char4 | 0.71 | 0.70 | 0.71 | 0.70 | 0.71 | ||
| Word1 | 0.74 | 0.76 | 0.75 | 0.75 | 0.75 | ||
| Word2 | 0.74 | 0.75 | 0.75 | 0.75 | 0.75 | ||
| B | LIWC | 0.62 | 0.55 | 0.61 | 0.58 | 0.61 | |
| Char2 | 0.58 | 0.58 | 0.58 | 0.58 | 0.58 | ||
| Char4 | 0.72 | 0.55 | 0.67 | 0.62 | 0.67 | ||
| Word1 | 0.70 | 0.44 | 0.63 | 0.54 | 0.63 | ||
| Word2 | 0.57 | 0.28 | 0.54 | 0.38 | 0.54 | ||
| P | LIWC | 0.60 | 0.59 | 0.61 | 0.59 | 0.61 | |
| Tags | 0.56 | 0.55 | 0.57 | 0.56 | 0.57 | ||
| H | LIWC | 0.51 | 0.46 | 0.52 | 0.48 | 0.51 | |
| Tags | 0.50 | 0.52 | 0.51 | 0.51 | 0.51 | ||
| Gradient Boosting Classifier | T | LIWC | 0.72 | 0.73 | 0.73 | 0.72 | 0.73 |
| Char2 | 0.65 | 0.62 | 0.65 | 0.63 | 0.65 | ||
| Char4 | 0.73 | 0.69 | 0.72 | 0.71 | 0.72 | ||
| Word1 | 0.89 | 0.86 | 0.88 | 0.87 | 0.88 | ||
| Word2 | 0.97 | 0.84 | 0.91 | 0.89 | 0.91 | ||
| B | LIWC | 0.58 | 0.58 | 0.59 | 0.58 | 0.59 | |
| Char2 | 0.59 | 0.58 | 0.60 | 0.59 | 0.60 | ||
| Char4 | 0.68 | 0.52 | 0.64 | 0.59 | 0.64 | ||
| Word1 | 0.67 | 0.36 | 0.60 | 0.47 | 0.59 | ||
| Word2 | 0.65 | 0.17 | 0.55 | 0.27 | 0.54 | ||
| P | LIWC | 0.59 | 0.55 | 0.59 | 0.57 | 0.59 | |
| Tags | 0.57 | 0.44 | 0.56 | 0.50 | 0.56 | ||
| H | LIWC | 0.50 | 0.41 | 0.50 | 0.45 | 0.50 | |
| Tags | 0.50 | 0.43 | 0.51 | 0.46 | 0.51 | ||
| RidgeClassifier | T | LIWC | 0.68 | 0.77 | 0.70 | 0.72 | 0.70 |
| Char2 | 0.64 | 0.63 | 0.64 | 0.63 | 0.64 | ||
| Char4 | 0.70 | 0.71 | 0.71 | 0.70 | 0.71 | ||
| Word1 | 0.74 | 0.73 | 0.74 | 0.73 | 0.74 | ||
| Word2 | 0.77 | 0.75 | 0.77 | 0.76 | 0.77 | ||
| B | LIWC | 0.62 | 0.55 | 0.61 | 0.58 | 0.61 | |
| Char2 | 0.56 | 0.58 | 0.57 | 0.57 | 0.57 | ||
| Char4 | 0.74 | 0.63 | 0.71 | 0.68 | 0.71 | ||
| Word1 | 0.70 | 0.50 | 0.65 | 0.58 | 0.64 | ||
| Word2 | 0.57 | 0.28 | 0.54 | 0.37 | 0.54 | ||
| P | LIWC | 0.61 | 0.59 | 0.61 | 0.60 | 0.61 | |
| Tags | 0.58 | 0.56 | 0.58 | 0.57 | 0.58 | ||
| H | LIWC | 0.52 | 0.46 | 0.52 | 0.49 | 0.52 | |
| Tags | 0.50 | 0.51 | 0.50 | 0.50 | 0.50 | ||
| Catboost | T | LIWC | 0.71 | 0.75 | 0.73 | 0.73 | 0.73 |
| Char2 | 0.65 | 0.63 | 0.65 | 0.64 | 0.65 | ||
| Char4 | 0.75 | 0.72 | 0.74 | 0.73 | 0.74 | ||
| Word1 | 0.89 | 0.86 | 0.88 | 0.87 | 0.88 | ||
| Word2 | 0.99 | 0.82 | 0.91 | 0.89 | 0.91 | ||
| B | LIWC | 0.58 | 0.58 | 0.58 | 0.57 | 0.58 | |
| Char2 | 0.59 | 0.58 | 0.59 | 0.58 | 0.59 | ||
| Char4 | 0.69 | 0.58 | 0.67 | 0.63 | 0.66 | ||
| Word1 | 0.65 | 0.40 | 0.60 | 0.50 | 0.60 | ||
| Word2 | 0.58 | 0.18 | 0.53 | 0.27 | 0.53 | ||
| P | LIWC | 0.58 | 0.54 | 0.58 | 0.56 | 0.58 | |
| Tags | 0.57 | 0.48 | 0.57 | 0.52 | 0.57 | ||
| H | LIWC | 0.51 | 0.43 | 0.51 | 0.47 | 0.51 | |
| Tags | 0.52 | 0.48 | 0.52 | 0.49 | 0.52 | ||
| MLP | T | LIWC | 0.71 | 0.77 | 0.73 | 0.74 | 0.73 |
| Char2 | 0.60 | 0.62 | 0.61 | 0.61 | 0.61 | ||
| Char4 | 0.68 | 0.69 | 0.68 | 0.68 | 0.68 | ||
| Word1 | 0.69 | 0.72 | 0.70 | 0.70 | 0.70 | ||
| Word2 | 0.69 | 0.72 | 0.70 | 0.70 | 0.70 | ||
| B | LIWC | 0.61 | 0.54 | 0.60 | 0.57 | 0.60 | |
| Char2 | 0.52 | 0.52 | 0.53 | 0.52 | 0.53 | ||
| Char4 | 0.86 | 0.79 | 0.83 | 0.82 | 0.83 | ||
| Word1 | 0.76 | 0.65 | 0.72 | 0.69 | 0.72 | ||
| Word2 | 0.51 | 0.79 | 0.52 | 0.62 | 0.53 | ||
| P | LIWC | 0.60 | 0.59 | 0.61 | 0.60 | 0.61 | |
| Tags | 0.58 | 0.52 | 0.58 | 0.55 | 0.58 | ||
| H | LIWC | 0.53 | 0.42 | 0.53 | 0.47 | 0.53 | |
| Tags | 0.51 | 0.51 | 0.52 | 0.51 | 0.52 |
Acc. = Accuracy, Prec. = Precision, T = Tweets, B = Bio-description, P = Profile picture, H = Header image
Fig. 8Comparison of the highest accuracy achieved on feature types by nine different classifiers
Fig. 9Comparison of the highest accuracy achieved by tweets analyzing methods
Fig. 10Comparison of the highest accuracy achieved by bios analyzing methods
Fig. 11ROC curves of the target classifiers
Fig. 12Comparison of the highest accuracy achieved by tweets analyzing methods via SVD approach
Fig. 13ROC curves of the target classifiers for SVD analysis