| Literature DB >> 35943770 |
Bin Cui1, Jian Wang1, Hongfei Lin1, Yijia Zhang2, Liang Yang1, Bo Xu1.
Abstract
BACKGROUND: Depression detection has recently received attention in the field of natural language processing. The task aims to detect users with depression based on their historical posts on social media. However, existing studies in this area use the entire historical posts of the users and select depression indicator posts. Moreover, these methods fail to effectively extract deep emotional semantic features or simply concatenate emotional representation. To solve this problem, we propose a model to extract deep emotional semantic features and select depression indicator posts based on the emotional states.Entities:
Keywords: depression detection; emotion-based reinforcement; emotional semantic features; sentence-level attention; social media
Year: 2022 PMID: 35943770 PMCID: PMC9399877 DOI: 10.2196/37818
Source DB: PubMed Journal: JMIR Med Inform
Figure 1Sample posts of a depressed user. The posts with red highlights are considered depression indicator posts.
Figure 2Architecture of the Emotion-Based Reinforcement Attention Network (ERAN). LSTM: long short-term memory network; RL: reinforcement learning.
Summary of the data sets.
| Data set | Label | User | Tweets |
|
| Depressed | 1402 | 292,564 |
|
| Nondepressed | >300 million | >10 million |
|
| Nonlabeled | 36,993 | 35,076,677 |
Values of hyperparameters.
| Hyperparameters | Value |
| Word embedding dimension | 300 |
| BiLSTMa hidden units | 200 |
| Dropout rate | 0.5 |
| Batch size | 128 |
| Learning rate | 0.001 |
aBiLSTM: bidirectional long short-term memory network.
Results compared with the baseline models.
| Model | Accuracy | Precision | Recall | |
| NBa [ | 0.636 | 0.724 | 0.623 | 0.588 |
| WDLb [ | 0.761 | 0.763 | 0.762 | 0.762 |
| MSNLc [ | 0.782 | 0.781 | 0.781 | 0.781 |
| MDLd [ | 0.790 | 0.786 | 0.786 | 0.786 |
| LSTMe | 0.797 | 0.812 | 0.813 | 0.812 |
| BiLSTMf | 0.805 | 0.817 | 0.818 | 0.817 |
| BiLSTM (Attg) | 0.817 | 0.828 | 0.828 | 0.828 |
| BERTh (base) [ | 0.845 | 0.883 | 0.825 | 0.853 |
| RoBERTai (base) [ | 0.851 | 0.902 | 0.837 | 0.868 |
| CNNj + RLk [ | 0.871 | 0.871 | 0.871 | 0.871 |
| LSTM + RL [ | 0.870 | 0.872 | 0.870 | 0.871 |
| MDHANl [ | 0.895 | 0.902 | 0.892 | 0.893 |
| ERANm (ours) | 0.906 | 0.912 | 0.897 | 0.904 |
aNB: naïve Bayesian.
bWDL: Wasserstein Dictionary Learning.
cMSNL: Multiple Social Networking Learning.
dMDL: Multimodal Depressive Dictionary Learning.
eLSTM: long short-term memory network.
fBiLSTM: bidirectional long short-term memory network.
gAtt: attention.
hBERT: Bidirectional Encoder Representation from Transformers.
iRoBERTa: Robustly Optimized BERT pre-training Approach.
jCNN: convolutional neural network.
kRL: reinforcement learning.
lMDHAN: multimodal depression detection with hierarchical attention network.
mERAN: emotion-based reinforcement attention network.
Figure 3Results of ablation experiments. Emotion-Based Reinforcement Attention Network (ERAN) is the proposed model, and the remaining three are the models after removing one module of ERAN. Acc: accuracy; EBatt: emotion-based BiLSTM (bidirectional long short-term memory network) attention network; ERN: emotion-based reinforcement learning network; F1: F1-score; P: precision; R: recall; RLAtt: reinforcement learning attention;
Figure 4Comparative results of BiLSTM trained on the selected posts, the unselected posts, and the original posts. Acc: accuracy; BiLSTM: bidirectional long short-term memory network; F1: F1-score; P: precision; R: recall.
Figure 5Examples of attention visualization. Different colors represent different weights. The deeper the color, the greater the weight of the post.