| Literature DB >> 35749157 |
Tahsin Mullick1, Ana Radovic2, Sam Shaaban3, Afsaneh Doryab1.
Abstract
BACKGROUND: Depression levels in adolescents have trended upward over the past several years. According to a 2020 survey by the National Survey on Drug Use and Health, 4.1 million US adolescents have experienced at least one major depressive episode. This number constitutes approximately 16% of adolescents aged 12 to 17 years. However, only 32.3% of adolescents received some form of specialized or nonspecialized treatment. Identifying worsening symptoms earlier using mobile and wearable sensors may lead to earlier intervention. Most studies on predicting depression using sensor-based data are geared toward the adult population. Very few studies look into predicting depression in adolescents.Entities:
Keywords: adolescent; depression; machine learning; mobile phone; uHealth
Year: 2022 PMID: 35749157 PMCID: PMC9270714 DOI: 10.2196/35807
Source DB: PubMed Journal: JMIR Form Res ISSN: 2561-326X
Papers on adolescent mental health prediction and how our work differs from the existing work.
| Paper | Study aim | Methods | Results | Difference from our work |
| Cao et al [ | Investigated the effectiveness of smartphone apps useful in evaluating and monitoring depression symptoms in a clinically depressed adolescent population compared with psychometric instruments (PHQ-9a, HAM-Db, and HAM-Ac); 13 participants aged 12 to 17 years | Used self-evaluation of adolescents and parents with smartphone data to improve predictions of PHQ-9 scores; used the SOLVD app installed only on Android phones; used only linear regressor and support vector regressor with polynomial kernel | Correlation between mood averaged over a 2-week period and biweekly psychometric score from PHQ-9, HAM-D, and HAM-A; combining self-evaluation from both parents and children along with smartphone sensor data resulted in PHQ-9 score prediction accuracy | Our work does not depend on self-evaluation by adolescents and parents to help improve predictions; instead, we consider a system where our reliance is exclusively on the captured sensor values to make predictions of PHQ-9 scores. We used universal and personalized modeling strategies with multiple machine learning algorithms. |
| Maharjan et al [ | StandStrong app used to assess feasibility and acceptability of sensing technologies for maternal depression treatment in low-resource settings for mothers aged between 15 and 25 years | They explored possible explanations for differences in successful data collection by time of day and sensor type along with description of qualitative results to illuminate these differences | The study mainly identified concerns related to technological barriers in passively sensed data collection. | The study was based on passively sensed data collection. It did not perform predictive modeling. The aim was to assess how well the app performed in data collection and the hurdles encountered therein. The study had 11 participants with depression with a mix of young and older participants, whereas our study was focused on adolescents, and all participants had been diagnosed with some form of depression. |
| MacLeod et al [ | Explored whether passively collected smartphone sensor data can be used to predict internalizing symptoms among youths in Canada; participants aged between 10 and 21 years | Self-reports of anxiety, depression, and attention-deficit hyperactivity disorder collected; N=122 for 2 weeks of passively sensed data; CES-DCd and SCAREDe anxiety assessments were used | Depressive symptoms correlated with time spent stationary, less mobility, higher light intensity during the night, and fewer outgoing calls. Anxiety correlated with less time spent stationary, greater mobility, and more time on-screen. Adding passively collected smartphone data to prediction models of internalizing symptoms significantly improved their fit. | This work was primarily focused on establishing correlations between self-reports. The study used passive sensor data to perform linear regressor model fitting for predictions of the CES-DC and SCARED values. Nonlinear modeling approaches were not considered, whereas we have explored and produced better results. |
aPHQ-9: Patient Health Questionnaire-9.
bHAM-D: Hamilton Depression Rating Scale.
cHAM-A: Hamilton Anxiety Rating Scale.
dCES-DC: Center for Epidemiological Studies Depression Scale for Children.
eSCARED: Screen for Child Anxiety Related Disorders.
Figure 1Demographic statistics: (A) gender distribution, (B) race distribution of the adolescents, (C) sexual orientation, (D) depression score distribution for each week of observation, and (E) depression score distribution for each participant.
Figure 2Feature extraction.
Figure 3Machine learning (ML) pipeline comprising exploratory data analysis that includes (A) check for skewness of data, (B) missing value assessment, (C) check of depression level distribution, (D) generation of correlation matrix and removal of features that are highly correlated, (E) k-nearest neighbors (KNN)-based missing value imputation, (F) aggregated data set creation, and (G) nonlinear and linear ML modeling of data.
Figure 4Depression score prediction approach. MAE: mean absolute error; MAPE: mean absolute percentage error; ML: machine learning; MSE: mean squared error; RMSE: root mean squared error.
Figure 5Machine learning approach for predicting change in depression level.
Figure 6Cross-validation strategies: (A) Leave One User Out, (B) Leave Week X Out, (C) Leave One Week One User Instance, and (D) Accumulated Weeks.
Depression score regression resultsa.
|
| LOPOb | LWXOc | ACCUd | LOWOUe |
| MAEf (SD) | 4.46 (0.62) | 3.43 (0.70) | 2.39 (0.10) | 2.53 (0.10) |
| MSEg (SD) | 30.74 (0.41) | 19.0 (0.39) | 10.28 (0.21) | 11.89 (0.25) |
| MAPEh (SD) | 0.55 (0.65) | 0.42 (0.52) | 0.27 (0.15) | 0.29 (0.20) |
| RMSEi (SD) | 5.07 (0.71) | 4.31 (0.65) | 2.83 (0.11) | 2.53 (0.17) |
| Feature set | Fitbit, calls, conversation, screen, location, and Wi-Fi | Calls, conversation, screen, location, and Wi-Fi | Fitbit, calls, screen, and location | Fitbit, calls, conversation, screen, location, and Wi-Fi |
| MLj algorithm | AdaBoost | Random forest | XGBoost | Random forest |
aThe values presented display evaluation metrics for depression score regression models. The best-performing machine learning models were AdaBoost, random forest, and XGBoost.
bLOPO: Leave One Participant Out.
cLWXO: Leave Week X Out.
dACCU: Accumulate Weeks.
eLOWOU: Leave One Week One User Instance Out.
fMAE: mean absolute error.
gMSE: mean squared error.
hMAPE: mean absolute percentage error.
iRMSE: root mean squared error.
jML: machine learning.
Figure 7Confusion matrix of depression levels based on depression score predictions.
Depression score change regression resultsa.
|
| LOPOb | LWXOc | ACCUd | LOWOUe |
| MAEf (SD) | 3.28 (0.70) | 3.24 (0.67) | 3.21 (0.20) | 3.12 (0.15) |
| MSEg (SD) | 21.35 (0.72) | 19.43 (0.63) | 20.13 (0.24) | 20.14 (0.22) |
| MAPEh (SD) | 8.33 (0.55) | 15.79 (0.61) | 13.69 (0.17) | 7.16 (0.20) |
| RMSEi (SD) | 4.2 (0.71) | 4.26 (0.66) | 3.86 (0.18) | 4.48 (0.21) |
| Feature set | Fitbit, calls, conversation, screen, location, and Wi-Fi | Calls, conversation, screen, location, and Wi-Fi | Fitbit, calls, and location | Fitbit, calls, conversation, screen, and location |
| MLj algorithm | AdaBoost | Random forest | XGBoost | Random forest |
aThe values presented display evaluation metrics for depression score regression models. The best-performing machine learning models were AdaBoost, random forest, and XGBoost.
bLOPO: Leave One Participant Out.
cLWXO: Leave Week X Out.
dACCU: Accumulate Weeks.
eLOWOU: Leave One Week One User Instance Out.
fMAE: mean absolute error.
gMSE: mean squared error.
hMAPE: mean absolute percentage error.
iRMSE: root mean squared error.
jML: machine learning.
Figure 8Confusion matrix for change in depression level into 7 classes (−3, −2, –1, 0, 1, 2, and 3) that represent transitions between higher and lower levels of depression.
Figure 9Variation in accuracy with increase in weeks of data trained on with a 2-point moving average to map the trend.
Figure 10Missing data percentage (missing%) versus depression score prediction root mean squared error (RMSE)-normalized.