Literature DB >> 34401386

Meta-analytic review of online guided self-help interventions for depressive symptoms among college students.

Liang Ma1, Cong Huang2,3, Rimeng Tao1, Zeshi Cui4, Philip Schluter5.   

Abstract

OBJECTIVE: This meta-analysis examines the effect of online guided self-help interventions for depressive symptoms among college students.
METHODS: We searched studies through PubMed, Embase, Web of Science, PsycINFO, and Cochrane Central. Effect estimates were reported as standardized mean differences (SMD) and data were pooled using random-effects models. Subgroup analyses were conducted to investigate the differential effects of these interventions by sample type, level of contact, use of incentive, length of intervention, and program content.
RESULTS: 24 comparisons (n = 3074) deriving from 19 trials were included in the meta-analysis. Intervention participants (n = 1620) indicated significant reductions in depressive symptoms at post-intervention compared to non-active control conditions (n = 1454). The weighted effect size was 0.46 (95% CI: 0.28-0.64), which dropped to 0.36 (95% CI: 0.26-0.45) after an outlier was removed. Subgroup analyses showed that the effects were significant among interventions using both selective and universal samples; among interventions of shorter (≤4 weeks), moderate (4-8 weeks), and greater length (≥8 weeks); among interventions with high, moderate, and low levels of contact; among interventions with and without incentive; and among interventions employing cognitive-behavioral therapy (CBT) and third-wave CBT.
CONCLUSION: This meta-analysis reinforces evidence to support the effectiveness of online guided self-help interventions in reducing depressive symptoms among college students. However, because of the generally variable and limited quality of current evidence, further research applying rigorous methods is needed to confirm and extend the findings of this meta-analysis.
© 2021 Published by Elsevier B.V.

Entities:  

Keywords:  College students; Depressive symptoms; Guided; Internet; Meta-analysis

Year:  2021        PMID: 34401386      PMCID: PMC8350612          DOI: 10.1016/j.invent.2021.100427

Source DB:  PubMed          Journal:  Internet Interv        ISSN: 2214-7829


Introduction

The college years are characterized by a developmentally challenging transition from adolescence to adulthood, which represents a peak age period for depression onset (Cuijpers et al., 2016; Ibrahim et al., 2013; Zivin et al., 2009). Elevated levels of depressive symptoms are common among college students, leading to considerable impairment and higher risk for depression (Ibrahim et al., 2013). According to the 2018 American College Health Survey, 41.9% of this population felt so depressed that it was difficult to function at least once within the previous year and 18.1% were diagnosed or treated for depression within the last 12 months (American College Health Association, 2018). However, very few seek help despite the fact that many colleges offer counseling services (Eisenberg et al., 2009; Zivin et al., 2009). Research has shown that many students prefer to manage mental health issues alone, due to concerns on confidentiality, a feeling that the problem is too personal, and the fear that no person or service could help (Ebert et al., 2019; Gould et al., 2002). In addition, fear of stigma, lack of time, financial costs, and inconvenient access to care have been identified as significant barriers to seeking help among college students (Gulliver and Bennett, 2015; Hunt and Eisenberg, 2010; Zivin et al., 2009). Because of these important hindrances, innovative interventions that overcome the obstacles are necessary to address depression for the student population. Online self-help programs may have the potential to improve the delivery of mental health service in colleges. In comparison with the general population, college students are more likely to use the internet and to seek health information online (Chiauzzi et al., 2008; Hanauer et al., 2004). Such internet-based interventions are easily accessible and available all the time to participants; do not have to involve a dedicated therapist, allowing participants to stay anonymous without adopting a patient role; and are cost-effective especially when involving some guidance (Donker et al., 2015; Health Quality Ontario, 2019; Mitchell et al., 2021). Since students experiencing mental health problems infrequently seek help from professionals due to various barriers but tend to be comfortable with modern information technologies, online self-help interventions may be of especial interest to the student population. It is possible that, for many students, online self-help interventions may be an acceptable alternative to traditional face-to-face formats. For instance, students who are reluctant to seek face-to-face counseling due to the real or perceived barriers to accessing services may be willing to use an online self-help program that can be accessed privately at a time that is convenient. Despite the advantages of online self-help interventions, one potential shortcoming of these interventions is low program completion rates (Andrews et al., 2018), which may diminish the effectiveness of an intervention. Multiple meta-analyses examining the efficacy of internet-based interventions suggest that users reap larger benefits from those including provision of human support and guidance (Andersson and Cuijpers, 2009; Baumeister et al., 2014; Heber et al., 2017). It is encouraging that recent meta-analyses suggest that guided self-help interventions for depression can have comparable effects to face-to-face interventions as well as equal adherence (Cuijpers et al., 2010a; van Ballegooijen et al., 2014; Carlbring and Andersson, 2018). In this review, we restricted the focus to guided online self-help interventions, which can be distinguished from other internet-based interventions by the support that is given by health professionals (e.g., public health nurses, psychotherapists) or program coaches to the participants when working through the intervention. To date, a number of internet-based interventions either guided or unguided have been developed to address depression among the college population. Although associations of these interventions with depression have been evaluated in randomized trials, results of these studies were frequently limited by small sample size, high dropout rates, and inconsistent findings. Only one qualitative review and two meta-analytic reviews have examined the effectiveness of internet-based interventions for depression targeting this population (Davies et al., 2014; Harrer et al., 2019; Lattie and Adkins, 2019). Although the aforementioned meta-analyses showed positive results on depressive symptoms with effect sizes in a small range, all authors combined guided and unguided self-help interventions in their investigations and there was substantial variation in the degree of success for included programs. As stated earlier, guided self-help interventions differ significantly from those unguided in terms of dropout rates and intervention effects, and should therefore be analyzed independently. However, no meta-analysis has specifically focused on effect studies of online guided self-help interventions for college students and examined potential effect moderators and other sources of heterogeneity. Restricting the focus to online guided self-help interventions in this meta-analysis will increase homogeneity, and enable us to assess their independent impact specific to this population. In addition to examining the overall effect size, meta-analysis also provides opportunities to calculate and compare the separate effect sizes for specific subgroups of interest. Such information would be useful as this may assist in identifying the optimal conditions for future intervention efforts. Research has shown that intervention programs for depression targeting at-risk individuals (selective and indicated programs) exhibited more support than those delivered universally (universal programs), possibly because high-risk participants are more motivated to engage in the program content and have a greater opportunity to show symptom reduction (Horowitz and Judy, 2006). Knowledge of such differences lends itself well to the suggestion that the effects of online guided self-help interventions for depression among college students may vary across universal, selective and indicated programs. Furthermore, whilst including therapist guidance could be beneficial, the increased cost associated with providing such professional support may prevent colleges with insufficient resources from adopting internet programs. Support has emerged that it is not necessary for the person providing such guidance to be a professionally trained therapist (Robinson et al., 2010; Titov et al., 2010). Thus it would be of value to evaluate whether online self-help interventions with different levels of contact (e.g., therapist, coach, and standardized emails) would produce equal effects for college students. Moreover, although meta-analyses on other problems such as smoking and eating disorder suggested superior effects for longer interventions to very brief interventions (Rooney and Murray, 1996; Stice and Shaw, 2004), larger effects emerged for shorter than longer interventions in a meta-analysis for adolescent depression (Stice et al., 2009). In the college context, it is possible that longer programs require significant time and may not be practical or appealing to college students who commonly have a crowded agenda. This meta-analysis therefore investigated the differential effects of such interventions by length of intervention. In addition, it is unclear whether increasing students' engagement through providing compensation (e.g., course credit, cash, and raffles) for participating in the program would produce greater intervention effects and it would be helpful for this meta-analysis to assess the difference between interventions with and without such incentive. Finally, intervention content may also affect whether a program produces effects, as elements that form the basis of different strategies have varied widely. Depression intervention programs can employ a broad range of strategies and examining them separately can help inform which strategy would produce greater benefit among the student population. Therefore, a meta-analytic review was undertaken to assess the effectiveness of online guided self-help interventions for depressive symptoms in college students. It was hypothesized that students assigned to these interventions would show lower depressive symptoms than those assigned to control conditions. A secondary aim was to examine the effectiveness of online guided self-help interventions in different contexts. We expected to find differential effects of these interventions by type of sample, level of contact, use of incentive, length of intervention, and intervention content.

Methods

This study was conducted in adherence to the PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health care interventions (Moher et al., 2009); see Appendix 1.

Identification and selection of studies

Studies were identified by searching PubMed, Embase, Web of Science, PsycINFO, and Cochrane Central for papers from inception to September 13 2019, and updated to April 1, 2021. The literature search was constructed around search terms for relevant participants (e.g., college students), search terms for relevant interventions (e.g., internet program), search terms for relevant outcomes (e.g., depression), and search terms for relevant design (e.g., randomized). The search strategy used in PubMed is displayed in Appendix 2, which was adapted for other databases as necessary. Reference lists from relevant reviews and retrieved studies were reviewed and searched to identify additional records. Initial records identified from electronic database searching, relevant reviews, and retrieved studies were imported into Endnote X7. After removal of duplicates, titles and abstracts were screened by the primary researcher (LM) and full-text articles for studies deemed potentially relevant were retrieved for further assessment by two independent researchers (LM and CH). Disagreements between the two researchers were resolved through discussion and consensus, and a third researcher (ZC) who was consulted if agreement between the two researchers could not be reached. We included (a) randomized controlled trials (RCTs) that compared (b) an online guided self-help intervention (which was considered eligible when the intervention used the Internet as delivery mode via computer, laptop, tablet, or other device, and included some degree of personal guidance or at least regular e-mail or other non-technical contacts with a therapist or coach) to (c) a non-active control condition including wait-list, assessment-only, and usual-care (in which participants were permitted to continue using the care as usual such as mental health counseling services) in terms of effects on (d) depressive symptoms as a primary or secondary (which was regarded as eligible only when the primary outcome was distress, anxiety or stress) outcome using (e) validated measures among (f) non-clinical samples at a (g) higher education facility (e.g., university, college, tertiary, or comparable post-secondary education).

Data extraction

Information was extracted using a piloted abstraction form for each study based on: (a) the characteristics of the trial including country and year of publication, sample size, outcome measures, and type of control; (b) the characteristics of the intervention including intervention content, follow-up in weeks, level of contact, use of incentive, and involvement of group technique; and (c) the characteristics of participants including student sample, recruitment strategy and setting, type of sample, mean age, and percentage of females. Type of sample was categorized as universal if the intervention was administered to all members of a population without explicit screening procedures on symptoms or risk factors of depression; selective if participants were included on the basis of endorsing a known risk factor of depression; and indicated if the intervention was delivered to individuals reporting elevated but subclinical levels of depressive symptoms (Mrazek and Haggerty, 1994). Level of contact was categorized as high if qualified psychotherapists monitored the intervention and provided support directly; moderate if the support was provided by research team members or program coaches such as trained students; and low if only standardized reminder emails were used to deliver the support. Length of intervention was categorized as brief (≤4 weeks), moderate (4–8 weeks), and long (≥8 weeks) following the categorization of a recent meta-analysis (Harrer et al., 2019). Use of incentive was categorized as yes if students were rewarded by cash, course credit, or other forms of compensation for their engagement in the program; and no if incentive was not used. Intervention content was categorized as CBT if the intervention focused on cognitive restructuring such as identifying and changing negative cognition as well as behavioral activation that emphasizes increasing pleasant activities (Beck, 1979); third-wave CBT if the intervention relied on components based on mindfulness, meditation, and/or acceptance principles (Ost, 2008) such as mindfulness-based therapy (MBT; Kabat-zinn, 2003) and acceptance and commitment therapy (ACT; Hayes, 2004); and others if the intervention utilized additional strategies (e.g., physical activity). Involvement of a group technique was indicated if the intervention included group-based components, such as live videoconferences. Recruitment strategy was categorized as online if the participants were recruited primarily via online advertisements, websites, or circular e-mails; and on-site if recruited through announcements in classes or other institutions. Study setting was categorized as campus if the participants were recruited from specified colleges; and public if recruited from the general population.

Risk of bias assessment

The methodological quality was determined using the Cochrane Collaboration's Risk-of-Bias Tool 2 (Sterne et al., 2019), which considers risk of bias across five domains: (1) the randomization process (e.g., adequate generation of random sequence and concealment of allocation to conditions); (2) deviations from the intended interventions (e.g., blinding of participants, carers and trial personnel); (3) missing outcome data (e.g., attrition rate < 10% or use of intention-to-treat analysis); (4) measurement of outcome (e.g., blinding of outcome assessors); and (5) selection of the reported results (e.g., published trial protocols). The risk of bias for each domain was scored as low, moderate, or high. Because we focused on trials that compared online guided self-help interventions with non-active controls such as wait-list, it was difficult in these contexts to ensure that the participants and people delivering the interventions were unaware of the interventions assignment. Therefore, we omitted the domain assessing the risk of bias due to deviations from the intended interventions. The risk of bias assessment was performed independently by two authors (LM and RT), with a third author (ZC) engaged in case of disagreement.

Statistical analysis

Intervention effects were defined as the difference of mean change scores in depressive symptoms between intervention and control conditions at post-intervention. We calculated the effect size for each study with standardized mean differences (SMD) and 95% confidence intervals (CI). Because a number of studies had small sample sizes, we corrected the effect sizes for small sample bias using Hedges' g (Hedges and Olkin, 1985): , where the pooled standard deviation (at pretest) was defined as and the correction factor was defined as . We used random-effects models and inverse variance weighted methods to evaluate the pooled effect sizes (DerSimonian and Laird, 1986). Effect sizes of 0.5 or lower were considered as small, effect sizes of 0.5–0.8 were interpreted as moderate, and effect sizes higher than 0.8 were regarded as large (Cohen, 2013). In some studies (namely: Ahmad et al., 2020; Ellis et al., 2011; Gibbel, 2010; Morris et al., 2016; Sethi et al., 2010), multiple intervention arms were included, resulting in two comparisons between an online guided self-help intervention and the control condition. We calculated the effect size for each individual comparison and included both comparisons in the overall meta-analysis. Because such comparisons are dependent, and may distort the pooled effect sizes by artificially reducing heterogeneity (Borenstein et al., 2011), we conducted a series of sensitivity analyses to assess the influence of studies with multiple comparisons on the pooled effect size. First, we recalculated the pooled effect size by including only the comparison with larger effect size from that study. Second, we recalculated the pooled effect size by including only the comparison with the smaller effect size from that study. Finally, we recalculated the pooled effect size by combining the effects of both intervention groups in that study to create a single comparison before being pooled in the overall meta-analysis. We evaluated the impact of individual studies on the pooled effect sizes by conducting additional influence analyses using a holdout approach whereby each individual study was sequentially omitted from the eligible study pool and analyzed. Publication bias was assessed using Egger regression symmetry tests, which indicate publication bias if asymmetry is suggested. We also performed the Duval-Tweedie trim and fill procedure, with an adjusted effect size being produced to account for missing studies that lead to publication bias (Duval and Tweedie, 2000). Heterogeneity was assessed using the I2-statistic that was distinguished as low, moderate, and high with values of 25%, 50%, and 75% (Higgins et al., 2003). Subgroup analyses were performed to explore potential sources of heterogeneity. We calculated the pooled effect sizes for subgroups with enough comparisons available (n ≥ 3), including type of sample (universal or selective or indicated), level of contact (high or moderate or low), use of incentive (yes or no), length of intervention (brief or moderate or long), and intervention content (CBT or third-wave). All the analyses were performed with Stata release 12 (StataCorp) and p ≤ 0.05 was considered statistically significant.

Results

Selection and inclusion of studies

A total of 3382 records were initially identified from database searching and reference review. After a careful assessment of the identified citations with the eligibility criteria, 19 studies were included in the current review. A PRISMA flowchart summarizing the process for inclusion of studies is presented in Fig. 1.
Fig. 1

PRISMA flowchart.

PRISMA flowchart.

Descriptive characteristics of included studies

Descriptive characteristics of the included studies are presented in Table 1. All studies were conducted after 2010, primarily in western countries. In total, the studies included 3074 participants with sample sizes ranging from 38 (Sethi et al., 2010) to 1162 (Viskovich & Pakenham, 2020). All studies but one (Cook et al., 2019) recruited the participants from specified colleges. The participants were primarily recruited online (n = 15), whilst the remaining 4 studies used on-site recruitment strategies. The mean age of participants was 22.9 years and the average proportion of females was 73.4%. Of the 19 studies, ten studies used traditional CBT, eight studies used third-wave CBT, and one study used physical activity. There were four studies including high levels of contact, 13 studies providing moderate levels of contact, and 2 studies involving low levels of contact throughout the intervention. The length of intervention ranged from 3 (Ellis et al., 2011; Levin, 2014; Sethi et al., 2010) to 12 weeks (Cook et al., 2019), including 6 brief interventions, 8 moderate interventions, and 5 long interventions. Eight were categorized as universal interventions, nine were categorized as selective interventions, and 2 were categorized as indicated interventions. Eleven studies used incentive to promote students' engagement in the program whilst 8 studies provided no such compensation. The interventions were predominantly administered individually, with three studies also involving group technique. The most commonly used measure for depressive symptoms was DASS (Depression, Anxiety, and Stress Scales; n = 6), PHQ (Patient Health Questionnaire; n = 6), and BDI (Beck Depression Inventory; n = 5). The majority (n = 14) of studies used a wait-list control condition, and the other studies used an assessment-only (n = 3) or usual-care (n = 2) control condition. For studies providing data on follow-up assessments (n = 7), support was shown in all these studies that the intervention effects were maintained at up to 12 months follow-up.
Table 1

Description and characteristics of included studies.

AuthorStudent sample (n)RecruitSettingAge (mean)Female%Sample typeOnline intervention (n)Control (n)Contact levelFollow-up (wk)IncentiveGroup formatSymptom measureCountry
Day et al., 2013Students with mild to moderate levels of anxiety, depression or stress: DASS ≥ 8, 10, 15 (66)OnlineCampus23.689SCBT (33)WL (33)Trained students (M)6, 24NoDASSCAN
Ellis et al., 2011Students with elevated scores on the Kessler Psychological Distress Scale: K10 < 30 (39)On-siteCampus19.777SOnline CBT: The MoodGYM intervention (13)Online Peer Support: The MoodGarden intervention (13)AO (13)Researcher (M)3YesDASSAUS
Eustis et al., 2018General (156)OnlineCampus2579UAcceptance-based behavioral intervention (78)WL (78)Doctoral candidate in clinical psychology (M)4YesDASSUSA
Gibbel, 2010Students with mildly and moderately depressed undergraduates: CES-D ≥ 10 and <25 (65)OnlineCampusNS83IiCBT (Moodgym) (24); Spiritual intervention (19)AO (22)Research team member (M)5, 7YesCES-DUSA
Harrer et al., 2018Students with elevated levels of stress: PSS-4 ≥ 8 (150)OnlineCampus24.175SiCBT for stress (StudiCare Stress) (75)WL (75)Trained students (M)7, 12NoBDIGER
Levin, 2014First-year students (76)On-siteCampus18.454UACT (37)WL (39)Regular contact (M)3, 6YesDASSUSA
Levin et al., 2017General (79)OnlineCampus2166UACT (40)WL (39)Research assistants (M)4YesCCAPS-DUSA
Mailey et al., 2010Students with receiving mental health counseling (47)On-siteCampus2568SPA (26)UC (25)Counselor (M)10NoBDIUSA
Morris et al., 2016General (138)OnlineCampus20.567UCBT programs: “Insomnia Relief” (48) and “Anxiety Relief” (43)WL (47)Standardized emails (L)6YesBDIUK
Mullin et al., 2015Students with self-identified as experiencing symptoms of anxiety or depression (55)OnlineCampus27.964SCBT (30)WL (23)Psychologist (H)6, 12YesPHQAUS
Räsänen et al., 2016Students with self-reporting as experiencing some form of psychological distress (68)OnlineCampus2485SACT (33)WL (35)Trained psychology students (M)7, 48NoBDIFIN
Richards et al., 2016Students with self-reported GAD symptoms: GAD-7 ≥ 10 (137)OnlineCampus23.877SCBT (70)WL (67)Psychologists (H)6NoBDIIRL
Sethi et al., 2010First-year students with low to moderate levels of depression and/or anxiety (38)On-siteCampus19.566SiCBT (Moodgym) (9);Combined face-to-face + online CBT (9)AO (10)Psychologists (H)3YesDASSAUS
Viskovich & Pakenham, 2020General (1162)OnlineCampus26.967.8UACT (596)WL (566)Standardized emails (L)4NoDASSAUS
Ritvo et al., 2021Undergraduates (154)OnlineCampus23.176UMindfulness virtual community (76)WL (78)Moderators (M)8YesLVCPHQCAN
Cook et al., 2019Students with elevated repetitive thoughts: PSWQ ≥ 50; RRS ≥ 40 (235)OnlinePublic20.483SCBT (82)UC (77)Clinician (H)12NoPHQUK
Salamanca-Sanabria et al., 2020Students with mild to moderately severe depressive symptoms: PHQ ≥ 10 and <19 (214)OnlineCampus22.271ICBT (107)WL (107)Postgraduate student in clinical psychology (M)7, 12NoPHQCO
Ahmad et al., 2020Undergraduates (119)OnlineCampus24.875UFull mindfulness virtual community (40); partial mindfulness virtual community (39)WL (40)Moderators (M)8YesLVCPHQCAN
EI Morr et al., 2020Undergraduates (160)OnlineCampus22.679UMindfulness virtual community (79)WL (80)Moderators (M)8YesLVCPHQCAN

Note. U = universal; S = selective; I = indicated; M = moderate; H = high; WL = wait-list; UC = usual-care; AO = assessment-only; LVC = live video conference. NS = not specified; BDI = Beck Depression Inventory; DASS = Depression, Anxiety, and Stress Scales; PHQ = Patient Health Questionnaire; CES-D = Centre for Epidemiologic Studies-Depression; CCAPS-D = Counseling Center Assessment of Psychological Symptoms-Depression.

Description and characteristics of included studies. Note. U = universal; S = selective; I = indicated; M = moderate; H = high; WL = wait-list; UC = usual-care; AO = assessment-only; LVC = live video conference. NS = not specified; BDI = Beck Depression Inventory; DASS = Depression, Anxiety, and Stress Scales; PHQ = Patient Health Questionnaire; CES-D = Centre for Epidemiologic Studies-Depression; CCAPS-D = Counseling Center Assessment of Psychological Symptoms-Depression. Table 2 presents a summary of risk of bias assessment for each of the included studies. Overall, the risk of bias was variable across studies: only one study fully met all the remaining four domains and was categorized as low risk; eight studies showed high risk for at least one domain and was categorized high risk; and 10 studies were categorized as moderate risk as concerns emerged for one or more domains.
Table 2

Risk of bias assessment.

AuthorThe randomization processMissing outcome dataMeasurement of the outcomeSelection of the reported resultOverall
Day et al., 2013+++~Moderate
Ellis et al., 2011~+~High
Eustis et al., 2018~+~High
Gibbel, 2010~+~High
Harrer et al., 2018+++~Moderate
Levin, 2014++~High
Levin et al., 2017~++~Moderate
Mailey et al., 2010++~High
Morris et al., 2016~++~Moderate
Mullin et al., 2015~+~~Moderate
Räsänen et al., 2016~++High
Richards et al., 2016++++Low
Sethi et al., 2010+~~High
Viskovich & Pakenham, 2020~++~Moderate
Ritvo et al., 2021+++~Moderate
Cook et al., 2019++~+Moderate
Salamanca-Sanabria et al., 2020++~High
Ahmad et al., 2020+++~Moderate
EI Morr et al., 2020+++~Moderate

Note. + = lower risk of bias; − = higher risk of bias; ~ = some concern.

Risk of bias assessment. Note. + = lower risk of bias; − = higher risk of bias; ~ = some concern.

Meta-analytic results

We were able to calculate the effect sizes indicating the difference between online guided self-help interventions and control conditions in 24 comparisons that derived from the 19 studies. Forest plot for the meta-analysis results on depressive symptoms is presented Fig. 2. At posttest, 15 comparisons (63%) produced non-significant effect sizes (the 95% CIs overlapped zero), 8 comparisons (33%) produced significant effect sizes in small-to-moderate range, and 1 comparison (4%) produced a significant and large effect size.
Fig. 2

Forest plot for the meta-analysis of online guided self-help interventions on depressive symptoms.

Forest plot for the meta-analysis of online guided self-help interventions on depressive symptoms. The pooled effect size for depressive symptoms at posttest was g = 0.46 (95% CI: 0.28–0.64), which was considered small but significant (the 95% CI did not overlap zero). Heterogeneity for depressive symptoms was high and significant (I2 = 77.6%; p < 0.001). Visual inspection of the screen plot indicated that the study by Salamanca-Sanabria et al. (2020) (which targeted students with mild to moderately severe depressive symptoms) produced the largest effect size and was discrepant with the overall distribution. After removal of this potential outlier, the pooled effect size dropped to 0.36 (95% CI: 0.26–0.45), and the heterogeneity turned non-significant and was reduced to a low level (I2 = 13.2%; p = 0.28). Because of its considerable influence on the pooled effect size, we excluded this study from subsequent analyses. In the sensitivity analyses of studies with multiple comparisons, the resulting effect size was g = 0.38 (95% CI: 0.27–0.48) when comparisons with the smaller effect size from the same study were removed; the resulting effect size was g = 0.35 (95% CI: 0.25–0.45) on exclusion of comparisons with the larger effect size; and the resulting effect size was g = 0.35 (95% CI: 0.26–0.45) when combined effects of multiple comparisons were considered in the meta-analysis. Heterogeneity was non-significant and remained at low levels (all I2 ≤ 25%; all p > 0.05) in these analyses. These results suggested no substantial changes in the pooled effect sizes compared with the ones found in the overall analysis. In the influence analyses of individual studies, the pooled effect size ranged from 0.31 (95% CI: 0.24–0.39) on exclusion of the study by Harrer et al. (2018) to 0.40 (95% CI: 0.30–0.50) after removal of the study by Viskovich and Pakenham (2020). There was no indication of significant heterogeneity (both I2 = 0; both p > 0.05) in the analyses in which either of the studies was omitted. Overall, these analyses indicated that removal of individual studies resulted in no substantial changes in the pooled effect sizes. We further assessed the impact of 5 studies in which wait-list control condition was not used, by omitting these studies from the analyses. The heterogeneity was non-significant and low (I2 = 20.5%; p = 0.23), and the resulting effect size was g = 0.38 (95% CI: 0.27–0.48), which did not change substantially compared with the overall analysis. Similar analyses were performed to evaluate the influence of three studies that involved group-based components. The resulting effect size was again g = 0.38 (95% CI: 0.27–0.48) with a non-significant and low level of heterogeneity (I2 = 18.3%; p = 0.23), which was also comparable to the effect size found in the overall meta-analysis. There was no evidence for asymmetry of the funnel plot in the overall meta-analysis as Egger's test was non-significant (p = 0.09), which indicated low levels of publication bias. Under the trim and fill procedure, the adjusted effect size was stable (g = 0.35; 95% CI: 0.26–0.43), with three imputed studies being incorporated into the model. Results for the subgroup analyses were available in Table 3. None of the hypothesized variables accounting for any significant between-subgroup differences (all p > 0.05). We focused on analyses evaluating whether the effects of these interventions were significant in specific subgroups of interest. Significant effects were replicated across interventions using both universal (g = 0.27; 95% CI: 0.18–0.36) and selective approaches (g = 0.47; 95% CI: 0.30–0.64); interventions of shorter (g = 0.29; 95% CI: 0.16–0.42), moderate (g = 0.52; 95% CI: 0.37–0.68), and greater length (g = 0.25; 95% CI: 0.09–0.41); interventions of high (g = 0.42; 95% CI: 0.10–0.74), moderate (g = 0.40; 95% CI: 0.29–0.52), and low levels of contact (g = 0.24; 95% CI: 0.13–0.34); interventions with (g = 0.36; 95% CI: 0.23–0.48) and without incentive (g = 0.39; 95% CI: 0.21–0.57); and interventions employing traditional (g = 0.42; 95% CI: 0.27–0.57) and third-wave CBT (g = 0.32; 95% CI: 0.21–0.44). Heterogeneity was low and non-significant (all I2 < 25%; all p ≥ 0.05) in these subgroup analyses.
Table 3

Subgroup analyses by hypothesized variables.

SubgroupNo.
Effect sizes
P value between subgroups
ComparisonsIntervention conditionControl conditionHedge's g (95% CI)Heterogeneity
Type of sample0.26
 Universal1010779670.27 (0.18, 0.36)I2 = 0, p = 0.71
 Selective113933580.47 (0.30, 0.64)I2 = 16.9%, p = 0.28
Length of intervention0.46
 Brief87957450.29 (0.16, 0.42)I2 = 6.1%, p = 0.38
 Moderate93753020.52 (0.37, 0.68)I2 = 0, p = 0.66
 Long63433000.25 (0.09, 0.41)I2 = 0, p = 0.56
Level of contact0.70
 High52001770.42 (0.10, 0.74)I2 = 43.8%, p = 0.13
 Moderate156265570.40 (0.29, 0.52)I2 = 0, p = 0.50
 Low36876130.24 (0.13, 0.34)I2 = 0, p = 0.98
Use of incentive0.92
 Yes165984690.36 (0.23, 0.48)I2 = 0, p = 0.64
 No79158780.39 (0.21, 0.57)I2 = 52.0%, p = 0.05
Intervention content0.99
 CBT124443560.42 (0.27, 0.57)I2 = 6.6%, p = 0.38
 Third-wave1010439660.32 (0.21, 0.44)I2 = 15.2%, p = 0.30
Subgroup analyses by hypothesized variables. We also conducted a subgroup analysis to examine the potential influence of study quality on the intervention effects. The effect size based on comparisons categorized as high risk of bias was g = 0.45 (95% CI: 0.27–0.63); and the effect size based on comparisons categorized as moderate risk of bias was g = 0.29 (95% CI: 0.20–0.39). Heterogeneity stayed low and non-significant in these analyses (both I2 < 10%; both p > 0.05), and no significant difference (p = 0.52) was produced between these subgroups, suggesting the results are robust despite the inclusion of studies with mixed quality. We also attempted to examine the long-term effectiveness of online guided self-help interventions for depressive symptoms among college students. However, owing to the small number of studies (n = 7) with variable follow-up intervals (which ranged from 6 to 60 weeks) and insufficient follow-up data (e.g., follow-up assessments were only conducted within the intervention condition), we were unable to calculate the follow-up effect sizes indicating the difference between online guided self-help interventions and control conditions in the longer term.

Discussion

This meta-analysis examined 19 studies using online guided self-help interventions to address depressive symptoms for a total of 3074 college students. We found that students who participated in these interventions report lower levels of depressive symptoms compared to those who received no intervention. The effect size found in this meta-analysis is modest but compares favorably to the one reported in Harrer's recent meta-analysis, which combined guided and unguided internet-based interventions targeting college students. Such strengthening effects may be attributable to the focus on guided self-help interventions in the current meta-analysis as stronger effects were found for guided versus unguided online interventions in a range of other target groups (Heber et al., 2017; Richards and Richardson, 2012; Karyotaki et al., 2021). Our analyses detected no significant difference for online guided self-help interventions with different levels of contact (e.g., therapist, coach, or standardized emails), suggesting that such interventions may produce equivalent effects. In support, two other meta-analyses have reported similar findings with regard to the equivalent efficacy of therapist versus non-therapist supported online interventions for depression among the general population (Richards and Richardson, 2012; Newby et al., 2016). However, this failure to find a significant difference may be due to a lack of power, since our analyses were based on a relatively small number of studies. Thus, further studies are needed to determine whether internet-based interventions involving different levels of contact would produce equal effects for addressing depressive symptoms among college students. Significant effects were exhibited for online guided self-help interventions among both universal and selective samples, with albeit non-significantly greater effects for selective interventions. In contrast to multiple meta-analyses of different depression intervention programs supporting superior effects among selective to universal samples (Horowitz and Judy, 2006; Merry et al., 2004; Stice et al., 2009), our analyses point at the potential of equivalent effects for online guided self-help interventions among these samples in the college context. It is possible that online guided self-help interventions are of greater appeal to college students especially when administered universally. However, drawing definitive conclusions based on current evidence is premature and more research is needed. The effect size of online guided self-help interventions with and without incentive was both significant and comparable in our analyses. Although providing incentive might attract more students to participate, the insignificant between-subgroup results underscore the importance for future studies to determine what is cost-effective for colleges to implement at a broader scale. It is possible that students who participate in the program on condition of compensation endorse lower levels of depressive symptoms compared with those purely volunteering to participate. As a result, the effect size of programs providing incentive may be mitigated by the inclusion of more students at low-risk status in contrast with the effect size of programs delivered on a voluntary basis. As more studies of online guided self-help interventions are conducted, it would be helpful to revisit the question of whether providing incentive truly affects the intervention effects. Our analyses indicated no significant difference between online guided self-help interventions of moderate length and those briefer or longer interventions. Such findings did not replicate the results of other meta-analyses which suggested significantly stronger effects for interventions of moderate length (Harrer et al., 2019; Heber et al., 2017; Richards and Richardson, 2012). Theoretically, longer interventions enable participants to more sufficiently reflect on the program content and practice the attitudinal and behavioral change skills. However, there could be an upper limit for this occasion as extremely long programs may not appeal to youth, leading to increased dropout rates and attenuating intervention effects. Thus, further trials are needed to establish whether intervention length is associated with improved outcomes, and, if so, the optimal duration of such interventions. This meta-analysis also examined the difference between traditional and third-wave CBT. In Harrer et al.'s meta-analysis, the authors found significantly higher effects on depressive symptoms among college students for programs based on CBT principles than other programs. However, they used a broader definition of CBT that combined both traditional and third-wave CBT in their analyses, and did not compare the effects between different CBT strategies. Thus, this meta-analysis provided support for Harrer et al.'s approach of combining traditional and third-wave CBT as we found no significant difference between these strategies. As research accumulates, it would be worth further exploring whether providing guidance for different intervention strategies would contribute to larger reductions of depressive symptoms. For instance, it is possible that third-wave CBT may be delivered more efficaciously if some guidance is provided, whereas increasing guidance may not lead to a significant influence on the effects of traditional CBT. Future research comparing directly traditional and third-wave CBT with the same level of guidance is also needed before we can draw definitive conclusions about the superiority of these different intervention strategies. This meta-analysis was not pre-registered and has several limitations. First, only English language papers prior to April 1, 2021 were identified, which may omit potentially germane studies (e.g., more current or non-English studies) from inclusion within the meta-analysis. Second, this meta-analysis focused on inactive control conditions and the quality of many studies included in the meta-analysis was variable, which may produce an inflated estimate for the efficacy of these interventions (Cuijpers et al., 2010b; Sterne et al., 2000; Thornton and Lee, 2000). Third, this meta-analysis was limited to examine the follow-up effects of online guided self-help interventions as many studies did not report longer-term outcomes and the studies that did report these data employed various follow-up intervals. Forth, this meta-analysis included a range of subgroup contrasts that were based on a small number of studies (n ≤ 10) and many of these studies had small samples, limiting our analyses to detect significant between-subgroup differences. This meta-analysis has several implications. First, it reinforces that online guided self-help interventions are associated with modest reductions in depressive symptoms among college students. Second, it underlines a range of research gaps. For instance, many of the studies used universal or selective samples with predominant females and provided data solely at the conclusion of intervention, whilst few studies were administered to indicated samples, focused on male students, or conducted follow-up assessments. Third, it provides preliminary information on the mechanisms of how online guided self-help interventions operate among college students by examining a range of hypothesized variables. Experimentally manipulating these variables such as level of contact, length of intervention, and use of incentive would be essential for future research to confirm such ostensive relations. These results suggest that online guided self-help interventions are likely to be a promising approach for addressing depressive symptoms among college students. However, because of the generally variable and limited quality of current evidence, further research applying rigorous methods is needed to confirm and extend the findings of this meta-analysis.

Declaration of competing interest

All the authors declare that they have no conflicts of interest.
PRISMA checklist
Section/topic#Checklist itemReported on page #
Title
Title1Identify the report as a systematic review, meta-analysis, or both.2



Abstract
Structured summary2Provide a structured summary including, as applicable: background; objectives; data sources; study eligibility criteria, participants, and interventions; study appraisal and synthesis methods; results; limitations; conclusions and implications of key findings; systematic review registration number.1



Introduction
Rationale3Describe the rationale for the review in the context of what is already known.2–6
Objectives4Provide an explicit statement of questions being addressed with reference to participants, interventions, comparisons, outcomes, and study design (PICOS).6–7



Methods
Protocol and registration5Indicate if a review protocol exists, if and where it can be accessed (e.g., Web address), and, if available, provide registration information including registration number.N/A
Eligibility criteria6Specify study characteristics (e.g., PICOS, length of follow-up) and report characteristics (e.g., years considered, language, publication status) used as criteria for eligibility, giving rationale.7
Information sources7Describe all information sources (e.g., databases with dates of coverage, contact with study authors to identify additional studies) in the search and date last searched.7
Search8Present full electronic search strategy for at least one database, including any limits used, such that it could be repeated.7, 25
Study selection9State the process for selecting studies (i.e., screening, eligibility, included in systematic review, and, if applicable, included in the meta-analysis).8
Data collection process10Describe method of data extraction from reports (e.g., piloted forms, independently, in duplicate) and any processes for obtaining and confirming data from investigators.8
Data items11List and define all variables for which data were sought (e.g., PICOS, funding sources) and any assumptions and simplifications made.8–9
Risk of bias in individual studies12Describe methods used for assessing risk of bias of individual studies (including specification of whether this was done at the study or outcome level), and how this information is to be used in any data synthesis.9–10
Summary measures13State the principal summary measures (e.g., risk ratio, difference in means).10
Synthesis of results14Describe the methods of handling data and combining results of studies, if done, including measures of consistency (e.g., I2) for each meta-analysis.11–12
Risk of bias across studies15Specify any assessment of risk of bias that may affect the cumulative evidence (e.g., publication bias, selective reporting within studies).11
Additional analyses16Describe methods of additional analyses (e.g., sensitivity or subgroup analyses, meta-regression), if done, indicating which were pre-specified.11–12



Results
Study selection17Give numbers of studies screened, assessed for eligibility, and included in the review, with reasons for exclusions at each stage, ideally with a flow diagram.12
Study characteristics18For each study, present characteristics for which data were extracted (e.g., study size, PICOS, follow-up period) and provide the citations.37–38
Risk of bias within studies19Present data on risk of bias of each study and, if available, any outcome level assessment (see item 12).13, 39
Results of individual studies20For all outcomes considered (benefits or harms), present, for each study: (a) simple summary data for each intervention group (b) effect estimates and confidence intervals, ideally with a forest plot.Fig. 2
Synthesis of results21Present results of each meta-analysis done, including confidence intervals and measures of consistency.14–16
Risk of bias across studies22Present results of any assessment of risk of bias across studies (see item 15).14–16
Additional analysis23Give results of additional analyses, if done (e.g., sensitivity or subgroup analyses, meta-regression [see item 16]).14–17



Discussion
Summary of evidence24Summarize the main findings including the strength of evidence for each main outcome; consider their relevance to key groups (e.g., healthcare providers, users, and policy makers).17–19
Limitations25Discuss limitations at study and outcome level (e.g., risk of bias), and at review-level (e.g., incomplete retrieval of identified research, reporting bias).20
Conclusions26Provide a general interpretation of the results in the context of other evidence, and implications for future research.21



Funding
Funding27Describe sources of funding for the systematic review and other support (e.g., supply of data); role of funders for the systematic review.N/A
  56 in total

Review 1.  Measuring inconsistency in meta-analyses.

Authors:  Julian P T Higgins; Simon G Thompson; Jonathan J Deeks; Douglas G Altman
Journal:  BMJ       Date:  2003-09-06

Review 2.  Economic evaluations of Internet interventions for mental health: a systematic review.

Authors:  T Donker; M Blankers; E Hedman; B Ljótsson; K Petrie; H Christensen
Journal:  Psychol Med       Date:  2015-08-03       Impact factor: 7.723

3.  Stigma and help seeking for mental health among college students.

Authors:  Daniel Eisenberg; Marilyn F Downs; Ezra Golberstein; Kara Zivin
Journal:  Med Care Res Rev       Date:  2009-05-19       Impact factor: 3.929

4.  Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement.

Authors:  David Moher; Alessandro Liberati; Jennifer Tetzlaff; Douglas G Altman
Journal:  Ann Intern Med       Date:  2009-07-20       Impact factor: 25.391

5.  Meta-analysis in clinical trials.

Authors:  R DerSimonian; N Laird
Journal:  Control Clin Trials       Date:  1986-09

Review 6.  Psychological and/or educational interventions for the prevention of depression in children and adolescents.

Authors:  S Merry; H McDowell; S Hetrick; J Bir; N Muller
Journal:  Cochrane Database Syst Rev       Date:  2004

7.  Internet treatment for depression: a randomized controlled trial comparing clinician vs. technician assistance.

Authors:  Nickolai Titov; Gavin Andrews; Matthew Davies; Karen McIntyre; Emma Robinson; Karen Solley
Journal:  PLoS One       Date:  2010-06-08       Impact factor: 3.240

Review 8.  A systematic review of studies of depression prevalence in university students.

Authors:  Ahmed K Ibrahim; Shona J Kelly; Clive E Adams; Cris Glazebrook
Journal:  J Psychiatr Res       Date:  2012-12-20       Impact factor: 4.791

9.  Internet-Delivered Cognitive Behavioural Therapy for Major Depression and Anxiety Disorders: A Health Technology Assessment.

Authors: 
Journal:  Ont Health Technol Assess Ser       Date:  2019-02-19

10.  Barriers of mental health treatment utilization among first-year college students: First cross-national results from the WHO World Mental Health International College Student Initiative.

Authors:  David Daniel Ebert; Philippe Mortier; Fanny Kaehlke; Ronny Bruffaerts; Harald Baumeister; Randy P Auerbach; Jordi Alonso; Gemma Vilagut; Kalina I Martínez; Christine Lochner; Pim Cuijpers; Ann-Marie Kuechler; Jennifer Green; Penelope Hasking; Coral Lapsley; Nancy A Sampson; Ronald C Kessler
Journal:  Int J Methods Psychiatr Res       Date:  2019-05-09       Impact factor: 4.035

View more
  1 in total

Review 1.  Meta-Analysis of Psychological Interventions for Reducing Stress, Anxiety, and Depression among University Students during the COVID-19 Pandemic.

Authors:  Romualdas Malinauskas; Vilija Malinauskiene
Journal:  Int J Environ Res Public Health       Date:  2022-07-27       Impact factor: 4.614

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.