| Literature DB >> 34733541 |
Martha Neary1, John Bunyi1, Kristina Palomares1, David C Mohr2, Adam Powell3, Josef Ruzek4,5, Leanne M Williams5, Til Wykes6,7, Stephen M Schueller1.
Abstract
OBJECTIVE: Given the increasing number of publicly available mental health apps, we need independent advice to guide adoption. This paper discusses the challenges and opportunities of current mental health app rating systems and describes the refinement process of one prominent system, the One Mind PsyberGuide Credibility Rating Scale (PGCRS).Entities:
Keywords: digital mental health; evaluation; mHealth; mental health; mobile health
Year: 2021 PMID: 34733541 PMCID: PMC8558599 DOI: 10.1177/20552076211053690
Source DB: PubMed Journal: Digit Health ISSN: 2055-2076
Figure 1.One Mind PsyberGuide evaluation metrics, as listed on onemindpsyberguide.org.
Figure 2.Process of rating scale developement.
Features assessed by the PsyberGuide Credibility Rating Scale (PGCRS).
| Domain | Feature | New in V2 | Rationale for addition |
|---|---|---|---|
| (1) Intervention Specificity | a. Clarity of proposed goal | ✓ | Without clear, measurable, specific goals, it is difficult to evaluate the success of a tool in meeting those goals. Goals should not only be clear, but achievable; a tool which over promises or makes lofty claims is unlikely to deliver on those goals (for example, “become more successful” or change your life”). |
| (2) Consumer Ratings | a. (i) Number of app store ratings | ||
| a (ii) Average value | ✓ | In addition to the number of consumer reviews, which serves as a proxy for popularity, the average value of reviews can help distinguish apps which consumers rate highly or poorly. | |
| (3) Research | a. Direct research evidence | ||
| b. Indirect research evidence | ✓ | There is value in a tool being grounded in indirect evidence and evidence-based practices. Ideally, tools will have both direct and indirect evidence. | |
| c. Research independence & review | |||
| (4) Development | a. Development processes (e.g. pilot, feasibility & acceptability data; stakeholder engagement) | ✓ | Valuable lessons can be learned through data collected during the initial development or piloting process which can inform product development. It is also important for developers to solicit feedback from the stakeholders on what the app is designed to help, to ensure the app is meaningful and relevant to them. |
| b. Efficacy of other products by same development team | ✓ | There is some knowledge and learning that comes from developing a previous MH app which can assist in the development of a subsequent product. Importantly, in order to receive points here the previous app needs to have been shown to be effective. This is consistent with FDA's pre-certification process that incorporates developmental process and team into its review. | |
| c. Clinical input in development | |||
| d. Ongoing maintenance and updates (date of last software update) |
Discrepancies between final and initial scores and inter-rater reliability for initial scores for training apps.
| Domain | 1 | 2 | 3 | 4 | |||||
|---|---|---|---|---|---|---|---|---|---|
| Feature | a | a | a | b | c | a | b | c | d |
|
| −0.26 | 0.00 | 0.00 | −0.14 | −0.09 | −0.17 | 0.07 | −0.17 | 0.00 |
|
| 0.53 | 0.00 | 0.54 | 0.39 | 0.53 | 0.54 | 0.31 | 0.38 | 0.00 |
|
| 0.31 | 0.00 | 0.23 | 0.17 | 0.20 | 0.31 | 0.10 | 0.17 | 0.00 |
|
| 0.42–0.62 | 1.00 | 0.63–0.77 | −0.39–0.15 | 0.63–0.67 | 0.10–0.18 | 0.68–0.81 | −0.10–0.19 | 1.00 |
Note: Feature numbers correspond to those listed in Table 1.
| Domain | Feature | Score | Criteria |
|---|---|---|---|
|
| a. Clarity of proposed goal | 2 | Product describes at least one mental health goal which is specific, measurable, achievable (e.g. reduce stress, reduce PTSD symptoms) |
| 1 | Product describes non-specific or hard to measure mental health goals (e.g. improve your life, improve your wellbeing) | ||
| 0 | No clear goals | ||
|
| |||
|
| a. Number/average ratings | 2 | Ratings exist from >1500 users with an average rating of 3.5+ |
| 1 | Ratings exist from 31–1500 users with an average rating of 3.5+ | ||
| 0 | Fewer than 30 user rating OR an average rating below 3.5 | ||
|
| |||
|
| a. Direct research evidence | 3 | Strong research support for the product (at least two between-group design experiments that show efficacy or effectiveness) |
| 2 | Some research support for the product (at least one experiment that shows efficacy or effectiveness) | ||
| 1 | Other research (e.g. single case designs, quasi-experimental methods demonstrating efficacy, or preliminary analyses) | ||
| 0 | No research | ||
| b. Indirect research evidence | 1 | The app uses evidence-based practices to achieve its goals | |
|
| 0 | The app does not use evidence-based practices to achieve its goals (or there are no goals described) | |
| c. Research Independence & Review | 2 | At least one research paper funded by government agency (e.g. NIH) or non-profit organization OR two articles published in peer-reviewed journals | |
| 1 | All research funded primarily by for-profit organizations or combined funding sources OR one article published in a peer-reviewed journal | ||
| 0 | No information about source of funding for the research AND No published, peer-reviewed papers | ||
|
| a. Development processes | 1 | Pilot, feasibility and acceptability data OR evidence of stakeholder |
| engagement in development | |||
| 0 | No pilot, feasibility and acceptability data AND no evidence of stakeholder engagement | ||
| b. Efficacy of Other Products | 1 | Developer/development team has developed other mental health interventions delivered via technological medium which demonstrate efficacy | |
| 0 | No other mental health technological interventions demonstrating efficacy have been developed by this team | ||
| c. Clinical Input in Development | 1 | Clinical leader with mental health expertise involved in development | |
| 0 | No clinical leader with mental health expertise involved in development | ||
| d. Ongoing maintenance & updates | 2 | The application has been revised within the last 6 months | |
Scoring Instructions
For mobile applications: Assign a score for each feature. Add each feature score to obtain total score. No items need to be reverse coded. To normalize to a 5 point scale, divide total score by 3.
For web-based tools: Omit items 2a and 4d. Assign a score for each feature. Add each feature score to obtain total score. No items need to be reverse coded. To normalize to a 5 point scale, multiply total score by .