Literature DB >> 27287964

Development and Validation of the User Version of the Mobile Application Rating Scale (uMARS).

Stoyan R Stoyanov1, Leanne Hides, David J Kavanagh, Hollie Wilson.   

Abstract

BACKGROUND: The Mobile Application Rating Scale (MARS) provides a reliable method to assess the quality of mobile health (mHealth) apps. However, training and expertise in mHealth and the relevant health field is required to administer it.
OBJECTIVE: This study describes the development and reliability testing of an end-user version of the MARS (uMARS).
METHODS: The MARS was simplified and piloted with 13 young people to create the uMARS. The internal consistency and test-retest reliability of the uMARS was then examined in a second sample of 164 young people participating in a randomized controlled trial of a mHealth app. App ratings were collected using the uMARS at 1-, 3,- and 6-month follow up.
RESULTS: The uMARS had excellent internal consistency (alpha = .90), with high individual alphas for all subscales. The total score and subscales had good test-retest reliability over both 1-2 months and 3 months.
CONCLUSIONS: The uMARS is a simple tool that can be reliably used by end-users to assess the quality of mHealth apps.

Entities:  

Keywords:  Australia; Internet; MARS; RCT; anxiety; anxiety disorders; app evaluation; app rating; app trial; cellphone; cognitive behavioral therapy; depression; depressive disorder; e-therapy; eHealth; ehealth; emental health; end user; evidence-informed; mHealth; mHealth evaluation; mHealth implementation; mental health; mhealth trial; mobile application; mobile health; online; randomized controlled trial; reliability; research translation; smartphone; telemedicine; user testing; well being

Year:  2016        PMID: 27287964      PMCID: PMC4920963          DOI: 10.2196/mhealth.5849

Source DB:  PubMed          Journal:  JMIR Mhealth Uhealth        ISSN: 2291-5222            Impact factor:   4.773


Introduction

Mobile health (mHealth) apps have the potential to provide around-the-clock access to evidence-based health information, education, and treatment to end users on a global scale. There are currently more than 165,000 mHealth apps (free and paid) publicly available [1], yet the accuracy of the health information contained in these apps is not scrutinized by regulatory bodies [2], which could compromise user health and safety [3-5]. Concerns about the quality, efficacy, reliability, and security of mHealth apps are also often raised. While meta-analytic studies have demonstrated the efficacy of mHealth apps targeting physical activity and weight loss [6,7], the evidence base for other types of mHealth apps is poor at best [8-10]. In response to these issues, we developed the Mobile App Rating Scale (MARS) to provide researchers, professionals, and clinicians with a brief tool for classifying and assessing the quality of mHealth apps [11]. The 23-item MARS contains 4 objective quality subscales—engagement, functionality, aesthetics, and information quality—and a subjective quality rating. The MARS has demonstrated high levels of interrater reliability for evaluating the quality of mHealth apps on well-being [11] and mindfulness [9]. However, training and expertise in mHealth and the relevant health field is required to administer it. This paper describes the development and pilot testing of a simpler, end user version of the MARS (uMARS) and provides preliminary evidence for its internal consistency and test-retest reliability.

Methods

Study 1: Development and Pilot Testing of the uMARS

The original MARS was simplified through the following process. The professional version was first reviewed by 2 researchers to remove complex terminology from its items and response scales. Three items requiring professional expertise, pertaining to evidence base, app goals, and accuracy of app description, were removed. Readability of the MARS and the draft uMARS was then determined using the Flesch Reading Ease test [12,13], which has a score range of 0-100, with higher scores indicating easier readability. This measure also provides the estimated US school grade required for reading comprehension. The draft uMARS was then pilot-tested with 13 young people, to ensure they understood the item content and response scales. The measure was embedded in prototype testing sessions of 2 mHealth apps: Ray’s Night Out [14] and Music eScape [15]. Ray’s Night Out uses a harm-minimization approach to increase young people’s alcohol knowledge and awareness of their drinking limits; Music eScape teaches young people how to identify and manage affect using music. Both are available on the iOS Apple app store. Eligible participants were Australian residents aged 16 to 25 years, who had access to an iPhone 4 or later model. The Ray’s Night Out group comprised 1 male and 8 females with a mean age of 20.7 years (SD 1.6). The Music eScape group comprised 3 males and 1 female, with a mean age of 21.5 years (SD 1.9). After testing the apps and rating them with the uMARS scale, participants were asked the question “Do you have any comments or suggestions about the uMARS rating scale?” to identify any unclear or difficult items.

Study 2: Testing the uMARS Internal Consistency and Test-Retest Reliability

The uMARS (Multimedia Appendix 1) provides a 20-item measure that includes 4 objective quality subscales—engagement, functionality, aesthetics, and information quality—and 1 subjective quality subscale. The reliability of the uMARS was evaluated as part of a randomized controlled trial (RCT), testing the efficacy and quality of Music eScape. The RCT sample comprised 164 Australians aged 19.8 years on average (SD 2.51); 34 males. The highest level of education completed by 59.8% of the sample was secondary school, and 24.4% had completed a bachelor’s degree or higher. Most participants (57.9%) were students and 35.4% had full-time, part-time, or casual employment. Participants were randomly allocated via a Web-based research management tool developed at the Queensland University of Technology to receive immediate or 1-month delayed access to the Music eScape app. Young people were asked to use the app as much as they liked over a month, using their own iPhones. The current iOS version at the time of the trial was iOS8. Participants received weekly text messages reminding them to do so. App ratings were collected using the uMARS at 1-, 3-, and 6-month follow-ups in the immediate access group. In the delayed access group, uMARS ratings were collected at 2-, 3-, and 6-month follow-ups (ie, after 1, 2, and 5 months of app access). At each assessment point, participants were asked if they had used the app since the last assessment, and only those who reported some use were included in analyses.

Data Analysis

The internal consistencies of the uMARS subscales and total score were calculated using Cronbach's alpha. For the purpose of analysis, the “N/A” answer option for items 13-16 of the information subscale was recoded as “system missing,” as this option represents a qualitatively different response. Test-retest reliabilities were calculated for the subscales and total scores of the uMARS after 1 month of app use and at 3 months post baseline (ie, a test-retest period of 1-2 months), and over 3 months (ie, between assessments at 3 and 6 months post baseline). Interclass correlation coefficients (ICCs) [16-18] were used, as they provide weighted values of rater agreement and assess proximity rather than equality of ratings. To calculate the ICCs, a random-effects average measures model with absolute agreement was utilized [16]. Data were analyzed with SPSS version 23 (SPSS Inc, Armonk, NY, USA).

Results

Study 1: Readability

Results of the Flesch-Kincaid readability tests are in Table 1. Scores indicated that the uMARS was written in plain English and that its required reading level was approximately grade 8.
Table 1

Readability ease and grade level scores of the original Mobile App Rating Scale and the simplified user version of the scale (uMARS).

MARSa versionReading easeReadability levelGrade levelReading age
Original47.2Difficult9.515-16 years old
uMARSa58.0Plain English – fairly difficult7.912-13 years old

a MARS: Mobile App Rating Scale; uMARS: user version of the MARS.

Pilot Participant Feedback

No suggestions for further scale improvement were made. Seven of the 13 participants who pilot-tested the scale left the comments or suggestions item blank, 4 wrote “no,” and 1 wrote “Well done. Good questions. Well explained.” Another wrote “I thought it was shorter/there is a brief or revised version of it?”

Study 2: uMARS Internal Consistency

A total of 152 of the 164 (92%) participants completed the survey after 1 month of app use. Of these, 19 indicated they never used the app, and were excluded from analyses. For the remaining 133 participants, the total uMARS score had excellent internal consistency (Cronbach alpha = .90). Internal consistencies of its subscales were also very high (engagement alpha = .80; functionality alpha = .70; aesthetics alpha = .71; information alpha = .78; and satisfaction alpha = .78).

uMARS Test-Retest Reliability

Test-retest reliabilities are presented in Table 2. A total of 113 participants completed the scale after 1 month of app use and at 3 months post baseline (ie, a test-retest period of 1-2 months), and 74 completed both the 3- and 6-month surveys (giving a 3-month test-retest period). All included participants had used the app at least once since the previous survey. The uMARS total score demonstrated good [16] levels of ICC of .66 and .70 over 1- to 2-month and 3-month periods, respectively. Levels for all subscales scores were similarly high.
Table 2

Test-retest reliability of the user version of the Mobile App Rating Scale (95% CI).

Subscale/item1- to 2-month period (N=113)3-month period (N=74)
Engagement.71 (.66-.76).73 (.67-.78)
1Entertainment.60 (.41-.72).75 (.61-.85)
2Interest.69 (.55-.79).67 (.48-.79)
3Customization.61 (.44-.73).53 (.25-.70)
4Interactivity.55 (.35-.69).69 (.51-.81)
5Target group.72 (.59-.80).73 (.57-.83)
Functionality.62 (.54-.68).69 (.61-.76)
6Performance.54 (.34-.69).71 (.53-.81)
7Ease of use.65 (.49-.76).72 (.55-.82)
8Navigation.62 (.45-.74).67 (.48-.79)
9Gestural design.61 (.44-.73).65 (.44-.78)
Aesthetics.58 (.48-.66).68 (.59-.76)
10Layout.39 (.11-.58).48 (.18-.67)
11Graphics.70 (.56-.79).77 (.63-.85)
12Visual appeal.63 (.46-.75).80 (.68-.87)
Information.48 (.38-.57).52 (.40-.62)
13Quality of information.48 (.24-.64).44 (.11-.65)
14Quantity of information.48 (.24-.64).32 (.08 to .57)
15Visual information.42 (.16-.60).75 (.61-.84)
16Credibility of source.51 (.29-.66).63 (.41-.77)
Total uMARSa.66 (.63-.68).70 (.67-.78)
Subjective items.70 (.64-.75).71 (.64-.77)
17Would you recommend.84 (.76-.89).75 (.60-.84)
18How many times.44 (.18-.61).48 (.17-.67)
19Would you pay.81 (.73-.87).82 (.71-.89)
20Overall (star) rating.71 (.59-.80).77 (.63-.85)

a uMARS: user version of the Mobile App Rating Scale.

Readability ease and grade level scores of the original Mobile App Rating Scale and the simplified user version of the scale (uMARS). a MARS: Mobile App Rating Scale; uMARS: user version of the MARS. Test-retest reliability of the user version of the Mobile App Rating Scale (95% CI). a uMARS: user version of the Mobile App Rating Scale.

Discussion

This study developed and tested an app user version of the original MARS to assist app developers and researchers with evaluating the quality of mHealth apps. The uMARS (Multimedia Appendix 1) provides a 20-item measure that includes 4 objective quality subscales—engagement, functionality, aesthetics, and information quality—and 1 subjective quality subscale. One further subscale, consisting of 6 items is added to measure users’ perceived impact of the evaluated app. The study demonstrated that the uMARS had excellent internal consistency for the full scale and good levels for all subscales. It is reassuring that even after a 3-month delay between ratings, test-retest reliability of the total score was good, and test-retest reliabilities of its subscales were fair to good, with the engagement and subjective subscales being particularly robust. These results indicate that the uMARS provides a reliable measure of app quality in target users. Replication of the current results with multiple types of mHealth apps is required to provide additional confidence in its performance. Tests of its sensitivity to improvements in app quality and an examination of its ability to predict outcomes of mHealth apps are also needed. As the uMARS may potentially have applications beyond mHealth, tests of its performance in other domains are also indicated. Current indications are that the uMARS will offer an unprecedented ability to readily obtain rich information from users about mobile apps. The scale can be used to obtain user feedback on the quality of mobile apps during the development and testing process, which may result in overall improvements in their quality.
  13 in total

1.  Feature-level analysis of a novel smartphone application for smoking cessation.

Authors:  Jaimee L Heffner; Roger Vilardaga; Laina D Mercer; Julie A Kientz; Jonathan B Bricker
Journal:  Am J Drug Alcohol Abuse       Date:  2014-11-14       Impact factor: 3.829

2.  A new readability yardstick.

Authors:  R FLESCH
Journal:  J Appl Psychol       Date:  1948-06

Review 3.  Intraclass correlations: uses in assessing rater reliability.

Authors:  P E Shrout; J L Fleiss
Journal:  Psychol Bull       Date:  1979-03       Impact factor: 17.737

4.  Towards a Framework for Evaluating Mobile Mental Health Apps.

Authors:  Steven Chan; John Torous; Ladson Hinton; Peter Yellowlees
Journal:  Telemed J E Health       Date:  2015-07-14       Impact factor: 3.536

5.  Guided Internet-based vs. face-to-face cognitive behavior therapy for psychiatric and somatic disorders: a systematic review and meta-analysis.

Authors:  Gerhard Andersson; Pim Cuijpers; Per Carlbring; Heleen Riper; Erik Hedman
Journal:  World Psychiatry       Date:  2014-10       Impact factor: 49.548

6.  Computing Inter-Rater Reliability for Observational Data: An Overview and Tutorial.

Authors:  Kevin A Hallgren
Journal:  Tutor Quant Methods Psychol       Date:  2012

7.  Review and Evaluation of Mindfulness-Based iPhone Apps.

Authors:  Madhavan Mani; David J Kavanagh; Leanne Hides; Stoyan R Stoyanov
Journal:  JMIR Mhealth Uhealth       Date:  2015-08-19       Impact factor: 4.773

Review 8.  The effectiveness of mobile-health technologies to improve health care service delivery processes: a systematic review and meta-analysis.

Authors:  Caroline Free; Gemma Phillips; Louise Watson; Leandro Galli; Lambert Felix; Phil Edwards; Vikram Patel; Andy Haines
Journal:  PLoS Med       Date:  2013-01-15       Impact factor: 11.069

9.  Increasing physical activity with mobile devices: a meta-analysis.

Authors:  Jason Fanning; Sean P Mullen; Edward McAuley
Journal:  J Med Internet Res       Date:  2012-11-21       Impact factor: 5.428

Review 10.  Mapping mHealth research: a decade of evolution.

Authors:  Maddalena Fiordelli; Nicola Diviani; Peter J Schulz
Journal:  J Med Internet Res       Date:  2013-05-21       Impact factor: 5.428

View more
  140 in total

Review 1.  Review of Use and Integration of Mobile Apps Into Psychiatric Treatments.

Authors:  Steven Chan; Haley Godwin; Alvaro Gonzalez; Peter M Yellowlees; Donald M Hilty
Journal:  Curr Psychiatry Rep       Date:  2017-10-30       Impact factor: 5.285

2.  Criteria for assessing the quality of mHealth apps: a systematic review.

Authors:  Rasool Nouri; Sharareh R Niakan Kalhori; Marjan Ghazisaeedi; Guillaume Marchand; Mobin Yasini
Journal:  J Am Med Inform Assoc       Date:  2018-08-01       Impact factor: 4.497

3.  Content and Usability Evaluation of Medication Adherence Mobile Applications for Use in Pediatrics.

Authors:  Julia K Carmody; Lee A Denson; Kevin A Hommel
Journal:  J Pediatr Psychol       Date:  2019-04-01

Review 4.  'Prostate Cancer Risk Calculator' mobile applications (Apps): a systematic review and scoring using the validated user version of the Mobile Application Rating Scale (uMARS).

Authors:  Ahmed Adam; Julian C Hellig; Marlon Perera; Damien Bolton; Nathan Lawrentschuk
Journal:  World J Urol       Date:  2017-12-08       Impact factor: 4.226

Review 5.  Computer-Assisted Cognitive-Behavior Therapy and Mobile Apps for Depression and Anxiety.

Authors:  Jesse H Wright; Matthew Mishkind; Tracy D Eells; Steven R Chan
Journal:  Curr Psychiatry Rep       Date:  2019-06-27       Impact factor: 5.285

6.  Review of the Content and Quality of Mobile Applications About Alzheimer's Disease and Related Dementias.

Authors:  Seul Ki Choi; Brooks Yelton; Victor K Ezeanya; Kristie Kannaley; Daniela B Friedman
Journal:  J Appl Gerontol       Date:  2018-07-26

7.  Acceptability and Use of a Dual HIV/Syphilis Rapid Test and Accompanying Smartphone App to Facilitate Self- and Partner-Testing Among Cisgender Men and Transgender Women Who Have Sex with Men.

Authors:  Iván C Balán; Javier López Rios; Cody Lentz; Siddarth Arumugam; Curtis Dolezal; Bryan Kutner; Christine Tagliaferri Rael; Alexander Wang Ying; Ongun Uzay Macar; Samuel K Sia
Journal:  AIDS Behav       Date:  2021-05-26

8.  Clinical and Cost-Effectiveness of PSYCHOnlineTHERAPY: Study Protocol of a Multicenter Blended Outpatient Psychotherapy Cluster Randomized Controlled Trial for Patients With Depressive and Anxiety Disorders.

Authors:  Harald Baumeister; Natalie Bauereiss; Anna-Carlotta Zarski; Lina Braun; Claudia Buntrock; Christian Hoherz; Abdul Rahman Idrees; Robin Kraft; Pauline Meyer; Tran Bao Dat Nguyen; Rüdiger Pryss; Manfred Reichert; Theresa Sextl; Maria Steinhoff; Lena Stenzel; Lena Steubl; Yannik Terhorst; Ingrid Titzler; David Daniel Ebert
Journal:  Front Psychiatry       Date:  2021-05-14       Impact factor: 4.157

9.  Parenting apps review: in search of good quality apps.

Authors:  Anila Virani; Linda Duffett-Leger; Nicole Letourneau
Journal:  Mhealth       Date:  2019-09-30

10.  Mental health apps in a college setting: openness, usage, and attitudes.

Authors:  Adam Kern; Victor Hong; Joyce Song; Sarah Ketchen Lipson; Daniel Eisenberg
Journal:  Mhealth       Date:  2018-06-30
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.