Literature DB >> 33718002

Systematic evaluation of content and quality of English and German pain apps in European app stores.

Yannik Terhorst1,2, Eva-Maria Messner2, Dana Schultchen3, Sarah Paganini4, Alexandra Portenhauser2, Anna-Sophia Eder2, Melanie Bauer3, Mike Papenhoff5, Harald Baumeister2, Lasse Bosse Sander6.   

Abstract

BACKGROUND AND
OBJECTIVE: Pain spans a broad spectrum of diseases and types that are highly prevalent and cause substantial disease burden for individuals and society. Up to 40% of people affected by pain receive no or inadequate treatment. Providing a scalable, time-, and location-independent way for pain diagnostic, management, prevention and treatment mobile health applications (MHA) might be a promising approach to improve health care for pain. However, the commercial app market is rapidly growing and unregulated, resulting in an opaque market. Studies investigating the content, privacy and security features, quality and scientific evidence of the available apps are highly needed, to guide patients and clinicians to high quality MHA.Contributing to this challenge, the present study investigates the content, quality, and privacy features of pain apps available in the European app stores.
METHODS: An automated search engine was used to identify pain apps in the European Google Play and Apple App store. Pain apps were screened and checked for systematic criteria (pain-relatedness, functionality, availability, independent usability, English or German). Content, quality and privacy features were assessed by two independent reviewers using the German Mobile Application Rating Scale (MARS-G). The MARS-G assesses quality on four objectives (engagement, functionality, aesthetics, information quality) and two subjective scales (perceived impact, subjective quality).
RESULTS: Out of 1034 identified pain apps 218 were included. Pain apps covered eight different pain types. Content included basic information, advice, assessment and tracking, and stand-alone interventions. The overall quality of the pain apps was average M = 3.13 (SD = 0.56, min = 1, max = 4.69). The effectiveness of less than 1% of the included pain apps was evaluated in a randomized controlled trial. Major problems with data privacy were present: 59% provided no imprint, 70% had no visible privacy policy.
CONCLUSION: A multitude of pain apps is available. Most MHA lack scientific evaluation and have serious privacy issues, posing a potential threat to users. Further research on evidence and improvements privacy and security are needed. Overall, the potential of pain apps is not exploited.
© 2021 The Authors. Published by Elsevier B.V.

Entities:  

Keywords:  MARS; Mobile health apps; Pain; e-Health

Year:  2021        PMID: 33718002      PMCID: PMC7933737          DOI: 10.1016/j.invent.2021.100376

Source DB:  PubMed          Journal:  Internet Interv        ISSN: 2214-7829


Introduction

Pain is a highly prevalent and manifold condition (Breivik et al., 2006; Jackson et al., 2016; Leadley et al., 2012). According to the global burden of disease study, pain is the number one cause for years lived with disability (YLD) (James et al., 2018). Low back pain alone caused 65 Million YLDs in the year 2017 (James et al., 2018), and is associated with high societal costs (Baumeister et al., 2012; Breivik et al., 2013; Gaskin and Richard, 2012). Multiple effective treatment options for pain exist (e.g. pharmacological, interventional, physiological or psychological treatments) (Turk et al., 2011; Veehof et al., 2016). Pain severity often results from a complex interaction between 1) pathological mechanisms, 2) past pain experiences and social environment, 3) cognitive factors and 4) emotional factors (Turk et al., 2011). Thus, an interdisciplinary and multidimensional treatment approach has been established as the gold standard (Kumar, 2007; Traue et al., 2013). However, 40% of affected individuals do not receive adequate treatment and are unsatisfied with the effectiveness of their treatment of long-lasting pain (Breivik et al., 2006). Insufficient availability and involvement of pain specialist, inadequate medication, infrastructural barriers (e.g., traveling time), as well as personal barriers (e.g., limited time resources) contribute to insufficient health care (Becker et al., 2017; Breivik et al., 2006, Breivik et al., 2013; Rod, 2016). Given their time- and location-independent nature, mobile health applications (MHA) might provide an opportunity to improve health care for people with pain (Albrecht, 2016; Donker et al., 2013; Linardon et al., 2019; Thurnheer et al., 2018; Weisel et al., 2019). Pain apps can be used in various ways to improve the current healthcare situation, including 1) diagnostics, 2) screening, 3) self-monitoring, 4) referral, 5) education, 6) pain-management, and 7) social support (Albrecht, 2016; Lalloo et al., 2015, Lalloo et al., 2017), accompanying on-site treatments or being used as standalone interventions (Ebert et al., 2017, Ebert et al., 2018; Messner et al., 2019a). This might be in particular a valuable addition to healthcare for pain during times of lockdown, social distancing, and other quarantine measurements that affect pain treatment and management (El-Tallawy et al., 2020). However, besides their structural advantages MHA and pain apps can also offer anonymous health care to individuals who fear stigmatization that may not seek healthcare otherwise (Andrade et al., 2014). The commercial app market is rapidly growing and multiple pain apps are available (Lalloo et al., 2015; Salazar et al., 2018; Wallace and Dhingra, 2014) and other health conditions (e.g. depression, anxiety, rheumatism, PTSD: (Knitza et al., 2019; Sander et al., 2019; Sucala et al., 2017; Terhorst et al., 2018, Terhorst et al., 2020)). In contrast to the multitude of apps available in the app stores, there commonly is no or only limited evidence for their effectiveness (Donker et al., 2013; Rathner and Probst, 2018; Sucala et al., 2017; Terhorst et al., 2018; Thurnheer et al., 2018). Besides the lack of evidence, several studies have highlighted issues with privacy policies (e.g., missing or insufficient) (Grundy et al., 2019; Huckvale et al., 2019), lack of health care professional involvement(Lalloo et al., 2015; Wallace and Dhingra, 2014) or negative side-effects in case of device failure (Luxton et al., 2011). Another concern is the limited availability of information about MHA in the commercial app stores: App description, user star ratings, and user comments are often the only, but an unreliable source for users to assess MHA content and quality (Armstrong, 2015; Shen et al., 2015; Stoyanov et al., 2015, Stoyanov et al., 2016; Terhorst et al., 2018): For users, perceived MHA quality is rather led by perception of MHA look and feel than on information and content quality, which are paramount from clinicians perspective (Armstrong, 2015; Bardus et al., 2016; Nicholas et al., 2015; Stoyanov et al., 2016; Terhorst et al., 2018). Consequently, users and health care providers are experiencing difficulties in identifying MHA suiting their conditions and needs (Shen et al., 2015). Hence, studies investigating the content, privacy and security features, quality and scientific evidence of the available apps are highly needed, to guide patients and clinicians to high quality MHA (Donker et al., 2013; Knitza et al., 2019; Sander et al., 2020; Schoeppe et al., 2017; Shen et al., 2015; Terhorst et al., 2018). In prior investigations Wallace and Dhingra (2014), and Lalloo et al. (2015) provided a comprehensive overview of the content of pain apps. However, they did not review the quality of MHA (Lalloo et al., 2015; Wallace and Dhingra, 2014). Extending the content analysis, Salazar et al. (2018) provided an overview on quality and content of 18 pain apps limited to MHA in the Spanish app stores. Overall, they reported average quality of the included MHA with the lowest quality in the dimensions engagement and information quality. However, next to the Spanish focus, their search term was limited to generic search term “pain” (both in English and Spanish), which seems too narrow given the broad spectrum of pain disorders. This limitation is also true for the previously mentioned reviews by Wallace and Dhingra (2014) (search term: “pain”), and Lalloo et al. (2015) (search term: “pain” and “pain management”). Hence, this study aims to extend the previous findings by an improved search, a more detailed evaluation of privacy and content features, reviewing the quality of pain apps, and investigating the relationship between standardized expert ratings of quality and user-star ratings. The following questions will be addressed: Which content is provided by pain apps in the European commercial app stores? What measures have been taken to ensure privacy and data protection? Of which quality are pain apps in the European commercial app stores regarding user engagement, functionality, aesthetics, and information content? Is the user star rating associated with the standardized MARS-G quality rating?

Methods

Search strategy and inclusion criteria

An automatic search engine (webcrawler) developed in context of the mobile health app database project (MHAD; http://mhad.science/) (Messner et al., 2021; Meßner et al., 2019; Paganini et al., 2019; Portenhauser et al., 2021; Sander et al., 2019; Schultchen et al., 2020; Terhorst et al., 2018) was used to systematically search the European Google Play and Apple App store for pain apps. The webcrawler searches and automatically extracts the available information from the app stores. Search terms included were: “chronic pain”, “pain”, “ache”, “migraine”, “headache”, “rheumatoid arthritis”, “back pain” and “pain diary”. All identified apps were listed in a central database, and duplicates were automatically removed. The search was conducted on October 24th, 2018. After identification, a two-step procedure was applied to assess the eligibility of identified apps: In a first step, titles and descriptions were screened by one reviewer (either DS, LS, MB, YT). Inclusion criteria were: (a) the app was developed for a pain-related context, (b) the app was available in a language spoken by the reviewer team (English, German), (c) the app was available for download, (d) the app was usable independently (e.g. access not restricted to study participation or a specific clinic). In a second step, all eligible apps were downloaded and checked regarding the criteria on basis of the actual app content. Not working (e.g., unable to start on the device) pain apps were excluded.

App characteristics

The Mobile Application Rating Scale (MARS) (Messner et al., 2019b; Stoyanov et al., 2015) was used to capture app characteristics: (a) app name, (b) store link (c) platform (android or iOS), (d) content-related subcategory, (e) aims, (f) price, (g) user rating.

Content and functions

The classification site of the MARS was used to assess content and functions of included pain apps (Messner et al., 2019b; Stoyanov et al., 2015).

Privacy and security features

The MARS comes with an assessment of privacy and security features on a descriptive level. All features were assessed based on the downloaded pain apps. External information (e.g. developer web-site) were not investigated, if not referenced within the app.

Quality rating

The Mobile Application Rating Scale – German (MARS-G), a translated version of the original MARS (Stoyanov et al., 2015) was used for the quality rating (Messner et al., 2019b). The MARS shows good to excellent between rater agreement, reliability, construct validity, and concurrent validity (Messner et al., 2019b; Stoyanov et al., 2015; Terhorst et al., 2020). Quality was rated on a 5-point scale (1-inadequate, 2-poor, 3-acceptable, 4-good, and 5-excellent). The MARS-G includes 19 items divided into four sub-dimensions: (A) engagement (5 items: fun, interest, individual adaptability, interactivity, target group), (B) functionality (4 items: performance, usability, navigation, gestural design), (C) aesthetics (3 items: layout, graphics, visual appeal), and (D) information quality (7 items: accuracy of app description, goals, quality of information, quantity of information, quality of visual information, credibility, evidence base). Item 19 on the information subscale evaluates the evidence base of a given app (“Has the app been trialled/tested; must be verified by evidence (in published scientific literature)?”). Studies providing evidence for an app were identified by app description, developer's and provider's web-sites, and searching the app name in google scholar and PubMed. Mean scores were calculated for each dimension separately. The overall score represents the mean of the engagement, functionality, aesthetics and information quality scales. In addition to the objective quality scales of the MARS, the MARS includes two subjective scales: (E) subjective quality (4 items: recommendation, frequency of use, willingness to pay, overall star-rating) and (F) perceived impact (6 items: awareness, knowledge, attitudes, intention to change, help-seeking, behavioural change).

Rater training and agreement

Psychologists (at least BSc level) conducted the quality assessment. Throughout the assessment, all raters were supervised by DS or LS (licensed psychotherapist). Both supervisors had experience with using the MARS (Messner et al., 2021; Portenhauser et al., 2021; Sander et al., 2019; Schultchen et al., 2020; Terhorst et al., 2018). All raters underwent standardized online training before the quality assessment (https://youtu.be/5vwMiCWC0Sc, last updated on January 2021). All pain apps were rated by two independent raters (MB, SP, AP, ASE, DS, YT, LS). Agreement between raters was evaluated by intra-class correlation (ICC) (Koo and Li, 2016). A cut-off of ICC above 0.75 was defined as the criteria for a satisfactory rater agreement (Fleiss, 1999).

Associations between user star rating and quality

Associations between MARS ratings and user star-ratings were investigated by means of correlations.

Data analyses

Rater agreement was examined by intra-class correlation based on a two-way mixed-effects model (Koo and Li, 2016). Consistency between raters was evaluated. For correlations analysis an alpha-level of 5% was defined. P-values were adjusted using the procedure proposed by Holm (1979).

Results

Search and app characteristics

1034 apps were identified (Apple App Store = 693, Play Store = 341). After screening and eligibility check, a total of 218 pain apps were eligible for inclusion (Apple App Store = 110, Play Store = 108) (see Fig. 1).
Fig. 1

Flow-chart of the in- and exclusion process.

Flow-chart of the in- and exclusion process. Of all 218 included pain apps, 108 pain apps had a user-star-rating (M = 4.05; SD = 0.69), while 110 were not rated by users. The majority of pain apps were free of charge (n = 170, 78%). Prices of fee-based pain apps ranged from 1.09 EUR to 21.99 EUR (M = 3.88 EUR, SD = 3.17). Pain apps were identified in eight store categories (multiple categories can be assigned to an app): “health and fitness” (n = 145, 67%), “medical” (n = 125, 57%), “lifestyle” (n = 14, 6%), “education” (n = 8, 4%), “books and reference” (n = 5, 2%), “news and magazines” (n = 3, 1%), “sport” (n = 1, 1%), “entertainment” (n = 1, 1%), “utilities” (n = 1, 1%). The included pain apps target a broad range of areas. Most pain apps focused on headache and migraine (n = 59, 27%), back pain (n = 53, 24%), arthritis and rheumatism (n = 30, 14%), chronic pain (n = 18, 8%), general/unspecified pain (n = 18, 8%), fibromyalgia (n = 3, 1%), neck (n = 2, 1%) and other specific pain types (n = 4, 2%). Pain apps offering a pain assessment or diary function without focus on a specific pain type were grouped under “Diary & Assessment” (n = 31, 14%). Overall most of the pain apps aimed for improved physical fitness (n = 151, 69%), general well-being (n = 130, 60%), goal-setting (n = 35, 16%), behavior change (n = 22, 10%), reduction of stress (n = 9, 4%), emotion regulation (n = 5, 2%), social-behavior (n = 5, 2%), reduction of anxiety (n = 4, 2%), entertainment (n = 4, 2%), reduction of depression (n = 3, 1%), addiction (n = 1, <1%) and other (n = 58, 27%). For a more nuanced representation of the aims in respect to the type of pain (e.g., back pain, headache and migraine) see Multimedia Appendix 1. Almost half (N = 106, 49%) of MHA provided information about pain, 47% (N = 102) offered advice, 44% (N = 95) included a tracking feature, and 29% (N = 64) included an assessment tool. Furthermore, frequent features were physical exercise (N = 59, 27%), feedback (N = 40, 18%), strategies (N = 36, 17%), and reminder (N = 35, 16%). An overview of the content and functions grouped by the type of pain is presented in Multimedia Appendix 1. Of all included pain apps, n = 76 (35%) had a legal disclosure, n = 55 (25%) a visible privacy policy, n = 32 (15%) required a log-in, n = 33 (15%) offered password protection, n = 21 (10%) informed about conflict of interests/financial background. Consent to data collection was required in an active form by n = 26 pain apps (12%) and in a passive form by 20 pain apps (9%). Of the pain apps including assessment (n = 56), active consent was required by 9 (16%), and passive consent was present in 13 (23%) pain apps. Agreement between raters was excellent (ICC = 0.84, 95%-CI: 0.83 to 0.85). The overall quality of pain apps was average M = 3.13 (SD = 0.56), ranging from 1.00 to 4.69. Quality ratings of the four sub-dimensions yielded: functionality M = 3.66 (SD = 0.62), aesthetics M = 3.11 (SD = 0.73), information quality M = 2.93 (SD = 0.69) and engagement M = 2.81 (SD = 0.75). The quality per type of pain is summarized in Table 1. A more detailed summary of pain apps with at least good overall quality (>4.0) is presented in Table 2. For the quality ratings of all pain apps, their content and privacy features see Multimedia Appendix A.
Table 1

Quality of MHA.

Area of applicationFrequencyEngagementFunctionalityAestheticsInformationOverall
Arthritis & rheumatism302.97 (0.77)3.70 (0.63)3.18 (0.78)3.39 (0.74)3.31 (0.60)
Back pain532.44 (0.71)3.44 (0.61)2.77 (0.83)2.71 (0.63)2.84 (0.58)
Chronic pain182.88 (0.74)3.65 (0.51)3.04 (0.86)2.83 (0.61)3.10 (0.55)
Diary & assessment313.06 (0.63)3.69 (0.55)3.36 (0.53)2.91 (0.61)3.25 (0.48)
Fibromyalgia32.57 (0.25)3.92 (0.62)2.89 (0.75)2.96 (1.04)3.08 (0.62)
General/unspecific182.94 (0.79)3.70 (0.57)3.28 (0.50)3.29 (0.68)3.30 (0.46)
Headache & migraine592.92 (0.73)3.87 (0.56)3.28 (0.60)2.87 (0.64)3.24 (0.45)
Neck22.50 (0.14)4.06 (0.09)3.33 (0.00)2.92 (0.35)3.20(0.15)
Other42.58 (1.05)2.66 (1.26)2.46 (1.01)2.50 (1.08)2.55 (1.05)
Total218 (100%)2.81 (0.75)3.66 (0.62)3.11 (0.73)2.93 (0.69)3.13 (0.56)
Table 2

High-quality MHA.

AppPain areaContentEngagementFunctionalityAestheticsInformationOverallSubjective qualityPerceived impact
Jointfully OsteoarthritisArthritis & rheumaAssessment, tracking, feedback, information, advice, goalsetting, strategies, resources, physical, reminder, tailored, medicine4.304.004.334.434.273.634.08
RheumaLiveArthritis & rheumaTracking, feedback3.604.754.004.334.173.251.42
Kaia: Back Pain Relief at HomeBack painTracking, information, advice, strategies, alternative medicine, physical4.304.504.673.714.303.633.67
Schmerztagebuch [eng.: pain diary] & Community CatchMyPainDiary & assessmentAssessment, tracking, information, advice,3.904.754.174.384.303.633.50
Schmerztagebuch [eng.: pain diary] - Pain TracerDiary & assessmentAssessment, tracking3.604.504.174.174.113.502.08
InjurymapGeneralFeedback, information, advice, goalsetting, strategies, physical4.904.634.504.754.693.503.67
Migraine XplainedHeadache & migraineAssessment, tracking, information, reminder4.105.004.674.174.483.753.58
Migräne App (Play Store) [eng.:]Headache & migraineTracking, feedback, information, advice, reminder, medicine4.204.504.504.174.343.633.08
Migräne App (App Store)Headache & migraineAssessment, tracking, feedback, information, advice, goalsetting, relaxation4.404.634.174.134.333.634.00
Migraine BuddyHeadache & migraineTracking, feedback4.604.254.003.214.023.253.25
Besides the four objective dimension the MARS includes a scale for the subjective quality and the perceived impact on health. Across all pain apps the trained reviewers rated subjective quality M = 2.19 (SD = 0.70) and the perceived impact on health M = 2.06 (SD = 0.69) as poor. See Multimedia Appendix 1 for the app-wise ratings.

Quality rating on evidence

The evidence item of the information quality section indicated that only a single pain app (=“Kaia: Back Pain Relief at Home”) was tested in a randomized controlled trial (Toelle et al., 2019) and the same pain app was also tested in a cluster randomized controlled trial (cRCT) (Priebe et al., 2020) showing its superiority compared to the control conditions. Overall this translated to an RCT evidence for 0.5% of all examined pain apps. Pain apps rated with at least 3 (=“App has been trialled (e.g., acceptability, usability, satisfaction ratings) and has positive outcomes in studies that are not RCTs, and there is no contradictory evidence.”; (Stoyanov et al., 2015)) were the English “GeoPain” (https://www.geopain.com/) and the German “Migräne App” (https://schmerzklinik.de/die-migraene-app/). The evidence item of a total of 199 pain apps (91.3%) were rated as “not applicable”, since no information on evidence were found. Quality of MHA. High-quality MHA.

Associations between user-rating and quality

User ratings were available for 108 pain apps (49.5%). No significant correlations between user star-ratings and standardized expert ratings were found (p > .05; overall r = 0.03, engagement r = 0.04, functionality r = −0.08, aesthetics r = 0.01 and information quality r = 0.00).

Discussion

This study investigated the content, privacy and security features, quality of pain apps. Identified pain apps cover a broad range of pain disorders (e.g. back pain, migraine). Besides the application areas, the content and functions also varied: Most often, they provide information and advice as well as some form of assessment and tracking. Moreover, there were multiple apps that included treatment elements (e.g., relaxation as well as physical, mindfulness, or breathing exercises). Overall the quality of pain apps was acceptable. However, we could only find one RCT and one cRCT evaluating one of the identified pain apps (Priebe et al., 2020; Toelle et al., 2019). Furthermore, for the vast majority (91%) could not find any evidence on the effectiveness and safety of the examined pain apps. The lack of scientific evidence is in line with previous studies in the field of app evaluation (Donker et al., 2013; Meßner et al., 2019; Sander et al., 2019; Sucala et al., 2017; Terhorst et al., 2018, Terhorst et al., 2020). Besides the need for scientific evidence, the present findings indicate that the quality of available pain apps could be improved by increasing the engagement of users. Engagement of users reflected the dimension with the weakest score across all subscales. The lack of user engagement is a common problem in MHA across different disorders (Meßner et al., 2019; Sander et al., 2019; Terhorst et al., 2018). Gamification, tracking, and feedback to the users conceptualized as persuasive design features could be promising measurements to overcome this problem (Bakker et al., 2016; Baumeister et al., 2019; Baumel and Yom-Tov, 2018). Moreover, many pain apps would benefit from an improved quality of the provided information. Wrong or misleading information in MHA are a threat to users' safety (Albrecht, 2016; Terhorst et al., 2018). Only if MHA provide evidence-based and high-quality information, they could become a helpful component in health care (Albrecht, 2016; Donker et al., 2013; Terhorst et al., 2018). In this study, we did not find a significant correlation between user-rating and the MARS-G. Hence, the user-star-rating cannot be considered a reliable measurement to determine the quality of apps. This finding is in line with prior investigations and underlines the need for systematic evaluations guiding users and health care providers to MHA fitting their needs and requirements (Armstrong, 2015; Bardus et al., 2016; Nicholas et al., 2015; Shen et al., 2015; Terhorst et al., 2018). Since available information in the app stores (e.g., user rating, app description) are insufficient for a reliable assessment of MHA content and quality, other objective platforms to inform users as well as health care providers are needed. Systematic evaluations using reliable and objective instruments to assess content and quality (e.g. MARS (Stoyanov et al., 2015), ENLIGHT (Baumel et al., 2017), APA App Evaluation Model [https://www.psychiatry.org/psychiatrists/practice/mental-health-apps/app-evaluation-model]) system could function as the foundation of such platforms. Considering the amount of existing reviews, which cover a broad range of health conditions, the building of an objective research-based international information platform for MHA might already be feasible (Bardus et al., 2016; Grainger et al., 2017; Knitza et al., 2019; Machado et al., 2016; Mani et al., 2015; Masterson Creber et al., 2016; Meßner et al., 2019; Salazar et al., 2018; Sander et al., 2019; Santo et al., 2016; Schoeppe et al., 2017; Terhorst et al., 2018; Thornton et al., 2017). However, no international information platform exists and only country specific projects have started to provide evidence-based information such as the German Mobile Health App Database (MHAD; http://mhad.science/), the German HealthOn platform (https://www.healthon.de/) or the US platform PsyberGuide (https://psyberguide.org/). Equally important to adequate information about MHA quality and content is information about privacy and security (Grundy et al., 2019; Huckvale et al., 2019; Torous et al., 2019). According to international standards and the European Data Protection Regulation MHA should provide a visible privacy policy, an imprint with contact information, and an opt-in consent for all data collection (THE EUROPEAN PARILIAMENT and THE COUNCIL OF THE EUROPEAN UNION, 2018; Torous et al., 2019). Furthermore, users have to be informed about their rights (e.g., right to inspect collected personal data, right to delete personal data, right to change personal information). However, for most MHA, those criteria were not met. Before MHA become a part of routine healthcare, these criteria should be met to ensure patient safety. Regarding privacy and security, it has to be kept in mind, that we only assessed the privacy and data security on a descriptive level. In a recent assessment of data sharing and privacy practices of MHA for depression and smoking cessation, a more elaborated procedure was applied, showing not only the lack of privacy policy, but also inadequate or insufficient information in present privacy policies (Huckvale et al., 2019). Future studies should build on this procedure and have a closer look at the privacy policies of MHA not only on a descriptive level, but actively monitoring the data traffic of MHA as well as trying to break or interfere with the data traffic to investigate if data protection criteria are met under attack. Besides the only descriptive analysis of privacy issues, some further limitations have to be considered. Firstly, this study might have missed relevant MHA due to the restriction to the European Google Play and Apple App store. However, Google Play and Apple App store cover about 99% of the total market, and the loss of relevant pain apps due to the selection of stores should be minimal (StatCounter, 2019). However, it cannot be guaranteed that all included pain apps are also available in the US stores and other major markets (e.g., India, China) or that the present results are transferable to the respective app versions in those stores. Secondly, the search results are limited to the selected search terms and the language restriction (English and German). The terms were selected by experts to cover a broad range of pain-related diseases, symptoms or content. Using this approach this study with N = 218 pain apps provides a by far more comprehensive analysis of pain app quality than in a previous review of pain apps (N = 18) (Salazar et al., 2018), nonetheless future studies could expand the present findings by including other languages and pain types. Thirdly, while this study makes a valuable contribution to the field by presenting the aims, content and function as well as quality of available pain apps for eight central types of pain (arthritis and rheumatism, back pain, chronic pain, fibromyalgia, headache and migraine, neck, other specific pain types and general pain) and thus, enables patients and health care providers to find a pain app fitting their needs, the study fell short in the comparison of provided content against established treatment guidelines. However, deriving the guidelines from the literature and comparing them with the content of pain apps was not the aim in the present study and doing so in a single study for pain apps in general might not be feasible due to the many pain apps and differences in the guidelines depending on the disorders. Hence, future studies with a closer focus on specific pain disorders (e.g. treatment of back pain) would be more suited to compare pain app content to treatment guidelines.

Conclusion

A multitude of pain apps is available. The overall quality of the included pain apps can be considered moderate at best. Particularly, the engagement features and the information quality need to be improved. The vast majority of available pain apps has no evidence and the effectiveness and safety of 99.5% of all examined pain apps have not been proven in an RCT. Above that, issues regarding the lack of privacy policies and information about data security are present. User ratings and app descriptions in the stores are insufficient to inform users. Independent information platforms offering reliable information for users and health care providers about content and quality (e.g., measured by the MARS) are highly needed. So far, the capabilities of pain apps to improve health care are not exploited. The following is the supplementary data related to this article.

Appendix 1

Aim, content and quality ratings of included apps.

Declaration of competing interest

EMM, YT, LS, HB developed and run the German Mobile Health App Database project. The MHAD is a self-funded project at Ulm University with no commercial interests. HB, LS and EMM received payments for talks and workshops in the context of e-mental-health. All other authors declare no conflict of interest.
  51 in total

Review 1.  A systematic review of smartphone applications for chronic pain available for download in the United States.

Authors:  Lorraine S Wallace; Lara K Dhingra
Journal:  J Opioid Manag       Date:  2014 Jan-Feb

2.  Predicting user adherence to behavioral eHealth interventions in the real world: examining which aspects of intervention design matter most.

Authors:  Amit Baumel; Elad Yom-Tov
Journal:  Transl Behav Med       Date:  2018-09-08       Impact factor: 3.046

3.  The efficacy of app-supported smartphone interventions for mental health problems: a meta-analysis of randomized controlled trials.

Authors:  Jake Linardon; Pim Cuijpers; Per Carlbring; Mariel Messer; Matthew Fuller-Tyszkiewicz
Journal:  World Psychiatry       Date:  2019-10       Impact factor: 49.548

Review 4.  Chronic diseases in the European Union: the prevalence and health cost implications of chronic pain.

Authors:  R M Leadley; N Armstrong; Y C Lee; A Allen; J Kleijnen
Journal:  J Pain Palliat Care Pharmacother       Date:  2012-12

5.  Global, regional, and national incidence, prevalence, and years lived with disability for 354 diseases and injuries for 195 countries and territories, 1990-2017: a systematic analysis for the Global Burden of Disease Study 2017.

Authors: 
Journal:  Lancet       Date:  2018-11-08       Impact factor: 79.321

Review 6.  Prevention of Mental Health Disorders Using Internet- and Mobile-Based Interventions: A Narrative Review and Recommendations for Future Research.

Authors:  David Daniel Ebert; Pim Cuijpers; Ricardo F Muñoz; Harald Baumeister
Journal:  Front Psychiatry       Date:  2017-08-10       Impact factor: 4.157

Review 7.  Smartphones for smarter delivery of mental health programs: a systematic review.

Authors:  Tara Donker; Katherine Petrie; Judy Proudfoot; Janine Clarke; Mary-Rose Birch; Helen Christensen
Journal:  J Med Internet Res       Date:  2013-11-15       Impact factor: 5.428

8.  The individual and societal burden of chronic pain in Europe: the case for strategic prioritisation and action to improve knowledge and availability of appropriate care.

Authors:  Harald Breivik; Elon Eisenberg; Tony O'Brien
Journal:  BMC Public Health       Date:  2013-12-24       Impact factor: 3.295

Review 9.  Benefits of Mobile Apps in Pain Management: Systematic Review.

Authors:  Simon E Thurnheer; Isaac Gravestock; Giuseppe Pichierri; Johann Steurer; Jakob M Burgstaller
Journal:  JMIR Mhealth Uhealth       Date:  2018-10-22       Impact factor: 4.773

Review 10.  Pain Management During the COVID-19 Pandemic.

Authors:  Salah N El-Tallawy; Rohit Nalamasu; Joseph V Pergolizzi; Christopher Gharibo
Journal:  Pain Ther       Date:  2020-08-25
View more
  7 in total

Review 1.  Do pain management apps use evidence-based psychological components? A systematic review of app content and quality.

Authors:  Megan MacPherson; A Myfanwy Bakker; Koby Anderson; Susan Holtzman
Journal:  Can J Pain       Date:  2022-06-03

2.  A systematic quality rating of available mobile health apps for borderline personality disorder.

Authors:  Lena Sophia Steubl; Josephin Reimann; Laura Simon; Yannik Terhorst; Michael Stach; Harald Baumeister; Lasse Bosse Sander; Eva-Maria Messner
Journal:  Borderline Personal Disord Emot Dysregul       Date:  2022-06-01

3.  Special issue on digital health interventions in chronic medical conditions: Editorial.

Authors:  Harald Baumeister; David D Ebert; Frank Snoek
Journal:  Internet Interv       Date:  2021-09-20

4.  Quality of Physical Activity Apps: Systematic Search in App Stores and Content Analysis.

Authors:  Sarah Paganini; Yannik Terhorst; Lasse Bosse Sander; Selma Catic; Sümeyye Balci; Ann-Marie Küchler; Dana Schultchen; Katrin Plaumann; Sarah Sturmbauer; Lena Violetta Krämer; Jiaxi Lin; Ramona Wurst; Rüdiger Pryss; Harald Baumeister; Eva-Maria Messner
Journal:  JMIR Mhealth Uhealth       Date:  2021-06-09       Impact factor: 4.773

5.  Standardized evaluation of the quality and persuasiveness of mobile health applications for diabetes management.

Authors:  A Geirhos; M Stephan; M Wehrle; C Mack; E-M Messner; A Schmitt; H Baumeister; Y Terhorst; L B Sander
Journal:  Sci Rep       Date:  2022-03-07       Impact factor: 4.996

6.  Review and Analysis of German Mobile Apps for Inflammatory Bowel Disease Management Using the Mobile Application Rating Scale: Systematic Search in App Stores and Content Analysis.

Authors:  Johannes Knitza; Till Orlemann; Maximilian Gerner; Nicolas Vuillerme; Timothée Aubourg; Eva-Maria Messner; Yannik Terhorst; Verena Hörmann; Ingo Ganzleben; Hannah Schenker; Georg Schett; Raja Atreya; Markus F Neurath
Journal:  JMIR Mhealth Uhealth       Date:  2022-05-03       Impact factor: 4.947

7.  Effects of a video intervention on physicians' acceptance of pain apps: a randomised controlled trial.

Authors:  Hauke Jeldrik Hein; Julia Anna Glombiewski; Winfried Rief; Jenny Riecke
Journal:  BMJ Open       Date:  2022-04-25       Impact factor: 3.006

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.