Literature DB >> 35277986

Understanding self-monitoring to inform a mobile intervention for binge eating and weight management: A proof-of-concept randomized trial.

Jianyi Liu1,2, Sean A Munson3, Angela Chang1,4, Claire Voss1, Andrea K Graham1,4.   

Abstract

OBJECTIVE: This study explored consumers' perspectives on self-monitoring, a common feature in behavioral interventions that helps inform consumers' progress and answer their questions, to learn what outcome metrics matter to consumers and whether self-selection of these metrics leads to greater engagement (i.e., compliance, satisfaction) in self-monitoring than monitoring only default options.
METHODS: In a proof-of-concept randomized trial, 48 adult participants were randomly assigned to "clinician-determined monitoring" or "clinician + self-determined monitoring" conditions. Before starting monitoring, all participants shared outcomes that would matter to them in a mobile intervention for binge eating and weight management. Then, for 3 weeks, participants in the "clinician-determined" condition monitored their weight and binge-eating episodes, and participants in the "clinician + self-determined" condition monitored these and another metric of their choosing. After, satisfaction and compliance were assessed.
RESULTS: Participants identified 116 metrics, grouped into 12 themes, that mattered to them. During monitoring, participants in the "clinician + self-determined" condition monitored 41 metrics. Surprisingly, participants in the "clinician-determined" condition also monitored metrics besides weight and binge eating. This resulted in a failure of our experimental manipulation, which represents a significant limitation of this research. No significant differences emerged in satisfaction or compliance between conditions. DISCUSSION: Although our proof-of-concept trial yielded null quantitative results, findings also suggested binge eating and weight management interventions may benefit from including an individually customizable monitoring option in addition to default metrics, warranting testing in future research. PUBLIC SIGNIFICANCE: Examining consumers' self-monitoring preferences for a mobile intervention for binge eating and weight management revealed a variety of metrics that matter to consumers, although binge eating and weight were still most valued. Findings from our proof-of-concept trial suggest design implications of encouraging an individually customizable monitoring option, in addition to default metrics, which needs to be tested in future research over a longer period and during actual mobile intervention delivery.
© 2022 The Authors. International Journal of Eating Disorders published by Wiley Periodicals LLC.

Entities:  

Keywords:  binge eating; mobile intervention; self-monitoring; user-centered design; weight management

Mesh:

Year:  2022        PMID: 35277986      PMCID: PMC9314870          DOI: 10.1002/eat.23700

Source DB:  PubMed          Journal:  Int J Eat Disord        ISSN: 0276-3478            Impact factor:   5.791


INTRODUCTION

Self‐monitoring, in which a person regularly observes and records their behavior(s) or the outcomes of behavior(s) within a behavior change strategy, is an evidence‐based technique for behavior change and a common feature of many evidence‐based interventions for mental and behavioral health problems (Michie et al., 2013). While this technique can be burdensome to people engaged in interventions (“consumers”), it can inform both clinicians and consumers about consumers' progresses (Fortney et al., 2017; Shimokawa et al., 2010; Torous & Roux, 2017; J. P. Tregarthen et al., 2015). In this paper, “consumer” refers to those who engage in interventions (e.g., using a mobile app). Other terms such as “users” have been associated with infantilizing and oversimplifying people, and “patients” may not be suitable for all experiences. Consumers and clinicians can have different interests regarding what metrics consumers should monitor in an intervention, as what matters to them can differ. Qualitative analyses of open‐ended responses from patients with a history of depression, informal caregivers, and health care professionals revealed 137 different outcomes domains that are of value to these individuals, 80 of which related to the benefits of treatment (Chevance et al., 2020). For example, the authors identified 19 domains related to mood and emotional symptoms (e.g., relief from mental pain, anxiety, sadness) and 18 domains associated with cognitive symptoms (e.g., motivation, cognitive distortion, and social interest). These distinctive outcomes extend beyond standard depression assessments, and the authors noted that clinical trials rarely assess for many of these domains (Chevance et al., 2020). Additionally, in another study that investigated patients' perspectives on benefits from a depression intervention, 20 participants reported 6 themes and 15 subthemes of personal benefits that matter to themselves (Folkersma et al., 2021). These studies suggest there might be a wide range of meaningful metrics from consumers' perspectives, making it important to examine what metrics matter to consumers. Research also has demonstrated that a large proportion of people who engage with behavior change interventions want customizable options for monitoring, different from the default options requested by clinicians (Karkar et al., 2017; Schroeder et al., 2018). Understanding these differences is important because not attending to the differences can negatively impact consumers' engagement with an intervention (Munson et al., 2020). In one study of digital wellness tracking programs in the workplace, not having customizable tracking options was a main reason why some participants decided not to use the app or stopped using the app (Chung et al., 2017). These participants indicated that if they could not track metrics that fit with their health goals, this program might not be the best choice. A separate study found that different interests in tracking goals and tools used in medical care settings could result in tension between patients and clinicians and contribute to patient disengagement (Chung et al., 2016). Conversely, having default monitoring options that are recommended by clinicians can be useful. While some patients have preferences for what they will track, others look to their health clinicians for guidance on what data are relevant and likely to lead to actionable insights (Chung et al., 2016; Lordon et al., 2020). In a study evaluating designs of a self‐tracking app for patients with migraines, participants wanted the app to recommend default options for monitoring, in addition to being able to add their own. Participants indicated that they might not choose the default option, but they appreciated the guidance provided when deciding what to monitor (Schroeder et al., 2019). Research in eating disorders has not yet focused on understanding what outcomes matter to consumers to monitor in an intervention. Such an examination is necessary because self‐monitoring is a core component of some evidence‐based interventions like cognitive‐behavioral therapy (CBT) for eating disorders (Fairburn, 2008), and completing self‐monitoring can have clinical benefits (Latner & Wilson, 2002). Yet, barriers can prevent monitoring or thwart sustaining monitoring (e.g., Cordeiro et al., 2015). Consequently, it is useful to understand how to make this technique most impactful and engaging to consumers. Additionally, previous research on a smartphone intervention with self‐monitoring for eating disorders (Kim et al., 2018) found that consumers who received a personalized version with content tailored to their baseline symptoms showed significantly better improvement in their Eating Disorder Examination Questionnaire scores compared to consumers who received the standardized CBT‐based version (J. Tregarthen et al., 2019). Therefore, having personalized content in a digital intervention with self‐monitoring could benefit consumers, and it is important to investigate what outcome matter to consumers to design personalized interventions. User‐centered design (also referred to as human‐centered design) is an approach that includes a myriad of methods for deeply understanding stakeholder needs and preferences and a given context to design compelling, effective, and engaging services and products. In user‐centered design, the end‐user (i.e., target consumer of a product) is placed at the center of the design process to ensure that the resulting product meets the needs and preferences of those who engage with and/or are impacted by it (Graham et al., 2019; Lyon & Koerner, 2016; Norman, 1988; Norman & Draper, 1986). We applied user‐centered design methods through a needs assessment and prototyping activity to inform the design of the self‐monitoring component of a mobile intervention for binge eating and weight management (Graham, Munson, et al., 2021; Graham, Neubert, et al., 2021; Weinheimer et al., 2020). With the needs assessment, our aim was to learn what metrics consumers want to monitor in an intervention besides standard (i.e., default) options (study aim 1). With the prototyping activity, our aim (study aim 2) was to conduct a proof‐of‐concept randomized trial that tested the concept of whether monitoring customized metrics determined by the consumer versus only clinical defaults leads to differences in engagement (i.e., compliance, satisfaction)—the intended outcome of user‐centered design. Our proof‐of‐concept trial hypothesis was that monitoring self‐determined metrics would lead to higher user satisfaction and compliance, compared to only monitoring clinician‐determined default metrics. To our knowledge, this is the first study to (1) identify metrics that are desirable to monitor among individuals seeking treatment for binge eating and weight management, and (2) test the impact of different monitoring strategies on design outcomes.

METHODS

Design

This study was a proof‐of‐concept randomized trial in which participants were randomly assigned to one of two study conditions. In the “clinician‐determined monitoring” condition, participants were asked to monitor their weight and the number of binge‐eating episodes they experienced for 3 weeks. In the “clinician + self‐determined monitoring” condition, participants monitored their weight, the number of binge‐eating episodes they experienced, and another metric of their choosing for 3 weeks. This study was approved by the Northwestern University Institutional Review Board, and the trial was registered at ClinicalTrials.gov (NCT04711577). All participants provided online informed consent.

Participants

Participants (N = 50) were recruited on dscout, an online qualitative research platform. Participants were eligible if they were nonpregnant, English‐speaking adults (age ≥ 18 years old) with self‐reported obesity (body mass index ≥30 kg/m2) and self‐reported recurrent binge eating (≥12 binge‐eating episodes over the past 3 months). Participants also needed to be interested in using an app to manage their weight and binge eating and to engage in self‐monitoring; although this study did not deploy an intervention besides self‐monitoring (see Section 2.3), these entry criteria aimed to ensure the sample reflected future consumers of a digital intervention and so findings could generalize to future implementation. Participants also were required to own or have access to a scale on a regular basis to measure their weight. Among those eligible, the final sample was selected based on sex, age, race, and ethnicity to ensure a diverse group. Two participants (one in each condition) dropped out before submitting any monitoring entries, and therefore were excluded from the final analyses since they did not have any monitoring entries to assess, resulting in a final sample of 48 participants who initiated and completed the proof‐of‐concept trial.

Procedure

Recruitment and enrollment

All study procedures occurred via dscout: an online qualitative research platform with >100,000 users who can voluntarily respond to advertisements for qualitative research studies. The key feature of dscout is facilitating digital diary studies that enable users to report on experiences in the moments and contexts in which they occur, in response to prompts created by researchers. Dscout allows researchers to ask various types of questions, including multiple‐choice questions, open‐ended text questions, and media‐based entries (i.e., videos, photographs). This study used the dscout research platform to mimic self‐monitoring; we did not create a monitoring app. No clinical intervention was delivered beyond asking participants to self‐monitor (e.g., no clinical content was provided, and participants did not receive feedback on their self‐monitoring entries). Questions asking participants to report on what they would like in a mobile intervention were intended to gauge their perspectives for a future digital intervention. Data S1 presents screenshots of a self‐monitoring entry. Participants were recruited via an advertisement on dscout for a paid opportunity to understand self‐monitoring behaviors in mobile interventions for weight management and binge eating in a 1‐month research study. Interested individuals provided online informed consent and then completed an online screening questionnaire to assess eligibility. Of the 821 individuals who initiated the screen, 50 eligible individuals were selected for and invited to participate. All invited individuals accepted the invitation to enroll. Figure 1 presents a consort diagram.
FIGURE 1

CONSORT participant flow diagram

CONSORT participant flow diagram

Sample size and randomization

To facilitate two study arms, two parallel projects were conducted in dscout, one project per study condition. Dscout's sample size limit is 25 participants per project. Since this study was a proof‐of‐concept study (i.e., aimed to answer whether this research question is worth asking, not to estimate effect sizes), we decided that the dscout project cap would be sufficient. Participants were randomized into one of two study conditions—“clinician‐determined monitoring” or “clinician + self‐determined monitoring”—prior to the start of the project. Participants were randomized using a computer‐generated sequence with a one‐to‐one ratio in blocks of four and six.

Study procedure

Study procedures are presented in Figure 2. During the first week, all participants, across both study conditions, were asked to share a list of outcomes that would matter to them regarding self‐monitoring, in addition to binge eating and weight, within a mobile intervention. They provided these data in a video recording and via an open‐ended text response.
FIGURE 2

Study procedures

Study procedures Participants were then invited to practice self‐monitoring for 3 weeks. Each week, participants were asked to submit two self‐monitoring entries (although they could submit more than one response for each entry) for a total of six requested entries. Entries were made available on the same days each week, such that participants had 3–4 days to complete each entry depending on when the entry was made available. Participants could submit their entry at any time during the requested window. Entries were made available sequentially so that participants would not submit all six entries at once. Participants received reminders before that entry's scheduled deadline. If the deadline passed and an entry still was not submitted, participants were asked to leave that entry blank and move on to the next one. However, expired entries could not be locked, so it was technically possible for participants to still submit entries past the deadline. Previous research has shown that the ability to enter missing information can be important for sustaining tracking motivation (Cordeiro et al., 2015). Participants' study condition differentiated what they were asked to self‐monitor for the 3 weeks. For the first entry, participants in the “clinician‐determined monitoring” condition monitored their weight and the number of binge episodes they had in the past week. Participants in the “clinician + self‐determined monitoring” condition monitored those same metrics plus a metric of their choosing. These participants were encouraged to select something that they could monitor for the three‐week duration. For each subsequent entry, monitoring instructions were the same except participants were asked to indicate the number of binge episodes that occurred since the previous entry. For each entry, participants were provided “fill‐in‐the‐blank” (open‐ended) response options for the metrics we requested of that study condition. Specifically, in the “clinician‐determined monitoring” condition, participants were given two blank spaces to report their weight and number of binge episodes; in the “clinician + self‐determined monitoring” condition, participants were given three blank spaces to report their weight, number of binge episodes, and the self‐determined metric. After reporting these metrics, participants from both conditions completed a video recording to capture their qualitative experiences with monitoring, and then responded to an open‐ended question asking if there was “anything else that would be useful for us to know.” After monitoring for 3 weeks, all participants were asked to reflect on their experience. To assess satisfaction with self‐monitoring, participants completed the seven‐item Satisfaction subscale of the Usefulness, Satisfaction, and Ease of Use (USE) Questionnaire (Lund, 2001), a validated measure that assesses usability (Gao et al., 2018). Items are rated on a scale of 1 (strongly disagree) to 7 (strongly agree), with higher scores indicating greater satisfaction. Also, participants indicated their satisfaction toward the frequency of monitoring (i.e., twice a week) by indicating if this frequency was “not enough,” “just right,” or “too much.” Additionally, participants were asked to indicate which of the following metric(s) they found most important to monitor: weight, number of binge‐eating episodes, nothing, something else (fill in), and the outcome they monitored over the past few weeks (assessed only among the “clinician + self‐determined monitoring” condition). Participants could select multiple outcomes. Finally, participants were asked to indicate their likelihood of continuing to monitor without monetary reward on a scale of 1 (not at all) to 10 (absolutely). Participants who completed all study procedures received a $75 reward.

Outcomes

The main outcomes were differences between study conditions in compliance and self‐reported satisfaction. Compliance was based on the number of monitoring entries completed of the expected six entries. Therefore, we defined compliance as the proportion of the sample who submitted at least six entries on time. Because individuals could submit multiple responses to each of the six entries, we also calculated the total number of entries submitted on time. Satisfaction was defined as the total score on the USE Questionnaire Satisfaction subscale. Secondary outcomes included participant satisfaction toward monitoring frequency, perception of the most important metric(s) to monitor, and likelihood of continuing to monitor without monetary rewards.

Analyses

Video recordings were automatically transcribed by dscout and were subsequently edited for accuracy and deidentification by an approved study team member. Transcripts from video recordings and open‐ended text responses were combined for qualitative analyses, which were conducted to assess what metrics participants want to monitor, evaluated among the full sample before the monitoring period and evaluated separately by condition for responses during the monitoring period. Using methods from thematic analysis (Braun & Clarke, 2006, 2021), we used an approach that involved becoming familiar with participants' entry data and video transcriptions, coding data, and conceptualizing and summarizing codes into meaningful domains. Regarding reflexivity, the analysis was completed by an independent rater (psychology Master's student with research experience in eating disorders), who routinely reviewed and discussed the codes with another rater (clinical psychologist researcher with expertise in eating disorders and weight management) to reach consensus. The research team's collective expertise in clinical science, digital intervention, and human‐computer interaction was used in understanding the results and developing the themes. The number of participants who mentioned and selected each outcome were counted. The median of the satisfaction ratings for each condition were calculated, and difference in median satisfaction between conditions was evaluated using a Mann Whitney test. A Fisher's exact test was performed to compare differences in the proportion of the sample who was compliant. Descriptives are presented for the secondary outcomes. These analyses were conducted using SPSS version 26. p Values of <.05 were considered statistically significant.

RESULTS

Sample characteristics

Detailed demographic information for the full sample and for each condition is presented in Table 1. Participants were adults with a mean age of 38.12 years (SD = 10.86). Over half of participants identified as White (n = 27; 58%), female (n = 25; 52%), and as college and post‐college graduates (n = 36; 75%).
TABLE 1

Demographics of the complete sample and by condition

CharacteristicComplete sample (N = 48)Clinician‐determined group (n = 24)Clinician + self‐determined group (n = 24)
Age (M (SD))38.12 (10.86)38.29 (10.36)37.54 (10.49)
GenderMan23 (48%)12 (50%)11 (46%)
Woman25 (52%)12 (50%)13 (54%)
Race/ethnicityAsian5 (10%)4 (17%)1 (4%)
Black or African American7 (15%)3 (13%)4 (17%)
Hispanic or Latinx8 (17%)2 (8%)6 (25%)
White27 (58%)14 (58%)13 (54%)
Prefer to self‐identify1 (2%)1 (4%)0 (0%)
EducationHigh school graduate3 (6%)1 (4%)2 (8%)
Some college9 (19%)3 (13%)6 (25%)
College graduate26 (54%)13 (54%)13 (54%)
Post graduate coursework10 (21%)7 (29%)3 (13%)
Household incomeLess than $25,0001 (2%)0 (0%)1 (4%)
$25,000–$49,9998 (17%)3 (13%)5 (21%)
$50,000–$74,99910 (21%)2 (8%)8 (33%)
$75,000–$99,99910 (21%)5 (21%)5 (21%)
$100,000–$124,9997 (15%)6 (25%)1 (4%)
$125,000–$149,9997 (15%)4 (17%)3 (13%)
Over 150,0005 (10%)4 (17%)1 (4%)
Demographics of the complete sample and by condition

What do people want to monitor?

Participants reported 116 distinct metrics that they were interested in monitoring (Data S2). As shown in Table 2, we grouped these metrics into 12 themes. The most common theme was metrics related to food and drink, comprised of 53 metrics indicated by 28 participants. The next most common theme was metrics related to physiology and body size (38 metrics indicated by 21 participants), followed by metrics related to physical activity (33 metrics indicated by 21 participants).
TABLE 2

Different themes that participants wanted to monitor

ThemeNo. of metrics per theme n (% a ) indicated this theme
Food and drink5328 (58%)
Physiology and body size3821 (44%)
Physical activity3321 (44%)
Mental health2613 (27%)
Sleep149 (19%)
Money/job53 (6%)
Nonspecific progress52 (4%)
Screen time42 (4%)
Time management33 (6%)
Comparisons to others22 (4%)
Spirituality22 (4%)
Miscellaneous11 (2%)

Percentages exceed 100% because participants could indicate as many metrics as they would like.

Different themes that participants wanted to monitor Percentages exceed 100% because participants could indicate as many metrics as they would like.

What do people choose to monitor?

Table 3 shows the comprehensive list of what participants chose to monitor, by study condition, during the 3‐week monitoring period. In the “clinician + self‐determined monitoring” condition, participants monitored 41 different metrics. Two‐fifths of participants in this condition (n = 10; 42%) chose to begin monitoring metrics related to food and drink. Over the 3 weeks, 50% of participants in this condition (n = 12) changed the metrics they were monitoring, despite instructions to monitor the same metric for 3 weeks.
TABLE 3

Metrics that participants monitored during the 3 weeks

MetricClinician + self‐determined groupClinician‐determined group
n (% a ) indicated this theme at entry n (% a ) indicated this theme over time n (% a ) indicated this theme
Theme: food and drinks
Food intake2 (8%)5 (21%)10 (42%)
Number of days met calorie goal1 (4%)1 (4%)
Number of days with solid eating1 (4%)
Day of binges episodes1 (4%)1 (4%)
Water intake1 (4%)2 (8%)1 (4%)
Time of meals1 (4%)1 (4%)1 (4%)
Number of meals per day1 (4%)1 (4%)
Food consumed during a binge1 (4%)2 (8%)
Number of takeout food orders1 (4%)1 (4%)1 (4%)
Whether chose soda or water for lunch1 (4%)
Number of nonwater beverages1 (4%)
Triggers of binge eating1 (4%)
Weekday vs. weekend binge episodes1 (4%)
Number of binge episodes prevented1 (4%)
Calorie intake1 (4%)6 (25%)
Eating log1 (4%)
Theme: physiology and body size
Heart rate2 (8%)2 (8%)
Blood pressure1 (4%)2 (8%)
Body fat percentage1 (4%)1 (4%)1 (4%)
Visceral fat percentage1 (4%)
Number of pounds lost1 (4%)1 (4%)
Number of pounds gained1 (4%)1 (4%)
Relation between changes in binge eating and weight1 (4%)
Menstrual cycle1 (4%)
Blood oxygen level1 (4%)
Waist measurement1 (4%)
Body measurement1 (4%)
Theme: physical activity
Amount of activity2 (8%)4 (17%)
Types of activity1 (4%)2 (8%)
Number of days active1 (4%)1 (4%)
Number of days of exercises1 (4%)1 (4%)
Steps1 (4%)1 (4%)1 (4%)
Exercises1 (4%)1 (4%)
Activity level1 (4%)
Theme: mental health
Number of meditation sessions1 (4%)1 (4%)
Number of days having “good willpower”1 (4%)
Emotional well‐being1 (4%)1 (4%)
Emotions1 (4%)1 (4%)
Feelings after a binge1 (4%)
Theme: sleep
Sleep time1 (4%)1 (4%)
Theme: money/job
Budget1 (4%)

Percentages exceed 100% because participants could indicate as many metrics as they would like.

Metrics that participants monitored during the 3 weeks Percentages exceed 100% because participants could indicate as many metrics as they would like. In the “clinician‐determined monitoring” condition, in which participants were instructed to only monitor weight and binge eating, nearly all participants (n = 23; 96%) indicated via their video recordings and/or open‐ended text responses that they monitored additional metrics on their own besides their weight and binge episodes over the monitoring period. Of these individuals, 10 participants (42%) chose to monitor their food intake, six (25%) chose to monitor calorie intake, and the rest monitored different outcomes.

Outcomes of monitoring

Table 4 presents the outcomes of self‐monitoring by study condition.
TABLE 4

Outcomes of monitoring

OutcomesClinician‐determined groupClinician + self‐determined group
Overall satisfaction (Median (IQR))5.00 (1.64)5.86 (1.89)
Compliance
Total entries on time135 (94%)137 (95%)
Total entries142 (99%)148 (103%)
Number submitted 6+ entries on time18 (75%)21 (88%)
Total people submitted 6+ entries22 (92%)23 (96%)
Satisfaction of monitoring frequency
Not enough13 (54%)11 (46%)
Just right11 (46%)10 (42%)
Too much0 (0%)3 (13%)
Likelihood of continuing to monitor (M (SD))8.33 (2.39)8.38 (2.39)
Perception of the most important metrics to monitor a
Weight19 (79%)21 (88%)
Number of binge episodes17 (71%)19 (79%)
Metric they monitored b NA13 (54%)
Other metric9 (38%)4 (17%)
Nothing0 (0%)0 (0%)

Percentages exceed 100% because participants could indicate as many metrics as they would like.

Assessed only among participants in the “clinician + self‐determined” group.

Outcomes of monitoring Percentages exceed 100% because participants could indicate as many metrics as they would like. Assessed only among participants in the “clinician + self‐determined” group.

Primary outcomes

Regarding overall satisfaction, there was no significant difference in satisfaction scores between groups (U = 232.5, p = .26). Regarding compliance, participants in the “clinician‐determined monitoring” condition submitted 135 entries on time, and 18 participants (75%) submitted at least six entries on time. Participants in the “clinician + self‐determined monitoring” condition submitted 137 entries on time, and 21 participants (88%) submitted at least six entries on time. There was no significant difference between study conditions in the proportion of compliant participants (p = .46).

Secondary outcomes

Regarding frequency of monitoring, the highest proportion of participants in both the “clinician‐determined monitoring” condition (54%) and the “clinician + self‐determined monitoring” condition (46%) indicated that monitoring twice a week was not enough; this was not significantly different (p = .34). Most participants in both conditions reported that weight was the most important metric to monitor (“clinician‐determined monitoring” condition: 79%; “clinician + self‐determined monitoring” condition: 88%). As for likelihood of continuing to monitor without monetary reward, there was no significant difference between groups (t(46) = 0.07, p = .94).

DISCUSSION

This study examined the metrics that consumers want to self‐monitor in a mobile intervention for binge eating and weight management and tested in a proof‐of‐concept randomized trial whether monitoring different metrics would lead to differences in satisfaction and compliance with self‐monitoring. Results inform the design of the self‐monitoring component within a future mobile intervention for binge eating and weight management. First, findings showed that consumers value a large variety of outcomes besides binge eating and weight. This finding aligns with a recent study on depression showing that substantially more outcomes matter to patients, informal caregivers, and healthcare professionals than the typical outcome domains assessed for depression (Chevance et al., 2020). Identifying the metrics consumers want to monitor is useful for informing the design of the self‐monitoring component of interventions and supports the larger goal of designing interventions that help consumers achieve their goals (Munson et al., 2020). Our results suggest it may be useful for binge eating and weight management interventions to engage consumers in identifying and monitoring additional metrics beyond binge eating and weight that matter to them. Although our study cannot confirm the optimal number of metrics that consumers should monitor (nor that there even is an optimal number of metrics), this work moves us closer toward the goal of designing interventions that meet the needs of its consumers to achieve engagement and improved clinical outcomes. To that end, we also tested whether encouraging consumers to monitor metrics that matter to them may help with engagement (i.e., satisfaction and compliance). Results of our proof‐of‐concept trial showed that monitoring a self‐determined metric, compared to only monitoring clinician‐determined default options, over a brief window of time did not lead to statistically significant differences in consumer satisfaction or compliance. Therefore, our concept was not proven, which might suggest against delivering self‐determined monitoring in this population. However, there were nuances of the current study that might have contributed to these null quantitative results and are important to consider. Our study period may have been too short to detect differences. Additionally, there could be a ceiling effect in participants' satisfaction scores. Another possible explanation is that nearly all participants in the “clinician‐determined” condition independently monitored extra metrics beyond the instructed two, which made their monitoring experience similar to participants in the “clinician + self‐determined” condition. This was an interesting, unexpected discovery. Our study was designed with two experimental conditions, and—like self‐monitoring worksheets or in digital interventions where there typically are pre‐set tracking options—each condition was administered the appropriate number of text boxes to enter their specified metrics. However, we learned through our qualitative questions that our manipulation became compromised, as participants' qualitative reflections revealed that they were tracking more metrics on their own than was requested. This is an interesting discovery because it reflects peoples' experiences as they occur in the “real world,” but would have been missed using the traditional quantitative self‐monitoring options. Therefore, our findings provide insight into what consumers are actually doing in their day‐to‐day lives, from which we can be better equipped to design monitoring tools that align with consumers' needs and preferences. Consequently, although our tested concept was not proven here, our results do seem to suggest that it may be beneficial to create opportunities for more flexibility for self‐monitoring that fit with consumers' own interests within digital interventions. Indeed, our finding that nearly everyone monitored additional metrics suggests that consumers may value being able to monitor both default and self‐determined metrics in a digital intervention; this is consistent with evidence for the importance of identifying individualized goals, and designing monitoring systems to support those goals, for individual health management (Munson et al., 2020). The variety of metrics chosen in our study suggests that an intervention may benefit from a design that offers consumers a customizable option for monitoring, as results did not point to prioritizing one particular metric aside from binge eating and weight, which most participants still perceived as most helpful to monitor. This finding, along with the lack of significant differences in satisfaction and compliance between conditions, supports sustaining standard self‐monitoring metrics in mobile interventions for binge eating and weight management. It also parallels findings in other self‐tracking domains that people often both want expert (i.e., clinical) guidance on relevant data to track as well as to have the option of tracking data in support of testing their own ideas about contributors and outcomes (Lordon et al., 2020; Schroeder et al., 2019). When digital interventions include a customizable monitoring option, our findings suggest consumers may also need support to make good decisions about whether, when, and how to switch metrics. Even in this short study, many participants switched the metrics they were monitoring during the monitoring period, despite instructions to monitor the same metrics over the 3 weeks. It could be problematic to change metrics before consumers have enough data to see and learn from trends. In other situations, it may be beneficial to switch, for example, if someone has learned all they can from a metric or have adjusted their behavior, then turning to new metrics may be helpful. Therefore, evaluating designs that support self‐monitoring consistency and/or decision rules for changing metrics, over a longer duration, represents a useful area for future research (e.g., Epstein et al., 2021).

Strengths and limitations

This study is the first to our knowledge to identify consumers' preferences for self‐monitoring in a digital intervention for binge eating and weight management. Furthermore, this study is the first proof‐of‐concept randomized trial that assessed the impact of different self‐monitoring designs on engagement. Finally, this study is strengthened by the use of mixed methods qualitative and quantitative data to enhance our understanding of participants' experiences. There were also study limitations. One, although our study condition manipulation was designed to create two separate groups, participants in the “clinician‐determined” condition monitored additional metrics on their own, compromising the manipulation. Two, participants were paid for completing the study, which may have affected their compliance. Three, the self‐monitoring activity was not conducted as part of a behavior change intervention; as a proof‐of‐concept prototyping activity, we only asked participants to monitor weight and binge eating, but evidence‐based treatments can involve monitoring more metrics. Also, unlike in treatment, participants monitored these metrics autonomously without support from a clinician or guide, which may have hindered their understanding of what to do with these metrics and which could have contributed to their switching behaviors. Four, our sample was derived from a registry of individuals who were signed up for an app, dscout, which advertises qualitative market research opportunities. Most participants were at least college educated, meaning that experiences of dscout users may not transfer to a broader national population despite more balance in other demographics (i.e., race, ethnicity, gender). Additionally, people who sign up for dscout may be more motivated to share in‐the‐moment insights of their real‐time experiences, which may limit generalizability to actual intervention delivery. However, most research studies require participants to be interested in and consent to participate, similar to how our participants needed to be interested in recording and sharing their real‐time data and behaviors related to self‐monitor to participate. Therefore, although the sample may be more motivated, self‐monitoring and sharing real‐time data are needed to some extent in real‐world interventions with self‐monitoring. Five, this study's findings are specific to people with recurrent binge eating and obesity; we cannot assume findings can be generalized to people with other eating disorders/weight statuses without additional study. Finally, the sample size was relatively small, so the quantitative analyses could be underpowered.

CONCLUSION

This study examined consumers' self‐monitoring preferences and the relationship among engagement (satisfaction, compliance) and selection of monitoring metrics as applied to a mobile intervention for binge eating and weight management. Our findings suggest that there are a large variety of metrics besides binge eating and weight that matter to consumers, although these two metrics are still of highest value. Our proof‐of‐concept trial results did not show statistically significant differences in compliance or satisfaction between the two monitoring conditions, indicating our concept under investigation was not proven. However, results suggested design implications of encouraging an individually customizable monitoring option and support for deciding when and how to switch metrics that are being monitored, which need to be tested in future research. Future studies also would benefit from evaluating self‐monitoring over a longer period during actual mobile intervention delivery. Data S1 Screenshots of a self‐monitoring entry that was administered in dscout, the online qualitative research platform used to conduct the study. Click here for additional data file. Data S2 Distinct metrics participants reported they were interested in monitoring. Click here for additional data file.
  21 in total

1.  Identifying outcomes for depression that matter to patients, informal caregivers, and health-care professionals: qualitative content analysis of a large international online survey.

Authors:  Astrid Chevance; Philippe Ravaud; Anneka Tomlinson; Catherine Le Berre; Birgit Teufer; Suzanne Touboul; Eiko I Fried; Gerald Gartlehner; Andrea Cipriani; Viet Thi Tran
Journal:  Lancet Psychiatry       Date:  2020-08       Impact factor: 27.083

2.  Enhancing treatment outcome of patients at risk of treatment failure: meta-analytic and mega-analytic review of a psychotherapy quality assurance system.

Authors:  Kenichi Shimokawa; Michael J Lambert; David W Smart
Journal:  J Consult Clin Psychol       Date:  2010-06

3.  TummyTrials: A Feasibility Study of Using Self-Experimentation to Detect Individualized Food Triggers.

Authors:  Ravi Karkar; Jessica Schroeder; Daniel A Epstein; Laura R Pina; Jeffrey Scofield; James Fogarty; Julie A Kientz; Sean A Munson; Roger Vilardaga; Jasmine Zia
Journal:  Proc SIGCHI Conf Hum Factor Comput Syst       Date:  2017-05-02

4.  Past, current, and future willingness to engage with treatment targets: Applying user-centered design to inform the design of a mobile behavioral intervention.

Authors:  Emilie A Weinheimer; Angela Chang; Sarah W Neubert; Jennifer E Wildes; Andrea K Graham
Journal:  Int J Eat Disord       Date:  2020-02-29       Impact factor: 4.861

5.  The behavior change technique taxonomy (v1) of 93 hierarchically clustered techniques: building an international consensus for the reporting of behavior change interventions.

Authors:  Susan Michie; Michelle Richardson; Marie Johnston; Charles Abraham; Jill Francis; Wendy Hardeman; Martin P Eccles; James Cane; Caroline E Wood
Journal:  Ann Behav Med       Date:  2013-08

6.  Integrating User-Centered Design and Behavioral Science to Design a Mobile Intervention for Obesity and Binge Eating: Mixed Methods Analysis.

Authors:  Andrea K Graham; Sean A Munson; Madhu Reddy; Sarah W Neubert; Emilie A Green; Angela Chang; Bonnie Spring; David C Mohr; Jennifer E Wildes
Journal:  JMIR Form Res       Date:  2021-05-10

7.  Patient-Driven Innovation for Mobile Mental Health Technology: Case Report of Symptom Tracking in Schizophrenia.

Authors:  John Torous; Spencer Roux
Journal:  JMIR Ment Health       Date:  2017-07-06

8.  Comparing a Tailored Self-Help Mobile App With a Standard Self-Monitoring App for the Treatment of Eating Disorder Symptoms: Randomized Controlled Trial.

Authors:  Jenna Tregarthen; Jane Paik Kim; Shiri Sadeh-Sharvit; Eric Neri; Hannah Welch; James Lock
Journal:  JMIR Ment Health       Date:  2019-11-21

9.  Understanding self-monitoring to inform a mobile intervention for binge eating and weight management: A proof-of-concept randomized trial.

Authors:  Jianyi Liu; Sean A Munson; Angela Chang; Claire Voss; Andrea K Graham
Journal:  Int J Eat Disord       Date:  2022-03-12       Impact factor: 5.791

View more
  1 in total

1.  Understanding self-monitoring to inform a mobile intervention for binge eating and weight management: A proof-of-concept randomized trial.

Authors:  Jianyi Liu; Sean A Munson; Angela Chang; Claire Voss; Andrea K Graham
Journal:  Int J Eat Disord       Date:  2022-03-12       Impact factor: 5.791

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.