Literature DB >> 34340693

Rapid deployment of a community engagement study and educational trial via social media: implementation of the UC-COVID study.

Lauren E Wisk1, Russell G Buhr2,3.   

Abstract

BACKGROUND: In response to the COVID-19 pandemic and associated adoption of scarce resource allocation (SRA) policies, we sought to rapidly deploy a novel survey to ascertain community values and preferences for SRA and to test the utility of a brief intervention to improve knowledge of and values alignment with a new SRA policy. Given social distancing and precipitous evolution of the pandemic, Internet-enabled recruitment was deemed the best method to engage a community-based sample. We quantify the efficiency and acceptability of this Internet-based recruitment for engaging a trial cohort and describe the approach used for implementing a health-related trial entirely online using off-the-shelf tools.
METHODS: We recruited 1971 adult participants (≥ 18 years) via engagement with community partners and organizations and outreach through direct and social media messaging. We quantified response rate and participant characteristics of our sample, examine sample representativeness, and evaluate potential non-response bias.
RESULTS: Recruitment was similarly derived from direct referral from partner organizations and broader social media based outreach, with extremely low study entry from organic (non-invited) search activity. Of social media platforms, Facebook was the highest yield recruitment source. Bot activity was present but minimal and identifiable through meta-data and engagement behavior. Recruited participants differed from broader populations in terms of sex, ethnicity, and education, but had similar prevalence of chronic conditions. Retention was satisfactory, with entrance into the first follow-up survey for 61% of those invited.
CONCLUSIONS: We demonstrate that rapid recruitment into a longitudinal intervention trial via social media is feasible, efficient, and acceptable. Recruitment in conjunction with community partners representing target populations, and with outreach across multiple platforms, is recommended to optimize sample size and diversity. Trial implementation, engagement tracking, and retention are feasible with off-the-shelf tools using preexisting platforms. TRIAL REGISTRATION: ClinicalTrials.gov NCT04373135 . Registered on May 4, 2020.
© 2021. The Author(s).

Entities:  

Keywords:  Coronavirus/COVID-19; Crisis Standards of Care; Educational intervention; Internet; Social media

Mesh:

Year:  2021        PMID: 34340693      PMCID: PMC8327053          DOI: 10.1186/s13063-021-05467-3

Source DB:  PubMed          Journal:  Trials        ISSN: 1745-6215            Impact factor:   2.279


Background

The novel severe acute respiratory syndrome coronavirus (SARS-CoV-2) publicly emerged in December 2019 and has since rapidly spread throughout the world, constituting a major pandemic. Early in the pandemic, concern for health system capacity and virus containment prompted many health officials, universities, and hospitals to undertake development of scarce resource allocation (SRA) policies [1]. These policies outline rules for distribution of limited resources, such as ventilators or hospital beds, and aim to do so in a way that is both consistent and ethical, with imperatives for maximizing benefits, equal treatment, and prioritization on instrumental value and need that can be operationalized in a number of ways [2]. Though medical ethicists have written extensively on how to construct such a framework, there are limited studies that aim to evaluate and intervene on knowledge of and agreement with these ethical principles during an active pandemic when the threat of their application is extant [3]. Conducting such a study requires reaching populations who may be the subject of allocation decisions and may be responsible for applying allocation decisions; both of which can be hard to reach during an active crisis. There is evidence that social media can be the best recruitment method for hard-to-reach populations and can be a similarly effective method as traditional recruitment in many cases [4-7]. Social media-delivered behavioral interventions have the potential to reduce the expense of more traditional interventions by eliminating the need for direct participant contact, also especially necessary during the pandemic, as well as increasing access to diverse participants that may not be accessible via clinic-based recruitment [8]. Although a burgeoning recruitment and delivery method, researchers are increasingly outlining rigorous methodological and ethical considerations for its use [9-11]. These methods have proven especially useful for the rapid recruitment into COVID-related studies during the current pandemic [12]. We sought to utilize Internet-based methods for engaging community participants and health care workers in completing a trial testing an educational intervention designed to influence knowledge of SRA policies and trust in institutional implementation of these policies during an ongoing pandemic. In this report, we outline detailed methods for recruitment strategies and implementation of an entirely online trial, including considerations for data fidelity and ethical oversight, with the use of widely available tools; we further summarize strengths and limitations of our approach and make recommendations for utilization of these methods in future work.

Methods

The UC-COVID (Understanding Community Considerations, Opinions, Values, Impacts, and Decisions for COVID-19) study is a community engagement study undertaken to characterize health and access to care during the COVID-19 pandemic. This study features an educational trial component to test the ability of a novel intervention to impact knowledge of and trust in institutional capacity to implement ethical allocation of scarce resources. We adopted a broad social media based recruitment strategy where we collaborated with community organizations, disease advocacy groups, and professional societies to invite study participation by direct messaging from organizations to their members; we also employed targeted social media posts and referrals from participants. Though our recruitment strategy primarily focused on groups in California (71% of our sample), eligibility was not restricted by location and all adults (age ≥ 18) were eligible. Recruitment for the survey opened on May 8, 2020, and closed on September 30, 2020; though 99% of respondents entered the survey between May and August. This study was approved by the Institutional Review Board and informed consent was obtained from all participants. This study was registered with ClinicalTrials.gov (NCT04373135).

Study design and procedures

To promote recruitment, we established partnerships with several disease advocacy groups (e.g., COPD Foundation, Taking Control of Your Diabetes, Pulmonary Hypertension Association, Vietnamese Cancer Foundation, AltaMed) and professional societies (e.g., California Thoracic Society, American Thoracic Society, Society for General Internal Medicine). Investigators contacted these groups, presented aligned goals, and proposed utilizing their networks for targeted study recruitment. Recruitment messages were primarily posted on social media accounts, message boards, and via direct newsletters to the email distribution lists of networks and partner groups. Broader study promotion was also achieved through “sharing” of study information via personal/professional social networks of study investigators, colleagues/institutions, and participants (website analytics revealed that visitors shared study information 279 times via embedded share applet—similar to snowball sampling). Recruitment messages included IRB-approved language to promote the study (brief descriptions, inclusion criteria) and provide a link directing participants to a hosted study website. The study website included general study information (including IRB and trial registration information), research team contact information, and a share applet that allowed users to send an email or create a social media post linking back to the landing page. We tracked study website traffic (user and page views) during the recruitment period, including device on which the page was accessed (mobile, tablet, desktop), referral source, language, and location. The study website directed participants to click an outbound link that transferred them to a Research Electronic Data Capture [13, 14] (REDCap, Nashville, TN; May 8, 2020, to June 19, 2020) or Qualtrics (Provo, UT) survey (June 19, 2020, to study close), hosted on secure servers at UCLA. Though we initially implemented our study in English only, to expand our ability to include diverse participants we utilized professional translation services (International Contact, Inc.; Berkeley, CA) to translate our study (website, consent, and survey) into the five most commonly spoken foreign languages in California (Spanish, Mandarin Chinese, Korean, Tagalog, and Vietnamese). As REDCap did not have native support for translation of the command buttons for the survey (e.g., “Next,” “Submit”), we migrated the survey to Qualtrics to facilitate full translation of the survey and interface. After entering the survey, respondents first viewed an online consent form with language included in a typical written consent. Of the 2844 survey initiations (Fig. 1), 362 (12.7%) entries from “bots” and 82 (2.9%) duplicate participant entries were excluded; 2384 respondents affirmed consent via the online form, 15 respondents declined consent, and 1 respondent exited the survey without affirming or declining consent. Of those who consented, 413 (17.3%) respondents did not continue the survey beyond that point, resulting in 1971 (82.7%) consented, active participants. One thousand five hundred forty (78.1%) participants completed the baseline assessment through at least part of the section on SRA policies and thus were eligible for pre/post comparison of key trial outcomes.
Fig. 1

UC-COVID CONSORT diagram. CONSORT flow diagram showing participant flow from recruitment and into first follow-up and trial randomization

UC-COVID CONSORT diagram. CONSORT flow diagram showing participant flow from recruitment and into first follow-up and trial randomization Baseline participants who did not complete on their first attempt (and who provided an email address) received a reminder email 4 weeks after their last activity, then weekly for 3 weeks (for up to four total invitations) thereafter to remind/encourage survey completion.

Assessing data validity

The Qualtrics platform has a built-in option (“Prevent Ballot Box Stuffing”) that is designed to prevent duplicative entries by placing a cookie on the browser of participants during their first entry into the survey. If the same respondent comes back on the same browser and device, without having cleared their cookies, they are flagged as a duplicate and not permitted to take the survey again. However, clearing browser cookies, switching to a different web browser, using a different device, or using a browser in “incognito” mode would all allow a participant to enter the survey again. As such, we additionally relied on embedded data to identify potential fraudulent entries for records attached to IP addresses that were duplicated in the data greater than four times; three of four instances were suspected to be the result of bots (fraudulent activity) [15] and discarded from the data. In the first instance, one IP address (geo-tagged to a location in China) contributed 172 attempted survey entries, none of which progressed in the survey beyond the consent page, that were all submitted within a 24-min window. In the second and third instances, two IP addresses contributed 121 and 69 attempted survey entries, none of which progressed past the demographics section of the survey, all included similarly formatted email addresses (random word + four random letters @ domain), including some emails that were duplicated across these two IP addresses. The final instance included 19 records with unique and valid emails; these records were determined to be valid and submitted by unique individuals using a shared server. Invalid records submitted by bots were largely consistent with each other (e.g., 100% identified as health care workers, 100% reported their age reported an age between 30 and 33) and compared to valid records were more likely to report younger age, male sex, divorced/widowed/separated, having a bachelor’s degree, currently working, and having a military background (data not shown).

Follow-up surveys

Following consent, respondents were asked to provide an email address for eligibility to receive a gift card and to receive follow-up surveys; respondents could still participate in the baseline survey if they did not provide an email (N = 222 did not provide an email). Follow-up invitations were sent in batches by month of baseline survey entry beginning in the second week of August so follow-up surveys were predominantly completed 2–3 months after the baseline survey; the first follow-up survey was closed in December. Participants received an email with a unique link to participate in the first follow-up survey; participants who did not complete the survey after the original invitation subsequently received a reminder email weekly for three additional attempts (for up to four total invitations) thereafter or until they completed. Of 1749 provided e-mail addresses, only 1550 invitations were initially sent as 18 e-mails were returned as undeliverable and by an unidentified error, 181 e-mails were marked as “not sent” at the close of the pre-programmed Qualtrics distribution. Of follow-up invitations sent, 19 respondents opted-out/declined to participate in the follow-up survey and 592 did not respond to the follow-up requests, resulting in 939 (60.6%) entries into the first follow-up survey. Participants were sent a second follow-up survey via automated email invitation in January 2021, with up to 4 automated reminders to complete. Participants who provided email addresses were entered in a raffle to win one of twenty-five (25) $100 gift cards for an online retailer; participants who complete two surveys receive one entry and those who complete all three surveys receive two entries.

Intervention

During the first follow-up survey, respondents from California were automatically randomized to receive either a brief educational video explaining SRA policies or no intervention using a randomization module programmed into the survey that executed a stratified randomization scheme based on health care worker status, gender, age, race, ethnicity, and education. As the intervention was based on the policy developed by the University of California system (one of the largest providers in the state, with 10 campuses, five medical centers, and three affiliated national laboratories) and our hypotheses focused on the intervention’s impact for California adults (i.e., those who could potentially be the subject of allocation decisions and/or may be responsible for applying allocation decisions per the University of California policy), only respondents who reported residing in California were randomized. As those outside California may be subject to different SRA policies, we elected not to expose them to information about a policy that was potentially not relevant to them and so did not randomize them (“negative controls”). Participants randomized to the intervention were automatically shown the intervention video, which was housed on a private Vimeo (New York, NY) channel and embedded in the survey. The 6:30-min long video was animated by a professional video production company (WorldWise Productions, Los Angeles, CA) and covered key topic of public health ethics, policy development, and a summary of how the University of California’s SRA policy would be implemented during a crisis. A copy of the video is available upon request. In addition to viewing the intervention video, participants randomized to treatment were also shown five additional survey questions to assess their impressions of the intervention; all other content of the follow-up survey was identical to controls.

Safety

At the completion of each survey, participants were directed to a “Thank you” page that additionally included a message directing them to contact the study team with any questions or concerns, including information on how to do so. Participants were also instructed to reach out to their personal health care or mental health provider if they experienced discomfort or distress and were provided with the website and phone number for the National Suicide Prevention Lifeline in the event they were in crisis.

External comparison

To determine the extent to which our non-probability sample is representative of the larger population from which our sample was drawn (primarily California adults, but also US adults), we compared our sample to respondents from 2019 Behavioral Risk Factor Surveillance System (BRFSS) [16]. Our survey used a number of BRFSS questions (see below) to facilitate comparison.

Survey measures

Survey data included information on demographics (sections 1 and 6), health and health behaviors (section 2), access to care (section 3), experience with COVID-19/coronavirus (section 4), and SRA policies (section 5). The baselines questionnaire was the longest (approximately 35 min to complete) while subsequent surveys were designed to be shorter (approximately 15 min to complete). Respondents first self-reported their status as a health care worker (Are you a health care professional? Examples include: physician/doctor, nurse, pharmacist, respiratory therapist, rehab specialist, psychologist, clinical social worker, or hospital chaplain. If you are a health professional student (pre-degree or certificate) please select “no” for the purposes of this survey.) Those who identified themselves as a health care worker received different survey items than non-health care workers. All participants were also asked to report their employment status, educational attainment, gender identity, year of birth, race, ethnicity, health insurance, place of residence, marital status, and if they had children. Health care workers were also asked to identify their specialty and tenure. A shorter version of this section was administered in the follow-up surveys to assess changes to employment and insurance. The second section of the survey ascertained information on health and health behaviors; the majority of these questions were drawn from the BRFSS [16]. Information included diagnosed chronic conditions, self-reported general health status (5-point Likert scale from “poor” to “excellent”), number of days in the past 30 days where mental health was “not good” (“Now thinking about your mental health, which includes stress, depression, and problems with emotions…”) and where physical health was “not good” (“Now thinking about your physical health, which includes physical illness and injury…”), screeners for depression (PHQ-2) [17] and anxiety (GAD-2) [18], a single item on sleep from the PHQ-9 [19], and respondents were asked to compare their mental health now to the same time last year. Respondents were asked about alcohol use in the past 30 days (number of days of use, number of drinks per occasion), cigarette [20] and e-cigarette [21] use in the past 30 days, and exercise in the past 30 days; they were also asked to report recent changes in these behaviors and if COVID-19 was a cause. A shorter version of this section was administered in the follow-up surveys to assess physical and mental health. Self-identified healthcare workers were also asked about burnout in subsequent surveys [22]. The third section focused on access to care; all participants were asked about receipt of an influenza vaccination for the 2019 season, if they had a personal doctor, and the time they last saw their personal doctor. Those who previously reported any chronic medical condition were asked a set of novel questions about the impact of COVID-19 and related social distancing on their disease management and symptoms. Non-health care workers were asked an additional series of novel questions about changes in their ability to access health care during COVID-19, delayed or forgone care during COVID-19, and changes in the use of prescription and over-the-counter medications during COVID-19. This section was not administered during the first follow-up survey and an abbreviated version was administered at the second follow-up survey. The fourth section focused on COVID impacts and asked respondents about their knowledge of government regulation of activities during the pandemic, personal protective behaviors, COVID-19 information seeking [23], COVID-19 related stress [23], and perceived personal risk from COVID-19. At follow-up, this section additionally contained questions about COVID testing and vaccination. The fifth and longest section focused on awareness/knowledge of SRA policies, alignment with values governing SRA policies, preferences for SRA implementation and communication, and trust/anxiety for SRA. All SRA questions were novel but demonstrated acceptable psychometric properties (e.g., Cronbach’s alphas ranging from 0.5666 to 0.8954; Appendix). A shorter version of this section was administered in the follow-up surveys. The final section asked optional personal questions regarding COVID-19 experiences [24] (exposure to COVID-19), disability status [25], advanced care planning, general sources of news/information [26], the experience of discrimination in health care, and other personal characteristics [27] (e.g., religion, sexual orientation, political identity). This section was not administered in the follow-up surveys.

Statistical analysis

All analyses were performed using SAS 9.4 (Cary, NC). Summary statistics were used to describe website traffic information based on analytics derived from the hosting platform (Square Space) and from Google Analytics; correlation between website traffic and respondent counts by geography were evaluated. To ascertain the representativeness of the sample, we calculated and compared descriptive statistics for participants (N = 1971), BRFSS respondents from California (N = 11,613), and BRFSS respondents from all 50 states and Washington DC (N = 409,810); BRFSS prevalence data are weighted to represent the US adult populations. Finally, differential non-responses were evaluated by comparing characteristics of those with complete (into SRA section) vs incomplete data at baseline. Differences were assessed using appropriate two-sided bivariate tests with a 0.05 alpha criterion.

Results

As about 1% of website traffic originated from organic searches, compared to the 55–60% from direct website entry or referral and 40–45% from social media recruitment, likelihood of un-invited survey participation (i.e., respondents who did not enter the survey as a direct result of targeted recruitment efforts) is relatively low (Table 1). Observed website traffic reflected the shift in recruitment strategy from outreach through partner organizations from May through July to broader social media based recruitment in August. Among social media platforms, Facebook was the highest yield recruitment source (85% of website traffic originating from social media). Website traffic was predominantly from California but still geographically diffuse, suggesting substantial national reach of recruitment efforts, and was highly correlated to place of residence reported by respondents, indicating similar response rates across geography (Fig. 2).
Table 1

Recruitment tracking via website analytics

TotalaMayJuneJulyAugustSeptember
Google Analytics
Usersb806927675509633427541
Sessionsc10,439354976112544312563
Sessions per user1.291.281.381.301.261.04
Pageviewsd15,7694081165919667157906
Pages per session1.511.152.181.571.661.61
User acquisition channel
 Social3692 (45%)1230 (44%)125 (23%)228 (24%)2096 (62%)13 (2%)
 Direct2946 (36%)1290 (47%)343 (64%)512 (54%)737 (22%)64 (12%)
 Referral1320 (16%)107 (4%)38 (7%)188 (20%)542 (16%)445 (85%)
 Email135 (2%)129 (5%)6 (1%)
 Organic Search60 (1%)9 (< 1%)22 (4%)14 (1%)13 (< 1%)2 (< 1%)
Users by language
 English7873 (96%)2746 (99%)535 (97%)872 (91%)3278 (96%)534 (99%)
 Other295 (4%)33 (1%)15 (3%)91 (9%)149 (4%)7 (1%)
Users by location
 United States7429 (92%)2564 (93%)508 (92%)887 (92%)3385 (99%)257 (48%)
 California5134 (64%)1135 (41%)290 (53%)655 (68%)2981 (87%)148 (27%)
 Other646 (8%)204 (7%)42 (8%)76 (8%)42 (1%)284 (52%)
Users by device
 Mobile4599 (56%)1336 (48%)233 (42%)464 (48%)2384 (70%)205 (38%)
 Desktop3061 (38%)1367 (49%)309 (56%)454 (47%)682 (20%)317 (59%)
 Tablet495 (6%)65 (2%)8 (1%)45 (5%)361 (11%)19 (4%)
Sessions by social source
 Facebook3528 (84%)975 (74%)73 (50%)161 (63%)2309 (94%)10 (59%)
 Twitter516 (12%)294 (22%)64 (44%)44 (17%)113 (5%)1 (6%)
 LinkedIn106 (3%)46 (3%)7 (5%)41 (16%)12 (< 1%)
 Instagram30 (1%)3 (2%)10 (4%)11 (< 1%)6 (35%)
Times study shared2791076713866
Square space analytics
Visitse10,330367988812605471292
Pageviewsf15,3504237151520619262336
Visit acquisition channel
 Social4024 (39%)1318 (36%)146 (16%)234 (19%)2313 (55%)13 (4%)
 Direct4157 (40%)1874 (51%)608 (68%)703 (56%)876 (21%)96 (33%)
 Referral2081 (20%)471 (13%)106 (12%)311 (25%)1013 (24%)180 (62%)
 Email4 (< 1%)4 (< 1%)
 Organic Search64 (1%)12 (< 1%)28 (3%)12 (1%)9 (< 1%)3 (1%)
Pageviews by language
 English15,398 (99%)4313(100%)1543 (95%)2027 (98%)7181 (99%)334 (98%)
 Other175 (1%)79 (5%)45 (2%)43 (1%)8 (2%)
Visits by location
 United States9563 (93%)3263 (89%)734 (83%)1135 (90%)4152 (99%)279 (96%)
 California6498 (63%)1531 (42%)391 (44%)837 (67%)3575 (85%)164 (56%)
 Other763 (7%)416 (11%)153 (17%)123 (10%)58 (1%)13 (4%)
Visits by device
 Mobile5570 (54%)1667 (45%)317 (36%)581 (46%)3493 (64%)93 (32%)
 Desktop4207 (41%)1930 (52%)564 (64%)634 (50%)1525 (28%)188 (64%)
 Tablet553 (5%)82 (2%)7 (1%)45 (4%)453 (8%)11 (4%)
Visits by browser
 Chrome2232 (22%)951 (26%)284 (32%)340 (27%)889 (16%)108 (37%)
 Other8094 (78%)2728 (74%)603 (68%)918 (73%)4579 (84%)184 (63%)
Visits by social source
 Facebook3430 (85%)974 (74%)72 (49%)156 (67%)2218 (96%)10 (77%)
 Twitter459 (11%)295 (22%)66 (45%)30 (13%)68 (3%)
 LinkedIn110 (3%)49 (4%)5 (3%)39 (17%)17 (1%)
 Instagram25 (1%)3 (2%)9 (4%)10 (< 1%)3 (23%)

aTotal may not be the sum of individual months as single users could access the website repeatedly across time. Similarly, user data may not sum to column totals if the same user accessed the site via different entry points

bGoogle Analytics defines users as the number of unique identifiers (assigned via a unique, randomly generate string that gets stored in a browser cookie) who have initiated at least one session during a given time period. Using cookies allows analytics to identify unique users across browsing sessions, but it cannot identify unique users across different browsers or devices

cGoogle Analytics defines sessions as the period of time a user is actively engaged with the site

dGoogle Analytics defines pageviews as the total number of pages viewed. Repeated views of a single page are counted

eSquareSpace defines a visit as a single browsing session, and can encompass multiple page views. Visits are tracked with a browser cookie that expires after 30 min; as such, any hit from a single user within that 30-min browsing session count as one visit and that one person can register multiple visits a day if they close their browser and return to the site at least 30 min later. For visitors using the “Do not track” Chrome browser option, every page they view is tracked as a new visitor so will inflate visit data

fSquareSpace defines a page view as how many actual page requests the site saw in a given time period. All full page loads count toward total page views, including views to separate pages within the site (e.g., study information page, survey exit/thank you page)

Fig. 2

Geographic coverage of UC-COVID recruitment. Google Analytics coverage map showing website traffic from May 1, 2020, to September 30, 2020, from metro areas of the USA (traffic from outside the USA not shown) with inset table showing Pearson’s correlation between traffic from SquareSpace, traffic from Google Analytics and respondents’ survey-reported residence at the country and state (USA only) levels

Recruitment tracking via website analytics aTotal may not be the sum of individual months as single users could access the website repeatedly across time. Similarly, user data may not sum to column totals if the same user accessed the site via different entry points bGoogle Analytics defines users as the number of unique identifiers (assigned via a unique, randomly generate string that gets stored in a browser cookie) who have initiated at least one session during a given time period. Using cookies allows analytics to identify unique users across browsing sessions, but it cannot identify unique users across different browsers or devices cGoogle Analytics defines sessions as the period of time a user is actively engaged with the site dGoogle Analytics defines pageviews as the total number of pages viewed. Repeated views of a single page are counted eSquareSpace defines a visit as a single browsing session, and can encompass multiple page views. Visits are tracked with a browser cookie that expires after 30 min; as such, any hit from a single user within that 30-min browsing session count as one visit and that one person can register multiple visits a day if they close their browser and return to the site at least 30 min later. For visitors using the “Do not track” Chrome browser option, every page they view is tracked as a new visitor so will inflate visit data fSquareSpace defines a page view as how many actual page requests the site saw in a given time period. All full page loads count toward total page views, including views to separate pages within the site (e.g., study information page, survey exit/thank you page) Geographic coverage of UC-COVID recruitment. Google Analytics coverage map showing website traffic from May 1, 2020, to September 30, 2020, from metro areas of the USA (traffic from outside the USA not shown) with inset table showing Pearson’s correlation between traffic from SquareSpace, traffic from Google Analytics and respondents’ survey-reported residence at the country and state (USA only) levels Characteristics of recruited participants differed from the population of California and from the US population as a whole (Table 2). Compared to the adult population of California, study participants (71.0% from California) were more likely to identify as female, less likely to identify as Hispanic, and more likely to report having a bachelor’s degree. Despite being slightly more likely to report health insurance coverage and having a personal health care provider, study participants did not differ from adults in California with respect to the prevalence of chronic conditions or the likelihood of having a routine check-up in the prior year. Similar distinctions were observed when comparing study participants to the adult population of the USA.
Table 2

Comparison of UC-COVID and BRFSS respondents: assessment of sample representativeness

UC-COVIDAll CA (BRFSS)All US (BRFSS)
Total N (unweighted)197111,613409,810
Age (years)
 18–3420.7%20.8%20.3%
 35–4931.1%35.4%32.9%
 50–6429.0%24.6%25.4%
 65+19.2%19.2%21.4%
Sex
 Male24.3%49.3%48.7%
 Female75.7%50.7%51.3%
Race/ethnicity
 Hispanic11.8%36.2%16.6%
 Black, non-Hispanic4.3%5.8%12.1%
 Asian/Pacific Islander, non-Hispanic12.1%16.2%5.7%
 AIAN/Other race, non-Hispanic3.3%2.6%2.7%
 White, non-Hispanic68.5%39.3%63.0%
Has bachelor’s degree or higher83.8%29.8%28.5%
Marital status
 Married or living with partner69.8%56.1%55.5%
 Divorced, widowed, or separated12.9%17.6%19.9%
 Never married17.3%26.4%24.6%
Has children (< 18) in household30.3%38.8%34.8%
Is a veteran or active duty military4.8%7.5%10.2%
Has health insurance coverage93.5%87.5%87.0%
Employment status
 Currently working63.4%59.0%57.7%
 Furloughed/on leave6.7%3.3%2.8%
 Retired or student17.8%23.0%24.8%
 Unable to work/Out of work12.1%14.7%14.7%
Has chronic conditiona49.5%48.4%54.3%
Has personal health care provider83.9%74.8%76.6%
Had routine check-up in past year73.8%71.8%76.6%
Completed survey in English98.9%86.8%93.8%

Study sample is compared to 2019 BRFSS respondents from California and from the whole Unites States (US). For BRFSS samples, survey weighted percentages are shown

AIAN, American Indian/Alaska Native

aBased on chronic conditions assessed in the BRFSS (ever diagnosis of): heart attack, heart disease, stroke, asthma, cancer, chronic obstructive pulmonary disease (COPD), arthritis, depression, kidney disease, or diabetes

Comparison of UC-COVID and BRFSS respondents: assessment of sample representativeness Study sample is compared to 2019 BRFSS respondents from California and from the whole Unites States (US). For BRFSS samples, survey weighted percentages are shown AIAN, American Indian/Alaska Native aBased on chronic conditions assessed in the BRFSS (ever diagnosis of): heart attack, heart disease, stroke, asthma, cancer, chronic obstructive pulmonary disease (COPD), arthritis, depression, kidney disease, or diabetes Respondents who did versus did not complete the baseline SRA questions also varied by sociodemographics (Table 3). Participants with complete SRA data were more likely to be younger, white non-Hispanic, have a bachelor’s degree, currently employed, married or partnered, insured, and living with a chronic condition.
Table 3

Comparison of UC-COVID respondents by baseline completeness

Baseline complete?a
YesNop value
Total N1540431
Health care worker0.018
 No69.0%74.9%
 Yes31.0%25.1%
Age (years)0.003
 18–3421.2%18.8%
 35–4932.6%25.8%
 50–6428.3%31.3%
 65+17.9%24.1%
Sex0.136
 Male25.1%21.6%
 Female74.9%78.4%
Race/ethnicity< 0.001
 Hispanic9.9%18.3%
 Black, non-Hispanic4.3%4.4%
 Asian/Pacific Islander, non-Hispanic12.7%9.7%
 AIAN/Other race, non-Hispanic2.1%7.7%
 White, non-Hispanic71.0%59.9%
Has bachelor’s degree or higher86.1%75.6%< 0.001
Marital status0.001
 Married or living with partner70.5%67.3%
 Divorced, widowed, or separated11.4%18.1%
 Never married18.1%14.6%
Has children (< 18) in household30.9%28.3%0.299
Is a veteran or active duty military5.0%3.9%0.363
Has health insurance coverage96.4%83.3%< 0.001
Employment status< 0.001
 Currently working65.7%55.2%
 Furloughed/on leave7.1%5.3%
 Retired or student16.1%23.7%
 Unable to work/Out of work11.1%15.8%
Has chronic conditionb51.4%42.9%0.002
Has personal health care provider83.1%89.0%0.046
Had routine check-up in past year74.5%69.5%0.168
Completed survey in English98.9%98.8%0.922
Place of residence< 0.001
 California68.5%79.8%
 Other31.5%20.2%
Provided e-mail for re-contact87.7%92.3%0.007

AIAN, American Indian/Alaska Native

aCompleteness refers to at least partial completeness for the scarce resource allocation policies questions

bEver diagnosis of heart attack, heart disease, stroke, asthma, cancer, chronic obstructive pulmonary disease (COPD), arthritis, depression, kidney disease, or diabetes

Comparison of UC-COVID respondents by baseline completeness AIAN, American Indian/Alaska Native aCompleteness refers to at least partial completeness for the scarce resource allocation policies questions bEver diagnosis of heart attack, heart disease, stroke, asthma, cancer, chronic obstructive pulmonary disease (COPD), arthritis, depression, kidney disease, or diabetes

Discussion

We demonstrate that rapid recruitment of a large cohort into a longitudinal educational trial via outreach through social media is feasible, efficient, and acceptable to participants. While recruitment through partner organizations was advantageous, broader social media-based outreach was useful for expanding the number of recruited participants; diversified recruitment efforts may also be useful for maximizing sample diversity. Tracking of initial engagement with recruitment posts is facilitated by off-the-shelf software and preexisting platforms. An emerging evidence base describes opportunities, such as rapidity, and shortcomings in Internet-enabled research, such as potential selection bias. This study is similarly subject to many of the potential biases reported elsewhere [28, 29]. Most notable are the potential issues with generalizability when the study sampling frame is an Internet-engaged audience [30]; though the digital divide has transformed over time [31], there are clear selection biases stemming both from Internet-based recruitment and voluntary participation with limited remuneration. Indeed, we found that our sample was more likely to be female and less likely to be ethnically diverse, as with many Internet-recruited samples, but a considerable source of deviation from characteristics of the general adult population potentially stemmed from targeted recruitment of health care workers [32, 33], and thus over-representation of college-educated and employed individuals. Examination of attrition revealed that loss from website engagement to survey participation was largely non-differential with respect to geography but that survey completion was linked to differences in many of the same characteristics that were associated with initial engagement. Despite these limitations, our sample was relatively representative with respect to age, certain racial subgroups, and, importantly, presence of chronic medical conditions. Overall, such comparisons within and between the sample and target population, while not definitive, are helpful to assess the extent to which otherwise diffuse (non-probability-based) recruitment was effective for constructing a minimally-biased sample (compared to expectations for Internet-recruited samples) and we strongly recommend designing similar comparisons in future work. Despite these common sampling limitations with participatory research, we were still able to recruit and retain a number of respondents from smaller subgroups given our larger sample size and concerted efforts to partner with community organizations representing diverse subpopulations. Any future work in this area should strive to adopt a truly community-based participatory research (CBPR) approach in order to maximize reach among communities who are typically under-represented in research. We identified that offering to share data summaries with partners to “close the loop” and ensure shared use of study information, consistent with CBPR principles, was one effective method for engaging recruitment partners. Further efforts to improve recruitment and retention across socioeconomic strata could be enhanced by employing guaranteed remuneration. Of note, both overall recruitment and recruitment of various subgroups may have been enhanced by the timeliness of this survey’s topic; for studies of other health issues, additional efforts highlighting the importance of the topic for target populations may be needed to boost recruitment. In prior work using similar recruitment methods [7, 15], identification of bots or other sources of invalid records is of particular concern and several methods for such identification have been utilized. We importantly identified that even with survey platform tools (such as the “Prevent Ballot Box Stuffing” option), additional attention to records’ metadata is critical to identify residual entries submitted from bots. Future studies should consider how these tools could be applied in consort in order to generate a maximally valid sample.

Conclusions

The unique challenges of diffuse, non-probability-based recruitment via social media can raise concerns about efficiency, generalizability, and validity. We demonstrate the feasibility of implementing a rapidly deployed yet rigorous trial that engaged and retained a large cohort using off-the-shelf tools. In light of these successes in virtually recruiting for and conducting an educational trial, researchers may wish to implement similar strategies and reporting methods for future studies.
Table 4

Psychometric properties of novel scarce resource allocation questions

AllLayHCW
Values for prioritization logisticsa0.60660.60680.6099
 They should try to save the most number of lives possible0.27870.20440.4528
 They should take life support away from some patients in order to give it to other patients who are more likely to survive− 0.0149− 0.05640.0904
 They should make decisions on a first-come, first-served basis0.08980.12440.0051
 They should apply the same rules to decide who gets life support to all patients equally0.70520.71830.6810
 The same rules should apply to all patients even if they were admitted to the hospital before the crisis started0.81950.83560.7709
 The same rules should apply to all patients even if they are in the hospital for reasons that are not related to the disaster or pandemic0.78880.80640.7483
 Hospital committees (instead of individual doctors) should make these decisions0.20570.24770.1213
 Hospital committees should not know the identities of the patients and use only medical information to make decisions0.32580.31290.3712
 Policies like this should be developed with input from patients and community members0.23930.28540.1379
Values for prioritization on health factorsb0.76980.75940.8523
 Patients who are deemed less likely to survive and make it out of the hospital alive0.69540.67770.7285
 Patients who have physical or intellectual disabilities0.53780.50960.5629
 Patients who have shorter expected lifespans because of chronic illness0.77690.74280.8268
 Patients who are elderly0.69780.69620.6795
 Patients who are children− 0.1289− 0.1233− 0.1597
 Patients expected to have a poor quality of life if they survive0.72300.70830.7906
 Patients expected to need life support for a long time to recover from their illness0.65690.70510.6198
 Patients who are chronically dependent on ventilators0.7013
 Patients in persistent vegetative or minimally conscious states0.7965
Values for prioritization on social factorsc0.88090.87200.8954
 People who are wealthy, famous, or in positions of power (for example: celebrities or politicians)0.28000.18130.4841
 People who are a racial or ethnic minority0.64350.68870.6315
 People who are LGBTQ+ (e.g, lesbian, gay, bisexual, or transgender)0.63330.67580.5865
 People who are prisoners0.66110.65900.6786
 People without health insurance0.65390.65110.6720
 People who are undocumented immigrants0.69780.70610.7012
 Patients who have shorter expected lifespans because of a serious health condition even if that condition is more common among people with a disability0.78620.76620.8025
 Patients who have shorter expected lifespans because of a serious health condition even if that condition is more common among racial or ethnic minorities0.82200.83970.7876
 Patients who have shorter expected lifespans because of a serious health condition even if that condition is more common among people living in poverty0.81140.81830.7923
 People who are philanthropic donors to the hospital or health system0.5067
Values for prioritization exemptionsd0.82840.82430.8627
 Patients who are pregnant in the first trimester0.52880.54930.4904
 Patients who are pregnant in the third trimester0.55880.58890.4672
 First responders (for example: police, fire fighters)0.75260.73520.7557
 Health care workers in general who are critical to caring for patients0.80460.79810.7473
 Health care workers specifically who are on the front lines and at increased risk of harm from the pandemic0.78020.79640.6686
 Patients who are participating in medical research studies0.48510.48690.5153
 Patients who are the sole or only caregiver of a family member (for example: a child or a disabled or elderly relative)0.58660.57880.6019
 Members of the military or veterans0.60480.58580.6495
 Public officials (for example: a mayor, governor, president, or congressperson)0.39310.33590.5992
 Patients who are on the list to get an organ transplant0.19840.24360.2232
 Patients who recently received an organ transplant0.38760.40260.4441
 Patients who have recently undergone major surgery (not related to a transplant)0.5039
 Patients who have had a complication from medical care (for example: a procedural or surgical complication or adverse reaction to a medication)0.4420
 Families or friends of critical health workers0.4583
Preferences for policy disclosureen/a0.57950.6676
 Hospitals should make this information public so patients like me know what their policy is even if I never have to go to that hospital0.56550.54340.3472
 I would want a hospital to tell patients like me about their policy only if I were admitted and in critical condition− 0.2361− 0.0380− 0.0380
 I would consider policies like this when deciding if I would go to a certain hospital0.46660.53970.8641
 I would feel more at ease if my doctor verbally explained how a policy like this works0.6488
 I would feel more at ease if my doctor provided a written explanation of how a policy like this works0.6744
 I would consider policies like this when deciding where I would tell my friends or family to seek care0.9026
 I would consider policies like this when deciding where I would refer my patients for hospital care0.9020
 I would feel comfortable verbally explaining how a policy like this works to patients0.2156
 I would feel comfortable providing a written explanation of how a policy like this works to patients0.2615
 I would feel more comfortable if someone else other than me explained this policy to patients0.1395
 I would be comfortable explaining that a patient had to be taken off of a ventilator due to a policy decision to that patient or their family0.0647
Trust in policy implementationf0.56660.56710.7856
 I trust that hospitals and doctors will apply policies like this in a fair and consistent way0.87250.88830.7092
 I trust hospitals and doctors to be honest and transparent about how resources are used in a crisis0.88970.89980.7316
 I feel anxious or worried when I think about policies like these− 0.1574− 0.1805− 0.1966
 I would trust my doctors to be honest with me about my chances for survival if I were extremely ill or in critical condition0.56360.58540.7118
 I feel that I could be honest with a patient about their chances for survival0.5245
 I feel that my colleagues could be honest with a patient about their chances for survival0.6869
 I feel like I could apply policies like this in a fair and consistent way0.7258
 I feel like my colleagues could apply policies like this in a fair and consistent way0.8112
 I feel like my employer would support me if I had to make these types of decisions0.5690
 I feel like I would be distressed or uncomfortable if I had to carry out policies like these− 0.0840
 I feel like I would have adequate legal protection from fallout if I had to carry out policies like these0.3615
 I would have a moral objection to carrying out a policy like this− 0.1630

Standardized factor loadings for each item and Cronbach’s alphas for entire scales (in italics) are shown for the sample overall (“All” column), among a lay (non-HCW) sample (“Lay” column), and among a health care worker (HCW) sample (“HCW” column). Exploratory factor analysis revealed that single factor solutions were appropriate for most scales (Eigenvalues from 0.5933 [for “Preferences for Policy Disclosure” among All respondents] to 4.5293 [for “Values for Prioritization on Social Factors” among HCW] with a median of 2.8018)

aRespondents were asked “Please tell us how you feel hospitals and health care workers should make decisions like these in general.” Items were rated from 1, I strongly disagree with this to 10, I strongly agree with this

bRespondents were asked “How strongly do you feel the following health factors should influence how hospitals and health care workers decide who receives life support during a crisis like a disaster or pandemic?” Items were rated from 1, Should be much less likely to get life support to 5, Should not influence one way or the other to 9, Should be much more likely to get life support

cRespondents were asked “How strongly do you feel the following factors not related to a patient’s health should be considered when making decisions about who should receive life support during a crisis like a disaster or pandemic?” Items were rated from 1, Should be much less likely to get life support to 5, Should not influence one way or the other to 9, Should be much more likely to get life support

dRespondents were asked “There may be some situations where exceptions are made for certain groups of people under policies like this. How strongly do you feel these groups should be considered when making decisions about who should receive life support during a crisis like a disaster or pandemic?” Items were rated from 1, Should be much less likely to get life support to 5, Should not influence one way or the other to 9, Should be much more likely to get life support

eRespondents were asked “How strongly do you agree or disagree with the following statements about how you would prefer to learn about such policies?” Items were rated from 1, I strongly disagree with this to 10, I strongly agree with this. Given that there were only three items similarly available for both lay and HCW samples, overall Cronbach’s alpha is not reported

fRespondents were asked “How strongly do you agree or disagree with the following statement about how policies like this would be applied?” Items were rated from 1, I strongly disagree with this to 10, I strongly agree with this

  24 in total

1.  The PHQ-9: validity of a brief depression severity measure.

Authors:  K Kroenke; R L Spitzer; J B Williams
Journal:  J Gen Intern Med       Date:  2001-09       Impact factor: 5.128

Review 2.  Screening for anxiety disorders with the GAD-7 and GAD-2: a systematic review and diagnostic metaanalysis.

Authors:  Faye Plummer; Laura Manea; Dominic Trepel; Dean McMillan
Journal:  Gen Hosp Psychiatry       Date:  2015-11-18       Impact factor: 3.238

3.  Denominator Issues for Personally Generated Data in Population Health Monitoring.

Authors:  Rumi Chunara; Lauren E Wisk; Elissa R Weitzman
Journal:  Am J Prev Med       Date:  2016-12-21       Impact factor: 5.043

4.  Using Social Media as a Research Recruitment Tool: Ethical Issues and Recommendations.

Authors:  Luke Gelinas; Robin Pierce; Sabune Winkler; I Glenn Cohen; Holly Fernandez Lynch; Barbara E Bierer
Journal:  Am J Bioeth       Date:  2017-03       Impact factor: 11.229

5.  Single item measures of emotional exhaustion and depersonalization are useful for assessing burnout in medical professionals.

Authors:  Colin P West; Liselotte N Dyrbye; Jeff A Sloan; Tait D Shanafelt
Journal:  J Gen Intern Med       Date:  2009-10-03       Impact factor: 5.128

6.  The Patient Health Questionnaire-2: validity of a two-item depression screener.

Authors:  Kurt Kroenke; Robert L Spitzer; Janet B W Williams
Journal:  Med Care       Date:  2003-11       Impact factor: 2.983

7.  Recruitment via social media: advantages and potential biases.

Authors:  Catherine Benedict; Alexandria L Hahn; Michael A Diefenbach; Jennifer S Ford
Journal:  Digit Health       Date:  2019-08-06

Review 8.  A systematic review of publications assessing reliability and validity of the Behavioral Risk Factor Surveillance System (BRFSS), 2004-2011.

Authors:  Carol Pierannunzi; Shaohua Sean Hu; Lina Balluz
Journal:  BMC Med Res Methodol       Date:  2013-03-24       Impact factor: 4.615

Review 9.  The Role of Social Media in Enhancing Clinical Trial Recruitment: Scoping Review.

Authors:  Ida Darmawan; Caitlin Bakker; Tabetha A Brockman; Christi A Patten; Milton Eder
Journal:  J Med Internet Res       Date:  2020-10-26       Impact factor: 5.428

10.  Fair allocation of scarce medical resources in the time of COVID-19: what do people think?

Authors:  Francesco Fallucchi; Marco Faravelli; Simone Quercia
Journal:  J Med Ethics       Date:  2020-10-12       Impact factor: 2.903

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.