Literature DB >> 36018628

Testing the Effectiveness of an Animated Decision Aid to Improve Recruitment of Control Participants in a Case-Control Study: Web-Based Experiment.

Sandro T Stoffel1,2, Jing Hui Law1,3, Robert Kerrison1,4, Hannah R Brewer5, James M Flanagan5, Yasemin Hirst1,5.   

Abstract

BACKGROUND: Participation in case-control studies is crucial in epidemiological research. The self-sampling bias, low response rate, and poor recruitment of population representative controls are often reported as limitations of case-control studies with limited strategies to improve participation. With greater use of web-based methods in health research, there is a further need to understand the effectiveness of different tools to enhance informed decision-making and willingness to take part in research.
OBJECTIVE: This study tests whether the inclusion of an animated decision aid in the recruitment page of a study website can increase participants' intentions to volunteer as controls.
METHODS: A total of 1425 women were included in a web-based experiment and randomized to one of two experimental conditions: one in which they were exposed to a simulated website that included the animation (animation; n=693, 48.6%), and one in which they were exposed to the simulated website without the animation (control; n=732, 51.4%). The simulated website was adapted from a real website for a case-control study, which invites people to consider taking part in a study that investigates differences in purchasing behaviors between women with and without ovarian cancer and share their loyalty card data collected through 2 high street retailers with the researchers. After exposure to the experimental manipulation, participants were asked to state (1) their intention to take part in the case-control study, (2) whether they would be willing to share their loyalty card for research, and (3) their willingness to be redirected to the real website after completing the survey. Data were assessed using ordinal and binary logistic regression, reported in percentages (%), adjusted odds ratio (AOR), and 95% confidence intervals.
RESULTS: Including the animation in the simulated website did not increase intentions to participate in the study (AOR 1.09; 95% CI 0.88-1.35) or willingness to visit the real study website after the survey (control 50.5% vs animation 52.6%, AOR 1.08; 95% CI 0.85-1.37). The animation, however, increased the participants' intentions to share the data from their loyalty cards for research in general (control 17.9% vs animation 26%; AOR 1.64; 95% CI 1.23-2.18).
CONCLUSIONS: While the results of this study indicate that the animated decision aid did not lead to greater intention to take part in our web-based case-control study, they show that they can be effective in increasing people's willingness to share sensitive data for health research. ©Sandro T Stoffel, Jing Hui Law, Robert Kerrison, Hannah R Brewer, James M Flanagan, Yasemin Hirst. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 26.08.2022.

Entities:  

Keywords:  animation; case-control; data; decision; effectiveness; epidemiology; experiment; health research; online; online experiment; recruitment; research; research participation; survey; volunteer; willingness

Mesh:

Year:  2022        PMID: 36018628      PMCID: PMC9463615          DOI: 10.2196/40015

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   7.076


Introduction

One of the most effective methods to test for exposure in epidemiological research is to conduct a case-control study in which people who have an illness are compared retrospectively with a matching population without the outcome [1]. Although it is a reliable methodology to test for associations, poor recruitment of population representative controls often undermines such studies [2]. Previous research reports the most common methods of recruiting control participants for case-control studies as follows: door-to-door recruitment, postal invitation, and random digit dialing [1,2]. More recently, with greater access to the internet, many cohort studies have moved their participant management and recruitment online (using unique websites), providing new opportunities to recruit participants, potentially improving diversity and ease of data collection [3,4]. Despite several advantages, however, caution needs to be exercised with the opportunistic recruitment of participants to web-based studies. For example, a recent study reported less than 4% of participants who visited a study recruitment website, after clicking on a targeted social media advertisement campaign, went on to sign up to the research [5]. This indicates that while individuals may be forming some interest to take part in research studies by clicking on a recruitment advertisement, their intention does not always translate to survey completion after they land on the research website. Evidence on the barriers and facilitators of web-based survey completion primarily relates to the completion of stand-alone web-based surveys, rather than the use of unique websites to recruit participants to case-control studies [6-11]. These studies suggest that individuals’ trust in the organization carrying out the research, whether they are early adopters of technology and high in literacy, and whether the research is in line with the individuals’ values and beliefs are positive predictors of individual participation. Recommendations to achieve better outcomes include clear communication of the research goals, transparency about how data will be used, and shorter survey length. Clinical trials have attempted to address some of the above (eg, transparency about how data will be used), using audio-visual decision aids to supplement the process of obtaining informed consent [12]. Communicating information via these mediums (enabled through web-based recruitment strategies), have the potential to reduce the associated cognitive load, facilitate further engagement with the research aims, generate positive attitudes toward the targeted behavior, and subsequently motivate engagement in the behavior itself [13-16]. To our knowledge, the potential impact of animated decision aids on intentions to take part in a web-based case-control study has not previously been investigated. This study aims to measure the effectiveness of an animated decision aid as a supplementary tool on a simulated website of a case-control study to encourage participation.

Methods

Setting

This study comprised a randomized web-based experiment, which assessed the effectiveness of adding an animated decision aid to a simulated website. The simulated website the animation was designed for (or added to) was the recruitment website for the case-control study: Cancer Loyalty Card Study (CLOCS) [17].

Cancer Loyalty Card Study

CLOCS is an observational case-control study that aims to investigate the self-care behaviors of patients with ovarian cancer prior to their cancer diagnosis. It seeks to do this by investigating differences in transactional data (such as medication purchasing) between women with and without ovarian cancer (the transactional data are collected through the loyalty cards of 2 UK-based high street retailers). Cases (ie, women with ovarian cancer) are recruited through participating National Health Service sites, while controls are recruited through the study website. Full details for CLOCS have previously been reported in the study protocol [17].

Animated Decision Aid

A key challenge for CLOCS recruitment has been communicating the research aims clearly, and our previous research highlights that the public often needs further explanations for how individual transactional data can be used in health research [18]. To improve public understanding and engagement with the aims of CLOCS, an animated decision aid was jointly prepared by Science Animated Limited, 2 patient representatives, and the CLOCS research team prior to the initiation of this web-based experiment. It aimed to convey key facts on ovarian cancer, the potential contribution of CLOCS in informing earlier diagnosis, as well as how women in the general population can play a role to aid its efforts (based on the participant information sheet tailored and approved by a National Health Service Research Ethics Committee [19/NW/0427-SA1], included in Multimedia Appendix 1). However, it should be noted that the animation was designed as a supplement to the main study materials, not as a key participant communication material to prompt informed decision-making. The animation features English subtitles and is 123 seconds (2 minutes and 3 seconds) long [19].

Procedure

The randomized web-based experiment was programmed in Survey Monkey. In July 2020, women who were eligible to take part in CLOCS as a control(ie, between the ages of 18 and 70, living in the United Kingdom, and without an ovarian cancer diagnosis) were recruited through a survey vendor (Dynata Limited). For those who were interested in taking part, information about the experiment, including a brief description of CLOCS, was presented and followed by the completion of the study consent form. If participants consented and were eligible, they were randomized (in a 1:1 ratio) to one of the following two experimental conditions: the simulated CLOCS website without the animation (control), or the simulated CLOCS website with the animation (animation) (Multimedia Appendix 2 and 3). Once everyone viewed the simulated website, they were asked to complete the survey, where they were required to indicate their intention to take part in CLOCS using a 4-point Likert scale (definitely yes, probably yes, probably not, and definitely not) adapted from previous research [20-22]. The study participants were then asked to indicate their loyalty card use by selecting from a list of high street retailers’ loyalty cards and whether they would be willing to share their loyalty card data for research purposes. The latter question was adapted from research on willingness to share electronic health data, which uses 4 commonly used models of consent for the use of data [23]. In the next step, the study participants were asked about their educational level, annual household income, and health literacy. For the latter, the eHealth Literacy Scale was used. This scale assesses individuals’ retrieval and judgement of health information on the internet [24]. It consists of 8 items rated on a 5-point scale, ranging from 1 to 5, and has demonstrated considerable reliability and validity [24]. The participants’ scores across the 8 items on the scale were summed and calculated for a sample mean. Individual scores below and above (or equal to) this sample mean were defined as low and high health literacy, respectively [25,26]. The final survey item was included as a behavior proxy. The participants were asked whether they want to be redirected to the actual CLOCS website for more information on how to take part. Those who responded that they would like to visit the website were provided with a link to the CLOCS website at the final page of the survey [22]. The website opened in a new tab for participants who clicked on the link. No data were collected from the study participants for their direct participation in CLOCS associated with the experiment.

Ethics Approval

Ethics approval for this study was granted by the UCL Research Ethics Committee (Project ID: 17823/001).

Data Analysis

A pilot study has been conducted beforehand for the purpose of sample size calculations. Based on the findings from the initial sample of 359 participants, with a 10% difference in intention to take part (definitely yes or probably yes versus definitely no or probably no), we determined the number of participants needed to achieve 95% certainty and 80% power was 650 per trial arm. Data from participants in both the pilot and final sample were combined for analysis. Sample characteristics were assessed using descriptive statistics (Table S1 in Multimedia Appendix 4). To aid interpretation, the participants’ income and educational levels were dichotomized in inferential analyses. For income, we used £30,000 (US $37,000) as the cut-off point, based on the average household income in the United Kingdom (reported by the Office for National Statistics, 2020) [27]. For education, we categorized participants into those with General Certificate of Secondary Education or A Levels and those with a university degree [20-22]. Differences in intentions to take part in CLOCS were assessed using univariate and multivariate ordinal logistic regression. Willingness to visit the actual website and willingness to share loyalty card data for research purposes, between groups, were assessed using univariate and multivariate binary logistic regressions. Odds ratios (ORs), 95% confidence intervals, and P values are presented in the results, with P values below .05 regarded as statistically significant. Participants who spent a short amount of time completing the survey (ie, survey speeders) were excluded from the analysis (Multimedia Appendices 5-8, based on the 50% cut-off points provided from previous research using median values) [28]. We report the analysis for the whole sample in Table S2 (Multimedia Appendix 9) and the distribution of the time spent on the survey before and after the exclusion of the speeders in Multimedia Appendices 5-8 [29]. Additional analyses for interaction between the intervention and health literacy led to null results and were not reported due to the unbalanced proportion of those with low health literacy and high literacy in the study population.

Results

Study Sample

Figure 1 demonstrates the flow of participants through the study. In total, 6034 women were invited to participate, of which 4609 (76.4%) were excluded as they were not eligible, dropped out, or discontinued the initial screening. The remaining 1425 (23.6%) were then randomized to one of the following two experimental conditions: 732 (51.4%) were randomly allocated to the control condition, and 693 (48.6%) to the animation condition. Across conditions, 131 (9.2%) did not finish the survey after randomization. Furthermore, 137 (9.6%) were excluded as they spent less than 50% of the median time. The analytical sample consisted of 1157 women—610 (52.7%) participants in the control condition, and 547 (47.3%) in the animation condition.
Figure 1

Flow through the study.

Most women in the analytical sample were aged between 55 and 70 years (n=417, 36.9%), did not have a university degree (n=686, 59.3%), and had an annual household income of less than £30,000 (US $37,000; n=622, 53.8%). The mean eHealth Literacy Scale score for the participants was 30.3 out of 40; thus, those who scored below this mean were classified as having low health literacy [20,21]. Post hoc comparisons revealed that sociodemographic characteristics were comparable between the two experimental conditions (Table S1 in Multimedia Appendix 4). Flow through the study.

Intentions to Participate in CLOCS

Intentions to participate in CLOCS were generally very high, with 69.7% (n=807) of women stating that they would probably or definitely participate. Figure 2 shows the distribution of the intentions to participate in CLOCS after exposure to the simulated website. The ordered logistic regressions in Table 1 show that the inclusion of the animation did not affect participation intentions (OR 1.12; 95% CI 0.90-1.39 and AOR 1.09; 95% CI 0.88-1.35). The regression further shows that older women aged 55-70 years stated lower intentions to participate (OR 0.56; 95% CI 0.42-0.74), while those with an income above average (OR 1.36; 95% CI 1.06-1.63), one or more existing loyalty cards (AOR 2.05; 95% CI 1.16-3.62), and low health literacy (AOR 1.26; 95% CI 1.01- 1.58) had higher intentions to take part in CLOCS.
Figure 2

Distribution for intention to take part in Cancer Loyalty Card Study.

Table 1

Ordered logistic regression on intention to participate in Cancer Loyalty Card Study (N=1157).

VariablesUnadjusted regressionAdjusted regression
AORa95% CIP valueAOR95% CIP value
Condition
ControlReferencebReference
Animation1.1200.904-1.389.301.0880.876-1.352.45
Age (years)
18-34ReferenceReference
35-440.9980.726-1.371.990.9520.691-1.311.76
45-540.8060.577-1.127.210.7990.569-1.122.20
55-700.5560.421-0.735<.0010.5910.445-0.786<.001
Education
Below or equal to GCSEcReferenceReference
University degree1.3201.060-1.643.011.1450.909-1.441.25
Income
Below averageReferenceReference
Above average1.3621.097-1.690.011.2631.006-1.584.04
Card
NoReferenceReference
Yes2.1141.202-3.719.0092.0511.164-3.616.01
Health literacy
High literacyReferenceReference
Low literacy1.5011.125-2.002.0061.3811.029-1.854.03

aAOR: adjusted odds ratio.

bNot applicable.

cGCSE: General Certificate of Secondary Education.

Distribution for intention to take part in Cancer Loyalty Card Study. Ordered logistic regression on intention to participate in Cancer Loyalty Card Study (N=1157). aAOR: adjusted odds ratio. bNot applicable. cGCSE: General Certificate of Secondary Education.

Willingness to Share Loyalty Card Data for Research

Most study participants stated that their data should be used but that they should have the option of saying no (control condition: n=246, 40.3%; and animation condition: n=181, 33.1%; Figure 3). However, significantly more participants in the animation condition, compared to control, indicated that they would be willing to provide the data if they were needed—as shown in the binary logistic regressions in Table 2 (control: n=109, 17.9% vs animation: n=142, 26.0%; OR 1.61; 95% CI 1.22-2.14 and AOR 1.64; 95% CI 1.23-2.18). Similarly, as with the intentions to participate in CLOCS, women aged 55-70 years were again less likely to state that their data should be used if needed compared to those aged 18-34 years (n=71, 17% vs n=72, 24.8%; AOR 0.62; 95% CI 0.43-0.89).
Figure 3

Distribution for willingness to share loyalty card data .

Table 2

Binary logistic regression on agreeing to share data from loyalty cards when needed (N=1157).

VariablesTotal, n (%)Unadjusted regressionAdjusted regression
AORa95% CIP valueAOR95% CIP value
Overall251 (21.7)b
Condition
Control109 (17.9)ReferenceReference
Animation142 (26.0)1.6121.216-2.136.0011.6401.232-2.184 .001
Age (years)
18-3472 (24.8)ReferenceReference
35-4455 (22.9)0.9000.602-1.346.610.8640.575-1.299.48
45-5453 (25.2)1.0220.678-1.540.921.0460.688-1.591.83
55-7071 (17.0)0.6210.430-0.899.010.6010.411-0.878.008
Education
Below or equal to GCSEc147 (21.4)ReferenceReference
University degree104 (22.1)1.0390.782-1.380.790.9320.690-1.260.65
Income
Below average income130 20.9)ReferenceReference
Above average income121 (22.6)1.1060.836-1.464.481.0520.782-1.414.74
Card
No9 (20.9)ReferenceReference
Yes242 (21.7)1.0480.496-2.216.901.0940.511-2.344.82
Health literacy
High literacy214 (22.0)ReferenceReference
Low literacy37 (20.0)0.8860.599-1.309.540.8070.541-1.204.30

aAOR: adjusted odds ratio.

bNot applicable.

cGCSE: General Certificate of Secondary Education.

Distribution for willingness to share loyalty card data . Binary logistic regression on agreeing to share data from loyalty cards when needed (N=1157). aAOR: adjusted odds ratio. bNot applicable. cGCSE: General Certificate of Secondary Education.

Willingness to Visit the CLOCS Website After the Survey

A slight majority of the study participants (n=596, 51.5%) indicated that they would like to visit the CLOCS website after the end of the survey (Multimedia Appendix 10). Table 3 shows that there was no difference between the two experimental conditions (control: n=308, 50.5% vs animation: n=288, 52.6%; OR 1.09; 95% CI 0.87-1.37 and AOR 1.08; 95% CI 0.85-1.37). Women with low health literacy, in comparison to those with high health literacy (n=111, 60% vs n=485, 49.9%; AOR 1.43; 95% CI 1.03-1.98), and women aged 35-44 years, in comparison to those aged 18-34 years, were more interested in visiting the study website (n=145, 60.4% vs n=148, 51%; AOR 1.49; 95% CI 1.05-2.12).
Table 3

Binary logistic regression on willingness to visit website after the survey (N=1157).

VariablesTotal, n (%)Unadjusted regressionAdjusted regression
AORa95% CIP valueAOR95% CIP value
Overall596 (51.5)b
Condition
Control308 (50.5)ReferenceReference
Animation288 (52.7)1.0900.865-1.374.461.0790.853-1.367.53
Age (years)
18-34 years148 (51.0)ReferenceReference
35-44 years145 (60.4)1.4641.036-2.071.031.4891.048-2.115.03
45-54 years118 (56.2)1.2310.861-1.758.251.3030.905-1.877.16
55-70 years185 (44.4)0.7650.566-1.033.080.8250.606-1.123.22
Education
Below or equal to GCSEc337 (49.1)ReferenceReference
University degree259 (55.0)1.2651.000-1.601.051.2140.946-1.559.13
Income
Below average income322 (51.8)ReferenceReference
Above average income274 (51.2)0.9780.776-1.233.850.8650.677-1.107.25
Card
No18 (41.9)ReferenceReference
Yes578 (51.9)1.4980.808-2.776.201.5210.810-2.854.19
Health literacy
High literacy485 (49.9)ReferenceReference
Low literacy111 (60.0)1.5061.094-2.074.011.4311.031-1.986.03

aAOR: adjusted odds ratio.

bNot applicable.

cGCSE: General Certificate of Secondary Education.

Binary logistic regression on willingness to visit website after the survey (N=1157). aAOR: adjusted odds ratio. bNot applicable. cGCSE: General Certificate of Secondary Education.

Discussion

Key Findings

This randomized web-based experiment examined the effectiveness of an animated decision aid to increase the willingness to participate in a case-control study. The results show that the animation did not increase intentions to participate in a real-world case-control study (CLOCS), or willingness to visit the real study website after the survey. However, the animation increased the participants’ willingness to share data from their loyalty cards for research. Interestingly, immediately after the completion of this web-based experiment, there was a spike in activity within the case-control study, with over 100 people signing up to participate in CLOCS.

Comparison With Previous Literature

Our findings reflect the mixed evidence currently available in the literature. While there has been some support for the effectiveness of animated decision aids in the context of health behavior research, many studies have focused on how the interventions can improve participants’ knowledge of the health behavior concerned, as their main outcome. When assessing participant intention to engage in the behavior (or an objective measurement of the behavior itself), the findings have been more inconsistent [30,31]. An interesting finding in our study is that, among all participants, those with lower health literacy scores were more interested to find out about the study compared with those with high health literacy scores. Our sample size calculations were based solely on the primary outcome; thus, we might have been underpowered to detect the interaction effects of health literacy and outcomes on the behavior proxy. However, the previous studies indicate that multimedia interventions are not always significant in individuals with lower educational levels [14] and low health literacy [15,16]. It has been previously shown in the cancer screening literature that gist-based supplementary materials could be used to enhance engagement with the main literature among people with low numeracy [32], and perhaps using animation and other easy-to-read materials could enhance participant recruitment in health research [33]. Furthermore, the positive association between low health literacy and willingness to visit the website may be explained by factors such as wanting to find more information, the salience of the research topic, and other factors that were not included in this web-based experiment. As such, future studies focusing on the comprehension of the materials among people with low health literacy using a think-aloud methodology could explain this outcome.

Strengths and Limitations

This study has some important limitations, which call for follow-up research. First, we did not include a comprehension assessment to check if participants fully understood or watched the animation. Second, we did not measure attitudes toward the simulated website and CLOCS. A recent systematic review on participant comprehension and informed consent in health research further highlighted that while there are efforts to improve participation rates using various methods, there is a lack of assessment of participant readability, literacy, and standardization of recruitment methods in health research for informed consent procedures [33]. While all the participant-facing CLOCS research materials have been reviewed by patient and public representatives, further assessment of participant comprehension prior to the animation experiment would have strengthened our methodology. However, the rationale for the exclusion of these measures was based on the assumptions that people often form immediate decisions about whether something is relevant to them using heuristic decision-making before establishing deliberative decisions [34]. As such, by excluding cognitive measures in our assessment to minimize judgement and bias, we tried to capture individuals’ potential reactions to the website as close to their reaction in real life. In this context, further studies using eye-tracking experiments on the simulated website will be highly informative to build a better understanding of the interaction with the website and the contents [35]. On the other hand, exposure to the animation increased the intentions to share loyalty card data for research in general, but not for intentions to participate in the CLOCS study; this suggests that there are study-specific characteristics that did not appeal to individuals (eg, actively signing up to provide information) or that the study participants were not eligible. However, a recent study also shows that only half of the population is willing to share shopping data for health research, highlighting the differences in sociodemographic characteristics of people who are willing to share their data for health research [36]. The characteristics of the participants who are willing to take part in CLOCS in this study mirror the results of this age-stratified survey employed in England, with older women less willing to take part. While self-sampling bias will continue to be a concern of case-control studies based on the differences in characteristics of the people who are willing to take part, recruitment strategies could be stratified and tailored to engage different populations who are less willing to take part in health research based on this evidence and the validation of public acceptability. Our results support this evidence further using experimental design with greater internal validity for potential barriers in recruiting participants to a case-control study.

Implications for Policy and Future Research

While such questions of generalizability are warranted, this study still poses important relevance and implications to current research contexts. Due to the COVID-19 pandemic, many research studies have been forced to consider the possibility of being adapted online. This may require researchers to derive additional strategies to reach web-based samples with different characteristics. Underrepresentation in health research is already an issue for minority populations, people with low literacy, and those with greater deprivation using traditional methods of recruitment, unless they are specifically targeted [37,38]; thus, there is a further need to ensure that web-based strategies can provide means for researchers to attract representative samples. Future studies should therefore continue exploring web-based methods to facilitate complex decision-making processes for potential participants of health research. The CLOCS animation has not been actively disseminated for participant recruitment following this dearth of evidence; however, unique findings might be obtained for research of a different nature. Other multimedia formats or mediums such as social media can be further explored in future studies, along with the consideration of potentially important variables such as participants’ willingness to share data for research purposes.

Conclusion

The results of this study indicate that the animated decision aid did not influence the participants’ intention to take part in CLOCS or visit the study website. The animation, however, increased the probability of individuals stating that they would share their loyalty card data for research. Future research should continue exploring methods that can effectively engage participants with low health literacy to participate in complex health research.
  29 in total

1.  Control recruitment in population-based case-control studies.

Authors:  Leslie Bernstein
Journal:  Epidemiology       Date:  2006-05       Impact factor: 4.822

Review 2.  Audio-visual presentation of information for informed consent for participation in clinical trials.

Authors:  Anneliese Synnot; Rebecca Ryan; Megan Prictor; Deirdre Fetherstonhaugh; Barbara Parker
Journal:  Cochrane Database Syst Rev       Date:  2014-05-09

3.  Association of eHealth literacy with cancer information seeking and prior experience with cancer screening.

Authors:  Hyejin Park; Mikyung Moon; Jung Hoon Baeg
Journal:  Comput Inform Nurs       Date:  2014-09       Impact factor: 1.985

4.  Educational video-assisted versus conventional informed consent for trauma-related debridement surgery: a parallel group randomized controlled trial.

Authors:  Yen-Ko Lin; Chao-Wen Chen; Wei-Che Lee; Yuan-Chia Cheng; Tsung-Ying Lin; Chia-Ju Lin; Leiyu Shi; Yin-Chun Tien; Liang-Chi Kuo
Journal:  BMC Med Ethics       Date:  2018-03-09       Impact factor: 2.652

5.  The impact of descriptive norms on motivation to participate in cancer screening - Evidence from online experiments.

Authors:  Christian von Wagner; Yasemin Hirst; Jo Waller; Alex Ghanouni; Lesley M McGregor; Robert S Kerrison; Wouter Verstraete; Ivo Vlaev; Monika Sieverding; Sandro T Stoffel
Journal:  Patient Educ Couns       Date:  2019-04-04

6.  Cancer Loyalty Card Study (CLOCS): protocol for an observational case-control study focusing on the patient interval in ovarian cancer diagnosis.

Authors:  Hannah R Brewer; Yasemin Hirst; Sudha Sundar; Marc Chadeau-Hyam; James M Flanagan
Journal:  BMJ Open       Date:  2020-09-08       Impact factor: 2.692

Review 7.  A reinvestigation of recruitment to randomised, controlled, multicenter trials: a review of trials funded by two UK funding agencies.

Authors:  Ben G O Sully; Steven A Julious; Jon Nicholl
Journal:  Trials       Date:  2013-06-09       Impact factor: 2.279

8.  Self-Care Behaviors of Ovarian Cancer Patients Before Their Diagnosis: Proof-of-Concept Study.

Authors:  James M Flanagan; Hanna Skrobanski; Xin Shi; Yasemin Hirst
Journal:  JMIR Cancer       Date:  2019-01-17

9.  Evidence on the effectiveness of health literacy interventions in the EU: a systematic review.

Authors:  Boudewijn B Visscher; Bas Steunenberg; Monique Heijmans; Jolien M Hofstede; Walter Devillé; Iris van der Heide; Jany Rademakers
Journal:  BMC Public Health       Date:  2018-12-29       Impact factor: 3.295

Review 10.  Online Patient Recruitment in Clinical Trials: Systematic Review and Meta-Analysis.

Authors:  Mette Brøgger-Mikkelsen; Zarqa Ali; John R Zibert; Anders Daniel Andersen; Simon Francis Thomsen
Journal:  J Med Internet Res       Date:  2020-11-04       Impact factor: 5.428

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.