Literature DB >> 21998626

Can simply answering research questions change behaviour? Systematic review and meta analyses of brief alcohol intervention trials.

Jim McCambridge1, Kypros Kypri.   

Abstract

BACKGROUND: Participant reports of their own behaviour are critical for the provision and evaluation of behavioural interventions. Recent developments in brief alcohol intervention trials provide an opportunity to evaluate longstanding concerns that answering questions on behaviour as part of research assessments may inadvertently influence it and produce bias. The study objective was to evaluate the size and nature of effects observed in randomized manipulations of the effects of answering questions on drinking behaviour in brief intervention trials. METHODOLOGY/PRINCIPAL
FINDINGS: Multiple methods were used to identify primary studies. Between-group differences in total weekly alcohol consumption, quantity per drinking day and AUDIT scores were evaluated in random effects meta-analyses. Ten trials were included in this review, of which two did not provide findings for quantitative study, in which three outcomes were evaluated. Between-group differences were of the magnitude of 13.7 (-0.17 to 27.6) grams of alcohol per week (approximately 1.5 U.K. units or 1 standard U.S. drink) and 1 point (0.1 to 1.9) in AUDIT score. There was no difference in quantity per drinking day.
CONCLUSIONS/SIGNIFICANCE: Answering questions on drinking in brief intervention trials appears to alter subsequent self-reported behaviour. This potentially generates bias by exposing non-intervention control groups to an integral component of the intervention. The effects of brief alcohol interventions may thus have been consistently under-estimated. These findings are relevant to evaluations of any interventions to alter behaviours which involve participant self-report.

Entities:  

Mesh:

Year:  2011        PMID: 21998626      PMCID: PMC3187747          DOI: 10.1371/journal.pone.0023748

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

The contribution of behavioural risk factors, such as physical inactivity, tobacco smoking, and unhealthy alcohol use, is estimated to be at least 20% of the total global burden of disease [1]. Accordingly there is increasing investment in the development of behavioural interventions. Attempts to influence behaviour have also gained a new prominence in wider public policy, for example in efforts to combat climate change and domestic terrorism. Trials and other evaluation studies typically involve asking study participants about their own behaviour over time, which in some cases may be validated with objective measures. Such data are fundamental to the behavioural sciences [2]. This process of reporting on one's own behaviour may itself induce reflection and actual change and this was the original reason for the introduction of control groups in behavioural research a century ago [3]. The Hawthorne effect, wherein participants change their behaviour in response to being monitored, has been widely discussed for three quarters of a century [4], [5], [6] and has entered “the folklore of behavioural science” [7]. Accounts of unexpected improvements apparently due to research assessments are often invoked as possible explanations for null findings in trials across a wide range of behaviours (see for example [8]). As the technological capacity for monitoring behaviour grows, for example through the use of pedometers in relation to walking, so does the need to better understand this phenomenon [9]. Longstanding recognition of the Hawthorne effect and the possible implications of answering questions in the context of research study assessments have not, however, led to any substantial tradition of experimental study in health sciences or elsewhere. Alongside some interesting non-experimental studies [10], [11], [12], there exist somewhat isolated trials of the effects of questionnaire completion on disparate health outcomes [13], [14], [15], [16], [17], [18], [19], [20], [21]. This situation has changed recently in the field of brief alcohol intervention trials in which individualised feedback, advice and brief counselling are evaluated for public health benefit [22], [23]. Assessment effects may have greater bias potential in these studies because of similarities with the evaluated interventions, which invariably require assessment, and because effect sizes are themselves small, their value deriving from potential for wide dissemination [24]. Assessment effects and brief interventions may also operate by similar mechanisms, acting upon the self-regulation of behaviour [21]. There have been no systematic reviews which investigate whether answering questions on a particular behaviour, which may be intrinsic to intervention study, subsequently impacts upon that behaviour. The objective of the present study is therefore to evaluate the size and nature of effects observed in randomized manipulations of the effects of answering questions on drinking behaviour in the non help-seeking populations who participate in brief intervention trials.

Methods

Study design & data collection

We excluded assessments undertaken with the specific purpose of changing behaviour, as these were judged likely to involve additional components, which may or may not have been reported. We are thus studying the effects of research assessments only. Peer-reviewed journal publications in any language were included and studies undertaken in alcohol treatment services excluded. There were no other selection criteria. This review has been reported in accordance with the PRISMA statement and was undertaken without a published protocol [25]. There have been many reviews of the brief alcohol intervention literature and we used these to identify relevant studies for this review (‘A’ in Figure 1). We contacted experts both individually and via three groups, the International Network on Brief Interventions for Alcohol Problems (INEBRIA), the Kettil Bruun Society for Social and Epidemiological Research on Alcohol, and the Research Society on Alcoholism (also included in ‘A’ in Figure 1). We searched PubMed using the terms “assessment” AND “alcohol” AND “reactivity”, with the final database searches taking place on 8th February 2011. The flowchart in Figure 1 summarises this process. Nine studies were excluded when the reports revealed the presence of non-assessment intervention components. Finally, three studies were excluded when author contact ascertained that assessments were undertaken specifically for intervention purposes [26], [27], [28].
Figure 1

Participant flowchart.

Outcomes & analyses

Various outcome measures are used in this literature. A priori we decided to select outcomes for quantitative study according to their availability: (1) overall total alcohol consumed within the past week or a typical recent week was reported or could be derived in all studies which provided quantitative data; (2) quantity consumed per drinking day was missing in only two cases; and (3) the WHO Alcohol Use Disorders Identification Test [29] (AUDIT) scores were available in half of the studies which provided quantitative data. It would have been possible also to have investigated the pooled effect on binary AUDIT outcome, though this was judged repetitious. All eight studies reported since 2005 provided unpublished data for inclusion in the meta analysis, with the authors of the earlier studies no longer having access to the raw data. These methods precluded certain forms of bias within studies, such as selective reporting of outcomes. All other outcomes were evaluated in a minority of available datasets. Outcomes 1 and 2 were converted into grams of ethanol [30]. Between-group mean differences in outcomes in the follow-up samples and their standard errors were calculated. Two trials had multiple follow-up intervals (at 1, 6 and 12, and 6 and 12 months respectively [31], [32]). Assessment effects were known to have been reported at 1 month and 12 months respectively and prior to analysis we decided that it was appropriately conservative to use the 6 month data as a summary measure in both studies to simplify the analyses. All data were meta-analysed in STATA version 10 with outcomes pooled in random effects models using the method of DerSimonian and Laird [33]. The I-squared statistic was used to evaluate the extent of heterogeneity [34].

Results

Ten trials were identified for inclusion in this systematic review [31], [32], [35], [36], [37], [38], [39], [40], [41], [42], [43]. Two of these trials, including one study which reported outcomes separately by gender [39], [40], [41] did not provide findings for meta-analysis as outcome data were unreported and datasets were no longer accessible (see below). The characteristics of the included studies are presented in Table 1. One trial was not individually randomised, allocation being cluster randomised in weekly groups for each general practitioner, though it was not described as such because the terminology was not in common use at that time [35].
Table 1

Characteristics of studies included in meta-analyses.

Richmond et al. 1995 Kypri & McAnally 2005 Carey et al. 2006 Daeppen et al. 2007 Kypri et al. 2007 McCambridge & Day 2008 Walters et al. 2009 Cherpitel et al. 2010
Country AustraliaNew ZealandUSASwitzerlandNew ZealandBritainUSAPoland
Setting General PracticeUniversity primary healthcare clinicUniversityEmergency Department (ED)University primary healthcare clinicUniversityUniversityEmergency Department (ED)
Study design 2 arm comparison nested within 4-arm trial2 arm comparison nested within 3-arm trial2 arm comparison nested within 6-arm trial2 arm comparison nested within 3-arm trial2 arm comparison nested within 4-arm trialDedicated 2-arm trial2 arm comparison nested within 5-arm trial2 arm comparison nested within 3-arm trial
Population age, gender composition 19–70 (mean 37.7); 43% women17–24 years (mean 20.2); 49% women18–25 years (mean 19.2); 67% women18 years and over (mean 36.7); 22% women17–29 years (mean 20.2); 52% women18–24 years (mean 20.6); 67% women18–38 years (mean 19.8); 66% women18 years and over (39% <30 years); 16% women
Eligibility/screening criteria >210/350 g ethanol per week for men/womenNone1+ episodes of heavy drinking (men: ≥5 drinks; women ≥4 drinks) in an average week or 4 episodes in the last month; class status not seniorInjury presentation to ED 11am–11pm, men under 65 years, 14 drinks per week and ≥5 drinks per session past 30 days, men over 65 years and women 7 drinks per week and ≥4 drinks per session past 30 daysScore of ≥8 on AUDITNone1+ episodes of heavy drinking (men: ≥5 drinks; women ≥4 drinks) in the preceding 2 weeksPresentation to ED 4pm-midnight, RAPS4 positive screen or ≥11 drinks per week for men, >6 for women or >4 drinks for men per drinking day, >3 for women
Exclusion of dependent drinkers Yes (physical dependence score>10 or MAST>20)NoneNoneYes, history of alcohol-related treatment in last 12 monthsNoneNoneNoneNone
Baseline sample size Control: 93Assessment: 93Control: 72Assessment: 74Control: 81Assessment: 89Control: 335Assessment: 343Control: 146Assessment: 147Control: 204Assessment: 217Control: 75Assessment: 72Control: 147Assessment: 152
Baseline drinking levels Drinks/weekControl: Mean 37.5 (SD 19.9)Assessment: Mean 34.7 (SD 18.2)% binge drinkersControl: not assessedAssessment: 28% binge drinkersDrinks/weekControl: Mean 19.3 (SD 11.2)Assessment: Mean 18.1 (SD 8.9)a)Days drinking/wkb) Drinks per occasionControl: a) Mean 3.6 (SD 2.3)b) Mean 3.7 (SD 2.8)Assessment:a) Mean 3.5 (SD 2.4)b) Mean 3.8 (SD 2.4)AUDIT scoreControl: Mean 15.1 (SD 5.5)Assessment: 14.9 (SD 5.0)History of Trauma Scale positiveControl: 12%Assessment: 10%Heavy drinking episodesControl: Mean 2.9 (SD 1.6)Assessment: Mean 3.3 (SD 1.9)Drinks per drinking day*Control: Mean 5.5 (SE 0.4)Assessment: Mean 5.6 (SE 0.4)

*1/6 measures of consumption, dependence and prior treatment in both groups. Abbreviations: MAST = Michigan Alcohol Screening Test; AUDIT = Alcohol Use Disorders Identification Test.

*1/6 measures of consumption, dependence and prior treatment in both groups. Abbreviations: MAST = Michigan Alcohol Screening Test; AUDIT = Alcohol Use Disorders Identification Test. Detailed information is presented in Table 2 on the assessment procedures being evaluated, blinding, and the consequent nature of the experimental contrasts employed. In some trials the experimental manipulations involved comparisons of longer versus shorter assessments [31], [32], whilst in others assessment was compared with minimal screening [35], [36], [37], or brief assessment with a screening instrument versus no screening at all [38]. The extent and nature of blinding and other potentially important aspects of study design were also variable across the studies. Table 3 comprises a summary of the primary study outcomes as they were reported. It is noteworthy that few of the statistically significant between-group differences attributed to answering questions are included in the present meta-analyses.
Table 2

Details of experimental contrasts.

Richmond et al. 1995 Kypri & McAnally 2005 Carey et al. 2006 Daeppen et al. 2007 Kypri et al. 2007 McCambridge & Day 2008 Walters et al. 2009 Cherpitel et al. 2010
Participants blind to…?
Study designNot clearYesNot clearNoYesNoYesNo
Group assignmentNot clear; judged likelyYesNot clearNoYesYesYesNo
Focus on drinkingControls: YesAssessment group: NoPartially (other health behaviours assessed)NoPartially (other health behaviours assessed)NoYesNoNo
HypothesisYesYesNot clear; judged likelyNot clear, judged likelyYesYesYesNot clear; judged likely
Content of experimental conditions
Control3-minute Health and Fitness Questionnaire: QF of drinking last 3 months, weight, smoking, exercise habitsBlood pressure measured, demographic detailsDemographic details, height, weight, Daily Drinking Questionnaire, maximum number of drinks in last month and duration of episode, frequency of heavy drinking, RAPIScreening only: 3 alcohol questions within a 10-item lifestyle questionnaireDemographic details, AUDIT, number of drinks consumed in heaviest episode in last 4 weeks, +an 8-page leaflet on the effects of alcohol (online)General health & sociodemographic questionnaireDemographic questions onlyRAPS-4 +3 questions on drinking (drinking days per week, drinks per average drinking day, maximum drinks in one occasion in past month
Assessment (As for controls +)Drinking history; 7-day diary; MAST; physical dependence scoreAge first drink, drank in last 12 months (Y/N), largest amount drunk in the last 4 weeks, AUDIT+non-alcohol measuresTLFB calendar for past 90 days: sequential assessment of alcohol use, drug use and sexual behaviourAUDIT, 7-day TLFB+non-alcohol measures4 weeks later: 14-day retrospective diary, APS, AREAS, perceived peer drinking normsAUDITAlcohol consumption, related problems, protective behaviours, readiness to change, and perceived normsDrinking in 6 hours before injury, feeling drunk at time of injury, attribution to alcohol, 30-day TLFB, SIP, readiness to change
Estimated times in minutes (control/assessment)3/155/15?/302/303/105/85/ 30 minutes (latter only repeated 3 m & 6 m)3/10 (latter only repeated 3 m)
Medium of assessment administrationNot clearComputer (Internet) self-completionFace-to-face interviewFace-to-face interviewComputer (Internet) self-completionPen and paper self-completionComputer (Internet) self-completionFace-to-face interview
Non-alcohol content in assessmentNonePhysical activity, fruit, vegetable consumption, smoking, mental health (from SF-36)Drug use and sexual behaviourInjury Severity Scale, presenting conditions, Quality of Life (SF-12)NoneNoneNoneAbbreviated Risk Taking/Impulsivity and Sensation Seeking Scales
Other features of assessmentCollateral interviews were conductedConsent given for saliva sampleBreath testing for assessment group; List of AAgroups and treatment services given to all
Reimbursement/PaymentNot statedAll participants given a pen (value NZ$0.50) at invitation to follow-upPaid US$20 and US$25 for 6 m and 12 m follow-up assessments.NoneParticipants given sandwich voucher (NZ$4.95) when invited for follow-upPaid £10 upon successful follow-up.Psychology course credit or US$20 at baseline; US $20 at 3 and 6 mo. for assessment group only, US$40 at 12 mo. for allNot stated

Abbreviations : MAST = Michigan Alcohol Screening Test; AUDIT = Alcohol Use Disorders Identification Test; APS = Alcohol Problems Scale; AREAS = Academic Role Expectations and Alcohol Scale; BAC = blood alcohol concentration; TLFB = Time Line Follow Back; RAPI = Rutgers Alcohol Problem Index; SF-12 = Short Form-12; SF-36 = Short Form-36; SIP = Short Index of Problems; RAPS-4 = Rapid Alcohol Problems Screen (4items).

Table 3

Study outcomes.

Richmond et al. 1995 Kypri & McAnally 2005 Carey et al 2006 Daeppen et al. 2007 Kypri et al. 2007 McCambridge & Day 2008 Walters et al. 2009 Cherpitel et al. 2010
Numbers analysedControl, Assessment6 m: 72, 66Control, Assessment6 weeks: 61, 65Control, Assessment1 m: 79, 886 m: 66, 6912 m: 59, 72Control, Assessment12 m: 257, 277Control, Assessment6 m: 124, 12212 m: 126, 126Control, Assessment2–3 m: 144, 156Control, Assessment12 m: 66, 63Control, Assessment12 m: 91, 99
Outcome measuresDrinks last 7 days• % Binge drinkers• Peak estimated BAC• Drinks/week• Drinks/drinking day• Heavy drinking days• Peak estimated BAC previous month• RAPI• Days drinking/week• Drinks/ drinking day• Number of binge drinking occasions• Drinks last 7 days• AUDIT• SF-12 Physical Comp. Score• SF-12 Mental Comp. Score• Drinking days last 14 days• Drinks/drinking day• Drinks last 14 days• Heavy episodes last 14 days• APS• AREAS• AUDIT score (only at 12 m)• AUDIT score• Drinks last 7 days• Drinking days last month• APS• AREAS• LDQ• % AUDIT>7• % 10+ drinks past 7 days• % Exceeded recommended weekly limit• AUDIT score• % <8 on AUDIT• Drinks per week• Peak estimated BAC in previous month• Protective Behaviors Score including 4 subscale scores and 15 item scores• RAPS-4• At-risk drinking (defined as for eligibility)• Drinking days per week• Drinks per drinking day• Maximum drinks per occasion• SIPS• Sought alcohol treatment
Duration of follow-up6 m6 weeks1 m, 6 m, 12 m12 m6 m, 12 m2–3 m12 m12 m
Summary of reported findings* 0/1 statistically significant differences0/2 statistically significant differences3/15 statistically significant differences:1 m• Drinks/wk• Drinks/drinking day• Peak BAC0/6 statistically significant differences4/13 statistically significant differences:12 m• Drinks last 14 days• Heavy episodes last 14 days• APS score• AUDIT4/9 statistically significant differences:• AUDIT score• LDQ score• % AUDIT>7• % 10+ drinks past 7 days7/25 statistically significant differences:• % <8 on AUDIT• Peak BAC• Protective Behaviors Score• Mixing beverages more weakly• Putting more ice in drink• Avoiding drinking games• Drinking slowly0/7 statistically significant differences

*In all cases, statistical significance defined as p<0.05.

Abbreviations: APS = Alcohol Problems Scale; AREAS = Academic Role Expectations and Alcohol Scale; AUDIT = Alcohol Use Disorders Identification Test; BAC = blood alcohol concentration; LDQ = Leeds Dependency Questionnaire; RAPI = Rutgers Alcohol Problem Index; SF-12 = Short Form-12; SF-36 = Short Form-36; SIP = Short Index of Problems; RAPS-4 = Rapid Alcohol Problems Screen (4items).

Abbreviations : MAST = Michigan Alcohol Screening Test; AUDIT = Alcohol Use Disorders Identification Test; APS = Alcohol Problems Scale; AREAS = Academic Role Expectations and Alcohol Scale; BAC = blood alcohol concentration; TLFB = Time Line Follow Back; RAPI = Rutgers Alcohol Problem Index; SF-12 = Short Form-12; SF-36 = Short Form-36; SIP = Short Index of Problems; RAPS-4 = Rapid Alcohol Problems Screen (4items). *In all cases, statistical significance defined as p<0.05. Abbreviations: APS = Alcohol Problems Scale; AREAS = Academic Role Expectations and Alcohol Scale; AUDIT = Alcohol Use Disorders Identification Test; BAC = blood alcohol concentration; LDQ = Leeds Dependency Questionnaire; RAPI = Rutgers Alcohol Problem Index; SF-12 = Short Form-12; SF-36 = Short Form-36; SIP = Short Index of Problems; RAPS-4 = Rapid Alcohol Problems Screen (4items). Meta-analytic findings are presented in Figures 2, 3, 4. For past week alcohol consumption (Figure 2), the pooled effect marginally exceeds the 5% probability threshold (z = 1.94, p = 0.053) and is equivalent to approximately 1.5 UK units, and just over 1 standard drink in the USA [30]. No statistical heterogeneity is observed in relation to this effect. These studies are, however, clinically heterogeneous. Five studies took place in university student populations with pro-active recruitment of volunteers [31], [32], [36], [38], [42] and three among adults attending clinical services [35], [37], [43]. All but one [36] of the former reported effects of brief interventions on alcohol consumption, whereas none of the latter did. This lack of effectiveness in these latter studies makes them somewhat unusual in the literature on brief interventions in primary care where effectiveness has been established [22]. In the absence of differences between randomised brief intervention and unassessed control groups, it would be surprising if there were differences between assessed and unassessed control groups, the comparison of interest here. When this analysis is restricted to the five studies undertaken with university students, the pooled effect is greater and statistically significant (21.8 grams [4.4 to 39.2] difference, z = 2.46, p = 0.014). These findings contrast with the no difference found in quantity per drinking day (z = 0.14, p = 0.89) for which there was evidence of statistical heterogeneity (Figure 3).
Figure 2

Meta-analysis of the effects of answering questions on total weekly drinking.

Figure 3

Meta-analysis of the effects of answering questions on quantity per drinking day.

Figure 4

Meta-analysis of the effects of answering questions on total AUDIT scores.

The studies by Daeppen and colleagues [37] and Cherpitel and colleagues [43] both took place in Emergency Departments where evidence of brief intervention effectiveness is more uncertain than in general practice [44]. The Daeppen study also differs clearly from other studies in Figure 4, introducing heterogeneity and reducing the pooled estimate of effect from approximately 1.5 to 1 point (z = 2.21, p = 0.027). Twelve month AUDIT data from the study by Kypri and colleagues [32] were used because this outcome was not assessed at the 6-month follow-up interval. The three studies with the largest estimated effects on past week drinking all involved alcohol-only assessments [32], [35], [38]. The trial by Anderson and Scott reported findings separately for men and women [40], [41]. Both published reports of this trial contained the statement that “there were no significant differences between the control group who received no assessment and the group who received assessment” and provided no further outcome data. Statistical power to detect differences was limited. If it is assumed that retention was equivalent in the non-assessed group and the assessed group, estimated total sample sizes at follow-up were 46 women and 109 men. Among men this provides approximately only 18% power to detect a small effect of 0.2 standard deviations. The trial by Gentilello and colleagues [39] did not refer to relevant outcomes in the published report. E-mail contact with the lead author ascertained that analyses had been undertaken and no differences in outcome detected between assessed and non-assessed groups (L. Gentillelo, personal communication). Again assuming that attrition was not different between the randomised groups, this sample of 307 provides approximately 41% power to detect a difference of 0.2 standard deviations.

Discussion

Ten trials of the effects of answering questions in research assessment procedures within brief alcohol intervention studies were identified. Outcome data were pooled on three specific measures of drinking behaviour drawn from eight trials with data available. This revealed somewhat equivocal evidence of small effects on two of the three outcomes across the studies as a whole. Evidence of assessment reactivity appears stronger if one restricts attention to the student literature. The possible effect on past weekly total consumption was not detected to a statistically significant level in any of the eight primary studies. This pattern of findings is behaviourally plausible where reduced overall alcohol consumption is caused by less frequent drinking with a consequent reduction in risk as assessed with the AUDIT, with quantity per drinking occasion remaining unaltered. Five of the eight studies included in the meta-analysis were undertaken in healthcare settings and five of the studies involved university students, with two studies taking place in student health services [32], [36]. The generalisability of these findings thus warrants careful scrutiny. This point is also reinforced by the detection of the two unpublished studies in general medical settings with no statistically significant effects, which incidentally gives reason for confidence in the completeness of our identification methods. Whilst the study by Anderson and Scott [40], [41] had very limited capacity to detect effects, this was much less true of the study by Gentilello and colleagues [39]. The methodological quality of included studies has not been formally assessed, making caution further necessary, as biases in trials will produce biased pooled effect estimates in meta-analyses. For example, though attrition is generally low, it is higher in some studies than in others and even small differences between groups across studies may introduce bias. We chose a single follow-up interval in the studies by Carey et al. [31] and Kypri et al. [32] and eschewed evaluation of binary AUDIT outcome which was statistically significant in the study by Walters and colleagues [42] in favour of continuous AUDIT score which was not. Close inspection of the data in the tables and figures suggests that both these decisions lead towards more conservative estimates of the effects of answering questions. We also deliberately ignored statistically significant effects within the primary studies on outcomes which have not been employed consistently across the studies. Findings of a small effect on drinking behaviour are coherent with data on various other outcomes in the relatively few individual trials that exist in the wider health sciences [13], [14], [15], [16], [17], [18], [20], [21]. These too have generally identified small effects, though they include one study which found no effects [20], and also one study which identified a large effect [19]. This pattern is found also in the wider non-health social science literature. For example prior questionnaire completion exerts a measurable small effect on voting behaviour [45], [46], as well as in laboratory-based social psychology experiments [47]. The outcome data in the present study were all self-reported, necessarily so given the target behaviour, and other investigations of this phenomenon also rely on such data [13], [14], [15], [18]. Importantly, however, studies do exist which identify effects of similar magnitude upon objectively assessed behaviours [16], [17], [21]. For example, the large effect obtained in a dental study was on plaque coverage ascertained using photography [19]. Furthermore, Godin and colleagues [17] observed both registrations at blood drives and blood donations, neither requiring self-report data, and effects have also been detected on attendances for screening in other studies [16], [21]. Intriguingly, a large effect on the amount of money deposited in an honesty box was also unobtrusively obtained in an experiment stimulating a sense of being observed [48]. Given the under-development of this area of study, there are many potential sources of bias which remain to be investigated to permit clear and confident causal inferences. Notwithstanding these cautionary remarks, what are the implications of the findings from the present study? The magnitude of an apparent assessment effect quantified here is small, though that should not be interpreted to mean that it is unimportant. The difference in past week consumption represents approximately 35% of the known effect of brief alcohol intervention in primary care [22]. As assessment is an integral component of a brief intervention, contamination has occurred, attenuating estimates of intervention effects. Brief interventions may thus be much more effective than has been previously understood. The present study needs to be replicated when the literature has further developed and the equivocal nature of the overall findings and the apparent discrepancies between student and non-student populations resolved. There is also a need to study whether answering questions on other behaviours generates similar reactivity in trials. The present findings suggest this form of contamination may be more common than has previously been appreciated. The small effects potentially attributable here to answering questions have been detected as unwanted artefacts of the research process. Almost all these questions are concerned with measurement of behaviour and its consequences. These questions have thus not been designed to elicit thinking about change, and thus to promote actual behaviour change. This is true also of the wider literature with the exception of the studies by Sandberg and Conner [21] and Godin and colleagues [17], in which questions specifically about anticipated regret and implementation intentions were asked. It is likely that selecting questions for their behaviour change potential may produce greater effects than have been seen here. Answering questions appears to exert a subtle influence on subsequent self-reported drinking behaviour among students. Other aspects of the research process such as randomisation [49], [50], [51] and consent [52], [53] also have psychological impacts. Their implications for subsequent behaviour remain to be evaluated [54], and may cumulatively generate greater bias. This impairs our ability to rule out reactivity to the research conditions themselves as a possible explanation for observed between-group differences in trials, thereby impeding secure inferences on the true effects of behavioural interventions [55]. These uncertainties are ironically produced by the unintended and largely overlooked consequences of undertaking research itself. Whilst behavioural science has had some awareness of these issues for some considerable time, rapid advances in understanding are now well overdue [56].
  46 in total

1.  Quantifying heterogeneity in a meta-analysis.

Authors:  Julian P T Higgins; Simon G Thompson
Journal:  Stat Med       Date:  2002-06-15       Impact factor: 2.373

2.  Calculating standard drink units: international comparisons.

Authors:  W R Miller; N Heather; W Hall
Journal:  Br J Addict       Date:  1991-01

3.  Should we ask our Children about Sex, Drugs and Rock & Roll?: Potentially Harmful Effects of Asking Questions About Risky Behaviors.

Authors:  Gavan J Fitzsimons; Sarah G Moore
Journal:  J Consum Psychol       Date:  2008-04-01

4.  Patient preferences and randomised clinical trials.

Authors:  C R Brewin; C Bradley
Journal:  BMJ       Date:  1989-07-29

5.  Cues of being watched enhance cooperation in a real-world setting.

Authors:  Melissa Bateson; Daniel Nettle; Gilbert Roberts
Journal:  Biol Lett       Date:  2006-09-22       Impact factor: 3.703

6.  Impact of the Hawthorne effect in a longitudinal clinical study: the case of anesthesia.

Authors:  D De Amici; C Klersy; F Ramajoli; L Brustia; P Politi
Journal:  Control Clin Trials       Date:  2000-04

7.  Alcohol treatment research assessment exposure subject reactivity effects: part I. Alcohol use and related consequences.

Authors:  Patrick R Clifford; Stephen A Maisto; Christine M Davis
Journal:  J Stud Alcohol Drugs       Date:  2007-07       Impact factor: 2.582

8.  Randomized controlled trial of the effects of completing the Alcohol Use Disorders Identification Test questionnaire on self-reported hazardous drinking.

Authors:  Jim McCambridge; Maria Day
Journal:  Addiction       Date:  2008-02       Impact factor: 6.526

9.  Effects of Study Design and Allocation on participant behaviour--ESDA: study protocol for a randomized controlled trial.

Authors:  Kypros Kypri; Jim McCambridge; Amanda Wilson; John Attia; Paschal Sheeran; Steve Bowe; Tina Vater
Journal:  Trials       Date:  2011-02-14       Impact factor: 2.279

10.  Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement.

Authors:  David Moher; Alessandro Liberati; Jennifer Tetzlaff; Douglas G Altman
Journal:  BMJ       Date:  2009-07-21
View more
  120 in total

1.  Assessment effects in educational and psychosocial intervention trials: an important but often-overlooked problem.

Authors:  Mi-Kyung Song; Sandra E Ward
Journal:  Res Nurs Health       Date:  2015-03-01       Impact factor: 2.228

2.  Ecological momentary assessment of daily discrimination experiences and nicotine, alcohol, and drug use among sexual and gender minority individuals.

Authors:  Nicholas A Livingston; Annesa Flentje; Nicholas C Heck; Allen Szalda-Petree; Bryan N Cochran
Journal:  J Consult Clin Psychol       Date:  2017-12

3.  Short-term risk of suicide attempt associated with patterns of patient-reported alcohol use determined by routine AUDIT-C among adults receiving mental healthcare.

Authors:  Julie E Richards; Susan M Shortreed; Greg E Simon; Robert B Penfold; Joseph E Glass; Rebecca Ziebell; Emily C Williams
Journal:  Gen Hosp Psychiatry       Date:  2019-12-18       Impact factor: 3.238

4.  An Exploratory RCT to Support Gamblers' Intentions to Stick to Monetary Limits: A Brief Intervention Using Action and Coping Planning.

Authors:  Simone N Rodda; Kathleen L Bagot; Victoria Manning; Dan I Lubman
Journal:  J Gambl Stud       Date:  2020-03

5.  Motivational Enhancement System for Adherence (MESA): pilot randomized trial of a brief computer-delivered prevention intervention for youth initiating antiretroviral treatment.

Authors:  Sylvie Naar-King; Angulique Y Outlaw; Moussa Sarr; Jeffrey T Parsons; Marvin Belzer; Karen Macdonell; Mary Tanney; Steven J Ondersma
Journal:  J Pediatr Psychol       Date:  2013-01-28

6.  A multifaceted intervention to improve mental health literacy in students of a multicampus university: a cluster randomised trial.

Authors:  Nicola J Reavley; Terence V McCann; Stefan Cvetkovski; Anthony F Jorm
Journal:  Soc Psychiatry Psychiatr Epidemiol       Date:  2014-05-06       Impact factor: 4.328

7.  Cannabis and Young Users-A Brief Intervention to Reduce Their Consumption (CANABIC): A Cluster Randomized Controlled Trial in Primary Care.

Authors:  Catherine Laporte; Hélène Vaillant-Roussel; Bruno Pereira; Olivier Blanc; Bénédicte Eschalier; Shérazade Kinouani; Georges Brousse; Pierre-Michel Llorca; Philippe Vorilhon
Journal:  Ann Fam Med       Date:  2017-03       Impact factor: 5.166

8.  Computerized versus in-person brief intervention for drug misuse: a randomized clinical trial.

Authors:  Robert P Schwartz; Jan Gryczynski; Shannon Gwin Mitchell; Arturo Gonzales; Ana Moseley; Thomas R Peterson; Steven J Ondersma; Kevin E O'Grady
Journal:  Addiction       Date:  2014-03-10       Impact factor: 6.526

9.  Reduction in Drinking was Associated With Improved Clinical Outcomes in Women With HIV Infection and Unhealthy Alcohol Use: Results From a Randomized Clinical Trial of Oral Naltrexone Versus Placebo.

Authors:  Robert L Cook; Zhi Zhou; Maria Jose Miguez; Clery Quiros; Luis Espinoza; John E Lewis; Babette Brumback; Kendall Bryant
Journal:  Alcohol Clin Exp Res       Date:  2019-07-10       Impact factor: 3.455

10.  The Impact of Asking About Interest in Free Nicotine Patches on Smoker's Stated Intent to Change: Real Effect or Artefact of Question Ordering?

Authors:  John A Cunningham; Vladyslav Kushnir; Jim McCambridge
Journal:  Nicotine Tob Res       Date:  2015-08-09       Impact factor: 4.244

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.