Literature DB >> 25537782

A usability study of two formats of a shortened systematic review for clinicians.

Laure Perrier1, M Ryan Kealey2, Sharon E Straus3.   

Abstract

OBJECTIVE: The aim of this study was to evaluate the usability of two formats of a shortened systematic review for clinicians.
MATERIALS AND METHODS: Usability of the prototypes was assessed using three cycles of iterative testing. 10 participants were asked to complete tasks of locating information or items within two prototypes and 'think aloud' while being audio taped. Interviews were also audio recorded and participants completed a systematic usability scale.
RESULTS: Revisions were made between each iteration in order to address issues identified by participants. Finding information relating to the number of studies in the meta-analysis, and locating the number of studies in the entire systematic review were revealed as areas needing attention during the usability evaluation.
CONCLUSIONS: Iterative testing combined with a multifaceted approach to usability testing offered essential insight into aspects of the prototypes that required modifications. Alterations were made in order to create finalised versions of the two shortened systematic review formats. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.

Entities:  

Keywords:  Evidence Based Practice; PRIMARY CARE; Review Literature as Topic

Mesh:

Year:  2014        PMID: 25537782      PMCID: PMC4275680          DOI: 10.1136/bmjopen-2014-005919

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


Two templates for a shortened format of a systematic review have been rigorously developed with input from end users. Errors were made by participants during each of the three iterative cycles of usability testing (cycle 1: five errors; cycle 2: eight errors; cycle 3: six errors) and highlighted areas of the template that required refinement, such as information on treatments that had no trials. A limitation is that the interplay between trait characteristics and end-user experiences could not be examined and may be revealed through the examination of personal (eg, age) or experiential factors which would be important considerations for future studies with larger populations.

Background

Systematic reviews are rigorous, comprehensive assessments of evidence that provide a synthesis of the literature.1 Many consider systematic reviews of high quality randomised controlled trials to rank as the highest in the hierarchy of evidence for interventions2 and the best source of information for making clinical decisions.3 Although systematic reviews are readily available in the published literature and through collected repositories such as the Cochrane Library, there is evidence that their use in clinical decision-making is not optimal.3 4 Searching, identifying and retrieving evidence-based resources paired with the lack of time experienced by busy clinicians is consistently identified as an obstacle to answering clinical questions.5–10 One solution to this obstacle is to create resources that use explicitly formulated methodological criteria so that information can be validated and refined in order to be read quickly.9 Numerous tools are available to clinicians that present summarised evidence-based information, either from a collection of sources (eg, UpToDate) or solely from systematic reviews and presented in a condensed format (eg, BMJ PICO abridged research articles). Prior to the development of our prototypes, we completed a systematic review examining the impact of interventions for seeking, appraising and applying evidence from systematic reviews in decision-making by clinicians.11 During the assessment of records, we specifically screened for studies that evaluated a different strategy for presenting a systematic review. Two trials were identified that reported promising results.12 13 However, on closer examination, their considerable limitations indicate that the results should be interpreted with caution due to their recruitment of a convenience sample affiliated with the Cochrane Collaboration and their use of a small sample size.12 13 Consequently, no studies providing a rigorous evaluation of the impact on end users were found in the literature. In an effort to address this, two shortened systematic review formats were developed to enhance their use by clinicians.14 15 The next step in the development process is to conduct a usability study with the prototypes so that they can be further refined and the design can be finalised.

Objective

To describe the usability testing of prototypes for two formats (case-based and evidence-expertise) that represent a shortened version of a full-length systematic review.

Methods

The approach to developing two alternate formats of a shortened systematic review to modify the presentation of information has been described in previous publications.14 15 In brief, prototypes for two formats of a shortened systematic review were developed using an explicit, rigorous process including a mapping exercise, an initial heuristic evaluation and a clinical content review.14 Following this, a series of focus groups were conducted with primary care physicians to further refine and determine the optimal format for the presentation of information.15 The focus groups provided a forum for clinicians to identify the essential components of a format for a shortened systematic review, including key features and content to aid in clinical decision-making.15 Support in the literature was found for the development of two shortened formats. The first format includes a case study to present contextualised information (case-based format), and the second format integrates evidence and clinical expertise (evidence-expertise format). The case-based format was designed to present a real-world example of how the clinical evidence would be used in decision-making. Text is easier to understand when it contains personalised elements, such as examples like case studies.16–20 This allows instructions and information to be embedded more succinctly and prompts readers to recall more information.19–21 The evidence-expertise format, guided by David Sackett's definition of evidence-based medicine, highlights the integration of clinical expertise and the best external evidence.22 An assessment of over one thousand systematic reviews showed that almost half presented results that neither supported nor refuted the interventions tested.23 In describing preferences for the presentation of evidence, primary care physicians expressed the need for an explicit statement when evidence was absent and the need for clinical expertise to bridge the gap.24 We chose a full-length systematic review to be used for the development of prototypes from a list of recently published systematic reviews supplied by the Health Information Unit at McMaster University (Canada).25 The systematic review had been rated by a pool of clinicians using the McMaster PLUS scale that allows them to identify articles they believe would be important to practising primary care physicians. Those that scored six or better (of seven) were reviewed by two physicians (one internal medicine physician and one family physician) who independently voted on the three most relevant to generalist physicians. The final review was selected by a third family physician independently. The final full-length systematic review that the prototypes were derived from for this study was van Zuuren et al.26 The final phase in the development process is to conduct a usability study using iterative cycles of testing with primary care physicians. Usability testing focuses on how well users can learn and use a product to achieve their goals and is defined as “how effectively, efficiently, and satisfactorily a user can interact with a user interface”.27 It does not test the comprehension of the content but rather provides direct information about how people use a tool and what their exact problems are with the tool being tested.

Usability testing

Participants

Primary care physicians practising full-time or part-time who were able to read and speak English were identified as eligible for participation. This group was chosen as systematic reviews summarise vast quantities of information on specialised topics which can be useful for generalist physicians. The sampling strategy for usability testing involved randomly selecting physicians practising in the Toronto, Canada area from a list available from the CPSO (College of Physicians and Surgeons of Ontario) database (representing 13 298 active family physicians and 3520 general internists). We also used snowball sampling28 which relies on referrals from the initial subjects to identify additional participants. Physicians were emailed and asked to reply indicating the time and date they were available to participate. Three iterative cycles of usability testing were completed and physicians could participate in one cycle of testing only.

Study design and procedures

Participants were assured of confidentiality in reporting the results. An honorarium was provided to participants. The usability testing of the two prototypes was run in three iterative cycles with 2–5 participants per iteration. This approach is a process of implementing a design or tool, seeking discussion and feedback and making subsequent refinements to the design or tool.29 Multiple testing is supported as the goal is to improve design, not just document weaknesses and sampling as little as five users has been shown to uncover substantial problems.30 After consent was obtained, participants were given instructions about the usability test. The participants were given the choice of viewing and testing the prototypes on a laptop brought by the investigator or on their own computer. Participants were given case scenarios along with the modified systematic reviews and asked to complete a task relevant to the scenario. Semistructured interviews were conducted and observations of their interaction with the reviews were recorded (LP). Participants were asked to ‘think aloud’ as they completed the task and these comments were audio recorded. Both prototypes were presented to each physician and a random sequence was generated by a computer technician using MySQL’s rand() function31 so that the order in which the prototypes were presented were randomised. At the end of this session, participants were asked a series of questions using a semistructured interview guide to learn about user satisfaction and to ask for suggestions on improving the document (see online supplementary appendix A). The System Usability Scale (SUS)32 33 was completed by the final eight participants (see online supplementary appendix B). This is a reliable 10-item questionnaire with five response options (from ‘strongly agree’ to ‘strongly disagree’) for respondents. This was introduced in the second and third iteration of testing after the tool had been modified. Field notes were also taken during the session (LP). Content analysis was conducted on the audio-recorded interviews and field notes. After each set of participants, changes were made to the prototypes based on problems identified from the data analysis. The prototypes were modified by a graphic designer. The revised prototypes were then used in the next iteration of usability testing and subsequent refinements were made.

Measures

Case scenarios: The case scenarios were constructed with a human factors engineer based on user tasks such as locating the original source for the document and identifying the number of studies in the systematic review. Although the two prototypes are distinct, essential information appears in the same position for both documents. For example, the table displaying the main results and meta-analysis information is positioned in the same location in the document for the case-based and the evidence-expertise formats. The majority of tasks (13 of 15 case scenarios) given to participants were relevant to prototypes and required users to locate information in the same place for either prototype. A total of two case scenarios were used for only one of the prototypes (one task was unique to the case-based prototype; one task was unique to the evidence-expertise prototype). An example of a case scenario (relevant to both prototypes) is as follows, A patient comes in with a clipping from a magazine that suggests tetracycline is the best treatment for more extensive skin lesions on the central part of their face. Can you locate information that would address their question? [Participants were prompted to think aloud and state when they had completed a task]. Since the order in which the two prototypes were presented to participants was randomised and the majority of questions (13 of 15) were identical, no attempt was made to report which prototype had more errors as viewing the first prototype provided insight as to the types of items participants would be asked to locate for the second prototype. Semistructured interview: The semistructured interview was used to provide further insight into the usability of the two prototypes. The questions were designed to move from general to specific information. The interviewer (LP) used probes to gain more specific information from the participants about the prototypes, for example, “can you give me an example?” The questions asked include the participants’ overall impressions of the document, what was liked best, what was liked least and items that should be added. The SUS is an industry standard that allows the evaluation of a wide variety of products and services, including hardware, software, mobile devices, websites and applications.33 The wording in this scale is geared towards applications and was altered by replacing the word ‘system’ with the word ‘document’. A score was calculated from the answers given. A SUS score above 68 would be considered above average and below 68 is below average in the evaluation of the usability of a product.33

Data analysis

The comments from participants during the tasks and the semistructured interviews were transcribed verbatim. Field notes were transcribed and included in the analytic process. Content analysis was done after each iterative cycle. Field notes and transcripts were reviewed and coded independently by two coders (LP, MRK) using a set of codes generated by the interviewer (LP) by initially reviewing terms of usability problems identified by medical information tools.34 A meeting with the second coder was used to identify discrepancies in coding. The number of case scenarios completed was tracked quantitatively.

Rigour and quality

Strategies identified in Lincoln and Guba's framework were used to enhance rigour and quality.35 Probing questions were used during the interviews and prompts were used during the case scenarios to increase the understanding of participants’ meaning.36 To increase validity, two investigators analysed and coded the transcripts independently, then discussed discrepancies until agreement was reached. Also, different data sources were used, for example, observations and interviews, plus all procedures were documented in order to create an audit trail.37 This process of triangulation ensured that discoveries and findings emerged from the data through consensus among investigators. Finally, the interviewer had no relationship with any of the participants.

Results

Ten participants took part in three iterative cycles of usability testing. One hundred fifty two recruitment emails were sent to potential participants. Six physicians were recruited with this method and four through snowball sampling, giving a response rate of 7%.

Demographics

The sample (n=10) included five women and five men. Nine participants were aged between 30 and 65 years and one was less than 30 years old (table 1). All participants were family physicians with the majority (90%) having five or more years of experience. Seven of 10 physicians practised in a private clinic. Four participants identified their practice as being in an inner city setting and six indicated that their practice was in an urban/suburban setting.
Table 1

Demographics and work profile of study participants

Cycle 1–3 participants: N=10
Age (years)
 <301
 30–391
 40–494
 50–593
 60–651
 >65
Sex
 Women5
 Men5
Area of practice
 Family medicine10
 General internal medicine
Years in practice
 <51
 5–101
 11–152
 16–252
 >254
Work setting*
  Private office/clinic (excluding free-standing walk-in clinics)7
 Community clinic/Community health centre
 Free-standing walk-in clinic
 Academic health sciences centre4
 Community hospital
 Nursing home/home for aged
 University/faculty of medicine1
 Administrative office
 Research unit
 Free-standing laboratory/diagnostic clinic
 Other (1 rehabilitation centre; 2 emergency room)2
Practice population
 Inner city4
 Urban/suburban6
 Small town
 Rural
 Geographically isolated/remote
 Other

*Participants listed all settings where they worked (up to three settings per person).

Demographics and work profile of study participants *Participants listed all settings where they worked (up to three settings per person).

Usability of the prototypes

Sessions took between 50 and 60 min to complete the case scenarios, interview and SUS. Two participants took part in the first iteration of testing in July 2013 with the original prototypes (see online supplementary appendix C). Recommendations were generated from this analysis and given to the graphic designer who modified the prototypes (see online supplementary appendix D). These recommendations focused on making the information that related to the meta-analysis, the clinical summary and the conditions with no evidence more prominent. These revised prototypes were used in the second iteration of testing which were held in August and September 2013. After three participants finished usability testing, two investigators (LP, MRK) independently coded these transcripts. Minor recommendations were given to the graphic designer and these modifications were implemented. These modifications concentrated on making the explanation for the evidence rating scale distinct and changing the header for the conditions with no trials (see online supplementary appendix E). The modified prototypes were taken into the third iteration of usability testing in January 2014. Five participants took part in usability testing and their transcripts were independently coded by two investigators (LP, MRK) who came to agreement that no new information was being learnt from participants that required major revisions to the prototypes. No further participants were recruited after this set of data was analysed. The completion of tasks based on the case scenarios were tracked quantitatively (table 2). Changes were made after the first iteration of usability testing based on this information (table 3). Locating the information for the meta-analysis was a challenge for all participants for all three sets of testing. After the first iteration, the graphic designer altered the prototypes in order to make the heading for the meta-analysis more prominent, and when the case scenario for this task was presented to participants in the second iteration of testing, it was emphasised that they were to look for information specific to the meta-analysis. Despite this, during the second iteration of usability testing, participants were not able to locate the meta-analysis information. Further modifications were made by the graphic designer; however, two of the final five participants in the third iteration of testing did not locate this information. Participants in the first iteration of usability testing could not locate the ‘take-home’ message. Modifications were made to make this information more distinct, and the wording of the case scenario given to participants was altered to clarify that the phrase ‘take-home’ message was a synonym for the word summary. This ceased to be an issue in the second and third iterations of testing and was completed by these participants without error. There were two case scenarios that were completed by all participants in the first iteration of testing; however, these same tasks were not completed by at least one participant in the second iteration of testing. One participant was not able to locate information about trials that had no evidence and attention was given to this by the graphic designer between the second and third iterations of testing. Two participants were not able to complete this task during the third iteration of testing; however, no further modifications were made as earlier work with focus groups15 had indicated that this information had lesser importance which limited graphics in their options for any further modifications, that is, there were no further ways to graphically display the information. Also, one participant could not locate information describing an odds ratio during the second iteration of testing. To address this, the three pieces of information offered along the bottom of the page (information describing the odds ratio, trials with no evidence, and the strength of evidence scale legend) were presented more distinctly as three separate units by using boxes and white space. Despite this, one participant stated that they could not complete the task asking them to locate the evidence rating scale legend in the third iteration of testing. No changes were made to the prototypes as on review of transcripts it was identified that the question may have been misinterpreted by the participant. The question was worded as, ‘Can you describe the evidence rating scale?’ (to prompt the participant to describe the scale as strong, moderate and weak). The participant's answer suggests that they may have thought that ‘evidence rating scale’ was a formal term as they stated, “I don't actually know what that term means…I think it means what kind of statistical analysis was done and I would imagine it has to do with this here, the odds ratio.”
Table 2

Case scenario completion

TaskNumber of tasks not completed
Cycle 1n=2Cycle 2n=3Cycle 3n=5
Information in same location for both prototypes:
 Number of studies contributing to meta-analysisTask given to participants: A colleague briefly describes the meta-analysis. Can you locate this information and describe how many studies it summarises?2 (100%)3 (100%)2 (40%)
 Locate evidence rating scaleTask given to participants: You want to present some of this information to others and need to confirm how the evidence rating scale works. Can you describe the evidence rating scale?1 (20%)
 Locate adverse event informationTask given to participants: You would like to prescribe rilmenidine and want to confirm any adverse events. Can you locate this information and describe it?1 (33%)
 State ‘take-home message’Task given to participants: You need to give a group of residents the clinical summary or a ‘take-home message’ from the document. What information would you share with them?2 (100%)
 Report explanation for odds ratioTask Given to Participants: You want to confirm the explanation for an odds ratio. Can you find this information and describe it?1 (33%)
 Treatment for ocular rosaceaTask given to participants: There are no trials for several treatments. Can you locate one treatment for ocular rosacea where there are no randomised controlled trials for evaluation and state it?1 (33%)2 (40%)
Information in different locations on prototypes:
 Number of studies contributing to reviewTask given to participants: You want a sense of how many studies contributed to the entire review. Locate this information and describe the number of studies1 (50%)2 (66%)1 (20%)
Total errors586
Table 3

Changes made to prototypes

Task
Number of studies contributing to meta-analysis
Task Given to Participants: A colleague briefly describes the meta-analysis. Can you locate this information and describe how many studies it summarizes?
Iteration/ Errors:
Cycle 1=2 errors – Change made: Heading moved from right-hand side to left-hand side
Cycle 2=3 errors – Change made: Text in heading bolded
Cycle 3=2 errors – Change made: No further changes
Task
Locate evidence rating scale
Task Given to Participants: You want to present some of this information to others and need to confirm how the evidence rating scale works. Can you describe the evidence rating scale?
Iteration/ Errors:
Cycle 3=1 error – Change made: No changes made - review of transcripts indicated misinterpretation of question
Task
State ‘take home message’ Task Given to Participants: You need to give a group of residents the clinical ‘take-home message’ from the document. What information would you share with them?
Iteration/ Errors:
Cycle 1=2 errors – Change made: Question re-worded to ‘summary or take-home message’
Task
Report explanation for odds ratio
Task Given to Participants: You want to confirm the explanation for an odds ratio. Can you find this information and describe it?
Iteration/ Errors:
Cycle 2=1 error – Change made: Information on bottom of document distinctly separated by adding boxes or white space
Task
Treatment for ocular rosacea
Task Given to Participants: There are no trials for several treatments. Can you locate one treatment for ocular rosacea where there are no randomized controlled trials for evaluation and state it?
Iteration/ Errors:
Cycle 2=1 error – Changes made: i) Text: colour changed from red to black; capitalized; made larger, and ii) Information on bottom of document distinctly separated by adding boxes or white space
Cycle 3=2 errors – Change made: No further changes
Task
Number of studies contributing to review Task Given to Participants: You want a sense of how many studies contributed to the entire review. Locate this information and describe the number of studies.
Iteration/ Errors:
Cycle 1=1 error – Change made: Efforts focused on making number of studies in the meta-analysis distinct
Cycle 2=2 errors – Change made: Efforts focused on making number of studies in the meta-analysis distinct
Cycle 3=1 error – Change made: No further changes
Case scenario completion Changes made to prototypes The SUS was administered in the second and third iterations of testing (table 4). An SUS score above 68 would be considered above average and anything under 68 is below average.33 The case-based shortened systematic review received an ‘above average’ score of 72.5 and 80 (2 of 3 participants) in the second iteration, and 87, 90 and 97.5 (3 of 5 participants) in the third iteration of testing. One participant gave it a score of 52.5 (below average) in the second iteration, and two people scored it 52.5 and 67.5 during the third iteration. All participants scored the evidence-expertise shortened systematic review ‘above average’ with scores of 77.5, 87.5 and 95 (second iteration) and 100, 77.5, 75, 70 and 70 (third iteration).
Table 4

System usability score: cycle 2 and cycle 3 participants

Participant numberSUS score
Cycle 2
 Evidence-expertise prototype395
477.5
587.5
 Case-based prototype352.5
480
572.5
Cycle 3
 Evidence-expertise prototype6100
777.5
870
975
1070
 Case-based prototype697.5
767.5
890
952.5
1087.5

Scores have a range of 0–100.

SUS, System Usability Scale.

System usability score: cycle 2 and cycle 3 participants Scores have a range of 0–100. SUS, System Usability Scale.

Discussion

The usability testing revealed features of the shortened systematic review prototypes that prompted modifications. Primary care physicians provided information on how the prototypes could be improved for more effective use. Changes that were made focused on making information prominent so that it could be found more readily. Two items that were consistently challenging to participants were locating the number of studies that contributed to the meta-analysis, and the number of studies that contributed to the entire systematic review. It is possible that physicians were not clear about the distinction between a systematic review and meta-analysis. Research indicates that physicians have low levels of knowledge with regard to research methodology and data analysis,38–41 and this may have contributed to making this case scenario challenging to complete. The SUS scores were above average in all instances with the exception of three participants and their scoring of the case-based shortened systematic review. The transcripts were reviewed for these participants to identify anything that required further attention. On initially viewing the case-based prototype, all three participants immediately declared a dislike for the case-based prototype. Examples of this included proactively stating a preference for the evidence-expertise version, repeating this preference more than once in the post-test interview, and stating that they would prefer to have it formatted exactly as the evidence-expertise prototype. Given this clear preference, it was decided not to make further alterations to the prototype based on these low SUS scores. Modifications to the prototypes were implemented based on the usability testing resulting in final versions for the two shortened formats of a systematic review. We plan to conduct a pilot study in order to assess the feasibility of a full-scale randomised controlled trial where participants will be asked to apply the evidence from either the full-length systematic review or one of the shortened formats to a patient in a clinical scenario.

Study limitations

The development process for the prototypes described in this paper needs to be considered within the context of certain limitations. The sample was small, which could be considered a threat to the generalisability of the results. However, it has been shown that with as few as five participants, the majority of usability problems and issues can be identified with a representative sample of end users.42 Also, the demographics of the sample provide some indication of a diverse group in that there is an equal representation of men and women, practising from less than 5 years to more than 25 years. Information on training in critical appraisal or experience in conducting systematic reviews was not collected from participants who may have indicated levels of expertise related to evidence-based information tools. However, these data would not have a great impact with regard to the usability testing as problems encountered with the use of the tool were being assessed rather than the content or comprehension. Estimates have shown that a single usability testing cycle of design/evaluation/redesign can lead to as much as a 10-fold reduction in usability problems.43 44 Moreover, we found that we were not obtaining new information with our interviews in the third iteration of testing and thus had reached saturation. Also, the sample was too small to allow the examination of whether personal (eg, age) or experiential factors were related to the use of the shortened systematic reviews. Future work may help determine how trait characteristics of users impact usability measures.

Conclusion

Iterative cycles of multifaceted usability testing provided insight into areas that needed to be refined for two formats (case-based and evidence-expertise) that represent a summary of a full-length systematic review. Usability testing included giving case scenarios to participants that involved them completing tasks of locating information or items within the prototypes, semistructured interviews and, for a portion of participants, completing a systematic usability scale. Changes were made by a graphic designer to the prototypes based on recommendations generated from the analysis of the usability testing. Alterations focused on making information more prominent so that information could be readily located by users. Conducting usability testing during the prototype phase with users provides the opportunity to address usability issues that may impact use of documents or tools. Incorporating their input before finalising the prototypes is done with the aim of increasing their usability and potential use in healthcare.
  27 in total

1.  Evaluation in the design of health information systems: application of approaches emerging from usability engineering.

Authors:  Andre Kushniruk
Journal:  Comput Biol Med       Date:  2002-05       Impact factor: 4.589

2.  Evaluation of ehealth systems and services.

Authors:  David H Gustafson; Jeremy C Wyatt
Journal:  BMJ       Date:  2004-05-15

3.  Answering physicians' clinical questions: obstacles and potential solutions.

Authors:  John W Ely; Jerome A Osheroff; M Lee Chambliss; Mark H Ebell; Marcy E Rosenbaum
Journal:  J Am Med Inform Assoc       Date:  2004-11-23       Impact factor: 4.497

4.  The influence of display and statistical factors on the interpretation of metaanalysis results by physicians.

Authors:  Parminder S Raina; Jamie C Brehaut; Robert W Platt; Terry P Klassen; David Moher; Phil St John; Dianne Bryant; Ray Viola; Ba Pham
Journal:  Med Care       Date:  2005-12       Impact factor: 2.983

5.  Technology induced error and usability: the relationship between usability problems and prescription errors when using a handheld application.

Authors:  Andre W Kushniruk; Marc M Triola; Elizabeth M Borycki; Ben Stein; Joseph L Kannry
Journal:  Int J Med Inform       Date:  2005-04-08       Impact factor: 4.046

Review 6.  How do primary care physicians seek answers to clinical questions? A literature review.

Authors:  Herma C H Coumou; Frans J Meijman
Journal:  J Med Libr Assoc       Date:  2006-01

7.  Users' guides to the medical literature. IX. A method for grading health care recommendations. Evidence-Based Medicine Working Group.

Authors:  G H Guyatt; D L Sackett; J C Sinclair; R Hayward; D J Cook; R J Cook
Journal:  JAMA       Date:  1995-12-13       Impact factor: 56.272

8.  Obstacles to answering doctors' questions about patient care with evidence: qualitative study.

Authors:  John W Ely; Jerome A Osheroff; Mark H Ebell; M Lee Chambliss; Daniel C Vinson; James J Stevermer; Eric A Pifer
Journal:  BMJ       Date:  2002-03-23

9.  Evidence based medicine: what it is and what it isn't.

Authors:  D L Sackett; W M Rosenberg; J A Gray; R B Haynes; W S Richardson
Journal:  BMJ       Date:  1996-01-13

10.  Information-seeking behaviors of practitioners in a primary care practice-based research network (PBRN).

Authors:  James E Andrews; Kevin A Pearce; Carol Ireson; Margaret M Love
Journal:  J Med Libr Assoc       Date:  2005-04
View more
  8 in total

1.  Visual Evidence: Increasing Usability of Systematic Reviews in Health Systems Guidelines Development.

Authors:  Connor J Smith; Rebecca M Jungbauer; Annette M Totten
Journal:  Appl Clin Inform       Date:  2019-10-02       Impact factor: 2.342

2.  Using a systematic review in clinical decision making: a pilot parallel, randomized controlled trial.

Authors:  Laure Perrier; Nav Persaud; Kevin E Thorpe; Sharon E Straus
Journal:  Implement Sci       Date:  2015-08-15       Impact factor: 7.327

Review 3.  Barriers and facilitators to uptake of systematic reviews by policy makers and health care managers: a scoping review.

Authors:  Andrea C Tricco; Roberta Cardoso; Sonia M Thomas; Sanober Motiwala; Shannon Sullivan; Michael R Kealey; Brenda Hemmelgarn; Mathieu Ouimet; Michael P Hillmer; Laure Perrier; Sasha Shepperd; Sharon E Straus
Journal:  Implement Sci       Date:  2016-01-12       Impact factor: 7.327

4.  Interactive Visual Displays for Interpreting the Results of Clinical Trials: Formative Evaluation With Case Vignettes.

Authors:  Jiantao Bian; Charlene Weir; Prasad Unni; Damian Borbolla; Thomas Reese; Yik-Ki Jacob Wan; Guilherme Del Fiol
Journal:  J Med Internet Res       Date:  2018-06-25       Impact factor: 5.428

Review 5.  Enhancing the uptake of systematic reviews of effects: what is the best format for health care managers and policy-makers? A mixed-methods study.

Authors:  Christine Marquez; Alekhya Mascarenhas Johnson; Sabrina Jassemi; Jamie Park; Julia E Moore; Caroline Blaine; Gertrude Bourdon; Mark Chignell; Moriah E Ellen; Jacques Fortin; Ian D Graham; Anne Hayes; Jemila Hamid; Brenda Hemmelgarn; Michael Hillmer; Bev Holmes; Jayna Holroyd-Leduc; Linda Hubert; Brian Hutton; Monika Kastner; John N Lavis; Karen Michell; David Moher; Mathieu Ouimet; Laure Perrier; Andrea Proctor; Thomas Noseworthy; Victoria Schuckel; Sharlene Stayberg; Marcello Tonelli; Andrea C Tricco; Sharon E Straus
Journal:  Implement Sci       Date:  2018-06-22       Impact factor: 7.327

6.  Redesigning printed educational materials for primary care physicians: design improvements increase usability.

Authors:  Agnes Grudniewicz; Onil Bhattacharyya; K Ann McKibbon; Sharon E Straus
Journal:  Implement Sci       Date:  2015-11-04       Impact factor: 7.327

7.  Development of consumer information leaflets for deprescribing in older hospital inpatients: a mixed-methods study.

Authors:  Natali Jokanovic; Parisa Aslani; Sophie Carter; Mai Duong; Danijela Gnjidic; Jesse Jansen; David Le Couteur; Sarah Hilmer
Journal:  BMJ Open       Date:  2019-12-11       Impact factor: 2.692

8.  Assessing how information is packaged in rapid reviews for policy-makers and other stakeholders: a cross-sectional study.

Authors:  Chantelle Garritty; Candyce Hamel; Mona Hersi; Claire Butler; Zarah Monfaredi; Adrienne Stevens; Barbara Nussbaumer-Streit; Wei Cheng; David Moher
Journal:  Health Res Policy Syst       Date:  2020-09-29
  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.