| Literature DB >> 32331530 |
Bharathy Kumaravel1, Jasmine Heath Hearn2, Leila Jahangiri3, Rachel Pollard4, Claire J Stocker5, David Nunan6.
Abstract
BACKGROUND: The importance of teaching the skills and practice of evidence-based medicine (EBM) for medical professionals has steadily grown in recent years. Alongside this growth is a need to evaluate the effectiveness of EBM curriculum as assessed by competency in the five 'A's': asking, acquiring, appraising, applying and assessing (impact and performance). EBM educators in medical education will benefit from a compendium of existing assessment tools for assessing EBM competencies in their settings. The purpose of this review is to provide a systematic review and taxonomy of validated tools that evaluate EBM teaching in medical education.Entities:
Keywords: Assessment; Competency; Evidence-based medicine; Medical education
Mesh:
Year: 2020 PMID: 32331530 PMCID: PMC7183115 DOI: 10.1186/s13643-020-01311-y
Source DB: PubMed Journal: Syst Rev ISSN: 2046-4053
Fig. 1PRISMA flowchart of the systematic review
Classification of tools against EBM steps evaluated
| Tool | EBM steps | ||||
|---|---|---|---|---|---|
| Ask | Acquire | Appraise | Apply | Assess | |
| Taylor’s questionnaire [ | Yes | Yes | |||
| Berlin [ | Yes | ||||
| Fresno [ | Yes | Yes | Yes | ||
| ACE [ | Yes | Yes | Yes | Yes | |
| Utrecht questionnaire U-CEP [ | Yes | Yes | Yes | ||
| MacRae examination [ | Yes | ||||
| EBM test [ | Yes | Yes | Yes | ||
| Educational prescription [ | Yes | Yes | Yes | Yes | |
| Mendiola-mcq [ | Yes | ||||
| Tudiver OSCE [ | Yes | Yes | Yes | ||
| Frohna’s OSCE [ | Yes | Yes | Yes | Yes | |
| BACES [ | Yes | ||||
Classification of tools against the seven educational outcome domains
| Outcome domains assessed by the twelve EBM instruments | |||||||
|---|---|---|---|---|---|---|---|
| Reaction to EBM teaching | Attitude | Self-efficacy | Knowledge | Skills | Behaviours | Patient benefit | |
| Taylor’s questionnaire | Yes | Yes | |||||
| Berlin | Yes | Yes | |||||
| Fresno | Yes | Yes | |||||
| ACE | Yes | Yes | |||||
| Utrecht questionnaire U-CEP | Yes | ||||||
| MacRae examination | Yes | Yes | |||||
| EBM test | Yes | Yes | |||||
| Educational prescription | Yes | Yes | |||||
| Mendiola | Yes | ||||||
| Tudiver OSCE | Yes | ||||||
| Frohna’s OSCE | Yes | ||||||
| BACES | Yes | Yes | |||||
High quality tools with ≥ 3 types of established validity
| Tool | Reported psychometric properties | |||||||
|---|---|---|---|---|---|---|---|---|
| Content validity | Interrater reliability | Internal validity | Responsive validity | Discriminative validity | Construct Validity | Internal reliability (ITC) | External validity | |
| Taylor’s questionnaire [ | Yes | Yes | Yes | Yes | ||||
| Berlin [ | Yes | Yes | Yes | Yes | ||||
| Fresno [ | Yes | Yes | Yes | Yes | ||||
| ACE [ | Yes | Yes | Yes | Yes | Yes | |||
| Utrecht questionnaire [ | Yes | Yes | Yes | Yes | Yes | Yes | Yes | |
| MacRae [ | Yes | Yes | Yes | Yes | Yes | |||
Details of studies where the high-quality tools (n = 6) were validated for use in evaluating EBM teaching in medical education
| Source instrument name and date | Instrument development-number of participants, level of expertise | EBM learning domains | Instrument description | EBM steps | Psychometric properties with results of validity and reliability assessment |
|---|---|---|---|---|---|
| Berlin questionnaire-Fritsche [ | 266 participants—43 experts in evidence-based medicine, 20 controls (medical students) and 203 participants in evidence-based medicine course (USA) | Knowledge and skills | Berlin questionnaire was developed to measure basic knowledge about interpreting evidence from healthcare research, skills to relate a clinical problem to a clinical question, the best design to answer it and the ability to use quantitative information from published research to solve specific patient problems. The questions were built around clinical scenarios and has two separate sets of 15 multiple-choice questions mainly focusing on epidemiological knowledge and skills (scores range from 0 to 15) | Appraise | Content validity Internal validity Responsive validity Discriminative validity The two sets of questionnaires were psychometrically equivalent: interclass correlation coefficient for students and experts 0.96 (95% confidence interval 0.92 to 0.98, |
| Fresno test-Ramos et al. [ | Family practice residents and faculty member ( | Knowledge and skills | Fresno test was developed and validated to assess medical professionals’ knowledge and skills. It consists of two clinical scenarios with 12 open-ended questions which are scored with standardised grading rubrics. Calculation skills were assessed by fill in the blank questions. | Ask, acquire and appraise | Content validity Interrater reliability Internal validity Discriminative validity Expert opinion Interrater correlations ranged from 0.76 to 0.98 for individual items Cronbach’s alpha was 0.88. ITC ranged 0.47–0.75. Item difficulties ranged from moderate (73%) to difficult (24%). Item discrimination ranged from 0.41 to 0.86. Construct validity, on the 212 point test, the novice mean was 95.6 and the expert mean was 147.5 ( |
| MacRae [ | Residents in University of Toronto General Surgery Program ( | Knowledge and skills | Examination consisted of three articles each followed by a series of short-answer questions and 7-point rating scales to assess study quality. | Appraise | Content validity Interrater reliability Internal validity Discriminative validity Construct validity Cronbach’s alpha 0.77 Interrater reliability—Pearson product moment correlation coefficient between clinical epidemiologist and non-epidemiologist-0.91 between clinical epidemiologist and nurse 0.78.Construct validity was assessed by comparing scores of those who attended the journal club versus those who did not and by postgraduate year of training ( |
Taylor [ Bradley et al. [ | 4 groups of healthcare professionals ( 2—undertaken CASP workshop within last 4 weeks; 3—undertaken CASP workshop in the last 12 months; 4—academics currently teaching EBP and attended 1997 Oxford CEBM workshop Later, Bradley et al. tried with 175 medical students in RCT of self-directed vs workshop-based EBP curricula (Norway) | Knowledge and attitudes | Questionnaire 11mcqs -true, false, do not know Correct responses given 1 Incorrect responses scored 1 Do not know 0 | Acquire and appraise | Content validity Internal validity Responsive validity Discriminative validity Cronbach’s alpha (0.72 for knowledge and 0.64 for attitude questions) Spearman’s correlation (internal consistency), total knowledge and attitudes scores ranged from 0.12 to 0.66, discriminative validity (novice and expert) Responsiveness (instrument able to detect change) |
| ACE tool- Dragan Ilic [ | 342 medical students—98 EBM-novice, 108 EBM-intermediate and 136 EBM-advanced participants (Australia) | Knowledge and skills | Assessing Competency in EBM (ACE )tool was developed and validated to evaluate medical trainees’ competency in EBM across knowledge, skills and attitudes—15 items, dochotomous outcome measure; items 1 and 2, asking the answerable question; items 3 and 4, searching literature; items 5–11 critical appraisal; items 12–15 relate to step 4 applying evidence to the patient scenario. | Ask, acquire, appraise and apply | Content validity Interrater reliability Internal validity Responsive validity Discriminative validity Construct validity—statistically significant linear trend for sequentially improved mean score corresponding to the level of training ( Item difficulty ranged from 36 to 84%, internal reliability ranged from 0.14 to 0.20, item discrimination ranged from 0.37 to 0.84, Cronbach’s alpha coefficient for internal consistency was 0.69 |
| Kortekaas-Utrecht questionnaire [ | Postgraduate GP trainees ( | Knowledge | Utrecht questionnaire on knowledge on clinical epidemiology (U-CEP): two sets of 25 questions and a combined set of 50 | Ask, appraise and apply | Content validity Internal validity Responsive validity Discriminative validity Content validity—expert opinion and survey Construct validity—significant difference in mean score between experts, trainees and supervisors Internal consistency—Cronbach alpha 0.79 for set A, 0.80 for set B and 0.89 for combined Responsive validity—significantly higher mean scores after EBM training than before EBM training Internal reliability—ITC using Pearson product, median 0.22 for set A, 0.26 for set B and 0.24 for combined Item Discrimination ability—median-0.35 for set A, 0.43 for set B and 0.37 for combined |
ITC item total correlation, RCT randomised controlled trial, CASP critical appraisal skills program, UCEP Utrecht questionnaire on knowledge on clinical epidemiology for evidence-based practice
Classification of the six high quality tools according to CREATE framework
| Assessment category | Type of assessment | Steps of EBM | |||||
|---|---|---|---|---|---|---|---|
| 7 | Benefits to patients | Patient-oriented outcomes | |||||
| 6 | Behaviours | Activity monitoring | |||||
| 5 | Skills | Performance assessment | Fresno ACE | Fresno ACE | Berlin’s Fresno ACE MacRae | ACE | |
| 4 | Knowledge | Cognitive testing | Fresno ACE U-CEP | Fresno ACE Taylor's | Taylor’s Berlins Fresno ACE UCEP MacRae | ACE UCEP | |
| 3 | Self-efficacy | Self-report/opinion | |||||
| 2 | Attitudes | Taylor's | Taylor’s | ||||
| 1 | Reaction to the educational experience | ||||||
Audience characteristic: students and trainees in medical education.
Assessment aims: formative
Details of studies which have used and validated six other tools identified as lower quality by this review for use in evaluating EBM teaching in medical education
| Source instrument name and date | Instrument development, number of participants, level of expertise | EBM learning domains | Instrument description | EBM steps | Psychometric properties with results of validity and reliability assessment |
|---|---|---|---|---|---|
| Educational Prescription-David Feldstein [ | 20 residents | Knowledge and skills | Educat academic GPs or clinical ional prescription (EP)—web-based tool that guides learners through the four As of EBM. Learners use the EP to define a clinical question, document a search strategy, appraise the evidence, report the results and apply evidence to the particular patient | Asking, acquiring, appraising, applying | Predictive validity Interrater reliability Interrater reliability on the 20 EPs showed fair agreement for question formation (k= 0.22); moderate agreement for overall competence ( |
| BACES-Barlow [ | Yes postgraduate medical trainees/residents—150 residents | Knowledge, skills | BACES-Biostatistics and Clinical Epidemiology Skills (BACES) assessment for medical residents-30 multiple-choice questions were written to focus on interpreting clinical epidemiological and statistical methods | Appraisal—interpreting clinical epidemiology and statistical methods | Content validity was assessed through a four person expert review Item Response Theory (IRT) makes it flexible to use subsets of questions for other cohorts of residents (novice, intermediate and advanced). 26 items fit into a two parameter logistic IRT model and correlated well with their comparable CTT (classical test theory) values |
| David Feldstein-EBM test [ | 48 internal medicine residents | Knowledge and skills | EBM test—25 mcqs-covering seven EBM focus areas: (a) asking clinical questions, (b) searching, (c) EBM resources, (d) critical appraisal of therapeutic and diagnostic evidence, (e) calculating ARR, NNT and RRR, (f) interpreting diagnostic test results and (g) interpreting confidence intervals | Asking, acquiring and appraising Asking clinical questions, searching, EBM resources, critical appraisal, calculations of ARR, NNT, RRR, interpreting diagnostic test results and interpreting confidence intervals. | Construct validity Responsive validity EBM experts scored significantly higher EBM test scores compared to PGY-1 residents ( |
| Frohna-OSCE [ | Medical students ( | Skills | A web-based 20-min OSCE-specific case scenario where students asked a structural clinical question, generated effective MEDLINE search terms and elected the most appropriate of 3 abstracts | Ask, acquire, appraise and apply | Face validity Interrater reliability Literature review and expert consensus Between three scorers, there was good interrater reliability with 84, 94 and 96% agreement ( |
| Tudiver-OSCE [ | Residents—first year and second year | Skills | Ask, acquire, appraise and apply | Content validity Construct validity Criterion validity Interrater reliability ICC 0.96 Internal reliability Cronbach’s alpha 0.58 | |
| Mendiola-mcq [ | Fifth year medical students | Knowledge | Appraise | Reliability of the mcq = Cronbach’s alpha 0.72 in M5 and 0.83 in M6 group Effect size in Cohen’s |
mcq multiple choice question, OSCE objective structured clinical examination, ICC intraclass correlation, NNT number needed to treat, ARR attributable risk ratio, RRR relative risk ratio
Assessment aims: formative