| Literature DB >> 29895284 |
André Strahl1,2, Christian Gerlich3, Georg W Alpers4, Katja Ehrmann3, Jörg Gehrke5, Annette Müller-Garnn5, Heiner Vogel3.
Abstract
BACKGROUND: The German quality assurance programme for evaluating work capacity is based on peer review that evaluates the quality of medical experts' reports. Low reliability is thought to be due to systematic differences among peers. For this purpose, we developed a curriculum for a standardized peer-training (SPT). This study investigates, whether the SPT increases the inter-rater reliability of social medical physicians participating in a cross-institutional peer review.Entities:
Keywords: Inter-rater reliability; Peer review; Quality assurance; Training curriculum; Work capacity evaluation
Mesh:
Year: 2018 PMID: 29895284 PMCID: PMC5998600 DOI: 10.1186/s12909-018-1233-z
Source DB: PubMed Journal: BMC Med Educ ISSN: 1472-6920 Impact factor: 2.463
Structure and learning objectives of peer-training
| Module | Time (min) | Educational objectives | Knowledge dimensions according to Bloom’s taxonomy | Methods |
|---|---|---|---|---|
| basic module | 90 | 1.1 The peer can describeb the basic principles of quality assurance in work capacity evaluation | Understanding | presentation, discussion, work in small groups |
| 60 | 1.2 The peer can explainb the content and structure of the peer review manual | Understanding | presentation, exercises | |
| 30 | 1.3 The peer can explainb the formal course of the peer review procedure in the routine application | Understanding | presentation | |
| 45 | 1.4 The peer can predictb the tasks in his role as a peer | Understanding | presentation, discussion | |
| advanced module | 30 | 2.1 The peer can applyb a uniform approach in evaluating a medical report along the prescribed review items | Applying | presentation, discussion, work in small groups |
| 360 | 2.2 The peer can applyb the six quality domains to individual reports by using the four-point ordinal rating scale | Applying | work in small groups | |
| 60 | 2.3 The peer can applyb the superordinate criterion to individual reports by using the three-point ordinal rating scale (traffic light assessment) | Applying | presentation, work in small groups | |
| ---a | 2.4 The peer can classifyb and clearly document quality deficiencies | Analysing | work in small groups | |
| 220 | 2.5 The peer can independently evaluateb a medical experts’ report on basis of the 23 review items | Evaluating | exercises, presentation, work in small groups |
aEducational objective 2.4 is not associated with a specified processing time. This objective has been integrated into the educational objectives 2.2 and 2.3. While working on these two objectives, the training instructor displays adequate documentation of quality deficiencies and discusses them with the peers
bKey action verb associated with the knowledge dimension according to Bloom’s Taxonomy to operationalize the educational objectives
Assessment of confidence in handling the components of the manual (n = 36)
| Individual confidence in handling the manual regarding the… | mode | min | max | mean |
|---|---|---|---|---|
| …superordinate criterion: confirmability of a medical report | 90 | 30 | 100 | 75.8 |
| …quality domain A: formal structure | 90 | 30 | 100 | 83.3 |
| …quality domain B: clarity | 70 | 40 | 100 | 73.9 |
| …quality domain C: transparency | 70 | 40 | 100 | 70.0 |
| …quality domain D: completeness | 80 | 40 | 100 | 71.7 |
| …quality domain E: medical-scientific principles | 90 | 20 | 100 | 79.4 |
| …quality domain F: efficiency | 90 | 50 | 100 | 83.9 |
| …documentation of quality deficiencies | 90 | 30 | 100 | 75.3 |
Percentage peer agreement of the standardized (SPT) and non-standardized (CG) peer-training
| standardized peer-training ( | control group ( | |||||
|---|---|---|---|---|---|---|
| report #A | report #B | report #C | report #D | report #E | report #F | |
| superordinate criterion | 78.8% | 60.1% | 41.8% | 52.9% | 35.7% | 33.3% |
| item A1 | 41.2% | 51.9% | 72.9% | 53.6% | 56.7% | 79.5% |
| item A2 | 37.5% | 46.4% | 61.5% | 70.6% | 89.5% | 79.5% |
| item B1 | 92.3% | 35.0% | 43.3% | 79.1% | 42.0% | 70.8% |
| item B2 | 31.4% | 25.9% | 25.9% | 79.1% | 29.8% | 30.4% |
| item B3 | 85.2% | 36.2% | 39.0% | 100% | 71.9% | 35.7% |
| item C1 | 92.3% | 39.6% | 47.0% | 30.1% | 26.3% | 21.6% |
| item C2 | 71.4% | 62.7% | 32.5% | 70.6% | 55.6% | 70.8% |
| item C3 | 41.2% | 41.9% | 25.4% | 69.3% | 52.9% | 30.4% |
| item D1 | 72.9% | 28.2% | 34.2% | 57.5% | 26.3% | 31.0% |
| item D2 | 78.8% | 67.5% | 58.4% | 78.4% | 33.9% | 62.0% |
| item D3 | 60.3% | 72.4% | 36.8% | 49.7% | 22.8% | 23.4% |
| item D4 | 60.3% | 85.8% | 100% | 88.2% | 47.5% | 77.9% |
| item D5 | 49.9% | 31.3% | 25.0% | 45.8% | 31.6% | 42.7% |
| item E1 | 100% | 100% | 100% | 88.9% | 89.5% | 89.5% |
| item E2 | 85.2% | 32.5% | 48.2% | 78.4% | 41.8% | 42.7% |
| item F1 | 92.3% | 100% | 61.5% | 88.9% | 100% | 44.4% |
| item F2 | 71.4% | 25.6% | 25.9% | 52.9% | 29.2% | 70.8% |
Inter-rater reliability among several assessors determined by Kappa coefficient of Fleiss (κm)