| Literature DB >> 21829574 |
Mark A Burgman1, Marissa McBride, Raquel Ashton, Andrew Speirs-Bridge, Louisa Flander, Bonnie Wintle, Fiona Fidler, Libby Rumpff, Charles Twardy.
Abstract
Expert judgements are essential when time and resources are stretched or we face novel dilemmas requiring fast solutions. Good advice can save lives and large sums of money. Typically, experts are defined by their qualifications, track record and experience. The social expectation hypothesis argues that more highly regarded and more experienced experts will give better advice. We asked experts to predict how they will perform, and how their peers will perform, on sets of questions. The results indicate that the way experts regard each other is consistent, but unfortunately, ranks are a poor guide to actual performance. Expert advice will be more accurate if technical decisions routinely use broadly-defined expert groups, structured question protocols and feedback.Entities:
Mesh:
Year: 2011 PMID: 21829574 PMCID: PMC3146531 DOI: 10.1371/journal.pone.0022998
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Characteristics of the expert groups.
| Workshop | Discipline | Numbe of experts | Location/Date | Range of years of relevant experience | Range of qualifications | Range of number of publications |
| 1 | Animal and plant biosecurity and quarantine | 21 | Melbourne, February 2010 | 0–37 | BSc, BASc, BVSc, BCom, Grad. Dip., MSc, PhD | 0–113 |
| 2 | Animal and plant biosecurity | 24 | Christchurch, March 2010 | 0–39 | BSc, MSc, MBA, MCom, PhD | 0–270 |
| 3 | Ecology, frog biology | 13 | Melbourne, March, 2010 | 0–42 | BA, BSc, BSc (Hons), M Env Studies, PhD | 0–45 |
| 4 | Public health, medicine | 25 | Canberra, May 2010 | 0–45 | BEng, BSc, BEcon, LLB, MBBS, Grad. Dip., MA, MSc, MBA, PhD | 0–220 |
| 5 | Risk analysis, biosecurity | 20 | Sydney, September 2010 | 0–40 | BEng, BSc, BVSc, MBBS, Grad. Dip., MA, MBA, PhD | 0–225 |
| 6 | Weed ecology | 20 | Melbourne, December 2010 | 0–50 | BSc, MSc, PhD | 1–220 |
Pearson correlations between peer assessments of expert performance and measures of expertise, and between self-assessed expertise and peer assessments of expertise.
| Workshop | Peer assessment versus years of experience | Peer assessment versus number of publications | Peer assessment versus qualifications | Self assessment versus peer assessment |
| 1 |
| 0.348 (n = 21) |
|
|
| 2 |
|
| 0.064 (n = 20) |
|
| 3 |
| −0.123 (n = 13) | 0.019 (n = 13) |
|
| 4 |
|
|
|
|
| 5 |
| 0.309 (n = 17) | 0.203 (n = 20) |
|
| 6 |
| 0.289 (n = 14) | 0.074 (n = 14) |
|
Statistically significant correlations (at α = 0.05, two-tailed) are in bold face.
Figure 1Self assessment versus peer assessment of expertise.
Data from all workshops (overall correlation, r = 0.85). Peer assessment is the average of the scores on the 11-point scale provided by each person's peers on the day of the workshop. The strong relationship was consistent across the five groups, where the correlations ranged from 0.67 to 0.94 (Table 3). The dashed line is parity (where self assessment and peer assessment are equal).
Correlations between peer assessments of expertise and the accuracy of predictions.
| Workshop, number of participants, number of questions | Peer assessment versus prediction accuracy – first round | Peer assessment versus prediction accuracy – second round |
| 1, n = 21, 10 questions | −0.391 | 0.119 |
| 2, n = 24, 10 questions | 0.215 | −0.009 |
| 3, n = 13, 8 questions | 0.190 | −0.470 |
| 4, n = 25, 5 questions | −0.360 | −0.148 |
| 5, n = 20, 6 questions | −0.305 | −0.441 |
| 6, n = 14, 8 questions | −0.367 | −0.332 |
A negative correlation indicates that more experienced and better-credentialed experts were closer to the truth. A positive correlation indicates the converse. None of the correlations were statistically significant (at α = 0.05, two-tailed).
Figure 2Peer assessment of expert knowledge versus actual performance for the participants in Workshop 3.
Prediction accuracy is calculated as ALRE (see text). Small values for prediction accuracy are better. Closed circles and the solid line are estimates from round 1 (r = 0.19). Crosses and the dashed line are estimates from round 2 (r = −0.47). Estimates closer to the x-axis indicate the answers are closer to the truth.
Figure 3The group average improvement in accuracy (ALRE) following discussion.
Change in estimates records distance from the truth, so that more strongly negative values improved more. The dots are the improvements in averages of best guesses. The error bars are 95% confidence intervals.
Figure 4Accuracy of group means compared to highly regarded experts.
Mean and 95% confidence intervals of standardized distance from the truth (ALRE) for the most highly regarded individual in each workshop prior to discussion (‘Highest status’), and the workshop group average following discussion (‘Group Average’).