| Literature DB >> 35403367 |
Anna Noel-Storr1, Gerald Gartlehner2,3, Gordon Dooley4, Emma Persad2, Barbara Nussbaumer-Streit5.
Abstract
BACKGROUND: Utilisation of crowdsourcing within evidence synthesis has increased over the last decade. Crowdsourcing platform Cochrane Crowd has engaged a global community of 22,000 people from 170 countries. The COVID-19 pandemic presented an opportunity to engage the community and keep up with the exponential output of COVID-19 research. AIMS: To test whether a crowd could accurately assess study eligibility for reviews under time constraints. OUTCOME MEASURES: time taken to complete each task, time to produce required training modules, crowd sensitivity, specificity and crowd consensus.Entities:
Mesh:
Year: 2022 PMID: 35403367 PMCID: PMC9088532 DOI: 10.1002/jrsm.1559
Source DB: PubMed Journal: Res Synth Methods ISSN: 1759-2879 Impact factor: 9.308
Key task characteristics
| Review | Eligible study types | Size of set | No. of included studies | No. of people invited | No. of people contributed | No. of records assessed/person (range) |
|---|---|---|---|---|---|---|
| Review 1: Quarantine | Observational modelling interventional | 5606 | 47 | 123 | 65 | 4–1201 |
| Review 2: IPC Adherence | Qualitative observational interventional | 3367 | 32 | 85 | 36 | 2–1500 |
| Review 3: Universal Screening | Observational (diagnostic) interventional | 4378 | 18 | 104 | 38 | 10–3168 |
| Review 4: Convalescent Plasma | Observational interventional | 948 | 12 | 122 | 12 | 1–711 |
| Total | 14,299 | 109 | 287 | 101 | 268 |
No. of included studies used in the evaluation datasets (some includes studies were used in the training modules so were not then included in the evaluation datasets).
Unique contributors.
Mean number of records assessed per crowd contributor.
FIGURE 1Screen shot of Review 1: Quarantine [Colour figure can be viewed at wileyonlinelibrary.com]
FIGURE 2Outcome measure: Time [Colour figure can be viewed at wileyonlinelibrary.com]
Crowd accuracy
| Review |
| TP | TN | FP | FN | Sensitivity | Specificity | Consensus |
|---|---|---|---|---|---|---|---|---|
| Review 1: Quarantine | 5606 | 45 | 3942 | 1617 | 2 | 95.7 | 70.9 | 72.02 |
| Review 2: IPC Adherence | 3367 | 31 | 2437 | 897 | 1 | 96.9 | 73.0 | 74.96 |
| Review 3: Universal Screening | 4378 | 17 | 3075 | 1285 | 1 | 94.4 | 70.5 | 71.34 |
| Review 4: Convalescent Plasma | 948 | 12 | 827 | 109 | 0 | 100.0 | 88.7 | 92.19 |
Note: TP = True Positive, the number of records correctly classified as possibly relevant; TN = True Negative, the number of records correctly classified as not relevant; FP = False Positive, the number of records incorrectly classified as possibly relevant; FN = False Negative, the number of records incorrectly classified as not relevant.
FIGURE 3Crowd consensus for included studies [Colour figure can be viewed at wileyonlinelibrary.com]
Crowdsourcing workflows
| Sensitivity maximising | Crowd assessment + author team dual assessment of conflicting crowd records + author team single assessment of |
|---|---|
| Speed maximising | Crowd assessment + author team single assessment of |
| Specificity maximising | Crowd assessment + crowd resolver |
A crowd resolver is a crowd contributor assesses only records that have received discordant classifications, and makes a final crowd classification on the record.