Nassr Nama1, Klevis Iliriani2, Meng Yang Xia1, Brian P Chen1, Linghong Linda Zhou1, Supichaya Pojsupap3, Coralea Kappel1, Katie O'Hearn3, Margaret Sampson4, Kusum Menon5, James Dayre McNally5. 1. Faculty of Medicine, University of Ottawa, Ottawa, ON, Canada. 2. School of Medicine, Trinity College, Dublin, Ireland. 3. Department of Pediatrics, Children's Hospital of Eastern Ontario, Ottawa, ON, Canada. 4. Department of Volunteers Communication and Information Resources, Children's Hospital of Eastern Ontario, Ottawa, ON, Canada. 5. Faculty of Medicine, University of Ottawa, Ottawa, ON, Canada;; Department of Pediatrics, Children's Hospital of Eastern Ontario, Ottawa, ON, Canada.
Abstract
BACKGROUND: Completing large systematic reviews and maintaining them up to date poses significant challenges. This is mainly due to the toll required of a small group of experts to screen and extract potentially eligible citations. Automated approaches have failed so far in providing an accessible and adaptable tool to the research community. Over the past decade, crowdsourcing has become attractive in the scientific field, and implementing it in citation screening could save the investigative team significant work and decrease the time to publication. METHODS: Citations from the 2015 update of a pediatrics vitamin D systematic review were uploaded to an online platform designed for crowdsourcing the screening process (http://www.CHEORI.org/en/CrowdScreenOverview). Three sets of exclusion criteria were used for screening, with a review of abstracts at level one, and full-text eligibility determined through two screening stages. Two trained reviewers, who participated in the initial systematic review, established citation eligibility. In parallel, each citation received four independent assessments from an untrained crowd with a medical background. Citations were retained or excluded if they received three congruent assessments. Otherwise, they were reviewed by the principal investigator. Measured outcomes included sensitivity of the crowd to retain eligible studies, and potential work saved defined as citations sorted by the crowd (excluded or retained) without involvement of the principal investigator. RESULTS: A total of 148 citations for screening were identified, of which 20 met eligibility criteria (true positives). The four reviewers from the crowd agreed completely on 63% (95% CI: 57-69%) of assessments, and achieved a sensitivity of 100% (95% CI: 88-100%) and a specificity of 99% (95% CI: 96-100%). Potential work saved to the research team was 84% (95% CI: 77-89%) at the abstract screening stage, and 73% (95% CI: 67-79%) through all three levels. In addition, different thresholds for citation retention and exclusion were assessed. With an algorithm favoring sensitivity (citation excluded only if all four reviewers agree), sensitivity was maintained at 100%, with a decrease of potential work saved to 66% (95% CI: 59-71%). In contrast, increasing the threshold required for retention (exclude all citations not obtaining 3/4 retain assessments) decreased sensitivity to 85% (95% CI: 65-96%), while improving potential workload saved to 92% (95% CI: 88-95%). CONCLUSIONS: This study demonstrates the accuracy of crowdsourcing for systematic review citations screening, with retention of all eligible articles and a significant reduction in the work required from the investigative team. Together, these two findings suggest that crowdsourcing could represent a significant advancement in the area of systematic review. Future directions include further study to assess validity across medical fields and determination of the capacity of a non-medical crowd.
BACKGROUND: Completing large systematic reviews and maintaining them up to date poses significant challenges. This is mainly due to the toll required of a small group of experts to screen and extract potentially eligible citations. Automated approaches have failed so far in providing an accessible and adaptable tool to the research community. Over the past decade, crowdsourcing has become attractive in the scientific field, and implementing it in citation screening could save the investigative team significant work and decrease the time to publication. METHODS: Citations from the 2015 update of a pediatrics vitamin D systematic review were uploaded to an online platform designed for crowdsourcing the screening process (http://www.CHEORI.org/en/CrowdScreenOverview). Three sets of exclusion criteria were used for screening, with a review of abstracts at level one, and full-text eligibility determined through two screening stages. Two trained reviewers, who participated in the initial systematic review, established citation eligibility. In parallel, each citation received four independent assessments from an untrained crowd with a medical background. Citations were retained or excluded if they received three congruent assessments. Otherwise, they were reviewed by the principal investigator. Measured outcomes included sensitivity of the crowd to retain eligible studies, and potential work saved defined as citations sorted by the crowd (excluded or retained) without involvement of the principal investigator. RESULTS: A total of 148 citations for screening were identified, of which 20 met eligibility criteria (true positives). The four reviewers from the crowd agreed completely on 63% (95% CI: 57-69%) of assessments, and achieved a sensitivity of 100% (95% CI: 88-100%) and a specificity of 99% (95% CI: 96-100%). Potential work saved to the research team was 84% (95% CI: 77-89%) at the abstract screening stage, and 73% (95% CI: 67-79%) through all three levels. In addition, different thresholds for citation retention and exclusion were assessed. With an algorithm favoring sensitivity (citation excluded only if all four reviewers agree), sensitivity was maintained at 100%, with a decrease of potential work saved to 66% (95% CI: 59-71%). In contrast, increasing the threshold required for retention (exclude all citations not obtaining 3/4 retain assessments) decreased sensitivity to 85% (95% CI: 65-96%), while improving potential workload saved to 92% (95% CI: 88-95%). CONCLUSIONS: This study demonstrates the accuracy of crowdsourcing for systematic review citations screening, with retention of all eligible articles and a significant reduction in the work required from the investigative team. Together, these two findings suggest that crowdsourcing could represent a significant advancement in the area of systematic review. Future directions include further study to assess validity across medical fields and determination of the capacity of a non-medical crowd.
Entities:
Keywords:
Crowdsourcing; citation screening; systematic reviews; vitamin D
Authors: Kaveh G Shojania; Margaret Sampson; Mohammed T Ansari; Jun Ji; Steve Doucette; David Moher Journal: Ann Intern Med Date: 2007-07-16 Impact factor: 25.391
Authors: Paul A Harris; Robert Taylor; Robert Thielke; Jonathon Payne; Nathaniel Gonzalez; Jose G Conde Journal: J Biomed Inform Date: 2008-09-30 Impact factor: 6.317
Authors: Robert G Badgett; Mohinder Vindhyal; Jason T Stirnaman; C Michael Gibson; Rim Halaby Journal: JAMA Pediatr Date: 2015-08 Impact factor: 16.193
Authors: Kumaravel Rajakumar; Charity G Moore; Jonathan Yabes; Flora Olabopo; Mary Ann Haralam; Diane Comer; Jaimee Bogusz; Anita Nucci; Susan Sereika; Jacqueline Dunbar-Jacob; Michael F Holick; Susan L Greenspan Journal: J Clin Endocrinol Metab Date: 2015-06-19 Impact factor: 5.958
Authors: Gal Dubnov-Raz; Netachen Livne; Raanan Raz; Avner H Cohen; Naama W Constantini Journal: Int J Sport Nutr Exerc Metab Date: 2015-08 Impact factor: 4.599
Authors: Nassr Nama; Margaret Sampson; Nicholas Barrowman; Ryan Sandarage; Kusum Menon; Gail Macartney; Kimmo Murto; Jean-Philippe Vaccani; Sherri Katz; Roger Zemek; Ahmed Nasr; James Dayre McNally Journal: J Med Internet Res Date: 2019-04-29 Impact factor: 5.428
Authors: Emily Allin; Nassr Nama; Michael A Irvine; Colleen Pawliuk; Marie Wright; Matthew Carwana Journal: BMJ Open Date: 2021-03-24 Impact factor: 2.692
Authors: Katie O'Hearn; Cameron MacDonald; Anne Tsampalieros; Leo Kadota; Ryan Sandarage; Supun Kotteduwa Jayawarden; Michele Datko; John M Reynolds; Thanh Bui; Shagufta Sultan; Margaret Sampson; Misty Pratt; Nick Barrowman; Nassr Nama; Matthew Page; James Dayre McNally Journal: BMC Med Res Methodol Date: 2021-07-08 Impact factor: 4.615
Authors: M Ellen Kuenzig; Kirles Bishay; Richard Leigh; Gilaad G Kaplan; Eric I Benchimol Journal: Clin Transl Gastroenterol Date: 2018-09-24 Impact factor: 4.488