| Literature DB >> 29888085 |
Cheng Ye1, Joseph Coco1, Anna Epishova1, Chen Hajaj1, Henry Bogardus1, Laurie Novak1, Joshua Denny1, Yevgeniy Vorobeychik1, Thomas Lasko1, Bradley Malin1, Daniel Fabbri1.
Abstract
Crowdsourcing services like Amazon Mechanical Turk allow researchers to ask questions to crowds of workers and quickly receive high quality labeled responses. However, crowds drawn from the general public are not suitable for labeling sensitive and complex data sets, such as medical records, due to various concerns. Major challenges in building and deploying a crowdsourcing system for medical data include, but are not limited to: managing access rights to sensitive data and ensuring data privacy controls are enforced; identifying workers with the necessary expertise to analyze complex information; and efficiently retrieving relevant information in massive data sets. In this paper, we introduce a crowdsourcing framework to support the annotation of medical data sets. We further demonstrate a workflow for crowdsourcing clinical chart reviews including (1) the design and decomposition of research questions; (2) the architecture for storing and displaying sensitive data; and (3) the development of tools to support crowd workers in quickly analyzing information from complex data sets.Entities:
Year: 2018 PMID: 29888085 PMCID: PMC5961774
Source DB: PubMed Journal: AMIA Jt Summits Transl Sci Proc
Figure 1:Overview of the crowdsourcing system.
Figure 2:Agenda for crowdsourcing workshop with researchers.
Figure 3:Example of Pybossa presenter with a text search engine.
Figure 4:An example helper library: highlighting similar words in a note
Figure 5:An example of expanded search terms for “diabetes”