Literature DB >> 25356929

Rapid grading of fundus photographs for diabetic retinopathy using crowdsourcing.

Christopher J Brady1, Andrea C Villanti, Jennifer L Pearson, Thomas R Kirchner, Omesh P Gupta, Chirag P Shah.   

Abstract

BACKGROUND: Screening for diabetic retinopathy is both effective and cost-effective, but rates of screening compliance remain suboptimal. As screening improves, new methods to deal with screening data may help reduce the human resource needs. Crowdsourcing has been used in many contexts to harness distributed human intelligence for the completion of small tasks including image categorization.
OBJECTIVE: Our goal was to develop and validate a novel method for fundus photograph grading.
METHODS: An interface for fundus photo classification was developed for the Amazon Mechanical Turk crowdsourcing platform. We posted 19 expert-graded images for grading by Turkers, with 10 repetitions per photo for an initial proof-of-concept (Phase I). Turkers were paid US $0.10 per image. In Phase II, one prototypical image from each of the four grading categories received 500 unique Turker interpretations. Fifty draws of 1-50 Turkers were then used to estimate the variance in accuracy derived from randomly drawn samples of increasing crowd size to determine the minimum number of Turkers needed to produce valid results. In Phase III, the interface was modified to attempt to improve Turker grading.
RESULTS: Across 230 grading instances in the normal versus abnormal arm of Phase I, 187 images (81.3%) were correctly classified by Turkers. Average time to grade each image was 25 seconds, including time to review training images. With the addition of grading categories, time to grade each image increased and percentage of images graded correctly decreased. In Phase II, area under the curve (AUC) of the receiver-operator characteristic (ROC) indicated that sensitivity and specificity were maximized after 7 graders for ratings of normal versus abnormal (AUC=0.98) but was significantly reduced (AUC=0.63) when Turkers were asked to specify the level of severity. With improvements to the interface in Phase III, correctly classified images by the mean Turker grade in four-category grading increased to a maximum of 52.6% (10/19 images) from 26.3% (5/19 images). Throughout all trials, 100% sensitivity for normal versus abnormal was maintained.
CONCLUSIONS: With minimal training, the Amazon Mechanical Turk workforce can rapidly and correctly categorize fundus photos of diabetic patients as normal or abnormal, though further refinement of the methodology is needed to improve Turker ratings of the degree of retinopathy. Images were interpreted for a total cost of US $1.10 per eye. Crowdsourcing may offer a novel and inexpensive means to reduce the skilled grader burden and increase screening for diabetic retinopathy.

Entities:  

Keywords:  Amazon Mechanical Turk; crowdsourcing; diabetic retinopathy; fundus photography; telemedicine

Mesh:

Year:  2014        PMID: 25356929      PMCID: PMC4259907          DOI: 10.2196/jmir.3807

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   5.428


Introduction

Since early diabetic retinopathy (DR) is often asymptomatic, detection of disease at this stage is either incidental or by deliberate screening. Screening for DR is both effective and cost-effective [1-4], but adherence rates to published guidelines for screening for DR are low, with only 35-60% of diabetic patients receiving an annual dilated fundus examination in the United States [5-8]. As a way to increase adherence, telehealth screening using non-mydriatic fundus photography and remote interpretation is increasing, especially in rural and remote settings [9-11]. Early diagnosis of DR and institution of appropriate therapy represents an enormous opportunity to prevent vision loss in a young, working-age demographic [3,4]. Telehealth, in particular, may be a way to control provider, payer, and societal costs. Among the costs of a telehealth program are the fundus camera, the telehealth software package, and the human resources needed for image acquisition and interpretation. Fundus photo interpretation costs in diabetic retinopathy screening may be high given the labor-intensive interpretation protocols and the need to interpret multiple images per patient. Computerized, semi-automated image analysis techniques have been developed that may be able to reduce physician workload and screening costs [12-14]; however, these methods are not FDA-approved, nor in wide use clinically at this time. If telehealth continues to expand, low-cost methods will be needed to interpret the large volume of fundus images expected with rising incidence of diabetes, especially in resource-poor settings and in large public health screenings. Crowdsourcing is defined by Brabham as “an online, distributed problem-solving and production model that leverages the collective intelligence of online communities to serve specific organizational goals” [15]. A subset of crowdsourcing, which he terms “distributed-human-intelligence tasking”, can involve subdividing larger tasks into small portions and then recruiting a group of individuals to each complete these small portions, and only collectively, the entire task [15]. The use of crowdsourcing in biomedical research is in its infancy, though some groups have used this method in public health research [16] and to interpret medical imaging. For example, malaria researchers have used a Web-based game to recruit untrained, anonymous volunteers to tag and count malaria parasites on digital images of blood smears [17]. The investigators showed that by combining the analyses of several users, they were able to achieve similar accuracy rates to expert microscopists. Crowdsourcing has recently been used to categorize a number of fundus photos with a variety of diagnoses as normal or abnormal [18]. In a trial conducted in the United Kingdom using untrained graders, the sensitivity was ≥96% for normal versus severely abnormal and between 61-79% for normal versus mildly abnormal [18]. The current research uses diabetic retinopathy as the test condition to explore whether a crowdsourcing interface can be used to train workers to classify human fundus photos as normal or abnormal and subsequently conduct diagnostic grading of images [19]. This project estimates the validity and reliability of crowdsourced grading of images for diabetic retinopathy, compared to the criterion standard of expert grading. Our hypothesis is that crowdsourced grading of fundus photography interpretation can be rapid, accurate, and reliable in the screening for diabetic retinopathy.

Methods

Crowdsourcing

An interface for fundus photo classification was developed for the Amazon Mechanical Turk (AMT [20]) crowdsourcing platform (Figure 1). AMT is an online labor market that allows access to thousands of people who can quickly accomplish small, discrete tasks for small amounts of money. Typical AMT tasks include tagging photos, translating words, or writing very short articles for websites. AMT has also been used to annotate photos of the tobacco point-of-sale retail environment [21], evaluate oral health promotion materials [22], investigate the relationship between depression and stigma [23], assess people’s reactions to frightening anti-smoking campaigns [24], and evaluate public awareness of ovarian cancer [25], among many other research-orientated applications [26-28]. Amazon Mechanical Turk has its own vocabulary used by AMT workers (Turkers) and AMT task administrators (Requestors). A Human Intelligence Task (HIT) is a small job that may be performed in a matter of seconds or minutes and, once the work is approved by the requestor, may pay US $0.01-$0.25 or more per task depending on the complexity of the HIT. A group of HITs is called a “batch” and is made up of similar HITs. Depending on the complexity of the task and the payment offered by the Requestor, a batch is often completed within minutes or hours of posting.
Figure 1

Screenshot of the Amazon Mechanical Turk Web interface for fundus photo grading.

AMT is a reputation-based economy such that Turkers may only access the most desirable HITs once they have a sufficient track record of previously accepted work [29]. High quality Turkers may avoid a new Requestor’s HITs until the Requestor has demonstrated his or her own fairness in approving and rejecting work. Indeed, a Turker’s reputation will suffer following rejection of even a small number of HITs. AMT is a complex ecosystem in which both high-quality work on the part of the Turkers and fairness on the part of the Requestor are rewarded. Turkers perform their work anonymously, but demographic studies have been conducted. In a survey of 1000 Turkers, Ipeirotis found that 46.8% of Turkers are located in the United States, 34% are in India, and the remaining 19.2% were from 64 other countries [30]. The majority of workers in the United States were women, most of whom reported AMT as a source of supplemental income, whereas in the majority of workers in India were men, and reported AMT as their primary source of income. Across nations, Turkers were younger and better educated than the general population [30]. Screenshot of the Amazon Mechanical Turk Web interface for fundus photo grading.

Design

For the current study, the United Kingdom national screening program grading scale [31] was used by 2 attending retinal surgeons (OPG, CPS) to categorize 19 anonymized teaching images. The same AMT interface and resolution of images used by the Turkers was used by the expert graders. Consensus was reached by discussion on images for which there was disagreement. For the purposes of the study, terms from the UK grading scale were translated into plain language: “background” retinopathy was called “mild”, “preproliferative” was called “moderate”, and “proliferative” was called “severe”. “Maculopathy” was defined as abnormal on a training image with otherwise moderate disease but was not coded separately. The AMT interface was designed to provide training on grading of DR within each HIT. This training included 6 images annotated with the salient features of each level of retinopathy in plain language. Turkers were presented with the following text: “This is a photo of the inside of the eye. We are looking to label eyes as healthy or unhealthy with respect to diabetes. Rate this eye.” Turkers could hover their mouse over 6 adjacent training images (1 normal, 1 mild, 1 moderate, 3 severe) while reviewing the active test image. This layout allowed for all of the training and grading to occur in one browser window. More examples of proliferative/severe disease were provided due to the heterogeneity of findings constituting this level of disease. There were no restrictions on the type of device or display/monitor used by Turkers to complete the task.

Phase I

In the first phase of the study, the 19 images were posted to AMT for grading by Turkers, with 10 repetitions per photo for an initial proof-of-concept. Four photos were re-posted with this initial batch for 10 repetitions to assess intragrader reliability. Turkers were paid US $0.10 per image, and a 10% commission was paid to Amazon. In order to be eligible to view and complete the HITs, Turkers needed to have successfully completed 100 prior HITs and have an overall HIT approval rate of 97%. In the initial batch, HITs were posted asking Turkers to grade images as normal (0) versus abnormal (1). In subsequent batches, Turkers were asked to grade the same 19 images using three categories (normal=0, mild to moderate=1, and severe=2) and then four categories (normal=0, mild=1, moderate=2, and severe=3). Percentage of images correctly classified was calculated. Sensitivity and specificity were calculated for all batches, collapsing all abnormal categories together for three- and four-category grading tasks using Stata 13. A worker consensus grade was assigned to each image based on the mode of the 10 Turker scores. Consensus grade using the mean of the 10 Turker scores was also calculated on an exploratory basis. For the two category tasks, a mean score <0.50 was defined as normal, and ≥0.50 was abnormal. For the three category tasks, <0.50 was defined as normal, ≥0.5 to <1.5 was defined as mild to moderate, and ≥1.5 was defined as severe. In the four category tasks, <0.50 was defined as normal, ≥0.5 to <1.5 was defined as mild, ≥1.5 to <2.5 was defined as moderate, and ≥2.5 was defined as severe.

Phase II

The purpose of Phase II was two-fold. First, Phase II sought to evaluate the emergent ability of the crowd to accurately distinguish between different levels of retinopathy, based on the idea that larger numbers of raters would increasingly coalesce around the correct answer. Second, Phase II sought to identify and ultimately confirm the threshold beyond which the contribution of one more rater would cease to provide additional information. To accomplish this, one prototypical image from each of the four grading categories was submitted to undergo 500 unique Turker interpretations. Jackknife resampling methods were then used to draw random subsamples from this “population” of 500 Turkers, beginning with 50 random samples of 2 Turkers, then 50 random samples of 3 Turkers, and so forth [21,32]. This made it possible to estimate the variance in accuracy derived from each randomly drawn sample of raters at each crowd size and to compute area under the curve (AUC) of the receiver-operator characteristic (ROC) curve, indicating their performance relative to the expert grading.

Phase III

In Phase III of the study, three additional iterative batches of the 19 images were run in an effort to improve Turker grading using the four diagnostic categories. The first batch used Turker feedback from all previous batches to modify the training image set. The second batch used more stringent criteria for Turkers in addition to the modified training images. A minimum of 500 completed and approved HITs was required as was an overall HIT approval rate of 99%. The third batch was conducted using Turkers holding an Amazon designation of “Photo Moderation Master” and raised the compensation to US $0.15 per image, in addition to the changes made for batches 1 and 2. The criteria necessary to achieve this designation are not published, but Masters are “elite groups of Workers who have demonstrated accuracy on specific types of HITs on the Mechanical Turk marketplace. Workers achieve a Masters distinction by consistently completing HITs of a certain type with a high degree of accuracy across a variety of Requesters” [33]. As in Phase I, percent correctly classified, worker consensus score, and average time to complete the HITs were estimated for these three iterative batches. The Wills Eye Institute Institutional Review Board ruled that approval was not required for this study.

Results

Two expert graders (OPG, CPS) coded 12 images as abnormal and 7 as normal (Table 1). Each of the three Phase I batches consisted of 23 photos (19 unique, 4 duplicates) with 10 unique graders for a total of 230 grading instances. Time-to-complete Turker grading of images varied with the number of grading categories. Two category (normal/abnormal) grading was completed in 20 minutes, three-category grading in 3 hours, and four-category grading in 2 hours. Because the images were interpreted rapidly and workers could complete as many or as few of the HITs as desired, most of the four duplicate images were rated by unique Turkers and therefore, we were unable to assess intragrader reliability.
Table 1

Turker grading of individual imagesa.

Image #Two-category ratingThree-category ratingFour-category rating
Expert ratingCorrect diagnosisb, %Turker consensusc Expert ratingCorrect diagnosisb, %Turker consensusc Expert ratingCorrect diagnosisb, %Turker consensusc
1Nor65Nor90Nor55
2Ab85M/M50SevMild0Sev
3Nor70Nor70Nor70
4Nor50AbNor40M/MNor60
5Nor80Nor70Nor50
6Ab100M/M90Mild20Mod
7Ab90Severe60Sev10Mod
8Nor50AbSev40M/MNor65
9Ab100Sev95Sev100
10Ab100Sev40M/MSev70
11Ab90Sev0M/MSev20Mild
12Nor90Nor80Nor90
13Ab100M/M30SevMod20Sev
14Ab80Sev40M/MSev10Mod
15Nor90Nor100Nor90
16Ab90Sev70Sev50
17Ab100M/M60Mild10Mod
18Ab100M/M100Mod95
19Ab90M/M80Mild20Mod
Correct, % 81.389.5 64.463.2 50.957.9
Sensitivityd, % 93.6100.0 96.3100.0 96.3100.0
Specificityd, % 67.871.4 66.771.4 66.7100.0

aNor=Normal; Ab=Abnormal; M/M=Mild or Moderate; Sev=Severe; Mod=Moderate.

bAt the level of the individual graders.

cConsensus rating presented only if it differed from the expert rating.

dCalculated for normal versus any disease level.

Across 230 grading instances of unique images in the two-category HITs (normal vs. abnormal) of Phase I, 187 (81.3%) of the images were correctly classified by Turkers (Table 1). Sensitivity and specificity were 93.6% and 67.8% respectively using individual Turker scores. Sensitivity and specificity were 100% and 71.4% respectively using Turker consensus scores. Average time to grade each image was 25 seconds, including time to review training images. At US $0.11 per grading, each image was graded for $1.10, and grading garnered an effective hourly wage of $14.31 (Table 2).
Table 2

Time to complete ratings (in seconds).

 Two-category ratingThree-category ratingFour-category ratingFour-category rating (improved training)Four-category rating (increased approval)Four-category rating (Master Graders)a
Mean time per HITs25.1650.8754.5250.9838.7944.14
95% CI21.93-28.3843.18-58.5546.15-62.8839.66-62.3031.65-45.9336.00-52.27
Hourly wage, $14.317.086.607.069.2812.23
Cost per image, $1.101.101.101.101.101.95

aMaster graders received US $0.15 per image, plus a 30% Amazon commission for a total cost of US $0.195/image.

Overall number of correctly classified images decreased with the addition of a third and fourth grading category to 64.4% (148/230) and 50.9% (117/230), respectively. Specificity and specificity for individual Turkers was 96.3% and 66.7% respectively for both three and four categories. At the level of Turker consensus, sensitivity was 100% for both three and four categories, and specificity was 71.4% and 100% for three and four categories, respectively. With additional grading criteria, the speed of grading decreased, as did the effective hourly wage. Average time to complete the three-category HITs was 51 seconds, for an effective hourly wage of $7.08. Average time to complete the four-category HITs was 55 seconds, for an effective hourly wage of $6.60 (Table 2). Turker grading of individual imagesa. aNor=Normal; Ab=Abnormal; M/M=Mild or Moderate; Sev=Severe; Mod=Moderate. bAt the level of the individual graders. cConsensus rating presented only if it differed from the expert rating. dCalculated for normal versus any disease level. Time to complete ratings (in seconds). aMaster graders received US $0.15 per image, plus a 30% Amazon commission for a total cost of US $0.195/image. Results of Phase II likewise indicate that sensitivity and specificity for overall ratings of abnormal versus normal was excellent, producing a highly significant AUC (0.98; Figure 2, Panel D). Turkers were not as accurate when asked to differentiate among four severity levels. Post hoc contrast analyses, however, indicate that Turkers performed well when asked to identify the abnormalities that were moderate in severity (ROC=0.85; Figure 2, Panel B). The pattern of results indicates that lower accuracy identifying mild (ROC=0.57; Figure 2, Panel A) and severe (AUC=0.73; Figure 2, Panel C) abnormalities was due to a tendency to rate all abnormalities as moderate in severity, rather than a failure to recognize normal versus mild and severe abnormalities more generally. Results also indicate that maximum AUC was usually achieved when crowd size reached a total of between 7 and 10 Turkers, confirming the validity of the crowd sizes used to rate the larger set of images (Figure 2). This affirms that the results of Phases I and III would not have been different had we sought a larger number of Turkers for each HIT.
Figure 2

Area under the curve (AUC) of the receiver-operator characteristic (ROC) curve for increasing numbers of Turker interpretations of a prototypical image from each severity level. Turkers had low accuracy for the Mild (Panel A) and Severe image (Panel C), but acceptable accuracy for the Moderate image (Panel B). When all four images were analyzed for absence or presence of disease only, Turkers performed well (Panel D) with a highly significant AUC.

Area under the curve (AUC) of the receiver-operator characteristic (ROC) curve for increasing numbers of Turker interpretations of a prototypical image from each severity level. Turkers had low accuracy for the Mild (Panel A) and Severe image (Panel C), but acceptable accuracy for the Moderate image (Panel B). When all four images were analyzed for absence or presence of disease only, Turkers performed well (Panel D) with a highly significant AUC. In Phase III, an additional normal training image was added due to Turkers’ interpreting visible choroidal vessels as abnormal during Phase I. The annotations were clarified to indicate that the presence of a single microaneurysm is considered abnormal and that hard exudates need not be present to achieve moderate or severe disease. Functionality to enlarge the image being graded was added. In the first batch using the new training, the percentage of correctly classified images using the consensus Turker scores was no better than previous: 42.1% (8/19 images) versus 57.9% (11/19 images), respectively by mode and 21.1% (4/19 images) versus 26.3% (5/19 images) by mean. In the second batch, with more stringent criteria for Turker selection, the proportion of correctly graded images improved to 52.6% (10/19 images) by mean (mode unchanged). This result was confirmed in a repeat run of this exact batch (data not shown). In the third batch, using “Photo Moderation Masters”, the proportion of correctly graded images decreased to 36.8% (7/19 images) by mean (mode unchanged). Throughout all batches, the diagnostic sensitivity for any level of disease was maintained at 100% (Table 3)
Table 3

Turker consensus in Phase III.

 Number correct (mean)a % correct (mean)Number correct (mode)a % correct (mode)Sensitivityb Specificityb
Phase I: Four-category rating526.31157.9100.0100.0
Phase 3: Trial 1 (improved training)421.1c 8d 42.1100.057.1
Phase 3: Trial 2 (raised approval rating)1052.611e 57.9100.0100.0
Phase 3: Trial 3 (Master Graders)736.81157.9100.0100.0

aCalculated by level (eg, Turker consensus matches expert designation as normal, mild, moderate, and severe).

bCalculated for normal versus any disease level using the mode consensus score.

cAfter excluding a single Turker with systematically higher scores, 42.1% correct.

dThree images had no mode and were considered incorrect for “Number Correct” and “% correct” but recoded as abnormal for sensitivity and specificity.

eOne image had no mode and was considered incorrect for “Number Correct” and “% correct” but recoded as abnormal for sensitivity and specificity.

Turker consensus in Phase III. aCalculated by level (eg, Turker consensus matches expert designation as normal, mild, moderate, and severe). bCalculated for normal versus any disease level using the mode consensus score. cAfter excluding a single Turker with systematically higher scores, 42.1% correct. dThree images had no mode and were considered incorrect for “Number Correct” and “% correct” but recoded as abnormal for sensitivity and specificity. eOne image had no mode and was considered incorrect for “Number Correct” and “% correct” but recoded as abnormal for sensitivity and specificity.

Discussion

Principal Findings

With minimal training, an anonymous, untrained workforce recruited through a public crowdsourcing platform can rapidly and correctly categorize fundus photos of diabetic patients as normal or abnormal. The Turkers in this study graded these images with a high sensitivity, which is ideal for a screening instrument. Critically, no false negative consensus results were generated in any of our batches, indicating that no cases of DR were missed. When asked to categorize photos by degree of retinopathy, Turkers improved with changes in the Turk interface, particularly with increasing prior approval rating needed to perform the HITs. The number of graders required to reach a reliable “consensus grade” was consistent across categories, and 10 grading instances per image was established as sufficient for valid results. Images were interpreted for a total cost of US $1.10 per eye. While abnormal images would still currently still require further evaluation to identify patients in need of a live examination, this cost is near the limit suggested by some investigators for feasible public health screening in developing economies [34]. Indeed, the reimbursement per image, which ranged from an effective hourly wage of US $6.60 to $14.31 in our study, could possibly be reduced, since a reasonable target hourly wage for AMT workers is the federal minimum wage of US $7.25 per hour or even lower [35]. Additionally, posting larger batches of images might allow for lower per image reimbursement, since Turker speed would likely increase after becoming more skilled at the task, allowing them to maintain the same effective wage. While there may not be a direct relationship between quality responses and high wages [36], there may be a link between high wages and rapid completion of image grading, so it may not be wise to dramatically reduce reimbursement. A post hoc analysis of individual Turkers’ scores revealed inconsistent use of the four grading categories by one Turker in the first batch of Phase 3 (Table 3). Several issues are brought to light by considering this specific batch. First, inconsistent use of all categories was a rare occurrence, demonstrating that Turkers are conscientious workers. This was also evident from comments made by Turkers as they completed HITs, which included thoughtful suggestions for improvement to the interface and concern over HITs that were felt to be ambiguous. Second, using the mean of crowdsourced responses may generate outputs that are rather sensitive to outliers. For this reason, using the mode to calculate consensus is generally preferable, though some images may not have a pure mode (Table 3), in which case the higher score of any “tie” would be used clinically. Third, and more broadly, AMT may be susceptible to Turker accounts that attempt to take advantage of the system by rapidly completing HITs with random responses either with live individuals or with automated programs or “bots” [29]. Moving forward, it may be necessary to analyze raw Turker scores for such phenomena and perhaps develop methods to exclude systematically unreliable scores. Since AMT is a reputation-based economy, Requestors can reject or block Turkers who are not performing appropriately. Both actions negatively impact the Turkers’ reputation, which in turn affects their ability to perform HITs in the future, so there is a strong incentive to perform tasks accurately and honestly. This is likely why increasing the prior HIT approval rating to 99% had the most dramatic impact on consensus accuracy. Adding the “Photo Moderation Master” qualification did not improve worker consensus. This may be due to the fact that the criteria Amazon uses to grant this qualification are not relevant to our task. Additionally, since only a fraction of Turkers have the qualification, requiring it reduces the available workforce, which can increase the time required to complete batches. Especially when factoring in the additional Amazon commission, use of the Master qualification may not be necessary or cost-effective for these types of tasks in the future. The current study was limited to a small set of hand-picked, high-quality, mydriatic fundus photos chosen to illustrate the key findings in diabetic retinopathy. Screening populations might have a subset of low-quality or uninterpretable images and would also be expected to have far more images of normal fundi. Identifying pathology in such sets would require extra vigilance on the part of Turkers to detect mild disease within large groups of normal photos. Larger datasets with more subtle pathology need to be tested with this methodology. Additionally, analyzing whether iterative improvements to the interface lead to better results is confounded by the fact that Turkers may have previously been exposed to the task and may be improving in their grading through practice. This is unlikely because the Turkers receive no feedback on their grading, so they do not know if they have correctly graded images or not. Moreover, while it is not currently feasible to “block” Turkers who have previously completed any of our HITs, it is possible to view their grading history within the project. Surprisingly, throughout all batches posted, most of our HITs were completed by Turkers otherwise naïve to our project. In the final batch of 190 HITs posted for this project, after approximately 3000 HITs using the same images had been posted, 170 (89.5%) were completed by Turkers who had never done any of our HITs before, and 20 HITs were done by 3 individual Turkers who had each graded only four images previously (data not shown). In future larger batches, adjusting for individual graders’ improvement over time could become necessary.

Future Considerations

While further refinement of the methodology is still needed to resolve the degree of retinopathy, the current model could possibly be used as a way to reduce the burden on skilled graders by eliminating most of the normal images in very large screenings and passing only those abnormal images on for further characterization. While the individuals who complete HITs on AMT are truly anonymous, they do have unique numerical identifiers and can be tracked across HITs and batches of HITs. Therefore, an intriguing possibility using a crowdsourcing interface could include developing a cadre of specific Turkers who demonstrate a track record of reliable grading. These graders might be capable of a higher level of categorization than the general pool of Turkers and could be recruited for more complex grading tasks. Additionally, it is likely that automated computer algorithms will also play a role in the analysis of fundus images for DR and other conditions in the future. This raises the possibility of an even more robust interaction between artificial intelligence and human intelligence. Images could be graded in a tiered fashion by one system, and then those graded ambiguously could be routed to the other for further validation. An unanticipated benefit of such a crowdsourcing program is that it might raise awareness of diabetes and diabetic retinopathy. Since our interface allowed Turkers to leave feedback for us to refine the instrument, we were able to capture comments such as, “I have learn about diabetes little bit [sic]”, “I really liked seeing the pics of the eye, very interesting”, and “This HIT was very good and a nice break from all of the bubbling surveys. Thank you!”, suggesting an interest in the subject matter beyond other HITs and beyond what we had expected at the outset. This finding is consistent with what has been termed “Virtual Citizen Science” in fields outside of biomedical research [37].

Conclusions

Crowdsourcing represents a novel and inexpensive means to rapidly identify diabetic retinopathy. Further refinements of the technique are required, as is external validation with larger image sets. Additionally, multiple medico-legal and ethical issues would need to be addressed prior to clinical use of this technology, but there may be a role for crowdsourcing medical imaging data in large public health screenings and other settings in the future.
  27 in total

1.  Vision loss among diabetics in a group model Health Maintenance Organization (HMO).

Authors:  Donald S Fong; Mohktar Sharza; Wansu Chen; John F Paschal; Reginald G Ariyasu; Paul P Lee
Journal:  Am J Ophthalmol       Date:  2002-02       Impact factor: 5.258

2.  Diabetic retinopathy in the developing world: how to approach identifying and treating underserved populations.

Authors:  David S Friedman; Ferhina Ali; Nickolas Kourgialis
Journal:  Am J Ophthalmol       Date:  2011-02       Impact factor: 5.258

3.  Patterns of adherence to diabetes vision care guidelines: baseline findings from the Diabetic Retinopathy Awareness Program.

Authors:  E R Schoenfeld; J M Greene; S Y Wu; M C Leske
Journal:  Ophthalmology       Date:  2001-03       Impact factor: 12.079

4.  Improving access to eye care: teleophthalmology in Alberta, Canada.

Authors:  Mancho Ng; Nawaaz Nathoo; Chris J Rudnisky; Matthew T S Tennant
Journal:  J Diabetes Sci Technol       Date:  2009-03-01

Review 5.  Progress towards automated diabetic ocular screening: a review of image analysis and intelligent systems for diabetic retinopathy.

Authors:  T Teng; M Lefley; D Claremont
Journal:  Med Biol Eng Comput       Date:  2002-01       Impact factor: 2.602

6.  Cost-effectiveness of detecting and treating diabetic retinopathy.

Authors:  J C Javitt; L P Aiello
Journal:  Ann Intern Med       Date:  1996-01-01       Impact factor: 25.391

7.  Longitudinal rates of annual eye examinations of persons with diabetes and chronic eye diseases.

Authors:  Paul P Lee; Zachary W Feldman; Jan Ostermann; Derek S Brown; Frank A Sloan
Journal:  Ophthalmology       Date:  2003-10       Impact factor: 12.079

8.  Crowdsourcing black market prices for prescription opioids.

Authors:  Nabarun Dasgupta; Clark Freifeld; John S Brownstein; Christopher Mark Menone; Hilary L Surratt; Luke Poppish; Jody L Green; Eric J Lavonas; Richard C Dart
Journal:  J Med Internet Res       Date:  2013-08-16       Impact factor: 5.428

9.  The cure: design and evaluation of a crowdsourcing game for gene selection for breast cancer survival prediction.

Authors:  Benjamin M Good; Salvatore Loguercio; Obi L Griffith; Max Nanis; Chunlei Wu; Andrew I Su
Journal:  JMIR Serious Games       Date:  2014-07-29       Impact factor: 4.143

10.  Crowdsourcing malaria parasite quantification: an online game for analyzing images of infected thick blood smears.

Authors:  Miguel Angel Luengo-Oroz; Asier Arranz; John Frean
Journal:  J Med Internet Res       Date:  2012-11-29       Impact factor: 5.428

View more
  29 in total

1.  Crowdsourcing to Evaluate Fundus Photographs for the Presence of Glaucoma.

Authors:  Xueyang Wang; Lucy I Mudie; Mani Baskaran; Ching-Yu Cheng; Wallace L Alward; David S Friedman; Christopher J Brady
Journal:  J Glaucoma       Date:  2017-06       Impact factor: 2.503

2.  Crowdsourcing for error detection in cortical surface delineations.

Authors:  Melanie Ganz; Daniel Kondermann; Jonas Andrulis; Gitte Moos Knudsen; Lena Maier-Hein
Journal:  Int J Comput Assist Radiol Surg       Date:  2016-06-27       Impact factor: 2.924

Review 3.  Crowdsourcing and Automated Retinal Image Analysis for Diabetic Retinopathy.

Authors:  Lucy I Mudie; Xueyang Wang; David S Friedman; Christopher J Brady
Journal:  Curr Diab Rep       Date:  2017-09-23       Impact factor: 4.810

Review 4.  Comparing Amazon's Mechanical Turk Platform to Conventional Data Collection Methods in the Health and Medical Research Literature.

Authors:  Karoline Mortensen; Taylor L Hughes
Journal:  J Gen Intern Med       Date:  2018-01-04       Impact factor: 5.128

Review 5.  Clinical Components of Telemedicine Programs for Diabetic Retinopathy.

Authors:  Mark B Horton; Paolo S Silva; Jerry D Cavallerano; Lloyd Paul Aiello
Journal:  Curr Diab Rep       Date:  2016-12       Impact factor: 4.810

6.  The wisdom of crowds for visual search.

Authors:  Mordechai Z Juni; Miguel P Eckstein
Journal:  Proc Natl Acad Sci U S A       Date:  2017-05-10       Impact factor: 11.205

7.  Defining Glioblastoma Resectability Through the Wisdom of the Crowd: A Proof-of-Principle Study.

Authors:  Adam M Sonabend; Brad E Zacharia; Michael B Cloney; Aarón Sonabend; Christopher Showers; Victoria Ebiana; Matthew Nazarian; Kristin R Swanson; Anne Baldock; Henry Brem; Jeffrey N Bruce; William Butler; Daniel P Cahill; Bob Carter; Daniel A Orringer; David W Roberts; Oren Sagher; Nader Sanai; Theodore H Schwartz; Daniel L Silbergeld; Michael B Sisti; Reid C Thompson; Allen E Waziri; Zoher Ghogawala; Guy McKhann
Journal:  Neurosurgery       Date:  2017-04-01       Impact factor: 4.654

8.  Scan, dwell, decide: Strategies for detecting abnormalities in diabetic retinopathy.

Authors:  Samrudhdhi B Rangrej; Jayanthi Sivaswamy; Priyanka Srivastava
Journal:  PLoS One       Date:  2018-11-16       Impact factor: 3.240

Review 9.  Crowdsourcing: an overview and applications to ophthalmology.

Authors:  Xueyang Wang; Lucy Mudie; Christopher J Brady
Journal:  Curr Opin Ophthalmol       Date:  2016-05       Impact factor: 3.761

Review 10.  Crowdsourcing in Surgical Skills Acquisition: A Developing Technology in Surgical Education.

Authors:  Jessica C Dai; Thomas S Lendvay; Mathew D Sorensen
Journal:  J Grad Med Educ       Date:  2017-12
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.