| Literature DB >> 28230179 |
Humayun Irshad1, Eun-Yeong Oh2, Daniel Schmolze3, Liza M Quintana1, Laura Collins1, Rulla M Tamimi4,5, Andrew H Beck1.
Abstract
The assessment of protein expression in immunohistochemistry (IHC) images provides important diagnostic, prognostic and predictive information for guiding cancer diagnosis and therapy. Manual scoring of IHC images represents a logistical challenge, as the process is labor intensive and time consuming. Since the last decade, computational methods have been developed to enable the application of quantitative methods for the analysis and interpretation of protein expression in IHC images. These methods have not yet replaced manual scoring for the assessment of IHC in the majority of diagnostic laboratories and in many large-scale research studies. An alternative approach is crowdsourcing the quantification of IHC images to an undefined crowd. The aim of this study is to quantify IHC images for labeling of ER status with two different crowdsourcing approaches, image-labeling and nuclei-labeling, and compare their performance with automated methods. Crowdsourcing- derived scores obtained greater concordance with the pathologist interpretations for both image-labeling and nuclei-labeling tasks (83% and 87%), as compared to the pathologist concordance achieved by the automated method (81%) on 5,338 TMA images from 1,853 breast cancer patients. This analysis shows that crowdsourcing the scoring of protein expression in IHC images is a promising new approach for large scale cancer molecular pathology studies.Entities:
Mesh:
Substances:
Year: 2017 PMID: 28230179 PMCID: PMC5322394 DOI: 10.1038/srep43286
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Crowdsourcing work flow for Image Labeling and Nuclei Labeling.
Figure 2Crowdsourcing application interface for Image Labeling.
The screenshot illustrates the interface for selecting the image class label.
Figure 3Crowdsourcing application interface for Nuclei Labeling.
The screenshot illustrates the interface for labeling the positive and negative nuclei separately.
Figure 4Inter-observer reliability of pathologist labels and agreement with crowd labels.
Figure 5Sensitivity analysis of crowd labels in the pilot study.
The analysis supports using 3 crowdsourced labels per image.
Comparison of three methods (Definiens, crowdsourced image labeling and crowdsourced nuclei labeling) for IHC image classification.
| Types | Methods | 3-Class Labeling | 2-Class Labeling | ||
|---|---|---|---|---|---|
| *Image Labeling | Crowd CV | 0.64 | |||
| Crowd CT | 0.68 | 0.81 | 0.61 | ||
| Crowd | 0.64 | 0.63 | 0.77 | 0.59 | |
| Crowd | 0.64 | 0.64 | 0.77 | 0.59 | |
| *Nuclei Labeling | Crowd | ||||
| Definiens | 0.70 | 0.51 | 0.81 | 0.48 | |
Crowdsourced Nuclei labeling reported higher percentage agreement (Ag) and Spearman correlation (ρ) as compared with Definiens and Crowdsourced image labeling.
Crowd performance on test questions in quiz mode and work mode.
| Crowdsourcing Jobs | Quiz Mode | Work Mode | ||
|---|---|---|---|---|
| Passed | Failed | Passed | Failed | |
| Image Labeling | 113 | 155 | 61 | 52 |
| Nuclei Labeling | 3,244 | 1,572 | 2,216 | 1,243 |
Confusion Matrix.
| Image Labels | Crowd Image Labeling | Crowd Nuclei Labeling | Definiens Labeling | Actual | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Neg | Low Pos | Pos | Neg | Low Pos | Pos | Neg | Low Pos | Pos | ||
| Neg | 174 | 127 | 145 | 72 | 125 | 214 | 452 | |||
| Low Pos | 15 | 160 | 21 | 151 | 11 | 174 | 222 | |||
| Pos | 13 | 48 | 11 | 34 | 7 | 32 | 1,179 | |||
| Predicted | 179 | 269 | 1,405 | 267 | 229 | 1,357 | 113 | 194 | 1,528 | 1,853 |
Neg represents ER Negative, Low Pos represents ER Low Positive and Pos represents ER Positive.
Figure 6Contributor trust scores analysis on Crowdsourcing jobs.
Figure 7Contributor learning effect over the period of time on image labeling job.