| Literature DB >> 23967384 |
Tyler Keay1, Catherine M Conway, Neil O'Flaherty, Stephen M Hewitt, Katherine Shea, Marios A Gavrielides.
Abstract
BACKGROUND: With the emerging role of digital imaging in pathology and the application of automated image-based algorithms to a number of quantitative tasks, there is a need to examine factors that may affect the reproducibility of results. These factors include the imaging properties of whole slide imaging (WSI) systems and their effect on the performance of quantitative tools. This manuscript examines inter-scanner and inter-algorithm variability in the assessment of the commonly used HER2/neu tissue-based biomarker for breast cancer with emphasis on the effect of algorithm training.Entities:
Keywords: Quantitative immunohistochemistry; reproducibility; whole slide imaging
Year: 2013 PMID: 23967384 PMCID: PMC3746414 DOI: 10.4103/2153-3539.115879
Source DB: PubMed Journal: J Pathol Inform
Figure 1Example of a field of view stained with a HER2/neu antibody, extracted from a whole slide image, digitized using: (a) The Aperio-CS (top), (b) The Aperio-T2 (middle), and (c) The Hamamatsu Nanozoomer (bottom) whole slide imaging systems. Images were extracted at ×20
Technical characteristics for the three whole slide imaging systems utilized in this study*
Parameter values of the membrane v9 algorithm for the quantitative assessment of HER2/neu expression, as used in this study
Classification score distribution in HER2/neu categories (1+, 2+, 3+) from pathologist panel, Algorithm 1 applied on image data from the three scanners, and Algorithm 2 applied on image data from the three scanners
Pair-wise agreement values using Kendall’s tau-beta (±SE) between algorithm classification results as well as between algorithms and scores from pathologist panel
Percent correct agreement between the scores derived from each algorithm applied to image data from the three scanners. Overall percent agreement along with percent agreement on each of the 1+, 2+, and 3+ categories are shown for each of Algorithm 1 and Algorithm 2
Percent correct agreement between the scores derived from each algorithm applied to image data from each of the three scanners and the scores from the pathologist panel. Overall percent agreement along with percent agreement on each of the 1+, 2+, and 3+ categories are shown for each of Algorithm 1 and Algorithm 2
Figure 2Example of two regions of interests, one where the majority of pathologists (four out of seven) scored it as 1+ whereas Algorithm 2 scored that case as 2+ (a), and another where the majority of pathologists (six out of seven) scored it as 2+ whereas Algorithm 2 scored it as a 1+ (b). Images were extracted at ×20
Contingency table of classification results comparing the scores of the pathologist panel, and Algorithm 1 applied to the three scanners
Contingency table of classification results comparing the scores of the pathologist panel, and Algorithm 2 applied to the three scanners