| Literature DB >> 34017063 |
Andrea Chatrian1,2, Richard T Colling3,4, Jens Rittscher5,6,7,8, Clare Verrill9,10,11, Lisa Browning4,12, Nasullah Khalid Alham13,14, Korsuk Sirinukunwattana13,14, Stefano Malacrino13,14,3, Maryam Haghighat13,14, Alan Aberdeen15, Amelia Monks13, Benjamin Moxley-Wyles16, Emad Rakha17, David R J Snead18.
Abstract
The use of immunohistochemistry in the reporting of prostate biopsies is an important adjunct when the diagnosis is not definite on haematoxylin and eosin (H&E) morphology alone. The process is however inherently inefficient with delays while waiting for pathologist review to make the request and duplicated effort reviewing a case more than once. In this study, we aimed to capture the workflow implications of immunohistochemistry requests and demonstrate a novel artificial intelligence tool to identify cases in which immunohistochemistry (IHC) is required and generate an automated request. We conducted audits of the workflow for prostate biopsies in order to understand the potential implications of automated immunohistochemistry requesting and collected prospective cases to train a deep neural network algorithm to detect tissue regions that presented ambiguous morphology on whole slide images. These ambiguous foci were selected on the basis of the pathologist requesting immunohistochemistry to aid diagnosis. A gradient boosted trees classifier was then used to make a slide-level prediction based on the outputs of the neural network prediction. The algorithm was trained on annotations of 219 immunohistochemistry-requested and 80 control images, and tested by threefold cross-validation. Validation was conducted on a separate validation dataset of 222 images. Non IHC-requested cases were diagnosed in 17.9 min on average, while IHC-requested cases took 33.4 min over multiple reporting sessions. We estimated 11 min could be saved on average per case by automated IHC requesting, by removing duplication of effort. The tool attained 99% accuracy and 0.99 Area Under the Curve (AUC) on the test data. In the validation, the average agreement with pathologists was 0.81, with a mean AUC of 0.80. We demonstrate the proof-of-principle that an AI tool making automated immunohistochemistry requests could create a significantly leaner workflow and result in pathologist time savings.Entities:
Mesh:
Year: 2021 PMID: 34017063 PMCID: PMC8376647 DOI: 10.1038/s41379-021-00826-6
Source DB: PubMed Journal: Mod Pathol ISSN: 0893-3952 Impact factor: 7.842
Fig. 1Examples of the classification of ambiguous prostate glands that prompt IHC ordering devised in this study (cf. Table 1).
Reason 1: An H&E section showing a short length of glands (e.g., <1 mm) that the pathologist is confident of calling cancer morphologically but wishes to confirm with IHC (A). This case was confirmed as cancer on CK5 staining (B). Reason 2: A small focus of glands that are only suspicious of are cancer (C). In this case, IHC confirmed cancer (D). Reason 3: Foci of glands with an unusual morphology and that the diagnosis would be uncertain (E). The CK5 in this case (F) demonstrated basal cell (brown) and the focus was deemed benign. Reason 4: A longer length of cancer (G) that is very well differentiated and needs to be confirmed cancer with IHC (H). Reason 5: Foci of glands that look atrophic but show atypia (I). In this example a few glands lack basal cells (J) but were considered benign as being admixed with this otherwise partially atrophic group. Reason 6: Small suspicious glands around PIN (K), in this example all were deemed benign (L). Reasons 7 and 8: not demonstrated here. These are cases used in this study and in some instances are annotated.
Fig. 2Prostate Cancer Diagnosis and Data Collection Workflows.
a Schema of typical workflow for diagnosis of prostate cancer from H&E-stained needle biopsies adopted by pathologists in the hospital. b Workflow after introduction of our tool for advance IHC requesting. The tool scans the H&E slide and requests IHC automatically, which is processed immediately after the H&E slide is scanned. The introduction of the IHC tool speeds up sign out of the case report. c Datasets used for training, testing and validating the network. The dataset used for training and testing the network is a snapshot of standard clinical decisions made independently by pathologists. The validation dataset was annotated retrospectively and blindly to the archived data original labels. The procedure replicates the decision-making that occurs in the hospital.
Reasons for ordering IHC staining on prostate core biopsy tissue slides with rates for each and resulting rates of malignancy. See also Fig. 1 for examples.
| Reason | Description | Number of foci identified | Number that were called cancer after IHC |
|---|---|---|---|
| 1 | A short length of cancer (e.g., <1 mm) of any glands that show convincing cancer morphologically but need confirming with IHC for completeness | 187 | 168 |
| 2 | Foci that are suspicious of cancer but only consist of a couple of glands such that we are unlikely to definitively call the focus cancer, but if lack basal cells we may consider a differential diagnosis of ASAP | 67 | 20 |
| 3 | Unusual morphology glands that are difficult to classify | 100 | 7 |
| 4 | A longer length of cancer which has an unusual appearance (e.g., well differentiated cancer or cancer with very few well differentiated glands widely spaced by benign/stroma.) Or variant/unusual tumour (e.g., clear cell change/atrophic variant/PIN like adenocarcinoma) | 44 | 42 |
| 5 | Foci that are atypical but probably benign (e.g., atrophy) | 144 | 6 |
| 6 | Small glands around PIN? small foci of invasion or ASAP/PIN | 64 | 6 |
| 7 | Diffuse cells where the differential diagnosis lies between inflammation or Gleason pattern 5 cancer | 32 | 0 |
| 8 | Other—glands that do not easily fit into the above | 3 | 0 |
Fig. 3Illustration of the IHC requesting algorithm.
a Conceptual of IHC requesting. The Venn diagram illustrates the relationship between the benign vs malignant nature of tissue and the diagnostic need for IHC. Most tissue presents clearly benign (e.g. regularly shaped glands) or clearly malignant morphology (e.g. amorphous tumour sheets). These cases cause relatively low diagnostic uncertainty, and IHC is not required to make a diagnosis (“certain” cases). A portion of benign and malignant cases presents morphology that cannot easily be placed in either class. Such “ambiguous” cases cause higher diagnostic uncertainty and require IHC for a diagnosis to be made. b, c Decision-making algorithm for ordering IHC for uncertain PCa cases. In part (b), a deep neural network ensemble is trained to recognise ambiguous epithelial morphology in tiles. The DNN also learns to compute representative features. In row (c) the morphology of whole slide images is characterised through the distribution extracted deep features, a gradient boosting classifier is trained to predict which cases require IHC for diagnosis.
Fig. 4Example application of the tile classifier on the tissue areas of an unseen whole H&E slide (from the validation dataset).
In a–d, the results are compared to foci annotated by pathologists. Red denotes tiles classified as certain, while blue denotes tiles classified as ambiguous. a–c Show the overlap between pathologist annotation and the classifier output. The classifier marks as ambiguous regions that were identified by pathologists. d is an example of the tile classifier output for a WSI.
(a) Tile-classifier metrics for the task of classifying foci of interest on H&E slides, and slide-level classifier metrics for the IHC request prediction task, for the three cross-validation splits of the dataset. (b) Slide-level classifier results for the task of IHC requesting evaluated on the validation set, for the three annotating pathologists. (c) Estimation of time savings and extra costs from introduction of the model (n = 380 IHC-requested cases).
| (a) | |||||
|---|---|---|---|---|---|
| Split number | Tile classification accuracy | Tile classification AUC | IHC order accuracy | IHC order AUC | |
| 0 | 0.86 | 0.91 | 0.99 | 0.99 | |
| 1 | 0.85 | 0.94 | 0.99 | 0.99 | |
| 2 | 0.91 | 0.93 | 0.99 | 0.99 | |
The false positive rate is averaged over the three pathologists. The minimum expected savings of 3 days 2 h for turnaround time and 11 min for reporting times were used in the calculations. We assume an IHC order cost of £11 per slide.
Fig. 5ROC curve for the tool vs the three pathologists (validation set).
Turnaround and reporting time savings, and extra incurred costs, for the three chosen operating points (at the 0.6, 0.7, and 0.9 values of specificity), for the 380 IHC-requested cases in the retrospective audit.
Fig. 6Results of two-step classification. The first step is the tile-level classifier.
a shows salience maps (blue for ambiguous and red for certain cases), which highlight some morphological characteristics used by the tile-level classifier to distinguish between the two classes (obtained through guided-backpropagation [30]). The second step is the slide-level IHC request decision. b, c The manifold of slide feature vectors projected onto its principal components. Each data point is the feature vector summarising tissue content for one slide. Feature vectors are labelled according to the dataset of origin (b) and the IHC request status (c).