| Literature DB >> 35014198 |
Noorul Wahab1, Islam M Miligy2,3, Katherine Dodd4, Harvir Sahota4, Michael Toss2, Wenqi Lu1, Mostafa Jahanifar1, Mohsin Bilal1, Simon Graham1, Young Park1, Giorgos Hadjigeorghiou1, Abhir Bhalerao1, Ayat G Lashen2, Asmaa Y Ibrahim2, Ayaka Katayama5, Henry O Ebili2, Matthew Parkin2, Tom Sorell6, Shan E Ahmed Raza1, Emily Hero4,7, Hesham Eldaly4, Yee Wah Tsang4, Kishore Gopalakrishnan4, David Snead4, Emad Rakha2, Nasir Rajpoot1, Fayyaz Minhas1.
Abstract
Recent advances in whole-slide imaging (WSI) technology have led to the development of a myriad of computer vision and artificial intelligence-based diagnostic, prognostic, and predictive algorithms. Computational Pathology (CPath) offers an integrated solution to utilise information embedded in pathology WSIs beyond what can be obtained through visual assessment. For automated analysis of WSIs and validation of machine learning (ML) models, annotations at the slide, tissue, and cellular levels are required. The annotation of important visual constructs in pathology images is an important component of CPath projects. Improper annotations can result in algorithms that are hard to interpret and can potentially produce inaccurate and inconsistent results. Despite the crucial role of annotations in CPath projects, there are no well-defined guidelines or best practices on how annotations should be carried out. In this paper, we address this shortcoming by presenting the experience and best practices acquired during the execution of a large-scale annotation exercise involving a multidisciplinary team of pathologists, ML experts, and researchers as part of the Pathology image data Lake for Analytics, Knowledge and Education (PathLAKE) consortium. We present a real-world case study along with examples of different types of annotations, diagnostic algorithm, annotation data dictionary, and annotation constructs. The analyses reported in this work highlight best practice recommendations that can be used as annotation guidelines over the lifecycle of a CPath project.Entities:
Keywords: annotations; computational pathology; guidelines; whole-slide images
Mesh:
Year: 2022 PMID: 35014198 PMCID: PMC8822374 DOI: 10.1002/cjp2.256
Source DB: PubMed Journal: J Pathol Clin Res ISSN: 2056-4538
Figure 1Manual versus automated process of histopathology image‐based diagnosis/prognosis. The dotted arrows show the manual process, whereas the solid arrows show the steps involved in automating the process.
Figure 2Proposed annotation workflow for a CPath project.
Figure 3The four proposed levels of annotation.
Proposed annotation QC metrics.
| Matric name | Purpose | Unit |
|---|---|---|
| Completeness | Are the annotations complete according to the defined protocol? | Yes/no |
| Exhaustiveness | What percentage of tissue is annotated in the defined box(es)? | Percentage area |
| Diversity | How many types of regions are annotated? | 1 to number of defined types in the protocol |
| Agreement |
How much the annotators agree on regions? How much the annotators agree on cells? |
Jaccard similarity index Cohen's kappa |
Figure 4Proposed annotation QC steps.
Figure 5A diagnostic algorithm for assigning a grade to a breast cancer histopathology image. T, tubule formation; P, nuclear pleomorphism; M, mitotic count; HPF, high‐power field.
Figure 6(A) An example of annotation variability between two pathologists (A1, tumour‐associated stroma; A2, tumour). Annotations by pathologist 1 (blue) and pathologist 2 (yellow). (B) Mean percentage of cells (B1) on which two pathologists agreed (B2), disagreed (B3), and missed by one pathologist (B4) in breast H&E cell annotations.