Literature DB >> 32195366

Pan-cancer diagnostic consensus through searching archival histopathology images using artificial intelligence.

Shivam Kalra1,2, H R Tizhoosh2,3, Sultaan Shah1, Charles Choi1, Savvas Damaskinos1, Amir Safarpoor2, Sobhan Shafiei2, Morteza Babaie2, Phedias Diamandis4, Clinton J V Campbell5,6, Liron Pantanowitz7.   

Abstract

The emergence of digital pathology has opened new horizons for histopathology. Artificial intelligence (AI) algorithms are able to operate on digitized slides to assist pathologists with different tasks. Whereas AI-involving classification and segmentation methods have obvious benefits for image analysis, image search represents a fundamental shift in computational pathology. Matching the pathology of new patients with already diagnosed and curated cases offers pathologists a new approach to improve diagnostic accuracy through visual inspection of similar cases and computational majority vote for consensus building. In this study, we report the results from searching the largest public repository (The Cancer Genome Atlas, TCGA) of whole-slide images from almost 11,000 patients. We successfully indexed and searched almost 30,000 high-resolution digitized slides constituting 16 terabytes of data comprised of 20 million 1000 × 1000 pixels image patches. The TCGA image database covers 25 anatomic sites and contains 32 cancer subtypes. High-performance storage and GPU power were employed for experimentation. The results were assessed with conservative "majority voting" to build consensus for subtype diagnosis through vertical search and demonstrated high accuracy values for both frozen section slides (e.g., bladder urothelial carcinoma 93%, kidney renal clear cell carcinoma 97%, and ovarian serous cystadenocarcinoma 99%) and permanent histopathology slides (e.g., prostate adenocarcinoma 98%, skin cutaneous melanoma 99%, and thymoma 100%). The key finding of this validation study was that computational consensus appears to be possible for rendering diagnoses if a sufficiently large number of searchable cases are available for each cancer subtype.
© The Author(s) 2020.

Entities:  

Keywords:  Cancer imaging; Data mining; Machine learning

Year:  2020        PMID: 32195366      PMCID: PMC7064517          DOI: 10.1038/s41746-020-0238-2

Source DB:  PubMed          Journal:  NPJ Digit Med        ISSN: 2398-6352


Introduction

Digital pathology is the virtual version of conventional microscopy utilized for the examination of glass pathology slides. In recent years, there has been accelerated adoption of digital pathology, whereby pathology laboratories around the world are slowly beginning to trade in their light microscopes for digital scanners, computers, and monitors. As a result, the pathology community has begun to scan many slides resulting in the creation of large databases of whole-slide images (WSIs). The emergence of deep learning and other artificial intelligence (AI) methods and their impressive pattern-recognition capabilities when applied to these digital databases has immensely added to the value proposition of digital pathology[1-3]. Computerized operations, such as segmentation of tissue fragments and cell nuclei, and classification of diseases and their grades become possible after pathology slides are digitized. These operations could assist with many diagnostic and research tasks with expert-like accuracy when trained with the proper level of labeled data[4]. The majority of recent studies in digital pathology have reported the success of supervised AI algorithms for classification and segmentation[4-7]. This overrepresentation compared with other AI algorithms is related to the ease of design and in-lab validation to generate highly accurate results. However, compared with other methods of computer-vision algorithms, AI-based image search and retrieval offers a new approach to computational pathology. Content-based image search[8-11] implies that the input for search software is not text (e.g., disease description in a pathology report), but rather the input is an image such that the search and retrieval can be performed based on image pixels (visual content). Content-based image search is inherently unsupervised, which means that its design and implementation may not need manual delineation of a region of interest in the images[12-14]. More importantly, image search does not make any direct diagnostic decision on behalf of the pathologist; instead, it searches for similar images and retrieves them along with the corresponding metadata (i.e., pathology reports), and displays them to the pathologist as decision support. Variability in the visual inspection of medical images is a well-known problem[15-17]. Both inter- and intra-observer variability may affect image assessment and subsequently the ensuing diagnosis[18-21]. A large body of work have reported high rates of diagnostic inaccuracy as a result of major discordance among participating physicians with respect to case target diagnoses, and propose a combination of “routine second opinions” and “directed retrospective peer review”[22-24]. As most proposed AI-driven solutions for digital pathology mainly focus on the concept of classification, it appears that algorithmic decision-making may not necessarily contribute to supporting concordance by providing a framework for consensus building. Most capable classification schemes trained with immense effort are supposed to be used for triaging cases in the pathology laboratory, and not for direct assistance in the pathologist’s office[4]. In contrast, instantly retrieving multiple diagnosed cases with histopathologic similarity to the patient’s biopsy about to be diagnosed offers a new generation of decision support that may even enable “virtual” peer review. Content-based image retrieval (CBIR) systems have been under investigation for more than two decades[25-27]. Recently, deep learning has gained a lot of attention for image search[28-30]. While CBIR systems of medical images have been well researched[11,31-33], only with the emergence of digital pathology[34,35] and deep learning[3,36,37] has research begun to focus on image search and analysis in histopathology[2,38-40]. In the past 3 years, an image search engine called Yottixel has been designed and developed for application in pathology[32,41-43]. Yottixel is a portmanteau for one yotta pixel alluding to the big-data nature of pathology images. The underlying technology behind Yottixel consists of a series of AI algorithms, including clustering techniques, deep networks, and gradient barcoding. By generating a “bunch of barcodes” (BoB) for each WSI, digitized pathology slides can be indexed for real-time search. In other words, the tissue patterns of a WSI are converted into barcodes, a process that is both storage-friendly and computationally efficient. In this paper, we report the outcome of a comprehensive validation of the Yottixel search engine. We used WSI data from The Cancer Genome Atlas (TCGA) repository provided by the National Cancer Institute (NCI)/National Institutes of Health (NIH). Almost 30,000 WSI files of 25 primary anatomic sites and 32 cancer subtypes were processed by dismantling these large slides into almost 20,000,000 image patches (also called tiles) that were then individually indexed employing ~3,000,000 barcodes. We employ the largest publicly available archive of WSIs to verify the performance of an image search engine for digital pathology.

Results

Performance measurement of search engine

In two major series of experiments, we calculated the “accuracy” of image search through “leave-one-patient-out” samplings. Whereas the literature of computer vision focuses on top-n accuracy (if any one of the n search results is correct, then the search is considered be to be successful), we calculated the majority-n accuracy (only if the majority among n search results were correct, the search was considered correct). Specifically, “correct” means that the tumor type (horizontal search) or tumor subtype within a specific diagnostic category (vertical search) was recognized correctly and matched by the majority of identified and retrieved cases. In order to avoid falsification of results through anatomic duplicates, we excluded all WSIs of the patient when one of the WSIs was the query.

Horizontal search: cancer-type recognition

The first series of experiments undertaken for all anatomic sites was horizontal search. The query WSI is compared against all other cases in the repository, regardless of anatomic site categorization. Of course, the primary anatomic site is generally known, and, in many cases, the cancer type may also be known to the pathologist. Thus, the purpose of the horizontal search (which is for either organ or cancer-type recognition) is principally a fundamental algorithmic validation that may also have applications like searching for origin of malignancy in case of metastatic cancer. The results of the horizontal search are depicted in Fig. 1 (see Appendix for details with Table 1 showing results for frozen section and Table 2 for permanent diagnostic slides). All experiments were conducted via “leave-one-patient-out” validation.
Fig. 1

Horizontal search for frozen sections (top) and permanent diagnostic slides (bottom).

Details are demonstrated in Tables 1 and 2 in the Appendix.

Table 1

Results for cancer-type recognition (horizontal search) among frozen slides.

Tumor typeWSI countPatient countHit rate (%)Majority-5Majority-10
Top-10Top-5Top-3AccuracyRecallAccuracyRecall
Brain1797108397.4495.2192.7682.2486.3783.8686.42
Gynecological2216145097.6093.5088.2267.9677.0368.8678.97
Pulmonary1634106895.3490.7583.9058.0165.6159.3067.99
Gastrointestinal tract1947121295.1287.9881.8661.3268.1662.8668.98
Breast1495107593.4488.5683.8765.6174.4566.3577.46
Prostate/testis75563491.9287.2884.6366.2274.3068.0773.77
Urinary tract1980130090.2583.4879.8962.6768.8964.5967.83
Endocrine76972984.7871.3961.8930.6844.0835.3743.56
Melanocytic malignancies53252983.8368.7957.5125.9339.8529.1339.85
Liver, pancreaticobiliary65960281.4873.2963.7330.3444.6135.3543.55
Hematopoietic18116978.4573.4869.0644.1955.2545.8549.17
Head and neck66346570.8857.1648.1122.3229.5626.2427.75
Mesenchymal25925556.3742.8533.5906.1716.2211.1915.44

Every whole-slide image was compared with all other slides in the repository regardless of the primary site. The table is sorted based on Top-10 hit rates. The accuracy and recall (sensitivity) for majority-5 and majority-10 among search results are provided as well.

Table 2

Results for cancer-type recognition (horizontal search) among diagnostic slides.

Tumor TypeWSI countPatient countHit rate (%)Majority-5Majority-10
Top-10Top-5Top-3AccuracyRecallAccuracyblackRecall
Brain169287098.9997.8196.6991.3794.3391.6094.80
Pulmonary1109101198.4696.1291.7075.8384.5876.1986.29
Prostate/testis70155097.4394.8692.1580.3186.7382.8885.31
Breast1116104995.9691.5787.0970.8778.7971.5078.61
Gastrointestinal tract1144110895.5490.7385.8365.1274.2567.9174.59
Urinary tract1374127595.4190.8285.5166.0174.5669.2173.84
Gynecological103993395.2890.3784.5063.7173.4066.8974.88
Endocrine93673294.5591.8888.6773.9381.4577.1381.34
Liver, pancreaticobiliary61858593.8587.3782.2063.7570.3264.7270.81
Head and neck46644690.5582.4075.9649.1460.9454.5057.94
Melanocytic malignancies55150988.2079.3170.4137.2051.9143.7352.09
Mesenchymal59425387.3780.6373.7350.8461.7853.7064.14
Hematopoietic22116384.6181.4476.4752.0364.2556.5661.09

Every whole-slide image was compared with all other slides in the repository regardless of the primary site. The table is sorted based on Top-10 hit rates. The accuracy and recall (sensitivity) for majority-5 and majority-10 among search results are provided as well.

Horizontal search for frozen sections (top) and permanent diagnostic slides (bottom).

Details are demonstrated in Tables 1 and 2 in the Appendix. Results for cancer-type recognition (horizontal search) among frozen slides. Every whole-slide image was compared with all other slides in the repository regardless of the primary site. The table is sorted based on Top-10 hit rates. The accuracy and recall (sensitivity) for majority-5 and majority-10 among search results are provided as well. Results for cancer-type recognition (horizontal search) among diagnostic slides. Every whole-slide image was compared with all other slides in the repository regardless of the primary site. The table is sorted based on Top-10 hit rates. The accuracy and recall (sensitivity) for majority-5 and majority-10 among search results are provided as well. The following observations can be made from the results: Provided there are sufficient number of patients, we observed that the more we retrieve the more likely it was to achieve the right diagnosis: top-10 is better than top-5, and top-5 is better than top-3. General top-n accuracy that is common in the computer-vision literature (top-3, top-5 and top-10 column in Tables 1 and 2) show high values, but may not be suitable in the medical domain as it considers the search to be a success if at least one of the search results has the same cancer type as the query image. The majority vote among top-n search results appears to be much more conservative and perhaps more appropriate, as it only considers a search task as successful if the majority of top-n search results show the same cancer type as the query image (majority-5 and majority-10 columns in Tables 1 and 2). With some exceptions, a general trend is observable that the more images/patients are available the higher the search-based consensus accuracy. The number of cases positively correlated with the majority-vote accuracy for both frozen sections and permanent diagnostic slides.

Vertical search: correctly subtyping cancer

In the second series of experiments, we performed vertical search. Given the primary site of the query slide we confined the search only to WSIs from that organ. Hence, the goal of the vertical search was to recognize the cancer subtype. For this purpose, only those primary anatomic sites in the data set with at least two possible subtypes were selected. Sample retrievals are illustrated in Appendix Fig. 2. The results for “leave-one-patient-out” validation are depicted in Figs 3 and 4 (details in Appendix, Table 3 for frozen sections and Table 4 for diagnostic slides).
Fig. 2

Sample retrievals for cancer subtype categorization through majority votes.

The top four slides are of permanent diagnostic slides whereas the bottom three slides are of frozen section slides. The misclassified and successful queries are marked with red and green boundaries, respectively (for abbreviations, see Table 5).

Fig. 3

Accuracy of vertical search for frozen sections.

Vertical search in frozen sections slides from different anatomic sites (a–j) with at least two cancer subtypes.

Fig. 4

Accuracy of vertical search for diagnostic slides.

Vertical search in permanent diagnostic slides from different anatomic sites (a–j) with at least two cancer subtypes.

Table 3

Accuracy and recall (sensitivity) for cancer subtype identification (vertical search) among frozen section slides.

Tumor typeWSI countPatient countMajority-5Majority-10Majority-20
AccuracyRecallAccuracyRecallAccuracyRecall
Brain
GBM110258294.1994.1992.7494.3792.9293.65
LGG69550182.5882.5980.2883.0281.0083.31
Endocrine
ACC818145.6746.9128.3948.1520.9835.80
PCPG17417085.6386.7886.2089.6683.9086.78
THCA51447897.0897.6797.4798.4497.8598.83
Gastrointestinal tract
COAD83044963.7369.4056.6274.1060.0078.43
ESCA16616525.9031.3312.0423.4909.0315.66
STAD62342871.1074.4865.4880.4267.4181.70
READ32817014.3219.2105.4814.6302.138.54
Gynecological
OV118458699.0799.2498.9899.1698.8199.16
CESC29829164.4268.1259.0665.4458.0563.42
UCS494910.2012.2404.0812.2402.042.04
UCEC68552490.0790.8089.0590.8089.3491.68
Hematopoietic
DLBC574591.2291.2380.7087.7273.6878.95
THYM12412497.5897.5895.1697.5895.1695.97
Liver, pancreaticobiliary
LIHC39237093.3693.8892.6094.6493.6294.90
CHOL515135.2945.1019.6047.0613.7227.45
PAAD21618191.6691.6790.7493.5290.7493.98
Melanocytic malignancies
SKCM46346098.7098.4998.4898.9299.5699.78
UVM696946.3746.3831.8839.1318.8427.54
Prostate/testis
TGCT15514986.4587.7483.8785.8181.2985.81
PRAD60048598.3398.3398.3398.3398.5098.67
Pulmonary
LUSC74548578.2578.7970.8777.9973.4277.72
LUAD80650068.2369.1164.1471.3466.1270.84
MESO838327.7132.5314.4526.5103.6121.69
Urinary tract
BLCA42040192.8594.2990.9594.2990.9595.00
KICH1388878.2681.1668.1177.5457.2473.19
KIRC105552997.8197.9197.2598.2097.6398.20
KIRP36728262.1267.3051.2263.7647.1358.86

Only those primary sites were considered for vertical search which had at least two subtypes in the repository. A positive correlation of 0.57 was measured between the number of patients and the highest accuracy.

Table 4

Accuracy and recall (sensitivity) for cancer subtype identification (vertical search) among permanent diagnostic slides.

Tumor typeWSI countPatient countMajority-5Majority-10Majority-20
AccuracyRecallAccuracyRecallAccuracyRecall
Brain
GBM85138191.1891.3087.8990.0188.1389.42
LGG84148989.7789.5488.5890.6189.1791.20
Endocrine
ACC2275693.8393.3994.2794.7194.7196.92
PCPG19617688.7788.7885.7190.3184.1889.29
THCA51350097.6697.8696.6896.8996.4996.70
Gastrointestinal tract
COAD43642876.1482.0069.7286.0074.3190.00
ESCA15715559.8769.4345.2264.3339.4955.41
READ15715610.1912.2000.633.6600.000.61
STAD39436975.1279.1967.7681.7367.0083.25
Gynecological
UCEC56650592.2293.9591.6995.2992.7595.97
CESC27726762.4564.0854.5164.7949.0958.80
UCS905642.2251.1132.2248.8927.7740.00
OV10610566.9867.9259.4367.9251.8862.26
Hematopoietic
DLBC434358.1353.4937.2058.1416.2727.91
THYM17812098.8798.8899.4399.44100.00100.00
Liver, pancreaticobiliary
CHOL393943.5843.5925.6435.9002.5617.95
LIHC37836493.6594.2193.6594.7494.4495.00
PAAD20118291.0493.5392.0395.0293.0399.00
Melanocytic malignancies
UVM808083.7583.7577.5082.5068.7572.50
SKCM47142999.5799.5899.5799.7999.5799.79
Prostate/testis
TGCT25414999.2199.6196.8598.8296.0696.06
PRAD44740198.4398.2198.2198.6698.4398.43
Pulmonary
LUAD52046570.9671.3563.2672.3164.4272.50
MESO867408.1312.7902.328.1400.001.16
LUSC50347281.7082.3178.1384.1083.3088.47
Urinary tract
BLCA45438495.8196.9394.2795.8393.6195.83
KIRC51651191.6693.0290.1192.4489.5392.64
KICH10810875.9282.4166.6674.0759.2570.37
KIRP29627267.2272.6453.0467.9148.3164.86

Only those primary sites were considered for vertical search which had at least two subtypes in the repository. A positive correlation of 0.49 was measured between the number of patients and the highest accuracy.

Sample retrievals for cancer subtype categorization through majority votes.

The top four slides are of permanent diagnostic slides whereas the bottom three slides are of frozen section slides. The misclassified and successful queries are marked with red and green boundaries, respectively (for abbreviations, see Table 5).
Table 5

The TCGA codes (in alphabetical order) of all 33 primary diagnoses and corresponding number of evidently diagnosed patients in the data set (TCGA = The Cancer Genome Atlas).

TCGA CodePrimary diagnosisNumber of patients
ACCAdrenocortical carcinoma86
BLCABladder urothelial carcinoma410
BRCABreast invasive carcinoma1097
CESCCervical squamous cell carcinoma and endocervical adenocarcinoma304
CHOLCholangiocarcinoma51
COADColon adenocarcinoma459
DLBCLymphoid neoplasm diffuse large B-cell lymphoma48
ESCAEsophageal carcinoma185
GBMGlioblastoma multiforme604
HNSCHead and neck squamous cell carcinoma473
KICHKidney chromophobe112
KIRCKidney renal clear cell carcinoma537
KIRPKidney renal papillary cell carcinoma290
LGGBrain lower-grade glioma513
LIHCLiver hepatocellular carcinoma376
LUADLung adenocarcinoma522
LUSCLung squamous cell carcinoma504
MESOMesothelioma86
OVOvarian serous cystadenocarcinoma590
PAADPancreatic adenocarcinoma185
PCPGPheochromocytoma and paraganglioma179
PRADProstate adenocarcinoma499
READRectum adenocarcinoma170
SARCSarcoma261
SKCMSkin cutaneous melanoma469
STADStomach adenocarcinoma442
TGCTTesticular germ cell tumors150
THCAThyroid carcinoma507
THYMThymoma124
UCECUterine corpus endometrial carcinoma558
UCSUterine carcinosarcoma57
UVMUveal melanoma80

Accuracy of vertical search for frozen sections.

Vertical search in frozen sections slides from different anatomic sites (a–j) with at least two cancer subtypes.

Accuracy of vertical search for diagnostic slides.

Vertical search in permanent diagnostic slides from different anatomic sites (a–j) with at least two cancer subtypes. Accuracy and recall (sensitivity) for cancer subtype identification (vertical search) among frozen section slides. Only those primary sites were considered for vertical search which had at least two subtypes in the repository. A positive correlation of 0.57 was measured between the number of patients and the highest accuracy. Accuracy and recall (sensitivity) for cancer subtype identification (vertical search) among permanent diagnostic slides. Only those primary sites were considered for vertical search which had at least two subtypes in the repository. A positive correlation of 0.49 was measured between the number of patients and the highest accuracy. Looking at the results of Figs. 3 and 4 (Tables 3 and 4), we can observe the following: For both frozen sections and permanent diagnostic slides, we continue to see a general trend whereby “the more patients the better” with both positive exceptions (KICH with 196 patients, and PCPG with 179 patients in Table 3) and negative exceptions (LUAD with 520 patients in Table 4). With majority-vote accuracy values for frozen sections (Table 3) in excess of 90% (KIRC, GBM, COAD, UCEC, PCPG), a search-based computational consensus appear to be possible when a large number of evidently diagnosed patients are available. With majority-vote accuracy values for diagnostic slides (Table 4) in excess of 90% (GBM, LGG, UCEC, KIRC, COAD, ACC, PCPG), a search-based computational consensus appear to be possible when a large number of evidently diagnosed patients are available. In most cases, it appeared that taking the majority of the top-7 search results provided the highest accuracy in most cases. However, the accuracy dropped drastically for subtypes with a small number of patients as we retrieved more and more images beyond six slides, as the majority in such cases were taken from incorrect cases (we do not filter any result; no threshold is used; hence, all search results are considered as valid results). Based on all observations, it seems that there is a direct relationship between the number of diagnosed WSIs in the data set and achievable consensus accuracy. For vertical search, we calculated positive correlations of for frozen sections (Table 3) and for permanent diagnostic slides (Table 4). This trend was more pronounced for horizontal search with positive correlation of for frozen sections slides (Table 1), and for permanent diagnostic slides (Table 2). In addition, the Cox-Stuart trend test[44] was used to check the upward monotonic trend of accuracy with respect to patients number. Having an increasing trend is considered as the null hypothesis for this test. The -values for the horizontal (vertical) search are 1 (0.9991) and 0.9844 (0.9713) for frozen and diagnostic slides, respectively. Since the -values are greater than the significance level (0.05), the null hypothesis is accepted. Consequently, there is a strong evidence of an upward monotonic trend.

Visualization of search results

Examining best, average, and worst cases for diagnostic slides, we randomly selected 3000 slides and visualized them using the T-distributed Stochastic Neighbor Embedding (t-SNE) method[45] (see Fig. 5). From this visualization, we can observe that several subtype groups have been correctly extracted through search (see groups a to f). We can also observe the presence of outliers (e.g., DLBC in groups a and b). The outliers may be a product of the resolution of these scans, at least in part. At 20× magnification, for example, recognizing a diffuse large B-cell lymphoma (DLBC) from other large cell, undifferentiated non-hematopoietic tumors may not always be immediately possible for pathologists. This typically requires serial sections examined at multiple magnifications with ancillary studies such as immunohistochemistry.
Fig. 5

T-distributed Stochastic Neighbor Embedding (t-SNE) visualization of pairwise distances of 3000 randomly selected diagnostic slides from six different primary sites.

These primary sites are selected to contain top, average, worst accuracy from the Table 2—lung, brain (top-2), kidney, liver (middle-2), lymph nodes, and pleura (bottom-2). Six different areas containing majority of the points from the same cancer subtype are assigned with unique alphabets—a, b, c, d, e, f. The random slides from the majority cancer subtype within each of the assigned areas are shown in Samples box (gray background). The outliers (not belonging to majority the cancer subtype or the primary site) are shown in the outliers box (red outline). For example, area a contains majority of scans from brain with glioblastoma multiforme (GBM), whereas its outliers are from lymph nodes with diffuse large B-cell lymphoma (DLBC). Without any explicit training, our technique maintains the semantic categories within the diagnostic slides as shows by the t-SNE plot of the pairwise distances. The kidney, liver, and brain form different isolated groups whereas lung, pleura, and lymph nodes are intermixed with each other.

T-distributed Stochastic Neighbor Embedding (t-SNE) visualization of pairwise distances of 3000 randomly selected diagnostic slides from six different primary sites.

These primary sites are selected to contain top, average, worst accuracy from the Table 2—lung, brain (top-2), kidney, liver (middle-2), lymph nodes, and pleura (bottom-2). Six different areas containing majority of the points from the same cancer subtype are assigned with unique alphabets—a, b, c, d, e, f. The random slides from the majority cancer subtype within each of the assigned areas are shown in Samples box (gray background). The outliers (not belonging to majority the cancer subtype or the primary site) are shown in the outliers box (red outline). For example, area a contains majority of scans from brain with glioblastoma multiforme (GBM), whereas its outliers are from lymph nodes with diffuse large B-cell lymphoma (DLBC). Without any explicit training, our technique maintains the semantic categories within the diagnostic slides as shows by the t-SNE plot of the pairwise distances. The kidney, liver, and brain form different isolated groups whereas lung, pleura, and lymph nodes are intermixed with each other.

The challenge of validating histologic similarity

One of the major benefits of using classification methods is that they can easily be validated; every image belongs to a class or not, a binary concept that can be conveniently quantified by counting the number of correctly/incorrectly categorized cases. It should be noted that through treating the image search as a classifier, we have not only used the primary diagnosis for “objective” evaluation of search results but also we are most likely ignoring some performance aspects of image search as search is a technology inherently suitable for looking at border cases and fuzziness of histologic similarity. The concept of similarity in image search is intrinsically a gradual concept (i.e., cannot be answered with a simple yes/no in many cases) and mostly a matter of degree (very similar, quite dissimilar, etc.). In addition, the similarity (or dissimilarity) between images is generally calculated using a distance metric/measure (in our case the Hamming distance[46]). The histologic similarity as perceived by pathologists may not correspond to tests where we used distance as a classification criterion. In other words, the classification-based tests that we run may be too harsh for search results and ignorant toward anatomic similarities among different organs. One of the possible ways of examining the performance of the search is to look at the heatmap[47] of the confusion matrix. The values to construct the heatmap can be derived from the relative frequency of every subtype among the top ten search results for a given subtype. A perfect heatmap would exhibit a pronounced diagonal with other cells being insignificant. Figure 6 shows the generated heatmap for all diagnostic subtypes in the data set. The ordering of subtypes along the -axis was done manually. It should be noted that our matching heatmap is not symmetrical like a correlation-based heatmap.
Fig. 6

Heatmap of re-scaled relative frequency of matched (red) and mismatched (pale) search results for each diagnosis from permanent diagnostic slides.

Re-scaling of frequencies was done through dividing each frequency by the total number of slides for each subtype.

Heatmap of re-scaled relative frequency of matched (red) and mismatched (pale) search results for each diagnosis from permanent diagnostic slides.

Re-scaling of frequencies was done through dividing each frequency by the total number of slides for each subtype.

Analysis of the heatmap

The pronounced diagonal in Fig. 6 shows that most disease subtypes have been correctly classified as they were very frequently retrieved among the top ten horizontal search results. Other obvious observations: MESO is a difficult diagnosis with almost absent diagonal values. READ and COAD build a confusion region of four squares; they are confused with each other frequently. The same observation can be made for LUAD and LUSC. The vertical values for LUAD and LUSC also show that they are present in many other searches, for instance, when we search for UESC, HNSC, and ESCA. LIHC is frequently among the search results for CHOL. For PRAD and BRCA we predominantly found PRAD and BRCA images, respectively. Of note, the observational analysis of the heatmap alone may be limited. If we cluster (group) the search result frequencies and construct the dendrograms for the relationships in order to create an advanced heatmap, we might more easily discover the benefits of the search (see Fig. 7). From there, we can observe:
Fig. 7

Recognizing structures through clustering.

Dendrograms of clustered relative search frequencies.

LGG and GBM are both glial tumors of the central nervous system. Rectum and colon cancer are gland forming tumors of the colon. Both uterine and ovarian carcinoma are grouped under gynecological. Gallbladder, stomach, and esophagus are upper gastrointestinal tumors. Adenocarcinoma and squamous cell carcinoma are both subtypes of lung tumors. Three kidney tumors appear close together.

Recognizing structures through clustering.

Dendrograms of clustered relative search frequencies. The errors (i.e., misclassifications) identified were still within the general grouping that the tumor originated from. Hence, from an image search perspective, it suggests that is it good at being close to the site of origin when it makes “classification” errors.

Chord diagram of image search

We used a chord diagram to further explore retrieved results. A chord diagram is the graphic display of the inter-relationships between numbers in a matrix. The numbers are arranged radially around a circle with the relationships between the data points generally visualized as arcs connecting the numbers/labels[48]. In Fig. 8a, the chord diagram of horizontal search (cancer-type recognition) for 11,579 permanent diagnostic slides of the TCGA data set is illustrated. We can observe the following:
Fig. 8

Horizontal search.

a Chord diagram of horizontal image search for diagnostic slides of the TCGA data set. Sample relations for (b) brain (LGG and GBM), (c) pulmonary (LAUD, LUSC, and MESO), and (d) gynecological (UCEC, UCS, and CESC). The chord diagram can be interactively viewed online: https://bit.ly/2k6g3k1.

Adenocarcinomas from several disparate organ systems match (e.g., colon, lung, stomach, and breast). This is not surprising, as adenocarcinomas formed by glandular structures of equivalent grade in most organs are morphologically similar. Certain tumors derived from the same organ are related (e.g., LGG and GBM, UCEC and CESC, and kidney RCC and KIRP). High-grade tumors from different anatomic locations appear to match (e.g., GBM and sarcoma). This may be attributed to the fact that such high-grade tumors likely display similar morphologic findings (e.g., necrosis). Squamous tumors from the head and neck and lung resemble urothelial carcinoma from the urinary bladder. In clinical practice, this differential diagnosis can be morphologically challenging to diagnose, and thus warrants the use of ancillary studies such as immunohistochemistry to determine tumor origin. Hepatocellular carcinoma and thyroid carcinoma appear to exhibit the greatest number of matches (eight to nine) to other tumor subtypes. The significance of this finding is unclear. The broad relationship demonstrated among certain tumor subtypes is unexpected (e.g., cutaneous melanoma to sarcoma, LUSC, and adenocarcinoma from several organs). Indeed, melanoma is known as the great mimicker in pathology given that these melanocytic tumors can take on many morphological appearances.

Horizontal search.

a Chord diagram of horizontal image search for diagnostic slides of the TCGA data set. Sample relations for (b) brain (LGG and GBM), (c) pulmonary (LAUD, LUSC, and MESO), and (d) gynecological (UCEC, UCS, and CESC). The chord diagram can be interactively viewed online: https://bit.ly/2k6g3k1. One has to emphasize that some relationships depicted in the chord diagram may disappear if distances are normalized and threshold applied. We did not filter any search results. No threshold was used. Hence, all search results were considered. The interactive version of TSNE plot is available online at http://dev1-kimia.uwaterloo.ca:5001/.

Discussion

The accelerated adoption of digital pathology is coinciding with and probably partly attributed to recent progress in AI applications in the field of pathology. This disruption in the field of pathology offers a historic chance to find novel solutions for major challenges in diagnostic histopathology and adjacent fields, including biodiscovery. In this study, we indexed and searched the largest publicly available data set of histopathology WSIs provided by the NIH/NCI. The question was whether one can build a computational consensus to potentially remedy the high intra- and inter-observer variability seen with diagnosing certain pathology tumors through search in a large archive of previously (and evidently) diagnosed cases. We performed a horizontal search to verify basic recognition capabilities of the image search engine. Furthermore, we performed leave-one-patient-out vertical searches to examine the accuracy of top search results for establishing a diagnostic majority for cancer subtypes. The results of this validation study show that building a computational consensus to assist pathologists with “virtual peer review” is possible if large and representative archives of well-characterized and evidently diagnosed cases are available. The ideal size of the data set appears to be in excess of several thousand patients for each primary diagnosis, and is most likely directly related to the anatomic complexity and intrinsic polymorphism of individual tissue types. Whereas one may need substantial computational power (i.e., a set of high-performance GPUs) to index a large existing repository from scratch, the usage of bunch-of-barcodes idea makes the continuous indexing and search quite feasible for any laboratory, clinic, and hospital. Since we used a mosaic (a set of patches) to represent and to retrieve WSIs, the search was guided to look for features present in multiple patches to classify the entire WSI. For detailed search, such as mitotic rates and grading applications, one needs a different data set and should also apply single-patch search to look for details. As well, regardless of implementation (e.g., onsite versus cloud), the validated search technology is completely safe toward patient-sensitive information as the barcodes do not contain any reversible information that could compromise patient privacy. Future research should look into subtype consensus for individual primary diagnoses in more details for carefully curated data sets. As well, the need for much larger curated archives in the pathology community is clearly evident, which includes additional tissue types such as hematological. Lastly, comprehensive discordance measurement for subtypes with and without computational consensus should be planned and carried out as the ultimate evidence for the efficacy of the image search as a supportive diagnostic tool. The intellectual property as well as the financial implications for related works emerging from sharing image repositories are certainly significant issues that need elaboration in future works.

Methods

Data collection

We used the publicly available data set of 30,072 WSIs from the TCGA project[49,50] (Genomic Data Commons GDC). Due to the retrospective nature of this study using only publicly available data, ethics approval was not required. All WSIs are tagged with a primary diagnosis. We removed 952 WSIs due to the following reasons: poor staining, low resolution, lack of all magnification levels in the WSI pyramid, large presence of out-of-focus regions, and/or presence of unreadable regions within an image. Most WSIs had a magnification of 20 or 40, some at lower magnifications. In total, we processed 29,120 WSIs at 20 magnification (approximately six terabytes in compressed form) for this study. The data set contains 25 anatomic sites with 32 cancer subtypes. Ten tumor types (brain, endocrine, gastrointestinal tract, gynecological, hematopoietic, liver/pancreaticobiliary, melanocytic, prostate/testis, pulmonary, and urinary tract) had more than one primary diagnoses. From the 29,120 WSIs, 26,564 specimens were neoplasms, and 2556 were non-neoplastic. A total of 17,425 files comprised frozen section digital slides, and 11,579 files were of permanent hematoxylin and eosin (H&E) sections. For the remaining 116 WSIs, the tissue section preparation was unspecified. We did not remove manual pen markings from the slides when present. The TCGA codes for all 32 cancer subtypes are provided in Table 5 in the appendix. The TCGA data set has a number of shortcomings[50]. Many of the cases are of frozen section in which tissue morphology may be compromised by frozen artifacts. Available cases may also reflect research bias in institutional biorepository collections. Furthermore, “tumors routinely subjected to neoadjuvant therapy may not have been able to be included in TCGA, because of limited availability of untreated specimens”[50]. Moreover, hematopathology is conspicuously absent from the TCGA data"set with just a few lymph nodes included. In spite of the shortcomings, the TCGA is the largest public data set that can support a pan-cancer validation of AI solutions for digital pathology. The TCGA codes (in alphabetical order) of all 33 primary diagnoses and corresponding number of evidently diagnosed patients in the data set (TCGA = The Cancer Genome Atlas).

The search algorithm

The Yottixel image search engine incorporates clustering, transfer learning, and barcodes and was used to conduct all experiments[30,32,41-43,51-54]. Before any search can be performed, all images in the repository have to be “indexed”, i.e., every WSI is catalogued utilizing a “bunch of barcodes” (BoB indexing). These barcodes are stored for later use and generally not visible to the user. This process contains several steps (Fig. 9):
Fig. 9

Yottixel image search engine: whole-slide images are segmented first to extract the tissue region by excluding the background (top block).

A mosaic of representative patches (tiles) is assembled through grouping of all patches of the tissue region using an unsupervised clustering algorithm (second block from the top). All patches of the mosaic are fed into a pretrained artificial neural network for feature mining (third block from the top). Finally, a bunch of barcodes is generated and added to the index of all WSI files in the archive (bottom block).

Tissue extraction—Every WSI contains a bright (white) background that generally contains irrelevant (non-tissue) pixel information. In order to process the tissue, we need to segment the tissue region(s), and generate a black and white image (binary mask) that provides the location of all tissue pixels as “1” (white). Such a binary mask is depicted in the top row of Fig. 9. Mosaicking—Segmented tissue now gets patched (divided into patches/tiles). These patches have a fixed size at a fixed magnification (e.g., 500 × 500 µm2 at 20 scan resolution). All patches of the WSI get grouped into a pre-set number of categories (classes) via a clustering method (we used -means algorithm[55]). A clustering algorithm is an unsupervised method that automatically groups WSI patches into clusters (i.e., groups) that contain similar tissue patterns. A small percentage (5–20%) of all clustered patches are selected uniformly distributed within each class to assemble a mosaic. This mosaic represents the entire tissue region within the WSI. A sample mosaic consisting of four patches is depicted in the second row of Fig. 9. Most WSIs we processed had a mosaic with around 70–100 patches. Feature mining—All patches of the mosaic of each WSI are now pushed through pretrained artificial neural networks (generally trained with natural images using data sets such as ImageNet[56]). The output of the network is ignored and the last pooling layers or the first connected layers are generally used as “features” to represent each mosaic patch. There could be ~1000–4000 features. The third row of Fig. 9 shows this process where the features (colored squares) are passed on to the next stage, namely BoB indexing. Bunch of barcodes—All feature vectors of each mosaic are subsequently converted into binary vectors using the MinMax algorithm[43]. This bunch of barcodes is the final index information for every query/input WSI that will be stored in the Yottixel index for future or immediate search. This is illustrated at the bottom of Fig. 9.

Yottixel image search engine: whole-slide images are segmented first to extract the tissue region by excluding the background (top block).

A mosaic of representative patches (tiles) is assembled through grouping of all patches of the tissue region using an unsupervised clustering algorithm (second block from the top). All patches of the mosaic are fed into a pretrained artificial neural network for feature mining (third block from the top). Finally, a bunch of barcodes is generated and added to the index of all WSI files in the archive (bottom block). In summary, Yottixel assigns “a bunch of barcodes” to each WSI to index the entire digital slide. The BoB indexing enables Yottixel to search a large archive of histopathology images very efficiently. The index can be easily shared among institutions if necessary. Technical details of Yottixel algorithms are described in a separate paper where its performance was tested with 2300 WSIs[41].

Reproducibility

Does image search generate the same results for the same WSI if fed into the Yottixel engine again? We ran indexing several times and the results did not change significantly. We observed slight changes in the order of search results affecting neither the hit rate nor the majority vote. The only component of our approach with some non-deterministic behavior is the K-means clustering algorithm. However, the K-means is run for as many iterations until it converges to a stable solution when we index WSIs. After a new WSI has been indexed its “bunch of barcodes” do not change anymore, and hence the same WSI as input (with unique patient ID) will generate the same results.
  32 in total

1.  Hierarchical edge bundles: visualization of adjacency relations in hierarchical data.

Authors:  Danny Holten
Journal:  IEEE Trans Vis Comput Graph       Date:  2006 Sep-Oct       Impact factor: 4.579

2.  Inter-observer variance and the need for standardization in the morphological classification of myelodysplastic syndrome.

Authors:  Keiko Sasada; Noriko Yamamoto; Hiroki Masuda; Yoko Tanaka; Ayako Ishihara; Yasushi Takamatsu; Yutaka Yatomi; Waichiro Katsuda; Issei Sato; Hirotaka Matsui
Journal:  Leuk Res       Date:  2018-04-09       Impact factor: 3.156

Review 3.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

4.  Inter-observer and intra-observer reliability for lung cancer target volume delineation in the 4D-CT era.

Authors:  Alexander V Louie; George Rodrigues; Jason Olsthoorn; David Palma; Edward Yu; Brian Yaremko; Belal Ahmad; Inge Aivas; Stewart Gaede
Journal:  Radiother Oncol       Date:  2010-02-01       Impact factor: 6.280

5.  Observer variability in the interpretation of contrast enhanced MRI of the breast.

Authors:  S Mussurakis; D L Buckley; A M Coady; L W Turnbull; A Horsman
Journal:  Br J Radiol       Date:  1996-11       Impact factor: 3.039

Review 6.  Inter-observer and intra-observer variability in hepatology.

Authors:  Betsy Winkfield; Christophe Aubé; Pascal Burtin; Paul Calès
Journal:  Eur J Gastroenterol Hepatol       Date:  2003-09       Impact factor: 2.566

7.  Second Opinion Reviews for Cancer Diagnoses in Anatomic Pathology: A Comprehensive Cancer Center's Experience.

Authors:  Carolina Strosberg; Julie Gibbs; Diana Braswell; Ronni R Leslie; Jane Messina; Barbara A Centeno; Domenico Coppola
Journal:  Anticancer Res       Date:  2018-05       Impact factor: 2.480

8.  Clinical-grade computational pathology using weakly supervised deep learning on whole slide images.

Authors:  Gabriele Campanella; Matthew G Hanna; Luke Geneslaw; Allen Miraflor; Vitor Werneck Krauss Silva; Klaus J Busam; Edi Brogi; Victor E Reuter; David S Klimstra; Thomas J Fuchs
Journal:  Nat Med       Date:  2019-07-15       Impact factor: 53.440

9.  A Fast and Refined Cancer Regions Segmentation Framework in Whole-slide Breast Pathological Images.

Authors:  Zichao Guo; Hong Liu; Haomiao Ni; Xiangdong Wang; Mingming Su; Wei Guo; Kuansong Wang; Taijiao Jiang; Yueliang Qian
Journal:  Sci Rep       Date:  2019-01-29       Impact factor: 4.379

Review 10.  Artificial Intelligence and Digital Pathology: Challenges and Opportunities.

Authors:  Hamid Reza Tizhoosh; Liron Pantanowitz
Journal:  J Pathol Inform       Date:  2018-11-14
View more
  5 in total

Review 1.  Deep Learning of Histopathology Images at the Single Cell Level.

Authors:  Kyubum Lee; John H Lockhart; Mengyu Xie; Ritu Chaudhary; Robbert J C Slebos; Elsa R Flores; Christine H Chung; Aik Choon Tan
Journal:  Front Artif Intell       Date:  2021-09-10

Review 2.  Multiplexed Immunohistochemistry and Digital Pathology as the Foundation for Next-Generation Pathology in Melanoma: Methodological Comparison and Future Clinical Applications.

Authors:  Yannick Van Herck; Asier Antoranz; Madhavi Dipak Andhari; Giorgia Milli; Oliver Bechter; Frederik De Smet; Francesca Maria Bosisio
Journal:  Front Oncol       Date:  2021-03-29       Impact factor: 6.244

3.  Human-interpretable image features derived from densely mapped cancer pathology slides predict diverse molecular phenotypes.

Authors:  James A Diao; Jason K Wang; Wan Fung Chui; Andrew H Beck; Hunter L Elliott; Amaro Taylor-Weiner; Victoria Mountain; Sai Chowdary Gullapally; Ramprakash Srinivasan; Richard N Mitchell; Benjamin Glass; Sara Hoffman; Sudha K Rao; Chirag Maheshwari; Abhik Lahiri; Aaditya Prakash; Ryan McLoughlin; Jennifer K Kerner; Murray B Resnick; Michael C Montalto; Aditya Khosla; Ilan N Wapinski
Journal:  Nat Commun       Date:  2021-03-12       Impact factor: 14.919

4.  Identification of Tissue Types and Gene Mutations From Histopathology Images for Advancing Colorectal Cancer Biology.

Authors:  Yuqi Jiang; Cecilia K W Chan; Ronald C K Chan; Xin Wang; Nathalie Wong; Ka Fai To; Simon S M Ng; James Y W Lau; Carmen C Y Poon
Journal:  IEEE Open J Eng Med Biol       Date:  2022-07-19

5.  Automated quality assessment of large digitised histology cohorts by artificial intelligence.

Authors:  Maryam Haghighat; Lisa Browning; Korsuk Sirinukunwattana; Stefano Malacrino; Nasullah Khalid Alham; Richard Colling; Ying Cui; Emad Rakha; Freddie C Hamdy; Clare Verrill; Jens Rittscher
Journal:  Sci Rep       Date:  2022-03-23       Impact factor: 4.379

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.