Literature DB >> 30450002

Imitating Pathologist Based Assessment With Interpretable and Context Based Neural Network Modeling of Histology Images.

Arunima Srivastava1, Chaitanya Kulkarni1, Kun Huang2, Anil Parwani3, Parag Mallick4, Raghu Machiraju1.   

Abstract

Convolutional neural networks (CNNs) have gained steady popularity as a tool to perform automatic classification of whole slide histology images. While CNNs have proven to be powerful classifiers in this context, they fail to explain this classification, as the network engineered features used for modeling and classification are ONLY interpretable by the CNNs themselves. This work aims at enhancing a traditional neural network model to perform histology image modeling, patient classification, and interpretation of the distinctive features identified by the network within the histology whole slide images (WSIs). We synthesize a workflow which (a) intelligently samples the training data by automatically selecting only image areas that display visible disease-relevant tissue state and (b) isolates regions most pertinent to the trained CNN prediction and translates them to observable and qualitative features such as color, intensity, cell and tissue morphology and texture. We use the Cancer Genome Atlas's Breast Invasive Carcinoma (TCGA-BRCA) histology dataset to build a model predicting patient attributes (disease stage and node status) and the tumor proliferation challenge (TUPAC 2016) breast cancer histology image repository to help identify disease-relevant tissue state (mitotic activity). We find that our enhanced CNN based workflow both increased patient attribute predictive accuracy (~2% increase for disease stage and ~10% increase for node status) and experimentally proved that a data-driven CNN histology model predicting breast invasive carcinoma stages is highly sensitive to features such as color, cell size, and shape, granularity, and uniformity. This work summarizes the need for understanding the widely trusted models built using deep learning and adds a layer of biological context to a technique that functioned as a classification only approach till now.

Entities:  

Keywords:  CNN; cancer histology; modeling; neural networks

Year:  2018        PMID: 30450002      PMCID: PMC6236488          DOI: 10.1177/1178222618807481

Source DB:  PubMed          Journal:  Biomed Inform Insights        ISSN: 1178-2226


Introduction

Automated classification of histology (hematoxylin and eosin stained) whole slide images (WSI) has been the subject of detailed research.[1,2] As a result, convolutional neural networks (CNNs) have recently gained steady popularity as the central technique to model and classify these images in the context of cancer.[3,4] They have shown to be precise and efficient classifiers in many different experiments pertaining to a variety of cancer subtypes. Our previous work assessing the predictive power of CNNs in the context of multiple patient attributes shows that CNN-based image models characterize disease staging more effectively than other trans-omics indicators.[5] Unfortunately, as is typical of sophisticated machine learning modeling techniques, one of the drawbacks the research community faces while utilizing CNNs for this purpose is the lack of interpretability of CNN-based features. Traditional CNNs learn incrementally from training data, generating an abstract set of features used by the network layers to classify regions in the image. These features do not have a precise translation to tissue structure, morphology or nuclei/cell organization, hence are not interpretable to clinicians or researchers who rely on the use of these indicators to characterize disease. While observing activation of features in each network layer illuminates relationships between CNN features and pathology driven features,[4] there still is a dearth of an attempt to conclusively find interpretable signatures from CNN features. This work makes an effort to extend our foundational exploration of deep histology image models by aiming to optimize the power of CNNs while mimicking the output produced by pathologists who classify histology WSIs using qualitative and observable disease-specific indicators (Figure 1).
Figure 1.

Histology slide assessment and use by (1) pathologists, whose protocol dictates honing in on a region of interest, classifying different structures in the tissue, analyzing the state and structure of cells and finally performing prediction and delivering prognosis; (2) traditional CNN modeling, which trains from input images with a corresponding label, and is used to predict the probability of a new image belonging to these labels by performing high-dimensional modeling with self engineered features; and (3) our approach to interpretable context based CNN modeling, which intelligently selects only disease-relevant input images for training and modeling, finally resulting in label prediction using CNNs as well as an observable qualitative label signature for each new image.

Histology slide assessment and use by (1) pathologists, whose protocol dictates honing in on a region of interest, classifying different structures in the tissue, analyzing the state and structure of cells and finally performing prediction and delivering prognosis; (2) traditional CNN modeling, which trains from input images with a corresponding label, and is used to predict the probability of a new image belonging to these labels by performing high-dimensional modeling with self engineered features; and (3) our approach to interpretable context based CNN modeling, which intelligently selects only disease-relevant input images for training and modeling, finally resulting in label prediction using CNNs as well as an observable qualitative label signature for each new image. The interpretability of CNNs is a daunting task. It essentially involves de-convolving sophisticated learning operations to identify and map features in existing decipherable space. Thus, our workflow, which aims to perform effective and interpretable CNN modeling, focuses on two main tasks. First, it aims to reduce noise and irrelevant variance in the training data by utilizing only disease-relevant regions within the whole slide images to perform modeling. This step is motivated by pathologists’ protocol where they demarcate regions of interest (ROIs) before analyzing tissue specimens.[6] As we model and analyze breast invasive carcinoma for the purposes of this work, we chose mitotic activity as a viable indicator for marking disease-relevant tissue region.[7,8] Second, we identify regions within the whole slide images that were most valued by the CNN when performing prediction of patient attributes (eg, American Joint Committee on Cancer [AJCC] stage). These “CNN relevant regions of interest (CNN-ROIs)” are further assessed in accordance with known image and morphology features. Both these tasks are achieved by utilizing a combination of specialized CNN architecture (AlexNet[9] with modulated parameters), methodology to visualize CNN learned weights (Class Activation Mapping—CAM[10]), and tools to extract image and morphology-specific features (pixel-wise k-means for dominant color extraction, Ilastik[11] and CellProfiler[12] for shape, size and texture assessment). The dataset used to train a model to identify mitosis is the tumor proliferation challenge (TUPAC 2016) WSI repository and patient attribute prediction is performed on and using the Cancer Genome Atlas’s Breast Invasive Carcinoma (TCGA-BRCA) histology dataset.[13] Intelligent, mitotic activity based, sampling of training data resulted in significant improvement in the automatic prediction of patient staging and node status. Extracting an interpretable signature relates three specific types of qualitative features to the regions that were most informative to the CNN predicting staging of breast invasive carcinomas. CNN relevant ROI consistently presented unique dominant (pixel-wise most frequently occurring) hues, specific shape, and size of cells and distinct morphological texture in the form of granularity and uniformity. While color/hue is an inherent property of staining different components of the tissue, aberrant cell size and shape is known to be a marker for tumor cells, and texture measures have proven to be highly distinct for histology images containing tumors and presenting specific disease subtypes.[14,15] This semantic context adds a layer of understandable characterization to CNN based models and helps identify critical components of histology images most relevant to this successful modeling.

Methods and Materials

The workflow to perform interpretable CNN modeling on breast invasive carcinoma is divided into four (4) distinct steps as visualized in Figure 2. Namely, the steps are (1) building a deep learning model to identify areas of high mitotic activity, (2) building a deep learning model to predict patients attributes (stages and node status) using intelligent sampling of training data where we retain areas of high mitotic activity only, (3) implementing a method for performing visualization and identification of “CNN relevant regions of interest (CNN-ROIs)” that are regions which were most informative to the CNN while predicting patient attributes and lastly (4) using CNN-ROIs to perform qualitative feature extraction which relates patient attributes to observable and interpretable features. Each of these steps is described in detail below in Figure 2.
Figure 2.

Workflow schematic for interpretable context-based CNN modeling of histology images. The main four steps of the workflow are as follows: (1) Build a model to select disease-relevant patches from the WSI based on tissue state; (2) perform intelligent sampling of WSI tiles for training of a deep learning model to predict patient attributes, based on whether they exhibit disease-relevant tissue state; (3) perform patient attribute prediction and extract CNN relevant regions of interest (CNN-ROIs), which were most informative to the deep learning model; and (4) assess these CNN-ROIs in terms of qualitative, observable features that associate model learning to interpretable features.

Workflow schematic for interpretable context-based CNN modeling of histology images. The main four steps of the workflow are as follows: (1) Build a model to select disease-relevant patches from the WSI based on tissue state; (2) perform intelligent sampling of WSI tiles for training of a deep learning model to predict patient attributes, based on whether they exhibit disease-relevant tissue state; (3) perform patient attribute prediction and extract CNN relevant regions of interest (CNN-ROIs), which were most informative to the deep learning model; and (4) assess these CNN-ROIs in terms of qualitative, observable features that associate model learning to interpretable features.

Predicting areas of high mitotic activity

Data and pre-processing

The data used to train a neural network to identify regions of high mitotic activity are the auxiliary dataset provided by the Tumor Proliferation Assessment Challenge 2016 (TUPAC 2016), which was one of Medical Image Computing and Computer Assisted Intervention (MICCAI) 2016 grand challenges. It consists of images from 73-breast cancer cases aggregated from three pathology centers. All cases are represented by a number of image regions stored as TIFF images, with the mitotic regions (as classified by two pathologists) annotated. The WSIs are produced at 40x magnification and at the spatial resolution of 0.25 µm/pixel. All whole slide images are tiled for parallel and faster processing. We used tile dimensions 224px x 224px to be cognizant of the structures we needed to identify while dispelling noise and artifacts. Since the average cell size in these tissues is ~40 to 100 pixels, (observationally), we concluded that 224px x 224px tile size would be apt as it would allow a minimum of ~2 to 3+ cells per tile taking into account varying distances between cells. Normalization was performed employing the widely used Macenko[16] technique of normalizing histology images for quantitative analysis. This technique has been utilized successfully during histopathology assessment.[4,17] It uses the highest varying optical density (a transformation of the RGB vector describing absorbance) in the image to create a color transform applied to all images, followed by appropriate changes to the image histogram to capture most of the intensity dynamic range.

Neural network modeling

To build a model which identifies areas of high mitotic activity, we chose to utilize the traditional AlexNet architecture (Supplementary Table 1) with a few key changes to adapt it to the nature of our problem (Supplementary Table 1). AlexNet is a popular and fundamental CNN architecture that has achieved widespread success in both traditional imaging challenges[9] and histology specific modeling.[3] Additionally, the resulting network is not extremely deep, which suits data that has a few underlying distinctive features in the presence of largely similar images. AlexNet uses rectified linear unit (ReLU) activation function to add non-linearity to the network and speed up training, and dropout instead of regularization to combat overfitting. Additionally, overlap pooling is employed to reduce the size of network. Our version of the model, the codebase, the necessary utilities and the complete trained models are available on request.

Selecting tiles presenting high mitotic activity

Utilizing the modified AlexNet architecture, and the mitosis annotated data from TUPAC2016, we built a model for identifying mitotic activity probability in tiles of histology images. This model was trained with tiles from TUPAC2016 whole slide images that were labeled according to the existence of mitotic cells within them. Upon successful training, the model was then employed to rank tiles from single whole slide images according to the probability of mitotic activity within that tile. The top 1% tiles that present high probability for mitotic activity are chosen for each whole slide image to represent disease-relevant areas for that patient.

Prediction of patient attributes based on regions of high mitotic activity

The TCGA breast cancer study[18] is a well-characterized and thoroughly comprehensive experimental study of breast invasive carcinoma.[19,20] It consists of +1000 whole slide images from tumor sites and associated clinical information detailing AJCC stage, tumor subtypes and relevant mutational status is available. This work utilizes the same high-quality 163 whole slide images (105 patients) from the TCGA-BRCA compendium that were analyzed in our previous work on trans-omics features[5] as these images were histopathologically documented by pathologists and thus had extensive clinical information available. Each image is digitized at 40x and contains upward of 10 billion pixels.[21] The TCGA-BRCA compendium images were tiled and pre-processed employing the same methodology as detailed in the section above. The mitotic activity prediction model described in the section above was employed on the 224px x 224px tiles from each TCGA-BRCA WSI and the top 1% of highly mitotic tiles are used to represent each whole slide image. Two separate models, using the AlexNet architecture as described above, were built. One using all the generated tiles (baseline) and the second using top 1% tiles showcasing mitotic activity, both training a predictor for patient staging and node status. The ensuing performance comparison ensured that the intelligent sampling of training data in accordance with disease-relevant tissue state was indeed producing superior results.

Visualizing CNN relevant ROIs

While there are multiple methods for visualizing a trained CNN’s feature weights and network filters with respect to an input image, we choose to use a method, namely—CAM,[10] which identifies, across the entire trained CNN, localized regions that contribute most to the classification task. This technique utilizes a global average pooling layer at the penultimate step of the CNN in order to identify discriminative localized regions for each class. Global average pooling enables a generalized view across all network layers of the optical cues in an image that drive the model to a certain classification. Figure 3 presents an example of the visualization mask we obtain using CAM and the mitotic activity prediction model, for a histology image tile containing mitosis. We observe that the region highlighted using the CAM visualization mask contained mitotic cells and other (non mitotic) cells were ignored. Visualization masks such as these were generated for tiles from TCGA-BRCA histology images on which the disease stage prediction model was employed. Those visualization masks hone in on regions informative to stage prediction.
Figure 3.

(A) Whole slide image tile containing mitotic cells. (B) Binary mask obtained using class activation mapping to highlight the discriminative localized region utilized by the CNN and (C) Composite tile, highlighting only the regions deemed “important” by the CNN, zeroing in on mitotic cells.

(A) Whole slide image tile containing mitotic cells. (B) Binary mask obtained using class activation mapping to highlight the discriminative localized region utilized by the CNN and (C) Composite tile, highlighting only the regions deemed “important” by the CNN, zeroing in on mitotic cells.

Image and morphological feature extraction

Once CAM enabled the visualization of CNN relevant ROIs, the concluding step involves extracting these regions and extracting qualitative features from them. We focus on three different types of features when assessing these CNN-ROIs to find a cohesive and interpretable signature that can be associated with the labels that a CNN model is aiming to predict. Namely, these three features are (a) color/hue, (b) cell size and shape, and (c) image texture. The procedure for extracting these features from CNN-ROIs is outlined below.

Finding dominant colors in CNN-ROI

The protocol for assessing histology images is highly dependent on the visible colors in the image (different colors of the staining mark for different structures within the tissue). It stands to reason that dominant colors visible in the CNN-ROIs evidence the predominance of a certain structure relevant to the CNN modeling. The dominant colors are extracted from an image by utilizing unsupervised k-means clustering across the RGB vectors of all the pixels of an image.[22-24] With k = 4(accounting for distinct visible colors observed across a sampling of whole slide images), we extract clusters consisting of the RGB vectors for each pixel. The color corresponding to the largest cluster’s centroid is then deemed the dominant color in the image. The method, corresponding sources, and our codebase is available on request. By identification of cells, tissues, and gaps in TCGA-BRCA whole slide images and subsequent extraction of RGB vectors from a sampling of these areas, we closely approximated the main colors visible and their corresponding RGB vectors. By euclidean distance-based proximity to these RGB vectors, we classify the dominant color to either be “purple” (cells), “pink” (muscle), or “white” (gaps or artifacts).

Assessing cell size and shape in CNN-ROI

Cells and their attributes are known to be relevant for pathologists to study and grade histology samples. We perform cell-specific segmentation in the tissue and analyze size and shape characteristics of the cells present. This is achieved by a combination of cell segmentation (performed by Ilastik) and object detection (performed by CellProfiler). Ilastik is a tool for image classification and segmentation. We train an Ilastik model on a subset of tiles from available whole slide images, where we manually demarcate cell regions. Ilastik then uses features based on color, intensity, and brightness and a random forest classifier to label pixels of the image if they are predicted as belonging to a cell. Using the results from Ilastik prediction, we extract a binary mask of the image, which identifies cell regions. This mask can be now used with CellProfiler and the “IdentifyPrimaryObjects” and “MeasureObjectSizeShape” module to extract size and shape features from the identified objects. The identification of objects is performed using a 2-class Otsu thresholding on the binary mask. These features include area, compactness, eccentricity, form factor, and Zernike features. For each CNN-ROI image, we extract the mean, median, and standard deviation of all of these features across the objects identified in the image for downstream analysis. This aggregation tuple is useful as due to the aberrant shape of tumor cells, the dynamic ranges of these features are highly relevant and distinctive. A full list of the relevant features and their descriptions is available in Supplementary Table 2.

Assessing CNN-ROI texture

Finally, to complete the qualitative signature for CNN-ROI, we extract the texture features of these images. This is also performed using a CellProfiler pipeline with the help of the module “MeasureTexture” and “MeasureGranularity.” We perform texture extraction after separating the native CNN-ROI image to Red, Green and Blue channels, using the “ColortoGray” module in CellProfiler. Features include well-characterized texture features such as Haralick features. Similar to size and shape features, we aggregate them for each histology image using mean, median and standard deviation due to the varying dynamic ranges of each feature. A full list of the relevant features and their descriptions is available in Supplementary Table 3.

Results

This work and the resulting exploration can be divided to four distinct findings (a) successful predictions of enhanced mitotic activity in whole slide image tiles, (b) prediction of patient attributes using a model built with selected tiles that display evidence of mitotic activity, (c) isolating the tile regions discriminative for each class using class activation mappings, and (d) performing morphological assessment of selected regions to extract interpretable signatures for each class. The sections below highlight the main results for each aforementioned section.

Prediction of regions containing high mitotic activity

The AlexNet architecture was utilized to build a model to predict probability of a histology image tile containing mitotic activity. This model utilized the training data from the annotated whole slide images from the TUPAC16 challenge. As mentioned previously, the model trains on tiles of size 224px x 224px, and the labels are generated based on the existence of mitotic activity on the slide. This model seemed to successfully isolate mitotic activity within tiles as evidenced by all performance measures (~82% Accuracy, precision, recall and F-score) when employed on the testing subset of the dataset (20% of all tiles). A subset of tiles was also presented to a pathologist and the correct identification of mitosis was verified.

Patient attribute prediction

Patient stage and node status prediction models were trained with all generated tiles in TCGA-BRCA histology set as well as when trained with selected 1% tiles presenting mitotic activity for each whole slide image. For the prediction of patient stages, between a model trained from all tiles and a model trained from selected tiles presenting high mitotic activity, accuracy increased from 42.67% to 44.73%, and precision, recall, and F-Score increased from 25.09% to 26.99%, 24.26% to 27.13%, and 24.45% to 26.69%, respectively. Displaying a similar trend, a model predicting node status in these patients, utilizing all tiles, compared to a model trained on select tiles showed an increase in accuracy, precision, recall, and F-Score (28.64% to 38.17%, 24.09% to 28.86%, 23.02% to 28.19%, and 22.29% to 28.1%, respectively). As we observe from these results, the approach of intelligently sampling training data based on relevant tissue state (eg, mitotic activity) is justified as it shows marked improvement in the performance of a prediction model. We can hypothesize that this reduces noise from artifacts as well as ignores non-tumor areas and hones in on regions that pathologists would ideally focus on when assigning attributes to a whole slide image, and consequently a patient.

Isolation and visualizing discriminative localized pixel regions

As described above, using the CAM technique, visualization masks were generated for each tile that highlighted the ROI to the CNN. The tiles selected by the mitotic activity predictor are used to train and test a CNN model to predict staging and for each prediction, an associated ROI mask is generated. The technique was tested by employing the CAM technique to the mitosis-predicting model as well, and on observation, we confirmed that the ROI masks were highlighting mitotic cells and ignoring typical circular non-mitotic cells. ROI masks are used to generate CNN-ROI images and qualitative features are extracted from these regions to assign an interpretable signature to the same.

Interpretable signatures of CNN-ROIs

In the penultimate step of this analysis, once we obtain the regions that were important to the CNN when defining the final prediction for patient stage, we can assess these regions interpretably to isolate signatures that represent the model and its encompassing labels. While multiple observable facets are available for exploration of the CNN-ROIs, we focus primarily on three aspects of these patches. Namely, (1) dominant colors, (2) cell size and shape, and (3) texture features. Utilizing the tools, as described in the methods section above, we assess these tuple attributes for both CNN-ROI and non-ROI patches for each tile. Comparing the two facilitates the identification of the unique signatures as distinguished by the CNN. Figure 4 presents an overall comparison of all three qualitative features, between the CNN-ROI and non-ROI patches from a sampling of 10,000 tiles, spanning 10 patients and multiple stages.
Figure 4.

Results for the qualitative feature extraction (dominant color, cell size and shape, and image texture) of a random sampling of WSI tiles (~10 000 tiles across 10 patients, spanning multiple stages). Each heatmap compares CNN-ROI image and its inverse (Non CNN ROI) across all tiles. About 54% of tiles show dominant color changes from between CNN-ROI and Non CNN ROI images. Zernike features, orientation and area are distinct for CNN ROI cells. Uniformity and Granularity are image texture features that characterize CNN-ROI Images.

Results for the qualitative feature extraction (dominant color, cell size and shape, and image texture) of a random sampling of WSI tiles (~10 000 tiles across 10 patients, spanning multiple stages). Each heatmap compares CNN-ROI image and its inverse (Non CNN ROI) across all tiles. About 54% of tiles show dominant color changes from between CNN-ROI and Non CNN ROI images. Zernike features, orientation and area are distinct for CNN ROI cells. Uniformity and Granularity are image texture features that characterize CNN-ROI Images.

Dominant colors comparison between CNN-ROI and Non-ROI

In over 54% of the samples tiles, the dominant colors between CNN-ROI and Non ROI patches were distinct. A majority of these listed “purple” as the dominant color in the CNN-ROI patch and “pink” or “white” as the dominant color in the corresponding non-ROI patch. This provides evidence for the fact that hyperchromaticity of cells is a factor of distinction during the decision-making of the CNN patient staging model.

Cell size and shape

Specific features relating to shape, area, and orientation were observed to be distinct (twofold change) between CNN-ROI and non-ROI patches for each tile across the sampled set. These features included—standard deviation of compactness (cells in CNN-ROI images have highly varying numbers of close and well-defined enclosed structures), mean, median and standard deviation of cell area (cells in CNN-ROI images are bigger and vary more in terms of area), standard deviation of minimum and maximum feret diameter (cells shape varies more in CNN-ROI images) and multiple moments of Zernike shape features. This provides evidence for the hypothesis that pleomorphic, aberrant, atypical, and large cells characterize patient staging according to the predictive CNN model.

Image texture

Lastly, the texture measures (multiple features of granularity and texture angular second momentum (ASM)[25] describing structure and uniformity of texture respectively, were consistently distinct (twofold change) across CNN ROI versus non-ROI patches. Consistently, all different moments of ASM, which describes uniformity, are drastically lower in CNN-ROIs versus the non-ROIs, which present high uniformity. This is consistent with our previous findings, as pleomorphic cells contained within CNN-ROI patches are not well ordered, which would result in this texture feature presenting lower values. Granularity on the other hand, showcases the opposite trend (higher in CNN-ROI vs. non- ROI), as it describes the size distribution of objects across a certain pixel scale, which provides further evidence that there is a higher concentration of cells within the CNN-ROI patches than the non-ROI patches. These findings are consistent with previous results and further establish the qualitative feature signature built for this model.

Discussion

The goal of this manuscript was to understand and interpret the unique signatures between subtypes of whole slide images, as understood and interpreted by a CNN model. We wished to quantify, whether controlling the input and assessing the output manifests in better performance and understanding of deep histology models. To this end, we identified meaningful regions in whole slide images by automatically classifying tissue state (high mitotic activity), which is a crucial facet of histologically assessing breast cancer. Following which, we performed experiments which predicted staging with the selected neural network model, using all tiles and only highly mitotic tiles. The performance enhancement in prediction confirmed our correct selection of whole slide image patches. Lastly, we used these selected tiles and ROIs as identified by the deep learning model to explore and understand the exact features of regions the CNN deemed interesting, and on which it based its predictions. We believe this work will enable the community to better understand the high dimensional neural network models that have slowly become the standard in automatic histology modeling. Additionally, it has the potential to identify new histopathological features that are markers of disease as understood by data driven deep modeling. Click here for additional data file. Supplemental material, Supplementary_Table_1 for Imitating Pathologist Based Assessment With Interpretable and Context Based Neural Network Modeling of Histology Images by Arunima Srivastava, Chaitanya Kulkarni, Kun Huang, Anil Parwani, Parag Mallick and Raghu Machiraju in Biomedical Informatics Insights Click here for additional data file. Supplemental material, Supplementary_Table_2 for Imitating Pathologist Based Assessment With Interpretable and Context Based Neural Network Modeling of Histology Images by Arunima Srivastava, Chaitanya Kulkarni, Kun Huang, Anil Parwani, Parag Mallick and Raghu Machiraju in Biomedical Informatics Insights Click here for additional data file. Supplemental material, Supplementary_Table_3 for Imitating Pathologist Based Assessment With Interpretable and Context Based Neural Network Modeling of Histology Images by Arunima Srivastava, Chaitanya Kulkarni, Kun Huang, Anil Parwani, Parag Mallick and Raghu Machiraju in Biomedical Informatics Insights
  17 in total

1.  Mitosis detection in breast cancer pathology images by combining handcrafted and convolutional neural network features.

Authors:  Haibo Wang; Angel Cruz-Roa; Ajay Basavanhally; Hannah Gilmore; Natalie Shih; Mike Feldman; John Tomaszewski; Fabio Gonzalez; Anant Madabhushi
Journal:  J Med Imaging (Bellingham)       Date:  2014-10-10

2.  Histology image analysis for carcinoma detection and grading.

Authors:  Lei He; L Rodney Long; Sameer Antani; George R Thoma
Journal:  Comput Methods Programs Biomed       Date:  2012-03-20       Impact factor: 5.428

3.  Cancer Digital Slide Archive: an informatics resource to support integrated in silico analysis of TCGA pathology data.

Authors:  David A Gutman; Jake Cobb; Dhananjaya Somanna; Yuna Park; Fusheng Wang; Tahsin Kurc; Joel H Saltz; Daniel J Brat; Lee A D Cooper
Journal:  J Am Med Inform Assoc       Date:  2013-07-25       Impact factor: 4.497

Review 4.  The Cancer Genome Atlas (TCGA): an immeasurable source of knowledge.

Authors:  Katarzyna Tomczak; Patrycja Czerwińska; Maciej Wiznerowicz
Journal:  Contemp Oncol (Pozn)       Date:  2015

5.  Identification of Gene-Expression Signatures and Protein Markers for Breast Cancer Grading and Staging.

Authors:  Fang Yao; Chi Zhang; Wei Du; Chao Liu; Ying Xu
Journal:  PLoS One       Date:  2015-09-16       Impact factor: 3.240

6.  Automated Analysis and Classification of Histological Tissue Features by Multi-Dimensional Microscopic Molecular Profiling.

Authors:  Daniel P Riordan; Sushama Varma; Robert B West; Patrick O Brown
Journal:  PLoS One       Date:  2015-07-15       Impact factor: 3.240

7.  Evaluation of mitotic activity index in breast cancer using whole slide digital images.

Authors:  Shaimaa Al-Janabi; Henk-Jan van Slooten; Mike Visser; Tjeerd van der Ploeg; Paul J van Diest; Mehdi Jiwa
Journal:  PLoS One       Date:  2013-12-30       Impact factor: 3.240

8.  Classification of breast cancer histology images using Convolutional Neural Networks.

Authors:  Teresa Araújo; Guilherme Aresta; Eduardo Castro; José Rouco; Paulo Aguiar; Catarina Eloy; António Polónia; Aurélio Campilho
Journal:  PLoS One       Date:  2017-06-01       Impact factor: 3.240

9.  CellProfiler: image analysis software for identifying and quantifying cell phenotypes.

Authors:  Anne E Carpenter; Thouis R Jones; Michael R Lamprecht; Colin Clarke; In Han Kang; Ola Friman; David A Guertin; Joo Han Chang; Robert A Lindquist; Jason Moffat; Polina Golland; David M Sabatini
Journal:  Genome Biol       Date:  2006-10-31       Impact factor: 13.583

10.  Predicting non-small cell lung cancer prognosis by fully automated microscopic pathology image features.

Authors:  Kun-Hsing Yu; Ce Zhang; Gerald J Berry; Russ B Altman; Christopher Ré; Daniel L Rubin; Michael Snyder
Journal:  Nat Commun       Date:  2016-08-16       Impact factor: 14.919

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.