Literature DB >> 31324828

Pan-Renal Cell Carcinoma classification and survival prediction from histopathology images using deep learning.

Sairam Tabibu1, P K Vinod2, C V Jawahar3.   

Abstract

Histopathological images contain morphological markers of disease progression that have diagnostic and predictive values. In this study, we demonstrate how deep learning framework can be used for an automatic classification of Renal Cell Carcinoma (RCC) subtypes, and for identification of features that predict survival outcome from digital histopathological images. Convolutional neural networks (CNN's) trained on whole-slide images distinguish clear cell and chromophobe RCC from normal tissue with a classification accuracy of 93.39% and 87.34%, respectively. Further, a CNN trained to distinguish clear cell, chromophobe and papillary RCC achieves a classification accuracy of 94.07%. Here, we introduced a novel support vector machine-based method that helped to break the multi-class classification task into multiple binary classification tasks which not only improved the performance of the model but also helped to deal with data imbalance. Finally, we extracted the morphological features from high probability tumor regions identified by the CNN to predict patient survival outcome of most common clear cell RCC. The generated risk index based on both tumor shape and nuclei features are significantly associated with patient survival outcome. These results highlight that deep learning can play a role in both cancer diagnosis and prognosis.

Entities:  

Mesh:

Year:  2019        PMID: 31324828      PMCID: PMC6642160          DOI: 10.1038/s41598-019-46718-3

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Kidney Cancer accounts for nearly 3.8% of adult cancers and is among the 10 most common cancers in both men and women. Recent estimate by American Cancer Society indicates that 63340 new cases and 14970 death will occur in 2018[1]. Renal Cell Carcinoma (RCC) is the most common (85%) malignant tumor of kidney and is a heterogeneous group of tumors with different histology, molecular characteristics, clinical outcomes and responses to therapy[2]. The major subtypes of RCC are clear cell, papillary and chromophobe accounting for 70–80%, 14–17% and 4–8% of RCC, respectively[3,4]. Clear cell RCC (KIRC) and papillary RCC (KIRP) originate from cells in the proximal convoluted tubules of the nephron[5]. KIRC is characterized with loss of chromosome 3p and mutation of the von Hippel–Lindau (VHL) gene while KIRP is characterized by trisomy of chromosomes and loss of chromosome 9p[6,7]. Chromophobe RCC (KICH) originates from intercalated cells in the distal convoluted tubules and is characterized by loss of chromosomes[8]. KIRC patients have an overall 5-year survival rate of 55–60%[9-11] whereas for KIRP patients, it varies from 80–90%[12,13] and for KICH patients, it is 90%[14]. Due to these distinct biological and clinical behavior of subtypes, accurate detection of RCC and its subtypes is vital for the clinical management of patients. RCC subtypes can be detected radiologically based on degree of tumor enhancement on multidetector computed tomography (CT) or magnetic resonance imaging (MRI)[15-19]. Further, the microscopic examination of Hematoxylin and Eosin (H & E) stained slides of biopsies continues to be a valuable tool for pathologist and clinicians[20]. Histological images contain markers of disease progression and phenotypic information that can have diagnostic and predictive values. Most pathologists follow simple decision tree-based approaches for tumor stratification. Major limitations in the examination of H & E images by pathologists are inter-observer discordance and time required to diagnose. Computer-aided approaches can help to overcome these limitations and to identify subtle morphological differences between clinical groups. The Cancer Genome Atlas (TCGA)[21,22] project has resulted in the creation of large repositories of digital H & E whole-slide images (WSI) of RCC. These images are acquired at 20x and 40x magnifications with size varying from 10000–100000 pixels, which are visually tricky to analyze and interpret accurately. Further, the high-resolution WSI also poses a challenge to develop computational model for classification. An automated system to detect RCC and its subtypes is still not available. Cheng et al.[23] extracted image features from TCGA tumor slides to develop a lasso-regularized Cox model for predicting survival of KIRC patients. In recent years, deep learning techniques especially convolutional neural networks (CNN’s) have significantly improved accuracy on a wide range of computer vision tasks such as image recognition[24-26], object detection[27,28] and semantic segmentation[27,29]. Further, CNN’s have also been successful in capturing the complex tissue patterns and have been widely used in biomedical imaging for segmentation as well as for classification tasks in cancers such as breast[30-32], lung[33-36] and prostate[37]. In this study, we demonstrate how deep learning can be applied to histopathological images of RCC. An automatic system for pan-RCC classification and survival prediction tasks was developed from WSI of TCGA. CNN’s trained on WSI classify each RCC subtype from normal tissues and three RCC subtypes, and extract features for survival prediction. In that process, we also characterized the high probability cancerous regions and tissue of origin differences and came up with a strategy for overcoming data imbalance in multi-class classification problems using histopathological images.

Results

CNN model to distinguish RCC from normal tissue using histopathology images

TCGA slide images of RCC and normal tissues were split into 70-15-15% for training, validation and testing (Supplementary Tables 1–3). Image patches of size 512 × 512 were extracted from the slides of both 20x and 40x resolution with 50% overlap (Fig. 1). The background patches were removed using pixel thresholding (See Methods Section). We used the pre-trained Resnet 18 and Resnet 34 by replacing the last layers with just 2 output layers and fine-tuned it on RCC data. We terminated the training after 30 epochs when the validation accuracy failed to improve. The images were processed patch-wise as well as slide-wise. We trained the model on only KIRC and KICH due to lack of sufficient normal tissue samples for KIRP. We obtained 93.39% and 87.34% (Tables 1 and 2) patch-wise accuracy on the test set for KIRC and KICH, respectively. To assess this accuracy, we performed a slide-wise analysis by counting the percentage of positively classified patches in a slide. We achieved a best Area Under Curve (AUC) of 0.98 for KIRC and 0.95 for KICH using Resnet 34 (Table 2).
Figure 1

Cancer Classification Pipeline. (1) Kidney Whole slide images. (2) 512*512 patches extracted from images with 50% overlap and background removed using pixel thresholding. (3) Patches from normal and cancerous slides fed to the deep network. (4) Patches classified as cancerous or non-cancerous. (5) High probability patches identified by the trained network and binary mask is applied. (6) The patches from three subtypes used to train a similar deep architecture for a three-way classification. (7) Features extracted from the penultimate layer of the network and fed to DAG-SVM and a three-way classification is performed by it.

Table 1

Cancer/Normal Classification (20x resolution).

ModelPatch-wise AccuracyPrecisionRecallSlide-wise AUC
Resnet-18 (KIRC) 90.20 90.2388.350.97
Resnet-34 (KIRC)89.2289.6186.840.98
Resnet-18 (KICH)84.6885.7882.920.92
Resnet-34 (KICH) 87.34 87.5786.340.98
Table 2

Cancer/Normal Classification (40x resolution).

ModelPatch-wise AccuracyPrecisionRecallSlide-wise AUC
Resnet-18 (KIRC) 93.39 93.4192.950.99
Resnet-34 (KIRC)93.6293.4793.370.99
Resnet-18 (KICH)79.0979.6180.060.95
Resnet-34 (KICH) 80.57 80.6581.170.95
Cancer Classification Pipeline. (1) Kidney Whole slide images. (2) 512*512 patches extracted from images with 50% overlap and background removed using pixel thresholding. (3) Patches from normal and cancerous slides fed to the deep network. (4) Patches classified as cancerous or non-cancerous. (5) High probability patches identified by the trained network and binary mask is applied. (6) The patches from three subtypes used to train a similar deep architecture for a three-way classification. (7) Features extracted from the penultimate layer of the network and fed to DAG-SVM and a three-way classification is performed by it. Cancer/Normal Classification (20x resolution). Cancer/Normal Classification (40x resolution).

RCC subtype detection using CNN

A challenging task is to distinguish the subtypes of RCC: KIRC, KIRP and KICH. We used the deep learning framework to distinguish these three subtypes (Fig. 1). However, there exists a major class imbalance with 43% KIRC, 14% KIRP and 43% KICH images. First, we tested whether the CNN can distinguish these subtypes irrespective of the data imbalance. We extracted the 512 × 512 patches from 20x resolution for each subtype and performed a three-class classification using Resnet-18 and Resnet-34. We obtained a patch-wise accuracy of 87.69% and a micro-average AUC of 0.91 with Kappa score of 0.794 (Table 3). However, the recall score is in the range of 83% suggesting that certain classes are misclassified. Individual analysis of the classification scores on each subtype (Supplementary Table 4) revealed that the CNN performs better on subtypes KIRC and KICH but not on KIRP (70%) which can be attributed to the severe data imbalance within the subtypes.
Table 3

Cancer subtype classification.

ModelPatch-wise AccuracyPrecisionRecallSlide-wise AUC
Resnet-1887.6988.8283.660.88
Resnet-3486.1988.3083.180.88
Resnet-18 + DAG-SVM 92.62 90.7889.07 0.93
Resnet-34 + DAG-SVM 91.96 88.9487.92 0.93
Cancer subtype classification. We used a DAG-SVM method[38] to deal with this data imbalance method without making any changes to the deep learning architecture. In this framework, the N-class classification problem is divided into N(N-1)/2 binary classification problem by replacing the softmax layer of deep architecture by binary classifiers for each pair of classes arranged in a directed acyclic graph (DAG) structure (Fig. 2). Such division enables the network to focus on individual binary problems which are relatively easier to tackle as well as to learn more pairwise discriminative features (See Methods Section). Using the representation matrix obtained by the CNN from the penultimate layer of the CNN for patch images, we trained the DAG-SVM and performed the subtype classification. We obtained a substantial increase in the patch-wise accuracy by 5% (Table 3). The increase in the micro-average AUC is 0.03 whereas macro-average AUC is 0.05. This demonstrates that the combined CNN and DAG-SVM model substantially improves the performance of discriminating the subtypes even in the given limitation of data imbalance. This is also evident from the increase in the classification score of the subtype KIRP (10%) (Supplementary Table 5).
Figure 2

Directed Acyclic Graph SVM (DAG-SVM)-Architecture of the model for a four-class problem where features learned by the deep network are used to train all the classifiers. Each node is a binary classifier for a pair of classes.

Directed Acyclic Graph SVM (DAG-SVM)-Architecture of the model for a four-class problem where features learned by the deep network are used to train all the classifiers. Each node is a binary classifier for a pair of classes. Further, we also applied two alternative strategies to overcome the class imbalance problem. First, data augmentation of the minority class was done to balance the class distribution. We obtained an accuracy of 91.47% with a Kappa score of 0.859. Second, a weighted resampling technique was used to train the CNN such that the minority class gets more weightage depending on the severity of the imbalance. We obtained an accuracy of 92.16% with a Kappa score of 0.871. Both these methods significantly improved the subtype classification accuracy (Supplementary Table 6). Furthermore, we applied DAG-SVM to both the data augmentation and sampled models and observed an increase in accuracy by 2.28% in case of augmented model and 2.35% in case of sampled model.

Prediction of tumor areas and the survival outcome for KIRC

Using the CNN model that distinguishes RCC from normal tissue, high probability tumor patches were identified, and a probability heat map was generated for the WSI. Based on the tumor region, several tumor and nuclei shape features such as area, perimeter etc. were extracted from the cancer slide images of each patient (Fig. 3). The risk index of each patients was calculated using lasso-regularized COX model for each feature and validated using two-level cross validation strategy (see methods).
Figure 3

Shape Features Extraction Pipeline. (A) Tissue slide image. (B) Heatmap generated after each patch is fed through the deep network. (C) High probability patches are highlighted. (i) Binary mask generated for high probability patches. (ii) Small patches removed using morphological operations. (iii) Tumor shape features extracted from final image. (a) High probability patches accumulated. (b) Nuclei segmented from each patch. (c) Nuclei shape features extracted from each patch.

Shape Features Extraction Pipeline. (A) Tissue slide image. (B) Heatmap generated after each patch is fed through the deep network. (C) High probability patches are highlighted. (i) Binary mask generated for high probability patches. (ii) Small patches removed using morphological operations. (iii) Tumor shape features extracted from final image. (a) High probability patches accumulated. (b) Nuclei segmented from each patch. (c) Nuclei shape features extracted from each patch. Figure 4 shows the survival curves of low and high-risk groups for different features. We found that 13 tumor shape features and 6 nuclei shape features are significantly (p value less than 0.05) associated with the patient survival (Table 4). Tumor shape features include total area (p value = 1.5e-6, HR = 2.398) and total perimeter (p value = 1.49e-5, HR = 2.485), and nuclei shape features include total convex area (p value = 2.2e-7, HR = 2.576) and total major axis (p = 0.000614, HR = 2.252) (Fig. 4A–C). Most predictive individual features were selected for integrative analysis to stratify patients into low risk and high-risk groups using Lasso-Cox model. We found significant (p value = 3.68e-6) association between combined image features and survival outcome (Fig. 4D). Multivariate analysis was performed with predicted risk indices, age, gender, grade and stage as covariates (Table 5). We observed that the survival outcome significantly depends on the predicted risk indices and stages of tumor. This suggest that tumor and cell nuclei shape features identified in this study can be used as a prognostic factor to predict the survival outcome of KIRC patients.
Figure 4

Tumor shape and cell shape features predict survival outcome of KIRC patients.

Table 4

Tumor shape and Nuclei features and their log-rank test p-values.

Feature typeImage featurep-value
Tumor shape featuresTotal area1.5e-6
Total convex area2.2e-7
Total perimeter1.49e-5
Total filled area8.76e-7
Total major Axis0.000614
Total minor axis0.000572
Total perimeter by Area0.00139
Main region area0.0042
Main region convex area1.58e-5
Main region perimeter1.1e-5
Main region perimeter by area4.29e-5
Main region major axis6.09e-5
Main region minor axis0.00161
Nuclei shape featurestotal area0.000106
Total convex area6.98e-5
Total perimeter0.00327
Total filled area0.000115
Total major axis0.00251
Total minor axis0.00748
Integrative model3.68e-6
Table 5

Multivariate analysis of predicted risk indices.

VariableHR(95% CI)p-value
Lasso Cox2.265 (1.5343–3.343)3.87e-5
Age1 (0.999–1.0001)3.71e-6
Gender1.095 (0.7870–1.524)0.590
Stage1.733 (1.4968–2.007)2.01e-13
Grade1.106 (0.9369–1.306)0.233
Tumor shape and cell shape features predict survival outcome of KIRC patients. Tumor shape and Nuclei features and their log-rank test p-values. Multivariate analysis of predicted risk indices.

Discussion

In this study, we developed an automatic system that can distinguish RCC from normal tissue and RCC subtypes using histopathology images. To the best of our knowledge, this is the first study to perform classification of RCC subtypes using deep learning framework. CNN’s successfully identified the cancerous tissue patterns as well as the inherent texture differences among RCC subtypes. We also introduced a method to deal with data imbalance which can be generalized to other cancer classification tasks. Finally, we developed a prognostic model based on the regions detected by the CNN to predict survival outcome. We used a two-level cross validation technique to validate our model. Due to the enormous amount of information contained in the WSI, the histopathological analysis is a laborious and time intensive task for the pathologist. With the growing number of cancer cases, fast and precise assessment is not feasible. Our study addresses this limitation by providing computational interpretation of WSI with high AUCs in the testing phase of various classification tasks. Since the pipeline was developed on the WSI of cancer, it eliminates the requirement of annotating tumor regions for the training purpose and provides a strategy that can used to study WSI of other tumors. Further, a high patch-wise AUC obtained for distinguishing tumor from normal tissue suggests that tumor regions are not localized and are spread across WSI. Our analysis also showed that a better performance is obtained for KIRC when trained on 40x images and for KICH when trained on 20x images (Tables 1 and 2). These results suggest the existence of complex morphological distinguishing features between cancers. Subtypes of RCC typically have varying morphological patterns and nuclear/cytoplasmic features at the microscopic level[39]. Different cases of KIRC have mixture of cells with clear cytoplasm and granular eosinophilic cytoplasm, pseudopapillary pattern, rich vascularity and areas of hemorrhage. KICH has polygonal cells with prominent cell membranes, cells of different size with either eosinophilic or foamy cytoplasm and incomplete vascular septae. KIRC has predominantly papillary growth pattern, foamy macrophages and cuboidal cells. However, there are instances were histopathology images are not conclusive leading to re-classification of some KIRC samples as KICH during the histological review[40]. The CNN model is able to distinguish these subtypes by capturing their unique characteristics with high AUC of 0.93 (Table 3). Further, we also tested whether the inherent difference in the texture of tissue of origin contribute to classification of subtypes by using normal tissue samples as a test set to RCC subtype model. We obtained subtypes classification with AUC of 0.85 (Supplementary Table 7). Together, these analyses suggest that both tumor regions and the inherent texture difference contributes to subtype classification. Texture classification methods[41] can be further used to visualize the difference (Fig. 5) and study their characteristics. Further, for subtype detection, we introduced a DAG-SVM classifier on the top of the deep network. This helped to break the multi-class classification task into multiple binary classification tasks which not only improved the performance of the model but also helped to deal with data imbalance. DAG-SVM is useful when the deep networks lack the scope for improvement without a major change in architecture. This method can also be applied to different deep architectures, since the method focuses only on the output layer.
Figure 5

Visualization of cancerous and normal tissue regions at 40x resolution.

Visualization of cancerous and normal tissue regions at 40x resolution. Finally, we also demonstrated the reliability of features extracted from tumor slide images in predicting the survival outcome. We showed that both tumor shape and nuclei features from the high probability regions identified by CNN are significantly associated with survival outcome of KIRC. Image features such as ratio of perimeter squared by area, which is a quantification of tumor surface[34], is correlated with survival outcome. A major finding that tumor shape features also predict survival extends the previous studies that uses genomic and nuclei features to predict survival of KIRC. It is also consistent with observation that tumor shape features can be used to predict the survival outcome of lung patients[34]. However, this model is developed using annotated images by pathologist, which in our case is done independent of any human bias. The lack of data and the class imbalance are major challenges in building survival models of other subtypes. Firstly, there is dearth of KIRP-matched normal slide images that can be used to develop a model to distinguish tumor from normal tissue and to identify high probability cancerous regions. Secondly, since the ratio of events (death) to total number of patients is few for KICH and KIRP (less than 15%) compared to KIRC (around 40%), the survival prediction for KICH and KIRP is affected. Further, subtypes such as renal oncocytoma and collecting duct (Bellini duct) of RCC have the rarest chance to occur (less than 1%) and therefore data is scarce to develop model including rare subtypes. On the other hand, we made several attempts to use only tumor slide images of KIRP to classify different tumor stages and grades, which did not yield satisfactory performance. Further study in this direction is required to see whether improvements in deep learning framework can help to detect the subtle differences between early and late stages of RCC and to integrate genomic features to improve the performance. Studies to understand the complex relationship between tissue histology and genetics can provide useful insights for the diagnosis and treatment.

Methods

Dataset and image processing

The whole slide images and clinical information were downloaded from TCGA data portal (https://gdc.cancer.gov/). Slides with reading and compatibility issues were removed (972 slides in total were removed from the whole dataset). We selected 1027 (KIRC), 303 (KIRP), and 254 (KICH) tumor slide images for our study. Further, corresponding 379, 47 and 83 normal tissue slide images for each subtype were selected. 512*512 sized tiles were extracted with 50% overlap ensuring multiple viewpoints within the tissue, at a magnification of 20x and 40x. For subtype classification, patches from 20x magnification were used. The patches were removed if the mean intensity of 50%-pixel values was larger than a threshold (in our case 210 for RGB channels).

Training architecture and framework

The Resnet 18 and 34 architectures[24] pre-trained on Imagenet dataset were used with input resized from 512*512 to 224*224. The final linear layer was modified to 2 for cancer/normal detection and 3 for subtype detection. The network was fine-tuned with stochastic gradient optimizer and cross entropy was used as the loss function. To introduce generalization, we also used data augmentation techniques (random horizontal flip and random Crop) while training. The images before being fed to the network were normalized using mean and standard deviation which was calculated on the training set. Batch size was set to 128 with learning rate 1e-5 and decay according to the validation error. The average number of epochs for both experiments lies in between 3–40 epochs. The training was terminated when the validation accuracy did not vary much for 4–5 epochs. Tests were implemented using Pytorch library[42]. For the 3-class classification problem in the subtype detection, we used a DAG-SVM on the top of the fully connected layer of the CNN to improve the multiclass accuracy. We removed the final softmax layer from the trained model and use the learned representation given by the CNN (which is essentially a 512*1 vector) to train the DAG-SVM. Linear kernel was used while training the SVMs. A two-step procedure was followed for data augmentation. First, random patches were chosen from the minority class and then three augmentation techniques: random vertical flip, rotation (−25 to +25 degrees) and adding noise to image were applied with a probability of 0.5 to the selected image. 100000 augmented images were generated using this technique from 74000 images. Further, a weighted resampling technique was also used to overcome the class imbalance. The weights were assigned according to the class distribution of the respective subtypes using the WeightedRandomSampler function in pytorch. The training parameters were kept the same.

Mapping high probability regions and survival analysis

The survival analysis was performed on KIRC WSI’s (Supplementary Table 8) The probability score for each patch was calculated by passing them through the KIRC CNN model (normal/cancer) trained on patches in 40x resolution, and a heatmap was constructed for each slide (Supplementary Figs 1 and 2). The high probability regions (greater than 0.95) were highlighted using binary masking. To remove the effects of very small patches, tissue regions which were less than 1/3rd of the main region were removed. 15 shape features including region area, convex area, perimeter, solidity, and eccentricity were extracted from the high probability patches. Further, the nuclei were segmented using a hierarchical multilevel thresholding approach[43] and 7 nuclei shape features were extracted (Supplementary Table 9). In this approach, the parameter values are set adaptively eliminating the need for parameter learning or training data. If a patient had one or more slides the values were averaged and patient level features were aggregated. We used these morphological features to build a Lasso regularized COX model. We estimated the risk index of each patient based on each morphological feature. A two-level cross validation strategy was used to validate our model[23]. Each patient was kept aside as a test case and the rest of the samples were used to train the COX model using a 10-fold cross validation. The model was then used to stratify the left out patient into either high risk or low risk group using the median score obtained from the training set. This process was repeated until each patient was used as a test set. From the finally obtained groups and survival indices, p-values were calculated using log-rank test. We used Kaplan-Meier plots to visualize the survival curves of low and high-risk groups. The survival analysis was performed using R packages “survival” and “glmnet”[44,45].
  31 in total

1.  Differentiation of renal clear cell carcinoma and renal papillary carcinoma using quantitative CT enhancement parameters.

Authors:  Andrea J Ruppert-Kohlmayr; Martin Uggowitzer; Thomas Meissnitzer; Georg Ruppert
Journal:  AJR Am J Roentgenol       Date:  2004-11       Impact factor: 3.959

2.  Chromophobe cell renal carcinoma.

Authors:  Y Megumi; K Nishimura
Journal:  Urol Int       Date:  1998       Impact factor: 2.089

3.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.

Authors:  Shaoqing Ren; Kaiming He; Ross Girshick; Jian Sun
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-06-06       Impact factor: 6.226

4.  Differentiation of renal cell carcinoma subtypes by multislice computerized tomography.

Authors:  Khaled Z Sheir; Mohamed El-Azab; Ahmed Mosbah; Mahmoud El-Baz; Atallah A Shaaban
Journal:  J Urol       Date:  2005-08       Impact factor: 7.450

Review 5.  Risk and prognosis in renal neoplasms. A pathologist's prospective.

Authors:  S M Bonsib
Journal:  Urol Clin North Am       Date:  1999-08       Impact factor: 2.241

6.  The Cancer Genome Atlas Pan-Cancer analysis project.

Authors:  John N Weinstein; Eric A Collisson; Gordon B Mills; Kenna R Mills Shaw; Brad A Ozenberger; Kyle Ellrott; Ilya Shmulevich; Chris Sander; Joshua M Stuart
Journal:  Nat Genet       Date:  2013-10       Impact factor: 38.330

7.  Multilevel Genomics-Based Taxonomy of Renal Cell Carcinoma.

Authors:  Fengju Chen; Yiqun Zhang; Yasin Şenbabaoğlu; Giovanni Ciriello; Lixing Yang; Ed Reznik; Brian Shuch; Goran Micevic; Guillermo De Velasco; Eve Shinbrot; Michael S Noble; Yiling Lu; Kyle R Covington; Liu Xi; Jennifer A Drummond; Donna Muzny; Hyojin Kang; Junehawk Lee; Pheroze Tamboli; Victor Reuter; Carl Simon Shelley; Benny A Kaipparettu; Donald P Bottaro; Andrew K Godwin; Richard A Gibbs; Gad Getz; Raju Kucherlapati; Peter J Park; Chris Sander; Elizabeth P Henske; Jane H Zhou; David J Kwiatkowski; Thai H Ho; Toni K Choueiri; James J Hsieh; Rehan Akbani; Gordon B Mills; A Ari Hakimi; David A Wheeler; Chad J Creighton
Journal:  Cell Rep       Date:  2016-03-03       Impact factor: 9.423

8.  Solid renal cortical tumors: differentiation with CT.

Authors:  Jingbo Zhang; Robert A Lefkowitz; Nicole M Ishill; Liang Wang; Chaya S Moskowitz; Paul Russo; Halley Eisenberg; Hedvig Hricak
Journal:  Radiology       Date:  2007-08       Impact factor: 11.105

9.  Precision histology: how deep learning is poised to revitalize histomorphology for personalized cancer care.

Authors:  Ugljesa Djuric; Gelareh Zadeh; Kenneth Aldape; Phedias Diamandis
Journal:  NPJ Precis Oncol       Date:  2017-06-19

10.  Classification and mutation prediction from non-small cell lung cancer histopathology images using deep learning.

Authors:  Nicolas Coudray; Paolo Santiago Ocampo; Theodore Sakellaropoulos; Navneet Narula; Matija Snuderl; David Fenyö; Andre L Moreira; Narges Razavian; Aristotelis Tsirigos
Journal:  Nat Med       Date:  2018-09-17       Impact factor: 53.440

View more
  28 in total

Review 1.  Deep learning in histopathology: the path to the clinic.

Authors:  Jeroen van der Laak; Geert Litjens; Francesco Ciompi
Journal:  Nat Med       Date:  2021-05-14       Impact factor: 53.440

2.  Survival prediction of patients suffering from glioblastoma based on two-branch DenseNet using multi-channel features.

Authors:  Xue Fu; Chunxiao Chen; Dongsheng Li
Journal:  Int J Comput Assist Radiol Surg       Date:  2021-01-18       Impact factor: 2.924

3.  A deep learning framework for automated classification of histopathological kidney whole-slide images.

Authors:  Hisham A Abdeltawab; Fahmi A Khalifa; Mohammed A Ghazal; Liang Cheng; Ayman S El-Baz; Dibson D Gondim
Journal:  J Pathol Inform       Date:  2022-04-18

4.  Classification of malignant tumors by a non-sequential recurrent ensemble of deep neural network model.

Authors:  Dipanjan Moitra; Rakesh Kr Mandal
Journal:  Multimed Tools Appl       Date:  2022-02-14       Impact factor: 2.577

5.  Deep Multi-task Prediction of Lung Cancer and Cancer-free Progression from Censored Heterogenous Clinical Imaging.

Authors:  Riqiang Gao; Lingfeng Li; Yucheng Tang; Sanja L Antic; Alexis B Paulson; Yuankai Huo; Kim L Sandler; Pierre P Massion; Bennett A Landman
Journal:  Proc SPIE Int Soc Opt Eng       Date:  2020-03-10

6.  Down-regulation of BCL2L13 renders poor prognosis in clear cell and papillary renal cell carcinoma.

Authors:  Fei Meng; Luojin Zhang; Mingjun Zhang; Kaiqin Ye; Wei Guo; Yu Liu; Wulin Yang; Zhimin Zhai; Hongzhi Wang; Jun Xiao; Haiming Dai
Journal:  Cancer Cell Int       Date:  2021-06-30       Impact factor: 5.722

7.  Integrated Analysis of the Prognosis-Associated RNA-Binding Protein Genes and Candidate Drugs in Renal Papillary Cell Carcinoma.

Authors:  Silin Jiang; Xiaohan Ren; Shouyong Liu; Zhongwen Lu; Aiming Xu; Chao Qin; Zengjun Wang
Journal:  Front Genet       Date:  2021-02-12       Impact factor: 4.599

8.  Exploring Histological Similarities Across Cancers From a Deep Learning Perspective.

Authors:  Ashish Menon; Piyush Singh; P K Vinod; C V Jawahar
Journal:  Front Oncol       Date:  2022-03-30       Impact factor: 6.244

Review 9.  Extracellular Vesicles in Renal Cell Carcinoma: Multifaceted Roles and Potential Applications Identified by Experimental and Computational Methods.

Authors:  Zhiyuan Qin; Qingwen Xu; Haihong Hu; Lushan Yu; Su Zeng
Journal:  Front Oncol       Date:  2020-05-07       Impact factor: 6.244

Review 10.  Artificial Intelligence in Lung Cancer Pathology Image Analysis.

Authors:  Shidan Wang; Donghan M Yang; Ruichen Rong; Xiaowei Zhan; Junya Fujimoto; Hongyu Liu; John Minna; Ignacio Ivan Wistuba; Yang Xie; Guanghua Xiao
Journal:  Cancers (Basel)       Date:  2019-10-28       Impact factor: 6.639

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.