| Literature DB >> 34579788 |
Khoa A Tran1,2, Olga Kondrashova1, Andrew Bradley3, Elizabeth D Williams2,4, John V Pearson1, Nicola Waddell5.
Abstract
Deep learning is a subdiscipline of artificial intelligence that uses a machine learning technique called artificial neural networks to extract patterns and make predictions from large data sets. The increasing adoption of deep learning across healthcare domains together with the availability of highly characterised cancer datasets has accelerated research into the utility of deep learning in the analysis of the complex biology of cancer. While early results are promising, this is a rapidly evolving field with new knowledge emerging in both cancer biology and deep learning. In this review, we provide an overview of emerging deep learning techniques and how they are being applied to oncology. We focus on the deep learning applications for omics data types, including genomic, methylation and transcriptomic data, as well as histopathology-based genomic inference, and provide perspectives on how the different data types can be integrated to develop decision support tools. We provide specific examples of how deep learning may be applied in cancer diagnosis, prognosis and treatment management. We also assess the current limitations and challenges for the application of deep learning in precision oncology, including the lack of phenotypically rich data and the need for more explainable deep learning models. Finally, we conclude with a discussion of how current obstacles can be overcome to enable future clinical utilisation of deep learning.Entities:
Keywords: Artificial intelligence; Cancer genomics; Cancer of unknown primary; Deep learning; Explainability; Molecular subtypes; Multi-modal learning; Pharmacogenomics; Precision oncology; Prognosis; Tumour microenvironment
Mesh:
Year: 2021 PMID: 34579788 PMCID: PMC8477474 DOI: 10.1186/s13073-021-00968-x
Source DB: PubMed Journal: Genome Med ISSN: 1756-994X Impact factor: 11.117
Fig. 1Deep learning may impact clinical oncology during diagnosis, prognosis and treatment. Specific areas of clinical oncology where deep learning is showing promise include cancer of unknown primary, molecular subtyping of cancers, prognosis and survivability and precision oncology. Examples of deep learning applications within each of these areas are listed. The data modalities utilised by deep learning models are numerous and include genomic, transcriptomic and histopathology data categories covered in this review
Fig. 2An overview of Deep Learning techniques and concepts in oncology. a Graph convolutional neural networks (GCNN) are designed to operate on graph-structured data. In this particular example inspired by [17–19], gene expression values (upper left panel) are represented as graph signals structured by a protein–protein interactions graph (lower left panel) that serve as inputs to GCNN. For a single sample (highlighted with red outline), each node represents one gene with its expression value assigned to the corresponding protein node, and inter-node connections represent known protein–protein interactions. GCNN methods covered in this review require a graph to be undirected. Graph convolution filters are applied on each gene to extract meaningful gene expression patterns from the gene’s neighbourhood (nodes connected by orange edges). Pooling, i.e. combining clusters of nodes, can be applied following graph convolution to obtain a coarser representation of the graph. Output of the final graph convolution/pooling layer would then be passed through fully connected layers producing GCNN’s decision. b Semantic segmentation is applied to image data where it assigns a class label to each pixel within an image. A semantic segmentation model usually consists of an encoder, a decoder and a softmax function. The encoder consists of feature extraction layers to ‘learn’ meaningful and granular features from the input, while the decoder learns features to generate a coloured map of major object classes in the input (through the use of the softmax function). The example shows a H&E tumour section with infiltrating lymphocyte map generated by Saltz et al. [20] DL model c multimodal learning allows multiple datasets representing the same underlying phenotype to be combined to increase predictive power. Multimodal learning usually starts with encoding each input modality into a representation vector of lower dimension, followed by a feature combination step to aggregate these vectors together. d Explainability methods take a trained neural network and mathematically quantify how each input feature influences the model’s prediction. The outputs are usually feature contribution scores, capable of explaining the most salient features that dictate the model’s predictions. In this example, each input gene is assigned a contribution score by the explainability model (colour scale indicates the influence on the model prediction). An example of gene interaction network is shown coloured by contribution scores (links between red dots represent biological connections between genes)
Summary of deep learning methods, their relevant applications and brief technical descriptions of each DL model
| Application | DL method | Reference | Description |
|---|---|---|---|
| Microscopy-based assessment of cancer | CNN | Ruy et al. [ Nir et al. [ Ström et al. [ Ehteshami Bejnordi et al. [ Vuong et al. [ El Achi and Khoury [ | Trained CNNs on pathology images to predict grading of prostate [ |
| CNN & explainability | Hägele et al. [ | LRP used to assigned feature contribution for cancer grade for each pixel of WSIs | |
| Semantic segmentation | Poojitha and Lal Sharma [ | A semantic segmentation technique called GAN was used to segment tissue maps for prostate cancer grade prediction | |
| Molecular subtyping | MLP | DeepCC [ | Gene set enrichment analysis used to transform gene expression input into functional spectra |
| CNN | imCMS [ Sirinukuwattana et al. [ Woerl et al. [ | Models trained on histopathology images to classify molecular subtypes of of lung [ | |
| GCNN | Rhee et al. [ | Utilised a hybrid GCNN model to organise input gene expression profiles into STRING PPI network [ | |
| Multimodal learning | Islam et al. [ | Two CNN models used to predict breast cancer molecular subtypes from CNAs and gene expression; Outputs of the last fully connected layer of each model concatenated for a final subtype prediction | |
| Cancer of unknown primary | MLP | Jiao et al. [ | Model trained to predict origins of 24 cancer types using somatic mutation patterns and driver genes |
| CNN | SCOPE [ CUP-AI-Dx [ | Both studies trained models to predict different cancer types from gene expression | |
| RNN & explainability | TOAD [ | RNN-based model called Attention was trained on WSIs to predict metastasis and cancer origin; Attention algorithm reveal image regions contributing most to predictions were mostly cancer cells | |
| Prognosis prediction | MLP | Cox-nnet [ DeepSurv [ | Cox regression used as the last layer of MLP models for prognosis prediction |
| MLP & AEs | AECOX [ | AE used to “compress” gene expression into low-dimensional embedding vector and used as an input for Cox-regression | |
| Explainability | PASNET [ Cox-PASNET [ | A pathway layer used between the input and the hidden layers with each node representing a known pathway; Analysis of weight differences in pathway layers reveal clinically actionable genetic traits | |
| MesoNet [ | Histopathology images split into tiles and scored by survival prediction contributions; Scores used to identify top-contributing regions, reviewed by pathologists | ||
| GCNN & explainability | Chereda et al. [ | Combine GCNN and explainability method LRP to identify biologically and therapeutically relevant genes in predicting metastasis of breast cancer | |
| Explainability with multimodal learning | PAGE-Net [ | CNN used to compress features from WSIs; Cox-PASNet used to incorporate gene pathway and provide cross-modal analysis with image features extracted by CNN | |
| PathME [ | AEs used to compress features from four omics modalities, which are combined to predict survival; SHAP used to assign each omics feature survival prediction contribution score | ||
| Precision Oncology | MLP | HER2RNA [ | Transcriptomic profiles inferred from histopathology images divided into tiles; Predictions added up for all tiles and compared with ‘ground truth’ transcriptomic profiles |
| CNN | Image2TMB [ | Ensemble of three CNNs to extract features from histopathological images at different resolutions (x5, x10 and x20); Extracted features are combined to infer TMB | |
| Kather et al. [ | TCGA histopathology images used to predict mutational status of key genes, molecular subtypes and gene expression of standard biomarkers | ||
| Tumour microenvironment | MLP | Scaden [ | Ensemble of three models with different filter sizes to predict TME composition from gene expression; Predictions from the models are averaged into a final prediction |
| Explainability with MLP | MethylNet [ | MLP and AE used to ‘compress’ CpG beta values into an embedding vector for predicting TME composition; SHAP used to assign feature contribution to each CpG site | |
| Semantic segmentation | Saltz et al. [ | Semantic segmentation model used on H&E images to localise spatial heterogeneity patterns of TIL and necrosis | |
| Spatial transcriptomics | CNN | ST-Net [ | Images split into tiles centred on spatial transcriptomics spots; Tiles used to train a CNN to predict expression of 250 target genes |
| Pharmacogenomics | CNN | CDRscan [ | Two models used to extract features from somatic mutational fingerprints and molecular profiles of drugs (cell lines); Feature vectors combined to predict efficacy of drugs based on genomic profiles |
| MLP | DeepSynergy [ | Cell line gene expression and chemical features of drugs in drug combinations used as input; Predicts ‘synergy score’ between the drug combinations and transcriptomic profiles | |
| GCNN | Jiang et al. [ | Utilised graph structure to integrate protein-protein, drug-drug and drug-protein interactions to predict synergistic drug combination for specific cell lines | |
| Multimodal learning | DeepDR [ | Collection of ten AEs to integrate ten drug-disease networks, which predict drug-disease associations | |
| CNN | DeepDTI [ | Protein sequence and drug fingerprint as input to predict drug protein-binding sites |
AE: autoencoder, CNA: copy number alterations, CNN: convolutional neural network, DL: deep learning, GCNN: graph convolutional neural network, H&E: haematoxylin and eosin, LRP: layer-wise relevance propagation, MLP: multilayer perceptron, RNN: recurrent neural netowrk, SHAP: SHapley Additive exPlanations, TIL: tumour-infiltrating lymphocytes, TMB: tumour mutational burden, WSI: whole slide image