| Literature DB >> 36230685 |
Ting-He Zhang1, Md Musaddaqul Hasib1, Yu-Chiao Chiu2,3, Zhi-Feng Han1, Yu-Fang Jin1, Mario Flores1, Yidong Chen4,5, Yufei Huang2,3.
Abstract
Deep learning has been applied in precision oncology to address a variety of gene expression-based phenotype predictions. However, gene expression data's unique characteristics challenge the computer vision-inspired design of popular Deep Learning (DL) models such as Convolutional Neural Network (CNN) and ask for the need to develop interpretable DL models tailored for transcriptomics study. To address the current challenges in developing an interpretable DL model for modeling gene expression data, we propose a novel interpretable deep learning architecture called T-GEM, or Transformer for Gene Expression Modeling. We provided the detailed T-GEM model for modeling gene-gene interactions and demonstrated its utility for gene expression-based predictions of cancer-related phenotypes, including cancer type prediction and immune cell type classification. We carefully analyzed the learning mechanism of T-GEM and showed that the first layer has broader attention while higher layers focus more on phenotype-related genes. We also showed that T-GEM's self-attention could capture important biological functions associated with the predicted phenotypes. We further devised a method to extract the regulatory network that T-GEM learns by exploiting the attributions of self-attention weights for classifications and showed that the network hub genes were likely markers for the predicted phenotypes.Entities:
Keywords: Transformer; cancer type prediction; immune cell type prediction; interpretable deep learning; phenotypes prediction
Year: 2022 PMID: 36230685 PMCID: PMC9562172 DOI: 10.3390/cancers14194763
Source DB: PubMed Journal: Cancers (Basel) ISSN: 2072-6694 Impact factor: 6.575
Figure 1Illustration of the proposed T-GEM model. (A) General structure of T-GEM. (B) Detailed structure of a head in the first attention layer.
Hyperparameter sets of G-TEM.
| Hyperparameter | Value of Hyperparameter |
|---|---|
| The number of layers: | 1, 2, 3, 4 |
| the number of heads | 1, 2, 3, 4, 5 |
| the activation function of the classification layer | No activation, ReLu, GeLu |
Performances of T-GEM and the benchmark models for TCGA cancer type classification.
| ACC | MCC | AUC | |
|---|---|---|---|
| CNN(AUTOKERAS) | 94.34% | 0.9411 | 0.9985 |
| SVM | 93.21% | 0.9292 | 0.9972 |
| RANDOM FOREST | 91.60% | 0.9123 | 0.9970 |
| DECISION TREE | 81.80% | 0.8097 | 0.9062 |
| T-GEM | 94.92% | 0.9469 | 0.9987 |
Accuracy (%) of pruning a T-GEM layer or head for TCGA-based cancer type classification (original accuracy is 94.92%).
| LAYER | HEAD 1 | HEAD 2 | HEAD 3 | HEAD 4 | HEAD 5 | |
|---|---|---|---|---|---|---|
| LAYER 1 | 57.48 | 87.19 | 94.92 | 95.10 | 94.92 | 94.83 |
| LAYER 2 | 94.83 | 95.01 | 94.92 | 94.61 | 94.92 | 94.92 |
| LAYER 3 | 94.92 | 94.97 | 94.92 | 94.92 | 94.92 | 94.97 |
Accuracies of each T-GEM layer/head output via SVM.
| LAYER | HEAD 1 | HEAD 2 | HEAD 3 | HEAD 4 | HEAD 5 | |
|---|---|---|---|---|---|---|
| LAYER 1 | 93.57% | 93.48% | 91.69% | 92.36% | 91.91% | 93.12% |
| LAYER 2 | 93.12% | 90.65% | 90.88% | 93.80% | 90.43% | 93.57% |
| LAYER 3 | 91.24% | 93.35% | 90.29% | 90.61% | 89.57% | 88.18% |
Figure 2Entropies of each head’s attention weights. Dot plots for different layers are split by black lines. Each dot represents the average entropy value for one query gene.
Figure 3Functions of T-GEM each layer for cancer type classification. A. snapshot of layer 3 after thresholding the weight attribution scores. The links are associated with scores larger than the threshold. Query genes with no links are non-informative genes. (B) Enriched functions of each layer for the classification of breast cancer (BRCA). (A) link connects an enriched pathway with a pathway in the previous layer if this pathway’s informative gene-associated Key genes are enriched the pathway at the previous layer. The size of the dots represents the number of enriched informative genes in each pathway genesets, and the color shows the enrichment significance (FDR). FDR is negative log2 transferred, red means more significantly enriched, and blue means less enriched. The genesets from cancer states are marked as orange, and genesets from KEGG pathway are black. The pathway is ranked based on the sum of −log2(FDR) for all 3 layers. Abbr. VLID (VALINE_LEUCINE_AND_ISOLEUCINE_DEGRADATION); BM (BUTANOATE_METABOLISM); SIVT(SNARE_INTERACTIONS_IN_VESICULAR_TRANSPORT); OGB (O_GLYCAN_BIOSYNTHESIS); AAGM(ALANINE_ASPARTATE_AND_GLUTAMATE_METABOLISM); GSTM (GLYCINE_SERINE_AND_THREONINE_METABOLISM); ATB(AMINOACYL_TRNA_BIOSYNTHESIS) SHB (STEROID_HORMONE_BIOSYNTHESIS); CMC(CARDIAC_MUSCLE_CONTRACTION); OT(OLFACTORY_TRANSDUCTION).
Figure 4T-GEM regulatory network of the last layer for BRCA. Nodes are informative genes and their associated Key genes. Links go from Key to Query genes. Genes with red color are hub genes.
Figure 5T-GEM regulatory network of the last layer for LUAD. Nodes are informative genes and their associated Key genes. Links go from Key to Query genes. Genes with red color are hub genes.
Performances of T-GEM and benchmark models for cell type classification using PBMC scRNA-seq data.
| ACC | MCC | AUC | |
|---|---|---|---|
| CNN(AUTOKERAS) | 89.00% | 0.8779 | 0.9945 |
| SVM | 90.70% | 0.8970 | 0.9913 |
| RANDOM FOREST | 82.53% | 0.8062 | 0.9870 |
| DECISION TREE | 74.00% | 0.7112 | 0.8556 |
| T-GEM | 90.73% | 0.8971 | 0.9964 |
Figure 6T-GEM regulatory network for NK cell. Nodes are informative genes and their associated Key genes. Links go from Key to Query genes. Genes with red color are hub genes.
Figure 7T-GEM regulatory network for B cell. Nodes are informative genes and their associated Key genes. Links go from Key to Query genes. Genes with red color are hub genes.