Literature DB >> 32657371

AITL: Adversarial Inductive Transfer Learning with input and output space adaptation for pharmacogenomics.

Hossein Sharifi-Noghabi1,2, Shuman Peng1, Olga Zolotareva3, Colin C Collins2,4, Martin Ester1,2.   

Abstract

MOTIVATION: The goal of pharmacogenomics is to predict drug response in patients using their single- or multi-omics data. A major challenge is that clinical data (i.e. patients) with drug response outcome is very limited, creating a need for transfer learning to bridge the gap between large pre-clinical pharmacogenomics datasets (e.g. cancer cell lines), as a source domain, and clinical datasets as a target domain. Two major discrepancies exist between pre-clinical and clinical datasets: (i) in the input space, the gene expression data due to difference in the basic biology, and (ii) in the output space, the different measures of the drug response. Therefore, training a computational model on cell lines and testing it on patients violates the i.i.d assumption that train and test data are from the same distribution.
RESULTS: We propose Adversarial Inductive Transfer Learning (AITL), a deep neural network method for addressing discrepancies in input and output space between the pre-clinical and clinical datasets. AITL takes gene expression of patients and cell lines as the input, employs adversarial domain adaptation and multi-task learning to address these discrepancies, and predicts the drug response as the output. To the best of our knowledge, AITL is the first adversarial inductive transfer learning method to address both input and output discrepancies. Experimental results indicate that AITL outperforms state-of-the-art pharmacogenomics and transfer learning baselines and may guide precision oncology more accurately.
AVAILABILITY AND IMPLEMENTATION: https://github.com/hosseinshn/AITL. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
© The Author(s) 2020. Published by Oxford University Press.

Entities:  

Mesh:

Year:  2020        PMID: 32657371      PMCID: PMC7355265          DOI: 10.1093/bioinformatics/btaa442

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.937


1 Introduction

The goal of pharmacogenomics (Evans and Relling, 1999) is to predict response to a drug given some single- or multi-omics data. Since clinical datasets in pharmacogenomics (patients) are small and hard to obtain, it is not feasible to train a computational model only on patients. As a result, many studies have focused on large pre-clinical pharmacogenomics datasets such as cancer cell lines as a proxy to patients (Barretina ; Iorio ). A majority of the current computational methods are trained on cell line datasets and then tested on other cell line or patient datasets (Ding ; Geeleher , 2017; Güvenç ; Mourragui ; Rampášek ; Sakellaropoulos ; Sharifi-Noghabi ). However, cell lines and patients data, even with the same set of genes, do not have identical distributions due to the lack of an immune system and the tumor microenvironment in cell lines, which means a model cannot be trained on cell lines and then tested on patients (Mourragui ). Moreover, in cell lines, the response is often measured by the drug concentration that reduces viability by 50% (IC50), whereas in patients, it is often based on changes in the size of the tumor and measured by metrics such as response evaluation criteria in solid tumors (RECIST; Schwartz ). This means that drug response prediction is a regression problem in cell lines but a classification problem in patients. As a result, discrepancies exist in both the input and output spaces in pharmacogenomics datasets. Therefore, a need exists for a novel method to address these discrepancies to utilize cell line and patient data together to build a more accurate model eventually for patients. Transfer learning (Pan and Yang, 2009) attempts to solve this challenge by leveraging the knowledge in a source domain, a large data-rich dataset, to improve the generalization performance on a small target domain. Training a model on the source domain and testing it on the target domain violates the i.i.d assumption that the train and test data are from the same distribution. The discrepancy in the input space decreases the prediction accuracy on the test data, which leads to poor generalization (Zhang ). Transductive transfer learning (e.g. domain adaptation) and inductive transfer learning both use a labeled source domain to improve the generalization on a target domain. Transductive transfer learning assumes an unlabeled target domain, whereas inductive transfer learning assumes a labeled target domain where the label spaces of the source and target domain are different (Pan and Yang, 2009). Many methods have been proposed to minimize the discrepancy between the source and the target domains using different distribution metrics such as maximum mean discrepancy (Gretton ) In the context of drug response prediction, Mourragui proposed PRECISE, a subspace-centric method based on principal component analysis to minimize the discrepancy in the input space between cell lines and patients. Recently, adversarial domain adaptation has shown great performance in addressing the discrepancy in the input space for different applications, and its performance is comparable to the metric-based and subspace-centric methods in computer vision (Chen ; Ganin and Lempitsky, 2015; Hosseini-Asl ; Long ; Pinheiro, 2018; Tsai ; Tzeng ; Zou ). However, adversarial adaptation that addresses the discrepancies in both the input and output spaces have not yet been explored neither for pharmacogenomics nor for other applications. In this article, we propose Adversarial Inductive Transfer Learning (AITL), the first adversarial method of inductive transfer learning. Different from existing methods for inductive transfer learning as well as methods for adversarial transfer learning, AITL adapts not only the input space but also the output space. In pharmacogenomics, the source domain is the gene expression data obtained from the cell lines and the target domain is the gene expression data obtained from patients. Both domains have the same set of genes (i.e. raw feature representation). Discrepancies exist between the gene expression data in the input space, and the measure of the drug response in the output space. AITL learns features for the source and target samples and uses these features as input for a multi-task subnetwork to predict drug response for both the source and the target samples. The output space discrepancy is addressed by the multi-task subnetwork by assigning binary labels, called cross-domain labels, to the source samples which only have continuous labels. The multi-task subnetwork also alleviates the problem of small sample size in the target domain by joint training with the source domain. To address the discrepancy in the input space, AITL performs adversarial domain adaptation. The goal is that features learned for the source samples should be domain-invariant and similar enough to the features learned for the target samples to fool a global discriminator that receives samples from both domains. Moreover, with the cross-domain binary labels available for the source samples, AITL further regularizes the learned features by class-wise discriminators. A class-wise discriminator receives source and target samples from the same class label and should not be able to predict the domain accurately. We evaluated the performance of AITL and state-of-the-art pharmacogenomics and transfer learning methods on pharmacogenomics datasets in terms of the area under the receiver operating characteristic curve (AUROC) and the area under the precision-recall curve (AUPR). In our experiments, AITL achieved a substantial improvement compared with the baseline methods, demonstrating the potential of transfer learning for drug response prediction, a crucial task of precision oncology. Finally, we showed that the responses predicted by AITL for The Cancer Genome Atlas (TCGA) patients (without the drug response recorded) for breast, prostate, lung, kidney and bladder cancers had statistically significant associations with the level of expression of some of the annotated target genes for the studied drugs. This shows that AITL captures biological aspects of the response.

2 Background and related work

2.1 Transfer learning

Following the notation of (Pan and Yang, 2009), a domain like DM is defined by a raw input feature space (this is different from learned features by the network) X and a probability distribution p(X), where and x is the ith raw feature vector of X. A task is associated with , where is defined by a label space Y and a predictive function which is learned from training data of the form (X, Y), where and . A source task that is associated with a labeled source domain is defined as . is learned from training data . Similarly, a target task that is associated with a labeled target domain is defined as . is learned from training data . Since , it is challenging to train a model only on the target domain. Transfer learning addresses this challenge with the goal to improve the generalization on a target task using the knowledge in DM and DM, as well as their corresponding tasks and . Transfer learning addresses this challenge with the goal to improve the generalization on a target task using the knowledge in DM and DM, as well as their corresponding tasks and . Transfer learning can be categorized into three categories: (i) unsupervised transfer learning, (ii) transductive transfer learning and (iii) inductive transfer learning. In unsupervised transfer learning, there is no label in the source and target domains. In transductive transfer learning, the source domain is labeled, whereas the target domain is unlabeled. In this category, domains can be either the same or different (domain adaptation), but the source and target tasks are the same. In inductive transfer learning, the target domain is labeled and the source domain can be either labeled or unlabeled. In this category, the domains can be the same or different, but the tasks are always different (Pan and Yang, 2009).

2.2 Inductive transfer learning

There are three approaches to inductive transfer learning: (i) deep metric learning, (ii) learning and (iii) weight transfer (Scott ). Deep metric learning methods are independent of the number of samples in each class of the target domain, denoted as k. Few-shot learning methods focus on small k (). Finally, weight transfer methods require a large k ( or ; Scott ). In drug response prediction, a limited number of samples are for each class is available in the target domain; therefore, few-shot learning is more suitable for such a problem. Few-shot learning involves training a classifier to recognize new classes, provided only a small number of examples from each of these new classes in the training data (Snell ). Various methods have been proposed for few-shot learning (Chen ; Scott ; Snell ). For example, ProtoNet (Snell ) uses the source domain to learn how to extract features from the input and applies the feature extractor in the target domain. The mean feature of each class, obtained from the source domain, is used as the class prototype to assign labels to the target samples based on the Euclidean distance between a target sample’s feature and class prototypes.

2.3 Adversarial transfer learning

Recent advances in adversarial learning leverage deep neural networks to learn transferable representation that disentangles domain- and class-invariant features from different domains and matches them properly (Long ; Peng ; Zhang ). Generative adversarial networks (Goodfellow ) attempt to learn the distribution of the input data via a minimax framework where two networks are competing: a discriminator D and a generator G. The generator tries to create fake samples from a randomly sampled latent variable z that fool the discriminator, whereas the discriminator tries to catch these fake samples and discriminate them from the real ones. Therefore, the generator wants to minimize its error, whereas the discriminator wants to maximize its accuracy: A majority of literature on adversarial transfer learning are for transductive transfer learning, often referred to as domain adaptation, where the source domain is labeled while the target domain is unlabeled. Various methods have been proposed for adversarial transductive transfer learning in different applications such as image segmentation (Chen ), image classification (Tzeng ), speech recognition (Hosseini-Asl ) and adaptation under label-shift (Azizzadenesheli ). The idea of these methods is that features extracted from source and target samples should be similar enough to fool a global- (Tzeng ) and/or class-wise discriminators (Chen ).

3 Materials and methods

3.1 Problem definition

Given a labeled source domain DM with a learning task and a labeled target domain DM with a learning task , where , and , where , we assume that the source and the target domains are not the same due to different probability distributions. The goal of AITL is to utilize the source and target domains and their tasks in order to improve the learning of on DM. In the area of pharmacogenomics, the source domain is the gene expression data obtained from the cell lines, and the source task is to predict the drug response in the form of log (IC50) values. The target domain consists of gene expression data obtained from patients, and the target task is to predict drug response in a different form—often change in the size of tumor after receiving the drug. In this setting, because cell lines are different from patients even with the same set of genes. Additionally, because for the target task , drug response in patients is a binary outcome, but for the source task , drug response in cell lines is a continuous outcome. As a result, AITL needs to address these discrepancies in both the input and output spaces.

3.2 Adversarial Inductive Transfer Learning

Our proposed AITL method takes input data from the source and target domains, and achieves the following three objectives: first, it makes predictions for the target domain using both of the input domains and their corresponding tasks, second, it addresses the discrepancy in the output space between the source and target tasks and third, it addresses the discrepancy in the input space. AITL is a neural network consisting of four components: The feature extractor receives the input data from the source and target domains and extracts salient features, which are then sent to the multi-task subnetwork component. The multi-task subnetwork takes the extracted features of source and target samples and maps them to their corresponding labels and makes predictions for them. This component has a shared layer and two task-specific towers for regression (source task) and classification (target task). Therefore, by training the multi-task subnetwork on the source and target samples, it addresses the small sample size challenge in the target domain. In addition, it also addresses the discrepancy in the output space by assigning cross-domain labels (binary labels in this case) to the source samples (for which only continuous labels are available) using its classification tower. The global discriminator receives extracted features of source and target samples and predicts if an input sample is from the source or the target domain. To address the discrepancy in the input space, these features should be domain-invariant so that the global discriminator cannot predict their domain labels accurately. This goal is achieved by adversarial learning. The class-wise discriminators further reduce the discrepancy in the input space by adversarial learning at the level of the different classes, i.e. extracted features of source and target samples from the same class go to the discriminator for that class and this discriminator should not be able to predict if an input sample from a given class is from the source or the target domain. The AITL cost function consists of a classification loss, a regression loss, and global- and class-wise discriminator adversarial losses and is optimized end-to-end. An overview of the proposed method is presented in Figure 1.
Fig. 1.

Schematic overview of AITL: first, the feature extractor receives source and target samples and maps them to a feature space in lower dimensions. Then, the multi-task subnetwork uses these features to make predictions for the source and target samples and also assigns cross-domain labels to the source samples. The multi-task subnetwork addresses the discrepancy in the output space. Finally, to address the input space discrepancy, global- and class-wise discriminators receive the extracted features and regularize the feature extractor to learn domain-invariant features. The feature extractor has one fully connected layer. The multi-task subnetwork has one fully connected shared layer followed by two fully connected layers for the regression task and one fully connected layer for the classification task. All the discriminators are single-layered fully connected subnetworks

Schematic overview of AITL: first, the feature extractor receives source and target samples and maps them to a feature space in lower dimensions. Then, the multi-task subnetwork uses these features to make predictions for the source and target samples and also assigns cross-domain labels to the source samples. The multi-task subnetwork addresses the discrepancy in the output space. Finally, to address the input space discrepancy, global- and class-wise discriminators receive the extracted features and regularize the feature extractor to learn domain-invariant features. The feature extractor has one fully connected layer. The multi-task subnetwork has one fully connected shared layer followed by two fully connected layers for the regression task and one fully connected layer for the classification task. All the discriminators are single-layered fully connected subnetworks

Feature extractor

To learn salient features in lower dimensions for the input data, we design a feature extractor component. The feature extractor receives both the source and target samples as input and maps them to a feature space, denoted as Z. We denote the feature extractor as : where Z denotes the extracted features for input X which is from either the source (S) or the target (T) domain. In pharmacogenomics, the feature extractor learns features for the cell line and patient data.

Multi-task subnetwork

After extracting features of the input samples, we want to use these learned features to (i) make predictions for target samples, and (ii) address the discrepancy between the source and the target domains in the output space. To achieve these goals, a multi-task subnetwork with a shared layer and two task-specific towers, denoted as and , is designed, where M is for regression (the source task) and M is for classification (the target task): The performance of the multi-task subnetwork component is evaluated based on a binary-cross entropy loss for the classification task on the target samples and a mean squared loss for the regression task on the source samples: where Y and Y are the true labels of the source and the target samples, respectively, and L and L are the corresponding losses for the target and the source domains, respectively. The multi-task subnetwork outputs (i) the predicted continuous labels for the source samples, (ii) the predicted binary labels for the target samples and (iii) the assigned cross-domain binary labels for the source samples. The assigned cross-domain binary labels are obtained via the classification tower in the multi-task subnetwork which assigns binary labels (responder or non-responder) to the source samples because such labels do not exist for the source samples. Therefore, the multi-task subnetwork adapts the output space of the source and the target domains by assigning cross-domain labels to the source domain. In pharmacogenomics, the multi-task subnetwork predicts log (IC50) values for the cell lines and the binary response outcome for the patients. Moreover, it also assigns binary response labels to the cell lines which are similar to those of the patients.

Global discriminator

The goal of this component is to address the discrepancy in the input space by adversarial learning of domain-invariant features. To achieve this goal, a discriminator receives source and target extracted features from the feature extractor and classifies them into their corresponding domain. The feature extractor should learn domain-invariant features to fool the global discriminator. In pharmacogenomics, the global discriminator should not be able to recognize if the extracted features of a sample are from a cell line or a patient. This discriminator is denoted as . The adversarial loss for is as follows:

Class-wise discriminators

With cross-domain binary labels available for the source domain, AITL further reduces the discrepancy between the input domains via class-wise discriminators. The goal is to learn domain-invariant features with respect to specific class labels such that they fool corresponding class-wise discriminators. Therefore, extracted features of the target samples in class i, and those of the source domain which the multi-task subnetwork assigned to class i, will go to the discriminator for class i. We denote such a class-wise discriminator as DC. The adversarial loss for DC is as follows: In pharmacogenomics, the class-wise discriminator for the responder samples should not be able to recognize if the extracted features of a responder sample are from a cell line or a patient (similarly for a non-responder sample).

Cost function

To optimize the entire network in an end-to-end fashion, we design the cost function as follows: where, λ and λ are adversarial regularization coefficients for the global- and class-wise discriminators, respectively.

AITL architecture

The feature extractor is a one-layer fully-connected subnetwork with batch normalization using the ReLU activation function. The multi-task subnetwork has a shared fully connected layer with batch normalization and the ReLU activation function. The regression tower has two layers (one hidden layer and one output layer) with the ReLU activation function in the first layer and the linear activation function in the second one. The classification tower has one fully-connected layer with the Sigmoid activation function that maps the features to the binary outputs directly. Finally, the global- and class-wise discriminators are one-layer subnetworks with the Sigmoid activation function.

3.3 Drug response prediction for TCGA patients

To study AITL’s performance, similar to (Geeleher ; Sharifi-Noghabi ), we employ the model trained on Docetaxel, Paclitaxel or Bortezomib to predict the response for patients in several TCGA cohorts for which no drug response was recorded. For each drug, we extract the list of annotated target genes from the PharmacoDB resource (Smirnov ). We excluded Cisplatin because there was only one annotated target gene for it in PharmacoDB. To study associations between the level of expression of the extracted genes and the responses predicted by ATIL for each drug in each TCGA cohort, we fit multivariate linear regression models to the gene expression of those genes and the responses to that drug predicted by AITL. We obtain P-values for each gene and correct them for multiple hypotheses testing, using the Bonferroni correction (). The list of annotated target genes for each drug is available in the Supplementary Material, Section S1.

3.4 Datasets

In our experiments, we used the following datasets (see Table 1 for more detail):
Table 1.

Characteristics of the datasets

DatasetResourceDrugTypeDomainSample sizeNumber of genesa
GSE55145 (Amin et al., 2014)Clinical trialBortezomibtargetedTarget6711 609
GSE9782-GPL96 (Mulligan et al., 2007)Clinical trialBortezomibtargetedTarget16911 609
GDSC (Iorio et al., 2016)Cell lineBortezomibtargetedSource39111 609
GSE18864 (Silver et al., 2010)Clinical trialCisplatinChemotherapyTarget2411 768
GSE23554 (Marchion et al., 2011)Clinical trialCisplatinChemotherapyTarget2811 768
TCGA (Ding et al., 2016)PatientCisplatinChemotherapyTarget6611 768
GDSC (Iorio et al., 2016)Cell lineCisplatinChemotherapySource82911 768
GSE25065 (Hatzis et al., 2011)Clinical trialDocetaxelChemotherapyTarget498119
GSE28796 (Lehmann et al., 2011)Clinical trialDocetaxelChemotherapyTarget128119
GSE6434 (Chang et al., 2005)Clinical trialDocetaxelChemotherapyTarget248119
TCGA (Ding et al., 2016)PatientDocetaxelChemotherapyTarget168119
GDSC (Iorio et al., 2016)Cell lineDocetaxelChemotherapySource8298119
GSE15622 (Ahmed et al., 2007)Clinical trialPaclitaxelChemotherapyTarget2011 731
GSE22513 (Bauer et al., 2010)Clinical trialPaclitaxelChemotherapyTarget1411 731
GSE25065 (Hatzis et al., 2011)Clinical trialPaclitaxelChemotherapyTarget8411 731
PDX (Gao et al., 2015)Animal (mouse)PaclitaxelChemotherapyTarget4311 731
TCGA (Ding et al., 2016)PatientPaclitaxelChemotherapyTarget3511 731
GDSC (Iorio et al., 2016)Cell linePaclitaxelChemotherapySource38911 731

Number of genes in common between the source and all of the target data for each drug.

Characteristics of the datasets Number of genes in common between the source and all of the target data for each drug. The Genomics of Drug Sensitivity in Cancer (GDSC) cell lines dataset, consisting of a thousand cell lines from different cancer types, screened with 265 targeted and chemotherapy drugs (Iorio ). The Patient-Derived Xenograft (PDX) Encyclopedia dataset, consisting of more than 300 PDX samples for different cancer types, screened with 34 targeted and chemotherapy drugs (Gao ). TCGA (Weinstein ) containing a total number of 117 patients with diverse cancer types, treated with Cisplatin, Docetaxel or Paclitaxel (Ding ). Patient datasets from nine clinical trial cohorts containing a total number of 491 patients with diverse cancer types, treated with Bortezomib (Amin ; Mulligan ), Cisplatin (Marchion ; Silver ), Docetaxel (Chang ; Hatzis ; Lehmann ) or Paclitaxel (Ahmed ; Bauer ; Hatzis ). For the categorical measures of the drug response such as RECIST, we consider complete response and partial response as responder (Class 1) and consider stable disease and progressive disease as non-responder (Class 0). TCGA cohorts including, breast (BRCA), prostate (PRAD), lung (LUAD), kidney (KIRP) and bladder (BLCA) cancers that do not have the drug response outcome. The GDSC dataset was used as the source domain, and all the other datasets were used as the target domain. For the GDSC dataset, raw gene expression data were downloaded from ArrayExpress (E-MTAB-3610) and response outcomes from https:/www.cancerrxgene.org release 7.0. Gene expression data of TCGA patients were downloaded from the Firehose Broad GDAC (version published on January 28, 2016) and the response outcome was obtained from (Ding ). Patient datasets from clinical trials were obtained from the Gene Expression Omnibus and the PDX dataset was obtained from the Supplementary Material of Gao . For each drug, we selected those patient datasets that applied a comparable measure of the drug response. For preprocessing, the same procedure was adopted as described in the Supplementary Material of Sharifi-Noghabi ) for the raw gene expression data (normalized and z-score transformed) and the drug response data. After the pre-processing, source and target domains had the same number of genes.

4 Results

4.1 Experimental design

We designed our experiments to answer the following four questions: Does AITL outperform baselines that are trained only on cell lines and then evaluated on patients (without transfer learning)? To answer this question, we compared AITL against Geeleher and MOLI (Sharifi-Noghabi ) which are state-of-the-art methods of drug response prediction that do not perform domain adaptation. The Geeleher is non-deep learning method based on ridge regression and MOLI is a deep learning-based method. Both of them were originally proposed for pharmacogenomics. Does AITL outperform baselines that adopt adversarial transductive transfer learning and non-deep learning adaptation (without adaptation of the output space)? To answer this question, we compared AITL against ADDA (Tzeng ) and Chen , state-of-the-art methods of adversarial transductive transfer learning with global- and class-wise discriminators, respectively. For the non-deep learning baseline, we compared AITL to PRECISE (Mourragui ), a non-deep learning domain adaptation method specifically designed for pharmacogenomics. Does AITL outperform a baseline for inductive transfer learning? To answer this question, we compared AITL against ProtoNet (Snell ) which is a state-of-the-art inductive transfer learning method for small numbers of examples per class. Finally, do the predicted responses by AITL for TCGA patients have associations with the targets of the studied drug? Based on the availability of patient/PDX datasets for a drug, we experimented with four different drugs: Bortezomib, Cisplatin, Docetaxel and Paclitaxel. It is important to note that these drugs have different mechanisms and are being prescribed for different cancers. For example, Docetaxel is a chemotherapy drug mostly known for treating breast cancer patients (Chang ), whereas Bortezomib is a targeted drug mostly used for multiple myeloma patients (Amin ). Therefore, the datasets we have selected cover different types of anti-cancer drugs. In addition to the experimental comparison against published methods, we also performed an ablation study to investigate the impact of the different AITL components separately. AITL-AD denotes a version of AITL without the adversarial adaptation components, which means the network only contains the multi-task subnetwork. AITL-D denotes a version of AITL without the global discriminator, which means the network only employs the multi-task subnetwork and class-wise discriminators. AITL-D denotes a version of AITL without the class-wise discriminators, which means the network only contains the multi-task subnetwork and the global discriminator. All of the baselines were trained on the same data, tested on patients/PDX for these drugs, and eventually compared with AITL in terms of prediction AUROC and AUPR. Since the majority of the studied baselines cannot use the continuous log (IC50) values in the source domain, binarized log (IC50) labels provided by (Iorio ) using the Waterfall approach (Barretina ) were used to train them. Finally, for the minimax optimization, a gradient reversal layer was employed by AITL and the adversarial baselines (Ganin ) which is a well-established approach in domain adaptation (Long ; You ; Zhang ). We performed 3-fold cross-validation in the experiments to tune the hyper-parameters of AITL and the baselines based on the AUROC. Two folds of the source samples were used for training and the third fold for validation; similarly, two folds of the target samples were used for training and validation, and the third one for the test. The reported results refer to the average and standard deviation over the test folds. The hyper-parameters tuned for AITL were the number of nodes in the hidden layers, learning rates, mini-batch size, the dropout rate, number of epochs and the regularization coefficients. We considered different ranges for each hyper-parameter and the final selected hyper-parameter settings for each drug and each method are provided in the (Supplementary Material, Section S2). Finally, each network was re-trained on the selected settings using the train and validation data together for each drug. We used Adagrad for optimizing the parameters of AITL and the baselines (Duchi ) implemented in the PyTorch framework, except for Geeleher which was implemented in R. We used the author’s implementations for Geeleher , MOLI, PRECISE and ProtoNet. For ADDA, we used an existing implementation from https://github.com/jvanvugt/pytorch-domain-adaptation, and we implemented Chen from scratch.

4.2 Input and output space adaptation via AITL improves the drug response performance

Table 2 and Figure 2 report the performance of AITL and the baselines in terms of AUROC and AUPR, respectively. To answer the first experimental question, AITL was compared with the baselines which do not use any adaptation (neither the input nor the output space), i.e. Geeleher and MOLI (Sharifi-Noghabi ), and AITL demonstrated a better performance in both AUROC and AUPR for all of the studied drugs. This indicates that addressing the discrepancies in the input and output spaces leads to better performance compared with training a model on the source domain and testing it on the target domain. To answer the second experimental question, AITL was compared with state-of-the-art methods of adversarial and non-deep learning transductive transfer learning, i.e. ADDA (Tzeng ), Chen and PRECISE (Mourragui ), which address the discrepancy only in the input space. AITL achieved significantly better performance in AUROC for all of the drugs and for three out of four drugs in AUPR [the results of Chen for Cisplatin were very competitive with AITL]. This indicates that addressing the discrepancies in the both spaces outperforms addressing only the input space discrepancy. Finally, to answer the last experimental question, AITL was compared with ProtoNet (Snell )—a representative of inductive transfer learning with input space adaptation via few-shot learning. AITL outperformed this method in all of the metrics for all of the drugs.
Table 2.

Performance of AITL and the baselines in terms of the prediction AUROC

Method/drugBortezomibCisplatinDocetaxelPaclitaxel
Geeleher et al. (2014) 0.480.580.550.53
MOLI (Sharifi-Noghabi et al., 2019b)0.570.540.540.53
PRECISE (Mourragui et al., 2019)0.540.590.520.56
Chen et al. (2017) 0.54 ± 0.070.60 ± 0.140.52 ± 0.020.58 ± 0.04
ADDA (Tzeng et al., 2017)0.51 ± 0.060.56 ± 0.060.48 ± 0.06Did not converge
ProtoNet (Snell et al., 2017)0.49 ± 0.010.40 ± 0.0030.40 ± 0.01Did not converge
AITL-ADa0.69 ± 0.030.57 ± 0.030.57 ± 0.050.58 ± 0.01
AITL-DGb0.69 ± 0.040.62 ± 0.10.48 ± 0.03 0.62 ± 0.02
AITL-DCc0.69 ± 0.030.54 ± 0.10.59 ± 0.070.59 ± 0.03
AITL 0.74 ± 0.02 0.66 ± 0.02 0.64 ± 0.04 0.61 ± 0.04

AITL with only the multi-task subnetwork (no AD).

AITL with only class-wise discriminators and the multi-task subnetwork (no global discriminator).

AITL with only the global discriminator and the multi-task subnetwork (no class-wise discriminator). Boldface in the table indicates the best performing of the corresponding drug.

Fig. 2.

Performance of AITL and the baselines in terms of the prediction AUPR

Performance of AITL and the baselines in terms of the prediction AUPR Performance of AITL and the baselines in terms of the prediction AUROC AITL with only the multi-task subnetwork (no AD). AITL with only class-wise discriminators and the multi-task subnetwork (no global discriminator). AITL with only the global discriminator and the multi-task subnetwork (no class-wise discriminator). Boldface in the table indicates the best performing of the corresponding drug. We note that the methods of drug response prediction without adaptation, namely Geeleher and MOLI, outperformed the method of inductive transfer learning based on few-shot learning (ProtoNet). Moreover, these two methods also showed a very competitive performance compared with the methods of transductive transfer learning (ADDA, Chen and PRECISE). For Paclitaxel, ADDA did not converge in the first step (training a classifier on the source domain), which was also observed in another study (Sharifi-Noghabi ). ProtoNet also did not converge for this drug. We observed that AITL, when all of its components are used together, outperforms additional baselines with modified versions of AITL. This indicates the importance of both input and output space adaptation. The only exception was for the drug Paclitaxel, where AITL-D outperforms AITL. We believe the reason for this is that this drug has the most heterogeneous target domain (see Table 1), and therefore, the global discriminator component of AITL causes a minor decrease in the performance. Our ablation study showed that the global- and the class-wise discriminators are not redundant and, in fact, each of them plays a unique constructive role in learning the domain-invariant representation. All these results indicate that addressing the discrepancies in the input and output spaces between the source and target domains, via the AITL method, leads to a better prediction performance.

4.3 AITL predictions for TCGA patients have significant associations with target genes

To answer the last experimental question, we applied AITL models (trained on Docetaxel, Bortezomib and Paclitaxel) to the gene expression data without known drug response from TCGA (breast, prostate, lung, kidney and bladder cancers) and predicted the response for these patients separately. Based on the corrected P-values obtained from multiple linear regression, there are a number of statistically significant associations between the target genes of the studied drugs and the responses predicted by AITL. For example, in breast cancer, we observed statistically significant associations in MAP4 () for Doxetaxel, BLC2 () for Paclitaxel and PSMA4 () for Bortezomib. In prostate cancer, we observed statistically significant associations in MAP2 () for Docetaxel, TUBB () for Paclitaxel, and RELA () for Bortezomib. For bladder cancer, NR1I2 (P  =  0.04) for Docetaxel, MAP4 () for Paclitaxel and PSMA4 (P  =  0.001) for Bortezomib were significant. In kidney cancer, BLC2 () for Docetaxel, MAPT () for Paclitaxel and PSMD2 () for Bortezomib were significant. Finally, in lung cancer, MAP4 () for Docetaxel, TUBB () for Paclitaxel and RELA () for Bortezomib were significant. The complete list of statistically significant genes and their importance is presented in the Supplementary Material, Sections S3 and S4. All these results suggest that AITL predictions capture biological aspects of the drug response.

4.4 Discussion

To our surprise, ProtoNet and ADDA could not outperform Geeleher , MOLI and PRECISE. For ProtoNet, this may be due to the depth of the backbone network. A recent study has shown that a deeper backbone improves ProtoNet performance significantly in image classification Chen . However, in pharmacogenomics, employing a deep backbone is not realistic because of the much smaller sample size compared with an image classification application. Another limitation for ProtoNet is the imbalanced number of training examples in different classes in pharmacogenomics datasets. Specifically, the number of examples per class in the training episodes is limited to the number of samples of the minority class as ProtoNet requires the same number of examples from each class. For ADDA, this lower performance may be due to the lack of end-to-end training of the classifier along with the global discriminator of this method. The reason is that end-to-end training of the classifier along with the discriminators improved the performance of the second adversarial baseline (Chen ) in AUROC and AUPR compared with ADDA. Moreover, Chen also showed a relatively better performance in AUPR compared with Geeleher and MOLI. In pharmacogenomics, patient datasets with drug response are small or not publicly available due to privacy and/or data sharing issues. We believe including more patient samples and more drugs will increase generalization capability. In addition, recent pharmacogenomics studies have shown that using multi-omics data work better than using only gene expression (Sharifi-Noghabi ). In this work, we did not consider genomic data other than gene expression data due to the lack of patient samples with multi-omics data and drug response data publicly available; however, in principle, AITL can be extended to work with such data by adding separate feature extractors for each omics data type. This approach is particularly crucial if the different data types have different dimensionalities. Last but not least, we used pharmacogenomics as our motivating application for this new problem of transfer learning, but we believe that AITL can also be employed in other applications. For example, in slow progressing cancers such as prostate cancer, large patient datasets with gene expression and short-term clinical data (source domain) are available; however, patient datasets with long-term clinical data (target domain) are small. AITL may be beneficial to learn a model to predict these long-term clinical labels using the source domain and its short-term clinical labels (Sharifi-Noghabi ). Finally, although we designed the multi-task subnetwork for a regression task on the source domain and a classification task on the target domain, in principle, AITL can easily be modified to incorporate different types of outputs. We observed that predictions for TCGA samples tend to have a low variance. We believe the reason for that is first, we created target domains by pooling together samples from different patient datasets treated with the same drug; however, in reality each dataset has its own discrepancies compared with the other datasets within each target domain. Second, we trained the model using pan-cancer cell lines, however, the patient samples were cancer specific due to the lack of pan-cancer patient data with drug response which makes the trained model less applicable for pan-cancer resources such as TCGA. For future research directions, we believe that the TCGA dataset consisting of gene expression data of more than 12 000 patients (without drug response outcome) can be incorporated in an unsupervised transfer learning setting to learn better features that are domain-invariant between cell lines and cancer patients. The advantage of this approach is that we can keep the valuable patient datasets with drug response as an independent test set and not use it for training/validation. Another possible future direction is to incorporate domain-expert knowledge into the structure of the model. A recent study has shown that such a structure improves the drug response prediction performance on cell line datasets and, more importantly, provides an explainable model as well (Snow ).

5 Conclusion

In this article, we introduced a new problem in transfer learning motivated by applications in pharmacogenomics. Unlike domain adaptation that only requires adaptation in the input space, this new problem requires adaptation in both the input and output spaces. To address this problem, we proposed AITL, an Adversarial Inductive Transfer Learning method which, to the best of our knowledge, is the first method that addresses the discrepancies in both the input and output spaces. AITL uses a feature extractor to learn features for target and source samples. Then, to address the discrepancy in the output space, AITL utilizes these features as input of a multi-task subnetwork that makes predictions for the target samples and assigns cross-domain labels to the source samples. Finally, to address the input space discrepancy, AITL employs global and class-wise discriminators for learning domain-invariant features. In pharmacogenomics, AITL adapts the gene expression data obtained from cell lines and patients in the input space, and also adapts different measures of the drug response between cell lines and patients in the output space. In addition, AITL can also be employed in other applications such as predicting long-term clinical labels for slow progressing cancers. We evaluated AITL on four different drugs and compared it against state-of-the-art baselines in terms of AUROC and AUPR. The empirical results indicated that AITL achieved a significantly better performance compared with the baselines showing the benefits of addressing the discrepancies in both the input and output spaces. Finally, we analyzed AITL’s predictions for the studied drugs on breast, prostate, lung, kidney and bladder cancer patients in TCGA. We showed that AITL’s predictions have statistically significant associations with the level of expression of some of the annotated target genes for the studied drugs. We conclude that AITL may be beneficial in pharmacogenomics, a crucial task in precision oncology. Click here for additional data file.
  25 in total

Review 1.  Pharmacogenomics: translating functional genomics into rational therapeutics.

Authors:  W E Evans; M V Relling
Journal:  Science       Date:  1999-10-15       Impact factor: 47.728

2.  Precision Oncology beyond Targeted Therapy: Combining Omics Data with Machine Learning Matches the Majority of Cancer Cells to Effective Therapeutics.

Authors:  Michael Q Ding; Lujia Chen; Gregory F Cooper; Jonathan D Young; Xinghua Lu
Journal:  Mol Cancer Res       Date:  2017-11-13       Impact factor: 5.852

3.  Evaluating the molecule-based prediction of clinical drug responses in cancer.

Authors:  Zijian Ding; Songpeng Zu; Jin Gu
Journal:  Bioinformatics       Date:  2016-06-09       Impact factor: 6.937

4.  Gene expression profiling and correlation with outcome in clinical trials of the proteasome inhibitor bortezomib.

Authors:  George Mulligan; Constantine Mitsiades; Barb Bryant; Fenghuang Zhan; Wee J Chng; Steven Roels; Erik Koenig; Andrew Fergus; Yongsheng Huang; Paul Richardson; William L Trepicchio; Annemiek Broyl; Pieter Sonneveld; John D Shaughnessy; P Leif Bergsagel; David Schenkein; Dixie-Lee Esseltine; Anthony Boral; Kenneth C Anderson
Journal:  Blood       Date:  2006-12-21       Impact factor: 22.113

5.  A genomic predictor of response and survival following taxane-anthracycline chemotherapy for invasive breast cancer.

Authors:  Christos Hatzis; Lajos Pusztai; Vicente Valero; Daniel J Booser; Laura Esserman; Ana Lluch; Tatiana Vidaurre; Frankie Holmes; Eduardo Souchon; Hongkun Wang; Miguel Martin; José Cotrina; Henry Gomez; Rebekah Hubbard; J Ignacio Chacón; Jaime Ferrer-Lozano; Richard Dyer; Meredith Buxton; Yun Gong; Yun Wu; Nuhad Ibrahim; Eleni Andreopoulou; Naoto T Ueno; Kelly Hunt; Wei Yang; Arlene Nazario; Angela DeMichele; Joyce O'Shaughnessy; Gabriel N Hortobagyi; W Fraser Symmans
Journal:  JAMA       Date:  2011-05-11       Impact factor: 56.272

6.  BAD phosphorylation determines ovarian cancer chemosensitivity and patient survival.

Authors:  Douglas C Marchion; Hope M Cottrill; Yin Xiong; Ning Chen; Elona Bicaku; William J Fulp; Nisha Bansal; Hye Sook Chon; Xiaomang B Stickles; Siddharth G Kamath; Ardeshir Hakam; Lihua Li; Dan Su; Carolina Moreno; Patricia L Judson; Andrew Berchuck; Robert M Wenham; Sachin M Apte; Jesus Gonzalez-Bosquet; Gregory C Bloom; Steven A Eschrich; Said Sebti; Dung-Tsa Chen; Johnathan M Lancaster
Journal:  Clin Cancer Res       Date:  2011-08-17       Impact factor: 12.531

7.  Identification of markers of taxane sensitivity using proteomic and genomic analyses of breast tumors from patients receiving neoadjuvant paclitaxel and radiation.

Authors:  Joshua A Bauer; A Bapsi Chakravarthy; Jennifer M Rosenbluth; Deming Mi; Erin H Seeley; Nara De Matos Granja-Ingram; Maria G Olivares; Mark C Kelley; Ingrid A Mayer; Ingrid M Meszoely; Julie A Means-Powell; Kimberly N Johnson; Chiaojung Jillian Tsai; Gregory D Ayers; Melinda E Sanders; Robert J Schneider; Silvia C Formenti; Richard M Caprioli; Jennifer A Pietenpol
Journal:  Clin Cancer Res       Date:  2010-01-12       Impact factor: 12.531

8.  The Cancer Cell Line Encyclopedia enables predictive modelling of anticancer drug sensitivity.

Authors:  Jordi Barretina; Giordano Caponigro; Nicolas Stransky; Kavitha Venkatesan; Adam A Margolin; Sungjoon Kim; Christopher J Wilson; Joseph Lehár; Gregory V Kryukov; Dmitriy Sonkin; Anupama Reddy; Manway Liu; Lauren Murray; Michael F Berger; John E Monahan; Paula Morais; Jodi Meltzer; Adam Korejwa; Judit Jané-Valbuena; Felipa A Mapa; Joseph Thibault; Eva Bric-Furlong; Pichai Raman; Aaron Shipway; Ingo H Engels; Jill Cheng; Guoying K Yu; Jianjun Yu; Peter Aspesi; Melanie de Silva; Kalpana Jagtap; Michael D Jones; Li Wang; Charles Hatton; Emanuele Palescandolo; Supriya Gupta; Scott Mahan; Carrie Sougnez; Robert C Onofrio; Ted Liefeld; Laura MacConaill; Wendy Winckler; Michael Reich; Nanxin Li; Jill P Mesirov; Stacey B Gabriel; Gad Getz; Kristin Ardlie; Vivien Chan; Vic E Myer; Barbara L Weber; Jeff Porter; Markus Warmuth; Peter Finan; Jennifer L Harris; Matthew Meyerson; Todd R Golub; Michael P Morrissey; William R Sellers; Robert Schlegel; Levi A Garraway
Journal:  Nature       Date:  2012-03-28       Impact factor: 49.962

9.  The extracellular matrix protein TGFBI induces microtubule stabilization and sensitizes ovarian cancers to paclitaxel.

Authors:  Ahmed Ashour Ahmed; Anthony D Mills; Ashraf E K Ibrahim; Jillian Temple; Cherie Blenkiron; Maria Vias; Charlie E Massie; N Gopalakrishna Iyer; Adam McGeoch; Robin Crawford; Barbara Nicke; Julian Downward; Charles Swanton; Stephen D Bell; Helena M Earl; Ronald A Laskey; Carlos Caldas; James D Brenton
Journal:  Cancer Cell       Date:  2007-12       Impact factor: 31.743

10.  Clinical drug response can be predicted using baseline gene expression levels and in vitro drug sensitivity in cell lines.

Authors:  Paul Geeleher; Nancy J Cox; R Stephanie Huang
Journal:  Genome Biol       Date:  2014-03-03       Impact factor: 13.583

View more
  9 in total

1.  Network-based machine learning approach to predict immunotherapy response in cancer patients.

Authors:  JungHo Kong; Doyeon Ha; Juhun Lee; Inhae Kim; Minhyuk Park; Sin-Hyeog Im; Kunyoo Shin; Sanguk Kim
Journal:  Nat Commun       Date:  2022-06-28       Impact factor: 17.694

Review 2.  Computational estimation of quality and clinical relevance of cancer cell lines.

Authors:  Lucia Trastulla; Javad Noorbakhsh; Francisca Vazquez; James McFarland; Francesco Iorio
Journal:  Mol Syst Biol       Date:  2022-07       Impact factor: 13.068

3.  Prokaryotic and eukaryotic promoters identification based on residual network transfer learning.

Authors:  Xiao Liu; Yuqiao Xu; Yachuan Luo; Li Teng
Journal:  Bioprocess Biosyst Eng       Date:  2022-03-13       Impact factor: 3.210

Review 4.  Drug sensitivity prediction from cell line-based pharmacogenomics data: guidelines for developing machine learning models.

Authors:  Hossein Sharifi-Noghabi; Soheil Jahangiri-Tazehkand; Petr Smirnov; Casey Hon; Anthony Mammoliti; Sisira Kadambat Nair; Arvind Singh Mer; Martin Ester; Benjamin Haibe-Kains
Journal:  Brief Bioinform       Date:  2021-11-05       Impact factor: 13.994

5.  Predicting patient response with models trained on cell lines and patient-derived xenografts by nonlinear transfer learning.

Authors:  Soufiane M C Mourragui; Marco Loog; Daniel J Vis; Kat Moore; Anna G Manjon; Mark A van de Wiel; Marcel J T Reinders; Lodewyk F A Wessels
Journal:  Proc Natl Acad Sci U S A       Date:  2021-12-07       Impact factor: 12.779

Review 6.  Effectiveness of Artificial Intelligence for Personalized Medicine in Neoplasms: A Systematic Review.

Authors:  Sorayya Rezayi; Sharareh R Niakan Kalhori; Soheila Saeedi
Journal:  Biomed Res Int       Date:  2022-04-07       Impact factor: 3.246

7.  Transfer learning of clinical outcomes from preclinical molecular data, principles and perspectives.

Authors:  Axel Kowald; Israel Barrantes; Steffen Möller; Daniel Palmer; Hugo Murua Escobar; Anne Schwerk; Georg Fuellen
Journal:  Brief Bioinform       Date:  2022-05-13       Impact factor: 13.994

8.  Identification of phenocopies improves prediction of targeted therapy response over DNA mutations alone.

Authors:  Hamza Bakhtiar; Kyle T Helzer; Yeonhee Park; Yi Chen; Nicholas R Rydzewski; Matthew L Bootsma; Yue Shi; Paul M Harari; Marina Sharifi; Martin Sjöström; Joshua M Lang; Menggang Yu; Shuang G Zhao
Journal:  NPJ Genom Med       Date:  2022-10-17       Impact factor: 6.083

9.  TUGDA: task uncertainty guided domain adaptation for robust generalization of cancer drug response prediction from in vitro to in vivo settings.

Authors:  Rafael Peres da Silva; Chayaporn Suphavilai; Niranjan Nagarajan
Journal:  Bioinformatics       Date:  2021-05-17       Impact factor: 6.937

  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.