| Literature DB >> 33093487 |
Yitan Zhu1, Thomas Brettin2, Yvonne A Evrard3, Alexander Partin2, Fangfang Xia2, Maulik Shukla2, Hyunseung Yoo2, James H Doroshow4, Rick L Stevens2,5.
Abstract
Transfer learning, which transfers patterns learned on a source dataset to a related target dataset for constructing prediction models, has been shown effective in many applications. In this paper, we investigate whether transfer learning can be used to improve the performance of anti-cancer drug response prediction models. Previous transfer learning studies for drug response prediction focused on building models to predict the response of tumor cells to a specific drug treatment. We target the more challenging task of building general prediction models that can make predictions for both new tumor cells and new drugs. Uniquely, we investigate the power of transfer learning for three drug response prediction applications including drug repurposing, precision oncology, and new drug development, through different data partition schemes in cross-validation. We extend the classic transfer learning framework through ensemble and demonstrate its general utility with three representative prediction algorithms including a gradient boosting model and two deep neural networks. The ensemble transfer learning framework is tested on benchmark in vitro drug screening datasets. The results demonstrate that our framework broadly improves the prediction performance in all three drug response prediction applications with all three prediction algorithms.Entities:
Mesh:
Substances:
Year: 2020 PMID: 33093487 PMCID: PMC7581765 DOI: 10.1038/s41598-020-74921-0
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Figure 1Analysis scenario framework. The analysis scheme on the left is ensemble transfer learning (ETL). The middle and right analysis schemes are standard cross-validation (SCV) and ensemble cross-validation (ECV), respectively, which do not apply transfer learning but instead analyze only the target dataset.
Figure 2Flowchart of ensemble transfer learning (ETL).
Figure 3Architectures of two DNN models used in the analysis. (a) Single-network DNN (sDNN) model. Gene expressions and drug descriptors are concatenated to form the input. (b) Two-subnetwork DNN (tDNN) model. The subnetworks take gene expressions and drug descriptors as separate inputs.
Comparison on the prediction performance of standard cross-validation (SCV), ensemble cross-validation (ECV), and ensemble transfer learning (ETL) for drug repurposing application.
| Target | Source | Model | RMSE (SCV) | RMSE (ECV) | RMSE (ETL) | P-value (RMSE, SCV vs. ETL) | P-value (RMSE, ECV vs. ETL) | Cor (SCV) | Cor (ECV) | Cor (ETL) | P-value (Cor, SCV vs. ETL) | P-value (Cor, ECV vs. ETL) |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| CCLE | CTRP | LightGBM | 0.0895 (0.0007) | 0.0872 (0.0009) | 0.0827 (0.0007) | 2.30E−11 | 1.36E−08 | 0.8313 (0.0029) | 0.8403 (0.0037) | 0.8581 (0.0023) | 4.30E−11 | 1.82E−08 |
| sDNN | 0.0895 (0.0013) | 0.0863 (0.0010) | 0.0812 (0.0007) | 3.13E−08 | 4.17E−07 | 0.8341 (0.0050) | 0.8466 (0.0045) | 0.8672 (0.0030) | 1.97E−08 | 4.72E−07 | ||
| tDNN | 0.0918 (0.0009) | 0.0867 (0.0009) | 4.96E−12 | 7.66E−11 | 0.8236 (0.0033) | 0.8435 (0.0030) | 2.85E−12 | 3.81E−11 | ||||
| CCLE | GDSC | LightGBM | 0.0895 (0.0007) | 0.0872 (0.0009) | 0.0839 (0.0009) | 5.06E−09 | 2.52E−07 | 0.8313 (0.0029) | 0.8403 (0.0037) | 0.8535 (0.0035) | 8.89E−10 | 3.13E−08 |
| sDNN | 0.0895 (0.0013) | 0.0863 (0.0010) | 0.0838 (0.0008) | 3.71E−07 | 2.43E−06 | 0.8341 (0.0050) | 0.8466 (0.0045) | 0.8562 (0.0037) | 6.10E−07 | 1.42E−05 | ||
| tDNN | 0.0918 (0.0009) | 0.0867 (0.0009) | 2.85E−10 | 9.30E−08 | 0.8236 (0.0033) | 0.8435 (0.0030) | 1.25E−10 | 1.55E−08 | ||||
| GCSI | CTRP | LightGBM | 0.1168 (0.0005) | 0.1142 (0.0007) | 0.1063 (0.0015) | 2.08E−09 | 3.89E−08 | 0.7889 (0.0018) | 0.7992 (0.0017) | 0.8293 (0.0048) | 2.11E−10 | 3.85E−09 |
| sDNN | 0.1167 (0.0025) | 0.1119 (0.0017) | 0.1051 (0.0014) | 7.05E−07 | 1.57E−06 | 0.7956 (0.0111) | 0.8118 (0.0057) | 0.8384 (0.0047) | 2.13E−06 | 1.26E−06 | ||
| tDNN | 0.1177 (0.0032) | 0.1109 (0.0014) | 7.93E−09 | 4.92E−09 | 0.7923 (0.0105) | 0.8133 (0.0050) | 1.72E−08 | 4.97E−09 | ||||
| GCSI | GDSC | LightGBM | 0.1168 (0.0005) | 0.1142 (0.0007) | 0.1059 (0.0015) | 1.99E−09 | 4.93E−08 | 0.7889 (0.0018) | 0.7992 (0.0017) | 0.8321 (0.0056) | 7.83E−10 | 7.97E−09 |
| sDNN | 0.1167 (0.0025) | 0.1119 (0.0017) | 0.1047 (0.0021) | 4.92E−06 | 1.57E−05 | 0.7956 (0.0111) | 0.8118 (0.0057) | 0.8419 (0.0048) | 2.43E−06 | 9.13E−07 | ||
| tDNN | 0.1177 (0.0032) | 0.1109 (0.0014) | 1.76E−09 | 1.12E−09 | 0.7923 (0.0105) | 0.8133 (0.0050) | 1.54E−09 | 9.77E−11 |
RMSE indicates the square root of mean square error. Cor indicates the Pearson correlation coefficient. In the RMSE and Cor columns, the number before a parenthesis is the average prediction performance and the number in a parenthesis is the standard deviation, calculated across 10 cross-validation trials. The p-values are generated by t-tests and indicate how significantly the prediction performance of ETL differs from those of SCV and ECV. SCV vs. ETL indicates comparison of SCV and ETL. ECV vs. ETL indicates comparison of ECV and ETL. The best average prediction performance for each transfer learning task is indicated with bold.
Comparison on the prediction performance of standard cross-validation (SCV), ensemble cross-validation (ECV), and ensemble transfer learning (ETL) for precision oncology application.
| Target | Source | Model | RMSE (SCV) | RMSE (ECV) | RMSE (ETL) | P-value (RMSE, SCV vs. ETL) | P-value (RMSE, ECV vs. ETL) | Cor (SCV) | Cor (ECV) | Cor (ETL) | P-value (Cor, SCV vs. ETL) | P-value (Cor, ECV vs. ETL) |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| CCLE | CTRP | LightGBM | 0.0913 (0.0015) | 0.0894 (0.0015) | 0.087 (0.0016) | 6.43E−06 | 1.08E−04 | 0.8245 (0.0045) | 0.8325 (0.0047) | 0.8419 (0.0056) | 3.75E−06 | 8.75E−05 |
| sDNN | 0.0915 (0.0014) | 0.0886 (0.0009) | 0.0858 (0.0014) | 5.99E−06 | 5.27E−07 | 0.8275 (0.0052) | 0.8385 (0.0041) | 3.67E−06 | 2.90E−05 | |||
| tDNN | 0.0909 (0.0013) | 0.0882 (0.0009) | 2.39E−06 | 3.89E−05 | 0.8293 (0.0040) | 0.8386 (0.0038) | 0.8476 (0.0037) | 2.69E−06 | 1.46E−04 | |||
| CCLE | GDSC | LightGBM | 0.0913 (0.0015) | 0.0894 (0.0015) | 0.0877 (0.0014) | 5.37E−07 | 1.58E−04 | 0.8245 (0.0045) | 0.8325 (0.0047) | 0.8389 (0.0045) | 6.01E−07 | 1.80E−04 |
| sDNN | 0.0915 (0.0014) | 0.0886 (0.0009) | 0.0888 (0.0013) | 3.27E−04 | 3.28E−01 | 0.8275 (0.0052) | 0.8385 (0.0041) | 0.8366 (0.0038) | 1.26E−05 | 4.86E−03 | ||
| tDNN | 0.0909 (0.0013) | 0.0882 (0.0009) | 2.55E−05 | 3.87E−03 | 0.8293 (0.0040) | 0.8386 (0.0038) | 1.22E−04 | 1.07E−02 | ||||
| GCSI | CTRP | LightGBM | 0.1186 (0.0023) | 0.116 (0.0026) | 0.1118 (0.0029) | 7.89E−05 | 1.75E−03 | 0.783 (0.0090) | 0.7929 (0.0094) | 0.8087 (0.0109) | 9.27E−05 | 1.69E−03 |
| sDNN | 0.123 (0.0043) | 0.1218 (0.0033) | 0.1118 (0.0016) | 1.25E−05 | 9.55E−06 | 0.7798 (0.0160) | 0.7938 (0.0082) | 0.8119 (0.0049) | 1.14E−04 | 5.43E−05 | ||
| tDNN | 0.1237 (0.0043) | 0.118 (0.0029) | 3.36E−06 | 5.19E−06 | 0.7804 (0.0084) | 0.7989 (0.0083) | 2.83E−07 | 2.44E−05 | ||||
| GCSI | GDSC | LightGBM | 0.1186 (0.0023) | 0.116 (0.0026) | 0.1099 (0.0020) | 8.89E−08 | 6.55E−06 | 0.783 (0.0090) | 0.7929 (0.0094) | 0.8162 (0.0072) | 1.59E−07 | 7.10E−06 |
| sDNN | 0.123 (0.0043) | 0.1218 (0.0033) | 0.1106 (0.0016) | 6.42E−06 | 3.48E−06 | 0.7798 (0.0160) | 0.7938 (0.0082) | 0.8156 (0.0063) | 3.76E−05 | 6.37E−07 | ||
| tDNN | 0.1237 (0.0043) | 0.118 (0.0029) | 2.34E−06 | 6.80E−07 | 0.7804 (0.0084) | 0.7989 (0.0083) | 4.06E−08 | 2.86E−07 |
RMSE indicates the square root of mean square error. Cor indicates the Pearson correlation coefficient. In the RMSE and Cor columns, the number before a parenthesis is the average prediction performance and the number in a parenthesis is the standard deviation, calculated across 10 cross-validation trials. The p-values are generated by t-tests and indicate how significantly the prediction performance of ETL differs from those of SCV and ECV. SCV vs. ETL indicates comparison of SCV and ETL. ECV vs. ETL indicates comparison of ECV and ETL. The best average prediction performance for each transfer learning task is indicated with bold.
Comparison on the prediction performance of standard cross-validation (SCV), ensemble cross-validation (ECV), and ensemble transfer learning (ETL) for the application of new drug development.
| Target | Source | Model | RMSE (SCV) | RMSE (ECV) | RMSE (ETL) | P-value (RMSE, SCV vs. ETL) | P-value (RMSE, ECV vs. ETL) | Cor (SCV) | Cor (ECV) | Cor (ETL) | P-value (Cor, SCV vs. ETL) | P-value (Cor, ECV vs. ETL) |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| CCLE | CTRP | LightGBM | 0.1828 (0.0249) | 0.1826 (0.0249) | 0.1589 (0.0125) | 2.32E−02 | 2.42E−02 | 0.0739 (0.0781) | 0.0778 (0.0815) | 0.3742 (0.1490) | 1.52E−04 | 1.45E−04 |
| sDNN | 0.2132 (0.0608) | 0.1964 (0.0460) | 0.158 (0.0152) | 3.20E−02 | 4.52E−02 | 0.0762 (0.0803) | 0.0685 (0.1638) | 0.4455 (0.0965) | 3.77E−05 | 8.99E−04 | ||
| tDNN | 0.206 (0.0637) | 0.205 (0.0602) | 4.98E−02 | 4.46E−02 | 0.0917 (0.1589) | 0.0937 (0.1446) | 1.20E−04 | 6.99E−05 | ||||
| CCLE | GDSC | LightGBM | 0.1828 (0.0249) | 0.1826 (0.0249) | 9.11E−05 | 9.57E−05 | 0.0739 (0.0781) | 0.0778 (0.0815) | 2.52E−09 | 2.80E−09 | ||
| sDNN | 0.2132 (0.0608) | 0.1964 (0.0460) | 0.146 (0.0201) | 1.04E−02 | 8.23E−03 | 0.0762 (0.0803) | 0.0685 (0.1638) | 0.5717 (0.0539) | 5.76E−08 | 8.65E−06 | ||
| tDNN | 0.206 (0.0637) | 0.205 (0.0602) | 0.1412 (0.0214) | 2.28E−02 | 1.92E−02 | 0.0917 (0.1589) | 0.0937 (0.1446) | 0.6124 (0.0638) | 1.09E−05 | 6.65E−06 | ||
| GCSI | CTRP | LightGBM | 0.2491 (0.0402) | 0.249 (0.0401) | 3.03E−03 | 3.02E−03 | 0.163 (0.0643) | 0.1707 (0.0599) | 1.41E−05 | 1.54E−05 | ||
| sDNN | 0.2804 (0.0584) | 0.3042 (0.0693) | 0.2243 (0.0346) | 1.54E−02 | 1.57E−02 | 0.0172 (0.2006) | − 0.2031 (0.1689) | 0.2231 (0.1606) | 8.78E−02 | 1.27E−03 | ||
| tDNN | 0.3043 (0.0931) | 0.2988 (0.0670) | 0.215 (0.0348) | 1.97E−02 | 6.88E−03 | − 0.1835 (0.1726) | − 0.1463 (0.2150) | 0.3707 (0.0751) | 8.90E−06 | 1.17E−04 | ||
| GCSI | GDSC | LightGBM | 0.2491 (0.0402) | 0.249 (0.0401) | 8.21E−03 | 8.23E−03 | 0.163 (0.0643) | 0.1707 (0.0599) | 3.11E−05 | 3.63E−05 | ||
| sDNN | 0.2804 (0.0584) | 0.3042 (0.0693) | 0.2255 (0.0239) | 2.42E−02 | 8.93E−03 | 0.0172 (0.2006) | − 0.2031 (0.1689) | 0.1092 (0.2477) | 4.89E−01 | 2.36E−03 | ||
| tDNN | 0.3043 (0.0931) | 0.2988 (0.0670) | 0.2148 (0.0295) | 2.13E−02 | 7.11E−03 | − 0.1835 (0.1726) | − 0.1463 (0.2150) | 0.3147 (0.0983) | 5.37E−06 | 2.87E−04 |
RMSE indicates the square root of mean square error. Cor indicates the Pearson correlation coefficient. In the RMSE and Cor columns, the number before a parenthesis is the average prediction performance and the number in a parenthesis is the standard deviation, calculated across 10 cross-validation trials. The p-values are generated by t-tests and indicate how significantly the prediction performance of ETL differs from those of SCV and ECV. SCV vs. ETL indicates comparison of SCV and ETL. ECV vs. ETL indicates comparison of ECV and ETL. The best average prediction performance for each transfer learning task is indicated with bold.