Literature DB >> 22295060

Shifting from population-wide to personalized cancer prognosis with microarrays.

Li Shao1, Xiaohui Fan, Ningtao Cheng, Leihong Wu, Haoshu Xiong, Hong Fang, Don Ding, Leming Shi, Yiyu Cheng, Weida Tong.   

Abstract

The era of personalized medicine for cancer therapeutics has taken an important step forward in making accurate prognoses for individual patients with the adoption of high-throughput microarray technology. However, microarray technology in cancer diagnosis or prognosis has been primarily used for the statistical evaluation of patient populations, and thus excludes inter-individual variability and patient-specific predictions. Here we propose a metric called clinical confidence that serves as a measure of prognostic reliability to facilitate the shift from population-wide to personalized cancer prognosis using microarray-based predictive models. The performance of sample-based models predicted with different clinical confidences was evaluated and compared systematically using three large clinical datasets studying the following cancers: breast cancer, multiple myeloma, and neuroblastoma. Survival curves for patients, with different confidences, were also delineated. The results show that the clinical confidence metric separates patients with different prediction accuracies and survival times. Samples with high clinical confidence were likely to have accurate prognoses from predictive models. Moreover, patients with high clinical confidence would be expected to live for a notably longer or shorter time if their prognosis was good or grim based on the models, respectively. We conclude that clinical confidence could serve as a beneficial metric for personalized cancer prognosis prediction utilizing microarrays. Ascribing a confidence level to prognosis with the clinical confidence metric provides the clinician an objective, personalized basis for decisions, such as choosing the severity of the treatment.

Entities:  

Mesh:

Year:  2012        PMID: 22295060      PMCID: PMC3266237          DOI: 10.1371/journal.pone.0029534

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Not all individuals respond to drug treatment in the same way. Accordingly, the development of personalized therapeutic regimens optimized for individual patients represents a major goal of 21st-century medicine [1]. Modern tools are being utilized to assist physicians in effectively treating patients as individuals and providing personalized drug intervention. Inter-individual variation in response to drug treatment is strongly influenced by a patient's physiological state at the time of treatment. This state can be characterized by gene expression profiles [2]. Therefore, microarray technology can guide the selection of drugs or therapeutic regimens and be employed to assess the susceptibility of a patient to certain diseases, enabling a personalized plan for prevention monitoring and treatment [3]. The prospective benefits of microarray technology in clinical applications have been demonstrated by several landmark studies [4]–[7]. Microarray-based predictive models (or genomic signatures) have shown utility in associating different subgroups of breast cancer with distinct clinical outcomes [8]–[13], such as MammaPrint™ [4], [5], a milestone in microarray-based prognosis for breast cancer [14]. The development of a microarray-based predictive model for tumor classification typically involves two sequential steps [4], [15]–[17]. First, the model is developed based on a training set of patients with known class labels (e.g., tumor status) and gene expression data. Next, the training model is validated using a validation set that contains patients with known class labels. The validity of the training model in performance on the validation set has been the focus of ‘class prediction’ research. To ensure the training model can be used in real-world clinical applications, it was suggested that the model must be assessed on a large number of independent samples in this external validation process [18]. It is important to note that the aforementioned external validation strategy assesses the performance of a training model using a population defined by the validation set. The average performance (e.g., specificity, sensitivity) over the population is used to assess whether the model can be a reliable diagnostic or prognostic test. This strategy is performed under the assumption that the model performs equally for everyone without considering the inter-individual variability. Thus, the average performance based on a population of patients cannot ensure its predictive ability for individual patients, which might result in potentially unreliable diagnoses or prognoses in the real-world application. This one-size-fits-all strategy needs to be modified from population-wide to personalized medicine in microarray data-based applications. We propose a metric called clinical confidence that measures the model's reliability in prediction performance on an individual basis. Clinical confidence can be useful in determining appropriate treatments; for example, patients with high confidence and poor prognosis may be given more rigorous treatments. Additionally, patients with lower clinical confidences may be prime candidates for further evaluation of their conditions with alternative methods. The accuracy of the clinical confidence metric was investigated on three large clinical datasets with total of six clinical endpoints [19]. Specifically, we first divided each dataset into two, i.e., the training and validation set. To mimic real-world clinical scenarios, we made the validation set that contains only the patients whose microarray data were generated at a later date than those in the training set. We derived the clinical confidence from the training model, followed by the assessment of its correlation with prediction accuracy for prognosis and the survival time of the patients in the validation set. To the best of our knowledge, this is the first attempt to provide a measure of confidence for individual patients in microarray-based “class prediction” research, which is an important step forward in personalized medicine.

Materials and Methods

Datasets

Three large-scale, clinical cancer datasets were used in this study: breast cancer (BR) [20], multiple myeloma (MM) [21], and neuroblastoma (NB) [22]. A concise summary of the datasets is given in . More detailed information of these datasets can be found in the main paper of the second phase of MicroArray Quality Control project (MAQC-II) [19].
Table 1

A concise summary of datasets.

Data Set codeEndpoint DescriptionEndpoint Code* Sample SizeRatio of eventsMicroarray Platform (number of channel)
TrainingValidationTrainingValidation
BRTreatment ResponseBR-pCR1301000.34 (33/97) 0.18 (15/85)Affymetrix U133A (1)
BR-erpos1301001.60 (80/50)1.56 (61/39)
MMOverall Survival Milestone OutcomeMM-OS3402140.18 (51/289)0.14 (27/187)Affymetrix U133Plus2.0 (1)
Event-free Survival Milestone OutcomeMM-EFS3402140.33 (84/256)0.19 (34/180)
NBOverall Survival Milestone OutcomeNB-OS2461770.32 (59/187)0.28 (39/138)Agilent NB Customized Array (2)
Event-free Survival Milestone OutcomeNB-EFS2461930.65 (97/149)0.75 (83/110)
ControlPositive controlNB-PC2462311.44 (145/101)1.36 (133/98)Agilent NB Customized Array (2)
MM-PC3402141.33 (194/146)1.89 (140/74)Affymetrix U133Plus2.0 (1)
Negative controlNB-NC2462531.44 (145/101)1.30 (143/110)Agilent NB Customized Array (2)
MM-NC3402141.43 (200/140)1.33 (122/92)Affymetrix U133Plus2.0 (1)

*BR - Breast Cancer; MM - Multiple Myeloma; NB - Neuroblastoma; pCR - Pathologic Complete Response; erpos – ER Positive; OS – Overall Survive; EFS – Event-free Survival; PC – Positive Control; NC – Negative Control.

Ratio of good to poor prognoses (i.e., good/poor prognoses).

*BR - Breast Cancer; MM - Multiple Myeloma; NB - Neuroblastoma; pCR - Pathologic Complete Response; erpos – ER Positive; OS – Overall Survive; EFS – Event-free Survival; PC – Positive Control; NC – Negative Control. Ratio of good to poor prognoses (i.e., good/poor prognoses). Each dataset has two clinical endpoints related to cancer prognosis (including survival data) or treatment: BR-pCR and BR-erpos in the treatment response dataset, NB-EFS and MM-EFS in the event-free survival dataset, and NB-OS and MM-OS in the overall survival dataset ( ). These three clinical datasets were studied in the MAQC-II project led by the FDA [19]. To emulate a real-world clinical scenario in applying genomic signatures, two independent populations of patients for each of the three clinical datasets were defined by the MAQC Consortium as the training and validation sets using a chronological approach where the samples in the validation sets were generated at a later date than those in the training sets. The sample sizes for the training sets varied between 130 and 340, and the ratio of positive events to negative events ranged from 0.18 to 1.60; meanwhile, the sample sizes in the validation sets ranged from 100 to 214, and the ratio of positive events to negative events varied between 0.14 and 1.56. Two positive and two negative control endpoints were also used in this study. The positive control endpoints, i.e., NB-PC and MM-PC, were derived from the NB and MM datasets separately, with the samples denoted by the gender. For the two negative control endpoints, i.e., NB-NC and MM-NC (which correspond to the NB and MM datasets, respectively), the sample labels (i.e., positive or negative events) were randomly generated. Using these two controls allow us to assess the performance of the clinically relevant endpoints against the expected maximum and minimum performance provided by the controls.

Clinical confidence

The clinical confidence measures the confidence of a sample being assigned to a specific class by a predictive model. For sample i, its clinical confidence metric () is the confidence level of a sample in being correctly assigned by a predictive model and is defined as:where and are the similarity measures between sample and samples in class 1 and class 2, respectively. The similarity measure varies according to classifiers used. Two well-studied classifiers for gene expression data were employed in this study, i.e., Nearest-Centroid classification rule (NC) [4] and k-nearest neighbors (kNN, k = 5) [23]. For the NC classifier, and were defined as the correlation coefficients of the unknown sample to the centroids of class 1 and class 2, respectively. The centroid is defined as vectors of the average expression values. For the kNN classifier, and are defined to be the number of nearest neighbors to the unknown sample belonging to class 1 and class 2, respectively. values range from 0.5 and 1 in which a value of 0.5 indicates that the prediction is due to chance. Increasingly larger values correspond to increasingly higher prediction confidence. For the sake of simplicity, all of the analysis was based on three confidence levels: low confidence (LC; 0.5≤≤0.6), medium confidence (MC; 0.6<≤0.8) and high confidence (HC; 0.8<≤1.0).

Statistical analysis

The general analysis workflow is depicted in ; additional details are provided in . The analysis protocol starts by developing a best classifier based on the training set, and ends by predicting the validation set. The predicted class and corresponding clinical confidences are recorded in matrices L and C, respectively. To ensure the statistical validity, the procedure is repeated 500 times, resulting in 500 different classifiers from the training sets and 500 predictions for the validation set. The performance of both training models and predictions is assessed using Matthews correlation coefficient (MCC) [24], [25].
Figure 1

Detailed workflow for correlation analysis of clinical confidence and model performance.

Additional details are provided in .

Detailed workflow for correlation analysis of clinical confidence and model performance.

Additional details are provided in . A permutation test was also employed to compare classifier prediction accuracy versus chance [26], [27]. In each permutation, the analysis protocol shown in was repeated with the exception that the class labels in the training set were randomized. In other words, models constructed with randomized training sets were utilized to predict the validation sets. After 500 repetitions, the degree of chance correlation and predictability of endpoints was computed with Cohen's d [28], which measures the standardized difference between two means.

Results

The cross-validation performance measured in MCC values for all the training models along with the average prediction performance on the validation sets are summarized in . The model performance follows the order of NB-PC, MM-PC, BR-erpos, NB-EFS, NB-OS, BR-pCR, MM-EFS, MM-OS, MM-NC, and NB-NC. The two positive controls performed best while the two negative controls perform worst, which is consistent with expectations from the experiment design.

Clinical confidence positively correlates with the model prediction performance

We first investigated the model performance on the validation set for patients falling into different categories of clinical confidence. As depicted in , a positive correlation is shown between the prediction accuracy and the confidence level for the six clinical and four control endpoints using the kNN classifier. Among the six clinical endpoints, the BR-erpos dataset showed the strongest correlation. For the BR-erpos, the average MCC value predictions with low confidence (LC) was only 0.19, while the average MCC value markedly increased to approximately 0.78 as the confidence level approached 1. Thus, compared to the overall MCC value (0.71) (), clinical confidence could successfully account for inter-individual variability in discriminating patients with lower or higher than average prediction accuracy.
Figure 2

Prediction MCC as a function of clinical confidence for ten datasets using kNN.

The Circle radii are scaled to the percentage of total samples in the clinical confidence level. The confidence levels are ‘0.6’, ‘0.8’, and ‘1’.

Prediction MCC as a function of clinical confidence for ten datasets using kNN.

The Circle radii are scaled to the percentage of total samples in the clinical confidence level. The confidence levels are ‘0.6’, ‘0.8’, and ‘1’. It is clear that the intrinsic predictability by gene expression profiles varies for different endpoints, as evidenced by the gradual decrease in the steepness of model performance for six clinical endpoints over different confidence intervals (i.e., the slopes in , data was shown in ) and the number of samples distributed across different confidence regions (i.e., the marker size in ). As shown in , a positive linear correlation was observed between the slope obtained from and the inherent predictability (quantified by Cohen's d [28]) of the six clinical and four control endpoints. The predictable endpoints (e.g., BR-erpos, NB-EFS) tended to have a larger percentage of patients (represented as the marker size in ) in the high confidence regions with high prediction accuracy than the less predictable endpoints (e.g., MM-EFS, MM-OS). Detailed information about sample distribution in each confidence region was given in . These observations were further verified using a different pattern recognition method (i.e., NC) (), and also a different sample splitting strategy (80/20 splitting, ).
Figure 3

Correlation between slope rate and Cohen's d for the kNN classifier.

The slopes are obtained from regression analysis based on the linear portion of the confidence-MCC curve, while Cohen's d represents the inherent predictability of the dataset.

Correlation between slope rate and Cohen's d for the kNN classifier.

The slopes are obtained from regression analysis based on the linear portion of the confidence-MCC curve, while Cohen's d represents the inherent predictability of the dataset. The results demonstrate that a higher inherent predictability of an endpoint is related to a higher percentage of patients that fall into higher confidence levels when using microarray-based predictive models. As the correlation of a genomic signature with a clinical outcome is rarely perfect, the clinical confidence could be useful to separate the patients into different groups for whom specific treatment procedures can be developed.

The relationship of clinical confidence with patient's survival time

We also evaluated whether clinical confidence is predictive of the survival rate for the patients in the validation set. The patients were divided into two prognosis groups (i.e., good and poor prognosis) for both NB and MM datasets with endpoints OS (overall survival) and EFS (event-free survival), respectively (). presents the OS curves for patients with different clinical confidences for both prognosis groups. Patients with high clinical confidence exhibited an increased survival rate in the good prognosis group and a decreased survival rate in the poor prognosis group, indicating that the clinical confidence enhanced the accuracy of prognosis derived from the predictive models. Taking MM-OS as an example, the survival rate is apparently higher for patients in the good prognosis group with high confidence (HC) compared to those with low (LC) (log-rank test p value<0.01) and medium (MC) ones (log-rank test p value 0.13), especially for each day mark more than 1000 days ( ). For patients with poor prognosis, more than 80% of those with low clinical confidence lived as long as 300 days, while approximately 30% of patients for those with high confidences survived at that time, respectively. Similar trends were also observed in the NB-OS dataset.
Figure 4

Overall survival (OS) curves for patients with different clinical confidences using kNN, where ‘LC’, ‘MC’, and ‘HC’ denote ‘low confidence (0.6)’, ‘medium confidence (0.8)’, and ‘high confidence (1)’, respectively.

, depicting the EFS curves for patients with different clinical confidences, demonstrates a similar trend as the OS curves presented in . The positive correlation of clinical confidence with EFS rate is clearly shown in for the patients with good prognosis. However, the correlation is less significant for patients with poor prognosis (). Corresponding results for 80/20 splitting was shown in , and conformed to those above-mentioned observations. The results demonstrate that once the patients were grouped into either good or poor prognosis groups by the predictive models, the clinical confidence can further characterize the survival rate of individual patients in each prognosis group.

Discussion

Several population-wide diagnostic/prognostic tests based on gene expression have been reported [4], [6], [7]. The population-based models provide only an average indication for the population with corresponding average population accuracy. In this study, we demonstrated that clinical confidence is both capable of separating patients that can be more reliably predicted from those that are less accurately predicted, and predictive of the survival rate for the patients after they are grouped into different prognostic groups. Thus, ascribing a confidence level to prognosis with the clinical confidence metric will provide the clinician a more personalized, objective basis for decisions when using biomarkers derived from microarray data. Specifically, we found that the clinical confidence provided a better estimation for the survival time when patients were classified into different prognosis categories based on both 70/30 and 80/20 sample assignments. For patients with good prognosis, higher clinical confidence was strongly correlated with longer survival time. Similarly, for patients with poor prognosis, the survival rate was significantly lower for those with high confidences than for the others. Taking endpoints MM-EFS and MM-OS as examples, despite the fact that they are rather difficult to be predicted, patients with high confidence display a significantly higher or lower survival rate when they are grouped in accordance with good or poor prognosis, respectively. Importantly, all patients in the high confidence group survived to 5000 days (), demonstrating that clinical confidence is an informative survival time prognosis tool. An important aspect of this study is that two positive (NB-PC, MM-PC) and two negative control (NB-NC, MM-NC) datasets were involved, which is essential to assess the performance of the clinically relevant endpoints against the theoretical maximum and minimum performance provided by the controls. Specifically, the positive correlation between model performance and clinical confidence for the two positive control datasets shown in confirmed the potential of clinical confidence to provide a measure of reliability for personalized medicine, while the negligible impact of clinical confidence in the two negative control datasets further limited the possibility of obtaining false positives. Thus, the inclusion of positive and negative control datasets in such an analysis would be of great help to ensure the reliability of the results. It remains enigmatic why some of the endpoints were more difficult than others to predict. and compare predictability across the three datasets and corresponding six endpoints. Readily predictable endpoints have a high percentage of patients who fall into the high confidence region. For example, the percentage of patients that showed high clinical confidence (74.70%) for the BR-erpos endpoint is much higher than that of the MM-EFS endpoint (37.51%) ( ), which may indicate that the BR-erpos endpoint contains a stronger gene expression signal than MM-EFS does. Additionally, the predictability of the dataset (Cohen's d) is directly related with the correlation coefficient between the confidence level and MCC prediction performance ( ). The ability to quantify clinical confidences may greatly enhance clinical decision-making processes based on microarray-based prediction models, especially for personalized treatment options. For example, the models presented here could test for potential treatment response with the high confidence and low confidence predictions being used in different ways. Patients with good prognosis and high confidences are candidates for applying routine protocols to avoid over-treatment, while rigorous strategies should be selected for those with poor prognosis and high confidences to prolong survival time as long as possible. However, for patients in the low confidence regions, additional evaluation using alternative methods should be considered. It is important to note that the strategy proposed in this study emphasizing the shift from population-based to personalized cancer prognosis does not negate the importance of population-based prediction, but rather builds upon its success. If the performance of a predictive model is not informative, such as seen in the two negative controls (i.e., MM-NC and NB-NC), the clinical confidence will not be predictive. Thus, model validation methods that include cross-validation and independent external validation are still essential to ensure the validity of microarray-based predictive models. However, since the population-based prediction does not provide an accurate assessment for each patient within the population, clinical confidence offers a means to measure reliability for individual predictions based on the population-based prediction. The benefits of personalized medicine in health care are well recognized [1]. It allows both the patient and the physician to be more aware of the benefits and risks of possible treatments and potential outcomes affected by genetic make-up or other environmental influences. Thus, informed, tailored, and health-related decisions can be made for each person [29]. Combining microarray technology capable of profiling the expression levels of hundreds of thousands of genes with pattern recognition techniques has been an important step toward individualized decision-making [30]. We presented examples applying confidence assessment to cancer prognosis and survival time prediction for models developed from microarray data. However, the approach can be generalized to biomarkers and models built based on data from other high throughput platforms. Moreover, the concept is generally applicable for all supervised learning classification methodologies that can define a clinical confidence. Prediction MCC as a function of clinical confidence for ten datasets using . Circle radii are scaled to the percentage of total samples in the clinical confidence level. The confidence levels are ‘0.5–0.6’, ‘0.6–0.8’ and ‘0.8–1’, respectively. (TIF) Click here for additional data file. Correlation between slope rate and Cohen's for the classifier. The slopes are obtained from regression analysis based on the linear portion of the confidence-MCC curve, while Cohen's d represents the inherent predictability of the dataset. (TIF) Click here for additional data file. Event-free survival (EFS) curves for patients with different clinical confidences using where ‘LC’, ‘MC’, and ‘HC’ denote ‘low confidence (0.6)’, ‘medium confidence (0.8)’, and ‘high confidence (1)’, respectively. (TIF) Click here for additional data file. Prediction MCC as a function of clinical confidence for ten datasets using 80/20 splitting and . The Circle radii are scaled to the percentage of total samples in the clinical confidence level. The confidence levels are ‘0.6’, ‘0.8’, and ‘1’. (TIF) Click here for additional data file. Correlation between slope rate and Cohen's for the classifier based on 80/20 sample assignment. The slopes are obtained from regression analysis based on the linear portion of the confidence-MCC curve, while Cohen's d represents the inherent predictability of the dataset. (TIF) Click here for additional data file. Overall survival (OS) curves for patients with different clinical confidences using 80/20 splitting and , where ‘LC’, ‘MC’, and ‘HC’ denote ‘low confidence (0.6)’, ‘medium confidence (0.8)’, and ‘high confidence (1)’, respectively. (TIF) Click here for additional data file. Event-free survival (EFS) curves for patients with different clinical confidences using 80/20 splitting and , where ‘LC’, ‘MC’, and ‘HC’ denote ‘low confidence (0.6)’, ‘medium confidence (0.8)’, and ‘high confidence (1)’, respectively. (TIF) Click here for additional data file. MCC performance for training and validation sets. (DOCX) Click here for additional data file. Slope and Cohen's d for each dataset. (DOCX) Click here for additional data file. Percentage of patients in low confidence (LC), medium confidence (MC) and high confidence (HC) regions. (DOCX) Click here for additional data file. Construction of the best classifier and calculate the correlation between clinical confidence and survival rate. (DOC) Click here for additional data file.
  27 in total

Review 1.  Assessing the accuracy of prediction algorithms for classification: an overview.

Authors:  P Baldi; S Brunak; Y Chauvin; C A Andersen; H Nielsen
Journal:  Bioinformatics       Date:  2000-05       Impact factor: 6.937

2.  A paradigm for class prediction using gene expression profiles.

Authors:  Michael D Radmacher; Lisa M McShane; Richard Simon
Journal:  J Comput Biol       Date:  2002       Impact factor: 1.479

Review 3.  Pitfalls in the use of DNA microarray data for diagnostic and prognostic classification.

Authors:  Richard Simon; Michael D Radmacher; Kevin Dobbin; Lisa M McShane
Journal:  J Natl Cancer Inst       Date:  2003-01-01       Impact factor: 13.506

4.  The MicroArray Quality Control (MAQC)-II study of common practices for the development and validation of microarray-based predictive models.

Authors:  Leming Shi; Gregory Campbell; Wendell D Jones; Fabien Campagne; Zhining Wen; Stephen J Walker; Zhenqiang Su; Tzu-Ming Chu; Federico M Goodsaid; Lajos Pusztai; John D Shaughnessy; André Oberthuer; Russell S Thomas; Richard S Paules; Mark Fielden; Bart Barlogie; Weijie Chen; Pan Du; Matthias Fischer; Cesare Furlanello; Brandon D Gallas; Xijin Ge; Dalila B Megherbi; W Fraser Symmans; May D Wang; John Zhang; Hans Bitter; Benedikt Brors; Pierre R Bushel; Max Bylesjo; Minjun Chen; Jie Cheng; Jing Cheng; Jeff Chou; Timothy S Davison; Mauro Delorenzi; Youping Deng; Viswanath Devanarayan; David J Dix; Joaquin Dopazo; Kevin C Dorff; Fathi Elloumi; Jianqing Fan; Shicai Fan; Xiaohui Fan; Hong Fang; Nina Gonzaludo; Kenneth R Hess; Huixiao Hong; Jun Huan; Rafael A Irizarry; Richard Judson; Dilafruz Juraeva; Samir Lababidi; Christophe G Lambert; Li Li; Yanen Li; Zhen Li; Simon M Lin; Guozhen Liu; Edward K Lobenhofer; Jun Luo; Wen Luo; Matthew N McCall; Yuri Nikolsky; Gene A Pennello; Roger G Perkins; Reena Philip; Vlad Popovici; Nathan D Price; Feng Qian; Andreas Scherer; Tieliu Shi; Weiwei Shi; Jaeyun Sung; Danielle Thierry-Mieg; Jean Thierry-Mieg; Venkata Thodima; Johan Trygg; Lakshmi Vishnuvajjala; Sue Jane Wang; Jianping Wu; Yichao Wu; Qian Xie; Waleed A Yousef; Liang Zhang; Xuegong Zhang; Sheng Zhong; Yiming Zhou; Sheng Zhu; Dhivya Arasappan; Wenjun Bao; Anne Bergstrom Lucas; Frank Berthold; Richard J Brennan; Andreas Buness; Jennifer G Catalano; Chang Chang; Rong Chen; Yiyu Cheng; Jian Cui; Wendy Czika; Francesca Demichelis; Xutao Deng; Damir Dosymbekov; Roland Eils; Yang Feng; Jennifer Fostel; Stephanie Fulmer-Smentek; James C Fuscoe; Laurent Gatto; Weigong Ge; Darlene R Goldstein; Li Guo; Donald N Halbert; Jing Han; Stephen C Harris; Christos Hatzis; Damir Herman; Jianping Huang; Roderick V Jensen; Rui Jiang; Charles D Johnson; Giuseppe Jurman; Yvonne Kahlert; Sadik A Khuder; Matthias Kohl; Jianying Li; Li Li; Menglong Li; Quan-Zhen Li; Shao Li; Zhiguang Li; Jie Liu; Ying Liu; Zhichao Liu; Lu Meng; Manuel Madera; Francisco Martinez-Murillo; Ignacio Medina; Joseph Meehan; Kelci Miclaus; Richard A Moffitt; David Montaner; Piali Mukherjee; George J Mulligan; Padraic Neville; Tatiana Nikolskaya; Baitang Ning; Grier P Page; Joel Parker; R Mitchell Parry; Xuejun Peng; Ron L Peterson; John H Phan; Brian Quanz; Yi Ren; Samantha Riccadonna; Alan H Roter; Frank W Samuelson; Martin M Schumacher; Joseph D Shambaugh; Qiang Shi; Richard Shippy; Shengzhu Si; Aaron Smalter; Christos Sotiriou; Mat Soukup; Frank Staedtler; Guido Steiner; Todd H Stokes; Qinglan Sun; Pei-Yi Tan; Rong Tang; Zivana Tezak; Brett Thorn; Marina Tsyganova; Yaron Turpaz; Silvia C Vega; Roberto Visintainer; Juergen von Frese; Charles Wang; Eric Wang; Junwei Wang; Wei Wang; Frank Westermann; James C Willey; Matthew Woods; Shujian Wu; Nianqing Xiao; Joshua Xu; Lei Xu; Lun Yang; Xiao Zeng; Jialu Zhang; Li Zhang; Min Zhang; Chen Zhao; Raj K Puri; Uwe Scherf; Weida Tong; Russell D Wolfinger
Journal:  Nat Biotechnol       Date:  2010-07-30       Impact factor: 54.908

5.  Gene expression profiling predicts clinical outcome of breast cancer.

Authors:  Laura J van 't Veer; Hongyue Dai; Marc J van de Vijver; Yudong D He; Augustinus A M Hart; Mao Mao; Hans L Peterse; Karin van der Kooy; Matthew J Marton; Anke T Witteveen; George J Schreiber; Ron M Kerkhoven; Chris Roberts; Peter S Linsley; René Bernards; Stephen H Friend
Journal:  Nature       Date:  2002-01-31       Impact factor: 49.962

6.  Gene expression patterns of breast carcinomas distinguish tumor subclasses with clinical implications.

Authors:  T Sørlie; C M Perou; R Tibshirani; T Aas; S Geisler; H Johnsen; T Hastie; M B Eisen; M van de Rijn; S S Jeffrey; T Thorsen; H Quist; J C Matese; P O Brown; D Botstein; P E Lønning; A L Børresen-Dale
Journal:  Proc Natl Acad Sci U S A       Date:  2001-09-11       Impact factor: 11.205

7.  Breast cancer classification and prognosis based on gene expression profiles from a population-based study.

Authors:  Christos Sotiriou; Soek-Ying Neo; Lisa M McShane; Edward L Korn; Philip M Long; Amir Jazaeri; Philippe Martiat; Steve B Fox; Adrian L Harris; Edison T Liu
Journal:  Proc Natl Acad Sci U S A       Date:  2003-08-13       Impact factor: 11.205

8.  Gene expression profiles predict complete pathologic response to neoadjuvant paclitaxel and fluorouracil, doxorubicin, and cyclophosphamide chemotherapy in breast cancer.

Authors:  M Ayers; W F Symmans; J Stec; A I Damokosh; E Clark; K Hess; M Lecocke; J Metivier; D Booser; N Ibrahim; V Valero; M Royce; B Arun; G Whitman; J Ross; N Sneige; G N Hortobagyi; L Pusztai
Journal:  J Clin Oncol       Date:  2004-05-10       Impact factor: 44.544

9.  A gene-expression signature as a predictor of survival in breast cancer.

Authors:  Marc J van de Vijver; Yudong D He; Laura J van't Veer; Hongyue Dai; Augustinus A M Hart; Dorien W Voskuil; George J Schreiber; Johannes L Peterse; Chris Roberts; Matthew J Marton; Mark Parrish; Douwe Atsma; Anke Witteveen; Annuska Glas; Leonie Delahaye; Tony van der Velde; Harry Bartelink; Sjoerd Rodenhuis; Emiel T Rutgers; Stephen H Friend; René Bernards
Journal:  N Engl J Med       Date:  2002-12-19       Impact factor: 91.245

10.  Repeated observation of breast tumor subtypes in independent gene expression data sets.

Authors:  Therese Sorlie; Robert Tibshirani; Joel Parker; Trevor Hastie; J S Marron; Andrew Nobel; Shibing Deng; Hilde Johnsen; Robert Pesich; Stephanie Geisler; Janos Demeter; Charles M Perou; Per E Lønning; Patrick O Brown; Anne-Lise Børresen-Dale; David Botstein
Journal:  Proc Natl Acad Sci U S A       Date:  2003-06-26       Impact factor: 12.779

View more
  5 in total

1.  Smad4 Loss Correlates With Higher Rates of Local and Distant Failure in Pancreatic Adenocarcinoma Patients Receiving Adjuvant Chemoradiation.

Authors:  Joseph M Herman; Salma K Jabbour; Steven H Lin; Matthew P Deek; Charles C Hsu; Elliot K Fishman; Sinae Kim; John L Cameron; Marina Chekmareva; Daniel A Laheru; Amol K Narang; Timothy M Pawlik; Ralph H Hruban; Christopher L Wolfgang; Christine A Iacobuzio-Donahue
Journal:  Pancreas       Date:  2018-02       Impact factor: 3.327

2.  Actions of Huangqi decoction against rat liver fibrosis: a gene expression profiling analysis.

Authors:  Gui-Biao Zhang; Ya-Nan Song; Qi-Long Chen; Shu Dong; Yi-Yu Lu; Ming-Yu Su; Ping Liu; Shi-Bing Su
Journal:  Chin Med       Date:  2015-12-18       Impact factor: 5.455

3.  A network pharmacology approach to evaluating the efficacy of chinese medicine using genome-wide transcriptional expression data.

Authors:  Leihong Wu; Yi Wang; Jing Nie; Xiaohui Fan; Yiyu Cheng
Journal:  Evid Based Complement Alternat Med       Date:  2013-05-12       Impact factor: 2.629

4.  CTen: a web-based platform for identifying enriched cell types from heterogeneous microarray data.

Authors:  Jason E Shoemaker; Tiago J S Lopes; Samik Ghosh; Yukiko Matsuoka; Yoshihiro Kawaoka; Hiroaki Kitano
Journal:  BMC Genomics       Date:  2012-09-06       Impact factor: 3.969

5.  Big data and computational biology strategy for personalized prognosis.

Authors:  Ghim Siong Ow; Zhiqun Tang; Vladimir A Kuznetsov
Journal:  Oncotarget       Date:  2016-06-28
  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.