Literature DB >> 30158596

A novel logistic regression model combining semi-supervised learning and active learning for disease classification.

Hua Chai1, Yong Liang2, Sai Wang1, Hai-Wei Shen1.   

Abstract

Traditional supervised learning classifier needs a lot of labeled samples to achieve good performance, however in many biological datasets there is only a small size of labeled samples and the remaining samples are unlabeled. Labeling these unlabeled samples manually is difficult or expensive. Technologies such as active learning and semi-supervised learning have been proposed to utilize the unlabeled samples for improving the model performance. However in active learning the model suffers from being short-sighted or biased and some manual workload is still needed. The semi-supervised learning methods are easy to be affected by the noisy samples. In this paper we propose a novel logistic regression model based on complementarity of active learning and semi-supervised learning, for utilizing the unlabeled samples with least cost to improve the disease classification accuracy. In addition to that, an update pseudo-labeled samples mechanism is designed to reduce the false pseudo-labeled samples. The experiment results show that this new model can achieve better performances compared the widely used semi-supervised learning and active learning methods in disease classification and gene selection.

Entities:  

Mesh:

Year:  2018        PMID: 30158596      PMCID: PMC6115447          DOI: 10.1038/s41598-018-31395-5

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Identifying disease related genes and classifying the disease type using gene expression data is a very hot topic in machine learning. Many different models such as logistic regression model[1] and support vector machines (SVM)[2] have been applied in this area. However these supervised learning methods need a lot of labeled samples to achieve satisfactory results. Nevertheless in many biological datasets there is only a small size labeled data and remaining samples are unlabeled. Labeling these unlabeled samples manually is difficult or expensive; hence many unlabeled samples are left in the dataset. On the other hand, the proportion of small size labeled samples may not represent the real data distribution, which makes the classifier difficult to get the expected accuracy. Trying to improve the classification performance, many incrementally learning technologies such as semi-supervised learning (SSL)[3] and active learning (AL)[4] have been designed which utilize the unlabeled samples. AL tries to train an accurate prediction model with minimum cost of labeling the unlabeled samples manually. It selects most uncertain or informative unlabeled samples and annotates them by human experts. These labeled samples are included to the training dataset to improve the model performance. Uncertainty sampling[5] is the most popular AL strategy in practice because it does not require significant overhead to use. However one problem is that using uncertainty sampling may make the model to be short-sighted or biased[6]. What is more, though AL reduces the manpower work, manually labeling the selected samples by AL in biological experiments still cost much. In another way, SSL uses unlabeled data together with labeled data in the training process without any manual labeling. Many different SSL methods have been designed in machine learning including transductive support vector machines[7], graph-based methods[8], co-training[9], self-training[10] and so on. However[11] pointed out that the pseudo-labeled samples are annotated based on the labeled samples in the dataset, and they are easy to be affected by the high noisy samples. That is why SSL may not achieve satisfactory accuracy in some places. Many researchers found the complementarity between AL and SSL. Song combined the AL and SSL to extract protein interaction sentences[12], the most informative samples which were selected by AL-SVM were annotated by experts and then the classifier was retrained using SSL technology by the new dataset[13]. used a SSL technology to help AL select the query points more efficiently and further reducing the workload of manual classification. In[14] a SVM classifier was proposed to manually label the most uncertain samples and at the same time the other unlabeled samples were labeled by SSL, thus a faster convergence result was gained. The recent study[15] proposed by Lin designed a new active self-paced learning mechanism which combines the AL and SSL for face recognition. However, most attention of the methods combing SSL and AL are paid to the SVM model. The logistic regression model which widely used for disease classification is seldom mentioned. And also in these existing methods, the most informative samples selected by AL are manually annotated, this work maybe very expensive or time consuming in disease classification. Hence we design a new logistic regression model combining AL and SSL which meets the following requirements: The new model should be easily understood and applied. Our method should not require significant engineering overhead to use. In this new logistic regression model, we use uncertainty sampling to select the most informative samples in AL. Uncertainty sampling is fairly easily generalized to probabilistic structure prediction models. For logistic regression model, the sample probability closed to the decision boundary (probability ≈ 0.5) will suffice. In the new logistic regression model, self-training is used as a complement to AL. Self-training is one of the popular technologies used in SSL because of its fast speed and simplicity, and this method is a good way to solve the short-sighted problem in AL. In self-training the classifier is first trained by using the small size labeled samples, and then the obtained classifier will be used to label the high confidence samples in the unlabeled samples pool. These selected samples will be included into the training set and the classifier will be retrained. The cycle repeats until all the unlabeled samples have been used. In the logistic regression model, the samples which the probability closed to 0 or 1 can be seen as the high confidence samples. In our model, uncertainty sampling is used for avoiding the classifier being misled by high noisy samples, and self-training can avoid the model to be short-sighted or biased because of the high confidence samples’ compactness and consistency in the feature space[15]. By the complementarity of uncertainty sampling and self-training, it is easy to build a select-retrain circulation mechanism based on the samples’ probabilities estimated by the logistic classifier. The new model can achieve a satisfactory accuracy while labeling the samples automatically without manual labeling. Sometimes labeling the disease samples manually is difficult, expensive or time consuming. In our model the uncertain samples selected by AL are labeled by the last classifier automatically, it significantly reduces the burden of manual labeling. However how to ensure the correctness of these uncertain samples? The most uncertain samples mean the false pseudo-labeled samples are easy to be generated. On the other hand the most uncertain samples can be seen as the most informative samples in the logistic model, and the misclassified samples will degenerate the model performance obviously. Considering these samples are not removed or corrected in the standard AL and SSL methods, we design an update mechanism for the pseudo-labeled samples which makes the misclassified samples have chances to be corrected based on the new classifiers which generated in later training interactions.

Method

Logistic regression model

Supposing the biological dataset has n samples, which includes n labeled samples and n unlabeled samples, n = n + n. And this dataset contains p genes. represents the coefficients between the disease type Y and gene expression X. represents the individual sample, where y is disease type, represents the gene expression data, c represents the sample is labeled or unlabeled. The basic logistic regression model can be expressed as: The log-likelihood of the logistic regression method can be expressed as: Trying to identify disease related genes in the gene expression data, L1-norm regularization (Lasso) is added in the model:Where is the L1-norm regularization part and λ is the tuning parameter.

Uncertainty sampling

In the active learning part of our method, we use uncertainty sampling to select samples in the unlabeled dataset. In the logistic regression model the sample which probability close to the decision boundary (probability ≈ 0.5) can be seen as the most uncertain sample in AL. Hence an AL logistic regression model can be expressed as:where v is the weight parameter of the unlabeled samples, and the represents the selection function which can be used to generate the v, a is the control parameter. The selected unlabeled samples will be labeled manually and then included into the training dataset. The can be expressed as following:

Self-training

In the logistic regression model the sample probability closest to 0 or 1 can be seen as the high confidence sample. It is easy to find that the difference between the self-training and uncertainty sampling is that the selection criteria of identifying the used unlabeled samples. Hence the self-training logistic regression model is shown as:where w is the weight parameter of the unlabeled samples, the represents the selection function of self-training and γ is the control parameter. The is shown as:

The logistic regression model combining semi-supervised learning and active learning

In this paper we propose a novel logistic regression model combining SSL and AL with an update mechanism. The high confidence unlabeled samples selected by self-training can avoid the classifier to be short-sighted. The low confidence samples selected by uncertainty sampling prevent the classifier to be misled by high noisy samples which are offered by self-training. The model can be expressed as:where w is the weight parameter of the unlabeled samples given by SSL, and the v is the weight parameter of the unlabeled samples obtained by AL. Different from the ordinary AL methods, the unlabeled samples selected in our model are labeled by the learned classifier automatically. Considering the uncertainty of classified samples, the misclassified samples should have the chances to be revised in latter training iterations. The update mechanism is described below: If the sample is selected by SSL and the label has been changed by the classifier, this sample will be returned to the unlabeled sample pool and wait to be selected again. If the sample is selected by AL and the label has been changed, we revise the label of this sample and it will be put into the training dataset directly. The work flow of our proposed logistic regression model is show in Fig. 1:
Figure 1

The work flow of proposed logistic regression model combining SSL and AL.

Step 1: Firstly the labeled data will be used to learn an initial logistic regression model. Step 2: The logistic regression model will be used to label the unlabeled samples and the high value samples which are selected by SSL or AL will be included into the training dataset. Step 3: Update the logistic regression model using the new training dataset. Step 4: Identify the false pseudo-labeled samples. If they are selected by SSL, return them to the unlabeled sample pool. Otherwise, change their labels and put them into the training dataset directly. Step 5: The cycle will continue until all the unlabeled samples have been labeled or the run time exceeds the maximum number of iteration. The work flow of proposed logistic regression model combining SSL and AL. The algorithm of our proposed logistic regression model combining SSL and AL is given in below: The algorithm of the semi-supervised logistic regression model. The maximum iteration C is computed based on the step size SZ which is the selection range for identifying high value samples based on the pseudo-labeled samples’ probabilities. The prob is defined as the probability of the i pseudo-labeled sample which is estimated by the logistic model. Here we give an example to discuss the convergence of this model: if the SZ is set 0.2, the C is 5 (SZ *C = 1). In first iteration only the pseudo-labeled samples meeting the following conditions will be used: 0 < prob < 0.05 and 0.95 < prob < 1 (selected by SSL) & 0.45 < prob < 0.55 (selected by AL), here the initial probability range is 0.2; in the second iteration the range will be increased to 0 < prob < 0.1 and 0.9 < prob < 1 (selected by SSL) & 0.4 < prob < 0.6 (selected by AL), the probability range is increased to 0.4, SZ = 0.2 means in every iteration the range of probability will increase by 0.2. And while C = 5, the probability range is increased to 1, it means all the pseudo-labeled samples will be used. The commonly C is set 10 (SZ = 0.1) or 20 (SZ = 0.05). Sometimes before the iteration reaches the maximum iteration C, all the pseudo-labeled samples have been selected, especially while the SZ is set very small. For saving the computing time and cost, the program will be terminated early.

Results

Simulation experiments

The datasets used in simulation experiments are generated as following: Step 1: Supposing the dataset has n samples, and the number of the genes is 4000. In these 4000 genes we set 10 disease related genes, and the coefficients of the remaining 3990 genes are set zero. Step 2: The correlation coefficient p is set 0.3. where (i = 1, …, n) are generated independently from standard normal distribution Step 3: The sample is generated as: , where β0 is the intercept and ε is the randomly generated Gauss white noise. Step 4: The unlabeled data points are selected randomly, supposing in the dataset there are n labeled samples and n2 unlabeled samples, where n = n1 + n2. In Group A we suppose n1 = 100, n2 = 200; and in Group B n1 = 150, n2=300. We recorded the (), c = 0 means the corresponding y is unlabeled. In this paper we compare six different methods: the single logistic model with Lasso, the AL logistic model with Lasso (AL-lo), the self-training logistic model with Lasso (SSL-lo), the logistic model combining with AL and SSL which needs manual labeling (ASSL-lo), the auto logistic model with Lasso combining with AL and SSL without manual labeling and update mechanism (Auto-ASSL(A)), and the logistic model with Lasso combining with AL and SSL without manual labeling but using update mechanism (Auto-ASSL(B)). In AL-lo and ASSL-lo, about 40% unlabeled samples are labeled manually. The classification accuracy of the unlabeled data is used to evaluate the classification performances of different models. The number of selected correct genes (NC), the number of selected genes (NS), sensitivity and specificity are used to evaluate the gene selection performances of the methods. Supposing true positive (TP) is the number of identified disease related genes, false positive (FP) is the number of selected unrelated genes, false negative (FN) is the number of disease related genes which are missed, and true negative (TN) is the number of the unrelated genes that are abandon by different models. The sensitivity and specificity can be expressed as: The gene selection performances of different methods in simulation experiments are shown in Table 1, the results are the average of 100 runs of the program. It is easy to find the specificity obtained by single logistic regression model is highest than any other methods, it means it doesn’t select too many unrelated genes. However the lowest sensitivity shows single logistic regression model selects the least disease related genes. The AL-lo achieves a closed specificity value compared to single logistic model, but it identifies more disease related genes. The SSL-lo selects more disease related genes than single logistic model, and meanwhile many unrelated genes are also selected. Through combining the AL and SSL, the ASSL-lo identifies most disease related genes, but the problem is that it also selects more disease unrelated genes than SSL-lo. Auto-ASSL(A) selects less correct genes compared to ASSL, and the numbers of selected unrelated genes are closed. Compared to the Auto-ASSL(A), the gene selection performance obtained by Auto-ASSL(B) is obviously improved. The sensitivity obtained by Auto-ASSL(B) is only less than ASSL-lo but higher than any other methods, and the specificity is even more than the ASSL-lo. It shows that the Auto-ASSL(B) can achieve a balance between the sensitivity and specificity, and it has a strong ability to identify the disease related genes meanwhile eliminates the interference of unrelated genes.
Table 1

The gene selection performances of different methods in simulation experiments.

Group AGroup B
NCNSsensitivityspecificityNCNSsensitivityspecificity
logistic3.1514.050.3150.9954.8226.580.4820.989
AL-lo3.6517.800.3650.9925.1928.160.5190.988
SSL-lo3.8723.610.3870.9905.5145.400.5510.979
ASSL-lo5.3263.900.5320.9716.7496.270.6740.955
Auto-ASSL(A)3.5957.680.3590.9735.2697.390.5260.953
Auto-ASSL(B)4.1727.450.4170.9885.7553.600.5750.976
The gene selection performances of different methods in simulation experiments. The values of classification accuracy obtained by different methods in the unlabeled data are shown in Fig. 2. The ROC curves obtained by different methods in one run of the program are shown in Fig. 3. And the AUC values corresponding to the ROC curves are given in Table 2. The ASSL logistic model achieves the best result through combining the AL and SSL, however it needs much time and cost for manual labeling. The performance obtained by Auto-ASSL(A) is even worse than SSL logistic model, this result proves the misclassified uncertain samples have significant bad effect on the classification performance and our update mechanism is very necessary for Auto-ASSL. The results show our method is advanced because it achieves higher accuracy than AL-lo or SSL-lo and only less than ASSL-lo, and meanwhile it doesn’t need any manual labeling.
Figure 2

The classification accuracy of different methods in simulation experiments.

Figure 3

The ROC curves of different methods in simulation experiments.

Table 2

The AUC obtained by different methods in simulation experiments.

AUClogisticAL-loSSL-loASSL-loAuto-ASSL(A)Auto-ASSL(B)
Group A0.95840.97230.97090.98740.94480.9810
Group B0.96820.98550.97960.99430.97380.9917
The classification accuracy of different methods in simulation experiments. The ROC curves of different methods in simulation experiments. The AUC obtained by different methods in simulation experiments. Hence the new logistic regression model combing SSL and AL can be seen as a very efficient method because it implements the following functions: It works without any manual intervention. This saves much cost and the results can be quickly obtained. It can achieve accuracy above 90% in disease classification. The experiments show our method can achieve a better accuracy than the AL and SSL logistic regression models. It can identify more disease related genes and at the same time less unrelated genes will be selected. This further saves the researchers’ time and cost.

Real data experiments

In real data experiments six methods are applied on four real gene expression datasets: Diffuse large B-cell lymphoma (DLBCL) dataset[16], Prostate cancer dataset[17], GSE21050[18] and GSE32603[19]. In these four datasets about 2/3 samples are treated as the unlabeled samples for evaluating the classification accuracy of unlabeled samples. The labeled samples and unlabeled samples are randomly selected in every runs of the program. More details of the datasets used in the experiments are shown in Table 3.
Table 3

Details of real datasets used in the experiments.

DatasetNumber of genesNumber of samplesNumber of labeled samplesDisease types
DLBCL26487726diffuse large b-cell lymphoma
Prostate213510234prostate cancer
GSE2105054613310103soft tissue sarcomas
GSE326031320023177breast cancer
Details of real datasets used in the experiments. The values of classification accuracy obtained by different methods in real datasets are shown in Table 4. The ROC curves obtained by different methods in one run of the program in different datasets are shown in Fig. 4, and the corresponding AUC are shown in Table 5. The SSL-lo performs better than single logistic and Auto-ASSL(A), but worse than the other three methods. It is obviously that the accuracy of ASSL logistic model is highest. The Auto-ASSL(A) does not perform well because the misclassified samples affect the accuracy. The classification accuracy obtained by Auto-ASSL(B) is better than any other methods except ASSL which proves that the update pseudo-labeled samples mechanism is a very important improvement for the model.
Table 4

The classification accuracy obtained by different methods in the real datasets.

MethodDLBCLProstateGSE21050GSE32603
logistic77.94%86.54%79.01%69.69%
AL-lo83.15%91.53%84.43%73.68%
SSL-lo81.82%88.97%80.92%70.57%
ASSL-lo87.14%94.42%89.34%80.63%
Auto-ASSL(A)80.67%88.55%78.33%68.48%
Auto-ASSL(B)85.62%93.36%86.37%76.46%
Figure 4

ROC curves obtained by different methods in real datasets (a) DLBCL (b) Prostate (c) GSE21050 (d) GSE32603.

Table 5

The AUC obtained by different methods in the real datasets.

MethodDLBCLProstateGSE21050GSE32603
logistic0.91990.95690.89750.7557
AL-lo0.92950.97490.93940.8328
SSL-lo0.89420.97080.92320.7962
ASSL-lo0.95830.98620.95960.9023
Auto-ASSL(A)0.83330.96460.88350.7757
Auto-ASSL(B)0.93910.97850.94320.8390
The classification accuracy obtained by different methods in the real datasets. ROC curves obtained by different methods in real datasets (a) DLBCL (b) Prostate (c) GSE21050 (d) GSE32603. The AUC obtained by different methods in the real datasets. The numbers of genes selected by different methods in real dataset are shown in Fig. 5. It is obvious that the single logistic method selects least genes. The numbers of selected genes obtained by ASSL and Auto-ASSL(A) are far more than other methods. Our method selects more genes than AL-lo and SSL-lo, but less than ASSL-lo.
Figure 5

The number of genes selected by different methods in real datasets.

The number of genes selected by different methods in real datasets. In order to further assess the correctness of the selected genes by different methods, the top-10 ranked genes selected by different methods in real datasets are listed in Tables 6–9, Table 9 is partly blank because the methods didn’t select so many genes. The genes in italic in the tables such as SELENOP, HPN, MTHFD2 and ROR2 are the ones which are selected by all the methods in the same datasets. The SELENOP in DLBCL can be seen as an extracellular antioxidant, and it may be potential non-invasive diagnostic markers for cancer. Some researches show that selenium could be seen as an anticancer therapy by affecting SELENOP[20]. The research has proved that expression of the encoded protein of HPN is related to the growth and progression of cancers, particularly prostate cancer. It may be associated with susceptibility to prostate cancer[21]. The MTHFD2 in GSE21050 is seen as a prognostic factor and a potential therapeutic target for future cancer treatments.[22]. The ROR2 in GSE32603 is reported that it can significantly reduce cell proliferation and induced apoptosis[23].
Table 6

The genes selected by different methods in DLBCL.

logisticAL-loSSL-loASSL-loAuto-ASSL(A)Auto-ASSL(B)
1 SELENOP SELENOP SELENOP SELENOP SELENOP MDM4
2KIF2CMT2APURACD34GPR18 SELENOP
3MT2AMIFMT2ATXNIPESDMIF
4MORC3 GLIPR1 GLIPR1 MT2A GLIPR1 MORC3
5TLE4SELLTLE4PURASELLTLE4
6SELLBMI1MIFTRIB2MYCLP1SELL
7N4BP2L1IFITM2N4BP2L1GAPDHTRIM23N4BP2L1
8 GLIPR1 GAPDHSELLMYCLP1TLE4 GLIPR1
9EFNA3CCL21CCL21 GLIPR1 KIF2CCCL21
10MYCLP1SMAD6ESDMIFGAPDHMT2A
Table 9

The genes selected by different methods in GSE32603.

logisticAL-loSSL-loASSL-loAuto-ASSL(A)Auto-ASSL(B)
1 ROR2 GRB2LOC642236 ROR2 ROR2 EMB
2LOC642236MS4A1 ROR2 EMB EMB ROR2
3GRB2 EMB GRB2ZSCAN9GRB2GRB2
4 EMB ROR2 EMB CDKN1BTMEM242TMEM242
5MAST1UBE2WUFC1HPSE
6C2orf70EPHB1 TPD52L2
7TAF8SUPT20H
8MTSS1ARL2BP
9PRDM4STK3
The genes selected by different methods in DLBCL. The genes selected by different methods in Prostate. The genes selected by different methods in GSE21050. The genes selected by different methods in GSE32603. On the other hand, our method also identified some special genes which other methods did not select. These genes are shown in bold in the Tables 6–9. The MDM4 in DLBCL plays a very important role in the proliferation of the cancer cells, and it is crucial for the establishment and progression of tumors[24]. JUNB plays a specific role in cancer cell proliferation, survival and drug resistance[25]. Single nucleotide polymorphism of TIPARP in Prostate has been proved to be related with cancer[26]. In[27] ENO2 is reported to be a risk factor for bone metastases in cancer. The TPD52L2 in GSE32603 encodes a member of the tumor protein D52-like family, and contributes to proliferation of cancer cells[28]. These genes mentioned in the literatures demonstrate that our new logistic regression model has a strong ability in gene selection.

Conclusion

In this paper we have designed a novel method which does not require significant engineering overhead to use and meanwhile achieves satisfying results by utilizing the unlabeled gene expression samples in disease classification. The novel logistic regression model is designed based on the complementarity of semi-supervised learning and active learning. In addition to that an update pseudo-labeled samples mechanism is embedded in this method to reduce the false pseudo-labeled samples. In conclusion, our method can achieve more accuracy results compared widely used SSL and AL logistic models, and it also has a good performance in identifying the disease related genes. In addition to that, this model can work without any manual labeling for saving much time and cost. We believe it will be an efficient tool to make contributions for disease classification and gene selection because of its high reliability and stability against noise and outliers.
Table 7

The genes selected by different methods in Prostate.

logisticAL-loSSL-loASSL-loAuto-ASSL(A)Auto-ASSL(B)
1 HPN TP63 TP63 TP63 PTGDS TP63
2 TP63 XBP1XBP1XBP1 HPN HPN
3MYOFNELL2 HPN HPN NELL2MYOF
4XBP1TGFB3PTGDSNELL2RRADXBP1
5PTGDS HPN NELL2RBM3HSBP1 JUNB
6NELL2ATP5MEMYOFPTGDSMYOFNELL2
7SERPINA3TRIM29ATP5MESDC1 TP63 SERPINA3
8RBM3MYOFSERPINA3CFDPDLIM5 TIPARP
9TGFB3RBM3TGFB3ATP5MEATP5METGFB3
10TRIM29SERPINA3TRIM29HSBP1SERPINA3TRIM29
Table 8

The genes selected by different methods in GSE21050.

logisticAL-loSSL-loASSL-loAuto-ASSL(A)Auto-ASSL(B)
1C15orf41 SNAPC1 SNAPC1 FADS1 MTHFD2 MTHFD2
2 SNAPC1 SNORD35BC8orf82SNORD35BADD3SNORD35B
3C8orf82 MTHFD2 MTHFD2 IFT43 SNAPC1 SNAPC1
4 MTHFD2 NFATC2IPSLC1A4C8orf82SNORD35BADD3
5LPAR1C8orf82PMLCDC42EP3FHL2C8orf82
6AKT2NUP155PLD1 MTHFD2 PCDH18XPO6
7XPO6XPO6WDHD1DCNNFATC2IPATP6V1D
8SLC1A4IFT43AKT2 SNAPC1 YEATS2IFT43
9PLD1PCDH18RPL13AXPO6LIMK2NUP155
10SNORD35BWDHD1NFATC2IPADD3SMAD4 ENO2
  16 in total

1.  Diffuse large B-cell lymphoma outcome prediction by gene-expression profiling and supervised machine learning.

Authors:  Margaret A Shipp; Ken N Ross; Pablo Tamayo; Andrew P Weng; Jeffery L Kutok; Ricardo C T Aguiar; Michelle Gaasenbeek; Michael Angelo; Michael Reich; Geraldine S Pinkus; Tane S Ray; Margaret A Koval; Kim W Last; Andrew Norton; T Andrew Lister; Jill Mesirov; Donna S Neuberg; Eric S Lander; Jon C Aster; Todd R Golub
Journal:  Nat Med       Date:  2002-01       Impact factor: 53.440

2.  Variants in the HEPSIN gene are associated with susceptibility to prostate cancer.

Authors:  H J Kim; J H Han; I H Chang; W Kim; S C Myung
Journal:  Prostate Cancer Prostatic Dis       Date:  2012-06-05       Impact factor: 5.554

3.  Towards Making Unlabeled Data Never Hurt.

Authors:  Yu-Feng Li; Zhi-Hua Zhou
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2015-01       Impact factor: 6.226

4.  The AP-1 transcription factor JunB is essential for multiple myeloma cell proliferation and drug resistance in the bone marrow microenvironment.

Authors:  F Fan; M H Bashari; E Morelli; G Tonon; S Malvestiti; S Vallet; M Jarahian; A Seckinger; D Hose; L Bakiri; C Sun; Y Hu; C R Ball; H Glimm; M Sattler; H Goldschmidt; E F Wagner; P Tassone; D Jaeger; K Podar
Journal:  Leukemia       Date:  2016-11-28       Impact factor: 11.528

5.  A genome-wide association study identifies susceptibility loci for ovarian cancer at 2q31 and 8q24.

Authors:  Ellen L Goode; Georgia Chenevix-Trench; Honglin Song; Susan J Ramus; Maria Notaridou; Kate Lawrenson; Martin Widschwendter; Robert A Vierkant; Melissa C Larson; Susanne K Kjaer; Michael J Birrer; Andrew Berchuck; Joellen Schildkraut; Ian Tomlinson; Lambertus A Kiemeney; Linda S Cook; Jacek Gronwald; Montserrat Garcia-Closas; Martin E Gore; Ian Campbell; Alice S Whittemore; Rebecca Sutphen; Catherine Phelan; Hoda Anton-Culver; Celeste Leigh Pearce; Diether Lambrechts; Mary Anne Rossing; Jenny Chang-Claude; Kirsten B Moysich; Marc T Goodman; Thilo Dörk; Heli Nevanlinna; Roberta B Ness; Thorunn Rafnar; Claus Hogdall; Estrid Hogdall; Brooke L Fridley; Julie M Cunningham; Weiva Sieh; Valerie McGuire; Andrew K Godwin; Daniel W Cramer; Dena Hernandez; Douglas Levine; Karen Lu; Edwin S Iversen; Rachel T Palmieri; Richard Houlston; Anne M van Altena; Katja K H Aben; Leon F A G Massuger; Angela Brooks-Wilson; Linda E Kelemen; Nhu D Le; Anna Jakubowska; Jan Lubinski; Krzysztof Medrek; Anne Stafford; Douglas F Easton; Jonathan Tyrer; Kelly L Bolton; Patricia Harrington; Diana Eccles; Ann Chen; Ashley N Molina; Barbara N Davila; Hector Arango; Ya-Yu Tsai; Zhihua Chen; Harvey A Risch; John McLaughlin; Steven A Narod; Argyrios Ziogas; Wendy Brewster; Aleksandra Gentry-Maharaj; Usha Menon; Anna H Wu; Daniel O Stram; Malcolm C Pike; Jonathan Beesley; Penelope M Webb; Xiaoqing Chen; Arif B Ekici; Falk C Thiel; Matthias W Beckmann; Hannah Yang; Nicolas Wentzensen; Jolanta Lissowska; Peter A Fasching; Evelyn Despierre; Frederic Amant; Ignace Vergote; Jennifer Doherty; Rebecca Hein; Shan Wang-Gohrke; Galina Lurie; Michael E Carney; Pamela J Thompson; Ingo Runnebaum; Peter Hillemanns; Matthias Dürst; Natalia Antonenkova; Natalia Bogdanova; Arto Leminen; Ralf Butzow; Tuomas Heikkinen; Kari Stefansson; Patrick Sulem; Sören Besenbacher; Thomas A Sellers; Simon A Gayther; Paul D P Pharoah
Journal:  Nat Genet       Date:  2010-09-19       Impact factor: 38.330

6.  Role of microRNA-7 and selenoprotein P in hepatocellular carcinoma.

Authors:  Marwa Tarek; Manal Louis Louka; Eman Khairy; Randa Ali-Labib; Doaa Zakaria Zaky; Iman F Montasser
Journal:  Tumour Biol       Date:  2017-05

7.  Increased MTHFD2 expression is associated with poor prognosis in breast cancer.

Authors:  Feng Liu; Yang Liu; Chuan He; Li Tao; Xiaoguang He; Hongtao Song; Guoqiang Zhang
Journal:  Tumour Biol       Date:  2014-05-29

8.  Gene expression correlates of clinical prostate cancer behavior.

Authors:  Dinesh Singh; Phillip G Febbo; Kenneth Ross; Donald G Jackson; Judith Manola; Christine Ladd; Pablo Tamayo; Andrew A Renshaw; Anthony V D'Amico; Jerome P Richie; Eric S Lander; Massimo Loda; Philip W Kantoff; Todd R Golub; William R Sellers
Journal:  Cancer Cell       Date:  2002-03       Impact factor: 31.743

9.  Neuron-specific enolase, histopathological types, and age as risk factors for bone metastases in lung cancer.

Authors:  Yang Zhou; Wen-Zhao Chen; Ai-Fen Peng; Wei-Lai Tong; Jia-Ming Liu; Zhi-Li Liu
Journal:  Tumour Biol       Date:  2017-07

10.  Ror2, a Developmentally Regulated Kinase, Is Associated With Tumor Growth, Apoptosis, Migration, and Invasion in Renal Cell Carcinoma.

Authors:  Chun-Ming Yang; Shan Ji; Yan Li; Li-Ye Fu; Tao Jiang; Fan-Dong Meng
Journal:  Oncol Res       Date:  2017-01-26       Impact factor: 5.574

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.