Literature DB >> 26057385

Comparison of High-Level Microarray Analysis Methods in the Context of Result Consistency.

Kornel Chrominski1, Magdalena Tkacz1.   

Abstract

MOTIVATION: When we were asked for help with high-level microarray data analysis (on Affymetrix HGU-133A microarray), we faced the problem of selecting an appropriate method. We wanted to select a method that would yield "the best result" (detected as many "really" differentially expressed genes (DEGs) as possible, without false positives and false negatives). However, life scientists could not help us--they use their "favorite" method without special argumentation. We also did not find any norm or recommendation. Therefore, we decided to examine it for our own purpose. We considered whether the results obtained using different methods of high-level microarray data analyses--Significant Analysis of Microarrays, Rank Products, Bland-Altman, Mann-Whitney test, T test and the Linear Models for Microarray Data--would be in agreement. Initially, we conducted a comparative analysis of the results on eight real data sets from microarray experiments (from the Array Express database). The results were surprising. On the same array set, the set of DEGs by different methods were significantly different. We also applied the methods to artificial data sets and determined some measures that allow the preparation of the overall scoring of tested methods for future recommendation.
RESULTS: We found a very low level concordance of results from tested methods on real array sets. The number of common DEGs (detected by all six methods on fixed array sets, checked on eight array sets) ranged from 6 to 433 (22,283 total array readings). Results on artificial data sets were better than those on the real data. However, they were not fully satisfying. We scored tested methods on accuracy, recall, precision, f-measure and Matthews correlation coefficient. Based on the overall scoring, the best methods were SAM and LIMMA. We also found TT to be acceptable. The worst scoring was MW. Based on our study, we recommend: 1. Carefully taking into account the need for study when choosing a method, 2. Making high-level analysis with more than one method and then only taking the genes that are common to all methods (which seems to be reasonable) and 3. Being very careful (while summarizing facts) about sets of differentially expressed genes: different methods discover different sets of DEGs.

Entities:  

Mesh:

Year:  2015        PMID: 26057385      PMCID: PMC4461299          DOI: 10.1371/journal.pone.0128845

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Microarrays are used to detect gene expression levels. Using this technology, we can simultaneously detect the expression levels of several thousand genes with one experiment [1]. Microarrays can also be used to determine how a disease or other external factors influence the level of gene expression in cells. To reach an appropriate conclusion, it is very important to analyze data (microarray readings) properly. Currently, many methods are used to detect differentially expressed genes (DEGs) from microarray data. However, there is no standardization and every scientist can select his or her preferred method. When we were asked for help with processing microarray data, we faced the problem of selecting an appropriate method. We were interested in finding a method that would yield "the best result". We found publications that provided comparisons of methods [2, 3, 4, 5]. However, such works did not answer all of our questions. All of the studies proved that methods are not consistent when taking the obtained results into account. At the same time, they did not provide recommendations, standard or procedure proposals or objective method (algorithm) assessments. We also noted that life scientists do not pay special attention to what method they use to analyze the results of microarray experiments (this is partly due to the use of commercial or ready-to-use software, where the information about which method adopted is described in the technical documentation) [6,7]. Based on this, we decided to determine how consistent the results are when examined by different methods of analysis of gene expressions [8, 9]. We decided to describe these results of our study with method evaluation. We decided to examine six commonly accepted and widely used methods for detecting DEGs [10,11]. The methods we tested were: Significance Analysis of Microarrays (SAM), Rank Products (RP), Bland-Altman (BA), Mann-Whitney test (MW), T Test (TT), and Linear Models for Microarray Data (LIMMA). Experiments were conducted using real data from eight microarray experiments (hereafter, Arraysets). We found that the first results were surprisingly divergent. Thus, we decided to test the methods on artificially prepared data sets (hereafter, Datasets) with known outstanding values (hereafter, aDEGs—artificial DEG) to be detected.

Microarray Experiment and Microarray Data Analysis

To obtain information about the types of microarrays and the principles of their operation, we referred to various sources [12, 13]. Fig 1 presents the steps of microarray experiments (reference to block number is given in brackets).
Fig 1

Microarray experiment steps (phases).

Aside from the usual steps that are common in most experiments—conception work, laboratory work (wet-lab) and closing work (blocks (1), (2), and (3) (Fig 1, S1 Fig) respectively)—in microarray experiments, three special steps (phases of data analysis) can be specified: Low-level data analysis (3a), where the intensity of fluorescence (raw data) is translated into numbers that reflect the fluorescence level for each probesetID for each microarray reading. High-level data analysis (3b, 3c), where we exclude probesets without expression changes and select the highest level of data analysis with probesets that undergo expression changes. Highest-level data analysis (4), where annotation, pathway analysis, interpretation, reporting, and visualization take place. In this study, we were only interested in high-level analysis methods, with a special interest in DEGs detection. Because tested method starts with a data table of microarray reading (the table of numbers and probeset identifiers as input data, without additional information), all factors concerning any biological or molecular mechanisms and tissue-specific questions were out of scope in this work. We started our examinations when we saw a normalized table with fluorescence levels for probesets and our goal was to determine which probesets represent genes with transcriptional activity chance (up—and down regulated—for which probeset identifier values changed). In addition, co-expression and pathway analysis was out of scope in our study; this can be done later, based on the results from the high-level analysis. As can be seen (Fig 1), all levels of analyses depended on the results of previous levels. Properly conducted low-level analysis is important for the results of high-level analysis. For highest-level analysis, results of both previous analyses (together with PCR validation of transcriptional activity of certain gene), as well as all biological, molecular mechanisms, and tissue specific issues, are crucial for the final results of experiment. In short, high-level analysis can be classified as: (3 b) simple methods, which are mainly based on statistics, and (3 c) complex methods, which are based on artificial intelligence and discrete mathematics [14, 15, 16].

2.1 Short characteristics of the methods

In the comparisons presented in this paper, we only focused on simple methods. Below, we provide a short characterization of each of the examined methods. Significant Analysis of Microarrays (SAM) [17, 18] SAM is a statistical method used to determine statistical significance in gene expressions between groups. In terms of mode of action, SAM reassembles a T test. However, SAM uses non-parametric statistics, due to the fact that microarray data are not normally distributed. Rank Product (RP) [19, 20, 14] RP is a statistical method for detecting gene expression changes. It belongs to non-parametric statistical tests and is based on ranks of fold changes. Bland-Altman (BA) [1, 21] BA analysis is a statistical method that allows the comparison of two groups of results. In addition to using BA on data from microarray experiments, it is also very popular in medical data analysis of medical data. Mann-Whitney (MW) [22] MW is a non-parametric test used to test the conformity between two populations. The null hypothesis is that the two populations are identical. It is one of the most popular tests used to check the conformity between groups. One of its usages is to detect gene expression changes in microarray data. Test T (TT) [23] TT is a statistical test that determines whether two sets differ from one another in a statistically significant way. This test is based on the average and variance of the population. It is one of the simplest and most frequently used statistical tests. Linear Models for Microarray Data (LIMMA) [24] LIMMA is available as a BioConductor package for analyzing gene expression in microarray data. It uses linear models to analyze microarray data. We also examined the frequency of using certain methods in scientific papers by searching Google Scholar and PubMed (Table 1).
Table 1

Frequency of hits: method name along with “differentially expressed genes” and “gene expression” phrases (Google Scholar, PubMed).

Method name"Differentially expressed genes""Gene expression"
SAM Significant Analysis of Microarrays 1 290; 303 516; 2 287
SAM 14 400; 252 190 000, 1 746
RP Rank Products 691; 15 948; 122
MW Mann-Whitney 7 210; 32 131 000; 511
BA Bland-Altman 106; 2 1 930; 14
TT test t 423; 1 6 420; 1 375
t-test 32 000; 285 1 390 000; 1 375
LIMMA Linear Models for Microarray Data 25; 74 38; 367
LIMMA 7 010; 77 11 600; 106
We searched the name of the method, along with (AND operator) two phrases. In the first search, we used “differentially expressed genes” and in the second, we used "gene expression".

Arraysets—Experiments

We compared the results of a high-level analysis for eight microarray (Affymetrix HG-U133A) experiments from ArrayExpress [25, 26]. All readings from this type of microarray contained 22,283 microarray probes (probesets) [12, 27]. The array sets contained microarray readings from a different number of samples. Table 2 presents brief information about the chosen array sets, along with information about the accession number, the number of samples in the array set, and the short characteristics of samples. With the exception of the first array set, all the others included readings from two types of samples: control (from healthy tissue) and tissue from affected tissue. We chose these experiments in order to verify the effectiveness of methods, both in the dependence of the number of samples in the microarray data and the different microarray experiments.
Table 2

Arraysets Characteristics.

ArraysetsAccession numberNumber of samplesCharacteristics
ArraySet1 E-GEOD-3207250all samples from cancerous tissue (lung)
ArraySet2 E-GEOD-14882168—control, 8—patients with MELAS syndrome
ArraySet3 E-GEOD-158528643—control, 43—lung cancer tissue
ArraySet4 E-MEXP-1690126—control, 6—ganglioglioma
ArraySet5 E-GEOD-56899455—control, 40 brain tissue affected by Alzheimer's
ArraySet6 E-GEOD-2252910482—chemoimmunotherapy patients, 22 from cancer tissue
ArraySet7 E-TABM-79410250—control, 52—prostate tumours
ArraySet8 E-GEOD-110387225—control, 47 tissue with leukemia
The first necessary step was to conduct a low-level data analysis. In order to perform a low-level analysis, we used standardized Robust MultiChip Average (RMA) [11, 23] method with a Bioconductor [28], which was done for all of the Arraysets. In this study, this type of low-level analysis is not a variable. We present (Fig 2) the process of preparation of Arraysets from the single array experiment data.
Fig 2

Arraysets preparation process.

For the purpose of conformity verification, we carried out a high-level analysis of array sets. The parameters for each method are presented in Table 3. To perform an analysis of tested methods, we used R packages, or R language built-in functions.
Table 3

Parameters that were fixed for each method of high-level analysis (for the purpose of experiments).

MethodsType of parameterValue of parameter
SAM fold change2.00
RP p-value0.01
MW p-value0.05
BA p-value0.02
TT p-value0.01
LIMMA p-value0.05
These parameters remained intact throughout all of the testing procedures (both for Arraysets and Datasets). The number of DEGs detected by different methods is presented in Table 4.
Table 4

DEGs detected in Arraysets by different methods (22,283 in all).

 SAMRPMWBATTLIMMA common DEGs
ArraySet1 3323114613752178222001340 11
ArraySet2 104314469521132153952 6
ArraySet3 460545512260174310922260 433
ArraySet4 1872184618489143201848 91
ArraySet5 1159020148401476448840 16
ArraySet6 6593380992977493992 95
Arrayset7 279831004797246825812684 254
ArraySet8 1716188517895051041633 122
From Table 4, it can be seen that we detected a small quantity of common DEGs—from 6 to 433—and the number of DEGs detected by various methods in each Arrayset was different. First, we suspected that the different quantity of detected DEGs was a result of using methods with distinct sensitivity levels. It is for this reason we also examined artificial datasets (see part 4). For each Arrayset, we decided to present the results as Venn diagrams. Because the visualization of common parts of more than five sets in the form of Venn diagrams is not trivial, we used the approach presented by [29]. We present the Venn diagram for Arrayset 1 (Fig 3) (Venn diagrams of the remaining Arraysets are presented in Supporting Information files (S1–S7 Figs)). The numbers represent the quantity of detected DEGs. Thus, the diagram also shows the result concordance of methods. The grayscale represents the number of methods, for which the common part has been computed.
Fig 3

Venn diagram for Arrayset1.

On the basis of the presented Venn diagrams, a very low concordance of results between tested methods can be noticed. A large part of DEGs identified by one of the methods was not considered as DEG by other methods. For each Arrayset, the number of DEGs detected by each method can be read in Table 4, and the number of common DEGs detected by methods can be read from the Venn diagrams (Fig 3, S1–S7 Figs). From the analysis, it is clear the diversified concordance between DEGs detected by certain methods on Arraysets. Such a low agreement level surprised us. As such, we decided to test the methods on artificially generated data to determine whether the methods themselves (or implementation of algorithms) work properly. We prepared special artificial datasets, with a priori known outstanding in value elements—aDEG (to mimic DEGs). We prepared two datasets with 2,000 elements, generated as normal distribution with mean = 10 and σ = 1.3. Next, in the case of Dataset1, 73 elements were changed with values from mean = 15, σ = 1.3 (as up regulated aDEGs) to mean = 5, σ = 1.3 (as down regulated aDEGs). The range of certain values of exemplary artificial array readings are shown on the boxplot chart in Fig 4 with usual ranges: min value, 1st quartile, median value, 3rd quartile, and max value. The values which "expression" does not change are marked as “other”, up regulated aDEGs are marked as “up regulated”, and down regulated aDEGs are marked as “down regulated”. It can be seen that all ranges are in this case completely separate.
Fig 4

Distribution of values in Dataset1.

Dataset2 had slightly outstanding values—ranging from mean = 13.5, σ = 1.3 to mean = 6.5, σ = 1.3 respectively (we also wanted to check the sensitivity of each method). Range of certain values for Dataset2 (“other” together with up and down regulated aDEGs) is shown on Fig 5 in a similar way as on Fig 4. In this case, one can see that down regulated are completely separate from “other”, only small amount of up regulated has common range with some maximum values of “other”.
Fig 5

Distribution of values in Dataset2.

Datasets are uploaded as Supplemental Materials; for more detailed information, refer to adeg_info.txt and ReadMe.txt file in Supplemental Materials. The results of aDEGs detection in both Datasets are presented in the next section.

Experiments on Datasets

In the case of the Arraysets, it was difficult to assess which method gave the best results, because we did not know which (and how many) values should be detected as DEGs. Thus, we needed a procedure that would allow scoring the methods used for DEGs detection. The problem of detecting DEGs can be regarded as a problem of classifying (grouping) data into three groups: DEGs without change, DEGs up regulated and DEGs down regulated.

4.1 Methods Quality Assessment [30, 31]

In the case of the artificially prepared Datasets, we knew what number of aDEGs to detect. Thus, we could make an algorithm assessment. We determined all true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN). Therefore, we assessed the quality of methods using measures that were used in computer science during the classification algorithms evaluation. These measures were: accuracy (acc), recall (rec), precision (prec), f-measure (fm), and Matthews correlation coefficient (MCC). MCC interpretation is similar to ROC/AUC, but it is presented in the form of one number. Furthermore, it is regarded to be more stable when class (groups) sizes may be different. The first used measure was accuracy. Accuracy is defined as: Accuracy describes the degree of conformity between values that should be detected and values that algorithms detect. An accuracy value close to 1 means the greater accuracy of the algorithm (it is better). An accuracy value equal to 1 means that the tested algorithm only found the values that should be detected. An accuracy of 0 means that the algorithm has not found any of the values that it should. The second measure used was recall, also known as true positive rate or sensitivity. Recall is defined as: Recall describes how many values are correctly detected by the tested algorithms, in proportion to all of the values that should be detected. Algorithms give the best results when the recall is equal to 1 and the worst results when the recall is 0. The next measure used was precision. Precision is described as: Precision describes how many values are correctly detected by the tested algorithms. As in recall, the best results are given when the precision is equal to 1 and the worst results are given when the precision is 0. The next metrics was the f-measure, which is described as: F-measure is a kind of compromise between recall and precision. When we only use recall or precision, we are not able to decide which algorithm gives better results, detects less "unwanted" values or misses a small amount of that which should be detected. F-measure is usually used for choosing an algorithm with optimal ratio of precision and recall (fm = 1 the best, 0 the worst). The last measure used in the algorithm evaluation was Matthews Correlation Coefficient, which. Matthews correlation coefficient is described as: Matthews Correlation Coefficient considers values in the range -1 to 1 (-1 the worst, 1 the best). This measure gives information similar to correlation, which inform us to what extent the values that are detected by the tested algorithm are similar to those we expected. Dataset1 and Dataset2 were examined with the same methods, i.e., SAM, RP, BA, MW, TT, and LIMMA (and the same parameters (see Table 3)).

4.2 Dataset1—results

For Dataset1, almost all of the methods detected at least 73 modified values. Furthermore, some recognized additional values. Only the BA method and LIMMA detected less than we expected (72 of 73). The distribution of values in Dataset1 is presented on the boxplot (Fig 4). A summary of values detected by all tested methods, as well as the measures for method evaluation, is presented in Table 5.
Table 5

Number of aDEGs detected and assessment parameters used for each method in Dataset1 (in bold—the best, in italics—the worst).

aDEGs detected (of all 73)
SAMRPBAMWTTLIMMA
Number of detected values7484761388172
True positives737372737372
True negatives192619161923186219191927
False positives11146580
False negatives001001
acc 0.995 0.9450.975 0.675 0.960 0.995
rec 1 1 0.986 1 1 0.986
prec0.9860.8690.947 0.528 0.901 1
f-measure 0.993 0.9290.966 0.691 0.948 0.993
MCC 0.989 0.8900.947 0.508 0.918 0.989
SAM and LIMMA had the best values in the case of four (for all five) measures. SAM had the highest values for acc, rec, fm, and MCC. LIMMA had the highest values for acc, prec, fm, and MCC. Moreover, SAM and LIMMA reached the same values for acc, fm, and MCC. However, they differed in rec and prec. SAM had the maximum value of rec parameter (equal to 1), which means that the algorithm had no false negatives. LIMMA had prec that equaled 1, meaning that it had no false positives. The worst results were obtained in the case of the MW method. Here, in four (of all five) measures, it had the worst values. Its measures had low values, so even maximum values for rec (equal to 1) would not have justified its usage. BA and TT reached relatively high overall scoring for all measures and the RP results were at an acceptable level. Table 6 presents a summary of aDEGs detected by various methods. This shows which of the additionally detected values were also detected (or not) by other methods. The first column contains the name of the method used and the second column is the number of additionally detected values, which are also detected by other methods. The other columns show whether additional aDEGs were detected (“Yes”, if detected by another method and “No” if not detected by another method). LIMMA detected less aDEGs than expected, with one false negative, and so it was omitted from this table. For example, in the first row, SAM detected one additional aDEG and this was also detected by MW; in the second row, RP detected one additional aDEG, which was also detected by MW and TT, but not SAM and BA.
Table 6

Summary of excessed aDEGs by each method.

ExcessedRecognized as aDEG by other method
aDEGSAMRPBAMWTT
SAM 1-NoNoYesNo
RP 1No-NoYesYes
4No-NoYesNo
1No-YesYesNo
4No-NoNoNo
BA 1NoYes-YesNo
1NoNo-YesYes
1NoNo-YesNo
4NoNo-NoNo
MW 1YesNoNo-No
1NoYesNo-Yes
1NoYesYes-No
5NoYesNo-No
1NoNoYes-Yes
1NoNoYes-No
7NoNoNo-Yes
49NoNoNo-No
TT 1NoYesNoYes-
1NoNoYesYes-
6NoNoNoYes-
It appears that the Dataset1 tested methods worked correctly and were able to detect almost all of the values that should be recognized as aDEGs. Based on the experiment of Dataset1, we can conclude that the best algorithms for use are SAMM and LIMMA (ex-equo), followed by BA and TT, and, eventually, as a last option, RP and BA. MW had the worst scoring and, therefore, we do not recommend it.

4.3 Dataset2—results

A similar evaluation procedure was carried out on Dataset2. Also, like Dataset1, the distribution of values in Dataset2 is presented on boxplot. Table 7 contains a summary of values detected by all of the tested methods, as well as the measures for method evaluation.
Table 7

Number of aDEGs detected and assessment parameters by each method in Dataset2 (in bold—the best, in italics—the worst).

aDEGs detected (of 73 all)
SAMRPBAMWTTLIMMA
Number of detected values6998501498576
True positives697346737173
True negatives192719021923185119131924
False positives025476143
False negatives4027020
acc0.9800.8750.845 0.620 0.920 0.985
rec0.945 1 0.630 1 0.972 1
prec 1 0.7440.920 0.489 0.8350.960
f-measure0.9710.8530.747 0.657 0.898 0.979
MCC0.9570.7730.665 0.443 0.839 0.968
The results for Dataset2 differ from those for Dataset1—they were not so sharp. Similarly to the Dataset1 result, the worst overall result had MW. RP and BA reached better but rather intermediate values in overall range. Additionally, BA had the lowest rec measure for all of the methods. On the other hand, RP had the maximum possible rec measure value. In the case of Dataset2, LIMMA reached the best scores for all of the methods. SAM only had slightly worse scoring. Again, they can be regarded as comparable. The third best-scored method was TT. It had worse results than LIMMA and SAM, but better than BA and RP. Table 8 presents the summary of information about aDEGs (which should be similarly read to Table 6). Due to the fact that SAM detected fewer values than expected, it was omitted from this table.
Table 8

Summary of excessed aDEGs by each method.

Excessed aDEGRecognized as aDEGs by other method
SAMRPBAMWTTLIMMA
RP 1NO-YESYESYESYES
5NO-NOYESYESNO
6NO-NOYESNONO
13NO-NONONONO
BA 1NOYES-YESYESYES
1NONO-YESYESNO
2NONO-NONONO
MW 5NOYESNO-YESNO
1NOYESYES-YESYES
6NOYESNO-NONO
1NONOYES-YESNO
7NONONO-YESNO
56NONONO-NONO
TT 5NOYESNOYES-NO
1NOYESYESYES-YES
1NONOYESYES-NO
7NONONOYES-NO
LIMMA 1NOYESYESYESYES-
2NONONONONO-
It is clear that almost all of the methods detected more aDEGs than was expected—the methods gave worse results on a less diversified Dataset. Based on the results obtained for Dataset2, the best results had LIMMA, followed by SAM and then TT. RP and BA were also acceptable. Similarly to the experiment with Dataset1, MW gave the worst results. A final summary of the scoring methods is shown in Table 9.
Table 9

Overall scoring of methods for the Datasets (one plus equals one point; the more, the better).

SAMRPBAMWTTLIMMA
Dataset1+ + +++ +-+ ++ + +
Dataset2+ + +++-+ ++ + +
overall scoring 6 23 0 4 6
Based on Table 9, we recommend methods SAM and LIMMA as the best choice and TT and BA as acceptable for high-level analysis.

Conclusions

The low level of concordant results for the Arraysets was surprising. When conducted in the Datasets, our analysis showed that, in most cases, the methods themselves (as well as their implementation) work properly (except the MW method, which gave unsatisfactory results). All of the evaluation measures used for scoring methods were better when the outstanding values were well “separated” (more diversified, as in Dataset1). Therefore, one possible conclusion is that microarray experiments data were slightly diversified (similar to Dataset2). In our opinion, such results show the need to recommend how studies based on microarray experiments should be carried out: The list of DEGs should be obligatory, published with precise specification concerning the high-level analysis (and software used). When selecting an algorithm for high-level analysis, it is important to take into account the measures for each method and choose a variant method based on special needs (best acc, best rec, best MCC, etc.). In general, when the list of DEGs is only obtained with the use of one high-level analysis, it should not be regarded as reliable and definitive. One could argue that an official recommendation about high-level analysis should also be carried out. A possible approach is to use a few methods and acknowledge DEGs as only those genes that are within an intersection of sets of DEGs obtained by different methods. Based on the overall method scoring presented (Table 9), we recommend at least LIMMA, SAM, and TT. The presented results should also be taken into account by authors of reviews (or those who search for DEGs under certain condition in different papers), while compiling results from different publications that describe a set of detected DEGs. It is very important to pay special attention to the methods of high-level analysis. This is because a resultant set of DEGs can vary, depending on the method used.

Dataset used in this study.

(ZIP) Click here for additional data file.

Venn diagram for Arrayset2.

(TIF) Click here for additional data file.

Venn diagram for Arrayset3.

(TIF) Click here for additional data file.

Venn diagram for Arrayset4.

(TIF) Click here for additional data file.

Venn diagram for Arrayset5.

(TIF) Click here for additional data file.

Venn diagram for Arrayset6.

(TIFF) Click here for additional data file.

Venn diagram for Arrayset7.

(TIF) Click here for additional data file.

Venn diagram for Arrayset8.

(TIF) Click here for additional data file.
  14 in total

1.  affy--analysis of Affymetrix GeneChip data at the probe level.

Authors:  Laurent Gautier; Leslie Cope; Benjamin M Bolstad; Rafael A Irizarry
Journal:  Bioinformatics       Date:  2004-02-12       Impact factor: 6.937

2.  Microarray data classified by artificial neural networks.

Authors:  Roland Linder; Tereza Richards; Mathias Wagner
Journal:  Methods Mol Biol       Date:  2007

3.  Statistical methods for assessing agreement between two methods of clinical measurement.

Authors:  J M Bland; D G Altman
Journal:  Lancet       Date:  1986-02-08       Impact factor: 79.321

4.  Should we abandon the t-test in the analysis of gene expression microarray data: a comparison of variance modeling strategies.

Authors:  Marine Jeanmougin; Aurelien de Reynies; Laetitia Marisa; Caroline Paccard; Gregory Nuel; Mickael Guedj
Journal:  PLoS One       Date:  2010-09-03       Impact factor: 3.240

5.  Bioconductor: open software development for computational biology and bioinformatics.

Authors:  Robert C Gentleman; Vincent J Carey; Douglas M Bates; Ben Bolstad; Marcel Dettling; Sandrine Dudoit; Byron Ellis; Laurent Gautier; Yongchao Ge; Jeff Gentry; Kurt Hornik; Torsten Hothorn; Wolfgang Huber; Stefano Iacus; Rafael Irizarry; Friedrich Leisch; Cheng Li; Martin Maechler; Anthony J Rossini; Gunther Sawitzki; Colin Smith; Gordon Smyth; Luke Tierney; Jean Y H Yang; Jianhua Zhang
Journal:  Genome Biol       Date:  2004-09-15       Impact factor: 13.583

6.  Rank products: a simple, yet powerful, new method to detect differentially regulated genes in replicated microarray experiments.

Authors:  Rainer Breitling; Patrick Armengaud; Anna Amtmann; Pawel Herzyk
Journal:  FEBS Lett       Date:  2004-08-27       Impact factor: 4.124

7.  Comparison and evaluation of methods for generating differentially expressed gene lists from microarray data.

Authors:  Ian B Jeffery; Desmond G Higgins; Aedín C Culhane
Journal:  BMC Bioinformatics       Date:  2006-07-26       Impact factor: 3.169

Review 8.  Statistical tests for differential expression in cDNA microarray experiments.

Authors:  Xiangqin Cui; Gary A Churchill
Journal:  Genome Biol       Date:  2003-03-17       Impact factor: 13.583

9.  Neural network analysis of lymphoma microarray data: prognosis and diagnosis near-perfect.

Authors:  Michael C O'Neill; Li Song
Journal:  BMC Bioinformatics       Date:  2003-04-10       Impact factor: 3.169

10.  Methods for evaluating gene expression from Affymetrix microarray datasets.

Authors:  Ning Jiang; Lindsey J Leach; Xiaohua Hu; Elena Potokina; Tianye Jia; Arnis Druka; Robbie Waugh; Michael J Kearsey; Zewei W Luo
Journal:  BMC Bioinformatics       Date:  2008-06-17       Impact factor: 3.169

View more
  11 in total

1.  Comment on "Nuclear localization of LDL receptor-related protein 1B in mammary gland carcinogenesis".

Authors:  Chengyu Yang; Long Zhao; Xuemei Liu; Ruixia Ma; Yan Xu
Journal:  J Mol Med (Berl)       Date:  2019-03-15       Impact factor: 4.599

2.  Bioinformatics Analysis of Differential Innate Immune Signaling in Macrophages by Wild-Type Vaccinia Mature Virus and a Mutant Virus with a Deletion of the A26 Protein.

Authors:  Chenyu Li; Haiyan Zhou; Hui Zhang; Shujuan Wang; Yan Xu
Journal:  J Virol       Date:  2017-11-30       Impact factor: 5.103

3.  Merging microarray studies to identify a common gene expression signature to several structural heart diseases.

Authors:  Olga Fajarda; Sara Duarte-Pereira; Raquel M Silva; José Luís Oliveira
Journal:  BioData Min       Date:  2020-07-08       Impact factor: 2.522

4.  Sexual Dimorphism and Aging in the Human Hyppocampus: Identification, Validation, and Impact of Differentially Expressed Genes by Factorial Microarray and Network Analysis.

Authors:  Daniel V Guebel; Néstor V Torres
Journal:  Front Aging Neurosci       Date:  2016-10-05       Impact factor: 5.750

5.  Letter to the Editor: Bioinformatics Analysis in Downstream Genes of the mTOR Pathway to Predict Recurrence and Progression of Bladder Cancer.

Authors:  Chenyu Li; Hang Liu; Yan Xu
Journal:  J Korean Med Sci       Date:  2018-01-22       Impact factor: 2.153

6.  Bioinformatics analysis of glial inflammatory responses to air pollution.

Authors:  Chenyu Li; Wei Jiang; Nina Tang; Yan Xu
Journal:  J Neuroinflammation       Date:  2017-08-15       Impact factor: 8.322

7.  The Author's Response: Bioinformatics Analysis in Downstream Genes of the mTOR Pathway to Predict Recurrence and Progression of Bladder Cancer.

Authors:  Chang Hyuk Yoo
Journal:  J Korean Med Sci       Date:  2018-01-22       Impact factor: 2.153

8.  Identification of a gene expression profile associated with the regulation of angiogenesis in endometrial cancer.

Authors:  Marcin Opławski; Mateusz Michalski; Andrzej Witek; Bogdan Michalski; Nikola Zmarzły; Agnieszka Jęda-Golonka; Maria Styblińska; Joanna Gola; Małgorzata Kasprzyk-Żyszczyńska; Urszula Mazurek; Andrzej Plewka
Journal:  Mol Med Rep       Date:  2017-06-28       Impact factor: 2.952

9.  Influence of Glucose Availability and CRP Acetylation on the Genome-Wide Transcriptional Response of Escherichia coli: Assessment by an Optimized Factorial Microarray Analysis.

Authors:  Daniel V Guebel; Néstor V Torres
Journal:  Front Microbiol       Date:  2018-05-23       Impact factor: 5.640

10.  Fold change based approach for identification of significant network markers in breast, lung and prostate cancer.

Authors:  Richa K Makhijani; Shital A Raut; Hemant J Purohit
Journal:  IET Syst Biol       Date:  2018-10       Impact factor: 1.615

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.