Literature DB >> 16970549

Normalization using weighted negative second order exponential error functions (NeONORM) provides robustness against asymmetries in comparative transcriptome profiles and avoids false calls.

Sebastian Noth1, Guillaume Brysbaert, Arndt Benecke.   

Abstract

Studies on high-throughput global gene expression using microarray technology have generated ever larger amounts of systematic transcriptome data. A major challenge in exploiting these heterogeneous datasets is how to normalize the expression profiles by inter-assay methods. Different non-linear and linear normalization methods have been developed, which essentially rely on the hypothesis that the true or perceived logarithmic fold-change distributions between two different assays are symmetric in nature. However, asymmetric gene expression changes are frequently observed, leading to suboptimal normalization results and in consequence potentially to thousands of false calls. Therefore, we have specifically investigated asymmetric comparative transcriptome profiles and developed the normalization using weighted negative second order exponential error functions (NeONORM) for robust and global inter-assay normalization. NeONORM efficiently damps true gene regulatory events in order to minimize their misleading impact on the normalization process. We evaluated NeONORM's applicability using artificial and true experimental datasets, both of which demonstrated that NeONORM could be systematically applied to inter-assay and inter-condition comparisons.

Entities:  

Mesh:

Year:  2006        PMID: 16970549      PMCID: PMC5054038          DOI: 10.1016/S1672-0229(06)60021-1

Source DB:  PubMed          Journal:  Genomics Proteomics Bioinformatics        ISSN: 1672-0229            Impact factor:   7.691


Introduction

Studying the cellular transcriptome and its dynamics using microarray technology has become a common application in modern biomedical research (, and has spurred over the past two decades many novel insights into the mechanisms of gene regulation and physiopathology 2., 3., 4., 5.. Microarray technology has seen considerable advances in the number of transcripts that can be detected simultaneously as well as the precision of the measurement. Several all-in-one commercial solutions and a multitude of array scanner systems in the public domain have emerged. However, these technologic solutions often diverge in the design of the probes and the method of transcript detection, making the statistical analysis of transcriptome data challenging and non-uniform 6., 7.. Microarray technology cannot be used for absolute but only relative quantification of the abundance of individual transcripts (. Obviously, this fact greatly complicates the successful analysis of microarray experiments. As for inter-assay comparisons, which hence need to be carried out for relative quantification between two different biologic samples, efficient and accurate normalization methods have to be developed 9., 10., 11.. In the first step, intra-assay normalization methods account for variations in print-tip quality, irregular sample distribution over the array surface, irregularities in the surface itself, camera aperture related distortions, non-linear detection-dye dynamic ranges, and (depending on the technology used) many other phenomena that contribute to technical variations for individual probes 10., 12., 13.. Then, inter-assay normalization techniques in the optimal case capture technical variations (which are due to sample preparation, extraction, amount, or labeling), quality variations, dye related differences (2 dye setup) and variations, chemical batch variations, array batch variations, and so on. If intra-array normalization methods have been successfully applied before, inter-assay normalization techniques usually require a global rescaling of the entire datasets relative to each other and thus are of homogenous nature 11., 14.. Many of the non-linear and basically all linear normalization methods thereby make two assumptions about the nature of the data: (1) invariance in the absolute quantity of a majority of transcripts, and (2) symmetry of the probe variation distribution 9., 10., 11., 12., 13., 14.. Note that normalization methods are based solely on control probes for which synthetic transcripts are added at different moments during the experimentation, assuming analogously invariance and symmetry of the control signals. Since solely internal control based normalization fails to capture sample related technical variation and is not very robust, it is rarely used by itself. The first invariance assumption for the majority of probes, in the limit of large probe sets, seems to hold and can be biologically justified. Furthermore, this assumption is necessarily required for the principle of inter-assay normalization based on probe signals to be meaningful. The second assumption about symmetry of the probe variation distribution, which is exploited in non-discriminant averaging normalizations, cannot be justified from a biological point of view. This assumption only holds true for technical replicates generated from a single biologic sample. Biological questions frequently pose concern about the dynamics in the transcriptome profile when comparing two different states of cell differentiation (, or the differences and/or similarities between a physiologic and a pathologic state of a cell (. When comparing distinct physiologic situations, the symmetry assumption seems to be poorly reflecting biological reality. A multitude of biologic processes can be cited that will consequently lead to asymmetric changes in the expression profile of a cell. For instance, inherently asymmetric processes such as apoptosis or mitotic repression will lead to a majority of genes being down-regulated and only a comparatively small fraction being induced 17., 18.. Furthermore, due to their very different mechanistic nature, transcription activation and repression mechanisms follow distinct dynamics. Consequently, such asymmetries are frequently observed in “real world” data 19., 20.. Since these asymmetries are not corrected in technical and/or biological replicates, they reflect true biological variations. Most inter-assay normalization methods do not account for such asymmetries and use non-discriminant averaging (mean, median) for normalization of these data. It is obvious that averaging methods, provided the asymmetry is of sufficient significance to exceed numerical precision, will lead to suboptimal normalization factor estimation. Such considerations, especially in the case of the so-called boutique arrays where the invariance assumption cannot be made, have lead to the development of discriminant normalization methods 21., 22.. Here, only either pre-defined or conditionally determined subsets of probes are used for the normalization process. For the former, either the so-called house-keeping genes that are thought to be invariant in their expression across a large number of cellular conditions, or external and synthetic probe/target pairs are added to the experimental pipeline at different points in time, and then are considered for normalization. However, such methods have significant disadvantages. The notion of a house-keeping gene is at best empiric and often circumstantial (. Dynamic determination (no a priori assumptions made) of the probes to be considered for normalization could be a solution to this problem. Unfortunately, few methods exist for such a type of inter-assay normalization 22., 24.. Those methods, however, are either expression rank-based 22., 25., 26., 27., or depend on statistical tests that are performed after inter-assay comparison/subtraction profiles have been calculated (. Especially in view of the increasing sensitivity of some microarray solutions, analysis methods that do not or only partially consider information on individual signal variance, will lead to suboptimal statistical interpretation of the transcriptome data (. While analyzing high-density kinetic transcriptome data for myeloid cell differentiation, we have realized that the asymmetries in the probe variance distributions did significantly compromise the median-based inter-assay normalization. We employed the novel AB1700 platform (Product Info: http://www.appliedbiosystems.com) for transcriptome analysis. As detailed in Materials and Methods, several non-linear and linear intra-assay normalization techniques were systematically and automatically applied during the primary analysis of raw image data. The resulting probe signal estimates were considered sufficiently normalized to directly proceed with inter-assay comparisons (http://www.appliedbiosystems.com). However, when carefully analyzing the phenomenon of asymmetric probe variance distributions, we realized that additional means of inter-assay normalization were required. Furthermore, after carefully reviewing the literature as well as the data generated by a different technology, we realized that this problem seems to be not limited to a particular technology or biologic model 19., 20.. Unsatisfied with existing solutions to the normalization problem, we developed a novel method, namely NeONORM, for inter-assay normalization that is insensitive to asymmetries in probe variation distributions. This study summarizes the development of NeONORM and its evaluation using synthetic test and “real world” data.

Results and Discussion

Derivation of the NeONORM error function

We specifically thought to develop an inter-assay normalization method for transcriptome or similar data that overcomes the problems associated with asymmetric heavy tails in fold-change distributions (Figure 1). This method would supposedly be applied once (generally non-linear) intra-assay normalizations (such as print-tip correction) have already successfully been applied. Our central assumption in the reasoning leading up to the NeONORM method is that except for technical replicates, the common hypothesis of symmetry of fold-change distributions is not well founded for inter-assay comparisons. In order to avoid implementing several different methods specifically tailored to and distinguishing explicitly between technical replicates, biological replicates, and comparisons of truly different conditions, ideally, this novel method would also intrinsically adapt according to the nature of the inter-assay comparison. Therefore, the method would behave in the limit of technical replicates and highly symmetric fold-change distributions, which is closely similar to the existing, averaging, and linear methods such as Median normalization.
Fig. 1

Asymmetric heavy tails in fold-change distributions. A histogram view of nine different “real-world” superimposed fold-change distributions is shown as calculated from recent experimental work in our laboratory. The arrow indicates asymmetric heavy tails that are observed to different degrees in such experimental data.

In addition to a quadratic error function (quadratic in the signal difference x), we construct a damping function such that it would restrict the maximum contribution of a large logQ (defined as the binary logarithm of the signal quotient) on the local quadratic error (small variation Δx of x, see Figure 2A). In the limit of Δx approaching zero, the influence of the damping function should disappear. For illustration, we have sketched the product of two such functions in Figure 2A. A simple multiplication of both the quadratic error function with the damping function obviously would result in lower error contributions once |x| exceeds the equilibrium point (EP) between the contributions of both functions, and can hence not be used. The NeONORM error function, however, approximates the multiplicative functions in the bounds of |x| EP, the error becomes essentially invariant (Figure 2A). Additionally, the individual contributed errors would be scaled as a function of the quality of measurement (cumulated variances over both individual signals). The formal derivation of this modified NeONORM error function can be found in the supporting online material (“NeONORMformalism.pdf”). In short:
Fig. 2

The NeONORM error function. A. Schematic representation of the standard quadratic error function (solid black), as well as a damping function (dashed black), the product of both (dashed red), and the novel NeONORM error function (dashed blue) that we indirectly derived from the former. B. NeONORM damping function (dashed black) for the first derivative of the NeONORM error function, and the first derivate of the standard quadratic (solid black) and NeONORM (dashed blue) error functions.

Damping the derivative of the quadratic error function with a Gaussian:yields the overall error function after integration and introduction of weights: Testing err(a) on real data produced very good results. Nevertheless, a parameter k to control the sensitivity was introduced to yield the final NeONORM error function: For any x, an individual minimum is obtained if k is sufficiently small, and the greater k is, the more err approaches the quadratic error function:and

Properties of the NeONORM error function

The NeONORM error function has very interesting properties with respect to the free parameter k. As can be seen in Equation (3), the entire error to be minimized is in fact a sum of individual contributing error functions err(a), each symmetrically centered at x and scaled by w. Sums of these individual error functions have interesting properties that shall be discussed here, referring in particular to Figure 3.
Fig. 3

Properties of the NeONORM error function. The NeONORM (composite) error function has exactly one minimum for τ ≤ 2 in the limit of identical weights for both individual error functions. Thereby τ denotes the absolute distance between the global minima of the individual error functions in units of k. Here k is the NeONORM sensitivity parameter identical to the absolute distance between the inclination points and the global minimum of each function. In the upper panel, we schematize the derivatives of the individual NeONORM contributing error functions (black: probe 1, blue: probe 2), and their composite (red) for different values of τ at constant k (shown are the first order derivates in the normalization factor a). The lower panel displays the original functions. Only when τ increases above 2, the NeONORM error function acquires two distinct minima.

The free parameter k, which we also refer to as the “sensitivity parameter”, has crucial influence on the shape of the total error function. For two values x and x, differences in their summation depending on their distance are depicted in Figure 3, where τ = |x − x|/k. The point p = (x + x)/2 is crucial for distinguishing three different cases of summation by inspecting the sign of the second derivative of err at p: In the second case, the saddle point is a “wide” minimum and indicates a bifurcation point. In the third case, two distinct minima exist, as can be easily seen in Figure 3. Note that the above holds true only for equal weights w and w. The parameter k thus determines a border of discrimination between clusters of points: no two points that have a distance less than 2k will create distinct minima. Unfortunately, at present, we are not able to derive a formal definition of k. The parameter therefore has to be empirically derived.

Sign plots of the first order derivatives of the NeONORM error function as visual indicators

In order to further illustrate the properties of the NeONORM error function at different values for k, we plotted the first derivative over the range of −2 < a < 2 and 10−3 < k < 101. In Figure 4A, we show a sign plot (blue=negative, red=positive) and a 3D-surface plot of the function over the same parameter space. We compared two technical replicates generated from total RNA of HT29 cells [HT29(1) vs. HT29(2), see Materials and Methods]. As becomes evident from these plots, and as expected, the NeONORM error function at large k possesses a single minimum (in the sign plot: blue/red border) corresponding to the minimum of the quadratic error function (mean). With decreasing k, the sole minimum bifurcates into several local minima. These bifurcations process onset at multiple a over a relatively short range of k. They hence reflect the growing (with decreasing k) influence of smaller and smaller clusters of co-varying probes generating local minima in the error landscape. At the limit of k approaching zero (the granularity or numerical resolution of the given data), individual probes will generate local minima in the error function, whereas in the limit of k approaching infinity, the data are “appraised” by the function in the same way as in the quadratic error function.
Fig. 4

The NeONORM sensitivity parameter k. A. Sign plot and 3D-surface plot for the first order derivative in a of the NeONORM error as a function of the sensitivity parameter k and the normalization factor a. Sign plots are used to illustrate the zero crossings of the function (blue-red and red-blue boundaries). The data represented are a subtraction from two HT29 technical replicates HT29(1) vs. HT29(2). B. As in panel A for a subtraction between the modified, artificial dataset HT29(1)mod vs. HT29(2). In order to generate the modified dataset, a random chosen quarter of all probe signals of the HT29(1) dataset was individually multiplied with a different random value drawn from a normal distribution with μ=21.5 and σ=0.1*μ. C. As in panel A, but only a sign plot for the second order derivative in a is shown. D. The NeONORM error function for selected increasing values for k is shown for the HT29(1) vs. HT29(2) (upper) and the HT29(1)mod vs. HT29(2) datasets. Note that k=0.02 is the flattest curve in both cases.

We then plotted the first derivative of the NeONORM error function for an inter-assay comparison where we used artificially distorted data (in order to generate asymmetry in the logQ distribution, see Materials and Methods). Briefly, the HT29(1)mod dataset was generated from HT29(1) by randomly choosing a quarter of all probe signals and multiplying each of them with a factor randomly drawn from a normal distribution with a mean of 21.5 and a variance of 0.1*21.5. As becomes evident from both plots in Figure 4B, the perturbation of the underlying data generates a massive distortion in the HT29(1)mod vs. HT29(2) comparison. The multiplication of a subset of probe signals with an almost constant factor essentially subdivides the probe set into two distinct clusters. This effect is noticeable at relative large k when compared to the bifurcations observed within very symmetric data (Figure 4A). The NeONORM error function hence captures strong clustering in the data. A second phenomenon can be observed in the sign plots of the artificial dataset. With decreasing k, the relative position of the global minimum of the error function shifts in a. Only at sufficiently small k, thus at sufficiently high sensitivity, the NeONORM error function returns the optimal normalization factor a. This obviously has major implications for the choice of k. At the same time, it illustrates the advantage of the NeONORM method in that averaging error functions (comparable to NeONORM in the limit of k approaching infinity) return suboptimal normalization factor a. For further illustration purposes, we also plotted the second derivative of the NeONORM error function over the same parameter space with the unmodified HT29 data. Here, the commencement of the multiple bifurcation zone already appears at larger k (Figure 4C). In Figure 4D, we show the NeONORM error function at selected k (increasing from top to bottom) for both datasets, the unmodified (top) and modified (bottom) HT29 inter-assay comparison. In the case of the asymmetric data, the appearance of a second minimum at small k is clearly visible.

Choosing NeONORM parameter k

Since at present we are unable to derive a formal description of k, we had to empirically derive a suitable value for k. Thereby an equilibrium between two considerations had to be found. Whereas for k tending to zero the sensitivity of the NeONORM error function is maximal, the global minimum in the error function is the result of fewer and fewer probes, making the estimation of the normalization factor a less and less robust against variations and numerical inadequacy of the data. Using the artificial datasets, as well as some one hundred different experiments performed in our own laboratory, we have found that k=0.20 provides a stable estimate for a, and should be considered the lower bound for k. The derivation of a from a at smaller k thereby is systematically inferior to the numerical precision imposed by the data. At the same time, k=0.20 should also be the upper bound for k since the maximal attainable sensitivity/precision of the NeONORM method is reached. Note that k=0.20 is found close to the commencement of the bifurcation zone in almost all the data we have analyzed. The algorithmic implementation of NeONORM is capable of correctly identifying the global minimum of the error function for arbitrarily small k, thus a is correctly estimated even within the bifurcation zone. Since we cannot rule out the possibility that k needs to be adjusted for different microarray platforms (for example, due to different precision of the numerical values), we have successfully applied the NeONORM method with k=0.20 to nine inter-assay comparisons of Affymetrix datasets (http://www.affymetrix.com). The evaluation of the NeONORM method presented below has been done at k=0.20.

“Asymmetric” test data

In order to evaluate the NeONORM method, several well defined test datasets had to be acquired. First, we were looking for “real world” inter-assay comparisons where the fold-change distributions were as close to symmetric as possible. To this end, we decided to use two technical replicates generated in our laboratory from the total RNA extracted from HT29 cells. These technical replicates, here denominated as HT29(1) and HT29(2), share a Pearson correlation of R=0.993, thus are indeed highly reproducible for the single biologic condition (Supporting online material: “HT29.txt”). In order to generate asymmetries in a controlled fashion in the fold-change distributions, the HT29(1) dataset was modified according to two distinct procedures to generate: (1) A dataset with ever increasing perturbations at ever larger distance to a=0, and with a constant and small variance over the perturbation (Figure 5; see Materials and Methods). This dataset was used to simultaneously test the performance and the robustness of the NeONORM method. Having up to a quarter of all probe signals artificially induced with a logQ average of up to 21.5, and a constant / ratio of 0.1, this dataset resulted in highly significant perturbations easily detectable in the sign plots shown in Figure 5 (Supporting online material: “HT29(1)robustness.txt”). (2) A second dataset where exponentially increasing perturbations with a constant and large variance are introduced (Table 1; see Materials and Methods). For small ratios of modified probe signals f (n=~32,500 for our AB1700 data), the perturbations were very modest (Supporting online material: “HT29(1)sensitivity.txt”). Hence, this dataset can be used to evaluate the sensitivity of the NeONORM method at k=0.20.
Fig. 5

Artificial “asymmetric” datasets. Sign plots of the NeONORM error function for the first series of artificial test data generated are shown (see Materials and Methods). A. From left to right: (1) Unmodified (original) dataset HT29(1) vs. HT29(2). (2) Asymmetrically modified dataset HT29(1) (random chosen quarter, μ=21.5 and σ = μ). (3) Double modified dataset HT29(1) was once more modified by subsequently choosing another random quarter of probe signals and multiplying them with a random value drawn from a normal distribution with μ=2−1.5 and σ=0.1*μ. This second operation generates almost symmetrically modified data, where the average total of modified probe signals is 7/16. B. As in panel A. To the right of each row the fraction of modified probe signals (1/16, 1/8, 1/4), and on the top of each column the average ratio change parameters (log2 of μ and the corresponding σ) are indicated.

Table 1

Pearson Correlations for Artificial Test Datasets

Pearson correlations
vs. HT29(1)vs. HT29(2)
HT29(1) unmodified1.0000.993
mod 1/256 1.5±1.50.9840.977
mod 1/128 1.5±1.50.9610.954
mod 1/64 1.5±1.50.9310.927
mod 1/32 1.5±1.50.8370.832
mod 1/16 1.5±1.50.7940.790
mod 1/8 1.5±1.50.7930.786
mod 1/4 1.5±1.50.7180.713
In conclusion, the artificial datasets generated here to evaluate the absolute and relative performance of the NeONORM method are of high quality as the nature of the perturbations covers and exceeds the entire range of feasible naturally occurring asymmetries, and the perturbations, due to the process of how they were generated, have similar statistical properties to real experimental data. Finally, we chose two datasets from retinoic acid (RA) treated and untreated NB4 cells obtained recently in our laboratory as an example of “real” biologic data (Supporting online material: “NB4.txt”). The effect of RA on changes in the gene expression profiles of myeloid cells and human acute promyelocytic leukemia [APL; the NB4 cell line was derived from an APL patient (] will be discussed in detail elsewhere (data not shown). Briefly, RA directly acts upon the transcriptional activity of a group of nuclear receptor transcription factors, the so-called retinoid receptors (, thereby switching their activity from a repressor state to a very potent activator state of gene expression (. The effects of RA on gene expression are very immediate and drastic (becoming apparent after as short as 15–30 min). Within the observed time scale of a couple of hours, RA thereby commits NB4 cells to reentry into differentiation, which is otherwise blocked by the presence of the PML-RARalpha oncogene (. Given the nature of the activity switch, as well as the rapidness of the physiologic response, we rightly assumed that these datasets would display some degree of asymmetry in the fold-change distributions. Note that all of these datasets had previously undergone identical multiple non-linear and linear normalization steps to account for systematic intra-assay variations before the artificial perturbations were introduced and before these data were used for evaluating NeONORM (see Materials and Methods).

Algorithmic implementation of the NeONORM method

Based on the above reasoning, we have developed an algorithm capable of generating and minimizing the negative second order exponential error functions used by the NeONORM method. For illustration of the algorithmic implementation, a pseudo-code description of the NeONORM algorithm is displayed in Figure 6.
Fig. 6

The pseudo-code for the implementation of the NeONORM algorithm. Additional details for the implementation can be made available upon written inquiry to the authors.

The algorithm consists of the following three steps: (1) Finding intervals I1 … I that each contains a minimum; (2) Iteratively approaching for each j = 1 … n the minimum contained in the interval I until a defined precision is reached, yielding a candidate value a for the minimum in I; (3) Calculating the error values err = err(a) for each candidate and selecting a that corresponds to the minimal error amongst err1 … err. The algorithm is implemented in Java in the framework of the ace.map suite for microarray statistical data analysis, which will be described in detail elsewhere. A value for k needs to be defined as well as a search interval s and a maximum tolerated derivative conv. A step size d is calculated as a function of k. Step 1: The interval s is sampled using [s/d] steps. In every step i, the sign of the derivative of err at x, namely s, is calculated. If s is positive and s is negative, a zero crossing that belongs to a minimum occurs between x and x. This interval is added to a vector of candidates. This first sampling step is only required for small k when one cannot assume that there is only a single minimum. Other methods to find multiple minima could be used here, but this method has the advantage to use a defined number of steps and is shown to be sufficient for the described purpose. Step 2: For every such candidate c, the minimum m is further approached iteratively, halving the size of the interval [xlow, xhigh] in every iteration step. In every step, the function calculates the error derivative d for the point p on half distance between xlow and xhigh. If d is negative, xlow is replaced by x, otherwise xhigh. The function stops once |d| < conv. Step 3: For every minimum m, the actual error is calculated. The errors for all minima are compared and the mininum that belongs to the smallest error is returned. Functions for calculating the cumulated error and its derivative sum up the results of corresponding functions that calculate individual error and error derivative for each gene, respectively. Assume there are n genes in a dataset. For every single error or derivative, n exponentials have hence to be calculated (plus n times some basic arithmetical operations that are not regarded here). The step size d is proportional to k. So the number of exponentials to be calculated for Step 1 is proportional to n · s/k. The iterative method in Step 2 will find exactly one minimum in the given interval if at least one exists. If there exist several minima, the method will find one of them ignoring the existence of others. The method is fast, stable, and rarely requires more than 30 iterations to converge under conv=10−4 for all considered cases so far. The algorithm was tested for k less than the numerical precision of data provided (10−4), distinguishing more than 500 minima in the interval [−0.1 : 0.1].

Comparative validation of the NeONORM method

We proceeded to evaluate the performance, robustness, and sensitivity of the NeONORM method on the different “real world” and artificial datasets by comparative validation of three methods, that is, inter-assay normalization by the Median, LOWESS (LOcally WEighted Scatterplot Smoothing), and NeONORM methods. Figure 7 summarizes the key comparison results of the Median and NeONORM methods for the artificial datasets from Figure 5. In all the three panels of Figure 7A, frequency plots of the logQ distributions between two input sample datasets are depicted. On the unmodified HT29(1) vs. HT29(2) inter-assay comparison (Figure 7A1), both methods perform highly similarly as expected. As we have shown above formally for the NeONORM error function in the limit of large k (≥ 10), the error estimate becomes virtually identical to the one obtained by Median normalization. We have furthermore contended that the NeONORM error function at k=0.20 still closely resembles the quadratic error function in case of symmetric fold-change distributions (Figure 5A). We find this assumption confirmed using the technical HT29 replicates since the normalization factors a are virtually identical for both methods (a=0.0000, a=−0.0045). Note that the small differences of both curves in the frequency plot are rather due to rounding differences in the binning procedure. When both NeONORM and Median normalization methods are applied to an assay comparison with an asymmetric heavy tail [Figure 7A2: HT29(1) mod vs. HT29(2)], a clear difference becomes apparent. Note that NeONORM correctly identifies the error minimum and normalizes the asymmetric data such that the maximum of the frequency distribution falls upon logQ=0, whereas Median normalization shifts the frequency distribution significantly to negative logQ values (a=−0.5472, a=0.0000). In case of the compensatory doubly modified dataset [HT29(1)mod and mod vs. HT29(2); Supporting online material: “HT29(1)double.txt”], the difference between both methods is less pronounced; however, NeONORM clearly much better normalizes this dataset as well (Figure 7A3: a=−0.1672, a=−0.0038). The fact that Median normalization does not lead to identical results here is due to the successively random choosing of probe subsets for modification, which, as can be easily shown, highly unlikely leads to perfectly symmetric data. Together with the initial unmodified data shown in Figure 7A1, this experiment demonstrates that NeONORM is insensitive to the presence of asymmetric heavy tails in the fold-change distributions, whereas averaging normalization methods for obvious reasons fail to correctly normalize the datasets. In Figure 7B, we look at Median (Figure 7B1) vs. NeONORM (Figure 7B2) normalization performance on the remaining nine artificial datasets shown in Figure 5B. As before, NeONORM correctly normalizes the different datasets such that the maxima of the increasingly perturbed datasets all superimpose at logQ=0. Median normalization, in contrast, leads to maxima being gradually (as a function of the perturbation of the artificial dataset) shifted towards negative logQ values (Figure 7B1). Figure 7B3 summarizes these findings by directly comparing both methods on the most severe perturbed artificial dataset [HT29(1)mod vs. HT29(2); a=−0.4774, a=−0.0060]. In conclusion, the NeONORM method performs very well even on highly asymmetric data, and is robust with respect to the grade of the asymmetry [the range covered is from near zero (technical replicates shown in Figure 7A1) to a point that certainly exceeds any natural occurring gene regulatory event (Figure 7B3)].
Fig. 7

Comparison of Median vs. NeONORM normalization. A. Direct comparison of Median vs. NeONORM on the datasets from Figure 5A. Frequency plots of logQ are shown simultaneously for both methods (Median in grey, NeONORM in red). B. Direct comparison of Median vs. NeONORM on the datasets from Figure 5B. Frequency plots of logQ are shown in the upper two panels simultaneously for all nine datasets (B1: Median normalized, B2: NeONORM normalized). B3: Frequency plots of logQ are shown simultaneously for both methods (Median in grey, NeONORM in red) on the HT29(1)mod vs. HT29(2) dataset.

In order to perform a sensitivity test of the NeONORM method, we used the second artificial test dataset. In this case, we additionally compared NeONORM performance relative to an implementation of the LOWESS algorithm 32., 33., 34.. LOWESS is a very well performing non-linear normalization method widely used for intra- and inter-assay normalization (Ref. 12., 32.; see Materials and Methods). The results of the direct comparisons are summarized in Figure 8A. The three panels show histogram plots for the logQ distributions of the second artificial dataset (Table 1; see Materials and Methods) as functions of the normalization methods. As was true for the previous dataset, Median normalization results in successive “migration” of the distribution maximum towards negative logQ values as a function of the severity of the perturbation (Figure 8A1). The LOWESS method seems to perform much better than Median normalization, as much as the maximum of the distribution is more stable around the logQ=0 point. However, when the perturbations of the original dataset become more and more severe (here especially evident for f=1/8 and f=1/4), the maximum of the distribution also shifts towards negative logQ values, indicating that the LOWESS method cannot correct for the asymmetries in the logQ distribution. Furthermore, due to the nature of the non-linear approach, the relative stability of LOWESS at low f values is counterbalanced by a significant distortion of the data (compare characteristic peaks at the maximum or the heavy tails between all distributions and with respect to the normalization methods). By contrast, NeONORM again performs equally well on all eight conditions, with the maxima of the distributions perfectly superimposed at logQ=0. Note that due to the linear nature of the normalization operation, no distortion of the distributions occurs. In order to demonstrate the significance of the differences in normalization factor a over this experimental series, we have generated a probe call table using arbitrary thresholds for logQ (as the nature and absolute value of the thresholds do not matter), and compared the differences in logQ calls (Table 2). It is apparent from this table that NeONORM, due to more accurate normalization, is the only method that shows constant logQ< −1 calls [see column with relative values, note that variation (negative values) is due to the normal distribution properties of the random drawing of factors during the perturbation of the data]. Furthermore, the logQ> 1 calls are proportional to the severity of the perturbation. Median normalized data, as expected by the nature of the operation, almost “grow” symmetrically over both thresholds, whereas LOWESS normalization generates satisfactory results for slightly biased test data and just starts to diverge significantly from NeONORM at f > 1/32. The distortion of the dataset also becomes apparent when comparing the total numbers of probe calls beyond the two thresholds.
Fig. 8

Comparison of Median, LOWESS, and NeONORM normalization. A. Direct comparison of Median, LOWESS, and NeONORM normalization on the artificial datasets summarized in Table 1. Histogram plots of logQ are shown in the three panels simultaneously for all the six modified datasets and the original dataset. B. Direct comparison of Median, LOWESS, and NeONORM normalization on the “real world” NB4 RA 4h vs. NB4 dataset. B1: the first order derivative in a of the NeONORM error function at k=0.20, for the range −2 < a < 2. B2: Normalization of the NB4 RA 4h vs. NB4 according to all the three methods (Histogram plot is shown, blue = Median, grey = LOWESS, red = NeONORM).

Table 2

Comparative Performance of Median, LOWESS, and NeONORM Methods on Artificial Microarray Data*

Profile [vs. HT29(2)]Median
logQ< −1
|logQ| ≤ 1
logQ> 1
absrelabsrelabsrel
HT29(1) unmodified2,139028,30002,1930
mod 1/256 1.5±1.52,138−128,213−872,28188
mod 1/128 1.5±1.52,1713228,144−1562,317124
mod 1/64 1.5±1.52,2187927,990−3102,424231
mod 1/32 1.5±1.52,42428527,673−6272,535342
mod 1/16 1.5±1.52,64050127,013−1,2872,979786
mod 1/8 1.5±1.53,4931,35425,477−2,8233,6621,469
mod 1/4 1.5±1.55,3053,16622,387−5,9134,9402,747

Profile [vs. HT29(2)]
LOWESS (f=0.3, iteration=2, δ=10E-5),
logQ< −1
|logQ| ≤ 1
logQ> 1
absrelabsrelabsrel

HT29(1) unmodified2,043028,33402,2550
mod 1/256 1.5±1.52,043028,246−882,34388
mod 1/128 1.5±1.52,046328,176−1582,410155
mod 1/64 1.5±1.52,0601728,008−3262,564309
mod 1/32 1.5±1.52,0783527,692−6422,862607
mod 1/16 1.5±1.52,0975427,092−1,2423,4431,188
mod 1/8 1.5±1.52,20816525,821−2,5134,6032,348
mod 1/4 1.5±1.52,69064723,493−4,8416,4494,194

Profile [vs. HT29(2)]
NeONORM (k=0.20, conv.=10E-4)
logQ< −1
|logQ| ≤ 1
logQ> l
absrelabsrelabsrel

HT29(1) unmodified2,144028,30102,1870
mod 1/256 1.5±1.52,143−128,214−872,27588
mod 1/128 1.5±1.52,144028,144−1572,344157
mod 1/64 1.5±1.52,143−127,973−3282,516329
mod 1/32 1.5±1.52,140−427,678−6232,814627
mod 1/16 1.5±1.52,129−1527,065−1,2363,4381,251
mod 1/8 1.5±1.52,142−225,751−2,5504,7392,552
mod 1/4 1.5±1.52,144023,267−5,0347,2215,034

Probe calls according to all three normalization methods are summarized for the second set of artificially generated data. Corresponding Pearson correlations are shown in Table 1. See Materials and Methods for a description of the modification procedure. abs = absolute number of probes, rel = relative (to the first row in each column) number of probes. Parameters for LOWESS and NeONORM methods are also indicated in the header row.

In order to further quantify the relative performance of the different normalization techniques and demonstrate the robustness of NeONORM when dealing with asymmetric microarray data, we calculated the Type I and Type II errors resulting from NeONORM and Median normalization methods (the latter was found to be the least appropriate method in the above assays, see Table 2) on a selection of the synthetic datasets introduced above. Since those synthetic data were generated from the HT29(1) dataset (see Materials and Methods), we can calculate the number of probe calls that should occur in a comparison with HT29(1) using the correct normalization factor a=0. Any deviation from this calculated number is to be considered an error introduced through the normalization method that was applied. According to the nature of the false call (false positive or false negative), we can determine the type of error [incorrect rejection of a true null-hypothesis (I) or failure to reject a false zero hypothesis (II) of non-change]. Table 3 lists the results where the two types of errors were calculated as percent false calls when subtracting the four increasingly asymmetric synthetic datasets from the original HT29(1) dataset. We thereby investigated the 99% confidence interval for our zero-hypothesis and set a cut-off for determining the number of probe calls at |logQ| > 1.00. Note that due to the use of random number generators when creating the synthetic files, the calculated expected probe calls are associated with a variance. Thus, only errors superior to 0.038% are significant (bold-face). In agreement with the observations made before (Table 2), NeONORM consistently outperforms Median normalization whether Type I or Type II error is considered. As a matter of fact, NeONORM is resistant against Type I errors even when Median normalization results in greater one percent of false positive calls, which corresponds here to 386 out of a total of 32,821 probes. When very significant asymmetries are introduced (up to a quarter of all probe signals being modified), NeONORM starts to generate some Type II errors; however, the values are still well below the Type II errors observed with Median normalization. NeONORM thus indeed significantly reduces the number of false calls due to inadequate or inefficient inter-assay normalization. Most importantly, NeONORM avoids Type I errors.
Table 3

Type I and Type II Error Analysis for Median and NeONORM Normalization Using the Synthetic Datasets*

HT29(1) vs.: (|logQ| >1.00, p <0.01)Absolute Error (%)
NeONORM
Median
Type IType IIType IType II
HT29(1) unmodified0.0000.0000.0000.000
mod 1/256 1.5±1.50.0300.0000.0120.000
mod 1/64 1.5±1.50.0060.0000.0210.015
mod 1/16 1.5±1.50.0000.0760.0820.189
mod 1/256 1.5±1.50.0000.2071.1761.222

Error is expressed as percent false probe calls at the given threshold and confidence interval for the zero-hypothesis of non-change in the subtraction profiles generated when comparing the modified datasets to the original HT29(1) dataset. Significant values are in bold type setting.

Finally, we also applied all the three normalization methods to a second “real world” dataset (NB4 RA 4h vs. NB4, see Materials and Methods, and also the paragraph on “Asymmetric test data”). As discussed above, given the nature of the physiologic response to RA, we had reasons to believe that such a dataset might be similar to some of the datasets presented in Figure 1 that show asymmetric heavy tails in the logQ distributions. Figure 8B summarizes the results we obtained by using the three normalization methods in direct comparison. In Figure 8B1, we show the first order derivative of the NeONORM error function at k=0.20 over the range from −2 < a < 2. The histogram plots of logQ distributions in Figure 8B2 recapitulate the results that were earlier obtained with the constructed test data. Median normalization leads to a very significantly shifted maximum (blue curve); LOWESS shows intermediate performance with respect to positioning of the maximum, but distorts the distribution (grey curve) when compared to NeONORM. Since these are non-controlled data (in the sense that we have not generated the asymmetry ourselves), we can only judge the relative performance. On the other hand, since the data are individually (intra-assay) median pre-normalized, the fact that the NeONORM normalization factor is close to zero [a=−0.4106, a=0.0000, note that a (given the non-linear nature) cannot be directly determined, and can only be estimated by renormalizing these data with NeONORM and determining the required normalization factor aa, which repositions the LOWESS maximum back to superimpose exactly with the NeONORM value aa=−0.0399] indicates that NeONORM here again performs close to optimum. As for the artificial data in Figure 8A, we also performed the probe call test (Table 4) to demonstrate the significant impact of choice of inter-assay normalization on potential results of comparative studies. Interestingly, due to the compression of the data during the LOWESS normalization, this method performs most poorly on the dataset (highest number of potentially false “down-regulated” calls, and lowest number of “up-regulated” calls) when compared to the other two methods. These results clearly can only be an indication of the relative performance of the three methods, as at present we can neither establish this dataset as representative for transcriptome studies, nor estimate the number of “down-” and “up-regulated” calls we should expect. However, given the detailed analysis we have performed above using artificially (in a well-controlled manner) modified data, NeONORM is the only method that shows robustness towards asymmetric logQ distributions. Given the fact that our two distinct strategies for generating these test datasets largely cover what could be expected to occur in real biologic samples as maximal asymmetry, and show that NeONORM scales very well with the degree of asymmetry and performs perfectly well in the limit of maximal symmetric data, we can draw the conclusion that NeONORM overcomes the problems associated with the standard symmetry hypothesis in linear inter-assay normalization.
Table 4

Probe Calls of All the Three Normalization Methods on the “Real World” Dataset

MethodNB4 RA 4h vs. NB4
logQ< −1|logQ| < 1logQ> 1
Median4,47919,4284,883
LOWESS4,05220,9563,792
NeONORM2,37518,7217,694

Applicability of NeONORM

Since NeONORM is basically invariant to the level of asymmetry in the data, perfectly symmetric data are also processed correctly. Hence, NeONORM can be applied to any comparative inter-assay study without any a priori knowledge about, and regardless of the level of symmetry. Given their similitude in nature, NeONORM should also find its application in other functional genomics applications such as comparative genome hybridizations or ChIP-on-chip experiments, which seem even more prone to asymmetries given an even reduced dynamic range of the signal response. The same prerequisites as for microarray experiments will probably have to be met there, as non-linear intra-assay normalization methods should for similar reasons be applied first to the individual experiments.

Future challenges

Since the estimate of the over-all error is used during the iterative minimization procedure, and also the minimum of the error function is found at different values for a (as a non-trivial function of k), the choice of k should not be arbitrary. It seems obvious from the sign plots presented in Fig. 4, Fig. 5 that k for practical reasons should be sufficiently large to assure that the error function is not estimated within the multiple bifurcation region where individual or small group of co-varying probes attain significant influence. On the other hand, k should be chosen as small as possible in order to ascertain maximum sensitivity of the NeONORM method with respect to asymmetric logQ distributions. While other methods such as LOWESS normalization sometimes also possess even several “free” parameters 32., 33., and much effort has been devoted to identify empirical optimal parameter ranges (, we feel that in the case of NeONORM, a formal way of definition might be possible as parameter k directly reflects an inherent property of the analyzed data. To derive a formal way of selecting parameter k will thus represent a major challenge towards the perfection of the NeONORM method. We specifically invite suggestions or ideas towards such a formal derivation of k. At the same time, we have taken great care in the empiric derivation of a suitable k as demonstrated during the evaluation of the overall performance of the NeONORM method. The range of possible asymmetries covered through our artificial test datasets should assure that NeONORM performs in a robust fashion with respect to true physiologic datasets. Finally, the integration of NeONORM with other methods of normalization deserves further interest. We clearly state that the NeONORM normalization can only be one of the several methods applied to current microarray data. However, the potential interplay of different non-linear intra-assay and linear inter-assay normalization methods does not seem to have been studied systematically yet. Potentially, a combination of particularly robust and situation/data adapted methods can be found and optimized to simplify, at a maximum of coherence, the microarray statistical analysis process.

Conclusion

The NeONORM method overcomes a current limitation of inter-assay normalization methods in that it is robust against asymmetries in the underlying fold change distributions. Such asymmetries reflect true changes in gene expression patterns rather than systematic experimental variation. NeONORM, a non-discriminant method, if combined with initial non-linear intra-assay normalization methods, could lead to better inter-assay normalization and thus better identification and estimation of gene regulatory phenomena in comparative transcriptome studies.

Materials and Methods

Cell culture and RNA extraction

Two different cell lines were used in the experiments. Human colon carcinoma HT29 cells (ATCC No: HTB-38) were cultivated under standard conditions in Dulbecco’s Modified Eagle Medium (DMEM; Gibco, ProdNo: 41965-039) supplemented with 4.5 g/L glucose, 10% foetal bovine serum (FBS; PAA Laboratories GmbH, ProdNo: A15-649), antibiotics (a mix of 10 U/mL penicillin G and 10 mg/L streptomycine; Gibco, ProdNo: 15140-122), and 1.5 mM L-glutamine (Gibco, ProdNo: 25030-024). Cells were harvested during the exponential growth phase and subjected to RNA extraction. We have previously described cultivation and all-trans RA treatment of NB4 cells (. Here, where appropriate, 10−6 M of all-trans RA (Sigma) was directly applied to the culture for 4 h prior to harvesting of the cells. RNA extraction was performed using the Qiagen RNeasy method according to the manufacture’s recommendations (ProdNo: 75144). Quality and quantity of the isolated total RNA was determined using an Agilent 2100 Bioanalyzer as well as standard spectrophotometry.

Microarray technology

We used the novel Applied Biosystems AB1700 (ProdNo: 4338036) oligonucleotide-based microarray technology (http://www.appliedbiosystems.com). For the present study, only Human Genome Survey Microarrays (V1.0, ProdNo: 4337467) were used, which contain probes for 29,918 validated human genes. Two assays with HT29 total RNA were performed as technical replicates [HT29(1) and HT29(2)]. One assay with NB4 cells that had been treated for 4h with RA (NB4 RA 4h) and the other from untreated NB4 cells (NB4) have further been specifically used.

RNA labeling, hybridization, and detection

RNA labeling, hybridization, and detection were performed following the protocols supplied by Applied Biosystems together with the corresponding kits. 20 μg of total RNA sample was subjected to Chemiluminescence (CL) RT Labeling (Applied Biosystems, ProdNo: 4339628). Labeled cDNAs were then hybridized and detected (Applied Biosystems, ProdNo: 4346875).

Data preprocessing and primary analysis

Applied Biosystems Expression Array System Software (V1.1.1, ProdNo: 4364137) has been used to acquire the CL and fluorescence (FL) images and primary data analysis. Briefly, the primary analysis consists of the following individual operations: (1) Image correction; (2) Global and local background correction; (3) Feature normalization; (4) Spatial normalization; (5) Global normalization. Note that we renormalize the resulting data according to the median once more after having removed probes for which the Applied Biosystems Software has set flags greater than 212, indicating compromised or failed measurements (as recommended by Applied Biosystems). This secondary normalization is implemented in the ace.map suite.

Generation of artificially modified test datasets

Two artificially modified test datasets were generated by modifying the HT29(1) dataset. The modifications were carried out in order to generate asymmetries in a controlled fashion in the fold-change distributions. The general procedure for the modification can be summarized as follows: A selected dataset X of size n is modified according to three parameters: f—the ratio of signals to be modified. Exactly [f · n] signals are modified; μ—the mean of the normal distribution from which the factor is drawn; σ—the variance of the normal distribution from which the factor is drawn. The routine is implemented in Java using the supplied pseudo random number generator and the transformation function [java.util.Random.next Gaussian()], which implements the Polar Method by Box et al. [http://java.sun.com/j2se/l.4.2/docs/api/java/util/Random.html#nextGaussian()]. The pseudo-code for the function can be given as: The two generated artificial datasets differ by the ratio f, the mean μ, and the variance σ. The first dataset shown in Figure 5B was generated at constant σ=0.1*μ for all combinations of f=1/16, 1/8, 1/4 and μ=0.5, 1.0, 1.5. The second dataset discussed in Table 1, Table 2 was generated with constant μ=21.5, constant σ=μ, and varying f=1/256 to 1/4. Simply, the first dataset creates ever larger perturbations (with increasing f) at ever larger distance to a=0 (with increasing μ) with a constant and small variance of the perturbation. The second dataset creates exponentially increasing perturbations (with increasing f) at a constant distance to a=0 with a constant and large variance.

Inter-array normalization

(1) Median normalization, linear. Assuming symmetry in the fold change; after calculating the fold changes, the median is subtracted from the log2-transformed signal quotients. (2) LOWESS normalization, non-linear. LOWESS is a method developed by Cleveland ( in 1979 and since then has been frequently improved and modified. It was applied to microarray data analysis for the first time by Yang et al. ( in 2001. For the original method, four parameters have to be specified (which normally happens more or less arbitrarily), and changes in each of them lead to different results. A suggestion for optimized parameter selection was published by Berger et al. ( in 2004. LOWESS performs very well for poorly preprocessed data. The LOWESS implementation used to compare to NeONORM is a Java port of Cleveland’s original FORTRAN code from 1985 freely available (http://netlib.bell-labs.com/netlib/go/lowess.f.gz), which was temporarily embedded in the ace.map platform for direct comparative testing. For the Median and NeONORM normalization methods, profiles are biased by a single additive value identical for all probes. LOWESS normalization is performed on Bland-Altman-/MA-plots and hence additionally uses the average of the log2-transformed signal values (M) corresponding to one gene and, roughly speaking, generates a bias function, possibly different for every m∈M. LOWESS parameters used for all normalizations were: f=0.3; polynome order=1; iterations=2; delta=0.00001.

Data representation

Histograms and frequency plots appearing in the figures were generated using ace.map, however, are standard means of data representation. The 3D-surface plots in Figure 4 were rendered using gnuplot 4.0. Sign plots (blue = negative, red = positive) were rendered by an ace.map plug-in. Subtraction profiles consist of log2-transformed quotients (logQ) of signal intensities for the intersection set of the probe IDs, which were contained in the two input sample files used for the subtraction.

Authors’ contributions

SN has significantly participated in the mathematical formulation of NeONORM, the generation of artificial data, the statistical data analysis, and manuscript preparation, as well as algorithmically implemented all of the three normalization methods. GB has contributed to the algorithmic implementation of NeONORM and manuscript preparation. AB has significantly participated in the mathematical formulation of NeONORM, the generation of artificial data, the statistical data analysis, and manuscript preparation. AB has designed and coordinated this study, and has acquired the experimental data. All authors read and approved the final manuscript.

Competing interests

The authors have declared that no competing interests exist.
  31 in total

1.  Identification of hTAF(II)80 delta links apoptotic signaling pathways to transcription factor TFIID function.

Authors:  B Bell; E Scheer; L Tora
Journal:  Mol Cell       Date:  2001-09       Impact factor: 17.970

2.  Analysis of variance for gene expression microarray data.

Authors:  M K Kerr; M Martin; G A Churchill
Journal:  J Comput Biol       Date:  2000       Impact factor: 1.479

3.  A comparison of normalization methods for high density oligonucleotide array data based on variance and bias.

Authors:  B M Bolstad; R A Irizarry; M Astrand; T P Speed
Journal:  Bioinformatics       Date:  2003-01-22       Impact factor: 6.937

4.  New normalization methods for cDNA microarray data.

Authors:  D L Wilson; M J Buckley; C A Helliwell; I W Wilson
Journal:  Bioinformatics       Date:  2003-07-22       Impact factor: 6.937

5.  Outcome signature genes in breast cancer: is there a unique set?

Authors:  Liat Ein-Dor; Itai Kela; Gad Getz; David Givol; Eytan Domany
Journal:  Bioinformatics       Date:  2004-08-12       Impact factor: 6.937

Review 6.  Mapping of genetic and epigenetic regulatory networks using microarrays.

Authors:  Bas van Steensel
Journal:  Nat Genet       Date:  2005-06       Impact factor: 38.330

7.  NB4, a maturation inducible cell line with t(15;17) marker isolated from a human acute promyelocytic leukemia (M3).

Authors:  M Lanotte; V Martin-Thouvenin; S Najman; P Balerini; F Valensi; R Berger
Journal:  Blood       Date:  1991-03-01       Impact factor: 22.113

8.  Optimized LOWESS normalization parameter selection for DNA microarray data.

Authors:  John A Berger; Sampsa Hautaniemi; Anna-Kaarina Järvinen; Henrik Edgren; Sanjit K Mitra; Jaakko Astola
Journal:  BMC Bioinformatics       Date:  2004-12-09       Impact factor: 3.169

9.  Early asymmetry of gene transcription in embryonic human left and right cerebral cortex.

Authors:  Tao Sun; Christina Patoine; Amir Abu-Khalil; Jane Visvader; Eleanor Sum; Timothy J Cherry; Stuart H Orkin; Daniel H Geschwind; Christopher A Walsh
Journal:  Science       Date:  2005-05-12       Impact factor: 47.728

10.  Rank Difference Analysis of Microarrays (RDAM), a novel approach to statistical analysis of microarray expression profiling data.

Authors:  Dietmar E Martin; Philippe Demougin; Michael N Hall; Michel Bellis
Journal:  BMC Bioinformatics       Date:  2004-10-11       Impact factor: 3.169

View more
  21 in total

1.  TLR5 signaling stimulates the innate production of IL-17 and IL-22 by CD3(neg)CD127+ immune cells in spleen and mucosa.

Authors:  Laurye Van Maele; Christophe Carnoy; Delphine Cayet; Pascal Songhet; Laure Dumoutier; Isabel Ferrero; Laure Janot; François Erard; Julie Bertout; Hélène Leger; Florent Sebbane; Arndt Benecke; Jean-Christophe Renauld; Wolf-Dietrich Hardt; Bernhard Ryffel; Jean-Claude Sirard
Journal:  J Immunol       Date:  2010-06-21       Impact factor: 5.422

2.  Early transcriptional programming links progression to hepatitis C virus-induced severe liver disease in transplant patients.

Authors:  Angela L Rasmussen; Nicolas Tchitchek; Nathan J Susnow; Alexei L Krasnoselsky; Deborah L Diamond; Matthew M Yeh; Sean C Proll; Marcus J Korth; Kathie-Anne Walters; Sharon Lederer; Anne M Larson; Robert L Carithers; Arndt Benecke; Michael G Katze
Journal:  Hepatology       Date:  2012-06-05       Impact factor: 17.425

3.  CTIP2 is a negative regulator of P-TEFb.

Authors:  Thomas Cherrier; Valentin Le Douce; Sebastian Eilebrecht; Raphael Riclet; Céline Marban; Franck Dequiedt; Yannick Goumon; Jean-Christophe Paillart; Mathias Mericskay; Ara Parlakian; Pedro Bausero; Wasim Abbas; Georges Herbein; Siavash K Kurdistani; Xavier Grana; Benoit Van Driessche; Christian Schwartz; Ermanno Candolfi; Arndt G Benecke; Carine Van Lint; Olivier Rohr
Journal:  Proc Natl Acad Sci U S A       Date:  2013-07-12       Impact factor: 11.205

4.  Nonpathogenic SIV infection of African green monkeys induces a strong but rapidly controlled type I IFN response.

Authors:  Béatrice Jacquelin; Véronique Mayau; Brice Targat; Anne-Sophie Liovat; Désirée Kunkel; Gaël Petitjean; Marie-Agnès Dillies; Pierre Roques; Cécile Butor; Guido Silvestri; Luis D Giavedoni; Pierre Lebon; Françoise Barré-Sinoussi; Arndt Benecke; Michaela C Müller-Trutwin
Journal:  J Clin Invest       Date:  2009-12       Impact factor: 14.808

5.  The microbiota mediates pathogen clearance from the gut lumen after non-typhoidal Salmonella diarrhea.

Authors:  Kathrin Endt; Bärbel Stecher; Samuel Chaffron; Emma Slack; Nicolas Tchitchek; Arndt Benecke; Laurye Van Maele; Jean-Claude Sirard; Andreas J Mueller; Mathias Heikenwalder; Andrew J Macpherson; Richard Strugnell; Christian von Mering; Wolf-Dietrich Hardt
Journal:  PLoS Pathog       Date:  2010-09-09       Impact factor: 6.823

6.  1918 Influenza virus hemagglutinin (HA) and the viral RNA polymerase complex enhance viral pathogenicity, but only HA induces aberrant host responses in mice.

Authors:  Tokiko Watanabe; Jennifer Tisoncik-Go; Nicolas Tchitchek; Shinji Watanabe; Arndt G Benecke; Michael G Katze; Yoshihiro Kawaoka
Journal:  J Virol       Date:  2013-02-28       Impact factor: 5.103

7.  7SK small nuclear RNA directly affects HMGA1 function in transcription regulation.

Authors:  Sebastian Eilebrecht; Guillaume Brysbaert; Thomas Wegert; Henning Urlaub; Bernd-Joachim Benecke; Arndt Benecke
Journal:  Nucleic Acids Res       Date:  2010-11-17       Impact factor: 16.971

8.  A chemokine gene expression signature derived from meta-analysis predicts the pathogenicity of viral respiratory infections.

Authors:  Stewart T Chang; Nicolas Tchitchek; Debashis Ghosh; Arndt Benecke; Michael G Katze
Journal:  BMC Syst Biol       Date:  2011-12-22

9.  TAF6delta orchestrates an apoptotic transcriptome profile and interacts functionally with p53.

Authors:  Emmanuelle Wilhelm; Mara Kornete; Brice Targat; Jimmy Vigneault-Edwards; Mattia Frontini; Laszlo Tora; Arndt Benecke; Brendan Bell
Journal:  BMC Mol Biol       Date:  2010-01-22       Impact factor: 2.946

10.  TAF6delta controls apoptosis and gene expression in the absence of p53.

Authors:  Emmanuelle Wilhelm; François-Xavier Pellay; Arndt Benecke; Brendan Bell
Journal:  PLoS One       Date:  2008-07-16       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.