Literature DB >> 24350232

Multiple hypotheses testing procedures in clinical trials and genomic studies.

Qing Pan1.   

Abstract

We review and compare multiple hypothesis testing procedures used in clinical trials and those in genomic studies. Clinical trials often employ global tests, which draw an overall conclusion for all the hypotheses, such as SUM test, Two-Step test, Approximate Likelihood Ratio test (ALRT), Intersection-Union Test (IUT), and MAX test. The SUM and Two-Step tests are most powerful under homogeneous treatment effects, while the ALRT and MAX test are robust in cases with non-homogeneous treatment effects. Furthermore, the ALRT is robust to unequal sample sizes in testing different hypotheses. In genomic studies, stepwise procedures are used to draw marker-specific conclusions and control family wise error rate (FWER) or false discovery rate (FDR). FDR refers to the percent of false positives among all significant results and is preferred over FWER in screening high-dimensional genomic markers due to its interpretability. In cases where correlations between test statistics cannot be ignored, Westfall-Young resampling method generates the joint distribution of P-values under the null and maintains their correlation structure. Finally, the GWAS data from a clinical trial searching for SNPs associated with nephropathy among Type 1 diabetic patients are used to illustrate various procedures.

Entities:  

Keywords:  false discovery rate; family wise error rate; global test; multiple hypotheses testing; resampling method; stepwise procedure

Year:  2013        PMID: 24350232      PMCID: PMC3859974          DOI: 10.3389/fpubh.2013.00063

Source DB:  PubMed          Journal:  Front Public Health        ISSN: 2296-2565


Introduction

When more than one hypotheses are tested at the same time, it is well known that the family wise type I error rate (FWER), that is, the probability of reporting at least one significant finding when the null hypotheses are true, will be inflated. Take J independent test statistics as an example. When each test controls its type I error rate at α level, the FWER is 1 − (1 − α). Table 1 lists the FWERs for different combinations of J and α. When J = 10 and α = 0.05, FWER goes up to 0.401. In cases of 100 or more simultaneous tests, it is almost sure to get false positive results.
Table 1

FWER versus number of tests and the size of individual tests.

αJFWER
0.0120.020
0.0150.049
0.01100.096
0.011000.634
0.0110001.000
0.0520.098
0.0550.226
0.05100.401
0.051000.994
0.0510001.000
FWER versus number of tests and the size of individual tests. Multiple hypotheses testing arises frequently both in clinical trials and in genomic studies. The different goals in these two settings result in different strategies. First, the hypotheses in clinical trials are often considered as a whole while those in genomic studies are more independent from each other. In clinic trials, multiple hypotheses are often considered jointly with a coherent theme. A few examples are given as follows. The symptoms of a complex disease often show up in different parts of the body or in different forms, such as different types of cancer. Multiple laboratory measurements monitor the underlying disease process, such as the viral loads and CD4 cell counts in HIV positive subjects. A treatment might have different responses in different patient sub-populations. On the other hand, multiple hypotheses in genomic studies arise because a large number of candidate markers are tested at the same time. Based on the number of tests carried out in the procedure, the multiple testing adjustment approaches can be grouped into global tests and stepwise procedures (1). Global tests summarize information from all endpoints/measurements/strata in one test statistic, while stepwise procedures carry out one test for each hypothesis. Therefore global tests are employed frequently in clinical trials while genomic studies almost always employ stepwise procedures. Second, the hypotheses in clinical trials are usually more specific with abundant prior information. In testing a specific treatment, with knowledge on the direction of the effects, directional tests with higher power can be employed. On the other hand, the genomic, epigenomic, transcriptomic, and proteomic network is much more complicated and often researchers screen for any signal, without knowing its direction or relationship to other markers. Third, the numbers of hypothesis in clinical trials are on a much smaller scale compared to the numbers in genomic studies – the numbers in clinical trials are usually less than ten, while the numbers of potential markers in genomic studies are sometimes over a million. In this manuscript, common procedures of multiple hypotheses adjustment in the two different settings are reviewed and compared. The effects of interests are usually inferred from regression coefficients. In linear regression for normally distributed outcomes, the coefficient represents the difference in the outcome values between the groups being compared. In generalized linear models with logit link for binary outcomes, the coefficient equals the logarithm of the odds ratio of the outcome in the treatment group relative to the control group. In Cox proportional hazards models for partially censored failure time data, the exponentiated coefficient represents the hazards ratio. This review focuses on the choice of proper multiple testing adjustment method after the estimation procedures. Hence, we assume that appropriate models are chosen for different data configurations and parameters and covariance matrix are consistently estimated. Suppose there are J hypotheses in total. Let β and denote the two J × 1 vectors of regression coefficients and their estimates, respectively, one element for each hypothesis. Furthermore, β = 0 corresponds to the jth null hypothesis, j  = 1, …, J.

Multiple Testing Procedures in Clinical Trials

SUM test

O’Brien (2) proposed a test derived from the generalized least squares principle where J is an J × 1 vector of 1’s and Σ is the covariance matrix of . When elements of are independent from each other, the O’Brien test statistic reduces to a linear combination of where each is weighted by inverse of its variances. Tests employing linear combinations of with different weights have been proposed (3–6), among which the SUM test is especially popular (7). The SUM test statistic has a simple sum form Under the null hypothesis , E(SUM) = 0. The SUM test is found to maximize the minimum power (maxmin test) for alternatives where all elements of β have the same sign (8, 9).

Two-step

When homogeneous effects are of interests, a two-step procedure is commonly used. In the first step, we test versus H: at least one pair for j ≠ j′ through Breslow-Day test or likelihood ratio test (LRT) (10, 11). Under the null, the LRT test statistic follows a Chi-square distribution with J − 1 degree of freedom asymptotically. If the null hypothesis of homogeneous treatment effects is not rejected, we proceed to the second step where data from different endpoints are pooled and an overall treatment effect is estimated and tested against zero with a Wald test. The second test is carried out conditionally on the acceptance of the null in the first step. When the type I error rates in the two steps, α1 and α2, both equal 0.05, the marginal probability that the Two-Step procedure concludes homogeneous non-zero treatment effects under H0 is 95% × 5% = 4.75%, while the probability of concluding non-zero treatment effect in at least one endpoint under H0 is 95% × 5% + 5% = 9.75%. Lachin and Wei (12) proposed to adjust α1 and α2 so that the overall type I error rate is α1 + α2(1 − α1) = 0.05.

Approximate likelihood ratio test

The Hotelling’s T test examines whether the vector β is a vector of zero Here n is the sample size in testing each hypothesis. Under H0, the Hotelling’s T test statistic has an asymptotic Chi-square distribution. Follmann (13) modified Hotelling’s T test for one-sided alternatives. His procedure rejects the null when the p-value of the Hotelling’s T test is less than twice its nominal level and the sum of the treatment effects is in the desired direction (positive or negative). Tang et al. (14) proposed an approximate likelihood ratio test (ALRT) for one-sided alternative hypotheses. A J × J matrix A which satisfies A′A  = Σ−1 and AΣA′ = I is calculated, where I denotes the identity matrix. Define where the vector is mapped into a new vector z with independent components z, j  = 1, …, J. For H: at least one β  > 0, the ALRT statistic is calculated as where negative z values contribute zero. Hence the absolute magnitude of negative z has no impact on ALRT. The ALRT statistic follows a mixed Chi-square distribution under H0.

MAX test

Another type of global tests employ the maximum of the standardized test statistics (15). The test statistic goes as follows where is the standard deviation of . Given the one-to-one relationship between and its p-value, an equivalent test statistic is the minimum of the P-values. The MAX test is powerful to detect alternatives where the treatment effects are non-zero in at least one endpoint/measurement/stratum.

Intersection-union test

Establishment of bioequivalency is required by the U.S. Food and Drug Administration (FDA) in approving generic drugs. The brand-name drug and its generic version are considered indifferent for the jth outcome if , where the indifferent range ε is decided clinically. FDA is interested in whether the generic drug is superior in at least one aspect while non-inferior in all aspects. Therefore, the alternative of interest goes as follows H: {max(β1, β2, …, β) > 0} ∩ {min(β1+ε1, β2+ε2, …, β+ε) > 0} where ∩ denotes intersection. The intersection-union test (IUT) (16–18) is most frequently used in these settings. It is a closed procedure which rejects the overall null hypothesis if and only if all null hypotheses included in the procedure are rejected. The ALRT is used to test against the alternative max (β1, β2, …, β) > 0. Non-inferiority in the jth endpoint is tested by Because the overall rejection region is the intersection of all rejection regions, the overall type I error will not exceed α if the type I error rates of individual tests are set at α. Although more than one tests are carried out in IUT, it is included in the category of global tests because it draws an overall conclusion, not multiple hypothesis-specific conclusions. The five global tests are summarized in Table 2.
Table 2

Comparison. of five global test statistics.

TestTest statistic
SUMSUM=j=1Jβ^j
Two-step: step oneLRT=2(L0La)
Two-step: step twoβ^
ALRTALRT=j=1Jmax(Zj,0)2
IUTALRT, Tj=β^j+εjSD(β^j) j = 1, …, J
MAXMAX=maxβ^1SD(β^1),β^2SD(β^2),,β^JSD(β^J)

L0 and L represent the maximum log likelihood under H and H respectively. And is the coefficient estimates from the pooled data.

Comparison. of five global test statistics. L0 and L represent the maximum log likelihood under H and H respectively. And is the coefficient estimates from the pooled data.

Comparison of rejection regions of the global tests

We take the special example with two coefficient estimates , which are are bivariate normal with mean (0, 0), variance 1 and 2 respectively, and correlation coefficient 0.3. The null and alternative hypotheses are H0: (β1 = 0) ∩ (β2 = 0) versus H: (β1 > 0) ∩ (β2 > 0). The rejection regions of the five global tests are shown in Figure 1, when α = 0.05 in each individual test. The five rejection regions imply that each test has optimal power against different alternatives. The Wei-Lachin SUM test rejects outside a straight line with slope –1 which represents a constant sum. The rejection region of the Two-Step test can be viewed as removing two sides from the rejection region of the SUM test. The MAX test and ALRT reject points with a large positive value in at least one dimension. The rejection region of the IUT eliminates points with negative or close to zero values in any endpoint compared to the rejection region of ALRT.
Figure 1

Comparison of rejection regions of five global tests.

Comparison of rejection regions of five global tests.

Simulation Studies

We simulate binary data following a logistic model to illustrate the global tests. Two different scenarios are examined – correlated multiple outcomes and independent stratified data. For correlated outcomes, each subject i has two endpoints. The independent data are from two strata. Two independent covariates are generated: a binomial variable X1 with equal probability to be zero or one and a normal variable X2 with mean 0 and standard deviation 5. The outcomes Y follow Bernoulli distribution specified by logit{p(Y = 1)} = η + β1 + θX2. Note the effects of the treatment X1 is reflected by two endpoint-specific regression coefficients, β1 and β2. Correlated binary outcomes are generated following Park, Park, and Shin method (19). The intercepts for endpoint 1 and 2 are η1 = 0.5, η2 = 0.2, and the coefficient for X2 is θ = 0.1 for both endpoints. In simulating the independent binary data, θ = 0.02. In case of unequal sample sizes in the two endpoints, observations in the endpoint with less subjects are missing completely at random. Maximum likelihood estimator for β1 and β2, as well as the covariance matrix are calculated through generalized estimating equations (20). One-sided alternatives H: (β1 > 0)∩(β2 > 0) are tested. Test statistics are calculated using , . Each setup is repeated 1000 times. In each iteration, all the test statistics are calculated using the same dataset. We examine and compare the powers and Type I error rates of all five tests for different true values (Table 3), different levels of correlations (Table 4), and different sample sizes at each endpoint (Table 5).
Table 3

Simulation results: size and power (%) with different true value positions.

True values
ρTesting procedure
β1β2SUMTwo-stepIUTALRTMAX
000.44.94.70.74.44.5
0.30.30.4212051818
0.60.60.45250234846
0.60.30.43533113536
0.600.4211733233
0.6−0.30.411713329

The Type I error 4.7% for Two-Step refers to cases rejected in Step Two. The Type I error for the three tests in IUT are 4.3%, 5.1%, and 4.4%, respectively. The intersection of the three rejection regions gives the overall Type I error rate.

Table 4

Simulation results: power (%) under different correlation between outcomes.

True values
ρTesting procedure
β1β2SUMTwo-stepIUTALRTMAX
0.60.606361185852
0.60.60.45250234846
0.60.60.84240273741
0.60.30444274038
0.60.30.43533113536
0.60.30.82624122830

ρ Represents the correlation coefficient between the outcomes, not the correlation between the estimated regression coefficient.

Table 5

Simulation results: power (%) with different sample sizes.

True values
Sample size
ρTesting procedure
β1β2n1n2SUMTwo-stepIUTALRTMAX
0.60.61001000.45250234846
0.60.6501500.44137154930
0.60.6251750.4302655116
0.60.31001000.43533113536
0.60.3501500.4272672321
0.60.3251750.418183239
0.60.610010006361185852
0.60.65015005550145537
0.60.6251750363495417
0.60.31001000444274038
0.60.3150500373374724
0.60.3175250242254810
Simulation results: size and power (%) with different true value positions. The Type I error 4.7% for Two-Step refers to cases rejected in Step Two. The Type I error for the three tests in IUT are 4.3%, 5.1%, and 4.4%, respectively. The intersection of the three rejection regions gives the overall Type I error rate. Simulation results: power (%) under different correlation between outcomes. ρ Represents the correlation coefficient between the outcomes, not the correlation between the estimated regression coefficient. Simulation results: power (%) with different sample sizes. The powers and Type I error rates for different (β1, β2) values are listed in Table 3. The correlation between Y i1 and Y i2 is set to be 0.4 and each endpoint has 100 observations. All tests except the IUT maintain the Type I error rates close to the nominal level 0.05. Without prior knowledge of the indifference range, we set the most restrictive indifference range where ε  = 0 for every endpoint which is equivalent to requiring all treatment effects to be positive, leading to low overall type I error rate and power. IUT tends to be more conservative than other methods because FDA is more concerned with false positives and only approves new treatment when there is significant evidence supporting its superiority. The procedures can be divided into two groups according to how the power changes when the difference between β1 and β2 gets larger. The first group includes Wei-Lachin SUM and Two-Step. They are more powerful than the other group when β1 = β2, but sensitive to non-homogeneous treatment effects. The power of the Wei-Lachin SUM test drops from 52 to 21% when β2 drops from 0.6 to 0 while β1 remains 0.6. The decreasing trend is even more obvious with the Two-Step. The second group includes ALRT and the MAX test. They are robust to non-homogeneous treatment effects. In Table 4, 100 correlated pairs (Y1, Y2) are generated with various correlation coefficients. All the methods incorporate information from both endpoints. When two outcomes are highly correlated, the treatment effects estimated from both endpoints, and , tend to be similar and provide less information compared to the independent case, hence lower power. However, the IUT has a reversed pattern because with higher positive correlation, the non-inferiority tests on the two endpoints tend to agree more, leading to higher overall rejection rates. Table 5 lists the different performance of the tests with unequal sample sizes for the two endpoints. When sample sizes are not balanced between the two endpoints, most tests have reduced power because the test statistics combine information from all endpoints and a large variance in one endpoint leads to large variance of the overall test statistic. ALRT is robust to unequal sample sizes. If treatment effects are equal in both endpoints (β1 = β2 = 0.6), the power of ALRT does not change with the distribution of samples into each two endpoints as long as the total sample size remains the same. When the treatment effects differ in the two endpoint (β1 = 0.6, β2 = 0.3), the power of ALRT could either increase or decrease depending on which endpoint has more subjects. If the endpoint with a larger treatment effects has a larger sample size, ALRT has higher power. If the endpoint with a smaller treatment effect gets more samples, the power decreases.

Controlling FWER/FDR in Genomic Studies

Stepwise procedures and FDR

Stepwise procedures are classified into one-step procedures and multi-step procedures. One-step procedures set a uniform threshold for all the unadjusted P-values while multi-step procedures set different thresholds for different hypotheses depending on the order of the unadjusted P-values. Multi-step procedures can be carried out step-down or step-up (21, 22). In step-down procedures, the hypothesis with the smallest P-value is tested first. And as long as one hypothesis fails to be rejected, all the hypotheses with larger unadjusted P-values will fail to be rejected. On the contrary, step-up procedures start from the largest unadjusted P-value and reject all smaller unadjusted P-values after the first one is rejected. In this manuscript, the FWER is preserved at nominal level in a strong sense, that is, FWER is no larger than the nominal level for testing any subset of the hypotheses set. Given the large number of hypotheses, researchers are often more interested in a more interpretable quantity, the FDR (23). FDR is the rate that the rejected or significant features are truly null. The numbers of true and false positives can be calculated directly from FDR. FDR helps to avoid a flood of false positives when most of the hypotheses are truly null or missing out significant features when the number of true alternative hypotheses is large. FDR can be estimated as where m is the total number of hypotheses being tested, π0 is the percent of true null among them, I is the indicator for a true statement in the bracket, and t is the cutoff value of p-values to call a feature significant. Although π0 is unknown, it can be estimated from the distribution of P-values. Benjamini and Hochberg (24) developed a step-up procedure to control FDR at level q*. For ordered unadjusted P-values P(1), P(2), …, P(), we reject the first j hypotheses with the smallest j P-values if .

Resampling method

Westfall and Young (25) and Troendle (26) developed resampling procedures which simulate the joint distributions of the P-values under the null while maintaining their correlation structure. The procedure starts with bootstrap or permutation under the null from the original sample. Then hypothesis-specific pivotal test statistics and the corresponding P-values are calculated on the simulated data. The steps are repeated a large number of times to achieve an empirical distribution of (P1, P2, …, P) under the null which maintains the correlation structure. The unadjusted P-values for the jth hypothesis is the percent of times the jth imputed test statistic is larger than or equal to the jth test statistic from the original data. Step-down or step-up procedures can be carried out on the unadjusted P-values based on resampling. There is a resampling option in SAS “multtest” procedure for several tests including the two-sample t-test, Cochran-Armitage test and Fisher’s exact test. However, this procedure does not allow covariate adjustments and can not be used in multiple comparisons in regressions.

Bonferroni adjustment

The Bonferroni adjustment is a one-step procedure which rejects the jth null hypothesis H0 when the p-value in testing the jth hypothesis . The FWER in the Bonferroni procedure is conserved at α level because where ∪ denotes union. Alternatively, researchers may compute adjusted P-values as and compare to the nominal level α. The Bonferroni adjustment is computationally straight forward because the threshold for significant P-values in each hypothesis is just the FWER divided by the number of hypotheses. However, Bonferroni procedure is conservative with low power. Wiens (27) and Huque and Alosh (28) modified the Bonferroni procedure with fixed testing sequence procedure. It allocates the overall Type I error rate sequentially and controls FWER at the nominal level. Let the sequence of hypotheses be . Assign type I error rate α to each of the null hypothesis such that . Furthermore, if the first hypothesis is not rejected, its portion of the type I error α1 will be passed onto the second hypothesis. That is, the type I error rate for the second hypothesis becomes α1 + α2 conditional on that fails to be rejected. On the contrary, if is rejected, the type I error rate of remains α2. In summary, the type I error rates of unrejected hypotheses accumulate and are passed onto the next hypotheses until a hypothesis is rejected or the last hypothesis .

Holm, Sidak, and Simes procedures

Holm method is a step-down procedure (29). First, it ranks all the observed p-values from smallest to largest P(1), P(2), …, P() Compare each P to starting from the smallest P(1). Let the first occurrence of be the kth ordered p-value. Then hypotheses corresponding to the first k − 1 p-values P(1), …, P(k−1) will be rejected and the hypotheses from the kth one on corresponding to P(), …, P() will not be rejected. Alternatively, researchers can also compute the Holm’s adjusted p-values and compare them to α. The adjusted p-values is based on the ordered p-values P(1), P(2), …, P() and where ∧ denotes taking the minimum. The adjusted p-values are capped at 1 by taking the minimum of and 1. Besides, the jth adjusted p-value is the maximum in the first j values, resulting in non-decreasing sequence of adjusted P-values. The Sidak (30) correction assumes that the J test statistics are mutually independent and replaces the element-wise p-value cutoff α/J by . It is less conservative than the Bonferroni correction because for n ≥ 1. Another set of thresholds combining the Holm threshold and Sidak correction, , also maintains FWER at α. Simes (31) procedure is also a step down procedure that rejects H0 when . Here P(1), P(2), …, P() are the ordered P-values from smallest to largest. Hochberg and Liberman (32) extended the Simes procedure by allocating different weights to the P-values depending on prior information on each hypothesis.

A Real Case: Genomic Studies Based on a Clinical Trial

We illustrate the stepwise procedures using the Genome Wide Association Study (GWAS) from the Diabetes Control and Complications Trial (DCCT) and Epidemiology of Diabetes Intervention and Complication (EDIC) trial. DCCT and EDIC are two clinical trials based on the same type 1 diabetes cohort in different time periods. The survival rate and life expectancy of type 1 diabetic patients have been improved greatly in recent years. However, chronic hyperglycemia status leads to deleterious changes in blood vessels. Cardiovascular diseases and microvascular complications are major threats to the long-term quality of life of type 1 diabetic patients. This study focuses on microvascular complications among type 1 diabetic patients. In EDIC, 1441 Type 1 diabetic patients enrolled from 1983 to 1989. They were randomized to either the intensive or conventional therapies, where participants in the intensive group monitored and regulated their blood glucose level constantly. DCCT ended in 1993 when significant reduction in the risk of microvascluar complications was found in the intensive therapy group (33). Of the 1441 DCCT participants, 1394 continued to the EDIC trial, where everyone receives the intensive therapy. The abnormalities in the capillaries lead to symptoms in different parts of the body – nephropathy, retinopathy, and neuropathy. The goal of this analysis is to validate fourteen SNPs associated with severe nephropathy and persistent microalbuminuria in Al-Kateb et al. (34). Urine glomerular filtration rates (GFR), which is an important clinical index of diabetic nephropathy, have been recorded annually in the DCCT/EDIC cohort. Log-transformed GFR values are employed as our main outcome. Linear regressions of the last GFR observation versus each of the fourteen SNPs are fitted, adjusting for age at randomization, gender and duration of diabetes at enrollment, stratified by the treatment group. Different SNP coefficients are assumed in the intensive and conventional treatment groups because patients under the two treatments were in quite different biophysical and metabolic statuses and the intensive control of the glucose level might suppress or activate SNP effects. Among the global tests for SNP effects in different strata, the SUM test is employed as the effects for each SNP are expected to be in the same direction across the two strata and the SUM test is the maxmin test under such conditions. Fourteen raw P-values are generated from the SUM tests, one for each SNP. To maintain the family wise Type I error rate or false discovery rate, four different stepwise procedures are performed – Bonferroni, Sidak, Hochberg, and FDR. All four procedures are directly available in SAS package “multtest.” P-values of various procedures are listed in Table 6. We can see that although some raw P-values are <0.05, none of the adjusted P-values remain significant. That is, after FWER is controlled, the seemingly significant results are not actually significant any more. Among the procedures controlling FWER, the Sidak and Hochberg procedures give smaller adjusted P-values and therefore are more powerful than the Bonferroni adjustment. Although researchers usually require the FWER no larger than 0.05, they might set higher cutoff value of FDR depending on the context of the research problem.
Table 6

Real data analysis: association between log(GFR) and 14 SNPs.

SNPMinorRawBonferroniSidakHochbergFDR
rs307806A0.010710.14990.13990.14990.1499
rs2279622T0.033830.47360.38230.43980.1939
rs4693614G0.063190.88460.59900.69240.1939
rs11715496A0.067020.93820.62140.69240.1939
rs8042694G0.069240.96930.63380.69240.1939
rs2259458T0.226391.00000.97250.97910.4245
rs3824935T0.235551.00000.97670.97910.4245
rs2027440C0.242561.00000.97950.97910.4245
rs2276768T0.309941.00000.99440.97910.4821
rs10497435C0.446261.00000.99970.97910.6248
rs3814995T0.520581.00001.00000.97910.6626
rs2705897T0.614451.00001.00000.97910.7169
rs7844961T0.735931.00001.00000.97910.7925
rs4900312A0.979141.00001.00000.97910.9791
Real data analysis: association between log(GFR) and 14 SNPs.

Discussion

This manuscript reviews methods for the multiple hypothesis testing problem. Five global tests widely used in clinical trials are reviewed: SUM test, Two-Step test, ALRT, IUT, and the MAX Test. The plots of the rejection regions illustrate the different alternatives to which the tests are directed. The SUM and Two-Step tests are powerful for alternatives with homogeneous effects. Two-Step test can be viewed as a modification of the SUM test that incorporates information on how different the treatment effects are and thus more sensitive to non-homogeneous treatment effects. ALRT is robust to not only unequal treatment effects but also unequal sample sizes from the endpoints. MAX test is also robust for non-homogeneous treatment effects. IUT provides information about the overall superiority and individual non-inferiority. In genomic studies, specific conclusions on individual hypotheses are desired and stepwise procedures are commonly used to control FWER or FDR. The Westfall and Young’s resampling method generates the joint distribution of P-values under the null and maintains the correlation structure between them. A selected SNP dataset from a clinical trial is used to illustrate the stepwise procedures. Finally, among the hundreds of papers on multiple hypothesis testing topic, only a selected few commonly used multiple hypothesis testing adjustment methods are reviewed here. Our goal is to introduce the classical methods and present the ideas behind them. They serve as the basis on which researchers may choose and develop their own method with careful consideration of the particular research setup and clinical questions.

Conflict of Interest Statement

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
  11 in total

1.  One-sided tests in clinical trials with multiple endpoints.

Authors:  D A Bloch; T L Lai; P Tubert-Bitter
Journal:  Biometrics       Date:  2001-12       Impact factor: 2.571

2.  Statistical significance for genomewide studies.

Authors:  John D Storey; Robert Tibshirani
Journal:  Proc Natl Acad Sci U S A       Date:  2003-07-25       Impact factor: 11.205

3.  A note on one-sided tests with multiple endpoints.

Authors:  Michael D Perlman; Lang Wu
Journal:  Biometrics       Date:  2004-03       Impact factor: 2.571

4.  Procedures for two-sample comparisons with multiple endpoints controlling the experimentwise error rate.

Authors:  W Lehmacher; G Wassmer; P Reitmeir
Journal:  Biometrics       Date:  1991-06       Impact factor: 2.571

5.  The analysis of multiple endpoints in clinical trials.

Authors:  S J Pocock; N L Geller; A A Tsiatis
Journal:  Biometrics       Date:  1987-09       Impact factor: 2.571

6.  On the design and analysis of randomized clinical trials with multiple endpoints.

Authors:  D I Tang; N L Geller; S J Pocock
Journal:  Biometrics       Date:  1993-03       Impact factor: 2.571

7.  Procedures for comparing samples with multiple endpoints.

Authors:  P C O'Brien
Journal:  Biometrics       Date:  1984-12       Impact factor: 2.571

8.  Statistical methods in cancer research. Volume I - The analysis of case-control studies.

Authors:  N E Breslow; N E Day
Journal:  IARC Sci Publ       Date:  1980

9.  The effect of intensive treatment of diabetes on the development and progression of long-term complications in insulin-dependent diabetes mellitus.

Authors:  D M Nathan; S Genuth; J Lachin; P Cleary; O Crofford; M Davis; L Rand; C Siebert
Journal:  N Engl J Med       Date:  1993-09-30       Impact factor: 91.245

10.  Multiple superoxide dismutase 1/splicing factor serine alanine 15 variants are associated with the development and progression of diabetic nephropathy: the Diabetes Control and Complications Trial/Epidemiology of Diabetes Interventions and Complications Genetics study.

Authors:  Hussam Al-Kateb; Andrew P Boright; Lucia Mirea; Xinlei Xie; Rinku Sutradhar; Alireza Mowjoodi; Bhupinder Bharaj; Michelle Liu; Jean M Bucksa; Valerie L Arends; Michael W Steffes; Patricia A Cleary; Wanjie Sun; John M Lachin; Paul S Thorner; Michael Ho; Amy Jayne McKnight; A Peter Maxwell; David A Savage; Kenneth K Kidd; Judith R Kidd; William C Speed; Trevor J Orchard; Rachel G Miller; Lei Sun; Shelley B Bull; Andrew D Paterson
Journal:  Diabetes       Date:  2007-10-03       Impact factor: 9.461

View more
  2 in total

1.  Gastric cancer drug trials - are women second class citizens?

Authors:  Manish A Shah
Journal:  Nat Rev Clin Oncol       Date:  2014-06-03       Impact factor: 66.675

2.  Applications of the Wei-Lachin multivariate one-sided test for multiple outcomes on possibly different scales.

Authors:  John M Lachin
Journal:  PLoS One       Date:  2014-10-17       Impact factor: 3.240

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.