Bongsong Kim1. 1. Department of Agronomy, Iowa State University, Ames, IA, USA.
Abstract
Hierarchical association coefficient algorithm calculates the degree of association between observations and categories into a value named hierarchical association coefficient (HA-coefficient) between 0 for the lower limit and 1 for the upper limit. The HA-coefficient algorithm can be operated with stratified ascending categories based on the average of observations in each category. The upper limit refers to a condition where observations are increasingly ordered into the stratified ascending categories, whereas the lower limit refers to a condition where observations are decreasingly ordered into the stratified ascending categories. An HA-coefficient represents how close an observed categorization is to the upper limit, or how distant an observed categorization is from the lower limit. To demonstrate robustness and reliability, the HA-coefficient algorithm was applied to 3 different simulated data sets with the same pattern in terms of the association between observations and categories. From all simulated data sets, the same result was obtained, indicating that the HA-coefficient algorithm is robust and reliable.
Hierarchical association coefficient algorithm calculates the degree of association between observations and categories into a value named hierarchical association coefficient (HA-coefficient) between 0 for the lower limit and 1 for the upper limit. The HA-coefficient algorithm can be operated with stratified ascending categories based on the average of observations in each category. The upper limit refers to a condition where observations are increasingly ordered into the stratified ascending categories, whereas the lower limit refers to a condition where observations are decreasingly ordered into the stratified ascending categories. An HA-coefficient represents how close an observed categorization is to the upper limit, or how distant an observed categorization is from the lower limit. To demonstrate robustness and reliability, the HA-coefficient algorithm was applied to 3 different simulated data sets with the same pattern in terms of the association between observations and categories. From all simulated data sets, the same result was obtained, indicating that the HA-coefficient algorithm is robust and reliable.
Entities:
Keywords:
GWAS; Genome-wide association study; HA-coefficient; Hierarchical Association Coefficient; QTL
A genome-wide association study (GWAS) is an analysis of categorical data. The GWAS data consist of categorical variables (categories) with patterned DNA sequences and a quantitative variable (observations) with real numbers for a trait of interest. This article introduces a new method for measuring association between categories and observations, named the hierarchical association coefficient (HA-coefficient) algorithm. The algorithm measures the association between categories and observations based on the degree of variance among the averages for all categories. If averages across n categories are similar, this suggests a situation where observations are randomly distributed into categories. If averages across different categories are clearly different, this suggests a situation where observations are assigned into categories by some criterion, and it can be said that categories and observations are associated. This foundation also applies to the F test which calculates a P value referring to the degree of variance among averages for all categories and is widely used for GWAS.[1-4]To measure the association between categories and observations, the HA-coefficient algorithm uses 2 sorting extremes: (1) observations being increasingly sorted into stratified ascending categories (HA-coefficient = 1), and (2) observations being decreasingly sorted into stratified ascending categories (HA-coefficient = 0). Note that the stratified ascending categories means a condition where observed categories are aligned in ascending order based on the average of observations in each category. The sorting extremes are conditions where the degree of variance among the averages for all categories are maximized. Meanwhile, the F test calculates a ratio of intercategorical variability to intracategorical variability, in which the greater the ratio, the more variance among the averages for all categories is found.[5] Simulations revealed that the HA-coefficient algorithm and F test produce similar results. The F test is a method for a hypothetical test, whereas the HA-coefficient algorithm calculates an objective measurement.
Theory and Methods
Hierarchical association distance
Given the whole population set has 2 or more members and is categorical, let us make the following conventions:Every member has a positive real number as an observation.Every member has a categorical identifier.Averages of observations in different categories are different.Then, the categories can be stratified based on the average of observations. On this basis, let us define:
Definition 1
“Hierarchical” means that all categories are stratified in ascending order based on the average of each category.
Definition 2
Suppose that all categorical boundaries in hierarchical stratification are fixed, and observations are permutable. “Top categorization” means a condition in which observations are arranged in ascending order in each category leading to ascending order across all categories.
Definition 3
Suppose that all categorical boundaries in hierarchical stratification are fixed, and observations are permutable. “Bottom categorization” means a condition in which observations are arranged in descending order in each category leading to descending order across all categories.
Definition 4
“Hierarchical association coefficient” means a proportion representing how close the top and observed categorizations are, or how distant the bottom and observed categorizations are.
Definition 5
Suppose that n categories are stratified in ascending order based on the average of each category from left to right, in which n = the number of all categories. There are n − 1 categorical boundaries. At each categorical boundary, we can make 2 categories by collapsing the other categorical boundaries. Let us call the result “hierarchical binary categorization” and designate the sum of the right subset as x1 and the sum of the left subset as x2 at any categorical boundary. The x1 is a representative value for a respective hierarchical binary categorization.Regarding Definitions 1 to 3, graphical instructions are shown in Figure 1. Definition 5 always assures that (1) x1 in the top categorization is equal to or greater than x1 in the observed categorization, and (2) x1 in the bottom categorization is equal to or less than x1 in the observed categorization. The use of x1 allows us to quantify the hierarchical association distance by substituting x1 as a value for an observed categorization for x in the following equation:
Figure 1.
Three categorizations (A, B, C) including 3 categories (blue, green, red). Each bar represents an observation. All categorizations contain the same observations. (A) Observed categorization in which categories are sorted in ascending order based on the average of each category. (B) The top categorization. (C) The bottom categorization.
where x is the variable, is the hierarchical association distance given x, g1 is the x1 in the top categorization, g2 is the x2 in the top categorization, and y is the sum of all observations.Three categorizations (A, B, C) including 3 categories (blue, green, red). Each bar represents an observation. All categorizations contain the same observations. (A) Observed categorization in which categories are sorted in ascending order based on the average of each category. (B) The top categorization. (C) The bottom categorization.Equation 1 can be derived as follows:Substitute x as a variable for r1 so that r2 = y − x. Then,where y is the sum of all observations, g1 is the x1 in the top categorization, g2 is the x2 in the top categorization, r1 is the x1 in the observed categorization, r2 is the x2 in the observed categorization, x is the variable, and d is the hierarchical association distance given x.It is always true that and so that . In the top categorization, , whereas the bottom categorization maximizes . At any categorical boundary, x1 and x2 must be different. Otherwise, is unsolvable. Figure 2 shows a graph for in which x1s for the bottom, observed, and top categorizations are 10, 25, and 40, respectively. The graph can be drawn only in quadrant I; that is, only positive real numbers can be observations.
Figure 2.
A curve for . The whole colored area is delimited by two x1s in the bottom and top categorizations. The orange area is delimited by two x1s in the bottom and observed categorizations.
A curve for . The whole colored area is delimited by two x1s in the bottom and top categorizations. The orange area is delimited by two x1s in the bottom and observed categorizations.
HA-coefficient algorithm
Given Equation 1, let us designate the area delimited between x1s at the bottom and top categorizations as W and the area delimited between x1s at the bottom and observed categorizations as R. The W and R represent cumulative hierarchical association distances and can be calculated as follows:Ultimately, the HA-coefficient can be calculated as follows:where HA is the HA-coefficient, W is the area delimited between x1s in the bottom and top categorizations, R is the area delimited between x1s in the bottom and observed categorizations, g1 is the x1 in the top categorization, g2 is the x2 in the top categorization, y is the sum of all observations, x is the variable, obs. is the x1 in the observed categorization, btm is the x1 in the bottom categorization, and top is the x1 in the top categorization.It is always true that 0 ≤ R ≤ W so that the HA-coefficient results in a proportion. If x1 in the observed categorization equals x1 in the bottom categorization, HA-coefficient = 0. If x1 in the observed categorization equals x1 in the top categorization, HA-coefficient = 1. Equation 4 calculates an HA-coefficient if the whole population set consists of 2 categories. If the whole population set consists of equal to or more than 2 categories, either of the following 2 algorithms can be used:HA-coefficient algorithm based on geometric meanHA-coefficient algorithm based on arithmetic meanwhere HA is the HA-coefficient, n is the total number of categories, k is the loop variable, y is the sum of all observations, x is the variable, obs.[k] is the x1 in the observed categorization given the kth categorical boundary, btm[k] is the x1 in the bottom categorization given the kth categorical boundary, and top[k] is the x1 in the top categorization given the kth categorical boundary.Equations 5 and 6 produce closely similar results. I recommend Equation 5 because unification of the equation is foundational in comparing multiple HA-coefficients. If 2 or more categories have the same average, the HA-coefficient algorithm is not applicable.
Prehierarchical and posthierarchical categorizations
Hierarchical stratification among categories can be determined independent of or dependent on observations.
Definition 6
If hierarchical stratification among categories is determined independent of observations, categories are “prehierarchical.” If hierarchical stratification among categories is determined dependent on observations, categories are “posthierarchical.”Prehierarchical categorization makes it feasible that HA-coefficient = 0, whereas posthierarchical categorization does not.
Simulations
To demonstrate robustness and reliability of the HA-coefficient algorithm, simple simulations were used. Figures 3A, 3B, and 3C refer to matrices of 1200 by 1201. The green triangle refers to the 1201st column, including 1200 natural numbers increasing by 1 from 1001 to 2200. Figures 3A, 3B, and 3C include 2, 3, and 4 couples of blue and yellow triangles, respectively. In each matrix, blue and yellow triangles are equal in shape and area. The number of blue triangles in each matrix equals the number of types of categorical identifiers. Figures 3A, 3B, and 3C have categorical identifiers of 2 (0, 1), 3 (0, 1, 2), and 4 (0, 1, 2, 3) types, respectively. In each matrix, the top blue triangle is filled with 0s, the next blue triangle is filled with 1s, and so on. The yellow triangles are filled with random categorical identifiers. As a column coordinate n changes from 1 to 1200, the HA-coefficient between nth and 1201st columns gradually increases to 1. The minimum HA-coefficient must be greater than 0 because each categorization is posthierarchical. The 100 times simulations were averaged into smooth plots and aim to answer the following questions:
Figure 3.
Three simulated data sets of 1200 by 1201. Green triangles refer to a vector containing 1200 observations increasing by 1 from 1001 to 2200. The 1200 by 1200 squares in (A), (B), and (C) are filled with categorical identifiers of 2, 3, and 4 types, respectively. In each matrix, the first blue triangle is filled with 0s, the next blue triangle is filled with 1s, and so on. The yellow triangles are filled with random categorical identifiers.
Question 1. Do the HA-coefficients from Figures 3A, 3B, and 3C increase from left to right?Question 2. Do the HA-coefficients from Figures 3A, 3B, and 3C coincide?Question 3. Do the HA-coefficients and P values calculated by the F test show a consistent pattern?Three simulated data sets of 1200 by 1201. Green triangles refer to a vector containing 1200 observations increasing by 1 from 1001 to 2200. The 1200 by 1200 squares in (A), (B), and (C) are filled with categorical identifiers of 2, 3, and 4 types, respectively. In each matrix, the first blue triangle is filled with 0s, the next blue triangle is filled with 1s, and so on. The yellow triangles are filled with random categorical identifiers.Regarding Question 3, the linear model (LM) for the F test was set as:i = 1,2,3, …, aj = 1,2,3, …, nwhere yij is the jth observation for ith category, µ is the mean of all observations, is the constant for ith category based on random deviation from µ, and is the random effect containing all uncontrolled sources of variability.[5]Through 100 times simulations, the resulting P values were averaged into smooth plots. If the answers to all questions are positive, the HA-coefficient algorithm is reliable and robust. All computations were conducted using R.[6] All R scripts are included in Supplementary R scripts.
Results and Discussion
Figure 4 shows 3 plots obtained by applying the HA-coefficient algorithm to data sets in Figures 3A, 3B, and 3C. This illustrates a common increasing pattern and gives a positive answer to Question 1. Each plot ranges between about 0.6 and 1.0. Because the simulated data sets are posthierarchical, it is infeasible that HA-coefficient = 0. The 3 different simulated data sets have the same observations at regular intervals and equal proportions of blue and yellow sections when comparing the same columns. Therefore, the 3 simulated data sets have the same pattern in terms of the association between categories and observations. If the HA-coefficient algorithm is robust and reliable, the same result must be produced from the 3 simulated data sets. Table 1 shows the Pearson correlation coefficients among the 3 plots coincide. This gives a positive answer to Question 2. The increasing pattern of all plots in Figures 4 and 5 gives a positive answer to Question 3. All answers to the above questions are positive. This indicates that the HA-coefficient algorithm is robust and reliable. The curves generated by the F test in Figure 5 are bent downward because the −log10 lifts small P values upward but pushes moderate P values downward. The F test (see Equation 7) has the following constraints:
Figure 4.
Plots (A), (B), and (C) represent patterns of the HA-coefficients obtained by applying the HA-coefficient algorithm to data sets in Figures 3A, 3B, and 3C, respectively.
Table 1.
Pearson correlation coefficients among the 3 plots in Figure 4.
Figure 3A
Figure 3B
Figure 3C
Figure 3A
1
0.999893
0.999845
Figure 3B
0.999893
1
0.999883
Figure 3C
0.999845
0.999883
1
Figure 5.
Plots (A), (B), and (C) represent patterns of the P values obtained by applying F test to data sets in Figures 3A, 3B, and 3C, respectively.
Constraint 1. Given the top categorization, P values = 0. It is impossible to represent −log10 (0).Constraint 2. Three assumptions for the LM are required: (1) conform the normal distribution, (2) have the same variance for each i, and are independent of each other and the [5]Plots (A), (B), and (C) represent patterns of the HA-coefficients obtained by applying the HA-coefficient algorithm to data sets in Figures 3A, 3B, and 3C, respectively.Pearson correlation coefficients among the 3 plots in Figure 4.Plots (A), (B), and (C) represent patterns of the P values obtained by applying F test to data sets in Figures 3A, 3B, and 3C, respectively.The above constraints do not apply to the HA-coefficient algorithm. Regarding Constraint 1, the graph lines obtained by the F test (Figure 5) do not reach the right end, while graph lines obtained by the HA-coefficient algorithm (Figure 4) are fully drawn from left to right ends. Regarding Constrain 2, the HA-coefficient algorithm produces an objective measurement; that is, assumptions for statistical inference are not needed. The simulations revealed that the HA-coefficient algorithm is faster than the F test based on the LM, e.g. when applied to Figure 3A, the former and the latter took 739 and 956 seconds (Intel i7-5600U CPU), respectively.
Conclusion
This study shows a comparison of the HA-coefficient algorithm and F test because both methods calculate the association between categories and observations based on the degree of variance among averages for all categories. The HA-coefficient algorithm’s objectivity, reliability, robustness, and speed enable the algorithm to become an alternative to the F test. When it comes to GWAS, the HA-coefficient algorithm will be suited for a population grown in the same environment because the same environment is fundamental in identifying unbiased QTL. Posthierarhichical categorizations are shown by the data sets in Figure 3. GWAS data sets have the posthierarchical categorization. The application of the HA-coefficient algorithm to a prehierarchical categorization is shown in Supplementary example. The HA-coefficient algorithm will be useful in many disciplines.
Authors: Sebastiaan M Bol; Perry D Moerland; Sophie Limou; Yvonne van Remmerden; Cédric Coulonges; Daniëlle van Manen; Joshua T Herbeck; Jacques Fellay; Margit Sieberer; Jantine G Sietzema; Ruben van 't Slot; Jeremy Martinson; Jean-François Zagury; Hanneke Schuitemaker; Angélique B van 't Wout Journal: PLoS One Date: 2011-02-25 Impact factor: 3.240
Authors: Naomi Duijvesteijn; Egbert F Knol; Jan W M Merks; Richard P M A Crooijmans; Martien A M Groenen; Henk Bovenhuis; Barbara Harlizius Journal: BMC Genet Date: 2010-05-20 Impact factor: 2.797
Authors: Youna Hu; Alena Shmygelska; David Tran; Nicholas Eriksson; Joyce Y Tung; David A Hinds Journal: Nat Commun Date: 2016-02-02 Impact factor: 14.919
Authors: Bongsong Kim; Xinbin Dai; Wenchao Zhang; Zhaohong Zhuang; Darlene L Sanchez; Thomas Lübberstedt; Yun Kang; Michael K Udvardi; William D Beavis; Shizhong Xu; Patrick X Zhao Journal: Bioinformatics Date: 2019-07-15 Impact factor: 6.937