Literature DB >> 21362183

Rule based classifier for the analysis of gene-gene and gene-environment interactions in genetic association studies.

Thorsten Lehr1, Jing Yuan, Dirk Zeumer, Supriya Jayadev, Marylyn D Ritchie.   

Abstract

BACKGROUND: Several methods have been presented for the analysis of complex interactions between genetic polymorphisms and/or environmental factors. Despite the available methods, there is still a need for alternative methods, because no single method will perform well in all scenarios. The aim of this work was to evaluate the performance of three selected rule based classifier algorithms, RIPPER, RIDOR and PART, for the analysis of genetic association studies.
METHODS: Overall, 42 datasets were simulated with three different case-control models, a varying number of subjects (300, 600), SNPs (500, 1500, 3000) and noise (5%, 10%, 20%). The algorithms were applied to each of the datasets with a set of algorithm-specific settings. Results were further investigated with respect to a) the Model, b) the Rules, and c) the Attribute level. Data analysis was performed using WEKA, SAS and PERL.
RESULTS: The RIPPER algorithm discovered the true case-control model at least once in >33% of the datasets. The RIDOR and PART algorithm performed poorly for model detection. The RIPPER, RIDOR and PART algorithm discovered the true case-control rules in more than 83%, 83% and 44% of the datasets, respectively. All three algorithms were able to detect the attributes utilized in the respective case-control models in most datasets.
CONCLUSIONS: The current analyses substantiate the utility of rule based classifiers such as RIPPER, RIDOR and PART for the detection of gene-gene/gene-environment interactions in genetic association studies. These classifiers could provide a valuable new method, complementing existing approaches, in the analysis of genetic association studies. The methods provide an advantage in being able to handle both categorical and continuous variable types. Further, because the outputs of the analyses are easy to interpret, the rule based classifier approach could quickly generate testable hypotheses for additional evaluation. Since the algorithms are computationally inexpensive, they may serve as valuable tools for preselection of attributes to be used in more complex, computationally intensive approaches. Whether used in isolation or in conjunction with other tools, rule based classifiers are an important addition to the armamentarium of tools available for analyses of complex genetic association studies.

Entities:  

Year:  2011        PMID: 21362183      PMCID: PMC3060133          DOI: 10.1186/1756-0381-4-4

Source DB:  PubMed          Journal:  BioData Min        ISSN: 1756-0381            Impact factor:   2.522


Background

Genetic association studies aim to identify the contribution of genetic polymorphisms to specific phenotypes such as disease status, drug responder status, and adverse drug reactions [1]. Association studies have been gaining interest as genotyping costs have significantly decreased and as more association-based success stories are reported [2,3]. In addition to the impact of a single genetic locus, more complex interactions between genetic polymorphisms and/or environmental factors, such as age, weight, and drug exposure can provide more accurate models for the prediction of complex phenotypes [4]. Several methods are currently available for the analysis of gene-gene and gene-environment interactions, e.g. random forests, focused interaction testing frameworks, stepwise logistic regression, explicit logistic regression, Multifactor Dimensionality Reduction (MDR) and Neural Networks (NN) [5,6]. One of the most frequently used algorithm, MDR, allows an exhaustive search for complex interactions [7]. MDR has a good sensitivity to detect gene-gene and gene-environment interactions; however, the algorithm faces two challenges. First, large computational resources are required to perform the exhaustive searching, and the search for higher interaction models can be time consuming for large SNP panels (such as genome-wide association studies). Second, continuous variables have to be binned into categories to be considered for analysis, potentially leading to a loss of information. However, there are some approaches to alleviate this problem [8,9]. An alternative approach for the analysis of genetic association studies is the use of neural networks (NN) and their modifications, e.g. genetic programming neural networks (GPNN) [10] and grammatical evolution neural networks (GENN) [11]. If well trained, neural network models yield good predictivity and allow for the incorporation of continuous variables. However, neural network models are often perceived to have a "black box character" where deconvoluting the contribution and interaction between genetic markers can be challenging. A recent comparison of methods showed that each method demonstrates strengths and weaknesses and there is still a need for alternative methods, because no single method will perform well in all scenarios [4]. Thus, rule based classifier algorithms which have proven performance with non-genetic datasets [12], could provide a valuable complementary method for the analysis of genetic association studies. Rule based classifiers generate classification models using a collection of "if ... then ..." rules [12]. The algorithms are computationally inexpensive, are capable of incorporating categorical and continuous variables and the developed models are usually easy to interpret. The aim of the current study was to evaluate the performance of three selected rule based classifier algorithms, RIPPER [13], RIDOR [14] and PART [15], for the analysis of genetic association studies. Simulated datasets with varying statistical power and three different case-control models were generated in order to perform these evaluations.

Methods

Algorithms

Three different deterministic rule based classifier algorithms, for which the mathematical background is extensively described in the literature, were evaluated in the current study: (a) RIPPER (Repeated Incremental Pruning to Produce Error Reduction) [13] (b) RIDOR (RIpple-DOwn Rule) [14] and (c) PART [15]. The RIPPER algorithm is a direct method, i.e. RIPPER extracts the rules directly from the data. The algorithm progresses through four phases: i) growth, ii) pruning, iii) optimization, iv) selection. In the growth phase, one rule is generated by greedily adding attributes to the rule until the rule meets stopping criteria. In the following prune phase, each rule is incrementally pruned, allowing the pruning of any final sequence of the attributes, until a pruning metric is fulfilled. In the optimization stage each generated rule is further optimized by a) greedily adding attributes to the original rule and b) by independently growing a new rule undergoing a growth and pruning phase, as described above. Finally, in the selection phase, the best rules are kept and the other rules are deleted from the model. RIDOR is also a direct method, first generating a default rule (e.g. "all patients are controls") and then exceptions ("except if rs5 = AB and rs10 = AB then patients are cases") to the default rule with the least error rate. The "best" exceptions for each exception are generated and iterated until pure. Thus, a tree-like expansion of exceptions is generated. The exceptions are a set of rules that predict classes other than the default. PART is an indirect method for rule generation. PART generates a pruned decision tree using the C4.5 statistical classifier [16] in each iteration. From the best tree, the leaves are translated into rules. In the current study, the WEKA open-source software was used to implement the three rule-based classifier methods [17]. In WEKA, the RIPPER algorithm is implemented and named as JRIP (i.e. Java implementation of RIPPER).

Datasets

For the evaluation of the algorithms, three different case-control models (A, B & C) (Table 1) were used for dataset simulation. Model A consisted of four case rules constructed by the interaction of two SNPs, rs5 and rs10. Model B consisted of two case rules constructed by the interaction of rs5 and rs10 and by the interaction between SNP rs15 and the area under the curve (AUC) of a hypothetical compound X, as an environmental factor. For model C cases were assigned randomly without any relationship to a genotype or environmental factor, i.e. a "null model".
Table 1

Case-Control Models

Model AModel BModel C
Rule1:Rule1:Rule1:
If rs5 = AA and rs10 = AB then CaseIf rs5 = BB and rs10 = AA then CaseIf random number > "threshold" then Case
Rule2:Rule2:Rule 2:
If rs5 = AB and rs10 = AA then CaseIf rs15 = AA and AUC >105 then CaseElse Control
Rule3:Rule3:
If rs5 = AB and rs10 = BB then CaseElse Control
Rule4:
If rs5 = BB and rs10 = AB then Case
Rule 5:
Else Control
Case-Control Models The case-control models A, B and C were used to simulate in total 42 various datasets: the number of subjects (300, 600), the number of SNPs (500, 1500, 3000) and the noise of the case-control model with varying false positive (FP) and false negative (FN) rates (5%, 10%, 20%) was varied for models A and B resulting in 18 different datasets for each case-control model (A and B) and 6 different datasets for model C (Table 2). SNPs simulated in the datasets were named with a consecutive dbSNP [18] reference SNP identifier number (rs) from rs1 to rs1200. The reported rs numbers refer to their biological functionality as described in the dbSNP database [18]. For all other SNPs not used in the case-control models, no influence on the phenotype (i.e. case status) was assumed. The genotype frequencies were simulated randomly, uncorrelated and under assumption of Hardy-Weinberg Equilibrium (HWE) [19-21]. For the minor allele frequency (MAF) [22] a uniform distribution was selected ranging from 0 - 0.5 with randomly varying minor alleles. The environmental factor AUC was simulated by a uniform distribution with medians at 110 and 95 resulting in a binomial distribution of the AUC. For the simulation of each of the 42 datasets different random seeds were used. Dataset generation was performed using SAS (SAS Institute Inc., Cary, NC, V 9.1.3). Simulated datasets are available upon request.
Table 2

Datasets investigated

DatasetModel# SNPs#Patients$Ratio Control/CaseFP [%]FN [%]
1A500300255
2A1500300255
3A3000300255
4A500600255
5A1500600255
6A3000600255
7A50030021010
8A150030021010
9A300030021010
10A50060021010
11A150060021010
12A300060021010
13A50030022020
14A150030022020
15A300030022020
16A50060022020
17A150060022020
18A300060022020

19B500150+150255
20B1500150+150255
21B3000150+150255
22B500300+300255
23B1500300+300255
24B3000300+300255
25B500150+15021010
26B1500150+15021010
27B3000150+15021010
28B500300+30021010
29B1500300+30021010
30B3000300+30021010
31B500150+15022020
32B1500150+15022020
33B3000150+15022020
34B500300+30022020
35B1500300+30022020
36B3000300+30022020

37C5003002n.a.n.a.
38C15003002n.a.n.a.
39C30003002n.a.n.a.
40C5006002n.a.n.a.
41C15006002n.a.n.a.
42C30006002n.a.n.a.

$ For case rule A patients are equally distributed for the 4 case rules; for case rule B the first number indicates the number for rule 1, the second number for rule 2; n.a: not applicable.

Datasets investigated $ For case rule A patients are equally distributed for the 4 case rules; for case rule B the first number indicates the number for rule 1, the second number for rule 2; n.a: not applicable.

Data Analysis

Data analyses were performed using WEKA, version 3.7.0. For each algorithm, a varying set of algorithm specific options was applied resulting in 18, 30, and 9 different settings for RIPPER, PART and RIDOR, respectively (Table 3). The settings were chosen based on theoretical evaluations and based on previous experiences with the algorithms on similar datasets. Each of the settings (Table 3) was applied to each of the 42 simulated datasets (Table 2) using WEKA (command line mode). A customized Perl script was used to extract the most important information (e.g. file name, model, statistics, etc.) from the WEKA result files and to summarize the information in a comma separated file. A grading system (A - D) was created (Table 4), to further compare the results at three different levels: a) the Model level, b) the Rules level, and c) the Attribute level.
Table 3

Settings of Algorithm Options

NrRIPPERRIDORPART
1-F 3 -N 2.0 -O 10-F 3 -S 1 -N 2.0 -A-R -B -M 2 -N 3
2-F 3 -N 5.0 -O 10-F 3 -S 1 -N 5.0 -A-R -B -M 5 -N 3
3-F 3 -N 10.0 -O 10-F 3 -S 1 -N 10.0 -A-R -B -M 10 -N 3
4-F 10 -N 2.0 -O 10-F 10 -S 1 -N 2.0 -A-R -B -M 2 -N 10
5-F 10 -N 5.0 -O 10-F 10 -S 1 -N 5.0 -A-R -B -M 5 -N 10
6-F 10 -N 10.0 -O 10-F 10 -S 1 -N 10.0 -A-R -B -M 10 -N 10
7-F 100 -N 2.0 -O 10-F 20 -S 1 -N 2.0 -A-R -B -M 2 -N 100
8-F 100 -N 5.0 -O 10-F 20 -S 1 -N 5.0 -A-R -B -M 5 -N 100
9-F 100 -N 10.0 -O 10-F 20 -S 1 -N 10.0 -A-R -B -M 10 -N 100
10-F 3 -N 2.0 -O 100-R -M 2 -N 3
11-F 3 -N 5.0 -O 100-R -M 5 -N 3
12-F 3 -N 10.0 -O 100-R -M 10 -N 3
13-F 10 -N 2.0 -O 100-R -M 2 -N 10
14-F 10 -N 5.0 -O 100-R -M 5 -N 10
15-F 10 -N 10.0 -O 100-R -M 10 -N 10
16-F 100 -N 2.0 -O 100-R -M 2 -N 100
17-F 100 -N 5.0 -O 100-R -M 5 -N 100
18-F 100 -N 10.0 -O 100-R -M 10 -N 100
19-B -M 2 -C 0.25
20-B -M 2 -C 0.1
21-B -M 5 -C 0.25
22-B -M 5 -C 0.1
23-B -M 10 -C 0.25
24-B -M 10 -C 0.1
25-M 2 -C 0.25
26-M 2 -C 0.1
27-M 5 -C 0.25
28-M 5 -C 0.1
29-M 10 -C 0.25
30-M 10 -C 0.1

RIPPER: F: number of folds for reduced error pruning; N: minimal weights of instances within a split; O: number of optimization runs

RIDOR: F: number of folds for reduced error pruning; S: number of shuffles for randomization; A: Flag set to use the error rate of all the data to select the default class in each step. N: minimal weight of instances within a split.

PART: C: confidence threshold for pruning; M: minimum number of instances per leaf; R: use reduced error pruning; N: number of folds for reduced error pruning; B: Use binary splits for nominal attributes

Table 4

Grading System

GradeModelRulesAttribute
A100% accordanceExample:If rs5 = BB and rs10 = AA then Case; If rs15 = AA and AUC >105 then Case; Else Control100% accordanceExample:If rs5 = BB and rs10 = AA then CaseAll attributes were present (i.e. detected) and ranked as most frequent, i.e. top 2 for model A and top 4 for model B.Example:rs5 Nr. 1; rs10 Nr. 3; rs15 Nr.2; AUC Nr. 4

BOne attribute was missing or an additional attribute was identified by the generated modelExample:If rs5 = BB and rs10 = AA and rs234 = BB then Case; If rs15 = AA and AUC >105 then Case; Else ControlOne attribute was missing or an additional attribute was identified by the generated modelExample:If rs5 = BB and rs10 = AA and rs234 = BB then CaseAll attributes were present but not ranked as most frequentExample:rs5 Nr. 5; rs10 Nr. 3; rs15 Nr.7; AUC Nr. 4

CTwo attributes were different between the generated and the true model,Example:If rs5 = BB and rs10 = AA and rs234 = BB then Case; If rs15 = AA and AUC >105 and rs56 = AA then Case; Else ControlTwo attributes were different between the generated and the true modelExample:If rs5 = BB and rs234 = BB then CaseOne attribute was not present, remaining attributes were present and rank was not consideredExample:rs5 not detected; rs10, rs15 and AUC detected

DThree attributes were different between the generated and the true modelExample:If rs5 = BB and rs234 = BB and rs56 = AA then Case; If rs15 = AA and AUC >105 then Case; Else ControlThree attributes were different between the generated and the true modelExample:If rs5 = BB and rs56 = AA and rs234 = BB then CaseTwo attributes were not present, remaining attributes were present and ranked as most frequentExample:rs5 and AUC not detected; rs10 and rs15 detected

Comparison of the generated versus true model, rules and attributes. (Model B was used for the example.)

Settings of Algorithm Options RIPPER: F: number of folds for reduced error pruning; N: minimal weights of instances within a split; O: number of optimization runs RIDOR: F: number of folds for reduced error pruning; S: number of shuffles for randomization; A: Flag set to use the error rate of all the data to select the default class in each step. N: minimal weight of instances within a split. PART: C: confidence threshold for pruning; M: minimum number of instances per leaf; R: use reduced error pruning; N: number of folds for reduced error pruning; B: Use binary splits for nominal attributes Grading System Comparison of the generated versus true model, rules and attributes. (Model B was used for the example.) On the Model level, each developed model was compared to the true model. At the Rules level, each rule was extracted individually from each developed model and compared to the true case-control rule. At both levels, the "best" grade was always assigned to each model, where A is better than B, B better than C, etc. On the Attribute level, all SNPs and other attributes used as case-control predictors were extracted from the respective models. The appearance of each marker was counted for each dataset and algorithm. The most frequent marker was ranked as 1st, the second most frequent marker was ranked as 2nd, and so forth. Subsequently, the grading system was applied (Table 4). If the attribute ranking fulfilled a grading specification, the dataset and algorithm was assigned the "best" grade, where A is better than B, B better than C, etc. For the evaluation of the "null model" (model C), the same procedure as described above was applied. The true model was assumed to be either model A and/or model B. For the rule and attribute evaluation, the respective rules and attributes from models A and/or B were applied.

Results

In total, 2394 models (= 42 datasets × 57 algorithm options) were generated and analyzed as described in the methods section. A summary statistic on the number of rules per model and the number of attributes per model is presented in Table 5. A qualitative summary of the results is shown in Table 6. Detailed quantitative results at the Model, Rule and Attribute level are presented in Tables 7, 8, &9.
Table 5

Statistics

RIPPERRIDORPART
Model AModel BModel CModel AModel BModel CModel AModel BModel C
#Min131231112
Attributes* per model5th Percentile231454247
Median7415161728242447
95th Percentile281342465070102109144
Max432163656273245240259

# rules$ per modelMin232331111
5th Percentile332442221
Median636889.5769
95th Percentile1171416171925.524.537.5
Max16918202020454346

* unique SNP or unique environmental variable; $ Case and control rules combined

Table 6

Qualitative Results - Summary

RIPPERRIDORPART
Evaluation LevelGradeModel AModel BModel CModel AModel BModel CModel AModel BModel C
ModelsA33% (6)56% (10)0% (0)0% (0)11% (2)0% (0)0% (0)33% (6)0% (0)
B33% (6)50% (9)0% (0)0% (0)6% (1)0% (0)0% (0)33% (6)0% (0)
C22% (4)56% (10)0% (0)0% (0)6% (1)0% (0)0% (0)22% (4)0% (0)
D6% (1)22% (4)0% (0)0% (0)17% (3)0% (0)6% (1)11% (2)0% (0)

RulesA83% (15)100% (18)0% (0)83% (15)94% (17)0% (0)44% (8)78% (14)0% (0)
B89% (16)72% (13)0% (0)94% (17)100% (18)0% (0)50% (9)100% (18)0% (0)
C67% (12)67% (12)0% (0)56% (10)100% (18)0% (0)78% (14)89% (16)0% (0)
D50% (9)56% (10)0% (0)67% (12)100% (18)0% (0)56% (10)94% (17)0% (0)

AttributesA83% (15)56% (10)0% (0)50% (9)72% (13)0% (0)11% (2)0% (0)0% (0)
B17% (3)28% (5)0% (0)50% (9)17% (3)0% (0)61% (11)78% (14)0% (0)
C0% (0)17% (3)0% (0)0% (0)11% (2)0% (0)6% (1)22% (4)0% (0)
D0% (0)0% (0)0% (0)0% (0)0% (0)0% (0)22% (4)0% (0)0% (0)

Summary of the results, separated by algorithm, case-control model, and grading. Number represents the percent frequency of datasets where a respective grade was achieved at least once (absolute number is in brackets).

Table 7

Quantitative Results - Models Level

RIPPERRIDORPART
ModelDatasetSNPsPatientsError [%]ABCDABCDABCD
A15003005017000000000
A215003005000000000002
A330003005000000000000
A45006005310000000000
A515006005750000000000
A630006005122000000000
A750030010000000000000
A8150030010010000000000
A9300030010000000000000
A1050060010801000000000
A11150060010000700000000
A12300060010101000000000
A1350030020000000000000
A14150030020000000000000
A15300030020000000000000
A1650060020630000000000
A17150060020000000000000
A18300060020000000000000

B19500150+15051170020101100
B201500150+15051503001002000
B213000150+1505442000030011
B22500300+30051404000005000
B231500300+30051800000000200
B243000300+30051710000001100
B25500150+150107100000001020
B261500150+15010022000010121
B273000150+15010032100000000
B28500300+300101800010001100
B291500300+300101710000000110
B303000300+300101310000000000
B31500150+15020007100010000
B321500150+15020001000000000
B333000150+15020004400000000
B34500300+30020041000000000
B351500300+30020001000000000
B363000300+30020000700000000

C37500300n.a.000000000000
C381500300n.a.000000000000
C393000300n.a.000000000000
C40500600n.a.000000000000
C411500600n.a.000000000000
C423000600n.a.000000000000

Summary of the results at the model level, separated by algorithm, case-control model, dataset and grading. Number expresses the absolute frequency of the respective grading assignment.

Table 8

Quantitative Results - Rules Level

RIPPERRIDORPART
ModelDatasetSNPsPatientsError [%]ABCDABCDABCD
A150030054414304623121322
A21500300503600416033412
A330003005017001000000
A4500600546140030200025282710
A515006005720001229412432
A6300060052832008270012621
A75003001023274725012320
A8150030010191303212022020
A9300030010111121101330000
A10500600104616001123702520112
A111500600103510156619112011169
A1230006001048520077000031
A135003002011218314360042
A14150030020516901300000
A15300030020301010070000
A1650060020481771861740210
A171500600202518262841540024
A183000600200191316240010

B19500150+15052943081355311792
B201500150+15053060088751213193
B213000150+15052751387511742188
B22500300+3005360008359341411510
B231500300+3005360009211252027114
B243000300+30053510091910101322258
B25500150+150102511114417201220141
B261500150+1501016519919147171333
B273000150+150101841114014390511
B28500300+30010360001789828181410
B291500300+300103510071416672774
B303000300+300103172079129217147
B31500150+15020180131043950132
B321500150+15020108211128560100
B333000150+150209718212790101
B34500300+3002018173134518132161
B351500300+300201809314312132472
B363000300+30020162232053581311

C37500300n.a.000000000000
C381500300n.a.000000000000
C393000300n.a.000000000000
C40500600n.a.000000000000
C411500600n.a.000000000000
C423000600n.a.000000000000

Summary of the results at the rules level, separated by algorithm, case-control model, dataset and grading. Number expresses the absolute frequency of the respective grading assignment.

Table 9

Quantitative Results - Attributes Level

RIPPERRIDORPART
ModelDatasetSNPsPatientsError [%]ABCDABCDABCD
A15003005100001000100
A215003005100010000100
A330003005010001000000
A45006005100010001001
A515006005100010000100
A630006005100010000100
A750030010100001000100
A8150030010100010000100
A9300030010100001000001
A1050060010100010001000
A11150060010100010000100
A12300060010100001000100
A1350030020100001000100
A14150030020010001000001
A15300030020010001000001
A1650060020100010000100
A17150060020100010000100
A18300060020100001000010

B19500150+1505100010000100
B201500150+1505100010000100
B213000150+1505100010000100
B22500300+3005100010000100
B231500300+3005100010000100
B243000300+3005100001000100
B25500150+15010100000100100
B261500150+15010010010000100
B273000150+15010001000100100
B28500300+30010100010000100
B291500300+30010100010000100
B303000300+30010100010000100
B31500150+15020010010000010
B321500150+15020001001000010
B333000150+15020001001000010
B34500300+30020010010000100
B351500300+30020010010000100
B363000300+30020010010000010

C37500300n.a.000000000000
C381500300n.a.000000000000
C393000300n.a.000000000000
C40500600n.a.000000000000
C411500600n.a.000000000000
C423000600n.a.000000000000

Summary of the results at the attribute level, separated by algorithm, case-control model, dataset and grading. A "1" reflects an affiliation to the respective grading, whereas a "0" reflects no affiliation.

Statistics * unique SNP or unique environmental variable; $ Case and control rules combined Qualitative Results - Summary Summary of the results, separated by algorithm, case-control model, and grading. Number represents the percent frequency of datasets where a respective grade was achieved at least once (absolute number is in brackets). Quantitative Results - Models Level Summary of the results at the model level, separated by algorithm, case-control model, dataset and grading. Number expresses the absolute frequency of the respective grading assignment. Quantitative Results - Rules Level Summary of the results at the rules level, separated by algorithm, case-control model, dataset and grading. Number expresses the absolute frequency of the respective grading assignment.

Models Level Analyses

The models generated had in median 4 to 47 unique attributes per model (Table 5). The RIPPER algorithm utilized the fewest number of attributes (median: 4 to 15) whereas RIDOR and PART had a significantly higher number of attributes per model (median: 16 to 47). The number of attributes was approximately two times higher for the "null model" (model C) compared to models A and B. The number of rules per model was comparable between all three algorithms (median: 3 to 9.5). The RIPPER algorithm performed well and was able to discover the true model (Grade A) at least once, 33% of the time with case-control model A and 56% of the time with case-control model B (Table 6). The RIDOR and PART algorithm performed poorly in model detection. Case-control model A was not discovered in any dataset by the two algorithms. Case-control model B was identified at least once in 11% and 33% of the datasets by RIDOR and PART, respectively. When minor deviations from the true model were considered (Grades B - D), RIPPER was again able to detect significantly more models compared to the RIDOR and PART algorithms. Overall, case-control model B was better discovered by all three algorithms, compared to case-control model A. None of the methods detected for model C a false positive finding. Table 7 provides a detailed summary of model level algorithm performance. In general, algorithms performed worse if the power of the dataset decreased, i.e. less subjects and/or more SNPs and/or higher FP/FN rates.

Rules Level Analyses

At the rules level, the RIPPER algorithm performed well and was able to discover the true case rules (Grade A) at least once, 83% of the time with case-control model A and 100% of the time with case-control model B (Table 6). The RIDOR and PART algorithms also performed well in rule detection. For case-control model A, the true rules were discovered at least once 83% and 44% of the time with the RIDOR and PART algorithms, respectively. With case-control model B the RIDOR and PART algorithms were able to discover the true rule at least once in 94% and 78% of the datasets, respectively. When deviations (Grades B - D) from the true case rules were allowed, all algorithms performed with high discovery rates. Overall the case rules of case-control model B were slightly better discovered by all three algorithms, compared to the case rules of case-control model A. None of the methods detected for the rules of model C a false positive finding. Table 8 provides a detailed summary of rule level algorithm performance. All algorithms performed worse in rule detection if the power of the dataset decreased, i.e. less subjects and/or more SNPs and/or higher FP/FN rates.

Attributes Level Analyses

At the attributes level, the RIPPER algorithm performed well. RIPPER yielded an "A" grade (all attributes present and ranked as the top 2 (model A) or the top 4 (model B) attributes) in 83% of the time with case-control model A and 56% of the time with case-control model B (Table 6). The RIDOR algorithm demonstrated satisfactory performance at the attribute level with 50% and 72% of the datasets yielding an "A" grade in case-control models A and B, respectively. The PART algorithm did not perform as well. With case-control model A, the grade "A" designation occurred only 11% of the time and with case-control model B, the "A" grade did not get assigned to any of the outputs. For model C, all grading (A-D) were zero across all methods. Table 9 provides a detailed summary of attribute level algorithm performance. All algorithms performed worse in attribute detection, if the power of the dataset decreased, i.e. less subjects and/or more SNPs and/or higher FP/FN rates. Quantitative Results - Attributes Level Summary of the results at the attribute level, separated by algorithm, case-control model, dataset and grading. A "1" reflects an affiliation to the respective grading, whereas a "0" reflects no affiliation.

Discussion

At the model level, the RIPPER algorithm outperformed the other two evaluated algorithms. RIPPER discovered the true model (grade "A") or a slight variation of the true model (grade "B") >40% of the time with case-control model A and >70% of the time with case-control model B. The RIDOR and PART algorithms were not able to detect the true case-control model A. Even under less stringent conditions of evaluation (Grade B), allowing minor deviations from the true model, the two algorithms were not able to discover the true case-control model. For case-control model B the RIDOR and PART algorithms performed better, however, RIPPER still outperformed the two. In general, the RIDOR and PART algorithms tended to build overly complex prediction models with a median of 2 to 3 times more predictors compared to the RIPPER algorithm. This complexity was probably caused by the nature of the algorithm. The PART algorithm derives rules from decision trees and decision trees tend to build overly complex models [12]. With the RIDOR algorithm, the "exception from the exception" principle is utilized for data investigation, this may not be appropriate to analyze complex gene-gene or gene-environment interactions. All three algorithms performed well in detecting case-control rules. As with model level analyses, the RIPPER algorithm was superior to the RIDOR and PART algorithms in rule level analyses. The RIPPER algorithm discovered rules at a much higher frequency, even when corrected for the number of options tested (Table 3). If the frequency of each rule is counted and the rules ordered according to their frequencies (analogous to the attribute level analyses), the RIPPER algorithm identified more top-ranked rules than the other two algorithms. At the attribute level, the RIPPER algorithm performed slightly better than RIDOR. The RIPPER algorithm discovered and top-ranked the true attributes 83% and 56% of the time for case-control models A and B, respectively. In contrast, the RIDOR algorithm discovered and top-ranked the true attributes 50% and 72% of the time for case-control models A and B, respectively. The PART algorithm performed the worst of the three algorithms but still identified the true attributes 72% and 78% of the time for case-control models A and B, respectively. The PART algorithm did not rank the true attributes in the top 2 (model A) or 4 (model B) as consistently as the RIPPER and RIDOR algorithms. In addition to the two case-control models A and B datasets without relationships between genotypes and phenotypes (model C, "null model") were also tested. The null model results were not useful in further discriminating between the three rule based classifiers. For all datasets and with all methods, no false positive number of models, attributes or rules were identified. Nevertheless, the models derived from the "null model" datasets showed that the median number of attributes per model was twice as high compared to the real case-control models A and B. The number of attributes per model might therefore be a mechanism to discriminate true models from false positive models. The RIPPER algorithm appears to be superior to the PART and RIDOR algorithm at all levels: models, rules and attributes. The tested set of algorithm options (Table 3) provides an adequate toolset for comparison; however, this work could be expanded and further optimized using simulated and non-simulated data. Based on the analysis presented, questions may arise on how to best translate the rule based classifier into practice and how to apply the classifiers to non-simulated data. In data analysis practice, one would apply a selected classifier, preferably RIPPER, to one dataset with a battery of options, e.g. as provided in Table 3. Thus, for RIPPER this would result in 18 different models, one for each setting. For each of the 18 models standard statistics, e.g. numbers of rules, sensitivity, specificity, accuracy, etc., are provided. The next steps would be mainly triggered by the purpose of the analysis, depending on whether the analyst aims to use the classifiers as model builders or as filters. If the classifier is to be used for model building, the revealed models should be further investigated, e.g. by thorough review considering statistics such as sensitivity, specificity and complexity of the models. If classifiers are to be used as filters for attributes, similar procedures as described in the methods section could be applied. Additional research is required regarding practical considerations and evaluation methods, such as cross validation or external prediction. In the presented analyses, the size of the datasets tested was limited to 3000 SNPs and 600 patients. Technically, rule based classifiers are neither limited by number of SNPs nor by number of patients. Thus, the methods should be scalable to whole genome levels. However, the size of the dataset analyzed in WEKA may be limited by the available computational memory and has to be taken into account.

Conclusions

The current analyses substantiate the utility of rule based classifiers such as RIPPER, RIDOR and PART for the detection of gene-gene and gene-environment interactions in genetic association studies. These methods could provide a valuable new method, complementing existing approaches, in the analysis of genetic association studies. The methods provide an advantage in being able to handle both categorical and continuous variable types, and since the outputs of the analyses are easy to interpret the rule based classifier approach could quickly generate testable hypotheses for further evaluation. In addition, since the algorithms are computationally inexpensive to run, they may serve as valuable tools for preselection of attributes to be used in more complex, computationally intensive approaches such as MDR. Whether used in isolation or in conjunction with other tools, rule based classifiers are an important addition to the armamentarium of tools available for analyses of complex genetic association studies. As a next step, the most promising algorithm RIPPER should be benchmarked against other popular analysis methods, such as MDR or Random Forests.

Competing interests

TL, DZ, JY, SJ and MDR have no competing interests.

Authors' contributions

TL designed the analysis, established the algorithms, performed the analysis and drafted the manuscript. JY contributed to the design of the simulated data and the scientific framing of queries to be performed. DZ developed the Perl summary scripts and contributed to the design of the analysis. SJ contributed to the establishment of the algorithms and the manuscript. MDR participated in the design and coordination of the analysis and helped to draft the manuscript. All authors read and approved the final manuscript.
  16 in total

1.  dbSNP: the NCBI database of genetic variation.

Authors:  S T Sherry; M H Ward; M Kholodov; J Baker; L Phan; E M Smigielski; K Sirotkin
Journal:  Nucleic Acids Res       Date:  2001-01-01       Impact factor: 16.971

Review 2.  Genetic association studies: design, analysis and interpretation.

Authors:  Cathryn M Lewis
Journal:  Brief Bioinform       Date:  2002-06       Impact factor: 11.622

Review 3.  Genetics, statistics and human disease: analytical retooling for complexity.

Authors:  Tricia A Thornton-Wells; Jason H Moore; Jonathan L Haines
Journal:  Trends Genet       Date:  2004-12       Impact factor: 11.639

4.  Genetic association studies.

Authors:  Heather J Cordell; David G Clayton
Journal:  Lancet       Date:  2005 Sep 24-30       Impact factor: 79.321

5.  A generalized combinatorial approach for detecting gene-by-gene and gene-by-environment interactions with application to nicotine dependence.

Authors:  Xiang-Yang Lou; Guo-Bo Chen; Lei Yan; Jennie Z Ma; Jun Zhu; Robert C Elston; Ming D Li
Journal:  Am J Hum Genet       Date:  2007-04-25       Impact factor: 11.025

6.  Analysis of gene-gene interactions.

Authors:  Jason H Moore
Journal:  Curr Protoc Hum Genet       Date:  2008-10

7.  An exact test for Hardy-Weinberg and multiple alleles.

Authors:  E J Louis; E R Dempster
Journal:  Biometrics       Date:  1987-12       Impact factor: 2.571

8.  The applicability of the Hardy-Weinberg principle in the study of populations.

Authors:  T Schaap
Journal:  Ann Hum Genet       Date:  1980-10       Impact factor: 1.670

Review 9.  A comprehensive review of genetic association studies.

Authors:  Joel N Hirschhorn; Kirk Lohmueller; Edward Byrne; Kurt Hirschhorn
Journal:  Genet Med       Date:  2002 Mar-Apr       Impact factor: 8.822

10.  Optimization of neural network architecture using genetic programming improves detection and modeling of gene-gene interactions in studies of human diseases.

Authors:  Marylyn D Ritchie; Bill C White; Joel S Parker; Lance W Hahn; Jason H Moore
Journal:  BMC Bioinformatics       Date:  2003-07-07       Impact factor: 3.169

View more
  4 in total

Review 1.  Challenges and opportunities in genome-wide environmental interaction (GWEI) studies.

Authors:  Hugues Aschard; Sharon Lutz; Bärbel Maus; Eric J Duell; Tasha E Fingerlin; Nilanjan Chatterjee; Peter Kraft; Kristel Van Steen
Journal:  Hum Genet       Date:  2012-07-04       Impact factor: 4.132

2.  CAMUR: Knowledge extraction from RNA-seq cancer data through equivalent classification rules.

Authors:  Valerio Cestarelli; Giulia Fiscon; Giovanni Felici; Paola Bertolazzi; Emanuel Weitschek
Journal:  Bioinformatics       Date:  2015-10-30       Impact factor: 6.937

3.  LAF: Logic Alignment Free and its application to bacterial genomes classification.

Authors:  Emanuel Weitschek; Fabio Cunial; Giovanni Felici
Journal:  BioData Min       Date:  2015-12-08       Impact factor: 2.522

4.  Supervised DNA Barcodes species classification: analysis, comparisons and results.

Authors:  Emanuel Weitschek; Giulia Fiscon; Giovanni Felici
Journal:  BioData Min       Date:  2014-04-11       Impact factor: 2.522

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.