| Literature DB >> 25870758 |
Benjamin Goudey1, Mani Abedini2, John L Hopper3, Michael Inouye4, Enes Makalic3, Daniel F Schmidt3, John Wagner5, Zeyu Zhou6, Justin Zobel7, Matthias Reumann1.
Abstract
Genome-wide association studies (GWAS) are a common approach for systematic discovery of single nucleotide polymorphisms (SNPs) which are associated with a given disease. Univariate analysis approaches commonly employed may miss important SNP associations that only appear through multivariate analysis in complex diseases. However, multivariate SNP analysis is currently limited by its inherent computational complexity. In this work, we present a computational framework that harnesses supercomputers. Based on our results, we estimate a three-way interaction analysis on 1.1 million SNP GWAS data requiring over 5.8 years on the full "Avoca" IBM Blue Gene/Q installation at the Victorian Life Sciences Computation Initiative. This is hundreds of times faster than estimates for other CPU based methods and four times faster than runtimes estimated for GPU methods, indicating how the improvement in the level of hardware applied to interaction analysis may alter the types of analysis that can be performed. Furthermore, the same analysis would take under 3 months on the currently largest IBM Blue Gene/Q supercomputer "Sequoia" at the Lawrence Livermore National Laboratory assuming linear scaling is maintained as our results suggest. Given that the implementation used in this study can be further optimised, this runtime means it is becoming feasible to carry out exhaustive analysis of higher order interaction studies on large modern GWAS.Entities:
Year: 2015 PMID: 25870758 PMCID: PMC4383059 DOI: 10.1186/2047-2501-3-S1-S3
Source DB: PubMed Journal: Health Inf Sci Syst ISSN: 2047-2501
Example contingency table.
| Genotype Frequencies | |||||
|---|---|---|---|---|---|
| Controls | ... | ||||
| Cases | ... | ||||
| Col. Counts | ... | ||||
2 × V-contingency table summarising the occurrence of genotype combinations for an arbitrary SNP interaction in a case-control GWAS study.
Figure 1Binary genotype representation. Example showing a) the conversion of a given SNP into the binary representation, b) computing the occurrence of a single genotype combination for a pairs of SNPs by taking their logical AND and counting the number of set bits in the resulting binary vector.
Figure 2Illustration of data decomposition and load balancing. Decomposition strategy. For any given SNP interaction study, the entire calculation is divided into equal-sized partitions. For each partition one MPI task is executed on an assigned CPU card. The further decomposition into small sub-tasks are handle by OpenMP dynamic scheduler.
Hardware, dataset size and previously reported runtimes for a variety of CPU- and GPU-based pairwise SNP analysis methods.
| Method | Hardware | Cores | Clock (Ghz) | SNPs (1000 s) | Samples (1000 s) | Time (min) | Estimated 3-way runtime | |
|---|---|---|---|---|---|---|---|---|
| Unscaled | Scaled | |||||||
| CPU-based | ||||||||
| Present method | PowerPC A2 | 262,144 | 1.60 | 1,100 | 2 | 8 | 8 | 5.8 years |
| Wang | Intel X3430 | 172 | 2.40 | 500 | 2 | 538 | 3 | >1000 years |
| Ma | Intel X5355 | 528 | 2.66 | 50 | 2 | 1158 | 1877 | >1000 years |
| GPU-based | ||||||||
| Goudey | Nvidia GTX470 | 448 | 1.22 | 450 | 5 | 13 | 31 | 21.7 years |
| GBOOST | Nvidia GTX285 | 240 | 1.48 | 351 | 5 | 80 | 220 | 219 years |
| Kam-Thong | Nvidia GTX295 | 960 | 1.24 | 4 | 10 | 2 | 19664 | >1000 years |
Hardware, dataset size and previously reported runtimes for a variety of CPU- and GPU-based pairwise SNP analysis methods. For each method, we show the make and model of processor or GPU card, the processor speed and the number of hardware cores. Runtime is additionally shown as originally reported as well as scaled to a dataset with 1.1 million SNPs, 2000 samples and using either the full Avoca system (CPU methods) or using the Nvidia GTX295 (GPU methods). In the final column we show the estimated time to processes all 3-way interactions for a 1.1 million SNP, 2000 sample dataset using the originally reported hardware and assuming perfectly linear scaling.
Figure 3Run time, scaling and efficiency analysis for strong scaling simulations. Total run times, scaling and efficiency (a., b. and c. respectively) as the number of hardware threads is increased for a 1.1 million SNP, 2000 sample dataset.
Figure 4Run times and scaling for varying size GWAS datasets on 64 and 1024 computing nodes. Runtime and scaling (a. and b. respectively) as the number of SNPs increases, using either 1024 or 64 nodes respectively. Subsets of the simulated data at increasing powers of 10 (103 - 106) are used. The scaling factor in subplot b. indicate the decrease in runtime as the number of pairs is reduced by a factor of 100.