| Literature DB >> 27307608 |
Haoyang Zeng1, Matthew D Edwards1, Ge Liu1, David K Gifford1.
Abstract
MOTIVATION: Convolutional neural networks (CNN) have outperformed conventional methods in modeling the sequence specificity of DNA-protein binding. Yet inappropriate CNN architectures can yield poorer performance than simpler models. Thus an in-depth understanding of how to match CNN architecture to a given task is needed to fully harness the power of CNNs for computational biology applications.Entities:
Mesh:
Substances:
Year: 2016 PMID: 27307608 PMCID: PMC4908339 DOI: 10.1093/bioinformatics/btw255
Source DB: PubMed Journal: Bioinformatics ISSN: 1367-4803 Impact factor: 6.937
Fig. 1.The basic architectural structure of the tested convolutional neural networks
The code and brief description of the 9 variants of CNN models compared in this work
| Code | Architecture (relative to 1layer structure) |
|---|---|
| 1layer | The basic structure as depicted in |
| 1layer_1motif | Use 1 convolutional kernels |
| 1layer_64motif | Use 64 convolutional kernels |
| 1layer_128motif | Use 128 convolutional kernels |
| 1layer_local_win9 | Use local maxpooling of window size 9 at top |
| 1layer_local_win3 | Use local maxpooling of window size 3 at top |
| 2layer | 2 layers with 16/32 kernels |
| 3layer | 3 layers with 16/32/64 kernels |
| 2layer_local_win3 | 2 layers with 16/32 kernels, use local maxpooling of window size 3 at top |
| 3layer_local_win3 | 3 layers with 16/32/64 kernels, use local maxpooling of window size 3 at top |
More detailed description of the models can be found on our supporting website.
Fig. 2.(A) The distribution of AUCs across 690 experiments in the motif discovery task. (B) The performance of our basic model (1layer) matches DeepBind. Blue points are the transcription factors with AUCs close to 0.5 for DeepBind but not for our basic model
Fig. 3.The distribution of AUCs across 690 experiments in the motif occupancy task
Fig. 4.(A) The effect of different architectures on AUC is experiment-specific. Each column of the heat map denotes one ChIP-seq experiment. Each row represents one model variant. Each value in the heatmap shows the change of AUC in motif discovery task from the basic model 1layer to a variant model. Four clusters from hierarchical clustering are colored in blue, green, red and violet. (B) The number of called ChIP-seq peaks in each of the fours clusters. The colors of the clusters match those in (A). (C) The distribution of AUCs in the motif discovery task across 112 ChIP-seq experiment with 40000 peaks
The training time for different model variants to train on 500 000 samples
| Model | Time for training on 500 000 samples (s) |
|---|---|
| 1layer | 64.34 |
| 1layer_64motif | 79.45 |
| 1layer_128motif | 94.47 |
| 1layer_local_win9 | 68.08 |
| 1layer_local_win3 | 73.12 |
| 2layer | 91.82 |
| 3layer | 124.12 |
| 2layer_local_win3 | 93.34 |
| 3layer_local_win3 | 125.5 |
Hyper-parameter sets tested
| Hyper-parameter | Choices |
|---|---|
| Dropout ratio | 0.75, 0.5, 0.1 |
| Momentum in AdaDelta optimizer | 0.9, 0.99, 0.999 |
| Delta in AdaDelta optimizer | 1e-4, 1e-6, 1e-8 |