Literature DB >> 18586730

Optimal pooling for genome re-sequencing with ultra-high-throughput short-read technologies.

Iman Hajirasouliha1, Fereydoun Hormozdiari, S Cenk Sahinalp, Inanc Birol.   

Abstract

New generation sequencing technologies offer unique opportunities and challenges for re-sequencing studies. In this article, we focus on re-sequencing experiments using the Solexa technology, based on bacterial artificial chromosome (BAC) clones, and address an experimental design problem. In these specific experiments, approximate coordinates of the BACs on a reference genome are known, and fine-scale differences between the BAC sequences and the reference are of interest. The high-throughput characteristics of the sequencing technology makes it possible to multiplex BAC sequencing experiments by pooling BACs for a cost-effective operation. However, the way BACs are pooled in such re-sequencing experiments has an effect on the downstream analysis of the generated data, mostly due to subsequences common to multiple BACs. The experimental design strategy we develop in this article offers combinatorial solutions based on approximation algorithms for the well-known max n-cut problem and the related max n-section problem on hypergraphs. Our algorithms, when applied to a number of sample cases give more than a 2-fold performance improvement over random partitioning.

Entities:  

Mesh:

Year:  2008        PMID: 18586730      PMCID: PMC2718651          DOI: 10.1093/bioinformatics/btn173

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.937


1 INTRODUCTION

After decades of research effort, the cost of sequencing an individual human genome via Sanger sequencing (Sanger et al., 1977, 1980, 1982) has now been reduced to the order of 10-million dollars for a 10-fold coverage and require 10 000 instrument days (1 year with 30 instruments) (Solexa web server, 2008). In order to perform fast and cost effective genome sequence comparisons for many biological and medical applications, the cost needs to be further reduced to the order of a few thousand dollars (i.e. a few dollars per megabase) and the time frame of sequencing must be reduced to a few days per instrument. For example, the Solexa sequencing-by-synthesis technology of Illumina (Bennett, 2004) offers about 100-fold improvements over Sanger sequencing in both cost and throughput. Similarly, the pyrosequencing technology of 454 Life Sciences (Margulies et al., 2005) promises massive parallelization of the sequencing process by the use of microchip sensors, improving the speed of Sanger sequencing by a factor of a few hundred. In addition to these, there are other commercial and pre-commercial products from Applied Biosystems, Helicos Biosciences and Visigen Biotechnology, among others that perform either clonal cluster sequencing or single molecule sequencing. Unfortunately, the massive increase in the throughput offered by the above technologies comes with a shortened read length, and shorter the read length, the more problematic it is to work with a genome that has many longer repeats. While Sanger sequencing offers 500–1000 bps per read, the read lengths of new technologies range from ∼25 (e.g. Solexa) to few hundred base pairs (e.g. 454) and as such, they are more suitable for re-sequencing studies. Although current developments in extending these technologies to produce longer, as well as paired end reads puts the de novo sequencing studies in the realm of the possible, for the rest of this report we concentrate on the problem of resequencing of individual genomes with short single end reads, and focus on the Solexa technology. The type of re-sequencing problem we consider in this work is based on bacterial artificial chromosome (BAC) libraries. Benefits of using BACs in re-sequencing studies is 2-fold. (1) It is possible to cluster the short reads obtained from each BAC into small local sets that represent the sequence of one BAC where most of the repeat sequences have a unique copy. (2) One can a priori determine the genomic neighbourhood the BAC is coming from; thus having the reference sequence provides an essential backbone. This can be achieved, for instance, by using the results of fingerprinting or BAC end sequencing experiments. Although on the face of it, a BAC-library-based sequencing effort defeats the benefits of massive parallelization offered by new sequencing technologies, it enables directed sequencing studies possible, where the interest is not on whole genome (re)sequencing, but on investigating a region of interest. Furthermore, recent developments bring the throughput in fingerprinting technology on par with new sequencing technologies (Mathewson et al., 2007). Solexa sequencing experiments are run on a flow cell with eight lanes, each yielding in the order of 108 bps of sequence per run. A typical BAC we consider has a length ranging between 150 Kbps and 250 Kbps. Thus, if we sequence one BAC on each lane, a single run would produce about a 1000-fold coverage per BAC, which is far beyond necessary in a re-sequencing study. Therefore, in order to maximize the throughput of the Solexa technology, hence minimize its cost, it is of key importance to utilize each lane in a more sensible way, such as by sequencing more than one BAC per lane. However, sequencing multiple BACs per lane introduces major difficulties due to repeat sequences that are present in two or more BACs, as they would cause tanglement in their assemblies. In order to minimize the repeat elements that are present in multiple BACs, novel algorithmic techniques must be developed. Available algorithms for DNA fragment assembly (Pevzner et al., 2001a,b; Roach et al., 1995), especially those for short reads (Chaisson et al., 2004; Chaisson and Pevzner, 2008; Warren et al., 2006) all suffer from the presence of repeats within the genome region to be assembled (Margulies et al., 2005). However, the high potential of high-throughput short-read technologies have promoted the development of novel protocols and algorithms such as the short read assembly protocol (SHARP; Sundquist et al., 2007) that aim to address the shortcomings of short-read technologies. Our goal in this article is to help the available fragment assembly methods, especially in the context of genome re-sequencing, by providing (near) optimal utilization of the multiple lanes available by the Solexa technology, while keeping the cost of fragment assembly at a minimum. For this purpose, we present algorithms that partition the input BACs into multiple lanes in a way to minimize potential assembly errors due to repeat sequences shared among BACs in each lane. Our algorithms are quite efficient in practice and provide significant improvement to the cost of fragment assembly over random partitioning strategies.

2 PROBLEM DEFINITION AND GENERAL ALGORITHMIC APPROACH

Given a set of BACs sampling in some known specific genomic regions, and the reference sequence, our ultimate goal is to construct the sequence of each BAC using the results of optimally designed Solexa sequencing experiments. Consider a set of m BACs sequenced with a read length of k (typically 25–50 bps); the problem we address in this work is, how can we partition this set into n groups (or pools) of approximately h=m/n BACs, such that individual reads can be attributed to the BACs they come from as correctly as possible. In other words, how can we minimize number of shared k-mers by multiple BACs in each pool in the overall configuration? Strictly speaking, the problem does not have a solution, because before conducting any sequencing experiment, it is not possible to know how many shared k-mers the BACs in question would have. However, note that our focus is on re-sequencing studies, and if we have even a crude idea on the genomic coordinates of the BACs, we can approximate that missing information by using the reference sequence. One other hurdle in designing a globally optimal experiment is the rapid proliferation of number of possible configurations. For instance, if we would like to pool m=150 BACs into groups of h=10, we would need to consider configurations in an exhaustive search. Since it is infeasible to search all these configurations for finding the global optimum, we propose an algorithmic approach to guide us to an ‘approximately’ optimal setup. Once again we note that our goal is to partition a given set of m BACs into pools of size h each, with the purpose of minimizing the number of potential assembly errors due to sequences that repeat in multiple BACs within a pool. Potential assembly errors due to sequences that repeat within a single BAC, on the other hand, are beyond the scope of this study and are not investigated here. We define our problem more formally as follows.

2.1 Pooling problem

Given a set of m BACs that are to be placed into n pools of approximately h=m/n BACs in each, let C be the number of BACs in a pool P that share a particular k-mer b. If we denote the cost of b as f(C), we can write an overall cost function for a given configuration as and the problem becomes one of selecting the optimum partitioning that minimizes the cost J. One can attribute alternative costs for shared k-mers b, two of which are For the remainder of this article, we will restrict our attention to these two formulations for reasons explained below. ; f(C)=C−1. The pooling problem under the cost function is a minimization problem for the number of k-mers that are shared between pairs of BACs which are in the same pool. This can be reduced to a well-known combinatorial problem called, n-clustering problem, as follows: construct a complete graph G where each BAC B is represented with a unique vertex v and given any pair of BACs B and B′, set the weight of the edge (V,V′) to the number of common k-mers in B and B′. The n-clustering problem is a problem of partitioning G into vertex sets, such that the sum of edge weights between vertices that belong to the same partition is minimized. Unfortunately even obtaining a constant factor approximation to the n-clustering problem is NP-hard (Sahni and Gonzalez, 1976). Thus in Section 3.1, we first reduce this problem to another combinatorial problem known as the max n-cut. Although this problem is also NP-hard (Sahni and Gonzalez, 1976), we solve it by the use of a simple local search procedure within an approximation factor of 1−1/n. For n=15, this implies an approximation factor of 0.93. Although this approximate solution to the max n-cut problem does not provide a guarantee on the approximation for the pooling problem, it gives good results in practice. An extension to the pooling problem is the balanced pooling problem, where we seek to minimize the cost of partitioning BACs into pools, such that the number of BACs in each pool is exactly h=m/n. As can be expected, even approximating the balanced pooling problem within a constant factor is NP-hard. Thus, in Section 3.2 we reduce the balanced pooling problem to the max n-section problem, which is the balanced version of the max n-cut problem. We again describe an algorithm to approximately solve this problem within a factor of 1−1/n. Although the latter algorithm does not provide a guarantee on the approximation factor it obtains for the balanced pooling problem, it yields good results in practice once again. The pooling problem under the second cost function f(C)=C−1 is a minimization problem for the number of BACs within a pool that shares each k-mer, summed over all k-mers. This is a generalized version of the pooling problem with the first cost function as will be explained below. The pooling problem under the second cost function can be reduced to a hypergraph partitioning problem as follows. Let G be a hypergraph where each BAC B is represented with a unique vertex v and each subset S of at most d vertices are connected with a hyperedge e. In the most general case of the problem d=m. The weight of e, namely w(e) is the number of k-mers that occur in all BACs in S and occur in no other BACs. Consider a partition of G into non-overlapping vertex sets. For a given subset S of vertices, let #(S) be the number of pools that have at least one vertex of S. Then the cost of e with respect to this partitioning is w(e)·(|S|−#(S)); here |S| denotes the number of BACs in set S. Our hypergraph partitioning problem defines a search for partitioning G into vertex sets/pools so as to minimize the total cost of the hyperedges with respect to this partition. Unfortunately, the above hypergraph partitioning problem requires space to just represent all the hyperedges. As this represents faster than exponential growth with the number of BACs, even setting up an instance of the problem on a computer is not feasible for the parameter values we are interested in. Notice that if we restrict the maximum number of vertices that can be incident to a hyperedge, d, to 2 (rather than m) then our hypergraph partition problem reduces to the n-clustering problem and thus to the pooling problem with the first cost function. Now we can consider versions of the hypergraph partition problem with d=3, 4, … as ‘approximations’ to our general hypergraph partitioning problem with d=m. Our hypergraph partitioning problem with d>1 is NP-hard (Sahni and Gonzalez, 1976) even to approximate within a constant factor. We reduce it to another hypergraph partitioning problem in which the cost of a hyperedge e with respect to a partitioning is w(e)· (#(S)−1) and the goal is to maximize the total cost of all hyperedges. In this article we show how to solve this problem approximately within a factor of 1−d/2n. For d=3 and n=15, our algorithm provides a 0.9 approximation factor, again sufficiently close to 1. The algorithm employs a greedy approach and is quite efficient. We also consider a balanced version of this hypergraph partitioning problem which asks for maximizing the total cost of all hyperedges with respect to a partition, provided that the number of vertices per each pool is exactly h=m/n. Again we provide a (1−d/2n)-approximation algorithm to this problem. This algorithm is quite involved, as further described in the next section, employing a solution to the minimum weighted bipartite matching towards a greedy selection of the vertices in each partition.

3 METHODS

In this section we give detailed descriptions of the approximation algorithms we use for solving the pooling problem, both balanced and unbalanced versions, under the two cost functions we presented earlier.

3.1 The pooling problem under

The pooling problem (unbalanced version) under our first cost function can be formulated as the well-known max n-cut problem as follows. Input: A weighted undirected graph G(V,w), with the vertex set V representing the set ℬ={B1,B2,…,B150} of BACs, and the edge weights w. For any vertex pair v,v′, w(v,v′) is the number of common k-mers in the corresponding BACs, B and B′. Output: A partitioning of V into pools P={P1,P2,···,P}, , which maximizes the following objective function:

3.1.1 A local search algorithm for max n-cut (LSMnC)

Randomly partition the vertex set V of the graph G into n different pools. If there exists a vertex v∈V such that it is assigned to pool P (v∈P) and there exists a pool P such that ∑w(v,u)≥∑w(v,x), then move vertex v from the pool P into the pool P. Repeat second step until no change can occur. The above simple local search algorithm, when applied to the general max n-cut problem may take too much time before it terminates. However, for our specific problem, the running time of the above algorithm is guaranteed to be polynomial with m and the maximum length of a BAC as shown below.

PROOF.

Let t be the total weight of the edges of the graph G, which is polynomial with m and the maximum number of k-mers in a BAC. It is clear that in each step of the local search algorithm, the total weight of edges going between pools increases by at least one. Therefore in the worst case, this algorithm terminates after t steps.▪ In practice, the running time of the above local search algorithm is in second order for m=150, n=15 and maximum BAC length of 250K. The approximation factor achieved by the above algorithm is 1−1/n as shown below. Consider an arbitrary vertex v∈V and assume P is the cluster containing v after the termination of the local search. We have: The expression in left-hand side of last equation represents the total weight of all edges in the ‘cut’ incident to vertex v. Also, the expression in the right-hand side of the same equation represents the total weight of all edges incident to vertex v. Since v has been chosen arbitrary from the vertex set V, we have: According to the above inequality, the total weight of the edges which are incident to a pair of vertices that do not belong to the same pool is at least (n−1)/n times the total weight of the edges in G, and thus the local search provides a 1−1/n approximation. ▪

3.2 The balanced pooling problem under

The balanced pooling problem asks to partition m BACs into n pools so as to minimize the above cost function, with the additional constraint that the number of BACs per each pool is exactly h=m/n. This is known as max n-section problem for which a local search algorithm by (Gaur et al., 2007) guarantees an approximation factor of 1−1/n. For each vertex u∈V and for each set of vertices belonging to a pool P, let w(u,P)=∑w(u,v). The local search algorithm for the max n-section problem works as follows.

3.2.1 Local search algorithm for max n-section (LSMnS)

Initialization. Partition the vertices V into n pools P1,P2, …,P uniformly at random such that |P1|≤|P2|≤ …≤|P|≤|P1|+1. Iterative step. Find a pair of vertices v∈P and u∈P (i≠j), such that w(v,P−v)+w(u,P−u)≥w(v,P−u)+w(u,P−v). If such a pair exists move u to the cluster P and v to the cluster P. Termination. If no such pair of vertices is found then terminate. This algorithm is an simple extension to the local search algorithm described in Section 3.1 and it is easy to show that it terminates in time polynomial with m and the maximum length of a BAC. Furthermore, it was shown in (Gaur et al., 2007) that this algorithm has a guaranteed approximation factor of 1−1/n.

3.3 The pooling problem under f(C)=C−1

We now focus on the cost function f(C)=C−1. As we discussed in the problem definition (Section 2), the pooling problem under cost function f(C)=C−1 can be reduced to a hypergraph partitioning problem as follows. Let G be a hypergraph where each BAC B is represented with a unique vertex v and each subset S of at most m vertices, are connected with a hyperedge e. The weight of e, namely w(e) is the number of k-mers that occur in all BACs in S and occur in no other BACs. Consider a partitioning of V, the vertex set of G, into non-overlapping pools P={P1,…,P}. For a given subset S of vertices, let #(S) be the number of pools of P that have at least one vertex of S. Then the cost of e with respect to P is w(e)·(|S|−#(S)) and the goal of the hypergraph partitioning problem is to minimize the total cost of all hyperedges. The ‘dual’ of this hypergraph partitioning would be another partitioning problem where the cost of e with respect to P is w(e)·(#(S)−1), and the goal is to maximize the total cost of all hyperedges. Input: A weighted hypergraph G(V,w), with vertex set V, and weights w(e) for each hyperedge e (which connects the set S⊆V for |S|≤d). Output: A partitioning of vertices V into pools P={P1,P2,···,P}, , which maximizes the following objective function: We give a greedy algorithm to solve the above hypergraph partitioning problem. The algorithm, at each iteration x randomly picks a vertex v∈V′, where V′ is the set of vertices not processed so far, and adds it to one of the pools P as described below. Before we provide further details, we give some definitions. Let P be the set of vertices in pool P before iteration x and let P={P,…,P}. (Thus .) Given some S⊂V, let #(S) denote the number of pools P∈P which include at least one vertex of S. Also let be a boolean function such that if P∩S=∅ and otherwise.

3.3.1 A greedy algorithm for hypergraph partitioning problem (GAHP)

The above algorithm achieves an approximation factor of 1−d/2n as shown below. As an initial step, set V0′=V and . In each iteration x∈{1,…,m}, randomly pick a vertex v∈V′ and put v into the pool (Thus, ∀i≤n,i≠k,P=P and P=P∪{v}). First, we need to find a lower bound on the total cost w(e)·(#(S)−1) with respect to the pool set P={P,P,·P} returned by the algorithm at the end of iteration m. It is not hard to see that for any set of vertices S (for the remainder of the proof, all sets S we consider will satisfy |S|≤d), for which v∈S: Thus, Now taking the sum of above equation for all possible hyperedges we would get: For bounding the left hand side of equation 8, we will consider an arbitrary iteration x. By adding up the right hand side of equation 9 over all values of x we get: For bounding equation 10 we first have to argue for any arbitrary S⊆V we have, Thus, Now using Equations 9–11 we will have: Utilizing Equations 12 and 8 we conclude: Now to find the approximation factor for this greedy algorithm we need to find an upper bound of the optimal solution. It is easy to see that for even optimal partitioning ∑w(e)·(#(S)−1)≤∑w(e)·(|S|−1). Thus, the approximation factor α can be bounded as: Now, as |S|

3.4 The balanced pooling problem under f(C)=C−1

Our last algorithm deals with the balanced pooling problem under the cost function f(C)=C−1, for which we give a greedy approximation algorithm. We remind the reader that there are m=nh vertices to be assigned into n pools, and eventually each pool must have exactly h vertices.

3.4.1 A greedy algorithm for balanced hypergraph partitioning (GABHP)

The algorithm starts with a set of n empty pools, P={P1,…,P}, and at each iteration x, it selects a set of n arbitrary vertices, say Y={y1,,…,y}, which are not assigned to any of the pools yet, and adds them to the pools such that each pool receives exactly one new vertex. Let the set of vertices in pool P at the beginning of iteration x be denoted by P and let P={P,…,P}. Thus, in iteration x, each y is assigned to exactly one of the pools P. For any set of vertices S∈V, let λ(y,P,S) be a boolean function defined as follows. Intuitively, for a given vertex y, a pool P, and a vertex set S,λ(y,P,S) is equal to zero if and only if no other vertex yℓ, incident to the hyperedge e has already been assigned to the pool P. Then, for each pool P, we define the marginal cost function μ(y,P) with respect to the potential assignment of y to P, as follows. We now construct a new complete bipartite graph H with vertex sets Y and P such that for any vertex y∈Y and pool P∈P, there exists an edge in H with weight μ(y,P). Then we find a perfect minimum weighted matching for H, i.e. a perfect matching where the sum of weights of the edges in the matching has the minimum possible value by using the well-known Hungarian algorithm (Munkres, 1957). For each j, 1≤j≤n, let π(y) be the pool which is matched to y in the prefect minimum bipartite matching of the graph H. We add the vertex y to the pool π. We run the above iterative step for x=1,…,h so as to assign each one of the m=nh vertices into one pool in a balanced manner. The above algorithm gives an approximation to the balanced hypergraph partitioning problem within a factor of 1−d/2n, where d is the maximum number of vertices that can be incident to a hyperedge. The proof for the approximation factor is similar to that for the unbalanced hypergraph partitioning problem and thus is omitted.

4 RESULTS AND DISCUSSION

We report results on two sets of BACs (with m=150) on which we tested our algorithms for both balanced and unbalanced pooling problem using both cost functions. We start by noting that for both data sets the results obtained by the balanced pooling algorithms turned out to be almost identical to those obtained by the unbalanced pooling algorithms for each of the two cost functions we used. In other words, the cost of the partition obtained by the LSMnC algorithm was almost identical to that of the LSMnS algorithm and the cost obtained by the GAHP algorithm was very similar to that of the GABHP problem. It is also interesting to note that the unbalanced pooling algorithms returned very balanced partitions. We compare the performance of these algorithms with that of random partitioning of BACs into pools. We measure the performance of our algorithms and that of random partitionings with respect to the general cost function f(C)=C−1. The total cost of a particular partitioning of a set of m BACs into pools P1,…,P, is . In order to compute the cost J for a partitioning, we construct, for each pool P, the joint trie of the k-mers (for this study k=50) of all BACs in P, denoted by T. The trie T can be constructed in time linear with the total lengths of the BACs in P as per the linear time algorithms for suffix tree construction (Mccreight, 1976; Sahinalp and Vishkin, 1994). During the construction of T, at each leaf corresponding to a k-mer b, we maintain the labels of the specific BACs that include b. By going through all leaves of each T, we compute J in time linear with the total lengths of the BACs. We used two data sets in our experiments, each consisting of 150 BACs. The first set of clones were collected in the high-resolution analysis of lymphoma genomes project at the BC Genome Sciences Centre. They represent regions of interest in the genome of a tumor sample, where there are marked local deviations from the reference human genome. The BAC coordinates are deduced by aligning BAC fingerprints to the reference genome (Krzywinski et al., 2007) and are confirmed by BAC end sequencing experiments. The second set is a synthetic library of clones with a mean size of 182 kb and a standard deviation of 34 kb, representing a random sampling of the finished regions of the reference human genome, hg18.

4.1 Pooling experiments with LSMnC/LSMnS algorithms

We first compare the performance of our local search methods LSMnC/LSMnS with random partitions (denoted ranPool). Although these methods were designed to minimize the cost of pooling with respect to the cost function , we report their performance with respect to the second cost function, f(C)=C−1, as, ultimately that is the cost function we would like to minimize. Note that ranPool is known to give an expected approximation factor of 1−1/n for the max n-cut problem. However, LSMnC will guarantee a worst case approximation factor of 1−1/n for the max n-cut problem. In Figures 1(a) and 2(a), we depict how the cost of the ranPool method for 5000 independent trials, as well as the cost of LSMnC/LSMnS method varies with respect to the number of pools on the lymphoma and the synthetic data sets (we show the mean, the highest and the lowest 25% cost for the ranPool method). We also depict how the ratio between the (average) cost of the ranPool and the cost of the LSMnC/LSMnS method changes with respect to the number of pools in Figures 1(b) and 2(b). We can observe that as the number of pools increases, the costs of both ranPool and LSMnC/LSMnS decrease for both data sets. Perhaps more interestingly, we can also observe that as the number of pools increases, the ratio between the (average) cost of ranPool and that of LSMnC/LSMnS increases as well.
Fig. 1.

A comparison of the cost of ranPool and LSMnC/LSMnS methods with respect to the number of pools on the lymphoma data set. (a) The actual cost of LSMnC/LSMnS method (red) and the cost of ranPool (green) with respect to the number of pools (mean, upper quartile and lower quartile results reported for 5000 independent runs) with respect to the number of pools. (b) The ratio of the average cost of ranPool (on 5000 independent runs) and the cost of LSMnC/LSMnS methods.

Fig. 2.

A comparison of the cost of ranPool and LSMnC/LSMnS methods with respect to the number of pools on the synthetic data set. (a) The actual cost of LSMnC/LSMnS method (red) and the cost of ranPool (green) with respect to the number of pools (mean, upper quartile and lower quartile results reported for 5000 independent runs) with respect to the number of pools. (b) The ratio of the average cost of ranPool (on 5000 independent runs) and the cost of LSMnC/LSMnS methods.

A comparison of the cost of ranPool and LSMnC/LSMnS methods with respect to the number of pools on the lymphoma data set. (a) The actual cost of LSMnC/LSMnS method (red) and the cost of ranPool (green) with respect to the number of pools (mean, upper quartile and lower quartile results reported for 5000 independent runs) with respect to the number of pools. (b) The ratio of the average cost of ranPool (on 5000 independent runs) and the cost of LSMnC/LSMnS methods. A comparison of the cost of ranPool and LSMnC/LSMnS methods with respect to the number of pools on the synthetic data set. (a) The actual cost of LSMnC/LSMnS method (red) and the cost of ranPool (green) with respect to the number of pools (mean, upper quartile and lower quartile results reported for 5000 independent runs) with respect to the number of pools. (b) The ratio of the average cost of ranPool (on 5000 independent runs) and the cost of LSMnC/LSMnS methods. For a specific number of pools (n=15) we depict the distribution of the costs obtained in the 5000 independent runs of both ranPool and the LSMnC/LSMnS methods on both the lymphoma and the synthetic data sets in Figures 3 and 4, respectively. As can be seen, the costs obtained by independent trials of the LSMnC/LSMnS method (each starting with a unique random partitioning) are highly concentrated around the mean. Furthermore, even the minimum cost obtained by the ranPool method is well above the maximum cost returned by the LSMnC/LSMnS method.
Fig. 3.

The distribution of cost obtained by ranPool and LSMnC/LSMnS for n=15 on the lymphoma data set for 5000 independent runs.

Fig. 4.

The distribution of cost obtained by ranPool and LSMnC/LSMnS for n=15 on the synthetic data set for 5000 independent runs.

The distribution of cost obtained by ranPool and LSMnC/LSMnS for n=15 on the lymphoma data set for 5000 independent runs. The distribution of cost obtained by ranPool and LSMnC/LSMnS for n=15 on the synthetic data set for 5000 independent runs. We can conclude that the results obtained by the LSMnC/LSMnS approach is much better than those obtained by ranPool. For example, at n=15 the LSMnC/LSMnS approach provides a factor 2 improvement over the average cost of 5000 independent runs of ranPool on the lymphoma data set. The cost improvement of LSMnC/LSMnS approach with respect to the minimum cost obtained by ranPool after 5000 independent trials is 1.4 on the same data set.

4.2 Pooling experiments with GAHP/GABHP algorithms

The local search algorithms LSMnC and LSMnS aim to ‘minimize’ the cost of pooling with respect to the cost function ; however, the cost obtained by these algorithms were considerably lower than that obtained by ranPool even with respect to the second cost function—which provides a more accurate performance measure. Our second set of algorithms, GAHP/GABHP are designed to ‘minimize’ the cost with respect to the second cost function. They are flexible in the sense that one can set up the value of d as desired; for d=n, the optimal solution to the hypergraph partitioning indeed minimizes the cost function f(C)=C−1. We tried the two algorithms for both d=2 and 3 in order to evaluate their advantage over the local search algorithms as well as random partitionings. The running time of both GAHP and the GABHP algorithms are exponential in d (the number of hyperedges grow exponentially with increasing d), thus it is of crucial importance to know up to which value of d, an improvement in performance could be expected. A significant performance improvement by GAHP/GABHP methods using d=3 over LSMnC/LSMnS methods (which solve the hypergraph partitioning problem for d=2) may imply that d should be increased to 4 or more for better performance. In Figures 5(a) and 6(a) we compare the (mean value as well as the highest and lowest 25%) costs obtained by 5000 independent runs of the ranPool method and that of the GAHP/GABHP approach on the lymphoma and the synthetic data sets. Figures 5(a) and 6(a) show how the costs change with respect to the increasing number of pools. We also show how the ratio between the average cost of the ranPool and the cost of GAHP/GABHP methods change with respect to the number of pools in Figure 7(a) and (b).
Fig. 5.

A comparison of the cost of ranPool and GAHP/GABHP (d=3) methods with respect to the number of pools on the lymphoma data set. (a) The actual cost of LSMnC/LSMnS method (red) and the cost of ranPool (green) with respect to the number of pools (mean, upper quartile and lower quartile results reported for 5000 independent runs) with respect to the number of pools. (b) The ratio of the average cost of ranPool (on 5000 independent runs) and the cost of GAHP/GABHP methods.

Fig. 6.

A comparison of the cost of ranPool and GAHP/GABHP (d=3) methods with respect to the number of pools on the synthetic data set. (a) The actual cost of LSMnC/LSMnS method (red) and the cost of ranPool (green) with respect to the number of pools (mean, upper quartile and lower quartile results reported for 5000 independent runs) with respect to the number of pools. (b) The ratio of the average cost of ranPool (on 5000 independent runs) and the cost of GAHP/GABHP methods.

Fig. 7.

A comparison of the cost of GAHP (GABHP) method for d=2 (red) and d=3 (green) with respect to the number of pools, (a) on the lymphoma data set, (b) on the synthetic data set.

A comparison of the cost of ranPool and GAHP/GABHP (d=3) methods with respect to the number of pools on the lymphoma data set. (a) The actual cost of LSMnC/LSMnS method (red) and the cost of ranPool (green) with respect to the number of pools (mean, upper quartile and lower quartile results reported for 5000 independent runs) with respect to the number of pools. (b) The ratio of the average cost of ranPool (on 5000 independent runs) and the cost of GAHP/GABHP methods. A comparison of the cost of ranPool and GAHP/GABHP (d=3) methods with respect to the number of pools on the synthetic data set. (a) The actual cost of LSMnC/LSMnS method (red) and the cost of ranPool (green) with respect to the number of pools (mean, upper quartile and lower quartile results reported for 5000 independent runs) with respect to the number of pools. (b) The ratio of the average cost of ranPool (on 5000 independent runs) and the cost of GAHP/GABHP methods. A comparison of the cost of GAHP (GABHP) method for d=2 (red) and d=3 (green) with respect to the number of pools, (a) on the lymphoma data set, (b) on the synthetic data set. We investigate the effect of using hypergraphs with d=3 over the use of ordinary graphs with d=2 on the GAHP/GABHP methods in Figure 7(a) and (b). As can be observed, the performance improvement achieved by increasing d form 2 to 3 is negligible for both data sets. It may be possible to explain this observation by investigating the distribution of repeat sequences among the BACs in the two data sets. The number of k-mers which are repeated in exactly two BACs are 100–200 times more than those repeated in three BACs or more; see Figure 8(a) and (b) for the distribution of hyperedge weights in the two data sets. Thus, the total weight of hyperedges incident to three vertices or more is insignificant in comparison to edges that are incident to exactly two vertices. As a result, on the two data sets we used, the hypergraph partitioning algorithms largely ‘ignore’ the hyperedges whose contribution to the total cost is very small. We expect that the performance of the GAHP/GABHP methods for d=3 will be superior to that for d=2 if highly repetitive BACs are sequenced.
Fig. 8.

The distribution of hyperedge weights (in log scale) among 5000 BACs, (a) from the lymphoma data set, (b) from the synthetic data set.

The distribution of hyperedge weights (in log scale) among 5000 BACs, (a) from the lymphoma data set, (b) from the synthetic data set. We finally compare the two alternative algorithmic methods we introduce in this article; LSMnC/LSMnS and GAHP/GABHP (for d=2). In Figure 9(a) and (b), a comparison of the two methods are provided for both data sets. Interestingly enough, the performance of LSMnC/LSMnS method is slightly better than the GAHP/GABHP method for both data sets. This indicates that LSMnC/LSMnS (which finds a local optimum) outperforms a greedy based method (GAHP/GABHP), which does not necessarily find a local optimum. However, we believe that if regions of DNA with high repetitions are used as the input, GAHP/GABHP (especially for d>2) would give better results compared to LSMnC/LSMnS and even GAHP/GABHP for d=2.
Fig. 9.

A comparison of the cost of LSMnC/LSMnS method (red) and GAHP/GABHP method for d=2 (green) with respect to the number of pools, (a) on the lymphoma data set, (b) on the synthetic data set.

A comparison of the cost of LSMnC/LSMnS method (red) and GAHP/GABHP method for d=2 (green) with respect to the number of pools, (a) on the lymphoma data set, (b) on the synthetic data set. Conflict of Interest: none declared.
  13 in total

1.  An Eulerian path approach to DNA fragment assembly.

Authors:  P A Pevzner; H Tang; M S Waterman
Journal:  Proc Natl Acad Sci U S A       Date:  2001-08-14       Impact factor: 11.205

2.  Solexa Ltd.

Authors:  Simon Bennett
Journal:  Pharmacogenomics       Date:  2004-06       Impact factor: 2.533

3.  Fragment assembly with short reads.

Authors:  Mark Chaisson; Pavel Pevzner; Haixu Tang
Journal:  Bioinformatics       Date:  2004-04-01       Impact factor: 6.937

4.  Genome sequencing in microfabricated high-density picolitre reactors.

Authors:  Marcel Margulies; Michael Egholm; William E Altman; Said Attiya; Joel S Bader; Lisa A Bemben; Jan Berka; Michael S Braverman; Yi-Ju Chen; Zhoutao Chen; Scott B Dewell; Lei Du; Joseph M Fierro; Xavier V Gomes; Brian C Godwin; Wen He; Scott Helgesen; Chun Heen Ho; Chun He Ho; Gerard P Irzyk; Szilveszter C Jando; Maria L I Alenquer; Thomas P Jarvie; Kshama B Jirage; Jong-Bum Kim; James R Knight; Janna R Lanza; John H Leamon; Steven M Lefkowitz; Ming Lei; Jing Li; Kenton L Lohman; Hong Lu; Vinod B Makhijani; Keith E McDade; Michael P McKenna; Eugene W Myers; Elizabeth Nickerson; John R Nobile; Ramona Plant; Bernard P Puc; Michael T Ronan; George T Roth; Gary J Sarkis; Jan Fredrik Simons; John W Simpson; Maithreyan Srinivasan; Karrie R Tartaro; Alexander Tomasz; Kari A Vogt; Greg A Volkmer; Shally H Wang; Yong Wang; Michael P Weiner; Pengguang Yu; Richard F Begley; Jonathan M Rothberg
Journal:  Nature       Date:  2005-07-31       Impact factor: 49.962

5.  Large-scale BAC clone restriction digest fingerprinting.

Authors:  Carrie A Mathewson; Jacqueline E Schein; Marco A Marra
Journal:  Curr Protoc Hum Genet       Date:  2007-04

6.  Nucleotide sequence of bacteriophage lambda DNA.

Authors:  F Sanger; A R Coulson; G F Hong; D F Hill; G B Petersen
Journal:  J Mol Biol       Date:  1982-12-25       Impact factor: 5.469

7.  Cloning in single-stranded bacteriophage as an aid to rapid DNA sequencing.

Authors:  F Sanger; A R Coulson; B G Barrell; A J Smith; B A Roe
Journal:  J Mol Biol       Date:  1980-10-25       Impact factor: 5.469

8.  DNA sequencing with chain-terminating inhibitors.

Authors:  F Sanger; S Nicklen; A R Coulson
Journal:  Proc Natl Acad Sci U S A       Date:  1977-12       Impact factor: 11.205

9.  Assembling millions of short DNA sequences using SSAKE.

Authors:  René L Warren; Granger G Sutton; Steven J M Jones; Robert A Holt
Journal:  Bioinformatics       Date:  2006-12-08       Impact factor: 6.937

10.  Whole-genome sequencing and assembly with high-throughput, short-read technologies.

Authors:  Andreas Sundquist; Mostafa Ronaghi; Haixu Tang; Pavel Pevzner; Serafim Batzoglou
Journal:  PLoS One       Date:  2007-05-30       Impact factor: 3.240

View more
  8 in total

1.  Overlapping pools for high-throughput targeted resequencing.

Authors:  Snehit Prabhu; Itsik Pe'er
Journal:  Genome Res       Date:  2009-05-15       Impact factor: 9.043

2.  Visual analysis of the quantitative composition of metagenomic communities: the AmphoraVizu webserver.

Authors:  Csaba Kerepesi; Balázs Szalkai; Vince Grolmusz
Journal:  Microb Ecol       Date:  2014-10-10       Impact factor: 4.552

3.  A statistical method for the detection of variants from next-generation resequencing of DNA pools.

Authors:  Vikas Bansal
Journal:  Bioinformatics       Date:  2010-06-15       Impact factor: 6.937

4.  A computer simulator for assessing different challenges and strategies of de novo sequence assembly.

Authors:  Bjarne Knudsen; Roald Forsberg; Michael M Miyamoto
Journal:  Genes (Basel)       Date:  2010-09-13       Impact factor: 4.096

5.  Empirical validation of pooled whole genome population re-sequencing in Drosophila melanogaster.

Authors:  Yuan Zhu; Alan O Bergland; Josefa González; Dmitri A Petrov
Journal:  PLoS One       Date:  2012-07-26       Impact factor: 3.240

6.  Combinatorial pooling enables selective sequencing of the barley gene space.

Authors:  Stefano Lonardi; Denisa Duma; Matthew Alpert; Francesca Cordero; Marco Beccuti; Prasanna R Bhat; Yonghui Wu; Gianfranco Ciardo; Burair Alsaihati; Yaqin Ma; Steve Wanamaker; Josh Resnik; Serdar Bozdag; Ming-Cheng Luo; Timothy J Close
Journal:  PLoS Comput Biol       Date:  2013-04-04       Impact factor: 4.475

7.  LDx: estimation of linkage disequilibrium from high-throughput pooled resequencing data.

Authors:  Alison F Feder; Dmitri A Petrov; Alan O Bergland
Journal:  PLoS One       Date:  2012-11-09       Impact factor: 3.240

8.  Improving mapping and SNP-calling performance in multiplexed targeted next-generation sequencing.

Authors:  Abdou Elsharawy; Michael Forster; Nadine Schracke; Andreas Keller; Ingo Thomsen; Britt-Sabina Petersen; Björn Stade; Peer Stähler; Stefan Schreiber; Philip Rosenstiel; Andre Franke
Journal:  BMC Genomics       Date:  2012-08-22       Impact factor: 3.969

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.