Literature DB >> 34965857

An improved approximation algorithm for the reversal and transposition distance considering gene order and intergenic sizes.

Klairton L Brito1, Andre R Oliveira2, Alexsandro O Alexandrino2, Ulisses Dias3, Zanoni Dias2.   

Abstract

BACKGROUND: In the comparative genomics field, one of the goals is to estimate a sequence of genetic changes capable of transforming a genome into another. Genome rearrangement events are mutations that can alter the genetic content or the arrangement of elements from the genome. Reversal and transposition are two of the most studied genome rearrangement events. A reversal inverts a segment of a genome while a transposition swaps two consecutive segments. Initial studies in the area considered only the order of the genes. Recent works have incorporated other genetic information in the model. In particular, the information regarding the size of intergenic regions, which are structures between each pair of genes and in the extremities of a linear genome. RESULTS AND
CONCLUSIONS: In this work, we investigate the SORTING BY INTERGENIC REVERSALS AND TRANSPOSITIONS problem on genomes sharing the same set of genes, considering the cases where the orientation of genes is known and unknown. Besides, we explored a variant of the problem, which generalizes the transposition event. As a result, we present an approximation algorithm that guarantees an approximation factor of 4 for both cases considering the reversal and transposition (classic definition) events, an improvement from the 4.5-approximation previously known for the scenario where the orientation of the genes is unknown. We also present a 3-approximation algorithm by incorporating the generalized transposition event, and we propose a greedy strategy to improve the performance of the algorithms. We performed practical tests adopting simulated data which indicated that the algorithms, in both cases, tend to perform better when compared with the best-known algorithms for the problem. Lastly, we conducted experiments using real genomes to demonstrate the applicability of the algorithms.
© 2021. The Author(s).

Entities:  

Keywords:  Approximation algorithms; Genome rearrangement events; Intergenic regions

Year:  2021        PMID: 34965857      PMCID: PMC8717661          DOI: 10.1186/s13015-021-00203-7

Source DB:  PubMed          Journal:  Algorithms Mol Biol        ISSN: 1748-7188            Impact factor:   1.405


Background

In the comparative genomics field, there are many ways to compare genomic features like DNA sequence, gene order, and genomic landmarks from different organisms. Genome rearrangement events are mutations that affect large stretches of the DNA sequence. Determining the shortest sequence of such events that can transform one genome into another is widely used as a metric to study evolutionary relationships among organisms, and to explain biological similarities and differences as well. The reversal and transposition are two of the most studied genome rearrangement events in the literature [1-3]. A reversal inverts a segment of a genome, and a transposition moves a segment of a genome to another position. One way to represent a genome is by using the gene order as the only genomic trait, which can be encoded as a sequence of elements, where each element represents a gene. When the compared genomes share the same set of genes and do not have replicated genes, we model them as permutations of natural numbers, such that each number in the sequence appears once. Furthermore, if the orientation of the genes is known, a plus or a minus sign (+ or −) is assigned to each element of the permutation to indicate its orientation, and we say that the permutation is a signed permutation. Otherwise, signs are omitted and the permutation is called unsigned. It is always possible to map the target genome in a permutation such that the elements are in increasing order. This permutation is called by identity permutation and denoted as and , considering unsigned and signed cases, respectively. Therefore, the transformation from a source genome to a target genome can be seen as a sorting problem. First studies in the genome rearrangement field considered a single type of rearrangement events, which led to solutions specific to that type. In particular, the reversal event leads to the sorting by reversals problem, which has a polynomial-time algorithm on signed permutations [1], whereas it is NP-hard on unsigned permutations [4] and the best algorithm has an approximation factor of 1.375 [2]. The transposition event leads to the sorting by transpositions problem, which is NP-hard [5] and the best algorithm has an approximation factor of 1.375 [3]. By allowing both reversal and transposition we have the sorting by reversals and transpositions problem, which is NP-hard on signed and unsigned permutations [6]. The best algorithms have approximation factors of 2 [7] and [8, 9] for signed and unsigned permutations, respectively. The gene order was fundamental to the initial development of rearrangement distance models. However, recent studies indicate that incorporating another genetic information apart from the gene order could generate more realistic models [10, 11]. In particular, the information regarding the size of intergenic regions (structures with a specific number of nucleotides between each pair of genes and in the extremities of genomes) was incorporated into the mathematical models. The Double Cut and Join (DCJ) is a rearrangement event that cuts the genome in two points and reassembles the stretches following a predetermined criterion. The problem of sorting by dcjs with intergenic regions is NP-hard [12], but one can find a polynomial-time algorithm when DCJs are used together with insertions and deletions on intergenic regions [13]. The Block-Interchange is a rearrangement event that swaps the position of two segments (not necessarily consecutive) of the genome. The sorting by intergenic block-interchange problem has a 2-approximation algorithm [14] and its complexity is unknown. Considering the reversal event, we have the sorting by intergenic reversals problem, which is NP-hard on signed and unsigned permutations [15, 16] and the best algorithms have approximation factors of 2 [15] and 4 [16], respectively. The sorting by intergenic transpositions is NP-hard and the best algorithm has an approximation factor of 3.5 [17]. The sorting by intergenic reversals and transpositions (SbIRT) is NP-hard on signed and unsigned permutations [16, 17] and the best algorithms have the approximation factors of 3 [17] and 4.5 [16], respectively. The problem with the generalized definition of the transposition event on signed permutations has an approximation algorithm with a factor of 2.5 [17]. The problem with an additional constraint that limits the number of genes affected by each operation, called super short operations, was investigated [18]. On signed permutations it was proposed a 5-approximation algorithm, while for unsigned permutations it was proposed a 3-approximation algorithm. In this work, we investigate the problem on signed and unsigned permutations. For the unsigned case, we present an improved algorithm based on intergenic breakpoints that guarantees an approximation factor of 4. We also show a 3-approximation algorithm for the problem on unsigned permutations considering the generalized definition of the transposition event. For the signed case, we show approximation algorithms with the same approximation factors as in the unsigned cases. Although the theoretical approximations for the signed case are superior to the previously known results, the tests with simulated data pointed that our algorithms tend to provide better practical results. We propose a greedy strategy to improve the algorithms’ performance and tested them using simulated and real data. This manuscript is organized as follows. "Definitions" section presents concepts and definitions used throughout the paper. "Theoretical results" section shows a lower bound and the approximation algorithms for the problem. "Practical results" section shows the experiments using real and simulated data. "Conclusion" section concludes the paper and introduces future directions.

Definitions

The problem we investigate uses information about source and target genomes. We assume that both genomes share the same set of genes and there are no replicated genes. Thus, given a linear genome with n genes and intergenic regions, we use (i) a permutation , representing the order of the genes, and (ii) a list of non-negative integer numbers , representing the sizes of intergenic regions. If the orientation of the genes is known, a “” or “−” sign is associated with each element of the permutation to indicate its orientation. We use , , to denote the element in position i of . Similarly, we denote by the size of the intergenic region in the left of . The intergenic region is on the right of . For convenience, we map the genes from the target genome to the identity permutation for the case where the orientation of the genes is unknown and otherwise. The permutation of the source genome can be mapped according to how we assigned elements to genes while mapping the target genome to the identity permutation, so the source and target genomes are represented as and , respectively. Since the identity permutation is fixed given the size of the genomes, an instance for the problem is composed by . Figure 1 shows the representation of a genome as .
Fig. 1

On the top we have a fictitious genome , with 5 genes and an intergenic region between each pair of genes and also in the extremities of . Each intergenic region has a specific number of nucleotides (represented by the letters A, C, G, or T). At the bottom, we have the genome representation for the problem, using a permutation and a list , representing the order of the genes and the sizes of the intergenic regions, respectively. Observe that for each intergenic region we have the information about the number of nucleotides inside it

On the top we have a fictitious genome , with 5 genes and an intergenic region between each pair of genes and also in the extremities of . Each intergenic region has a specific number of nucleotides (represented by the letters A, C, G, or T). At the bottom, we have the genome representation for the problem, using a permutation and a list , representing the order of the genes and the sizes of the intergenic regions, respectively. Observe that for each intergenic region we have the information about the number of nucleotides inside it From now on, we will refer to an instance of the problem in which the orientation of the genes is known and unknown by signed and unsigned instances, respectively. In this work, initially, we consider two rearrangement events: intergenic reversal and intergenic transposition. In the following, we formally describe them and show how they affect a given genome representation.

Definition 2.1

Given a genome , let i, j, x,  and y be integers such that , , and . An intergenic reversal applied to splits the intergenic regions (into ) and (into ), where and , and it reverses the sequence , generating as follows:such that and . If the orientation of the genes is known, the signs of the genes in the affected segment are flipped as follows:

Definition 2.2

Given a genome , let i, j, k, x, y,  and z be integers such that , , and . An intergenic transposition applied to splits the intergenic regions (into ), (into ), and (into ), where , , and , and swaps the adjacent segments and , generating as follows:such that , , and . Given a genome and an operation , represents the operation applied on . Similarly, given a sequence of operations , we use to denote . We hereafter refer to intergenic reversals and intergenic transpositions simply as reversals and transpositions, respectively. Note that both reversal and transposition are conservative events, i.e., they do not insert or remove genes nor nucleotides. Thus, an instance from the problem is valid if the following equality is satisfied:Given an instance of the problem, the minimum number of operations needed to transform into is called the distance and is denoted by and for the signed and unsigned cases, respectively. The extended form of is obtained by adding the elements and at the beginning and at the end of , respectively. We hereafter assume that permutations are in extended form, and we refer to them simply as permutations. Following, we present concepts and definitions that are used in previous works [16] regarding the problem.

Definition 2.3

Given an unsigned instance of the problem, a pair of elements , such that , is an intergenic breakpoint type one if one of the following cases occur: . and , such that .

Definition 2.4

Given an unsigned instance of the problem, a pair of elements is an intergenic adjacency if and is not an intergenic breakpoint type one. In other words, an intergenic breakpoint type one indicates a region that must be affected by a rearrangement event to fix the order of the genes or the size of the intergenic region to reach the target genome. On the other hand, an intergenic adjacency indicates a pair of genes that are consecutive in the target genome and the intergenic regions between them have the same size. From now on, we will refer to intergenic breakpoint and intergenic adjacency as breakpoint and adjacency, respectively.

Definition 2.5

A breakpoint type one , such that , is overcharged if , such that , and undercharged otherwise.

Definition 2.6

A pair of breakpoints type one and is connected if the following conditions are met: The pair of elements , , , , , or are consecutive in and do not form an adjacency in . , such that is the intergenic region size between the consecutive elements (from condition 1) in . A pair of connected breakpoints indicate that it is possible to form an adjacency using only the nucleotides from the intergenic regions of the two breakpoints. Note that a pair of connected breakpoints has at least one pair of consecutive elements between , , , and . Besides, the number of nucleotides in both breakpoints () is at least the size of the intergenic region between the consecutive elements in . Now, we introduce new definitions which are used to derive the results.

Definition 2.7

A breakpoint type one is called hard if it is overcharged or undercharged and soft otherwise. Note that in hard breakpoints the pair of genes are consecutive in the target genome, but the intergenic region between them is not the same as in the target genome.

Definition 2.8

A pair of breakpoints type one and is called softly connected if they are connected and both breakpoints are soft.

Definition 2.9

A hard breakpoint is called super hard if one of the following cases occur: or . or is a hard breakpoint or an adjacency. Note that a super hard breakpoint is in one of the extremities of the genome, or immediately before or after the breakpoint exists a hard breakpoint or an adjacency.

Definition 2.10

Given an unsigned instance , strips are maximal sequences of consecutive elements of without soft breakpoints. A strip with only one element is called a singleton, and it is defined as increasing if , and as decreasing otherwise. A strip with more than one element is called increasing if its elements form an increase sequence; and it is called decreasing otherwise. For a signed instance of the problem, we have the following definition:

Definition 2.11

Given a signed instance of the problem, a pair of elements , such that , is an intergenic breakpoint type two if one of the following cases occur: . and , such that . Given an unsigned instance of the problem, the total number of hard and soft breakpoints are denoted by and , respectively, and the total number of breakpoints type one is denoted by . The variation in the number of breakpoints type one after applying a rearrangement event to is denoted by , where . Similarly, given a signed instance of the problem, the total number of breakpoints type two is denoted by and the variation in the number of breakpoints type two after applying a rearrangement event to is denoted by , where .

Remark 2.1

The only unsigned instance I of the problem such that is . Similarly, the only signed instance of the problem, such that is . Thus, to transform into it is necessary to remove all the breakpoints of an instance. Figure 2 shows the concepts using a representation of the source and target genomes.
Fig. 2

An instance , where and . On top we have the source genome and in the bottom the target genome . Breakpoints are indicated above the source genome. Note that and , so . The hard breakpoints and are overcharged and undercharged, respectively. Breakpoints and are soft. The breakpoints and are connected, while and are not. Besides, the pair of breakpoints and are softly connected. The instance I has the increasing strips and ), and the decreasing strip

An instance , where and . On top we have the source genome and in the bottom the target genome . Breakpoints are indicated above the source genome. Note that and , so . The hard breakpoints and are overcharged and undercharged, respectively. Breakpoints and are soft. The breakpoints and are connected, while and are not. Besides, the pair of breakpoints and are softly connected. The instance I has the increasing strips and ), and the decreasing strip

Theoretical results

In this section, we show lower bounds and present approximation algorithms for both cases of the problem. We start by showing how many breakpoints a reversal and a transposition can remove in the best scenario.

Lemma 3.1

Given an unsigned instance and a signed instance of the problem, and for any reversal , respectively.

Proof

Recall that a reversal affects two pair of consecutive elements of . In the best case, and are breakpoints, and the reversal removes them.

Lemma 3.2

Given an unsigned instance and a signed instance of the problem, and for any transposition , respectively. The proof is similar to the one described in Lemma 3.1, and considering that a transposition can affect up to three breakpoints. Using the above lemmas we define a lower bound for the problem.

Proposition 3.1

(Proposition 7 [16]) Given an unsigned instance , .

Proposition 3.2

Given a signed instance , . Directly by Remark 2.1 and lemmas 3.1 and 3.2.

Approximation algorithms for the unsigned case of the problem

In this section, we investigate the unsigned case of the problem and present a 4-approximation algorithm considering the reversal and transpositions events. Besides, we show a 3-approximation algorithm incorporating a generalized definition of the transposition event. We show a sequence of lemmas that will be used by the algorithms as subroutines.

Lemma 3.3

(Lemma 19 [16]) It is possible to perform any redistribution of nucleotides within intergenic regions , , and using two consecutive transpositions in the format: In this section, we will refer to breakpoint type one simply as a breakpoint. Now let us show how to remove breakpoints from an unsigned instance depending on how many overcharged breakpoints an instance has.

Lemma 3.4

Given an unsigned instance for the problem, if there are at least two overcharged breakpoints then there exists a sequence of two transpositions that removes at least two breakpoints.

Proof

First note that a third breakpoint must exist in , otherwise the total number of nucleotides within intergenic regions of the source genome would be greater than the number of nucleotides within intergenic regions of the target genome. By Lemma 3.3, it is possible to make a redistribution of nucleotides within three intergenic regions using two consecutive transpositions. Without loss of generality, assume that two of these intergenic regions are between the two overcharged breakpoints, and that the third intergenic region is between an existing third breakpoint. In this case, the extra nucleotides from the two overcharged breakpoints are moved to the third breakpoint, and the lemma follows.

Lemma 3.5

Given an unsigned instance for the problem, if there is a pair of softly connected breakpoints then there exists a reversal or a transposition that removes at least one breakpoint. Brito et. al. [16, lemmas 14 and 20] showed how to remove a breakpoint from a pair of connected breakpoints. In particular, when both breakpoints and are soft, we have one of the following three possibilities to form at least one adjacency from them:For each one of the cases, a reversal or a transposition can be applied to remove at least one breakpoint, and the lemma follows. Case 1: or are consecutive in . Case 2: are consecutive in . Case 3: are consecutive in . Figure 3 shows, for each case in Lemma 3.5, a reversal or a transposition that can be applied to remove at least one breakpoint. In Case 3, a transposition is applied to the pair of soft breakpoints and in a third breakpoint, which can be located before or after the pair of soft breakpoints.
Fig. 3

The possibilities that can arise when a pair of softly connected breakpoints exists. In this case, one operation can be applied to remove at least one breakpoint. The pair of elements that are consecutive in the identity permutation is represented with a grayscale color

Remark 3.1

Note that Case 2 of Lemma 3.5 is the only one in which a hard breakpoint can be removed as a result of the operation applied ( and ). However, Lemma 3.5 cannot remove a super hard breakpoint. The possibilities that can arise when a pair of softly connected breakpoints exists. In this case, one operation can be applied to remove at least one breakpoint. The pair of elements that are consecutive in the identity permutation is represented with a grayscale color

Lemma 3.6

Given a valid unsigned instance for the problem, if and there is no pair of softly connected breakpoints, then there must be at least one overcharged breakpoint. Assume that there are no overcharged breakpoints in I. We will show by contradiction that , which contradicts the fact that I is a valid instance. Since there is no pair of softly connected breakpoints, it follows that for each soft breakpoint , we have , where , otherwise I has at least one pair of softly connected breakpoints. Let be the set of soft breakpoints from I. We have that , which means that there are not enough nucleotides in soft breakpoints to remove all of them while not turning them into undercharged breakpoints. Besides, for each undercharged breakpoint we also have that , where . Since I has no overcharged breakpoints, it follows that , and I is not a valid instance.

Lemma 3.7

Given a valid unsigned instance for the problem, if I has only one overcharged breakpoint , one undercharged breakpoint , and there is no pair of softly connected breakpoints, then , where and . By contradiction, assume that . Since no pair of softly connected breakpoints exist in I, it follows that there are no soft breakpoints in or there are not enough nucleotides in the soft breakpoints to remove them. In both cases, moving the excess of nucleotides from the overcharged breakpoint to the undercharged breakpoint is not enough to remove two breakpoints (). So, the instance remains with at least one undercharged breakpoint and possibly with soft breakpoints with not enough nucleotides to remove them, which contradicts the fact that .

Lemma 3.8

Given an unsigned instance for the problem, if I has only one overcharged breakpoint , at least one undercharged breakpoint , and there is no pair of softly connected breakpoints, then there is a sequence of two operations that removes at least two breakpoints. By Lemma 3.7 we have that , where and . If , assume without loss of generality that . We apply a sequence of two reversals to move the exceeding nucleotides from to , and both breakpoints are removed. If , then at least a third breakpoint must exist. By Lemma 3.3, it is possible to redistribute the nucleotides within intergenic regions , , and using two consecutive transpositions. Initially, we verify if there is a soft breakpoint to receive nucleotides. Note that adding or removing nucleotides to a soft breakpoint does not turn it into a hard breakpoint. If the soft breakpoint exists, then the overcharged and undercharged breakpoints will be removed and it will receive the exceeding nucleotides after applying two consecutive transpositions. Otherwise, the third breakpoint must be an undercharged breakpoint, which can be removed or turned into an overcharged breakpoint after receiving the exceeding nucleotides. In the worst case, two breakpoints are removed after applying a sequence of two operations, and the lemma follows. Note that the sequence of operations from Lemma 3.8 generates at most one overcharged breakpoint after two consecutive transpositions, but if it occurs the instance will have no soft breakpoints.

Lemma 3.9

Given an unsigned instance for the problem such that and with no pair of softly connected breakpoints, it is possible to create a hard undercharged breakpoint keeping the instance with no pair of softly connected breakpoints, or create a super hard undercharged breakpoint after applying one operation of reversal or transposition. If there is at least one decreasing strip in , then must exist a pair of soft breakpoints and , with , such that or are consecutive in [19]. If are consecutive in , then we apply a reversal . Otherwise, we apply a reversal . Note that in both cases all the nucleotides are moved to the hard undercharged breakpoint created, which guarantees that the instance remains with no pair of softly connected breakpoints. If there is no decreasing strip in , it is always possible to find three soft breakpoints , , and , such that a transposition creates a hard undercharged breakpoint and no hard breakpoint is removed [7]. Besides, since the instance has only increasing strips, it guarantees that the hard undercharged breakpoint created (joining two increasing strips) is a super hard undercharged breakpoint, and the lemma follows.

Lemma 3.10

Given an unsigned instance for the problem such that there is only one overcharged breakpoint, no undercharged breakpoints, and there is no pair of softly connected breakpoints, then there is a sequence of at most three operations that removes at least two breakpoints or a sequence of at most four operations that removes at least three breakpoints. Note that , since it is impossible to create a valid instance with only one overcharged breakpoint and one soft breakpoint. Applying Lemma 3.9 we have two possibilities: (i) a hard undercharged breakpoint is created keeping the instance with no pair of softly connected breakpoints, then Lemma 3.8 can be applied (resulting in two breakpoints removed after applying three operations); (ii) a super hard undercharged breakpoint is created. In this case, if there are no pair of softly connected breakpoints in I, then Lemma 3.8 can be applied (also resulting in two breakpoints removed after applying three operations). Otherwise, Lemma 3.5 can be applied. Note that, by Remark 3.1, the super hard undercharged breakpoint remains untouched, and one of the following cases can occur: A new overcharged breakpoint is created, and Lemma 3.4 can be applied (three breakpoints removed after applying four operations). A pair of softly connected breakpoints is created, and Lemma 3.5 can be applied (two breakpoints removed after applying three operations). There is no pair of softly connected breakpoints in I, and Lemma 3.8 can be applied (three breakpoints removed after applying four operations).

Remark 3.2

Note that if only two breakpoints are removed by Lemma 3.10, then it implies that the resulting genome is different from . Now consider Algorithm 1, which consists of four cases depending on the number of overcharged breakpoints or the existence of a pair of softly connected breakpoints. Note that at each iteration of Algorithm 1, at least one breakpoint is removed, so eventually will be transformed into and the algorithm stops. Besides, each step is performed in linear time using the auxiliary structures of a breakpoint list and the inverse permutation of (i.e., a permutation that indicates the position of each element i in ). Since , the running time of Algorithm 1 is .

Lemma 3.11

Given an unsigned instance for the problem, Algorithm 1 transforms into using at most operations. Algorithm 1 can be analyzed considering the following cases: Note that, if the algorithm reaches cases 3 or 4, there is exactly one overcharged breakpoint. Otherwise, case 1 would be performed first or the instance is not a valid one (Lemma 3.6). I has at least two overcharged breakpoints (lines 3 to 6). I has at least one pair of softly connected breakpoints (lines 7 to 10). I has only one overcharged breakpoint, at least one undercharged breakpoint, and there is no pair of softly connected breakpoints (lines 12 to 15). I has only one overcharged breakpoint, no undercharged breakpoints, and there is no pair of softly connected breakpoints (lines 16 to 19). Cases 1, 2, and 3 remove, on average, one breakpoint per operation. If the worst case of Case 4 is performed (where two breakpoints are removed with three operations), we have by Remark 3.2 that , and cases 1, 2, or 3 will be applied subsequently, and all guarantees a sequence of operations that will remove, on average, one breakpoint per operation. Thus, on average, each breakpoint is removed by using at most operations, and the lemma follows.

Theorem 3.1

Algorithm 1 is a 4-approximation algorithm for the unsigned case of the problem. Given an unsigned instance for the problem, we have by Proposition 3.1 that . By Lemma 3.11, Algorithm 1 transforms into using at most operations. Thus, we obtain the following approximation ratio:

Incorporating generic transpositions

In this section we use a more generalized definition of transpositions and design a 3-approximation algorithm for the sorting by intergenic reversals and transpositions problem using that definition. Let us start with a formal definition of intergenic moves and generic transpositions, that include intergenic transpositions and intergenic moves.

Definition 3.1

An intergenic move , with , , , , and cuts into three parts of sizes x, and , and cuts after z nucleotides, and inserts the segment from of size into after the z-th nucleotide. This means that results in , with if , , and . Similarly, an intergenic move , with , , , , and cuts after x nucleotides, cuts into three parts of sizes y, and nucleotides, and inserts the segment from of size into after the x-th nucleotide. This means that results in , with if , , and .

Definition 3.2

A generic transposition , with , is an intergenic move (as in Definition 3.1), if or , or is an intergenic transposition (as in Definition 2.2), where . We note that an intergenic move modifies only two intergenic regions of an instance. Now we show how generic transpositions affect the number of breakpoints from an instance .

Lemma 3.12

Given an unsigned instance and a signed instance of the problem, and for any generic transposition , respectively.

Proof

The proof is similar to the one described in Lemma 3.1, and considering that an intergenic transposition can affect up to three breakpoints and an intergenic move can affect up to two breakpoints. In the following lemma we explain how to remove an overcharged breakpoint using one intergenic move.

Lemma 3.13

Given an unsigned instance for the problem, if I has one overcharged breakpoint, then it is possible to remove at least one breakpoint using an intergenic move. Let , with , be the overcharged breakpoint, and let , such that . We note that another breakpoint , with and , must exist in , otherwise the instance is not valid. We can use an intergenic move to transfer w nucleotides from to , and the overcharged breakpoint is removed. If , we can apply the intergenic move (Fig. 4, Case 1); otherwise we can apply the intergenic move (Fig. 4, Case 2).
Fig. 4

Illustration of an intergenic move applied to remove an overcharged breakpoint. The overcharged breakpoint is represented with a grayscale color. On the top (Case 1), the intergenic move is applied to move the excess of nucleotides from the overcharged breakpoint to the breakpoint , such that . Similarly, at the bottom (Case 2), the intergenic move is applied but

Illustration of an intergenic move applied to remove an overcharged breakpoint. The overcharged breakpoint is represented with a grayscale color. On the top (Case 1), the intergenic move is applied to move the excess of nucleotides from the overcharged breakpoint to the breakpoint , such that . Similarly, at the bottom (Case 2), the intergenic move is applied but

Lemma 3.14

Given a valid unsigned instance for the problem, if and there are no overcharged breakpoints, then there must be at least one pair of softly connected breakpoints. Note that, since there is no overcharged breakpoint, and , then at least two soft breakpoints must exist, otherwise the instance has only undercharged breakpoints and it is not valid. We can use a similar argument as the proof of Lemma 3.6 to show that at least one pair of soft breakpoints must be connected, otherwise I is not a valid instance. Algorithm 2 consists of two cases: one occurs when there is an overcharged breakpoint and the other is applied when there are only soft and undercharged breakpoints. At each iteration of Algorithm 2 at least one breakpoint is removed using one reversal or one generic transposition, so eventually will be transformed into and the algorithm ends. The same argument of Algorithm 1 can be used to show that the running time of Algorithm 2, which is .

Lemma 3.15

Given an unsigned instance for the problem, Algorithm 2 transforms into using at most operations. Algorithm 2 has only two cases: (i) I has at least one overcharged breakpoint (lines 3 to 6) and (ii) I has at least one pair of softly connected breakpoints (lines 7 to 10). In both cases at least one breakpoint is removed per operation, and the lemma follows.

Theorem 3.2

Algorithm 2 is a 3-approximation algorithm for the unsigned case of problem incorporating generic transpositions. Since Lemma 3.12 has the same lower bound for as in Lemma 3.2, Proposition 3.1 is also valid when considering intergenic reversals and generic transpositions. Given an unsigned instance for the problem, we have by Proposition 3.1 that . By Lemma 3.15, Algorithm 2 transforms into using at most operations, and the lemma follows.

Greedy strategy

To improve the practical performance of algorithms 1 and 2, we search at the beginning of each iteration for one of the following operations. The search is performed in linear time knowing where each element is placed in . Therefore, it does not increase the asymptotic time complexity of algorithms 1 and 2. Besides, this strategy does not affect the theoretical approximation factors of algorithms 1 and 2, since the applied operations remove at least two breakpoints each. A transposition that removes three breakpoints. A reversal or transposition that removes two breakpoints.

Approximation algorithms for the signed case of the problem

In this section, we show how to obtain approximation algorithms for the signed case of the problem based on a reduction from a signed instance into an unsigned instance. The algorithms are designed following three steps: (i) Initially, we describe a polynomial time function that maps a signed instance of the problem into a valid unsigned instance . (ii) Then, we use the algorithms 1 or 2 to provide an solution for the instance , and (iii) we show a polynomial time function that maps a solution into a valid solution S(I) for I. Lastly, we prove the theoretical approximation factor obtained by adopting this process. Function works as follows: for each element of the source genome , we map it into two new elements: ), if , and , otherwise. In both cases, a new intergenic region with size zero is inserted between these two new elements. We apply the same procedure in the target genome . This procedure doubles the size of the instance but note that , since each breakpoint type two is mapped into a breakpoint type one. Besides, the function takes linear time to complete the mapping. Function uses the fact that algorithms 1 and 2 act only over breakpoints to map a solution for into a valid solution S(I) for I. It maps each reversal in into such that and , and each transposition (or generic transposition) in into such that , , and . Recall that this mapping is only possible because algorithms 1 and 2 do not create breakpoints of type one during the process that transform the source genome into the target genome. Furthermore, note that solutions S(I) and have the same number of operations. Since solution is , where n is the number of elements of , then function takes linear time to complete the solution mapping. Figure 5 shows an example using the functions and . The signed instance of the problem (at the top) is mapped into an unsigned instance (at the bottom) using the function . Moreover, the function is used to map an solution for into a valid solution S of same size for .
Fig. 5

On top, we have a signed instance of the problem, with , , and , which is mapped by the function into an unsigned instance (at the bottom), such that , , and . The function maps a solution for the instance into a valid solution , with same size, for the instance

On top, we have a signed instance of the problem, with , , and , which is mapped by the function into an unsigned instance (at the bottom), such that , , and . The function maps a solution for the instance into a valid solution , with same size, for the instance Algorithms 3 and 4 show the steps to obtain a solution for the signed case of the problem. Note that the functions and take linear time. Thus, the running time of algorithms 3 and 4 are . Now we show that algorithms 3 and 4 guarantee the approximation factors of 4 and 3 considering the reversal and transposition events and incorporating the generic transposition, respectively.

Lemma 3.16

Given a signed instance for the problem, we have that , where . Direct by the construction of the function.

Lemma 3.17

Given a signed instance for the problem, Algorithm 3 transforms into using up to reversals and transpositions. By Lemma 3.16, we have that , where . Besides, a solution for is obtained using up to reversals and transpositions. Since a valid solution S(I) for I generated by the function has the same size as , the lemma follows.

Theorem 3.3

Algorithm 3 is a 4-approximation algorithm for the signed case of the problem. Given an signed instance for the problem, we have by Proposition 3.2 that . By Lemma 3.17, we have that Algorithm 3 transforms into using at most operations of reversal and transposition, and the theorem follows.

Lemma 3.18

Given a signed instance for the problem, Algorithm 4 transform into using at most operations of reversal and transposition. The proof is similar to the one described in Lemma 3.17 but considering that a solution for the instance is obtained using at most operations of reversal and transposition.

Theorem 3.4

Algorithm 4 is a 3-approximation algorithm for the signed case of the problem incorporating generic transpositions. Given an signed instance for the problem, we have by lemmas 3.1, 3.2, and 3.12 the following lower bound: . By Lemma 3.18, we have that Algorithm 4 transforms into using at most operations of reversal and transposition, and the theorem follows.

Practical results

In this section, we compare the proposed algorithms using simulated datasets. Besides, we perform an experiment using marine and brackish picocyanobacteria genomes from Cyanorak 2.1 [20] system.

Results with unsigned simulated datasets

To assess algorithms 1 and 2, we compare them with the 4.5-approximation algorithm for the unsigned case of the problem presented by Brito et. al. [16]. We hereafter refer to the 4.5-approximation algorithm [16], Algorithm 1, and Algorithm 2 by 4.5, 4, and 3SbIRGT, respectively. We used the following datasets of simulated genomes:The DS1 dataset explores scenarios considering instances of same size and where the number of breakpoints tends to increase as the number of random operations used to generate each instance grows. The DS2 dataset explores scenarios considering groups of instances with different sizes and, by the random process of construction, they tend to have a higher number of breakpoints. DS1: This dataset was presented by Brito et. al. [16]. It is divided into groups according to the number of random operations (reversal or transposition) used to create each instance in the dataset. Each group contains 10.000 instances of size 100. Instances are created as follows: the target genome is composed by the identity permutation and the intergenic region sizes in the target genome are randomly chosen in the range [0..100]. The source genome was obtained after applying a sequence of random operations in the target genome. The number of random operations ranged from 5 up to 100, in intervals of 5. Reversals and transpositions can be selected with the same probability to create each instance. This dataset has a total of 200.000 instances. DS2: This dataset contains groups of instances with sizes 100, 200, 300, 400, and 500. Each group contains 10.000 instances. Instances are created as follows: the target genome is again composed by the identity permutation with intergenic region sizes randomly chosen in the range [0..100]. The source genome was obtained by shuffling the lists of genes and intergenic region sizes from the target genome independently, in order to create instances with a large number of breakpoints. This dataset has a total of 50.000 instances. Tables 1, 2, and 3 consider the DS1 dataset and they use, respectively, algorithms 4.5, 4, and 3SbIRT. Columns OP, Default Implementation, and Greedy Strategy represent the number of random operations used to create the instances, the result with no greedy strategy, and the result with the greedy strategy, respectively.
Table 1

Comparison between the default implementation and the greedy strategy of the 4.5 algorithm using the DS1 dataset

OPDefault implementationGreedy strategy
DistanceApprox. ratioDistanceApprox. ratio
Min.Avg.Max.Min.Avg.Max.Min.Avg.Max.Min.Avg.Max.
0548.64151.252.013.0045.67141.001.343.00
101017.54261.502.253.00813.09251.001.693.00
151726.72381.732.463.001320.80331.151.913.00
202535.72471.862.633.081627.67401.332.032.85
253343.81562.062.743.072333.45451.392.092.79
303950.97662.322.813.072838.50531.582.122.79
354357.06702.502.863.063042.74571.582.142.82
404962.43772.592.883.103446.40601.642.142.68
455467.11802.622.903.093949.66631.712.152.76
505971.14832.672.923.093952.25691.692.142.76
556074.69892.732.933.084254.70691.652.142.67
606477.80912.722.943.084456.75701.742.142.62
656880.50932.702.943.084458.51731.742.132.59
707182.89952.792.943.104860.06741.732.132.57
757185.07962.802.953.074861.50761.762.132.57
807586.91982.802.953.075062.67761.772.122.54
857688.55992.832.953.075263.80791.772.122.55
957990.001002.812.953.075264.71781.732.122.52
907791.271002.812.963.105465.45801.782.112.58
1008392.381012.842.963.105366.23811.772.112.55
Table 2

Comparison between the default implementation and the greedy strategy of the 4 algorithm using the DS1 dataset

OPDefault implementationGreedy strategy
DistanceApprox. ratioDistanceApprox. ratio
Min.Avg.Max.Min.Avg.Max.Min.Avg.Max.Min.Avg.Max.
050510.17151.252.373.0045.61131.001.332.75
101120.54291.572.643.00812.68231.001.633.00
151929.81381.912.743.101319.94301.151.832.80
202838.11492.142.803.081626.52371.331.952.69
253345.37582.352.833.072332.16431.392.012.71
303951.91662.532.863.002737.10491.552.042.58
354557.58702.482.883.003041.29531.582.062.58
404962.69772.622.903.003444.90561.622.072.53
455467.19812.702.913.003748.07601.692.082.55
505871.11832.712.923.043950.68631.692.072.50
556174.56872.722.923.044153.07641.652.082.50
606477.64922.732.933.004355.11681.742.082.46
656980.32932.792.933.004456.85711.722.072.44
707182.68962.792.943.044658.43721.732.072.46
757284.84952.792.943.044759.83721.762.072.48
807586.70982.802.943.004961.01731.772.072.46
857688.30992.812.943.005162.11761.742.072.43
957889.75992.832.953.035163.06751.732.072.45
907891.001002.832.953.005263.80771.732.062.45
1008292.13992.832.953.035464.54761.752.062.43
Table 3

Comparison between the default implementation and the greedy strategy of the 3SbIRGT algorithm using the DS1 dataset

OPDefault implementationGreedy strategy
DistanceApprox. ratioDistanceApprox. ratio
Min.Avg.Max.Min.Avg.Max.Min.Avg.Max.Min.Avg.Max.
0559.89141.252.302.8045.46121.001.292.75
101120.19281.572.592.89712.34221.001.592.83
151929.41381.912.702.921319.53301.151.802.70
202737.68482.142.772.941626.06361.271.912.67
253344.94572.292.802.952231.66421.391.982.64
304051.47662.502.832.952736.57481.552.012.53
354457.13702.482.852.963040.75531.582.042.53
404962.24772.642.872.963444.33561.642.042.53
455466.73802.682.882.963747.54601.672.052.50
505970.66822.672.892.963950.12631.682.052.48
556174.10872.712.902.974152.52651.622.052.50
606377.17902.762.912.974354.56671.702.052.46
656879.85922.792.912.974456.31701.702.052.44
707182.21952.782.912.974657.84711.692.052.46
757184.38942.792.922.974759.26721.722.052.43
807486.21972.802.922.974960.43721.702.052.43
857587.83982.772.922.975161.54751.732.052.42
957889.28982.812.932.975162.48741.712.052.41
907790.541002.832.932.975263.21761.732.042.42
1008291.65992.802.932.975263.95761.722.042.40
From Table 1, we note that the greedy strategy significantly improved the results of the 4.5 algorithm. The minimum, average, and maximum metrics for the distance and the approximation ratio using the greedy strategy presented lower values when compared with the algorithm default implementation, except for the minimum distance when OP . The average approximation ratio tends to increase as OP increases. When no greedy strategy is applied, the values ranged from 2.01 (OP ) to 2.96 (OP ). Using the greedy strategy the values ranged from 1.34 (OP ) to 2.11 (OP ). Besides, by adopting the greedy strategy we were able to find at least one optimal solution in the groups where OP and OP , indicated by the minimum approximation ratio column with value 1.00. Table 2 shows a similar behavior for 4 regarding the increase of the average approximation ratio as OP grows, and the improvement obtained by the greedy strategy. Using no greedy strategy, the average distance of 4 is better than the average distance of 4.5 algorithm when the number of random operations (OP) is greater than or equals to 50. It indicates that the default implementation of the 4 algorithm tends to provide better results when the instance has many breakpoints. When we compare both algorithms using the greedy strategy, the 4 algorithm provides better results for the vast majority of the groups and metrics. Considering all groups and using the greedy strategy, the maximum approximation ratio obtained by both algorithms (4.5 and 4) was 3.00, which is considerably less than the theoretical approximation factor proven for them. Table 3 shows that 3SbIRGT provided results similar to those presented by 4. Considering the average distance and average approximation ratio columns, we can see a slight improvement for all values of OP compared with the practical results of 4. This fact results from the inclusion of the intergenic move operation, which can reduce the number of operations needed to transform a genome into another. Besides, considering the versions without and with the greedy strategy, respectively, the maximum approximation ratios regarding all groups were 2.97 and 2.83. Using the greedy strategy, the average approximation ratio of 3SbIRGT ranged from 1.29 to 2.05, which is significantly less than the theoretical approximation factor. Table 4 shows the results for the DS2 dataset using 4.5, 4, and 3SbIRGT. The average distances of the algorithms without greedy strategy were close to the instance sizes in all groups. Computing the absolute difference between the average distance and the instance sizes, the highest values provided by the 4.5, 4, and 3SbIRGT algorithms were 4.00 (Size=500), 0.42 (Size=500), and 0.08 (Size=100), respectively. The greedy strategy also led to important improvement of the results for all the algorithms and groups. With and without greedy strategy, the best results were provided by 3SbIRGT followed by the 4 and 4.5 algorithms regarding the average distance and average approximation ratio metrics.
Table 4

Results of the 4.5, 4, and 3SbIRGT algorithms considering the default implementation and the greedy strategy using the DS2 dataset

SizeDefault ImplementationGreedy strategy
DistanceApprox. ratioDistanceApprox. ratio
Min.Avg.Max.Min.Avg.Max.Min.Avg.Max.Min.Avg.Max.
4.5SbIRT
10099103.471122.913.043.296071.77851.762.112.50
200199203.732152.973.033.21122137.831571.822.052.34
300299303.893152.963.003.12185202.532221.832.002.21
400399403.974122.983.013.07245266.512911.831.982.17
500499504.005132.993.013.07307330.053561.841.972.13
4SbIRT
10098100.381022.882.953.006070.02811.762.062.38
200198200.412022.962.993.01121135.481521.812.022.27
300298300.413022.952.973.01181199.642191.791.972.17
400397400.434022.972.983.00241263.212901.801.962.16
500498500.425022.982.993.01303326.293521.811.952.11
3SbIRT
1009799.921002.882.932.975969.45811.742.042.38
200197199.932002.942.982.99120134.891501.792.012.24
300298299.933002.952.972.99182199.072181.801.972.17
400397399.934002.972.982.99243262.572851.811.962.13
500498499.945002.982.982.99304325.683501.821.952.10
Table 5 shows the average running time, in seconds, of the 4.5, 4, and 3SbIRGT algorithms per instance, comparing the default implementation (DI) and the greedy strategy (GS) using the DS2 dataset. Note that the greedy strategy is more time-consuming than the default implementation. The maximum average running time of an algorithm without greedy strategy was less than 0.20 seconds, while using the greedy strategy it was 0.65 seconds. Observing the improvement in the results given by the greedy strategy in Table 4, we highlight that the additional running time is a good trade-off regarding running time and solution quality.
Table 5

The average running time of the 4.5, 4, and 3SbIRGT algorithms, in seconds, considering the default implementation (DI) and the greedy strategy (GS) using the DS2 dataset

Size4.5\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\textsc {SbIRT}}$$\end{document}SBIRT4\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\textsc {SbIRT}}$$\end{document}SBIRT3SbIRGT
DIGSDIGSDIGS
1000.010.030.010.030.010.02
2000.030.100.030.100.030.10
3000.060.230.060.230.060.22
4000.130.400.120.410.130.43
5000.180.630.190.640.180.64
Based on the results, the practical approximation ratio provided by the algorithms tends to be better than the theoretical approximation factors. Besides, it is noteworthy that the greedy strategy has brought a significant improvement on both datasets. Since incorporating this strategy does not change the asymptotic time complexity nor the theoretical approximation of the algorithms, it becomes an excellent alternative to obtain better results.

Results with signed simulated datasets

To assess algorithms 3 and 4, we compare them with the 3-approximation and the 2.5-approximation algorithms for the signed case of the problem, respectively, which were presented by Oliveira et. al. [17]. We hereafter refer to the 3-approximation algorithm [17], 2.5-approximation algorithm [17], Algorithm 3, and Algorithm 4 by 3, 2.5, 4, and 3, respectively. The results of the 4 and 3 algorithms were obtained adopting the greedy strategy. We used the and datasets presented by Oliveira et. al. [17], and they have the following characteristics: Each dataset started with 100 target genomes , such that has 100 elements, and each value of , with , being chosen randomly and uniformly in interval [0..100]. After that, from each source genome were generated 100 instances by applying:The parameters of each applied operation were randomly generated considering the range of valid values. The value of d ranged from 10 up to 100, in intervals of 10. For each value of d, a group with 10,000 instances was generated. DBSIRIT and DBSIRGT datasets have a total of 100,000 instances each. DBSIRIT: d random operations of reversals and transpositions (being 50% of each) in each source genome . DBSIRGT: d random operations of reversals and generic transpositions (being 50% of reversals, 40% of transpositions, and 10% of moves) in each source genome . Tables 6 and 7 show the practical results of the algorithms using the DBSIRIT and DBSIRGT datasets, respectively. The approximation ratio for each instance was computed using the lower bound based on the weighted cycle graph structure [17, Theorems 3.8 and 7.6].
Table 6

Comparison between the and algorithms using the DBSIRIT dataset

d\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3{\textsc {SbI}}\overline{\text {R}}{\textsc {T}}{}$$\end{document}3SBIR¯T\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$4{\textsc {SbI}}\overline{\text {R}}{\textsc {T}}{}$$\end{document}4SBIR¯TSmall (%)Small or equal (%)
Distance (Avg.)Approx. ratio (Avg.)Distance (Avg.)Approx. ratio (Avg.)
1012.601.6812.191.6351.3768.70
2026.371.7725.601.7154.7266.48
3040.231.8136.911.6678.6085.53
4053.291.8546.101.6095.8497.62
5063.781.8753.211.5699.4699.74
6071.881.8958.971.5599.8199.92
7077.831.9063.291.5599.9799.98
8082.451.9166.731.5599.9599.99
9085.981.9269.481.5599.9799.99
10088.781.9371.641.5699.9999.99
Table 7

Comparison between the 2.5 and 3 algorithms using the DBSIRGT dataset

d2.5\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\textsc {SbI}}\overline{\text {R}}{\textsc {GT}}$$\end{document}SBIR¯GT3\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\textsc {SbI}}\overline{\text {R}}{\textsc {GT}}$$\end{document}SBIR¯GTSmall (%)Small or equal (%)
Distance (avg.)Approx. ratio (avg.)Distance (avg.)Approx. ratio (avg.)
1011.791.5712.251.6432.3053.18
2024.971.6825.571.7234.7749.58
3038.211.7436.691.6763.3173.75
4050.471.7945.611.6289.6093.69
5060.461.8152.651.5897.6998.75
6068.171.8358.191.5699.4599.78
7074.061.8462.541.5599.7599.90
8078.681.8565.961.5599.9399.96
9082.091.8568.671.5599.9299.97
10084.851.8670.821.5599.97100.00
Table 6 compares the results obtained by the and algorithms. The columns Small and Small or Equal indicate, for each group, the percentage of solutions provided by the algorithm with strictly smaller size and with small or equal size, respectively, when compared to the solutions provided by the algorithm. From Table 6, it is possible to observe that the algorithm, in all the groups, was able to provide better results considering the metrics of average approximation ratio and average distance. Besides, considering the groups with d greater than 20, the algorithm provided better solutions in more than 75% of the instances (column Small). Considering the groups with d greater than 30, the algorithm provided better or equivalent size solutions (column Small or Equal) in more than 96% of the instances. It is important to note that, as the value of d increases, the absolute difference between the average distance provided by the and algorithms also increases significantly. When d is greater than 50, the absolute difference between the average distances is superior to 10, which indicates that the algorithm tends to provide better solutions in scenarios where a higher number of operations were used. Table 7 compares the results obtained by the 2.5 and 3 algorithms. The columns Small and Small or Equal indicate, for each group, the percentage of solutions provided by the 3 algorithm with strictly smaller size and with small or equal size, respectively, when compared to the solutions provided by the 2.5 algorithm. From Table 7, we can note that the 2.5 algorithm, when compared to the 3SbIRGT algorithm, showed a slightly better result regarding the average approximation ratio and distance in the groups with and . Considering these two groups ( and ), the absolute difference between the average distance provided by the algorithms was less than 0.61. Besides, by column Small, we can notice that in the groups and the 3 algorithm provided better solutions in 32.30% and 34.77% of the instances, respectively. This shows that the 3 algorithm can act in a complementary way with the 2.5 algorithm, even in the cases where both provide similar results. Since better estimates tend to outcome in enhanced analysis, selecting the better result between each algorithm is a good alternative to assist in this task. Regarding the groups where d is greater than 20, the 3 algorithm provided better results considering the average approximation ratio and distance. Furthermore, in the same groups, the 3 algorithm provided better or equivalent size solutions (column Small or Equal) in more than 73% of the instances. From Tables 6 and 7, it is possible to note that and 3 algorithms are robust and tend to provide practical results better than the theoretical bounds.

Results with real genomes

To assess the 3 algorithm and analyze the behavior with real genomes, we compared it with the 2-approximation algorithm for the problem considering reversals and transpositions on signed permutations (ignoring the intergenic regions), presented by Walter et al. [7]. We hereafter refer to the 2-approximation algorithm [7] by 2sb t. We used 97 genomes from Cyanorak 2.1 [20], which is a system for the visualization and curation of marine and brackish picocyanobacteria genomes. The system encompasses 51 synechococcus, 3 cyanobium, 41 prochlorococcus genomes, and 2 prochlorococcus metagenome-assembled genomes. For each genome, the number of genes ranged from 1834 to 4391, and replicated genes correspond to less than 5% of the total genes, on average. We performed a preprocessing stage to ensure that the data fits the model constraints, which is divided in two steps: After the preprocessing stage, we obtained for each pairing an instance . Note that the 2SbRT algorithm requires as input only the permutation , since it was not designed to consider the intergenic regions. Map the sequence of genes and the intergenic regions into : For each genome, we mapped the first occurrence of the genes into a permutation and computed the size of the intergenic regions to obtain . Pairing: For each pair of genomes, we performed a pairing so that the genes and conserved blocks shared by both genomes were kept while the remaining genes were removed through a process that simulates a sequence of deletions. Finally, 3 with the greedy strategy and the 2Sb t were applied to each pairing. The number of genome rearrangement events for each pairing was computed by the total of deletions used in the preprocessing stage (step 2) plus the size of the sequence of reversals and (generic) transpositions provided by the algorithms. These numbers were fed into a matrix of pairwise distances. We constructed two phylogenetic trees based on the matrix of pairwise distances computed from the algorithms and using the Circular Order Reconstruction method [21]. To analyze the topological characteristics of the phylogenetic trees, we performed a comparison with the phylogenetic tree presented by Laurence et al. [20] using a tool [22] based on the maximum agreement subtrees (MAST) to determine the topological congruence between two phylogenetic trees. Table 8 shows the obtained results.
Table 8

Analysis of the topological characteristics of the phylogenetic trees generated by the results of the 2Sb t and 3 algorithms compared with the phylogenetic tree presented by Laurence et al. [20]

MAST\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$I_{cong}$$\end{document}IcongP-value
2sb\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\text {R}}$$\end{document}R¯ t46 Leaves3.176.76e−22
3\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\textsc {SbI}}\overline{\text {R}}{\textsc {GT}}$$\end{document}SBIR¯GT51 Leaves3.522.77e−25
Table 8 indicates that both phylogenetic tree have a high concordance with the phylogenetic trees presented by Laurence et al. [20], with the phylogenetic tree obtained from the 3 algorithm providing a MAST with more leaves and consequently a better value for and P-value. It is important to mention that the objective of this experiment using real genomes is to demonstrate the applicability of our algorithm, which considers the information regarding the genes and the size of the intergenic regions, compared with a similar model that considers only the order and orientation of the genes. We used the same data preprocessing stage and reconstruction method to provide a fair comparison. However, the results may differ especially considering genomes with different characteristics and the adopted reconstruction method. Figure 6 shows a phylogenetic tree constructed using the Circular Order Reconstruction method [21] with the matrix of pairwise distances from the 3 algorithm.
Fig. 6

Phylogenetic tree based on genome rearrangements using the 3 algorithm with the greedy strategy and 97 genomes from the Cyanorak 2.1 system

Phylogenetic tree based on genome rearrangements using the 3 algorithm with the greedy strategy and 97 genomes from the Cyanorak 2.1 system Comparison between the default implementation and the greedy strategy of the 4.5 algorithm using the DS1 dataset Comparison between the default implementation and the greedy strategy of the 4 algorithm using the DS1 dataset Comparison between the default implementation and the greedy strategy of the 3SbIRGT algorithm using the DS1 dataset Results of the 4.5, 4, and 3SbIRGT algorithms considering the default implementation and the greedy strategy using the DS2 dataset The average running time of the 4.5, 4, and 3SbIRGT algorithms, in seconds, considering the default implementation (DI) and the greedy strategy (GS) using the DS2 dataset Comparison between the and algorithms using the DBSIRIT dataset Comparison between the 2.5 and 3 algorithms using the DBSIRGT dataset Analysis of the topological characteristics of the phylogenetic trees generated by the results of the 2Sb t and 3 algorithms compared with the phylogenetic tree presented by Laurence et al. [20] From Fig. 6 (created using treeio R package [23]), we observe that the approach separates the organisms considering the species and performed good groupings. It is worth mentioning that the tree was based exclusively on rearrangement event information.

Conclusion

We studied the sorting by intergenic reversals and transpositions problem on signed and unsigned permutations. We presented, for both cases, a 4-approximation algorithm, improving the 4.5 approximation factor previously known for the unsigned case. Besides, we generalized the transposition event and presented a 3-approximation algorithm to the problem that arises, which is more realistic in scenarios that consider intergenic regions. We developed a greedy strategy to improve the practical performance of the algorithms and conducted a comparison using datasets with different features. Considering the signed case of the problem, the tests indicated that our algorithms, in the vast majority of the cases, tend to provide better practical results compared with the previous known results. Moreover, we carried out an experiment using real genomes to verify the applicability of the proposed algorithms. From the theoretical point of view, the algorithms proposed for the unsigned case of the sorting by intergenic reversals and transpositions problem bring an important improvement considering the approximation factor. On the other hand, the results for the signed case of the problem have the practical potential of enhancing the estimates for the distance of compared genomes, and consequently, the analysis regarding the genome rearrangements. In future works, one can incorporate non-conservative events (e.g., insertion and deletion of genes or nucleotides) into the model.
  12 in total

1.  A 1.375-approximation algorithm for sorting by transpositions.

Authors:  Isaac Elias; Tzvika Hartman
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2006 Oct-Dec       Impact factor: 3.710

2.  A congruence index for testing topological similarity between trees.

Authors:  Damien M de Vienne; Tatiana Giraud; Olivier C Martin
Journal:  Bioinformatics       Date:  2007-10-12       Impact factor: 6.937

3.  On the Complexity of Sorting by Reversals and Transpositions Problems.

Authors:  Andre Rodrigues Oliveira; Klairton Lima Brito; Ulisses Dias; Zanoni Dias
Journal:  J Comput Biol       Date:  2019-05-23       Impact factor: 1.479

4.  Sorting by Genome Rearrangements on Both Gene Order and Intergenic Sizes.

Authors:  Klairton Lima Brito; Géraldine Jean; Guillaume Fertin; Andre Rodrigues Oliveira; Ulisses Dias; Zanoni Dias
Journal:  J Comput Biol       Date:  2019-12-18       Impact factor: 1.479

5.  Treeio: An R Package for Phylogenetic Tree Input and Output with Richly Annotated and Associated Data.

Authors:  Li-Gen Wang; Tommy Tsan-Yuk Lam; Shuangbin Xu; Zehan Dai; Lang Zhou; Tingze Feng; Pingfan Guo; Casey W Dunn; Bradley R Jones; Tyler Bradley; Huachen Zhu; Yi Guan; Yong Jiang; Guangchuang Yu
Journal:  Mol Biol Evol       Date:  2020-02-01       Impact factor: 16.240

6.  Algorithms for computing the double cut and join distance on both gene order and intergenic sizes.

Authors:  Guillaume Fertin; Géraldine Jean; Eric Tannier
Journal:  Algorithms Mol Biol       Date:  2017-06-05       Impact factor: 1.405

7.  Sorting Signed Permutations by Intergenic Reversals.

Authors:  Andre Rodrigues Oliveira; Geraldine Jean; Guillaume Fertin; Klairton Lima Brito; Laurent Bulteau; Ulisses Dias; Zanoni Dias
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2021-12-08       Impact factor: 3.710

8.  Breaking Good: Accounting for Fragility of Genomic Regions in Rearrangement Distance Estimation.

Authors:  Priscila Biller; Laurent Guéguen; Carole Knibbe; Eric Tannier
Journal:  Genome Biol Evol       Date:  2016-05-22       Impact factor: 3.416

9.  Super short operations on both gene order and intergenic sizes.

Authors:  Andre R Oliveira; Géraldine Jean; Guillaume Fertin; Ulisses Dias; Zanoni Dias
Journal:  Algorithms Mol Biol       Date:  2019-11-05       Impact factor: 1.405

10.  Cyanorak v2.1: a scalable information system dedicated to the visualization and expert curation of marine and brackish picocyanobacteria genomes.

Authors:  Laurence Garczarek; Ulysse Guyet; Hugo Doré; Gregory K Farrant; Mark Hoebeke; Loraine Brillet-Guéguen; Antoine Bisch; Mathilde Ferrieux; Jukka Siltanen; Erwan Corre; Gildas Le Corguillé; Morgane Ratin; Frances D Pitt; Martin Ostrowski; Maël Conan; Anne Siegel; Karine Labadie; Jean-Marc Aury; Patrick Wincker; David J Scanlan; Frédéric Partensky
Journal:  Nucleic Acids Res       Date:  2021-01-08       Impact factor: 16.971

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.