Background

DNA motifs are short sequences in the genome that play important functional roles in gene regulation. Due to their short length, it is difficult to identify these regions using features intrinsic in their composition. Assuming that the motifs are conserved in closely related species due to the importance of their function, it is possible to discover them by comparing the respective DNA sequences to identify the sub-sequences that are very similar to each other.

There are two common combinatorial formulations that identify the motifs: The first is the consensus motif problem which made its first appearance in 1984 [1], while the second is the planted (l, d)-motif problem that was presented in 2000 [2]. It is worth noting that the latter formulation is a special case of the former. The exact definitions are as follows:

Given a set of t sequences s i where 1 ≤ it defined over an alphabet ∑. The consensus motif problem is to find an l-length motif sequence M such that in each sequence s i , 1 ≤ it, there is at least one subsequence p i differing with at most d mismatches from M; i.e., d H (p i , M)d, where d H is the hamming distance between p i and M.

The planted (l, d) motif problem is a special case of the consensus problem in which we restrict that p i occurs only once in s i .

Due to its combinatorial nature, the consensus motif problem and its variant defined above is extremely challenging. Over a benchmark data of 20 sequences, each of length 600 characters, large instances of (15, 5), (17, 6), (19, 7) and (21, 8) have been addressed and many algorithms have been developed to solve them one after another. These algorithms can be classified into two major categories: approximation algorithms [212] and exact algorithms [1330]. Approximation algorithms are based on probabilistic local search techniques, such as Gibbs Sampling, Expectation Maximization, etc. Although these algorithms may solve the challenging instances in practice, there is no guarantee that the motif can be found even when l is short.

Exact algorithms are based on exhaustive search techniques. The brute force algorithm proceeds by testing all possible motifs of length l using pattern matching, leading to O (l n t 4l) time complexity. This algorithm, however, is not suitable for discovering long motifs in practice, and many algorithms have been developed to provide faster solutions. Examples of these algorithms are CENSUS [23], PMS1 [26], PMSP [27], PMSprune [29], PMS5 [30], SMILE [19], RISO [24], RISOTTO [28], and Voting [25]. In the following we briefly review the most efficient ones and the ones related to our work.

The algorithms SMILE [19], RISO [24], and RISOTTO [28] are based on the use of suffix tree. The time complexity of these algorithms is the same and it is O(t2Nv(l, d)), where v ( l , d ) = d i = 0 C i l 3 i is the size of the d-mismatch neighbourhood of motifs of length l and N = t i = 1 n i , n i is the length of sequence i from input sequences. RISOTTO improved the time complexity of SMILE and RISO in the average case and solved some challenging instances such as (15, 5) and (17, 6).

PMSP [27] is based on exploring the neighbourhood of the l-mer of the first sequence and checking whether the elements of such neighbourhoods are (l, d) motifs. The time complexity is O ( 1 w t n 2 v ( l , d ) ) . It is able to solve some challenging instances such as (15, 5) and (17, 6). PMSprune [29, 31] is an improved version of the PMSP algorithm, based on the branch and bound strategy. Although it has the same worst-case time complexity as PMSP algorithm, it is more efficient in practice and it could tackle the (17, 6) and (19, 7) instances for the first time. PMS5 [30] is based on computing the common d-neighbourhood of three l-mers using integer programming formulation. It combines this novel idea with the algorithms PMS1 and PMSPrune. PMS5 can tackle the large challenging instances (19, 7), (21, 8) and (23, 9). The only drawback of PMS5, it requires larger amount of internal memory to finish computation.

Our contribution

In a previous work [32, 33], we have introduced an idea composed of two stages to speed up the exact algorithms: In the first stage, we generate a set of candidate motifs by applying one of the exact algorithms based on the neighbourhood method (like Voting [25] or PMSP [27] algorithms) using qt sequences. In the second stage, for each candidate motif we check if it is a valid motif or not using pattern matching on the reminder (t - q) sequences. This dramatically reduces the search space and leads to significant speed up. The bottleneck in this approach, however, was the determination of the q value that yields the fastest running time. That is, the user has to guess the value of q, which might lead to non-optimal running time and even no speed up compared to the traditional methods. Also, the authors in [34] have used the same idea on PMS1, RISOTTO, and PMSprune algorithms.

In this paper, we present a theoretical method which can be used to determine the appropriate value of q. Then we apply this strategy on PMSprune algorithm and solve some big challenging instances such as (21, 8). Furthermore, we propose a parallel version of our algorithm to present a practical solution to the challenging instances of the motif problem. Our parallel version further speeds up the solution of the (21, 8) instance.

Definitions and related work

In this section, we introduce some notations and definitions that will help us to describe our algorithm and related work in a concise manner.

Definition 1 adapted from [29]: For any string x, with |x| = l, let B d (x) = {y: |y| = l, d H (y, x) ≤ d}, where d H denotes the Hamming distance and B d (x) denotes the set of neighbourhoods of x. We also write v(l, d) to refer to |B d (x)|.

Definition 2 adapted from [29]: Let s denote a string of length n and let x denote another string of length l, l <n. We define the minimum distance between s and x as d ̄ H ( x , s ) = min x l s d H ( x , x ) , where x l s denotes that x is a substring of s with length l.

Definition 3 adapted from [29]: Given an l-length string x and a set of strings S = {s1, ..., s t } with |s i | = n for i = 1, ..., t and l <n, we define the distance between S and x as d ̄ H ( x , S ) = max i = 1 t { d ̄ H ( x , s i ) } = max i = 1 t { min r l s i { d H ( x , r ) } }

Definition 4 adapted from [29]: A string x is an (l, d) motif for a set of sequences S = {s1, ..., s t }, if:

  1. 1)

    d ̄ H ( x , S ) d .

  2. 2)

    y l s 1 : x B d ( y ) d ̄ H ( x , { s 1 , . . . . , s t } ) d .

Proposition 1 adapted from [10]: Let u and v be two random strings of length l over an alphabet of 4 characters with equal probability of occurrence. The probability p d that d H (u, v) ≤ d is p d = i = 0 d l i 3 / 4 i 1 / 4 l - i , and the probability that d ̄ H ( x , S ) d is (1-(1-p d )n-l+1)t. The expected number of l-length motifs that occur at least once in each of the t sequences with up to d substitutions is E(l, d, t, n) = 4l(1-(1-p d )n-l+1)t.

PMSprune Algorithm

Because the first stage of our method will depend on the PMSprune algorithm. We will review the basic steps of it in the notions presented above.

The main strategy of PMSprune is to generate B d (y), for every l-mer y in s1, using a branch and bound technique. An element xB d (y) is a motif only if d ¯ H x , S d. The step of verifying that d ¯ H x , S d is achieved by scanning all substrings of S. For fixed values of t, n, and l, the expected time complexity of PMSprune is equal to

T P M S p r u n e = O t n - l + 1 2 l + p 2 d i = 1 2 d - d + 1 l i 3 i
(1)

where p2dis the probability that the hamming distance between two strings is at most 2d, and it is defined in Proposition 1. For fixed values of t, n, and l, value d' was estimated such that the probability of d ̄ H ( x , S ) d is close to 1. (The probability of d ̄ H ( x , S ) d is given in Proposition 1 and it is 1 - 1 - p d n - l + 1 t ).

Implementation

Our proposed strategy

Our new strategy, referred to as hybrid exact pattern motif search (HEP), is composed of three steps: first, we determine the value q, corresponding to the size of a subset of input sequences, as explained below. Second, we apply an exact exhaustive algorithm £ (like, PMSprune) on the set of q sequences to find the set of d-neighbourhood B d (x) (review definition 1 for exact definition of d-neighbourhood). We call this set the candidate motif set. Finally, we apply a pattern search algorithm over the remaining sequences to verify each motif. Note that our algorithm is generic in the sense that it takes the program £ also as input in addition to the input sequences and user parameters. A pseudo code for this strategy using the exact algorithm £ is as follows:

Algorithm 2: HEP (£, s1,..., s t , n, l, d)

Begin

  1. 1)

    Determine the number of sequences q using the method given below.

  2. 2)

    Implement the exact algorithm £ on q input sequences. Let C be the set of candidate motifs found in the q sequences.

  3. 3)

    For each pattern v in C, check if v is a valid motif or not in the reminder (t - q) input sequences using pattern matching Algorithm.

End.

Theorem 1: Algorithm 2 correctly finds all (l, d) motifs in a given t input sequences.

Proof: Step 2 of the algorithm is exhaustive and finds the whole set of d-neighborhood for the q sequences. Therefore, and by definition of the (l, d) motif problem, any (l, d) motif belongs to this set, even if q = 1. In Step 3, each candidate motif is verified by comparison to each substring in the remaining sequences. This step is conducted by an approximate pattern matching algorithm for each l-length substring in the candidate motif set and each l-length substring in the remaining sequences such that the hamming distance between these two substrings is ≤ d. This guarantees that no motif is missing.

Theorem 2: The running time of the HEP is equal to

T H E P = T £ ( q ) + l ( t - q ) ( n - l + 1 ) E ( l , d , q , n )
(2)

where T£(q)is the running time of step 2 involving the use of an exact algorithm £ on the q input sequences and l(t - q) (n - l + 1) E(l, d, q, n) is the running time of step 3 such that E(l, d, q, n) is the number of elements in the set C, which is estimated to be 4l(1- (1 - p d )n - l + 1)q. Note that the complexity of step 1 takes constant time, as we will explain below. Note that the running time of the brute force algorithm is acquired if q = 0 in equation 2. The running time of the exact algorithm £ is acquired if q = t in equation 2.

Determination of the best q

The range of the number of sequences q, enhancing the performance of the exact motif finding problem is calculated by solving the following inequality for the parameter q:

T H E P T £
(3)

Definition 5: We define mns as the minimum number of sequences q that yields better running time; i.e., the first value of q that verifies the inequality. We also define ons as the optimal number of sequences q that yields the best running time; i.e., the value of q such that T HEP is minimum over 1 ≤ qt.

Implementing HEP based on PMSprune

We decided to use PMSprune for implementing the first step in our method, because of its superiority compared to other algorithms as discussed in [31]. However, we stress that our approach is generic and can be used with any better algorithm that appears in future. In the following, we will refer to our method based on PMSprune as HEP_PMSprune. If q = mns we will denote it with HEP_PMSprune(mns), and if q = ons we will denote it with HEP_PMSprune(ons).

Determining mns for PMSprune

Replacing T£(q)by the time of PMSprune on q sequences, Equations (1) and (2) can be rewritten as follows:

T P M S p r u n e = q ( n - l + 1 ) 2 ( l + p 2 d i = 1 2 d - d + 1 l i 3 i ) + ( t - q ) ( n - l + 1 ) 2 ( l + p 2 d i = 1 2 d - d + 1 l i 3 i )
T H E P _ P M S p r u n e = q ( n - l + 1 ) 2 ( l + p 2 d i = 1 2 d - d + 1 l i 3 i ) + l ( t - q ) ( n - l + 1 ) E ( l , d , q , n )

Replacing T HEP with T HEP_PMSprune and T£ with T PMSprune in the inequality (3)results in the following variation:

l ( t - q ) ( n - l + 1 ) E ( l , d , q , n ) < ( t - q ) ( n - l + 1 ) 2 ( l + p 2 d i = 1 2 d - d + 1 l i 3 i )

Substituting the value of E(l, d, q, n) with the value given in Proposition 1 in the left hand side yields

4 l ( 1 - ( 1 - p d ) n - l + 1 ) q < ( n - l + 1 ) ( l + p 2 d i = 1 2 d - d + 1 l i 3 i ) l

Dividing both sides by 4l and taking the logarithm,

q > log ( n - l + 1 ) ( l + p 2 d i = 1 2 d - d + 1 l i 3 i ) - log ( l 4 l ) log ( 1 - ( 1 - p d ) n - l + 1 )
(4)

The inequality (4) provides the range of the values of q that makes the running time of HEP using PMSprune less than the running time of the original PMSprune over the all set of sequences. The minimum value of q in the range of the inequality is called mns and it is equal to:

log ( n - l + 1 ) ( l + p 2 d i = 1 2 d - d + 1 l i 3 i ) - log ( l 4 l ) log ( 1 - ( 1 - p d ) n - l + 1 ) + 1

Determining ons for PMSprune

For fixed values of t, n, l and d, ons can be calculated for PMSprune by selecting the value of q that minimizes the total number of operations THEP _PMSprunefor 1 ≤ qt. The following algorithm computes the value of ons for each instance (l, d).

Algorithm 3: Find ons

Begin

  1. 1)

    q = ons = 1

  2. 2)

    E ( l , d , q , n ) = 4 l ( 1 - ( 1 - ( i = 0 d l i ( 3 / 4 ) i ( 1 / 4 ) l - i ) ) n - l + 1 ) q

  3. 3)

    T min = q ( n - l + 1 ) 2 ( l + p 2 d i = 1 2 d - d + 1 l i 3 i ) +l ( t - q ) ( n - l + 1 ) E ( l , d , q , n )

  4. 4)

    for q = mns to t do

    E ( l , d , q , n ) = 4 l ( 1 - ( 1 - ( i = 0 d l i ( 3 / 4 ) i ( 1 / 4 ) l - i ) ) n - l + 1 ) q
    T = q ( n - l + 1 ) 2 ( l + p 2 d i = 1 2 d - d + 1 l i 3 i ) +l ( t - q ) ( n - l + 1 ) E ( l , d , q , n )

if T <Tmin then

Tmin = T

ons = q

  1. 5)

    return ons

End

The above algorithm computes q in O(t) time. In practice, the time for computing q takes negligible time with respect to the rest of motif finding steps; it took maximum one second for all experiments included in this paper with simulated and real datasets. To save some time, our implementation includes a look-up table containing pre-computed values of q for different values of l, n, and d, where l < 20, d < 3, and selected values of n with n = 300, n = 350, 400, ..., n = 700. For other values of l, n, and d, we compute the best q using the above algorithm.

Parallel version of HEP_PMSprune(ons)

We propose a parallel version for HEP_PMSprune(ons) called PHEP_PMSprune(ons). The two main steps of HEP_PMSprune(ons) can be parallelized as follows:

We parallelize the PMSprune algorithm by assigning a set of l-mers from s1 to each processor for establishing the set of neighboring motifs. The resulting sets are stored in candidate motif lists C i , i ∈ {1, 2, ..., p}, where p is the number of processors. After each processor finishes computation, the C i lists are merged together in a larger set C, such that each motif is represented once in this list; i.e., all repetitions are removed. Creating the C list is done in linear time with respect to the number of candidate motifs and it is achieved as follows:

We incrementally construct the partial list C j that contains the L j lists, 1 ≤ j ≤ p, by appending the list L j at the end of the list Cj- 1such that all elements in L j existing in Cj-1are discarded. This continues until j = p; i.e., C p is C. Discarding a repeated element is done efficiently as follows: For small values of l, we create a look-up table with size Σl, where Σ is the alphabet size. Each possible l-length string can be mapped to a number in the range between zero and Σl in O(l) time. The ith entry in this table contains one if a string in Cj- 1is mapped to i. Otherwise, it contains zero. The strings in C j are queried against this look-up table to discard repetitions and set entries they are mapped to with value one. For longer values of l, we use the Aho-Corasick automaton to index all l-length motifs in Cj- 1, and check if a strings in C j exists in the automaton or not and add the new strings of C j to the automaton. For these string matching algorithms, we refer the reader to [35].

In the second step, we validate each candidate motif independently in parallel over the available processors. The running time of this algorithm is O(Ts/p +|C|), where T s is the sequential running time and |C| is the size of set C.

The first step in the parallel algorithm does not lead to loss of any motifs. This is because the set C includes the d-neighborhood set of the q-sequences. The reason is that we run PMSprune in parallel against the strings (x, s2, s3, ..., s q ), where x is a substring of s1. That is, each substring is not processed. The second step in the parallel algorithm is also correct, because the elements in C are independent of each other and checking the validity of each candidate motif can be safely run in parallel. Our experimental results confirm the correctness of our parallelization procedure.

Results and discussion

Experiments on simulated datasets

We used the simulated data sets that are used in many articles [2530, 3234] with t = 20 sequences and n = 600 characters, where the alphabet size is 4. Each (l, d) input instance dataset is generated as follows: We generate random strings with length (n-l) each, where the characters appear randomly with equal probability. Then we generate randomly an l-length string M and plant a copy of it in each sequence at random position after mutating it with at most d random mutations. We tested the algorithms for varying n, l, and d values and for the following challenging instances: (11, 3), (13, 4), (15, 5), (17, 6), (19, 7), and (21, 8).

Experiments overview

Our experiments address three major issues: The first is the performance of our method compared to the use of PMSprune only. The second, we show that our method for selecting q, already achieves the best running time. The third is the performance of the parallel version and its scalability. The algorithms are implemented on a 2 Quad-core processors (2.5 GHz each) machine. The programs are coded in C language. In the parallel version, we use openMP directives for parallelizing the code.

Performance of HEP on PMSprune

Tables 1 and 2 show the performance of the algorithms HEP_PMSprune(mns) and HEP_PMSprune(ons) with respect to PMSprune algorithm respectively. The last column in Tables 1 and 2 displays the improvement in PMSprune which equals to T P M S p r u n e - T H E P _ P M S p r u n e ( m n s ) T P M S p r u n e and T P M S p r u n e - T H E P _ P M S p r u n e ( o n s ) T P M S p r u n e respectively. We used the notations 's', 'm', 'h', and 'dy' in computing the time for seconds, minutes, hours, and days, respectively. The results confirm that, the algorithms HEP_PMSprune(mns) and HEP_PMSprune(ons) significantly reduced the running time compared to the standard PMSprune algorithm in all challenging instances.

Table 1 Time Comparison of PMSPrune and HEP_PMSprune(mns) with the Challenging Instances
Table 2 Time Comparison of PMSPrune and HEP_PMSprune(ons) with the Challenging Instances

Evaluating the choice of q

In this section, we experimentally evaluate our algorithm for determining the best q that minimizes the running time of the HEP_PMSprune(q) algorithm. To achieve this, we will follow the following steps:

1. We run HEP_PMSprune(q), mnsqt for the problem instances (11, 3), (13, 4), (15, 5), (17, 6), (19, 7), and (21, 8) and determine the value of q that minimizes the running time; we will refer to this value with onsexp.

2. Compare the onsexp against our ons computed theoretically.

Figure 1, which plots the running time against different q values, shows the results of applying these steps. We observe the value of ons is equal or very close to the value of onsexp.

Figure 1
figure 1

Performance of our method for different challenging instances. Behavior of HEP_PMSprune(q) for different (l, d) instances such that q ∈{mns,..., t}. (a): (11, 3), (b): (13, 4), (c): (15, 5), (d): (17, 6), (e): (19, 7). We used the following remarks in the figures: 1) Black-triangle-down to indicate the runing time of HEP_PMSprune(mns). 2) Black-star to indicate the running time of PMSprune or HEP_PMSprune(t). 3) White-box to indicate the running time of HEP_PMSprune(ons); i.e., using theoretically estimated q.

We also conducted another experiment, where the problem instances were generated with different n and l and d. Table 3 shows the results for many of these instances, where the number of sequences t = 20. We can observe that our algorithm finds the optimal q in all these instances. We also observe improvement of the running time with respect to the PMSprune algorithm in most of the cases. The cases with no improvement in the running time are attributed to the fact that the expected number of motifs is very low and the original algorithm runs already fast in these cases.

Table 3 The performance of the HEP_PMSprune(ons) for different values of n and l

Note that it was not feasible to list the results for all possible values n, l, and d in Table 3. But in other instances with different values of n, l, and d, we found that ons and its time were consistent with onsexp and its time published in this table.

Performance of PHEP_PMSprune(ons)algorithm

In Table 4, we show the results of applying the parallel version of our algorithm PHEP_PMSprune(ons) using different number of processors and for different problem instances. The running time of the difficult instance (21, 8) has been decreased from 6.68 days to about 20.42 hours using 8 processors. Figure 2 shows the scalability results for the algorithm where speedup = T H E P P M S p r u n e ( o n s ) T H E P P M S p r u n e ( o n s ) . From Table 4 and Figure 2 we note that PHEP_PMSprune(ons) reduce the time of HEP_PMSprune(ons) and the speedup achieved scales well with the increasing number of processors.

Table 4 Running time of PHEP_PMSprune(ons) using different number of processors p for some challenging instances
Figure 2
figure 2

Scalability plot of the parallel version. The plots show speed-up for different number of processors and problem instances.

Experiments on real datasets

We used two collections of real datasets used in previous research papers [10, 26, 29, 36]. The first collection is a dataset including a number of the upstream regions of yeast genes [37] affected by certain transcription factors. The transcription factors are from the SCPD [38] database and the paper [39]. The upstream DNA sequences were extracted using the Saccharomyces Genome Database [37]. The second collection includes the dataset of Blanchette [36] which includes the upstream DNA regions of many genes from different species. This dataset is available at http://bio.cs.washington.edu/supplements/FootPrinter and a copy of it is available with our software tool for testing.

Tables 5 and 6 show the motifs found by our method compared to the published ones for both collections. In each table, we give a reference to the published motif. Our program could detect all published motifs. It is also interesting to note that our program could detect extra novel motifs in the case of the Interleukin-3 problem instance in Table 6. These motifs look interesting, because they are 20 bp long with hamming distance zero; an observation that calls for further biological investigation.

Table 5 Application of the PHEP_PMSprune(ons) on the real yeast dataset
Table 6 Application of the PHEP_PMSprune(ons) on the Blanchette real dataset

Tables 5 and 6 also include the running times (in seconds) of running our method for the listed problem instances and the improvement in time compared to the PMSprune method. The running time for one problem instance is the time needed to run our program in the (l, d) parameters range from (6, 0) until (21, 3), i.e., there are 64 invocations of our program. The results show that our program is superior to the PMSprune for large instances.

Conclusions

In this paper, we introduced an efficient method that can enhance the performance of exact algorithms for the motif finding problem. Our method depends on dividing the sequence space into two sets. Over the first set, we generate a set of candidate motifs. Then, we use the remaining set of sequences to verify if each candidate motif is a real one. The experimental results show that our method is superior to the best methods available so far and could tackle large problems like (21, 8). Finally, we introduced a scalable and efficient parallel version for the proposed method. Our tool is available for free for academic research at http://www.nubios.nileu.edu.eg/tools/hymotif.

Availability and requirements

Project name: hymotif.

Project home page: http://www.nubios.nileu.edu.eg/tools/hymotif

Operating system(s): Linux.

Programming language: C.

Other requirements: C/C++ libraries.

License: GPL.

Any restrictions to use by non-academics: No restrictions.