# Towards a Runtime Comparison of Natural and Artificial Evolution

## Abstract

Evolutionary algorithms (EAs) form a popular optimisation paradigm inspired by natural evolution. In recent years the field of evolutionary computation has developed a rigorous analytical theory to analyse the runtimes of EAs on many illustrative problems. Here we apply this theory to a simple model of natural evolution. In the Strong Selection Weak Mutation (SSWM) evolutionary regime the time between occurrences of new mutations is much longer than the time it takes for a mutated genotype to take over the population. In this situation, the population only contains copies of one genotype and evolution can be modelled as a stochastic process evolving one genotype by means of mutation and selection between the resident and the mutated genotype. The probability of accepting the mutated genotype then depends on the change in fitness. We study this process, SSWM, from an algorithmic perspective, quantifying its expected optimisation time for various parameters and investigating differences to a similar evolutionary algorithm, the well-known (1+1) EA. We show that SSWM can have a moderate advantage over the (1+1) EA at crossing fitness valleys and study an example where SSWM outperforms the (1+1) EA by taking advantage of information on the fitness gradient.

### Keywords

Runtime analysis Evolutionary algorithms Natural evolution Population genetics Theory Strong selection weak mutation regime## 1 Introduction

Evolutionary algorithms are popular general-purpose heuristics that have found countless applications in design and optimisation. For many large-scale problems they provide good results in reasonable time where many exact techniques fail [6]. They are a method of choice in black-box optimisation, where no information about the problem at hand is available, and the only way to obtain knowledge is to evaluate candidate solutions.

In the last 20 years evolutionary computation has developed a number of algorithmic techniques for the analysis of evolutionary and genetic algorithms. These methods typically focus on rigorously bounding the expected time required to reach a global optimum, or other well-specified high-fitness solutions. Problems studied include illustrative example functions from pseudo-Boolean optimisation that isolate characteristics found in more complex problems as well as problems from combinatorial optimisation. Various studies showed that the simple evolutionary algorithm called (1+1) EA evolving a single search point (see Sect. 2) finds global optima across a range of combinatorial optimisation problems in polynomial expected time, including shortest paths [29], sorting (as maximising sortedness) [29], minimum spanning trees [18], matroid optimisation problems [26], and Eulerian cycles [16]. The (1+1) EA (with restarts) further constitutes a polynomial-time randomised approximation scheme for the NP-hard Partition problem [35].

The runtime analysis of evolutionary algorithms has become one of the dominant concepts in evolutionary computation, leading to a plethora of results for evolutionary algorithms [1, 10, 20] as well as novel optimisation paradigms such as swarm intelligence [19, 20, 31] and artificial immune systems [11].

Interestingly, although evolutionary algorithms are heavily inspired by natural evolution, these methods have seldom been applied to natural evolution as studied in mathematical population genetics. This is a missed opportunity: the time it takes for a natural population to reach a fitness peak is an important question for the study of natural evolution. The kinds of results obtained from runtime analysis, namely how the runtime scales with genome size and mutation rate, are of general interest to population genetics. Moreover, recently there has been a renewed interest in applying computer science methods to problems in evolutionary biology with contributions from unlikely fields such as game theory [2], machine learning [33] and Markov chain theory [3]. Here, we present a first attempt at applying runtime analysis to the so-called Strong Selection Weak Mutation regime of natural populations.

*N*reflects the size of the underlying population (not to be confused with the population size in population-based EAs) and \(\beta \) represents the selection strength. One can think of

*f*as defining a phenotype that is under selection to be maximized; \(\beta \) quantifies how strongly a unit change in

*f*is favoured. This probability was first derived by Kimura [14] for a population of

*N*individuals that are sampled binomially in proportion to their fitness.

This choice of acceptance function introduces two main differences to the (1+1) EA. First, while the (1+1) EA never decreases its current fitness (a property called *elitism*), SSWM accepts solutions of lower fitness (worsenings) with some positive probability. This is reminiscent of the Metropolis algorithm (Simulated Annealing with constant temperature) which can also accept worsenings (see e. g. [12]). Second, and in contrast to the Metropolis algorithm, solutions of higher fitness can be rejected, since they are accepted with a probability that is roughly proportional to the relative advantage they have over the current solution.

We cast this model of natural evolution as a stochastic process, referred to as SSWM, using common mutation operators from evolutionary algorithms. We then present first runtime analyses of this process using techniques from the analysis of randomised and evolutionary algorithms.

to explore the performance of natural evolution in the context of runtime, comparing it against simple evolutionary algorithms like the (1+1) EA,

to investigate the non-elitistic selection mechanism implicit to SSWM and its usefulness in the context of evolutionary algorithms, and

to show that techniques for the analysis of evolutionary algorithms can be applied to simple models of natural evolution, aiming to open up a new research field at the intersection of evolutionary computation and population genetics (see [24] for recent work on unifying these two fields).

*n*being the problem size, SSWM is an effective hill climber as it optimises OneMax in expected time \(\mathord {O}\mathord {\left( (n \log n)\left( 1+\frac{1}{2\beta }\right) \right) }\). However, when \(2N\beta \) is smaller than \(\ln n\) by any constant factor, we encounter a phase transition and SSWM requires exponential time even on OneMax.

We then illustrate the particular features of the selection rule in more depth. In Sect. 4, we consider a function \({\textsc {Cliff}}_{d} \) where a fitness valley of Hamming distance *d* needs to be crossed in order to reach the global optimum. For \(d = \omega (\log n)\) the (1+1) EA needs time \(\varTheta (n^d)\), but SSWM is faster by a factor of \(e^{\varOmega (d)}\) because of its ability to accept worse solutions. Finally, in Sect. 5 we illustrate on the function Balance [27] that SSWM can drastically outperform the (1+1) EA because the fitness-dependent selection drives it to follow the steepest gradient. While the (1+1) EA needs exponential time in expectation with \(\varOmega (1)\) probability, SSWM with overwhelming probability finds an optimum in polynomial time.

The main technical difficulties are that in contrast to the simple (1+1) EA, SSWM is a non-elitist algorithm, hence fitness-level arguments based on elitism are not applicable. Popular techniques such as level-based theorems for non-elitist populations [4] are not applicable either because they require population sizes larger than 1. Moreover, while for the (1+1) EA transition probabilities to better solutions are solely determined by probabilities for flipping bits during mutation, for SSWM these additionally depend on the fitness difference. The analysis of SSWM is more challenging than the analysis of the (1+1) EA, and requires tailored proof techniques. We hope that these techniques will be helpful for analysing other evolutionary algorithms with fitness-based selection schemes.

## 2 Preliminaries

We define the optimisation time of SSWM as the number of the first iteration at which the optimum is accepted as a new individual.

*probability of fixation*, that is, the probability that a gene initially present in one copy in a population of

*N*individuals is eventually present in all individuals.

Since the acceptance function in this algorithm depends on the difference in fitness between genotypes, we include a parameter \(\beta > 0\) that effectively scales the fitness function and that in population genetics models the strength of selection on a phenotype. By incorporating \(\beta \) as a parameter of this function (and hence of the algorithm) we avoid having to explicitly rescale the fitness functions we analyse, while allowing us to explore the performance of this algorithm on a family of functions.

*N*.

*global mutations*) or flipping a single bit chosen uniformly at random (which we call

*local mutations*). The SSWM model is a good approximation of natural evolution when the expected number of new mutants in the population is much less than one, which implies that local mutations are a better approximation for this regime. However, we also consider global mutations in order to facilitate a comparison with evolutionary algorithms such as the (1+1) EA (Algorithm 2), which uses global mutations. Most of our analyses apply to both mutation operators; the analysis in Sect. 4 applies to global mutations only.

Next, we derive upper and lower bounds for \(p_{\mathrm {fix}}(\Delta f) \) that will be useful throughout the manuscript. The bounds for \(\Delta f>0\) show that \(p_{\mathrm {fix}}\) is roughly proportional to the fitness difference between solutions \(\beta \Delta f\).

### Lemma 1

### Proof

In the following we frequently use \(1 + x \le e^x\) and \(1-e^{-x}\le 1\) for all \(x \in {\mathbb {R}}\) as well as \(e^x \le \frac{1}{1-x}\) for \(x < 1\).

The next lemma shows that the probability of accepting an improvement of \(\Delta f\) is exponentially larger (in \(N\beta \Delta f\)) than accepting its symmetric fitness variation \(-\Delta f\).

### Lemma 2

### Proof

## 3 SSWM on OneMax

The function \({\textsc {OneMax}} (x) := \sum _{i=1}^n x_i\) has been studied extensively in natural computation because of its simplicity. It represents an easy hill climbing task, and it is the easiest function with a unique optimum for all evolutionary algorithms that only use standard bit mutation for variation [30]. Showing that SSWM can optimise OneMax efficiently serves as proof of concept that SSWM is a reasonable optimiser. It further sheds light on how to set algorithmic parameters such as the selection strength \(\beta \) and the population size *N*. To this end, we first show a polynomial upper bound for the runtime of SSWM on OneMax for a selection strength of \(2( N - 1) \beta \ge \ln ( cn )\). We then show that SSWM exhibits a phase transition on its runtime as a function of \(2N\beta \); decreasing this parameter by a constant factor below \(\ln n\) leads to exponential runtimes on OneMax.

Another reason why studying OneMax for SSWM makes sense is because not all evolutionary algorithms that use a fitness-dependent selection perform well on OneMax. Neumann et al. [17] as well as Oliveto and Witt [23] showed that evolutionary algorithms using fitness-proportional selection, including the Simple Genetic Algorithm, fail badly on OneMax even within exponential time, with very high probability.

### 3.1 Upper Bound for SSWM on OneMax

We first show the following simple lemma, which gives an upper bound on the probability of increasing or decreasing the number of ones in a search point by *k* in one mutation.

### Lemma 3

*i*ones creates an offspring with \(i\pm k\) ones. Then

The proof can be found in the appendix; it uses arguments from the proof of Lemma 2 in [30]. The second inequality follows immediately from the first one due to the symmetry \({{\mathrm{mut}}}(i,i-k)={{\mathrm{mut}}}(n-i,n-i+k)\).

Now we introduce the concept of drift and find some bounds for its forward and backward expression.

### Definition 1

### Lemma 4

### Proof

*j*is now the number of new zeros. Bounding \({{\mathrm{mut}}}(i,i-j)\) from above by Lemma 3 and bounding \(i/n \le 1-1/n\) yields

Finally, the case for local mutations is straightforward since the probability of a local mutation increasing the number of ones is \(\frac{n-i}{n}\) and that of decreasing it is at most 1. \(\square \)

The following theorem shows that SSWM is efficient on OneMax whenever \({2(N-1) \beta } \ge \ln (cn)\) for some constant \(c>1.2\), since then \(p_{\mathrm {fix}}(1)\) starts being greater than \({n\cdot p_{\mathrm {fix}}(-1)}\), allowing for a positive drift even on the hardest fitness level (\(n-1\) ones).

### Theorem 5

The factor \(\frac{1}{p_{\mathrm {fix}}(1)} \le 1+\frac{1}{2\beta }\) (by Lemma ) on the runtime bound represents the extra time paid due to the probability of rejecting a better search point. For small selection strength \(\beta \ll 1\) the upper bound essentially increases with \(1/(2\beta )\). This makes sense as for small \(\beta \) (and \(N\beta \gg 0\)) we have \(p_{\mathrm {fix}}(1) \approx 2\beta \) (cf. Lemma ). In this regime absolute fitness differences are small and improvements are only accepted with a small probability.

### Proof of Theorem 5

We only give a proof for global mutations here. The analysis for local mutations follows the same way, with simpler calculations, and without the factor “*e*” in the running time bound.

*O*(

*n*) bound is trivial) we obtain

*h*(

*z*) then the expected optimisation time, starting in state \(X_0\), is bounded as

*z*is the number of zeros \((n-i)\),

*T*is the optimisation time and \(z_{\min }\) is the smallest positive state. Here we have \(z_{\min }=1\), \(X_0 \le n\) and by using

### 3.2 A Critical Threshold for SSWM on OneMax

The upper bound from Theorem 5 required \(2(N-1) \beta \ge \ln (cn)\), or equivalently, \(2N\beta \ge \ln (n) + \ln (c) + 2\beta \). This condition is vital since if \(N \beta \) is chosen too small, the runtime of SSWM on OneMax is exponential with very high probability, as we show next.

If \(2N \beta \) is smaller than \(\ln (n)\) by a factor of \(1-\varepsilon \), for some constant \(\varepsilon > 0\), the optimisation time is exponential in *n*, with overwhelming probability. SSWM therefore exhibits a phase transition behaviour: changing \(N \beta \) by a constant factor makes a difference between polynomial and exponential expected optimisation times on OneMax.

### Theorem 6

If \(2 \le 2N\beta \le (1-\varepsilon ) \ln n\) for some \({0< \varepsilon < 1}\), then the optimisation time of SSWM with local or global mutations on OneMax is at least \(2^{c n^{\varepsilon /2}}\) with probability \({1-2^{-\varOmega (n^{\varepsilon /2})}}\), for some constant \(c > 0\).

The condition \(N\beta \ge 1\) is used to ease the presentation; it is not essential and we believe it can be dropped when using more detailed calculations. The idea behind the proof of Theorem 6 is to show that for all search points with at least \(n - n^{\varepsilon /2}\) ones, there is a negative drift for the number of ones. This is because for small \(N \beta \) the selection pressure is too weak, and worsenings in fitness are more likely than steps where mutation leads the algorithm closer to the optimum.

We then use the negative drift theorem with self-loops presented in Rowe and Sudholt [28] (an extension of the negative drift theorem [22] to stochastic processes with large self-loop probabilities). It is stated in the following for the sake of completeness. The theorem uses “\(p_{k, k \pm d} \le x\)” as a shorthand for “\(p_{k, k+d} \le x\) and \(p_{k, k-d} \le x\)”.

### Theorem 7

*a*,

*b*with \(0< a < b \le m\) and \(\varepsilon > 0\) such that for all \(a \le k \le b\) the drift towards 0 is

*k*. Further assume there exist constants \(r, \delta > 0\) (i. e. they are independent of

*m*) such that for all \(k \ge 1\) and all \(d \ge 1\)

*T*be the first hitting time of a state at most

*a*, starting from \(X_0 \ge b\). Let \(\ell = b-a\). Then there is a constant \(c > 0\) such that

### Proof of Theorem 6

We only give a proof for global mutations; the same analysis goes through for local mutations with similar, but simpler calculations.

The drift theorem, Theorem 7, will be applied to the number of zeros at the current point in time as distance function to the optimum, when the number of zeros is in the interval \([0, n^{\varepsilon /2}]\). By Chernoff bounds, SSWM starts with a fitness of at most \(n-n^{\varepsilon /2}\) with probability \(1-2^{-\varOmega (n)}\). We assume in the following that this happens; the claimed probability bound then follows from a union bound of the failure probability \(2^{-\varOmega (n)}\) and a failure probability \(2^{-\varOmega (n^{\varepsilon /2})}\) which will result from an application of the drift theorem.

Let \(p_{k, j}\) be the probability that SSWM will make a transition from a search point with *k* ones to one with *j* ones. Note that, although the drift theorem applies to the number of zeros, our notation of transition probabilities \(p_{k, j}\) refers to numbers of ones for simplicity and consistency with other parts of the paper. Throughout the remainder of the proof we assume \(k \ge n - n^{\varepsilon /2}\).

*n*, for positive \(x \le 0.09\) we have the inequality \(\sum _{j=1}^\infty j^2 \cdot x^j = \frac{x(1+x)}{(1-x)^3} \le x (1+ 5x)\), which together with \(N\beta \ge 1\) implies

*k*, denoted \(\Delta (k)\), is hence at most

*k*and all \(j \in \mathbb {N}\),

*n*is large enough)

*n*and for \(j > 1\) we additionally use \(j \cdot \left( n^{\varepsilon /2-1}\right) ^{j-1} \le 2^{-j+1}\). This proves the second condition (5) for \(\delta := 1\) and \(r := 2\). Applying the drift theorem completes the proof. \(\square \)

We remark that for many evolutionary algorithms, such as the (1+1) EA and the (1+\(\lambda \)) EA, a lower bound on OneMax transfers to all functions with a unique global optimum. The reason is that in these cases OneMax is an easiest function amongst those with a single optimum [30]. This generalisation does not apply to SSWM. In fact, OneMax is probably not the easiest function with a single global optimum, even when the fitness range is normalised to [0, *n*]. The reason here is that acceptance is determined by absolute fitness differences, and a convex fitness curve (beautifully illustrated in [32, Fig. 1]) might amplify fitness differences close to the optimum, in order to compensate for small probabilities of mutation creating fitness improvements. We leave the discovery of an easiest function for SSWM (with a unique optimum) as an open topic for future work.

## 4 On Traversing Fitness Valleys

We have shown that with the right parameters, SSWM is an efficient hill climber. On the other hand, in contrast to the (1+1) EA, SSWM can accept worse solutions with a probability that depends on the magnitude of the fitness decrease. This is reminiscent of the Metropolis algorithm—although the latter accepts every improvement with probability 1, whereas SSWM may reject improvements.

Jansen and Wegener [12] compared the ability of the (1+1) EA and a Metropolis algorithm in crossing fitness valleys and found that both showed similar performance on *smooth integer* functions: functions where two Hamming neighbours have a fitness difference of at most 1 [12, Sect. 6].

*d*and to traverse a fitness valley of Hamming distance

*d*to the optimum (see Fig. 3). Note that

*d*may depend on

*n*.

### Definition 2

The (1+1) EA typically optimises \({\textsc {Cliff}}_{d} \) through a direct jump from the top of the cliff to the optimum, which takes expected time \(\varTheta (n^d)\).

### Theorem 8

The expected optimisation time of the (1+1) EA on \({\textsc {Cliff}}_{d} \), for \(2 \le d \le n/2\), is \(\varTheta (n^d)\).

In order to prove Theorem 8, the following lemma will be useful for showing that the top of the cliff is reached with good probability. More generally, it shows that the conditional probability of increasing the number of ones in a search point to *j*, given it is increased to some value of *j* or higher, is at least 1 / 2.

### Lemma 9

The proof of this lemma is presented in the appendix.

### Proof of Theorem 8

From any search point with \(i < n-d\) ones, the probability of reaching a search point with higher fitness is at least \(\frac{n-i}{en}\). The expected time for accepting a search point with at least \(n-d\) ones is at most \(\sum _{i=0}^{n-d-1} \frac{en}{n-i} = O(n \log n)\). Note that this is \(O(n^d)\) since \(d \ge 2\).

We claim that with probability \(\varOmega (1)\), the first such search point has \(n-d\) ones: with probability at least 1 / 2 the initial search point will have at most \(n-d\) ones. Invoking Lemma 9 with \(j := n-d\), with probability at least 1 / 2 the top of the cliff is reached before any other search point with at least \(n-d\) ones.

Once on the top of the cliff the algorithm has to jump directly to the optimum to overcome it. The probability of such a jump is \(\frac{1}{n^d} \left( 1-\frac{1}{n}\right) ^{n-d}\) and therefore the expected time to make this jump is \(\varTheta (n^d)\). \(\square \)

SSWM with global mutations also has an opportunity to make a direct jump to the optimum. However, compared to the (1+1) EA its performance slightly improves when considering shorter jumps and accepting a search point of inferior fitness. The following theorem shows that for large enough cliffs, \(d = \omega (\log n)\), the expected optimisation time is by a factor of \(e^{\varOmega (d)} = n^{\omega (1)}\) smaller than that of the (1+1) EA. Although both algorithms need a long time for large *d*, the speedup of SSWM is significant for large *d*.

### Theorem 10

The expected optimisation time of SSWM with global mutations and \(\beta =1, N = \frac{1}{2}\ln (9n)\) on \({\textsc {Cliff}}_{d} \) with \(d = \omega (\log n)\) is at most \(n^{d}/e^{\varOmega (d)}\).

### Proof

We define *R* as the expected time for reaching a search point with either \(n-d\) or *n* ones, when starting with a worst possible non-optimal search point. Let \(T_{{\mathrm {cliff}}}\) be the random optimisation time when starting with any search point of \(n-d\) ones, hereinafter called the *top of the cliff* or a *local peak*. Then the expected optimisation time from any initial point is at most \(R + \text {E}\left( T_{{\mathrm {cliff}}}\right) \).

*trial*. After the end of a trial, taking at most

*R*expected generations, with probability \(1-p_\mathrm {success}\) SSWM returns to the top of the cliff again, so

*t*, then the conditional probability of moving to a local peak—when from a search point with \(i < n-d\) ones either a local peak or a non-optimal search point in \(S_2\) is reached—is

The remainder of the proof now shows a lower bound on \(p_\mathrm {success}\), the probability of a trial being successful. A sufficient condition for a successful trial is that the following events occur: the next mutation creates a search point with \(n-d+k\) ones, for some integer \(1 \le k \le d\) chosen later, this point is accepted, and from there the global optimum is reached before returning to the top of the cliff.

We estimate the probabilities for these events separately in order to get an overall lower bound on the probability of a trial being successful.

*k*that have \(n-d+k\) ones. Considering only such mutations, the probability of a mutation increasing the number of ones from \(n-d\) by

*k*is at least

*k*:

*b*, a state

*a*or less is unlikely to be reached in polynomial time.

We first show that the drift is typically equal to that on OneMax. For every search point with more than *a* ones, in order to reach \(S_1\), at least *k* / 2 bits have to flip. Until this happens, SSWM behaves like on OneMax and hence reaches either a global optimum or a point in \(S_1\) in expected time \(O(n \log n)\). The probability for a mutation flipping at least *k* / 2 bits is at most \(1/(k/2)! = (\log n)^{-\varOmega (\log n)} = n^{-\varOmega (\log \log n)}\), so the probability that this happens in expected time \(O(n \log n)\) is still \(n^{-\varOmega (\log \log n)}\).

*a*ones. From the proof of Theorem 5 and (3) we know that the drift at

*i*ones for \(\beta =1\) is at least

*i*ones to one with

*j*ones. The probability of decreasing the current state is at most \(p_{\mathrm {fix}}(-1) = O(1/n)\) due to Lemma . The probability of increasing the current state is at most \((n-i)/n\) as a necessary condition is that one out of \(n-i\) zeros needs to flip. Hence for \(i \le b\), which implies \(n-i = \omega (1)\), the self-loop probability is at least

*i*to \(i+j\), \(j \ge 1\), from Lemma 3 and (11):

*n*large enough,

*a*or less in \(n^{\omega (1)}\) steps is \(n^{-\omega (1)}\).

This implies that following a length-*k* jump, a trial is successful with probability \(1-n^{-\omega (1)}\). This establishes \(p_\mathrm {success}:= \mathord {\varOmega }\mathord {\left( n^{-d+1/2} \cdot \left( \frac{5}{4}\right) ^{d}\right) }\). Plugging this into (10), adding time *R* for the time to reach the top of the cliff initially, and using that \(O(n^{1/2}\log n) \cdot (4/5)^d = e^{-\varOmega (d)}\) for \(d = \omega (\log n)\) yields the claimed bound. \(\square \)

## 5 SSWM Outperforms (1+1) EA on Balance

Finally, we investigate a feature that distinguishes SSWM from the (1+1) EA as well as the Metropolis algorithm: the fact that larger improvements are more likely to be accepted than smaller improvements.

To this end, we consider the function Balance, originally introduced by Rohlfshagen et al. [27] as an example where rapid dynamic changes in dynamic optimisation can be beneficial. The function has also been studied in the context of stochastic ageing by Oliveto and Sudholt [21] and it goes back to an earlier idea by Witt [36].

In its static (non-dynamic) form, Balance can be illustrated by a two-dimensional plane, whose coordinates are determined by the number of leading ones (LO) in the first half of the bit string, and the number of ones in the second half, respectively. The former has a steeper gradient than the latter, as the leading ones part is weighted by a factor of *n* in the fitness (see Fig. 4).

### Definition 3

The function is constructed in such a way that all points with a maximum number of leading ones are global optima, whereas increasing the number of ones in the second half beyond a threshold of 7*n* / 16 (or decreasing it below a symmetric threshold of *n* / 16) leads to a trap, a region of local optima that is hard to escape from.

Rohlfshagen et al. [27, Theorem 3] showed the following lower bound for the (1+1) EA. The statement is specialised to non-dynamic optimisation and slightly strengthened by using a statement from their proof.

### Theorem 11

([27]) With probability \(\varOmega (1)\) the (1+1) EA on Balance reaches a trap, and then needs at least \(n^{\sqrt{n}}\) further generations in expectation to find an optimum from there. The expected optimisation time of the (1+1) EA is thus \(\varOmega (n^{\sqrt{n}})\).

We believe that the probability bound \(\varOmega (1)\) can be strengthened to \(1-e^{-\varOmega (n^{1/2})}\) with a more detailed analysis, which would show that the (1+1) EA gets trapped with an overwhelming probability.

We next show that SSWM with high probability finds an optimum in polynomial time. For appropriately small \(\beta \) we have sufficiently many successes on the LO-part such that we find an optimum before the OneMax-part reaches the region of local optima. This is because for small \(\beta \) the probability of accepting small improvements is small. The fact that SSWM for \(\beta <1\) is slower than the (1+1) EA on OneMax by a factor of \(O(1/\beta )\) turns into an advantage over the (1+1) EA on Balance.

The following lemma shows that SSWM effectively uses elitist selection on the LO-part of the function in a sense that every decrease is rejected with overwhelming probability.

### Lemma 12

For every \(x = ab\) with \(n/16< |b|_1 < 7n/16\) and \(\beta = n^{-3/2}\) and \(N \beta = \ln n\), the probability of SSWM with local or global mutations accepting a mutant \(x'=a'b'\) with \({\textsc {LO}}(a') < {\textsc {LO}}(a) \) and \(n/16< |b'|_1 < 7n/16\) is \(O(n^{-n})\).

### Proof

The following lemma establishes the optimisation time of the SSWM algorithm on either the OneMax or the LO-part of Balance.

For global mutations we restrict our considerations to *relevant steps*, defined as steps where no leading ones in the first half of the bit string is flipped. The probability of a relevant step is always at least \((1-1/n)^{n/2} \approx e^{-1/2}\). When using local mutations, all steps are defined as relevant.

### Lemma 13

### Proof

We use the method of typical runs [34]: we consider the typical behaviour of the algorithm, and show that events where the algorithm deviates from a typical run are very unlikely. A union bound over all such failure events proves the claimed probability bound.

Consider a relevant step, implying that global mutations will leave all leading ones intact. With probability 1 / *n* a local or global mutation will flip the first 0-bit. This increases the fitness by \(k \cdot n - \Delta _{\mathrm {OM}}\), where \(\Delta _{\mathrm {OM}}\) is the difference in the OneMax-value of *b* caused by this mutation and *k* is the number of consecutive 1-bits following the first 0-bit, after mutation. The latter bits are called *free riders* and it is well known (see [15, Lemma 1 and proof of Theorem 2]) that the number of free riders follows a geometric distribution with parameter 1 / 2, only capped by the number of bits to the end of the bit string *a*.

The probability of flipping at least \(\sqrt{n}\) bits in one global mutation is at most \(1/(\sqrt{n})! = e^{-\varOmega (\sqrt{n})}\) and the probability that this happens at least once in *T* relevant steps is still of the same order (using that \(T = \text {poly}\left( n\right) \) as \(p_{\mathrm {fix}}(n-\sqrt{n}) \ge 1/N \ge 1/\text {poly}\left( n\right) \)). We assume in the following that this does not happen, which allows us to assume \(\Delta _{\mathrm {OM}} \le \sqrt{n}\). We also assume that the number of leading ones is never decreased during non-relevant steps as the probability of accepting such a fitness decrease is \(O(n^{-n})\) by Lemma 12 and the expected number of non-relevant steps before *T* relevant steps have occurred is *O*(*T*).

Now, a LO-value of *n* / 2 is reached if (event *A*) in *T* relevant steps at least \(n/4 + n^{3/4}/8\) improvements happen and if (event *B*) the first \(n/4 + n^{3/4}/8\) improvements lead to a total of at least \(n/4 - n^{3/4}/8\) free riders (unless the number of leading ones hits *n* / 2). Note that these two events are independent, as improvements are due to the current mutation and the number of free riders is due to the uniform random distribution of bits following the first 0-bit [15, Lemma 1].

By Chernoff bounds [5] , the probability that the typical event *A* does not occur, that is, less than \(n/4 + n^{3/4}/8\) improvements happen, is \(e^{-\varOmega (n^{1/2})}\). For the same reason also the probability of event *B* not occurring is \(e^{-\varOmega (n^{1/2})}\). Taking the union bound over all rare failure probabilities proves the claim. \(\square \)

We now show that the OneMax part is not optimised before the LO part.

### Lemma 14

Let \(\beta = n^{-3/2}\), \(N \beta = \ln n\), and *T* be as in Lemma 13. The probability that SSWM starting with \(a_0b_0\) such that \(n/4 \le |b_0|_1 \le n/4 + n^{3/4}\) creates a search point *ab* with \(|b|_1 \le n/16\) or \(|b|_1 \ge 7n/16\) in *T* relevant steps is \(e^{-\varOmega (n^{1/2})}\).

It will become obvious that in *T* relevant steps SSWM typically makes a progress of *O*(*n*) on the OneMax part. The proof of Lemma 14 requires a careful and delicate analysis to show that the constant factors are small enough such that the stated thresholds for \(|b|_1\) are not surpassed.

### Proof of Lemma 14

We only prove that a search point with \(|b|_1 \ge 7n/16\) is unlikely to be reached with the claimed probability. The probability for reaching a search point with \(|b|_1 \le n/16\) is clearly no larger, and a union bound for these two events leads to a factor of 2 absorbed in the asymptotic notation.

The proof is divided into two parts: we first estimate the increase of \(|b|_1\) in steps where the number of leading ones in *a* does not change. We refer to these as *regular steps*. Steps where mutation increases the number of leading ones in *a* are called *special steps*; during these steps every mutation of *b* is accepted as the fitness gain through additional leading ones guarantees that any change in *b* will be accepted. In the following, we first show that the progress in \(|b|_1\) in regular steps is close to 1.14*n* / 9 and then we show that the progress in special steps is bounded by \(O(n^{3/4})\) with high probability.

*Bounding the progress in regular steps:*note that for \(\beta = n^{-3/2}\) we have

*improving*if the number of ones in

*b*increases and the step is accepted.

*k*by

*j*, adapting Lemma 3 to a string of length

*n*/ 2, is at most

*j*, given an improving step, is then

Now, by Chernoff bounds, the probability of having more than \(S := (1+n^{-1/4}) \cdot p^+ \cdot T\) improving steps in *T* relevant steps is \(e^{-\varOmega (n^{1/2})}\). Using a Chernoff bound for geometric random variables [5, Theorem 1.14], the probability of *S* improving steps yielding a total progress of at least \({(1+n^{-1/4}) \cdot 4/3 \cdot S}\) is \(e^{-\varOmega (n^{1/2})}\).

*Bounding the progress in special steps:*we also have at most

*n*/ 2 steps where the number of leading ones increases. If the number of leading ones increases by \(\delta \ge 1\), the fitness increase is \(\delta n + |b'|_1 - |b|_1\). Hence the above estimations of jump lengths are not applicable. These special steps are unorthodox as the large fitness increase makes it likely that any mutation on the OneMax part is accepted. We show that the progress on the OneMax part across all special steps is \(O(n^{3/4})\) with high probability.

We grant the algorithm an advantage if we assume that, after initialising with \(|b|_1 \ge n/4\), no search point with \(|b|_1 < n/4\) is ever reached.^{1} Under this assumption we always have at least as many 1-bits as 0-bits in *b*, and mutation in expectation flips at least as many 1-bits to 0 as 0-bits to 1.

Then the progress in \(|b|_1\) in one special step increasing the number of leading ones by \(d\ge 1\) can be described as follows. Imagine a matching (pairing) between all bits in *b* such that each pair contains at least one 1-bit. Let \(X_i\) denote the random change in \(|b|_1\) by the *i*-th pair. If the pair has two 1-bits, \(X_i \le 0\) with probability 1. Otherwise, we have \(X_i = 1\) if the 0-bit in the pair is flipped, the 1-bit in the pair is not flipped, and the mutant is accepted (which depends on the overall \(|b|_1\)-value in the mutant). The potential fitness increase is at most \(dn + n/2\) as the range of \(|b|_1\)-values is *n* / 2. Likewise, we have \(X_i = -1\) if the 0-bit is not flipped, the 1-bit is flipped, and the mutant is accepted (which again depends on the overall \(|b|_1\)-value in the mutant). The fitness increase is at least \(dn - n/2\). With the remaining probability we have \(X_i = 0\). Hence for global mutations (for local mutations simply drop the \(1-1/n\) term) the total progress in a special step increasing \({\textsc {LO}}(a) \) by \(d\) is stochastically dominated by a sum of independent variables \(Y_1, \dots , Y_{n/4}\) where \(\mathrm {Pr}\left( Y_i = \pm 1\right) = 1/n \cdot (1-1/n) \cdot p_{\mathrm {fix}}(dn \pm n/2)\) and \(Y_i = 0\) with the remaining probability.

*x*and \(e^x \le 1+2x\) for \(0 \le x \le 1\).

*n*large enough.

The total progress in all *m* special steps is hence stochastically dominated by a sequence of \(m \cdot n/4\) random variables \(Y_i\) as defined above, with \(d:= 1\). Invoking Lemma 17 (basically a Hoeffding bound on the non-zero outcomes of the variables), stated in the appendix, with \(\delta := n^{3/4}\), the total progress in all special steps is at most \(\delta + m \cdot n/4 \cdot \text {E}\left( Y_i\right) = \delta + O(n^{1/2}) = O(n^{3/4})\) with probability \(1-e^{-\varOmega (n^{1/2})}\).

Hence the net gain in the number of ones in all special steps is at most \(n^{3/4} + O(mn/4 \cdot n^{-3/2}) = O(n^{3/4})\) with probability \({1-e^{-\varOmega (n^{1/2})}}\).

Together with all regular steps, the progress on the OneMax part is at most \(1.14n/9 + O(n^{3/4})\), which for large enough *n* is less than the distance \(7n/16 - (n/4+n^{3/4})\) to reach a point with \(|b|_1 \ge 7n/16\) from initialisation. This proves the claim. \(\square \)

Finally, we put the previous lemmas together into our main theorem that establishes that SSWM can optimise Balance in polynomial time.

### Theorem 15

With probability \(1-e^{-\varOmega (n^{1/2})}\) SSWM with \(\beta = n^{-3/2}\) and \(N \beta = \ln n\) optimises Balance in time \(O(n/\beta ) = O(n^{5/2})\).

### Proof

By Chernoff bounds, the probability that for the initial solution \(x_0 = a_0 b_0\) we have \(n/4 - n^{3/4} \le |b_0|_1 \le n/4 +n^{3/4}\) is \(1-e^{-\varOmega (n^{1/2})}\). We assume pessimistically that \(n/4 \le |b_0|_1 \le n/4 +n^{3/4}\). Then Lemma 14 is in force, and with probability \(1-e^{-\varOmega (n^{1/2})}\) within *T* relevant steps, *T* as defined in Lemma 13, SSWM does not reach a trap or a search point with fitness 0. Lemma 13 then implies that with probability \(1-e^{-\varOmega (n^{1/2})}\) an optimal solution with *n* / 2 leading ones is found.

The time bound follows from the fact that \(T = O(n/\beta )\) and that, again by Chernoff bounds, we have at least *T* relevant steps in 3*T* iterations of SSWM, with probability \(1-e^{-\varOmega (n^{1/2})}\). \(\square \)

## 6 Conclusions

The field of evolutionary computation has matured to the point where techniques can be applied to models of natural evolution. Our analyses have demonstrated that runtime analysis of evolutionary algorithms can be used to analyse a simple model of natural evolution, opening new opportunities for interdisciplinary research with population geneticists and biologists.

Our conclusions are highly relevant for biology, and open the door to the analysis of more complex fitness landscapes in this field and to quantifying the efficiency of evolutionary processes in more realistic scenarios of evolution. One interesting aspect of our results is that they impose conditions on population size (*N*) and strength of selection (\(\beta \)) which represent fundamental limits to what is possible by natural selection. We hope that these results may inspire further research on the similarities and differences between natural and artificial evolution.

From a computational perspective, we have shown that SSWM can overcome obstacles such as posed by \({\textsc {Cliff}}_{d} \) and \({\textsc {Balance}} \) in different ways to the (1+1) EA, due to its non-elitistic selection mechanism. We have seen how the probability of accepting a mutant can be tuned to enable hill climbing, where fitness-proportional selection fails, as well as tunnelling through fitness valleys, where elitist selection fails. For Balance we showed that SSWM can take advantage of information about the steepest gradient. The selection rule in SSWM hence seems to be a versatile and useful mechanism. Future work could investigate its usefulness in the context of population-based evolutionary algorithms.

## Footnotes

- 1.
Otherwise, we restart our considerations from the first point in time where \(|b|_1 \ge n/4\) again, replacing

*T*with the number of remaining steps. With overwhelming probability we will then again have \(|b|_1 \le n/4 + n^{3/4}\).

## Notes

### Acknowledgments

The research leading to these results has received funding from the European Union Seventh Framework Programme (FP7/2007-2013) under Grant Agreement No. 618091 (SAGE). The authors thank the anonymous reviewers for their many constructive comments.

### References

- 1.Auger, A., Doerr, B. (eds.): Theory of Randomized Search Heuristics-Foundations and Recent Developments. Series on Theoretical Computer Science, vol. 1. World Scientific, Singapore (2011)MATHGoogle Scholar
- 2.Chastain, E., Livnat, A., Papadimitriou, C., Vazirani, U.: Algorithms, games, and evolution. Proc. Natl. Acad. Sci.
**111**(29), 10620–10623 (2014)MathSciNetCrossRefMATHGoogle Scholar - 3.Chatterjee, K., Pavlogiannis, A., Adlam, B., Nowak, M.A.: The time scale of evolutionary innovation. PLoS Comput. Biol.
**10**(9), 1–7 (2014)Google Scholar - 4.Corus, D., Dang, D.-C., Eremeev, A.V., Lehre, P.K.: Level-based analysis of genetic algorithms and other search processes. In: Parallel Problem Solving from Nature (PPSN), Springer, Berlin, pp. 912–921 (2014)Google Scholar
- 5.Doerr, B.: Analyzing Randomized Search Heuristics: Tools from Probability Theory. In: [1], pp. 1–20. World Scientific, Singapore (2011)Google Scholar
- 6.Eiben, A .E., Smith, J .E.: Introduction to Evolutionary Computing, 2nd edn. Springer, Berlin (2015)CrossRefMATHGoogle Scholar
- 7.Ewens, W.J.: Mathematical Population Genetics 1: Theoretical Introduction, 2nd edn. Springer, New York (2004)CrossRefMATHGoogle Scholar
- 8.Gillespie, J.H.: Molecular evolution over the mutational landscape. Evolution
**38**(5), 1116–1129 (1984)CrossRefGoogle Scholar - 9.Jägersküpper, J., Storch, T.: When the plus strategy outperforms the comma strategy and when not. In: Proceedings of IEEE Foundations of Computational Intelligence (FOCI 2007), pp. 25–32. IEEE (2007)Google Scholar
- 10.Jansen, T.: Analyzing Evolutionary Algorithms. The Computer Science Perspective. Springer, Berlin (2013)CrossRefMATHGoogle Scholar
- 11.Jansen, T., Oliveto, P.S., Zarges, C.: On the analysis of the immune-inspired B-Cell algorithm for the Vertex Cover problem. In: Proceedings of the International Conference on Artificial Immune Systems (ICARIS ’11), Springer, Berlin, pp. 117–131 (2011)Google Scholar
- 12.Jansen, T., Wegener, I.: A comparison of simulated annealing with a simple evolutionary algorithm on pseudo-Boolean functions of unitation. Theor. Comput. Sci.
**386**(1–2), 73–93 (2007)MathSciNetCrossRefMATHGoogle Scholar - 13.Johannsen, D.: Random Combinatorial Structures and Randomized Search Heuristics. Ph.D. thesis, Universität des Saarlandes, Saarbrücken, Germany and the Max-Planck-Institut für Informatik (2010)Google Scholar
- 14.Kimura, M.: On the probability of fixation of mutant genes in a population. Genetics
**47**(6), 713–719 (1962)Google Scholar - 15.Lehre, P.K., Witt, C.: Black-box search by unbiased variation. Algorithmica
**64**(4), 623–642 (2012)MathSciNetCrossRefMATHGoogle Scholar - 16.Neumann, F.: Expected runtimes of evolutionary algorithms for the Eulerian cycle problem. Comput. Oper. Res.
**35**(9), 2750–2759 (2008)MathSciNetCrossRefMATHGoogle Scholar - 17.Neumann, F., Oliveto, P.S., Witt, C.: Theoretical analysis of fitness-proportional selection: landscapes and efficiency. In: Proceedings of the 2009 Genetic and Evolutionary Computation Conference (GECCO ’09), pp. 835–842, ACM (2009)Google Scholar
- 18.Neumann, F., Wegener, I.: Randomized local search, evolutionary algorithms, and the minimum spanning tree problem. Theor. Comput. Sci.
**378**(1), 32–40 (2007)MathSciNetCrossRefMATHGoogle Scholar - 19.Neumann, F., Witt, C.: Runtime analysis of a simple ant colony optimization algorithm. Algorithmica
**54**(2), 243–255 (2009)MathSciNetCrossRefMATHGoogle Scholar - 20.Neumann, F., Witt, C.: Bioinspired Computation in Combinatorial Optimization—Algorithms and Their Computational Complexity. Springer, Berlin (2010)MATHGoogle Scholar
- 21.Oliveto, P.S., Sudholt, D.: On the runtime analysis of stochastic ageing mechanisms. In: Proceedings of the 2014 Genetic and Evolutionary Computation Conference (GECCO ’14), ACM Press, pp. 113–120 (2014)Google Scholar
- 22.Oliveto, P.S., Witt, C.: Simplified drift analysis for proving lower bounds in evolutionary computation. Algorithmica
**59**(3), 369–386 (2011)MathSciNetCrossRefMATHGoogle Scholar - 23.Oliveto, P.S., Witt, C.: On the runtime analysis of the simple genetic algorithm. Theor. Comput. Sci.
**545**, 2–19 (2014)MathSciNetCrossRefMATHGoogle Scholar - 24.Paixão, T., Badkobeh, G., Barton, N., Çörüş, D., Dang, D.-C., Friedrich, T., Lehre, P.K., Sudholt, D., Sutton, A.M., Trubenová, B.: Toward a unifying framework for evolutionary processes. J. Theor. Biol.
**383**, 28–43 (2015)Google Scholar - 25.Paixão, T., Pérez Heredia, J., Sudholt, D., Trubenová, B.: First steps towards a runtime comparison of natural and artificial evolution. In: Proceedings of the 2015 Genetic and Evolutionary Computation Conference (GECCO ’15), pp. 1455–1462, ACM (2015)Google Scholar
- 26.Reichel, J., Skutella, M.: Evolutionary algorithms and matroid optimization problems. Algorithmica
**57**(1), 187–206 (2010)MathSciNetCrossRefMATHGoogle Scholar - 27.Rohlfshagen, P., Lehre, P.K., Yao, X.: Dynamic evolutionary optimisation: an analysis of frequency and magnitude of change. In: Proceedings of the 2009 Genetic and Evolutionary Computation Conference (GECCO ’09), ACM Press, pp. 1713–1720 (2009)Google Scholar
- 28.Rowe, J.E., Sudholt, D.: The choice of the offspring population size in the (1,\(\lambda \)) evolutionary algorithm. Theor. Comput. Sci.
**545**, 20–38 (2014)MathSciNetCrossRefMATHGoogle Scholar - 29.Scharnow, J., Tinnefeld, K., Wegener, I.: The analysis of evolutionary algorithms on sorting and shortest paths problems. J. Math. Modell. Algorithms
**3**(4), 349–366 (2004)MathSciNetCrossRefMATHGoogle Scholar - 30.Sudholt, D.: A new method for lower bounds on the running time of evolutionary algorithms. IEEE Trans. Evol. Comput.
**17**(3), 418–435 (2013)CrossRefGoogle Scholar - 31.Sudholt, D., Thyssen, C.: A simple ant colony optimizer for stochastic shortest path problems. Algorithmica
**64**(4), 643–672 (2012)MathSciNetCrossRefMATHGoogle Scholar - 32.Traulsen, A., Iwasa, Y., Nowak, M.A.: The fastest evolutionary trajectory. J. Theor. Biol.
**249**(3), 617–623 (2007)MathSciNetCrossRefGoogle Scholar - 33.Valiant, L .G.: Evolvability. J. ACM
**56**(1), 3:1–3:21 (2009)MathSciNetCrossRefMATHGoogle Scholar - 34.Wegener, I.: Methods for the analysis of evolutionary algorithms on pseudo-boolean functions. In: Sarker, R., Mohammadian, M., Yao, X. (eds.) Evolutionary Optimization, volume 48 of International Series in Operations Research & Management Science, chapter 14. Kluwer Academic Publishers, Dordrecht, pp. 349–369 (2003)Google Scholar
- 35.Witt, C.: Worst-case and average-case approximations by simple randomized search heuristics. In: Proceedings of the 22nd Symposium on Theoretical Aspects of Computer Science (STACS ’05), Springer, Berlin, pp. 44–56 (2005)Google Scholar
- 36.Witt, C.: Population size versus runtime of a simple evolutionary algorithm. Theor. Comput. Sci.
**403**(1), 104–120 (2008)MathSciNetCrossRefMATHGoogle Scholar

## Copyright information

**Open Access**This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.