Self-adaptive potential-based stopping criteria for Particle Swarm Optimization with forced moves

We study the variant of Particle Swarm Optimization that applies random velocities in a dimension instead of the regular velocity update equations as soon as the so-called potential of the swarm falls below a certain small bound in this dimension, arbitrarily set by the user. In this case, the swarm performs a forced move. In this paper, we are interested in how, by counting the forced moves, the swarm can decide for itself to stop its movement because it is improbable to find better candidate solutions than the already-found best solution. We formally prove that when the swarm is close to a (local) optimum, it behaves like a blind-searching cloud and that the frequency of forced moves exceeds a certain, objective function-independent value. Based on this observation, we define stopping criteria and evaluate them experimentally showing that good candidate solutions can be found much faster than setting upper bounds on the iterations and better solutions compared to applying other solutions from the literature.


Introduction
Background Particle Swarm Optimization (PSO) is a meta-heuristic for so-called continuous black box optimization problems, which means that the objective function is not explicitly known in form of a, e. g., closed formula.PSO produces good results in a variety of different real world applications.The classical PSO as first introduced by Eberhart and Kennedy [KE95,EK95] in the year 1995 works in (solution candidates improving) iterations and can be very easily implemented and adapted to the users' applications, which lead to increased attention not only among computer scientists.In order to further improve the performance, many authors present changes to the original, "plain," or classical PSO scheme (for exact definitions, see Sec. 2) to improve the quality of the returned solution.
A serious problem of PSO that can be sometimes observed is the phenomenon of premature stagnation, i. e., the convergence of the swarm to a non-optimal solution.This phenomenon has been theoretically addressed by Lehre and Witt in [LW13].To overcome such stagnation, the authors propose Noisy PSO that adds a "noise" term to the velocity at every move.They prove that for the Noisy PSO started on a certain simple 1-dimensional objective function the first hitting time of the δ -neighborhood of the global optimum is finite.However, as proved in [SW13a], premature stagnation of classical PSO does not occur at all when the search space is 1-dimensional and the objective function is continuous, i. e., in the 1-dimensional case, PSO provably finds at least a local optimum, almost surely (in the well defined sense of probability theory).Furthermore, [SW13a] shows a similar result for a slightly modified PSO in the general D-dimensional case (for stagnation-related results regarding the unmodified PSO, see [RSW15]).This slightly modified PSO assigns a small random velocity in solely one dimension only if the so-called potential of the swarm -a fundamental, measurable quantity of the swarm -of all particles in this dimension falls below a certain (arbitrary and small) bound δ .In the following, we call such random velocity particle moves forced steps and the PSO variant f-PSO.The f-PSO provably finds a local optimum almost surely [SW13a].Monitoring the swarm's potential and increasing it from time to time by the forced steps is the key ingredient to mathematically proving successful convergence to a (local) optimum.In this paper, we will use the forced steps of f-PSO and the potential for overcoming another problem of heuristics, the question when to stop the algorithm.
Problem and new contribution An important problem that arises in the context of PSO and other iterative optimization methods (see [SW81,p. 23f] and further papers, referenced in the related work section below) is when to stop the iterations and to return the best found admissible solution.Usually, the process is terminated (i) when an upper limit on the number of iterations is reached, or (ii) when an upper limit on the number of evaluations of the objective function is reached, or (iii) when the chance of achieving significant improvements in further iterations is extremely low.The choice of the mentioned two upper limits obviously depends on the concrete objective function which means that the user has to have detailed knowledge of the objective function and to "intervene" by hand.In (iii), it is desired and advantageous that the algorithm adaptively decides when to stop, so external intervention is not necessary anymore.To achieve this for PSO, many criteria were introduced in the literature.For a short overview, see below.A commonly used criterion is the swarm diameter, i. e., the maximum distance between two particles, as a measure for the expansion and thus the movement capability of the swarm.When PSO is extended with forced steps, this and also other criteria of this kind do not work because the expansion of the overall swarm is forcefully kept above a certain value and can no longer converge like it does for the classical PSO algorithm.Just the globally best position found by the swarm converges to optimum.
The goal of this paper is to characterize the behavior of the f-PSO when it is close to a (local) optimum.Our experiments and mathematical investigations show that the number of forced moves does not only increase significantly when the distance to the next local optimum falls below a certain bound, but that additionally the number of forced moves performed close to a local optimum is independent of the objective function.In particular, we prove by potential arguments that the swarm "pulsates" in a cloud around the best solution candidate found so far.Note that a similar behavior has been described recently in [YSG18] for the classical PSO.Therefore, by measuring the frequency of occurrences of forced moves, this frequency can act as a stopping criterion, so the swarm may come to a self-determined halt.All findings are experimentally supported.

Related work
Adaptive stopping criteria for PSO have been investigated by Zielinski et al. [ZPL05] and by Zielinski and Laur [ZL07].In these papers, a list of upper limit-based and adaptive termination criteria is presented and in experiments applied to (2-dimensional) established benchmark functions and a real-world problem, resp.In [KHL + 07], Kwok et al. introduce a stopping criterion for PSO based on the rate of improvements found by the swarm in a given time interval.Quality of the termination is assured using the non-parametric sign-test enforcing a low false-positive rate.A further stopping approach due to Ong and Fukushimah [OF15] combines PSO with gene matrices.
To name some work beyond PSO, Ribeiro et al. [RRS11] introduce a stopping criterion that stops the GRASP (Greedy Randomized Adaptive Search Procedures) algorithm when the probability of an improvement is below a threshold under an experimentally fitted normal distribution.
Safe et al. [SCPB04] present a study of various aspects associated with the specification of termination conditions for simple genetic algorithms.In [AK00], Aytug and Koehler introduce a stopping criterion that stops a genetic algorithm when the optimal solution is found with a specified confidence and thus no real further progress can be expected.
In a very general, ground breaking investigation, Solis and Wets [SW81] consider random search in general and also address the question of stopping criteria.In this context, Dorea [Dor90] presents two adaptive stopping criteria.
Stopping criteria in the context of multi-objective optimization is presented by Martí et al. [MGBM16].
Organization of paper Sec. 2 presents the necessary description of the classical and forced PSO (f-PSO) and the definition of the potential of a swarm.In Sec. 3, we experimentally and mathematically analyze the behavior of f-PSO when it comes close to a (local) optimum.In Sec. 4, based on the mathematical analysis in Sec. 3, we present the new termination criteria and show their practicability by experimental evaluations.

Definitions
Due to the wide variety of existing PSO variants, we first state the exact "classical" PSO algorithm on which our work is based.
Definition 1 (Classical PSO Algorithm).A swarm S of N particles moves through the D-dimensional search space R D with D = {1, . .., D} being the set of dimensions.Each particle n ∈ S consists of a position X n ∈ R D , a velocity V n ∈ R D and a local attractor L n ∈ R D , storing the best position particle n has visited so far.Additionally, the particles of the swarm share information via the global attractor G ∈ R D , describing the best point any particle has visited so far, i. e., as soon as a particle has performed its move1 , it possibly updates the global attractor immediately.
The actual movement of the swarm is governed by the following movement equations where χ, c 1 , c 2 ∈ R + are some positive constants to be fixed later, and r and s are drawn u. a. r. from [0, 1] D every time the equation is applied.
Here, ⊙ denotes entrywise multiplication (Hadamard product).The application of the equations on particle n is called the move of n.When all particles have executed their moves, the swarm has executed one iteration.Now we repeat the definition of a swarm's potential measuring how close it is to convergence, i. e., we describe a measure for its movement.A swarm with high potential should be more likely to reach search points far away from the current global attractor, while the potential of a converging swarm approaches 0. These considerations lead to the following definition [SW13b] (the original definition in [SW13a] is more complex just for technical reasons): Definition 2 (Potential).Fix a moment of the computation of swarm S .For d ∈ D, the current potential Φ d of S in dimension d is is the total potential of S , and φ n d is the contribution of particle n to the potential of S in dimension d.
Note that the potential of the swarm has an entry in every dimension.The swarm comes to a halt if Φ → (0, . .., 0).So, if the particles come close to G, the swarm may stop even if G is a non-optimal point, an incident that is called (premature) stagnation.The single values of Φ can be actually computed and, hence, be used for decisions on the swarm.Between two different dimensions, the potential difference might be large, and "transferring" potential from one dimension to another is not possible due to the movement equations.On the other hand, along the same dimension the particles influence each other and can transfer potential from one particle to the other.This is the reason why there is no potential of individual particles, but only their contribution to the potential.
To address the phenomenon of stagnation, [SW15] slightly modified the PSO movement equations from Definition 1 as follows by "recharging" potential and, hence, keeping the swarm moving furthermore.
Definition 3 (f-PSO).The modified movement of the swarm is governed by the following movement equations where χ, c 1 , c 2 , δ ∈ R + are some positive constants to be fixed later, r, s and t are drawn u. a. r. from [0, 1] for every move and dimension of a particle.
If the forced velocity update applies to a particle, we call its move and the corresponding dimension in the move forced.An iteration of the swarm is called forced if during this iteration at least one particle performs a forced move.The whole method is called f-PSO.
If it is necessary to identify the values at the beginning of iteration i, we write Algorithm 1 provides a formal and detailed overview over f-PSO.The introduction of forced velocity updates guarantees that the swarm (or more precisely, the global attractor G) almost surely does not converge to a non-optimal point, but finds a local optimum [SW13a].In our analysis and for the experiments, we used the common parameter settings χ = 0.72984, c 1 = 1.49617, c 2 = 1.49617 and N > 1 as suggested and used in [CK02], which are parameter settings that are widely used in the literature.

Behavior of the f-PSO algorithm
The idea of the modification of the classical PSO algorithm is to help the swarm overcome (premature) stagnation.The modification is implemented in the calculation of the new velocity during a move of a particle (see Def. 3 and Lines 9 and 10 of Algorithm 1).The new velocity of a particle in the current dimension d is drawn u. a. r. from [−δ , δ ] when for all particles n ′ ∈ S their contributions φ n ′ d to the potential in dimension d is less than δ , which means that the range in which the swarm can optimize is small in dimension d.
The main topic of this section is to examine if at such a situation the particle swarm uses from now on only forced moves or, more desirable, the swarm will recover and continue using regular velocity updates.In [SW15], it is shown that indeed the latter is the case, unless the global attractor G is already in the neighborhood of a local optimum.As in this case almost all moves are forced, we will see in Sec. 4 that this can be translated into a criterion to stop the PSO's execution.

Experiments
We first introduce the notion of the forcing frequency to quantify the number of applications of the forced velocity update (Line 10 of Algorithm 1).To explore the behavior of the particle swarm with respect to the absolute forcing frequency σ (I, d), we performed a series of experiments on the well known benchmark functions SCHWE-FEL, ROSENBROCK, RASTRIGIN, H. C. ELLIPTIC and SPHERE (for a comprehensive overview of these and many other benchmark functions, see [Hel10, Sec 4.2]).SPHERE and H. C. ELLIPTIC will be analyzed in detail in Sec.3.2.The tests were performed with Raß' HiPPSO [Raß17], a high precision implementation of PSO, in order to rule out the influence of insufficient computer  is relatively small if f (G) is far away from the (unique) optimum value and increases considerably when f (G) approaches the optimum.Additionally, one can see that the gradient of the absolute forcing frequency, i. e., the relative forcing frequency, tends to be constant for all i ≥ i 0 after some value i 0 .This is a showcase and remains true for most of the other tested benchmark functions.

Definition 4 (Absolute and relative forcing frequency). Let I denote a (time) interval of |I| iterations during a run of the
ROSENBROCK function During the tests on the standard benchmark functions, processing the ROSENBROCK function showed a slightly different behavior.Metaphorically speaking, there is a small "banana-shaped" valley in this function that leads from a local optimum to the global optimum.If the attractors are in this valley (what they are quite early), the chance to improve in some dimensions is quite small and improvements can easily be voided by setbacks in other dimensions.If this happens, forced velocity updates will be applied in some dimensions while in the other dimensions the particles use the usual, regular velocity updates, which can be seen clearly  in Fig. 2. The chance to get moving into the optimum's direction again is not zero, but quite small which leads to a long phase of near-stagnation as also can be seen in Fig. 2. Eventually the swarm will re-start moving again, but it may take a long time to do so.
RASTRIGIN function Another phenomenon may occur on functions like RASTRIGIN, where there are many local optima.Here, it might happen that the global attractor G is already close to a (good) local optimum, but the local attractor L n of some particle n is close to a worse local optimum.In such a situation, the potential is governed by the distance between G and L n , and the condition for executing a forced velocity update is not satisfied, even though a good local nearoptimum value f (G) has already been found.But our experiments showed that this happens hardly ever.

Experiments on the SPHERE Function, Phases, and Theoretical Analysis
We now analyze the swarm's final behavior with the help of experiments on SPHERE (the results for H. C. ELLIPTIC are similar).These experiments show that the final behavior can be classified into two main phases: the approaching phase and the pulsation phase that in turn has three distinct sub-phases: forced phase, lockout phase, recovery phase.These findings hold for all objective functions, as the swarm is in this final phase a kind of a blind searching cloud.
So, when the global attractor reaches the δ -neighborhood of a local optimum, the number of global attractor updates decreases and the distance the global attractor moves becomes small.At that time, the whole swarm will start to approach the global attractor.At some point during this approaching phase, the local attractors will be close to the global attractor and the swarm will loose much of its potential.As described in [SW15], the introduction of forced velocity updates avoids that the actual positions of the particles approach each other closer than a value of about δ and the swarm enters a kind of pulsation phase: periodically it contracts and expands.However, the global attractor still converges (in the mathematical "infinite time" sense) to the local optimum.Also, the positions of the local attractors may further contract.
It can be observed that when the attractors are all almost at the same position and the particles converge to this position, the number of dimensions that are forced begins to increase.During the pulsation phase almost all particles are forced and the relative forcing frequency σ (I, d)/|I| begins to stagnate at a certain value ∂ σ d (the gradient in the figures).Hence, also σ (I)/|I| begins to stagnate at ∂ σ .When this stagnation is reached, the f-PSO will become a periodic pulsating process and will behave almost like a blind search in a box with side-length of order of magnitude δ around the global attractor.Hence, the objective function f will become irrelevant and the value ∂ σ is independent of f .

Experimental Identification of parameters influencing ∂ σ
In order to identify the parameters that influence ∂ σ , a series of experiments were performed with the SPHERE function.To study a pure pulsation phase, the global and local attractors were initially all set to 0, which is the global optimum of this function.With the swarm parameters χ, c 1 and c 2 being fixed, the two crucial remaining parameters that might influence ∂ σ are • the number D of dimensions and • the number N of particles, whereas the choice of δ has no influence at all.By changing δ the range of φ d causing a forced step is influenced for a dimension d ∈ D. However, the random velocity assigned by the forced step in this dimension d is changed in the same magnitude.Therefore the change of δ has no influence on the absolute forcing frequency as these two effects cancel each other.
Fig. 3 and 4 show the effect of changes in the dimension number D and swarm size N, resp., on the stagnation value ∂ σ for fixed |I|.The use of a forced move in a given dimension depends only on the velocities and positions of all particles relative to the global attractor G in this dimension.The dimensions are independent of each other, therefore the stagnation value (recall that |I| is fixed) for each number D of dimensions does not change with an increased number of dimensions as shown in Fig. 3.The same is not true for varying the number of particles, see Fig. 4. If the length |I| of the interval and the number D of dimensions are fixed, increasing the number of particles increases the number of applications of the movement equations by |I| • D for each new particle.Therefore, more dimensions can be forced in a given interval, but there are also more particles that must have a partial potential φ less than δ .

Mathematical analysis of the sub-phases for arbitrary objective functions
We can identify three sub-phases during the pulsation phase which are repeated in a fixed order for a given dimension.
(i) Starting with a forced phase, where each particle has its move in this dimension forced, (ii) followed by a lockout phase with the probability of a forced move being zero and (iii) finally a recovery phase in which the particles attempt to converge to the global attractor until one move of a particle gets forced again and the cycle repeats itself.
To fully understand why this leads to a less forced frequency σ (I) per particle, see Fig. 4, we have to take a closer look at the three sub-phases of the pulsation phase.As described in Def. 3, a dimension d ∈ D is forced during the move of a particle when all particles n ′ including the particle itself have contribution φ n ′ d less than δ in this dimension d.For i ∈ N, n ∈ S , and d ∈ D let F i,n,d be the {0, 1}-indicator variable being 1 iff in iteration i and dimension d particle n has the velocity update forced.The global and the local attractors are constant and in particular, it can be assumed that G i d = G for fixed G. We get the following observations regarding a forced move in a given dimension.
The first phase we analyze is the lockout phase.Recall that φ n Proof.With the attractors being constant, the only way for particle n to have contribution φ n d less than δ is to make a swarm move that reduces the sum of the velocity and the distance to the global attractor to less than δ .The earliest time this is possible is in the next iteration.
By Lemma 1, we get the length of the lockout phase as N, the number of particles.For the forced phase, we have to look at the probability that for a particle n with a forced move in dimension d, φ n d < δ still applies in the next iteration.
As each dimension is independent and the same calculations are performed in each dimension, we may restrict our analysis to one dimension d and omit the index d in the following proofs.
With this probability we can now analyze the length of the forced phase.
Lemma 3. Given F i,n,d = 1, define Y as the number of consecutive particle steps, during which dimension d of the respective particle is forced, starting with particle n at iteration i.Then for k ≥ 1 holds P[Y = k] = 1 2 k−1 .Proof.Again, we omit the dimension d from the variables.
By Lemma 1, we know that if for one particle |V n,i+1 |+|G−X n,i+1 | ≥ δ holds, the next particle cannot be forced and with Lemma 2, P[F i,n+1 = 1|F i,n = 1] = 1 2 .By assumption, P[F i,n = 1] = 1.By induction on k, we have: Finally for the recovery phase we cannot give an exact length, as it depends on the positions of the particles in the search space.The following Lemma 4 and Fig. 6 however give us at least an idea of the behavior of the particle during this phase.
Lemma 4. The probability for a particle n to get a partial potential φ n d greater than δ when not forced for the first time after a chain of forced moves is Proof.By assumption, L n = G.Again, d is omitted from the variables.
Applying the PSO movement equations leads to Under the assumption F i−1,n = 1, we have V n,i ∼ U (−δ , δ ) and |V n,i | ∼ U (0, δ ) (U denotes the continuous uniform distribution).Therefore we get α ≥ 1 − 1 2χ and can derive β = 1 2 by the following equations.Note that the lower bound on the probability in Lemma 4 is negative for χ < 1 2 .The actual probability depends on the position in the search space.With χ > 1 2 , the probability for a particle n to get partial potential φ n > δ is positive independent of the position of the particle.Therefore, the length of the recovery phase will decrease drastically with smaller χ values.We may conclude that for a given dimension we get the following behavior in the pulsation phase.
Theorem 1.Let S be a swarm of particles at a (local) optimum of an arbitrary function f , i. e., ∀n ∈ S : L n = G = Opt( f ) with Opt( f ) being a (local) optimum of the function f .The swarm repeats three phases independently for each dimension: • the forced phase of length k ≥ 1 with probability 1 2 k−1 , • the lockout phase of length N, • and the recovery phase of length ℓ ≥ 0 leading to a stagnation of σ (I)/|I| → ∂ σ at a (local) optimum depending on N and D.
Starting at one forced particle, the chance for the next k particles to be forced has probability 2 −k+1 as stated in Lemma 3 which is independent of the number of particles.At some particle u, this sequence of forced particles ends, and in this moment the forced phase ends.With Lemma 1, at least the following N particles will not be forced which constitutes the lockout phase.The next particle that can be forced is (u + 1) mod N = v, and for that all particles n that move between u and v need to keep partial potential φ n d < δ .All these particles use the regular swarm movement equations from Def. 3, and, with G being constant, they are independent of each other.As shown in Lemma 4, the chance for a particle n to get partial potential φ n d > δ can be quite probable depending on the actual swarm parameters.Hence, with increasing N the chance for a particle n to get partial potential φ n d > δ increases exponentially in the order of the probability of the complementary event.The exact probability for this event depends on the position of the particle relative to the global attractor.The same is true for the number of iterations a particle needs to fix its partial potential φ , but with normal swarm behavior the particle starts to converge to the global attractor and to decrease the potential.Fig. 6 shows the distance of the particles to the global attractor at the global optimum.We can see that the particles tend to converge to the attractor with high probability thus ending the recovery phase.Therefore, the particles once again start to get forced and the same pattern repeats itself.

Termination criteria
In Sec. 3 above, we discussed the behavior of the particle swarm when the global and local attractors are directly at a local optimum.In [SW15], it is shown that with f-PSO the global attractor converges (in the infinite-time sense) to an (at least local) optimum almost surely.However, this convergence can be very slow.Therefore, the particle swarm should be stopped at some time and either be re-started with, when indicated, different parameter settings, or, if the current global attractor does not satisfy the desired quality, to use a different algorithm.As the swarm does not know when it is close to a local optimum and due to the forced steps, the swarm will just converge up to a δ -neighborhood of the global attractor which in turn approached a (local) optimum.With our observations from Sec. 3 this means that the forcing frequency will exceed a fixed value which can be measured.
In the following, we will define a new stopping criterion that stops the algorithm at time when the f-PSO algorithm reaches this frequency.Note that reaching this forcing frequency is just an indication that the swarm is close to an optimum.

Full stagnation criterion
By Theorem 1, we know that the absolute and relative forcing frequency become fixed if the swarm would be at a (local) optimum, independent of the objective function.So one can use an arbitrary known function with known optimum to experimentally determine this frequency by starting the swarm in the optimum, as we did in Sec.3.2 for SPHERE.
We will use this observation to describe the first simple termination criterion: we just stop the swarm when the measured absolute forcing frequency begins to stagnate and is close to the fixed forcing frequency around a (local) optimum.
Definition 5. Let σ stag (N, D, µ) be the objective function-independent, fixed absolute forcing frequency around an optimum when intervals are considered with interval length µ, and let γ ∈ N. The criterion "(γ, µ)-FULL STOP" is to terminate the f-PSO execution iff σ stag (N, D, µ) −σ (I) ≤ γ for some interval I with |I| = µ .
We tested (γ, µ)-FULL STOP in Line 16 of Algorithm 1, i. e., f-PSO, on our benchmark functions with N = 5 particles, D = 15 dimensions, δ = 10 −7 and interval length µ = 50 000.The standard swarm parameters were used.The stagnation frequency σ stag (N, D, µ) = 318 350 was measured on the SPHERE function with all particles initialized at the global optimum.We used γ = 1350, so we terminated the execution when σ (I) ≥ 317 000 for intervals I of length µ.If better solutions would be required, one could reduce γ.We compared the solutions obtained with f-PSO that used (γ, µ)-FULL STOP with the results obtained by f-PSO that run for 15 000 000 iterations.Table 1 shows the median (not the average) and standard deviation on 500 runs of the norm of the gradient of the global attractor on these benchmark functions with and without the use of criterion (γ, µ)-FULL STOP.The less the gradient, the better the global attractor.On simple benchmark functions like SPHERE criterion (γ, µ)-FULL STOP produces sufficiently good output.We see that the time of termination is significantly earlier than with pre-specified number of iterations and that the difference in the gradient at termination is negligible.

Partial stagnation criterion
On more complex functions like ROSENBROCK and RASTRIGIN, we can identify two scenarios that interfere with the FULL STOP termination criterion.
The first scenario can be observed on the ROSENBROCK function.It may occur that some dimensions are already highly optimized, i. e., close to an optimal coordinate, while others are still far away from optimum coordinates, a situation which leads to late termination.With ROSEN-BROCK when the particles and the attractors are in the valley between the local and the global attractor, the chance to improve on the not yet optimized dimensions is voided by the worsening in the already good dimensions.Therefore the swarm will not change the local (and, thus, the global) attractors and the attractors are far away from each other.In this case the optimized dimensions reached the average stagnation frequency σ stag (N, D, µ)/D, but the other dimensions will not get forced much or not at all as shown in Fig. 2.
The second scenario becomes visible on the RASTRIGIN function.This function has many local optima with a global optimum at 0. Unlike to ROSENBROCK there is not much difference in the behavior of the particles in the different dimensions.Here, the problem originates from many local optima with only a small difference in function values.This can lead to the global attractor being in a better local optimum than some of the local attractors.Given this situation the particles have a very small probability to actually improve their local attractor.At this point, the algorithm should terminate, but given the different positions of the global and the local attractors these particles cannot decrease their velocity below the necessary bounds to reach the probability of a forced step at the stagnation frequency.The current σ will stay around some value, which is different from the stagnation frequency of each dimension and with a fluctuation larger than the fluctuation at the stagnation frequency.However, note that the probability for such an event is quite small and was not encountered in our experiments.
To account for such situations we formulate an extended termination criterion.
Definition 6.Let σ stag (N, D, µ) be as defined in Def. 5, and let κ ∈ [1, D].The criterion "(κ, γ, µ)- Def. 6 introduces the additional parameter κ.It can be interpreted as how many dimensions are required to reach their stagnation frequency such that f-PSO is stopped.Actually, if κ = D, we have (γ, µ)-FULL STOP.By this parameter, the user can choose the degree of convergence necessary before terminating.On complex objective functions that take a long time to evaluate, a small κ may be favorable leading to earlier termination and, hence, saving a lot of running time for applying further subsequent optimization methods to further improve the solution quality.On simple, easily to be evaluated functions such as SPHERE, one may choose κ = D, which results in (γ, µ)-FULL STOP, as already mentioned.With (κ, γ, µ)-PARTIAL STOP, both interfering scenarios are accounted for.There is no differentiation between dimensions, so it makes no difference that like in the first scenario some dimensions d are forced such they have reached (or are close to) their stagnation value ∂ σ d , while in other dimensions still almost no forcing takes place.Similarly, it might happen, like in the second scenario, that in all dimensions the frequency is less than the stagnation value, but the overall sum has reached the stagnation value.We executed a first series of experiments similar to the case with the criterion (γ, µ)-FULL STOP.Again, N = 5 particles, D = 15 dimensions, δ = 10 −7 , interval length µ = 50 000, σ stag (N, D, µ) = 318 350, and γ = 1350.Tables 2 and 3 show the results of the experiments.We tested (κ, γ, µ)-PARTIAL STOP with κ = 2 and κ = 8, and for the fixed iteration limit, we used 15 000 000 iterations.We repeated the experiments 500 times.We conclude that in most cases a small value of κ is sufficient without a significant loss in the quality of the returned solution.However, the ROSENBROCK measurement suggests that on complex objective functions, a small value of κ might lead to too early termination.
Tables 4 and 5.As presumed for the simple functions, f-PSO with (κ, γ, µ)-PARTIAL STOP terminates significantly earlier with only a small decline in the quality of the returned solution.For ROSENBROCK, the difference in the quality of the returned solution with different choices of the parameter κ becomes more distinct.Furthermore it should be noted that as shown in Fig. 5 the standard deviation increases drastically with shorter intervals.In practice a small tolerance is used when testing if the stagnation value is reached.Given this standard deviation on small intervals, a tolerance that compensates this standard deviation can lead to a too early termination when the stagnation value is not reached and the swarm would recover from the stagnation.This leads to a tradeoff between interval length and the risk of a termination when the swarm has not reached or is close to a local optimum.1).Geometric Mean over 500 runs of the stopped f-PSO with N = 5, D = 15, µ = 5 000 µ = 5 000 µ = 5 000, σ σ σ stag (N, D, µ) = 31 835 (N, D, µ) = 31 835 (N, D, µ) = 31 835, γ = 1350, and δ = 10 −7 , and κ = 2 and κ = 8, resp., against an 'unstopped' f-PSO terminated after 15 000 000 iterations.

Conclusions
This paper focused on developing stopping criteria for the f-PSO algorithms.We showed that at an optimum point the swarm behaves like a blind searching algorithm and that therefore the number of forced steps in a time interval, the absolute forcing frequency, is larger than an objection functionindependent number.Thus, reaching this frequency it can be presumed that the swarm is close to an optimum solution, and it can be stopped.
f-PSO algorithm.(a) Let σ (I, d) denote the number of times forced velocity updates (Line 10) have been executed in interval I for dimension d, d ∈ {1, . . ., D}. σ (I, d) is called the absolute forcing frequency in dimension d over interval I. (b) σ (I) = ∑ D d=1 σ (I, d) is the total number of forced velocity updates (Line 10) in interval I counted over all dimensions.σ (I) is called the absolute forcing frequency over interval I. Analogously, the relative forcing frequency is σ (I, d)/|I| and σ (I)/|I|, resp.

Figure 2 :
Figure 2: Optimizing the ROSENBROCK function: development of the absolute forcing frequencies σ (I i , d), d ∈ {1, . .., 30}, over the intervals I i = [0 . . .50 000 • i] compared to the development of the function value f (G) of the global best position G after 50 000 • i iterations (single red line).N = 3 particles and δ = 10 −7 were used.Noticeable is the constant behavior of σ (I i , d) with f (G) far away from the optimal value of 0, and the separation into forcing dimensions and non-forcing dimensions.

Figure 3 :Figure 4 :Figure 5 :
Figure 3: Optimizing the SPHERE function: Absolute forcing frequency σ (I) over intervals I of length |I| = 50 000 relative to the number D of dimensions at the global optimum with N = 5 particles and δ = 10 −7 , varying the number D of dimensions.Shown are the average value and standard deviation of 100 trials taking the average values of 10 intervals per trial.
Figure6: Frequency of the distance to the global attractor of one particle in one dimension each iteration for 5 000 000 iterations in a run with 5 particles and 15 dimensions and δ = 10 −7 initialized at the global optimum.