A multitier adaptive grid algorithm for the evolutionary multiobjective optimisation of complex problems
 827 Downloads
Abstract
The multitier Covariance Matrix Adaptation Pareto Archived Evolution Strategy (mCMAPAES) is an evolutionary multiobjective optimisation (EMO) algorithm for realvalued optimisation problems. It combines a nonelitist adaptive grid based selection scheme with the efficient strategy parameter adaptation of the elitist Covariance Matrix Adaptation Evolution Strategy (CMAES). In the original CMAPAES, a solution is selected as a parent for the next population using an elitist adaptive grid archiving (AGA) scheme derived from the Pareto Archived Evolution Strategy (PAES). In contrast, a multitiered AGA scheme to populate the archive using an adaptive grid for each level of nondominated solutions in the considered candidate population is proposed. The new selection scheme improves the performance of the CMAPAES as shown using benchmark functions from the ZDT, CEC09, and DTLZ test suite in a comparison against the \((\mu + \lambda )\) MultiObjective Covariance Matrix Adaptation Evolution Strategy (MOCMAES). In comparison with MOCMAES, the experimental results show that the proposed algorithm offers up to a 69 % performance increase according to the Inverse Generational Distance (IGD) metric.
Keywords
Multiobjective optimisation Evolutionary algorithms Evolution strategies Covariance matrix adaptation Adaptive grid archiving1 Introduction
The quality of the set of candidate solutions to a multiobjective optimisation problem can be assessed using three criteria: proximity to the true Pareto front (i.e. how close the set of candidate solutions is to the true global solution set), diversity (i.e. how well distributed the set of candidate solutions is over the true Pareto optimal front) and pertinency (i.e. how relevant the set of candidate solutions is to a decision maker). An ideal approximation set should be uniformly spread across the true Pareto optimal front (Deb 2001), or—in realworld problems at least–that part of it that represents a useful subset of solutions to the problem^{1} (Purshouse and Fleming 2007).
The vast majority of the current stateoftheart Evolutionary Multiobjective Optimisation (EMO) algorithms employ elitism to enhance convergence to the true Pareto optimal front. Elitism ensures some or all of the fittest individuals in a population at generation g are inserted into generation \(g+1\). Using this method, it is possible to prevent the loss of the fittest individuals which are considered to have some of the most valuable chromosomes in the population. However, in many multiobjective optimisation problems, solutions exist which may not be considered elite due to their objective value in regard to the population, but may contain useful genetic information. This genetic information can be utilised later in the search to move into unexplored areas of the objective space, but due to elitism and nondominated sorting schemes it may be abandoned in the early stages of the search.
The aim of this study is to counter the potential negative effects resulting from elitist approaches to selection (for example, the bounded Pareto archive used in the Covariance Matrix Adaption—Pareto Archived Evolution Strategy (CMAPAES)) by not only preserving elite solutions but also focusing part of the function evaluation budget on nonelitist solutions that have the potential to contribute useful genetic information in the future. To achieve this, a novel multitier adaptive grid selection scheme is developed and combined with the existing CMAPAES algorithm, in a new augmented algorithm named the Multitier Covariance Matrix Adaptation Pareto Archived Evolution Strategy. This novel algorithm sacrifices a portion of the function evaluation budget in favour of producing diverse approximation sets consisting of solutions from areas of the objective space which are difficult or impossible to obtain with an elitism approach. With this feature, the final approximation set offers a better representation of the tradeoff surface, therefore allowing the decision maker to make a more informed selection. The performance of this new algorithm is then evaluated on several benchmarking test suites from the literature.
The paper is organised as follows: Sect. 2 introduces the field of evolutionary multiobjective optimisation and its performance characteristics, Sect. 3 introduces the CMAPAES algorithm and novel multitier adaptive grid algorithm, Sect. 4 contains the experimental setup and methods of performance assessment, Sect. 5 presents and discusses the results, and Sect. 6 draws some conclusions as well as suggesting future research direction.
2 Background
2.1 Evolutionary algorithms
Evolutionary algorithms (EAs) are an optimisation technique inspired by some of the concepts behind natural selection and population genetics and are capable of iteratively evolving a population of candidate solutions to a problem (Goldberg 1989). They both explore the solution space of a problem (by using variation operators such as mutation and recombination) and exploit valuable information present in the previous generation of candidate solutions (by using a selection operator which gives preference to the best solutions in the population when creating the next generation of solutions to be evaluated).
One of the main reasons evolutionary algorithms are applicable across many different problem domains (including those where conventional optimisation techniques struggle) is their direct use of evaluation function information, rather than derivative information or other auxiliary knowledge. Derivative information (for example) can be extremely difficult to calculate in many realworld problems because the evaluation of candidate solutions can be expensive. Evolutionary algorithms are also robust to noisy solution spaces because of their populationbased nature. This means that each generation contains more information about the shape of the fitness landscape than would be available to conventional, nonpopulationbased optimisation methods (Michalewicz and Fogel 2000).
Evolutionary algorithms have also been used in combination with other approaches to optimisation to form hybrid algorithms which have been applied successfully to realworld problems (Sfrent and Pop 2015). Hyperheuristics are a methodology in search and optimisation which are concerned with choosing an appropriate heuristic or algorithm in any given optimisation context (Burke et al. 2003), and can operate on metaheuristics. Hybrid algorithms indicate the benefits of using an approach which aim to combine existing algorithms and heuristics such that a more general approach can be taken to optimisation.
2.2 Evolutionary multiobjective optimisation

The proximity of the approximation set to the true Pareto front.

The diversity of the distribution of solutions in the approximation set.

The pertinence of the solutions in the approximation set to the decision maker.
Conventional multiobjective optimisation methods often fail to satisfy all these requirements, with methods such as the weighted sum method (Hwang and Masud 1979) and the goal attainment method (Gembicki 1974) only capable of finding a single point from the approximation set rather than a diverse distribution of potential solutions. This means that such algorithms do not fully capture the shape of the tradeoff space without running the optimisation routine many times. In contrast, evolutionary algorithms (EAs) iteratively evolve a population of candidate solutions to a problem in parallel and are thus capable of finding multiple nondominated solutions. This results in a diverse set of potential solutions to choose from, rather than a single solution that may not meet the required performance criteria.
2.3 Obtaining good proximity
The primary goal in evolutionary multiobjective optimisation is finding an approximation set that has good proximity to the Pareto front. This ensures that the candidate solutions in this approximation set represent optimal tradeoffs between objectives. The early approaches to evolutionary multiobjective optimisation were primarily concerned with guiding the search towards the Pareto front, reflecting the importance of this goal.
Convergence to the Pareto front is mainly driven by selection for variation, where the best candidate solutions are assigned the highest fitness (and thus have the best chance of contributing to the next generation). Several techniques have been proposed to solve the problem of assigning scalar fitness values to individuals in the presence of multiple objectives—with Paretobased methods generally being considered the best. Several variants of Paretobased fitness assignment methods exist (see Zitzler et al. 2004 for more information), but the general procedure is to rank individuals in the approximation set according to some dominance criterion, and then map fitness values to these ranks (often via a linear transformation). Mating selection then proceeds using these fitness values.
The proximity of the approximation set to the true Pareto front can be enhanced by the use of elitism. Elitism aims to address the problem of losing good solutions during the optimisation process (Zitzler et al. 2004), either by maintaining an external population of nondominated solutions (commonly referred to as an archive), or by using a \((\mu + \lambda )\) type environmental selection mechanism. Studies have shown that elitist MOEAs perform favourably when compared to their nonelitist counterparts (Zitzler and Thiele 1999; Zitzler et al. 2000a). Elitism has also been shown to be a theoretical requirement to guarantee convergence of an MOEA in the limit condition (Rudolph and Agapie 2000).
In archivebased elitism, the archive can be used either just to store good solutions generated by the MOEA or can be integrated into the algorithm with individuals from the archive participating in the selection process. Some mechanism is often needed to control the number of nondominated solutions in the archive, since the archive is usually a finite size and the number of nondominated individuals can potentially be infinite. Densitybased measures to preserve diversity are commonly used in this archive reduction—for example, the Pareto Archived Evolution Strategy (Knowles and Corne 2000a) uses an adaptive crowding procedure to preserve diversity (see later).
An alternative elitist strategy is the \((\mu + \lambda )\) population reduction scheme, where the parent population and the child population compete against each other for selection. This scheme originated in Evolution Strategies and forms the basis of the environmental selection scheme used in algorithms such as NSGAII (Deb et al. 2002a) and, more recently, the MultiObjective Covariance Matrix Adaptation Evolutionary Strategy (MOCMAES) (Igel et al. 2007). In both these algorithms, a two level sorting process is used, with Pareto dominance as the primary sorting criteria and population density as a secondary sorting criteria (used as a tiebreaker amongst individuals having the same level of nondominance).
MOCMAES is a stateoftheart elitist multiobjective evolutionary optimisation technique that builds upon the powerful covariance matrix adaptation evolution strategy (CMAES) realvalued singleobjective optimiser (Hansen and Ostermeier 2001; Hansen et al. 2003). The key features of CMAES are that it is invariant against linear transformations of the search space, performs extremely well across a broad spectrum of problems in the continuous domain (Auger and Hansen 2005), and is robust to the initial choice of parameters (due to its advanced selfadaptation strategy). These make the CMAES algorithm an excellent choice to base a multiobjective evolutionary optimisation on.
Two variants of MOCMAES exist in the literature: the sMOCMAES which achieves diversity using the contributing hypervolume measure (or smetric) introduced by (Zitzler and Thiele 1998), and the cMOCMAES which achieves diversity using the crowdingdistance measure introduced in NSGAII. Whilst initial results have shown that MOCMAES is extremely promising, it is as yet mostly untested on realworld engineering problems. Some results show that MOCMAES struggles to converge to good solutions on problems with many deceptive locally Pareto optimal fronts—a feature that can be common in realworld problems (Voß et al. 2010).
In the original MOCMAES, a mutated offspring solution is considered to be successful if it dominates its parent. In contrast, (Voß et al. 2010) introduces a new MOCMAES variant which considers a solution successful if it is selected to be in the next parent population, introduces a new update rule for the selfadaptive strategy, and conducts a comparison of MOCMAES variants on synthetic test functions consisting of up to three objectives. MOCMAES with the improved update rule is shown to perform substantially better than the original algorithm and thus is used for comparison in Sect. 5 of this paper.
2.4 Obtaining good diversity
Most EMO algorithms use density information in the selection process to maintain diversity in the approximation set. However, diversity preservation has often been seen as a secondary consideration (after obtaining good proximity to the Pareto front). This is because, as Bosman and Thierens (2003) state:
“...since the goal is to preserve diversity along an approximation set that is as close as possible to the Pareto optimal front, rather than to preserve diversity in general, the exploitation of diversity should not precede the exploitation of proximity”.
Goldberg (1989) initially suggested the use of a niching strategy in EMO to maintain diversity, with most of the first generation of Paretobased EMO algorithms using the concept of fitness sharing from singleobjective EA theory (Fonseca and Fleming 1993; Horn et al. 1994; Srinivas and Deb 1994). However, the success of fitness sharing is strongly dependent on the choice of an appropriate niche size parameter, \(\sigma _\mathrm{share}\). Whilst several authors proposed guidelines for choosing \(\sigma _\mathrm{share}\) (Deb and Goldberg 1989; Fonseca and Fleming 1993), Fonseca and Fleming (1995) were the first to note the similarity between fitness sharing and kernel density estimation in statistics which then provided the EMO community with a set of established techniques for automatically selecting the niche size parameter, such as the Epanechnikov estimator (Silverman 1986).
A large number of the second generation of Paretobased MOEAs include advanced methods of estimating the population density, inspired by statistical density estimation techniques. These can be mainly classified into histogram techniques (such as that used in PAES (Knowles and Corne 2000a)) or nearest neighbour density estimators (such as that used in SPEA2 (Zitzler et al. 2001) and NSGAII (Deb et al. 2002a)). Other approaches to diversity preservation include the use of hybrid algorithms, such as the Hybrid Immune Genetic Algorithm (HIGA) Istin et al. (2011), which uses an immune component to continuously evolve new solutions and then inject them back into the population of an EA. These estimates of population density can be used in both mating selection and environmental selection. In mating selection, these density estimates are commonly used to discriminate between individuals of the same rank. Individuals from a less dense part of the population are assigned higher fitness and thus have a higher chance of contributing to the next generation.
Density estimation in environmental selection is commonly used when there exists more locally nondominated solutions than can be retained in the population. For example, in archivebased elitism, densitybased clustering methods are often used to reduce the archive to the required size. Nondominated solutions from sparser regions of the search space are again preferred over those from regions with higher population densities, with the aim being to ensure that the external population contains a diverse set of candidate solutions in close proximity to the Pareto front.
When an archive has reached capacity and a new candidate solution is to be archived, the information tracked by the AGA is used to replace a solution in the grid location containing the highest number of solutions. When a candidate solution is nondominated in regard to the current solution and the archive, the grid information is used to select the solution from the least populated grid location as the current (and parent) solution.
The AGA concept used in PAES later inspired several researchers and was altered and deployed in multiple EMO algorithms such as the Pareto Envelopebased Selection Algorithm (PESA) (a populationbased version of PAES) (Corne et al. 2000), the Micro Genetic Algorithm (Coello Coello and Pulido 2001), and the Domination Based MultiObjective Evolutionary Algorithm (\(\epsilon \)MOEA) (Deb et al. 2005).
2.5 Issues with elitism
Whilst elitism has been almost universally adopted in the current state of the art for evolutionary multiobjective optimisers, in many multiobjective optimisation problems solutions may exist which are not considered elite due to their objective value in regard to the population but may still contain useful genetic information. This genetic information can be utilised later in the search to move into unexplored areas of the objective space but, due to elitism and nondominated sorting schemes, it may be abandoned in the early stages of the search.
By observing this twoobjective plot of the approximation set, it can be seen that the elitist EMO algorithm has converged to an approximation set which is missing three distinct areas containing solutions in comparison with the true Pareto optimal front plotted in Fig. 4. The genetic information which would have potentially found these missing areas was discarded by the algorithm during the search process due to the use of elitism and nondominated sorting. This is a difficulty that occurs in the CEC09 UF1 test problem because of its complicated Pareto optimal set, which has some regions that are easier to reach. In these cases, elitist EMO algorithms will focus selection on these more dominant solutions and converge further into that area of the Pareto optimal set, discarding individuals which may have been only a few generations away from producing nondominated solutions in unexplored areas of the objective space.
3 CMAPAES
The algorithm execution life cycle for CMAPAES has been illustrated in Fig. 6. CMAPAES begins by initialising the algorithm variables and parameters; these include the number of grid divisions used in the AGA, the archive for storing Pareto optimal solutions, the parent vector Y, and the covariance matrix. An initial current solution is then generated at random, which is evaluated and then the first to be archived (without being subjected to the PAES archiving procedure). The generational loop then begins, and the square root of the covariance matrix is resolved using Cholsky decomposition (as recommended by Beyer and Sendhoff 2008) which offers a less computationally demanding alternative to spectral decomposition. The \(\lambda \) candidate solutions are then generated using copies of the current solution and the CMAES procedure for mutation before being evaluated. The archive is then merged with the newly generated offspring and subjected to Pareto ranking, and this assigns a rank of zero to all nondominated solutions, and a rank reflecting the number of solutions that dominate the inferior solutions. The population is then purged of the inferior solutions so that only nondominated solutions remain before being fed into the PAES archiving procedure. After the candidate solutions have been subjected to the archiving procedure and the grid has been adapted to the new solution coverage of objective space, the archive is scanned to identify the grid location with the smallest population, this is considered the lowest density grid population (ldgp). The solutions from the lowest density grid population are then spliced onto the end of the first \(\mu ldgp\) of the Pareto rankordered population to be included in the adaptation of the covariance matrix, with the aim to improve the diversity of the next generation by encouraging movement into the least dense area of the grid. After the covariance matrix is updated, the generational loop continues onto its next iteration until the termination criteria is satisfied (maximum number of generations).
CMAPAES has been benchmarked against NSGAII and PAES in Rostami and Shenfield (2012) on the ZDT synthetic test suite. Two performance metrics were used to compare the performance in terms of proximity (using the generational distance metric) and diversity (using the spread metric). CMAPAES displayed superior performance (the significance of which was supported with randomisation testing) in returning an approximation set close to or on the true Pareto optimal front as well as maintaining diversity amongst solutions in the set.
CMAPAES has also been benchmarked against the MOCMAES algorithm in Rostami (2014), using the hypervolume indicator as a measure of performance. In this study, both the algorithms considered demonstrated comparable performance across multiple test problems. The significance of which was supported by the use of nonparametric testing.
3.1 A novel multitier adaptive grid algorithm
The new multitier AGA aims to prevent a population from prematurely converging as a result of following only the dominant (i.e. elite) solutions which may be discovered early in the optimisation process. This common optimisation scenario often results in genetic drift and consequently a final approximation set with solutions clustered around these elite solutions. This prevention is achieved by dividing an optimisation function evaluation budget and investing a percentage of this budget in to nonelite solutions. These solutions which appear nonelite early on in the optimisation process may potentially contain genetic information that would contribute to finding undiscovered areas of the objective space later in the search.
The algorithm pseudocode for this new multitier approach is listed in Algorithm 1, which is executed from line 14 of the Multitier Covariance Matrix Adaptation Pareto Archived Evolution Strategy (mCMAPAES) execution life cycle presented in Algorithm 2. This new optimisation algorithm which builds upon the algorithmic components (AGA and CMA) outlined in Fig. 6 is referred to as the mCMAPAES. First, the candidate population is divided into subpopulations based on their nondominated rank using NSGAII’s fast nondominated sort. If the size of any subpopulation exceeds \(\mu \), then the standard AGA scheme is applied to it with a maximum archive capacity of \(\mu \), resulting in a number of rankordered archives each with a maximum capacity of \(\mu \). Then, a single population of size \(\mu \) plus the budget for nonelite individuals \(\beta \) is produced, for example, if \(\beta \) is set as 10 % for a \(\mu \) population of 100, then a population of size \(100 \times 1.10\) is to be produced. Next, the multitier archives containing the first \(\mu \times \beta \) solutions are merged with no size restriction (meaning the merged archive size can be greater than \(\mu \times \beta \)). This merged archive is then subjected to a nonelite AGA (ensuring nonelite solutions are not instantly discarded) with an archive capacity of \(\mu \), producing a population of individuals to be selected as parents for the next generation.
4 Experimental design and performance assessment
4.1 Experimental setup
In order to evaluate the performance of mCMAPAES on multiobjective test problems, a pairwise comparison between mCMAPAES and MOCMAES on selected benchmark problems from the literature (consisting of upto 3 objectives) has been conducted. MOCMAES (as outlined in Sect. 2.3) is a stateoftheart algorithm which uses the CMA operator for variance much like mCMAPAES.
Algorithm configurations used when benchmarking MOCMAES and mMAPAES
Parameter  MOCMAES  mCMAPAES 

\(\mu \)  2D(100), 3D(300)  2D(100), 3D(300) 
\(\lambda \)  2D(100), 3D(300)  2D(100), 3D(300) 
Archive capacity  –  2D(100), 3D(300) 
Multitier budget  –  10 % 
Divisions  –  10 
The ZDT, DTLZ and CEC2009 test suites have been selected for the benchmarking and comparison of mCMAPAES and MOCMAES (see Sect. 4.2). These test suites will pose both MOEAs with difficulties which are likely to be encountered in many realworld multiobjective optimisation problems, in both twodimensional and threedimensional objective spaces (allowing for feasible comparison with MOCMAES which relies on the hypervolume indicator for secondary sorting and is thus computationally expensive in highdimensional search spaces).
The metric used for performance assessment is the popular Inverted Generational Distance (IGD) indicator described in Sect. 4.3. The IGD indicator will be used at each generation in order to assess performance and compare both algorithms on not just the IGD quality of the final approximation set, but also the IGD quality over time. In order to comply with the CEC2009 competition rules (as described in Zhang et al. 2008b), both mCMAPAES and MOCMAES have been executed 30 times on each test function to reduce stochastic noise. This sample size is seen as sufficient because of the limited benefit of producing more than 25 samples (discussed in Sect. 4.1.1).
4.1.1 Sample size sufficiency
Selecting a sufficient number of samples when comparing optimisers is critical. The sample size of 25, in order to reduce stochastic noise, is reoccurring in the evolutionary computation literature (e.g. Yang et al. 2008; Zamuda et al. 2007; Falco et al. 2012; García et al. 2009; Wang et al. 2011). To prove the sufficiency of this sample size, a large number of hypervolume indicator value samples have been produced by executing mCMAPAES 200 times on the DTLZ1 synthetic test problem.
sizes, evolution paths and covariance matrices of the successful solutions are updated.
4.2 Multiobjective test suites
The performance of the novel mCMAPAES algorithm is compared to MOCMAES across three different realvalued, synthetic, test suites: the widely used ZDT biobjective test suite proposed in Zitzler et al. (2000b), the scalable DTLZ multiobjective test suite proposed in Deb et al. (2002b) and the unconstrained functions from the CEC2009 multiobjective competition test suite proposed in Zhang et al. (2008b). The configurations used for these test problems and some of their salient features are shown in Table 2
Parameter configurations used for the ZDT, DTLZ and CEC09 test suites
Problem  # Var  # Obj  Salient features 

ZDT1  30  2  Convex front 
ZDT2  30  2  Concave front 
ZDT3  30  2  Disconnected front 
ZDT4  10  2  Convex front, many local optima 
ZDT6  10  2  Concave front, nonuniform distribution 
DTLZ1  7  3  Linear front 
DTLZ2  12  3  Spherical front 
DTLZ3  12  3  Spherical front, many local optima 
DTLZ4  12  3  Spherical front, nonuniform distribution 
DTLZ5  12  3  Spherical front, difficult to find true front 
DTLZ6  12  3  Disconnected front 
DTLZ7  22  3  Disconnected front 
UF1  30  2  Nonlinear decision space 
UF2  30  2  Nonlinear decision space 
UF3  30  2  Many local fronts 
UF4  30  2  Nonconvex front 
UF5  30  2  Discrete points on a linear hyperplane 
UF6  30  2  Disconnected front 
UF7  30  2  Linear hyperplane front 
UF8  30  3  Spherical front 
UF9  30  3  Disconnected front 
UF10  30  3  Spherical front 
4.3 Performance assessment
IGD results from 30 executions of mCMAPAES and MOCMAES on the ZDT and CEC09 test suites with two problem objectives
2D  mCMAPAES  MOCMAES  

Best  Mean  Worst  Best  Mean  Worst  p value  %IGD  
ZDT1  0.00628  0.00657  0.00686  0.00813  0.00936  0.01031  1.4e−09  69.23  + 
ZDT2  0.00592  0.00614  0.00639  0.00989  0.01172  0.01511  1.4e−09  60.72  + 
ZDT3  0.00574  0.00609  0.00676  0.00552  0.00594  0.00643  0.0625  \(\)12.10  = 
ZDT4  1.80044  6.17983  11.44563  2.85512  8.35397  14.56593  0.0232  17.03  + 
ZDT6  0.01132  0.01279  0.01406  0.04788  0.08938  0.20901  1.4e−09  38.74  + 
UF1  0.03762  0.05824  0.06579  0.05044  0.07228  0.12375  1.1e−06  17.00  + 
UF2  0.01359  0.02006  0.02687  0.02117  0.03496  0.05235  5.5e−08  38.44  + 
UF3  0.04869  0.07992  0.12647  0.06044  0.08129  0.10133  0.7269  1.76  = 
UF4  0.05925  0.06431  0.06942  0.07661  0.08261  0.09722  1.4e−09  48.20  + 
UF5  0.49880  0.72982  1.04816  0.87997  1.04873  1.26644  8.3e−09  41.54  + 
UF6  0.08817  0.12736  0.22802  0.09314  0.11268  0.22469  0.010432  \(\)10.50  – 
UF7  0.01791  0.02431  0.03226  0.03306  0.06434  0.12773  1.4e−09  36.45  + 
IGD results from 30 executions of mCMAPAES and MOCMAES on the DTLZ and CEC09 test suites with three problem objectives
3D  mCMAPAES  MOCMAES  

Best  Mean  Worst  Best  Mean  Worst  p value  %IGD  
UF8  0.13308  0.18188  0.23023  0.16091  0.23432  0.24924  3.6e−08  45.14  + 
UF9  0.07381  0.07877  0.08795  0.06755  0.07440  0.07911  5.4e−05  \(\)21.45  – 
UF10  0.64046  0.97907  1.34102  1.33073  1.90805  2.89107  1.6e−09  41.28  + 
DTLZ1  0.60928  3.11971  5.72913  2.11988  10.1829  20.9531  1.2e−06  34.72  + 
DTLZ2  0.03919  0.04005  0.04077  0.04207  0.04491  0.04939  1.4e−09  7.06  + 
DTLZ3  22.4023  50.7571  102.51  171.175  188.531  229.147  1.4e−09  66.64  + 
DTLZ4  0.02459  0.03090  0.04093  0.03181  0.04411  0.07016  5.5e−08  28.99  + 
DTLZ5  0.00152  0.00174  0.00201  0.00190  0.00213  0.00259  8.3e−09  11.21  + 
DTLZ6  0.11059  0.32162  0.65582  0.19705  0.42455  0.71631  0.01701  16.99  + 
DTLZ7  0.05268  0.05783  0.06449  0.05824  0.06653  0.07449  2.9e−08  39.89  + 
The IGD metric measures how well the obtained approximation set represents the true Pareto optimal front which is provided as a large reference set. This is calculated by finding the minimum Euclidean distance of each point of the approximation set to points in the reference set. Lower IGD values indicate a better quality approximation set with IGD values of 0, indicating all the solutions in the approximation set are in the reference set and cover all the Pareto front.
The IGD measure has been employed in the performance assessment of algorithms in much of the multiobjective optimisation and evolutionary computation literature (e.g. Zhang et al. 2008a, 2010; Tiwari et al. 2009; Chen et al. 2009; Nasir et al. 2011).
4.4 Statistical comparison of stochastic optimisers
Nonparametric testing is becoming more commonly used in the literature to statistically contrast the performance of evolutionary algorithms in many experiments (García et al. 2010; Derrac et al. 2012; Li et al. 2012; Epitropakis et al. 2012; Hatamlou 2013; Civicioglu 2013).
5 Results
The results from the experiments described in Sect. 4 have been produced and presented in a number of formats in order to allow for a better assessment of each algorithms performance.
Overall, mCMAPAES outperformed MOCMAES on all but 3 (ZDT3, UF6 and UF9) of the 22 test functions, producing better performing worst, mean and best approximation sets.
The mean of the IGD metric at each generation has been plotted and presented in Figs. 9 and 10 for the twoobjective test functions and Figs. 11 and 12 for the threeobjective test functions. These plots illustrate the rate of IGD convergence from the initial population to the final population.
mCMAPAES significantly outperforms the MOCMAES on most of the test functions used in this comparison. However, as a consequence of investing a percentage of the maximum number of function evaluations in nonelite solutions, it can be observed in Figs. 9, 10, 11, 12 that the convergence of the algorithm is slower in most cases (more so in the twoobjective test functions). This suggests that in experiments where the number of function evaluations is not constrained to a low number, the mCMAPAES will outperform MOCMAES.
It can be observed in Figs. 9, 10, 11, 12 that the mean IGD for MOCMAES oscillates or rises on some test functions over time. This issue is most visible on UF4 (where the mean IGD for MOCMAES can be seen to oscillate over time) and on DTLZ3 (where the mean IGD for MOCMAES can be seen to improve in performance until 200 generations and then worsen gradually until termination). This issue is due to MOCMAES being dependent on the hypervolume indicator entirely for diversity preservation which, when paired with its elitism scheme, ends up gradually reducing the IGD quality of an approximation set once a difficult area of the search space is encountered.
On the UF3 test function, it can be observed (in Fig. 13) that, although the MOCMAES median IGD outperforms mCMAPAES, mCMAPAES achieved a better interquartile range and a far better total range—achieving the best approximation set for that test function. A similar result can be seen in the performance on UF6 where mCMAPAES also achieves the best approximation set but is outperformed by MOCMAES on the median values of the IGD results.
6 Conclusion
In this paper, a multitier AGA scheme has been introduced and incorporated into the CMAPAES algorithm to create mCMAPAES. mCMAPAES improves the quality of the produced final approximation set by investing a percentage of the allowed function evaluation budget in nonelite but potentially successfully solutions. With this approach, mCMAPAES is able to find portions of the Pareto optimal front which remain unexplored by elitist approaches. Experiments and statistical analysis presented in this study show that with CEC09 competition compliant benchmarking configurations, mCMAPAES significantly outperforms MOCMAES on all but 4 of the 22 considered synthetic test problems, and out of these 4, MOCMAES only performs statistically significantly better on 2 test functions.
When observing the IGD values at each generation, it can be seen that in some cases the IGD of the final population is higher than some of the generations before it, this is due to the nonelite solutions invested in at each generation being a factor right to the end of the algorithm. This suggests that in further work the algorithm may benefit from either an offline archive which the algorithm selects from at the end of the optimisation process or a final approximation set selection scheme which uses the last two generations of the optimisation process, including nondominated solutions only.
The results indicate a clear tradeoff between mCMAPAES and MOCMAES. In the majority of the benchmarks, MOCMAES appears to offer a faster rate of convergence. However, this comes at the cost of premature convergence very early in the optimisation process. In contrast, mCMAPAEs offer a slower rate of convergence throughout the entire optimisation process, with steady improvement until the end of the function evaluation budget. Unlike MOCMAES, mCMAPAES does not subject the entire nondominated population to the contributing hypervolume indicator. By not doing so, mCMAPAES remains computationally lightweight, unlike MOCMAES which becomes computationally infeasible as the number of problem objectives increase. By investing a portion of the function evaluation budget in nonelite solutions, areas of the Pareto optimal front which are difficult to obtain can be discovered later on in the optimisation process. This results in improved diversity and coverage in the produced approximation sets.
Future works will further investigate the possibility for selfadaptation of the mCMAPAES algorithm parameter which defines the budget for nonelite individuals (\(\beta \)). A current limitation of mCMAPAES requires the manual configuration of the \(\beta \) parameter, which may result in inefficient usage of the function evaluation budget when parameters such as the population size and the number of problem objectives change.
Footnotes
 1.
For example, in the design of automotive engines, there is typically a tradeoff between torque generated and emissions produced. Designs at the extreme ends of this tradeoff surface (i.e. with good emissions but poor torque—or vice versa) are usually not very useful for production automobiles.
 2.
The t value is the difference between the means of the datasets divided by the standard error.
Notes
Compliance with ethical standards
Conflict of interest
The authors declare that they have no conflict of interest.
Ethical approval
This article does not contain any studies with human participants or animals performed by any of the authors.
References
 Auger A, Hansen N (2005) A restart CMA evolution strategy with increasing population size. In: The 2005 IEEE congress on evolutionary computation, 2005, IEEE, vol 2, pp 1769–1776Google Scholar
 Beyer H, Sendhoff B (2008) Covariance matrix adaptation revisited–the cmsa evolution strategy. In: Parallel problem solving from nature–PPSN X. Springer, Berlin, pp 123–132Google Scholar
 Bosman PAN, Thierens D (2003) The balance between proximity and diversity in multiobjective evolutionary algorithms. IEEE Trans Evol Comput 7(2):174–188CrossRefGoogle Scholar
 Burke E, Kendall G, Newall J, Hart E, Ross P, Schulenburg S (2003) Hyperheuristics: An emerging direction in modern search technology. In: International series in operations research and management science, pp 457–474Google Scholar
 Chen CM, Chen YP, Zhang Q, (2009) Enhancing moea, d with guided mutation and priority update for multiobjective optimization. In: IEEE congress on evolutionary computation, (2009) CEC’09, IEEE, pp 209–216Google Scholar
 Civicioglu P (2013) Backtracking search optimization algorithm for numerical optimization problems. Appl Math Comput 219(15):8121–8144MathSciNetMATHGoogle Scholar
 Coello Coello CA, Cortés NC (2005) Solving multiobjective optimization problems using an artificial immune system. Genet Program Evol Mach 6(2):163–190CrossRefGoogle Scholar
 Coello Coello CA, Pulido GT (2001) A microgenetic algorithm for multiobjective optimization. In: Evolutionary multicriterion optimization. Springer, Berlin, pp 126–140Google Scholar
 Corne DW, Knowles JD, Oates MJ (2000) The Pareto envelopebased selection algorithm for multiobjective optimization. In: Parallel problem solving from nature PPSN VI. Springer, Berlin, pp 839–848Google Scholar
 De Falco I, Della Cioppa A, Maisto D, Scafuri U, Tarantino E (2012) Biological invasioninspired migration in distributed evolutionary algorithms. Inf Sci 207:50–65CrossRefGoogle Scholar
 Deb K (2001) Multiobjective optimization. Multiobjective optimization using evolutionary algorithms, pp 13–46Google Scholar
 Deb K, Pratap A, Agarwel S, Meyarivan T (2002a) A fast and elitist multiobjective genetic algorithm: NSGAII. IEEE Trans Evol Comput 6(2):182–197CrossRefGoogle Scholar
 Deb K, Mohan M, Mishra S (2005) Evaluating the \(\varepsilon \)domination based multiobjective evolutionary algorithm for a quick computation of Paretooptimal solutions. Evol Comput 13(4):501–525CrossRefGoogle Scholar
 Deb K, Goldberg DE (1989) An investigation of niche and species formation in genetic function optimization. In: Schaffer JD (ed) Proceedings of the third international conference on genetic algorithms, Morgan Kaufmann, New York, pp 42–50Google Scholar
 Deb K, Thiele L, Laumanns M, Zitzler E (2002b) Scalable multiobjective optimization test problems. In: Proceedings of the congress on evolutionary computation (CEC2002), Honolulu, pp 825–830Google Scholar
 Derrac J, García S, Molina D, Herrera F (2011) A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol Comput 1(1):3–18CrossRefGoogle Scholar
 Derrac J, Cornelis C, García S, Herrera F (2012) Enhancing evolutionary instance selection algorithms by means of fuzzy rough set based feature selection. Inf Sci 186(1):73–92CrossRefGoogle Scholar
 Epitropakis M, Plagianakos V, Vrahatis M (2012) Evolving cognitive and social experience in particle swarm optimization through differential evolution: A hybrid approach. Inf Sci Int J 216:50–92Google Scholar
 Fonseca CM, Fleming PJ (1993) Genetic algorithms for multiobjective optimization: Formulation, discussion and generalization. In: Forrest S (ed) Proceedings of the fifth international conference on genetic algorithms. Morgan Kaufmann, New York, pp 416–423Google Scholar
 Fonseca CM, Fleming PJ (1995) Multiobjective genetic algorithms made easy: Selection, sharing, and mating restriction. In: Zalzala AMS (ed) Proceedings of the first international conference on genetic algorithms in engineering systems: innovations and applications (GALESIA), pp 42–52Google Scholar
 García S, Molina D, Lozano M, Herrera F (2009) A study on the use of nonparametric tests for analyzing the evolutionary algorithms behaviour: a case study on the cec2005 special session on real parameter optimization. J Heuristics 15(6):617–644MATHCrossRefGoogle Scholar
 García S, Fernández A, Luengo J, Herrera F (2010) Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power. Inf Sci 180(10):2044–2064CrossRefGoogle Scholar
 Gembicki FW (1974) Vector optimization for control with performance and parameter sensitive indices. PhD thesis, Case Western Reserve University, ClevelandGoogle Scholar
 Goldberg DE (1989) Genetic algorithms in search, optimization and machine learning. AddisonWesley, ReadingMATHGoogle Scholar
 Hansen N, Müller SD, Koumoutsakos P (2003) Reducing the time complexity of the derandomized evolution strategy with covariance matrix adaptation (cmaes). Evol Comput 11(1):1–18CrossRefGoogle Scholar
 Hansen N, Ostermeier A (2001) Completely derandomized selfadaptation in evolution strategies. Evol Comput 9(2):159–195CrossRefGoogle Scholar
 Hatamlou A (2013) Black hole: a new heuristic optimization approach for data clustering. Inf Sci 222:175–184MathSciNetCrossRefGoogle Scholar
 Horn J, Nafpliotis N, Goldberg DE (1994) A niched Pareto genetic algorithm for multiobjective optimization. In: Proceedings of the congress on evolutionary computation (CEC) 1994. IEEE Press, New York, pp 82–87Google Scholar
 Hwang CL, Masud ASM (1979) Multiple objective decision making—methods and applications, vol 164, Lecture notes in economics and mathematical systems. Springer, BerlinGoogle Scholar
 Igel C, Hansen N, Roth S (2007) Covariance matrix adaptation for multiobjective optimization. Evol Comput 15(1):1–28CrossRefGoogle Scholar
 Istin M, Pop F, Cristea V (2011) Higa: Hybrid immunegenetic algorithm for dependent task scheduling in large scale distributed systems. In: 10th international symposium on parallel and distributed computing (ISPDC), 2011, IEEE, pp 282–287Google Scholar
 Knowles J, Corne D (1999) The pareto archived evolution strategy: A new baseline algorithm for pareto multiobjective optimisation. In: Proceedings of the 1999 congress on evolutionary computation, 1999, CEC 99, IEEE, vol 1Google Scholar
 Knowles JD, Corne DW (2000a) Approximating the nondominated front using the Pareto archived evolution strategy. Evol Comput 8(2):149–172CrossRefGoogle Scholar
 Knowles JD, Corne DW (2000b) Approximating the nondominated front using the Pareto archived evolution strategy. Evol Comput 8(2):149–172CrossRefGoogle Scholar
 Li K, Kwong S, Cao J, Li M, Zheng J, Shen R (2012) Achieving balance between proximity and diversity in multiobjective evolutionary algorithm. Inf Sci 182(1):220–242MathSciNetCrossRefGoogle Scholar
 Michalewicz Z, Fogel DB (2000) How to solve it: modern heuristics. Springer, BerlinMATHCrossRefGoogle Scholar
 Nasir M, Mondal AK, Sengupta S, Das S, Abraham A (2011) An improved multiobjective evolutionary algorithm based on decomposition with fuzzy dominance. In: IEEE congress on evolutionary computation (CEC), 2011, IEEE, pp 765–772Google Scholar
 Purshouse RC (2003) On the evolutionary optimisation of many objectives. PhD thesis, Department of Automatic Control and Systems Engineering, University of Sheffield, Sheffield, UK, S1 3JDGoogle Scholar
 Purshouse RC, Fleming PJ (2007) On the evolutionary optimization of many conflicting objectives. IEEE Trans Evol Comput 11(6):770–784CrossRefGoogle Scholar
 Rostami S (2014) Preference focussed manyobjective evolutionary computation. PhD thesis, Manchester Metropolitan University, ManchesterGoogle Scholar
 Rostami S, Shenfield A (2012) Cmapaes: Pareto archived evolution strategy using covariance matrix adaptation for multiobjective optimisation. In: Computational intelligence (UKCI), 2012, IEEE, pp 1–8Google Scholar
 Rudolph G, Agapie A (2000) Convergence properties of some multiobjective evolutionary algorithms. In: Proceedings of the congress on evolutionary computation (CEC) 2000, IEEE Press, pp 1010–1016Google Scholar
 Sfrent A, Pop F (2015) Asymptotic scheduling for many task computing in big data platforms. Inf Sci 319:71–91MathSciNetCrossRefGoogle Scholar
 Silverman BW (1986) Density estimation for statistics and data analysis. Chapman and Hall, LondonMATHCrossRefGoogle Scholar
 Srinivas N, Deb K (1994) Multiobjective optimization using nondominated sorting in genetic algorithms. Evol Comput 2(3):221–248CrossRefGoogle Scholar
 Tiwari S, Fadel G, Koch P, Deb K (2009) Performance assessment of the hybrid archivebased micro genetic algorithm (amga) on the cec09 test problems. In: IEEE congress on evolutionary computation (2009) CEC’09, IEEE, pp 1935–1942Google Scholar
 Voß T, Hansen N, Igel C (2010) Improved step size adaptation for the mocmaes. In: Proceedings of the 12th annual conference on genetic and evolutionary computation. ACM, New York, pp 487–494Google Scholar
 Wang Y, Cai Z, Zhang Q (2011) Differential evolution with composite trial vector generation strategies and control parameters. IEEE Trans Evol Comput 15(1):55–66CrossRefGoogle Scholar
 Wilcoxon F (1945) Individual comparisons by ranking methods. Biometrics Bull 1(6):80–83CrossRefGoogle Scholar
 Yang Z, Tang K, Yao X (2008) Large scale evolutionary optimization using cooperative coevolution. Inf Sci 178(15):2985–2999MathSciNetMATHCrossRefGoogle Scholar
 Zamuda A, Brest J, Boskovic B, Zumer V (2007) Differential evolution for multiobjective optimization with self adaptation. In: IEEE congress on evolutionary computation, pp 3617–3624Google Scholar
 Zhang Q, Zhou A, Jin Y (2008a) Rmmeda: a regularity modelbased multiobjective estimation of distribution algorithm. IEEE Trans Evol Comput 12(1):41–63CrossRefGoogle Scholar
 Zhang Q, Liu W, Tsang E, Virginas B (2010) Expensive multiobjective optimization by moea/d with Gaussian process model. IEEE Trans Evol Comput 14(3):456–474CrossRefGoogle Scholar
 Zhang Q, Zhou A, Zhao S, Suganthan PN, Liu W, Tiwari S (2008b) Multiobjective optimization test instances for the CEC 2009 special session and competition. University of Essex, Colchester, UK and Nanyang Technological University, Singapore, Special Session on Performance Assessment of MultiObjective Optimization Algorithms, Technical ReportGoogle Scholar
 Zitzler E, Deb K, Thiele L (2000a) Comparison of multiobjective evolutionary algorithms: empirical results. Evol Comput 8(2):173–195CrossRefGoogle Scholar
 Zitzler E, Deb K, Thiele L (2000b) Comparison of multiobjective evolutionary algorithms: empirical results. Evol Comput 8(2):173–195CrossRefGoogle Scholar
 Zitzler E, Laumanns M, Bleuler S (2004) A tutorial on evolutionary multiobjective optimization. In: Gandibleux X, Sevaux M, Sörensen K, T’Kindt V (eds) Metaheuristics for multiobjective optimisation. Lecture Notes in Economics and Mathematical Systems, vol 535. Springer, BerlinGoogle Scholar
 Zitzler E, Laumanns M, Thiele L (2001) SPEA2: improving the strength Pareto evolutionary algorithm. TIK Report 103, TIK Institut fur Technische Informatik und Kommunikationsnetze, Computer Engineering and Networks Laboratory, ETH, Swiss Federal Institute of Technology, Gloriastrasse 35, 8092 ZurichGoogle Scholar
 Zitzler E, Thiele L (1998) An evolutionary algorithm for multiobjective optimization: the strength pareto approach. CiteseerGoogle Scholar
 Zitzler E, Thiele L (1999) Multiobjective evolutionary algorithms: a comparative case study and the strength Pareto approach. IEEE Trans Evol Comput 3(4):257–271CrossRefGoogle Scholar
Copyright information
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.