# Multi objective optimization of computationally expensive multi-modal functions with RBF surrogates and multi-rule selection

## Abstract

GOMORS is a parallel response surface-assisted evolutionary algorithm approach to multi-objective optimization that is designed to obtain good non-dominated solutions to black box problems with relatively few objective function evaluations. GOMORS uses Radial Basic Functions to iteratively compute surrogate response surfaces as an approximation of the computationally expensive objective function. A multi objective search utilizing evolution, local search, multi method search and non-dominated sorting is done on the surrogate radial basis function surface because it is inexpensive to compute. A balance between exploration, exploitation and diversification is obtained through a novel procedure that simultaneously selects evaluation points within an algorithm iteration through different metrics including Approximate Hypervolume Improvement, Maximizing minimum domain distance, Maximizing minimum objective space distance, and surrogate-assisted local search, which can be computed in parallel. The results are compared to ParEGO (a kriging surrogate method solving many weighted single objective optimizations) and the widely used NSGA-II. The results indicate that GOMORS outperforms ParEGO and NSGA-II on problems tested. For example, on a groundwater PDE problem, GOMORS outperforms ParEGO with 100, 200 and 400 evaluations for a 6 dimensional problem, a 12 dimensional problem and a 24 dimensional problem. For a fixed number of evaluations, the differences in performance between GOMORS and ParEGO become larger as the number of dimensions increase. As the number of evaluations increase, the differences between GOMORS and ParEGO become smaller. Both surrogate-based methods are much better than NSGA-II for all cases considered.

## Keywords

Function approximation Multi objective optimization Radial basis functions Expensive optimization Global optimization Evolutionary optimization Parallel Metamodel## 1 Introduction

Multi-objective optimization (MO) approaches involve a large number of function evaluations, which make it difficult to use MO in simulation–optimization problems where the optimization is multi-objective and the nonlinear simulation is computationally expensive and has multiple local minima (multi modal). Many applied engineering optimization problems involve multiple objectives and the computational cost of evaluating each objective is high (e.g minutes to days per objective function evaluation by simulation) [12, 32]. We focus on special algorithms that aim to produce reasonably good results within a limited number of expensive objective function evaluations.

Many authors (for instance, Deb et al. [5] and Zhang et al. [52]) have successfully employed evolutionary strategies for solving multi-objective optimization problems. Even with the improvement over traditional methods, these algorithms require, typically, many objective function evaluations which can be infeasible for computationally expensive problems. Added challenges to multi-objective optimization of expensive functions arise with increase in dimensionality of the decision variables and objectives.

The use of iterative response surface modeling or function approximation techniques inside an optimization algorithm can be highly beneficial in reducing time for computing objectives for multi-objective optimization of such problems. Since the aim of efficient multi-objective optimization is to identify good solutions within a limited number of expensive function evaluations, approximating techniques can be incorporated into the optimization process to reduce computational costs. Gutmann [15] introduced the idea of using radial basis functions (RBF) [3] for addressing single objective optimization of computationally expensive problems. Jin et al. [18] appears to be the first journal paper to combine a non quadratic response surface with a single objective evolutionary algorithm by using neural net approximations. Regis and Shoemaker [36] were the first to use Radial Basis Functions (not a neural net) to improve the efficiency of an evolutionary algorithm with limited numbers of evaluations. Later they introduced a non-evolutionary algorithm Stochastic-RBF [37] , which is a very effective radial basis function-based method for single objective optimization of expensive global optimization problems. These methods have been extended to include parallelism [40]; high-dimensional problems [42]; constraints [38]; local optimization [49, 50]; integer problems [30, 31] and other extensions [39, 41]. Kriging-based methods have also been explored for addressing single objective optimization problems [9, 16, 19]. Jones et al. [19] introduced Efficient Global Optimization (EGO), which is an algorithm for single objective global optimization within a limited budget of evaluations.

Response surface methods have also become popular for multi-objective optimization problems, with kriging-surrogate techniques being the most popular. Various authors have used kriging-based methods by extending the EGO framework for multi-objective optimization of computationally expensive problems. For instance, Knowles [21] combined EGO with Tchebycheff weighting to convert an MO problem into numerous single objective problems. Tsang et al. [53] and Emmerich et al. [10] combined EGO with evolutionary search assistance. Ponweiser et al. [33] and Beume et al. [2] explored the idea of maximizing expected improvement in hypervolume. Authors have also explored the use of other function approximation techniques inside an optimization algorithm, including Radial Basis Functions (RBFs) [13, 20, 25, 45, 51], Support Vector Machines [24, 43] and Artificial Neural Networks [7]. Evolutionary algorithms are the dominant optimization algorithms used in these methods. Some papers also highlight the effectiveness of local search in improving performance of response surface based methods [20, 23, 45].

Various authors have indicated that RBFs could be more effective than other approximation methods in multi-objective optimization of computationally expensive problems with more than 15 decision variables [8, 29, 43]. Various authors have employed RBFs for surrogate-assisted multi-objective optimization of expensive functions with focus on problems with more than 15 decision variables. For instance, Karakasis and Giannakoglou [20] employ RBFs within an MOEA framework and use an inexact pre-evaluation phase (IPE) to select a subset of solutions for expensive evaluations in each generation of an MOEA. They also recommend the use of locally fitted RBFs for surrogate approximation. Georgopoulou and Giannakoglou [13] build upon [20] by employing gradient based refinements of promising solutions highlighted by RBF approximation during each MOEA generation. Santana et al. [45] divide the heuristic search into two phases where the first phase employs a global surrogate-assisted search within an MOEA framework, and rough set theory is used in the second phase for local refinements.

This paper focuses on the use of RBFs and evolutionary algorithms for multi-objective optimization of computationally expensive problems, where the number of function evaluations are limited relative to the problem dimension (e.g. to a few hundred evaluations for the example problems tested here). An added purpose of the investigation is to be able to solve MO problems where the number of decision variables varies between 15 and 25.

To this effect we propose a new algorithm, GOMORS, that combines radial basis function approximation with multi-objective evolutionary optimization, within the general iterative framework of surrogate-assisted heuristic search algorithms. Our approach is different from prevalent RBF based MO algorithms that use evolutionary algorithms [13, 20, 25, 45, 51]. Most RBF based evolutionary algorithms employ surrogates in an inexact pre-evaluation phase (IPE) in order to inexpensively evaluate child populations after every MOEA generation. By contrast our proposed methodology employs evolutionary optimization within each iteration of the algorithm framework to identify numerous potential points for expensive evaluations. Hence, multiple MOEA generations evolve via surrogate-assisted search in each algorithm iteration in GOMORS.

The novelty of the optimization approach is in the amalgamation of various evaluation point selection rules in order to ensure that various value functions are incorporated in selecting some points for expensive evaluations from the potential points identified by surrogate-assisted evolutionary search during each algorithm iteration. The combination of multiple selection rules targets a balanced selection between exploration, exploitation and front diversification. The selection strategy incorporates the power of local search and also ensures that the algorithm can be used in a parallel setting to further improve its efficiency.

## 2 Problem description

**Domination** A solution \(x_{1} \! \in \! {\fancyscript{D}}\) dominates another solution \(x_{2} \! \in \! {\fancyscript{D}}\) if and only if \(f_{i} (x_{1}) \le f_{i} (x_{2})\) for all \(1\! \le \! i \! \le \! k\), and \(f_{i} (x_{1}) < f_{i} (x_{2})\) for at least one \(i \in \{1, \dots , k \}\).

**Non-domination** Given a set of solutions \(S= \{ x_{i} \mid x_{i} \! \in \! {\fancyscript{D}}, 1 \! \le \! i \! \le \! n \}\) , a solution \(x^{*} \! \in \! S\) is non-dominated in \(S\) if there does not exist a solution \(x_j \! \in \! S\) which dominates \(x^{*}\).

**Pareto optimality** A candidate solution \(x^{*} \! \in \! S\) which is non-dominated in \(S\) is called a globally Pareto-optimal solution if \(S = {\fancyscript{D}}\), i.e., \(S\) is the entire feasible domain space of the defined problem.

## 3 Algorithm description

### 3.1 General framework

The algorithm initiates with selection of an initial set of points for costly function evaluations of the objective function set, \(F\). Latin hypercube sampling, a space-filling experimental design scheme proposed by McKay et al. [26] is used for selection of the initial evaluation set. The iterative framework of GOMORS (Step 2) follows, and consists of three major segments: (1) Building a Response Surface Model for approximating expensive objectives (Step 2.1), (2) Applying an Evolutionary algorithm for solving multi-objective response surface problems (Steps 2.2 and 2.3) and (3) Selecting multiple points for expensive evaluations from the solutions of the multi-objective response surface problems, and evaluating them in parallel (Step 2.4). The algorithm terminates after \(M\) expensive evaluations and the output is a non-dominated set of solutions \(P_{M}\), which is an approximation to the true Pareto solution of the problem i.e \(P^{*}\). Sections 3.2–3.4 and the online supplement discuss details of the major steps of the algorithm.

### 3.2 Response surface modeling

The first major component of the iterative framework is the procedure used for approximating the costly functions, \(F\), by the inexpensive surrogates, \(\widehat{F}_{m}(x) = [ \widehat{f}_{m,1}, \dots ,\widehat{f}_{m,k} ]\). A response surface model based on the evaluated points is generated in Step 2.1 of the algorithm and subsequently updated after each iteration. For example artificial neural networks [35], Support Vector Machines (SVM) [47], kriging [44], and radial basis functions (RBFs) [3, 34] could be employed to generate the response surface model.

While kriging has been used in MO [21, 33, 53], the number of parameters to be estimated for kriging meta-models increases quickly with an increase in the number of decision variables [16, 17], and hence, re-evaluating kriging surrogates in each iteration of a kriging based algorithm may itself become a computationally expensive step. Various authors [29, 42] including Manriquez et al. [8] have reported that kriging-based surrogate optimization is not effective for high dimensional problems (approximately defined as problems with more than 15 decision variables). In [8] they demonstrate the relative effectiveness of RBF approximation in tackling such high dimensional problems. The GOMORS algorithm proposed in this paper hence makes use of RBFs as the surrogate model for approximating the costly functions, although other surrogate models could be used in the GOMORS strategy.

### 3.3 Evolutionary optimization

*Surrogate-assisted global search*) and 2.3 (

*Gap Optimization*) aim to find near optimal fronts given that the expensive functions, \(F(x)\), are replaced by the corresponding response surface model approximations, i.e., \( \widehat{F}_{m}(x)\) (see Table 1). In Steps 2.2 and 2.3, two different multi-objective optimization problems are solved on the surrogate surface.

Definitions of sets and variables

Item | Description |
---|---|

\(F(x)= [f_{1}(x),\dots ,f_{k}(x)]\) | Expensive objectives for the multiple-objective optimization problem |

\(P^*=\{x_{1}^{*},\dots ,x_{n}^{*} \}\) | Set of Pareto-optimal solutions given the objectives \(F\) |

\(S_{m} = \{x_1, \dots , x_m \}\) | Set of points which have been evaluated via costly simulation until iteration \(m\) of the optimization algorithm |

\(\widehat{F}_{m}(x) = [ \widehat{f}_{m,1}(x), \dots ,\widehat{f}_{m,k}(x) ]\) | \(\widehat{f}_{m,i}\) is the inexpensive surrogate approximation of the \(i^{th}\) objective, \(f_i\), generated using the set of points, \(S_{m}\) |

\(P_{m}=\{ x \! \in \! S_m \}\) | The |

\(x^{crowd} = \mathop {{{\mathrm{{arg\,max}}}}}\limits _{x \in P_m} [d(x)] \) | The least crowded element in \(P_m\), according to the crowding definition, \(d(.)\), where \(d(x_j)\) indicates the function by [6] to compute crowding distance for \(x_j\) given \((x_j, F(x_j))\) for all \(x_j \in P_m\) |

\(\widehat{P}_{m}^A= \{ x_{1}, \dots , x_{n_A}\}\) | Candidate points obtained by |

\(\widehat{P}_{m}^B= \{ x_{1}, \dots , x_{n_B}\}\) | Candidate points obtained by |

\(H_V(P)\) | Hypervolume [1] of the objective space dominated by the objective vectors corresponding to the set \(P\). The hyperspace is such that \(y \in {\mathbb {R}}^k\) and \(y\) is bounded by a reference vector \(b\) (see Fig. 1 for illustration) |

Steps 2.2 and 2.3 solve the MO problems depicted in Equations 2 and 3 respectively. Since the objective functions of these problems are derived from the surrogate model, solving the MO problems is a relatively inexpensive step. However, the MO problems of Equations 2 and 3 are not trivial and could potentially have non-linear and multi-modal objectives. Hence, we employ evolutionary algorithms for MO optimization of surrogates in Steps 2.2 and 2.3.

Three algorithm were considered as potential alternatives for optimization of surrogates in Steps 2.2 and 2.3, namely, NSGA-II, [5], MOEA/D [52] and AMALGAM [48]. NSGA-II handles the evolutionary search optimization process by ranking and archiving parent and child populations according to a non-domination sorting. MOEA/D [52] uses aggregate functions, and simultaneously solves many single-objective Tchebycheff decompositions of multi-objective problems in an evolutionary generation. AMALGAM [48] is a multi-method evolutionary algorithm, which incorporates search mechanics of various algorithms. Extensive computer experiments on test problems were performed on GOMORS with either of NSGA-II, MOEA/D and AMALGAM as embedded evolutionary schemes (see Section C of the online supplement for a detailed discussion on the experiments) and AMALGAM was identified as the best performing evolutionary algorithm (although the differences were small) embedded in GOMORS.

### 3.4 Expensive evaluation point selection: Step 2.4

Step 2.4 of the GOMORS algorithm determines which of the candidate points are evaluated by the expensive objective functions, \(F(x)\), within an algorithm iteration. The candidate points are obtained from Steps 2.2 and 2.3, and are denoted as \(\widehat{P}_m^A\) and \(\widehat{P}_m^B\) respectively. As mentioned earlier, selection of points for expensive evaluation is a critical step in the algorithm because this is usually by far the most computationally expensive step.

A balance between *exploration* [19], *exploitation* and *diversification* [6] is crucial for selecting points for expensive evaluations from candidate points, \(\widehat{P}_m^A\) and \(\widehat{P}_m^B\). *Exploration* of the decision space aims at selecting points in unexplored regions of the decision space. *Exploitation* aims at exploiting the inexpensive response surface approximations of Step 2.1 to assist in choosing appropriate points for expensive evaluations. *Diversification* strives to ensure that the non-dominated evaluated points are spread out in the objective space.

*exploration*,

*exploitation*and

*diversification*. A detailed framework of Step 2.4 of the algorithm is given below which gives an overview of the selection rules (The \(i^{th}\) rule is referred as Rule \(i\)) (Refer Table 1 for definitions of sets and symbols):

The number of points to be selected for expensive evaluation via each rule i.e, \(t_i\) may vary. However, we performed all our computer experiments with \(t_i= 1\), for Rules 1–4. A point is selected via Rule 0 in each algorithm iteration with a probability of 0.1. Hence, either four or five points are selected for expensive function evaluations in each iteration of the algorithm. The expensive evaluations of points are performed in parallel to further speed up the algorithm.

In order to assess the individual effectiveness of the rules, we performed computer experiments on eleven test problems with the exclusive use of all rules (i.e, one point is selected from one rule, in all algorithm iterations). Furthermore, we also performed experiments with the idea of cycling between all rules in subsequent iterations of GOMORS framework. These different selection strategies were compared against the multiple rule selection strategy for simultaneous selection of evaluation points (described above). Results of the computer experiments indicated that if the value of parallelization is considered, the multiple rule selection strategy outperforms the other strategies employed on our analysis (See Section D of the online supplement for details.). However, if GOMORS is used in a serial setting, i,e, one point is evaluated in each algorithm iteration, cycling between rules, and use of Rule 3 are most beneficial (see section D of online supplement).

## 4 Test problems

In order to test the performance of GOMORS, computational experiments were performed on various test functions and a groundwater remediation problem. Certain characteristics of various problems can lead to inadequate convergence or poor distribution of points on the Pareto front. These characteristics include high dimensionality, non-convexity and non-uniformity of the Pareto front [6], the existence of multiple locally optimal Pareto fronts [4], low density of solutions close to the Pareto front, and existence of complicated Pareto sets (this implies that the Pareto solutions are defined by a complicated curve in the decision space) [22]. We have employed eleven test problems in our experimental analysis which incorporate the optimization challenges mentioned above. Five test problems are part of the ZDT test suite [4], while six were derived from the work done by Li and Zhang [22]. Mathematical formulations and optimization challenges of the test problems are discussed in detail in Section B.1 of the online supplement to this document. These problems are collectively referred as synthetic test problems in subsequent discussions.

### 4.1 Groundwater remediation design problem

Optimization problems pertaining to groundwater remediation models usually require solving complex and computationally expensive PDE systems to find objective function values for a particular input [27]. The groundwater remediation problem used in our analysis is based on a PDE system which describes the movement and purification of contaminated groundwater given a set of bioremediation and pumping design decisions [28]. Detailed description of the problem is provided in Section B.2 of the online supplement. The decision variables of the problem are the injection rates of remediation agent at 3 different well locations during each of \(m\) management periods. The input dimension size ranges between 6 and 36 variables, depending upon the number of management periods incorporated in the numerical computation model.

Two MO variants, GWDA and GWDM, of the groundwater optimization problem (Sect. 4.1)

GWDA | GWDM |
---|---|

minimize: \(f_1(X) = \sum _{t=1}^T C_t(X_t)\) | minimize: \(f_1(X) = \sum _{t=1}^T C_t(X_t)\) |

minimize: \(f_2(X) = \frac{\sum _{m=1}^M s_{m,T}(X)}{M}\) | minimize: \(f_2(X) = \max _{m=1}^M s_{m,T}(X)\) |

subject to: \(X \in [0,1]^{N\times T}\) | subject to: \(X \in [0,1]^{N\times T}\) |

## 5 Results and analysis

### 5.1 Experimental setup

We tested our algorithm on the test functions and groundwater problems discussed in the previous section. The performance of GOMORS was compared against the Non-Dominated Sorting Algorithm-II (NSGA-II) proposed by Deb et al. [5] (discussed in Sect. 3.3) and the kriging-based ParEGO algorithm proposed by Knowles [21]. All three algorithms are quite different. ParEGO is a multi-objective version of EGO [19] where the multi-objective problem is converted into many single objective optimization problems through Tchebycheff weighting [46]. One single objective problem is chosen at random from the predefined set of decomposed problems and EGO is applied to it to for selection of one point for evaluation per algorithm iteration. ParEGO is not designed for high dimensional problems (more than 15 variables). GOMORS on the other hand embeds RBFs within an evolutionary framework, and selects multiple points (from various rules defined in Sect. 3.4) for simultaneous (parallel) evaluations in each iteration and is designed for low and higher (15–25 decision variables) dimensional problems.

Since the objective of GOMORS is to find good solutions to MO problems within a limited function evaluation budget, our experiments were restricted to 400 function evaluations. Since all algorithms compared are stochastic, ten optimization experiments were performed for each algorithm, on each test problem, and results were compared via visual analysis of fronts and a performance metric based analysis. A detailed description of the experimental setup, including parameter settings for all algorithms and source code references for ParEGO and NSGA-II, is provided in Section E of the online supplement.

The **uncovered hypervolume** metric [1] was used to compare the performance of various algorithms. Uncovered hypervolume is the difference between the total feasible objective space (defined by the reference and ideal points in Fig. 1a) and the objective space dominated by estimate of the Pareto front obtained by an algorithm. A lower value of the metric indicates a better solution and the ideal value is zero.

Results from the synthetic test problems were analyzed in combination through the metric. Experiments were performed for each test problem with 8 decision variables, 16 decision variables and 24 decision variables to highlight performance differences of GOMORS, ParEGO and NSGA-II with varying problem dimensions. Results of all synthetic test problems were compiled for analysis by summing the metric values of each of the ten optimization experiments performed on each test problem. Since the uncovered hypervolume metric values obtained for each individual test problem and algorithm combination could be considered as independent random variables, a sum of them across a single algorithm is another random variable which is a convolution of the independent random variables. This convolution based metric summarizes overall performance of an algorithm on all synthetic test problems and is used as basis of our analysis methodology.

Results from the two variants of the groundwater remediation problem (GWDA and GWDM) were also analyzed through the uncovered hypervolume metric in a similar manner. Experiments were performed with 6 decision variables, 12 decision variables and 24 decision variables for each groundwater problem. Results from individual subsequent experiments performed on each problem were summed to obtain a convoluted metric analysis of the groundwater problems. Performance of GOMORS and ParEGO on the GWDM test problem was further assessed through visual analysis of non-dominated solutions obtained from each algorithm (median solution). The numerical comparisons are based on the number of objective function evaluations and hence do not incorporate parallel speedup. ParEGO does not have a parallel implementation so its wall-clock time is much longer than GOMORS. However, the comparisons here focus on the number of function evaluations and evaluate an estimate of total CPU time. These comparisons do not consider the additional advantage of GOMORS (parallel) in wall-clock time.

### 5.2 Results: synthetic test problems

Figure 2 aims to (1) visualize speed of convergence of all algorithms (by comparing results of 200 and 400 objective function evaluations) and (2) understand the effect of increasing decision space dimensions on performance of algorithms (by comparing problems with 8, 16 and 24 decision variables). The box plot visualization of each algorithm within each sub-plot of Fig. 2 corresponds to the uncovered hypervolume metric values summed over the eleven test problems (see Sect. 5.1). Lower values of the metric signify superiority of performance and a lower spread within a box plot depicts robustness of performance. Each sub-plot within Fig. 2 compares performance of all three algorithms for a specified number of decision variables (number of decision variables vary between 8 and 24) and a fixed number of function evaluations (either 200 or 400). Traversing from bottom to top, one can visualize the change in performance of algorithms as function evaluations increase (from 200 to 400), and a left to right traversal can help in visualizing performance differences with an increase in decision space dimensions (8, 16 and 24 decision variables).

Figure 2 clearly illustrates that GOMORS outperforms ParEGO for the 24 variable versions of the synthetic problems, both in terms of speed of convergence (at 200 evaluations) and at algorithm termination after 400 function evaluations (see top right sub-plot of Fig. 2). GOMORS’ convergence to a good solution is faster than ParEGO for the 8 and 16 variable versions of the synthetic test problems, but the difference is not as distinguishable (from Fig. 2) as in the case of the 24 variable versions. The Wilcoxon rank-sum test [14] (at 5 percent significance) confirms that performance of GOMORS is better than ParEGO for the 8, 16 and 24 variable versions of the test problems after 100, 200 and 400 function evaluations. Figure 2 also indicates that both GOMORS and ParEGO significantly outperform NSGA-II with reference to the synthetic problems when evaluations are limited to 400.

### 5.3 Results: groundwater remediation design problem

Figure 4 provides a visual comparison of non-dominated trade-offs obtained from GOMORS and ParEGO with application to the GWDM groundwater problem. The red line within each sub-plot is an estimate of the Pareto front of GWDM. Since knowledge of the true front is not known, we obtained our estimate of true Pareto front through a single trial of NSGA-II with 50,000 function evaluations. The green dots within each sub-plot correspond to the non-dominated solutions obtained via application of the algorithm referenced in the sub-plot. There are two sub-figures within Fig. 4 and four sub-plots within each sub-figure. Figure 4a corresponds to the median (the fronts ranked fifth in ten experiments of each algorithm) non-dominated fronts obtained from each algorithm for GWDM with 6 decision variables and Fig. 4b corresponds to the median fronts obtained from each algorithm for GWDM with 24 decision variables. A clock-wise traversal of sub-plots within each sub-figure depict results in the following order: (1) GOMORS after 100 evaluations, (2) GOMORS after 200 evaluations, (3) ParEGO after 200 evaluations, and (4) ParEGO after 100 evaluations.

## 6 Conclusion

GOMORS is a new parallel algorithm for multi-objective optimization of black box functions that improves efficiency for computationally expensive objective functions by obtaining good solutions with a relatively small number of evaluations (\(<\)500). GOMORS accomplishes this with the construction in each iteration of a surrogate response surface based on all the values of the objective functions computed in the current and previous iterations. Then evolution and non-dominated sorting are applied to the inexpensive function describing this surrogate surface. It is then possible to evaluate a large number of points on the surrogate inexpensively so multiple Rules can be used to select a diversity of points for expensive evaluation on parallel processors. The use of these multiple Rules is the innovative aspect of GOMORS and contributes to its effectiveness and robustness with limited numbers of evaluations (\(<\)500). GOMORS is very different from ParEGO, which solves multiple single objective problems.

Our numerical results indicate that GOMORS was significantly more effective than ParEGO when the number of evaluations was quite limited relative to the difficulty of the problem (Figs. 2, 3, 4) for both the test functions and the groundwater partial differential equation models. Computational demands for nonlinear optimization will grow rapidly as the dimension increases, so the effectiveness of GOMORS becomes more obvious on higher dimensional problems, as is evident in Figs. 2, 3, 4.

There are many real application models with computational times that are so large that the evaluations will be greatly limited, especially for multi-objective problems. For example, analysis of a carbon sequestration monitoring problem required global optimization with seven decisions of a nonlinear, multiphase flow transient PDF model that took over 2 hours per simulation [11]. So even 100 evaluations for a problem like this is probably more evaluations than are feasible. Hence the GOMORS approach to multi-objective optimization is a contribution in the area of surrogate-assisted multi-objective optimization of objective functions based on computationally expensive, multimodal, black box models.

## Notes

### Acknowledgments

The research was supported by a Fulbright Fellowship to Akhtar and grants to Shoemaker from NSF (CISE 1116298) and to Mahowald and Shoemaker from DOE (SciDAC).

## Supplementary material

## References

- 1.Bader, J.M.: Hypervolume-Based Search for Multiobjective Optimization: Theory and Methods. CreateSpace, Paramount (2010)Google Scholar
- 2.Beume, N., Naujoks, B., Emmerich, M.: SMS-EMOA: multiobjective selection based on dominated hypervolume. Eur. J. Oper. Res.
**181**(3), 1653–1669 (2007)MATHCrossRefGoogle Scholar - 3.Buhmann, M.D.: Radial Basis Functions. Cambridge University Press, New York (2003)MATHCrossRefGoogle Scholar
- 4.Deb, K.: Multi-objective genetic algorithms: problem difficulties and construction of test problems. Evol. Comput.
**7**(3), 205–230 (1999)CrossRefGoogle Scholar - 5.Deb, K., Agrawal, S., Pratap, A., Meyarivan, T.: A fast elitist non-dominated sorting genetic algorithm for multi-objective optimization: Nsga-ii. In: Proceedings of the 6th International Conference on Parallel Problem Solving from Nature. PPSN VI, pp. 849–858. Springer, London, UK (2000)Google Scholar
- 6.Deb, K., Kalyanmoy, D.: Multi-Objective Optimization Using Evolutionary Algorithms, 1st edn. Wiley, New York (2001)MATHGoogle Scholar
- 7.Deb, K., Nain, P.: An evolutionary multi-objective adaptive meta-modeling procedure using artificial neural networks. In: Yang, S., Ong, Y.S., Jin, Y. (eds.) Evolutionary Computation in Dynamic and Uncertain Environments, vol. 51, chap. 13, pp. 297–322. Springer, Berlin (2007)CrossRefGoogle Scholar
- 8.Diaz-Manriquez, A., Toscano-Pulido, G., Gomez-Flores, W.: On the selection of surrogate models in evolutionary optimization algorithms. IEEE Congr. Evol. Comput. 2155–2162 (2011)Google Scholar
- 9.Emmerich, M., Giotis, A., Özdemir, M., Bäck, T., Giannakoglou, K.: Metamodel-assisted evolution strategies. In: Parallel Problem Solving from Nature VII. Springer, pp. 361–370 (2002)Google Scholar
- 10.Emmerich, M.T.M., Giannakoglou, K., Naujoks, B.: Single- and multiobjective evolutionary optimization assisted by gaussian random field metamodels. IEEE Trans. Evol. Comput.
**10**(4), 421–439 (2006)CrossRefGoogle Scholar - 11.Espinet, A., Shoemaker, C.A., Doughty, C.: Estimation of plume distribution for carbon sequestration using parameter estimation with limited monitoring data. Water Resour. Res.
**49**(7), 4442–4464 (2013)CrossRefGoogle Scholar - 12.Fang, H., Rais-Rohani, M., Liu, Z., Horstemeyer, M.F.: A comparative study of metamodeling methods for multiobjective crashworthiness optimization. Comput. Struct.
**83**(25–26), 2121–2136 (2005)CrossRefGoogle Scholar - 13.Georgopoulou, C.A., Giannakoglou, K.C.: A multi-objective metamodel-assisted memetic algorithm with strength-based local refinement. Eng. Optim.
**41**(10), 909–923 (2009)CrossRefGoogle Scholar - 14.Gibbons, J.D., Chakraborti, S.: Nonparametric Statistical Inference. Chapman and Hall, Boca Raton (2010)Google Scholar
- 15.Gutmann, H.M.: A radial basis function method for global optimization. J. Global Optim.
**19**, 201–227 (2001)MATHMathSciNetCrossRefGoogle Scholar - 16.Huang, D., Allen, T.T., Notz, W.I., Zeng, N.: Global optimization of stochastic black-box systems via sequential kriging meta-models. J. Global Optim.
**34**(3), 441–466 (2006)MATHMathSciNetCrossRefGoogle Scholar - 17.Jin, R., Chen, W., Simpson, T.W.: Comparative studies of metamodelling techniques under multiple modelling criteria. Struct. Multidiscip. Optim.
**23**(1), 1–13 (2001)CrossRefGoogle Scholar - 18.Jin, Y., Olhofer, M., Sendhoff, B.: A framework for evolutionary optimization with approximate fitness functions. IEEE Trans. Evol. Comput.
**6**(5), 481–494 (2002)CrossRefGoogle Scholar - 19.Jones, D.R., Schonlau, M., Welch, W.J.: Efficient global optimization of expensive black-box functions. J. Global Optim.
**13**(4), 455–492 (1998)MATHMathSciNetCrossRefGoogle Scholar - 20.Karakasis, M.K., Giannakoglou, K.C.: On the use of metamodel-assisted, multi-objective. Eng. Optim.
**38**(8), 941–957 (2006)MathSciNetCrossRefGoogle Scholar - 21.Knowles, J.: ParEGO: a hybrid algorithm with on-line landscape approximation for expensive multiobjective optimization problems. IEEE Trans. Evol. Comput.
**8**(5), 1341–66 (2006)Google Scholar - 22.Li, H., Zhang, Q.: Multiobjective optimization problems with complicated pareto sets, MOEA/D and NSGA-II. IEEE Trans. Evol. Comput.
**13**(2), 284–302 (2009)CrossRefGoogle Scholar - 23.Lim, D., Jin, Y., Ong, Y.S., Sendhoff, B.: Generalizing surrogate-assisted evolutionary computation. IEEE Trans. Evol. Comput.
**14**(3), 329–355 (2010)CrossRefGoogle Scholar - 24.Martínez, S.Z., Coello, C.A.C.: A memetic algorithm with non gradient-based local search assisted by a meta-model. In: Proceedings of the 11th International Conference on Parallel Problem Solving from Nature: Part I. PPSN’10. Springer, Berlin, Heidelberg, pp. 576–585 (2010)Google Scholar
- 25.Martinez, S.Z., Coello, C.A.C.: Combining surrogate models and local search for dealing with expensive multi-objective optimization problems. In: IEEE Congress on Evolutionary Computation, pp. 2572–2579. IEEE (2013)Google Scholar
- 26.McKay, M.D., Beckman, R.J., Conover, W.J.: A comparison of three methods for selecting values of input variables in the analysis of output from a computer code. Technometrics
**21**(2), 239–245 (1979)MATHMathSciNetGoogle Scholar - 27.Minsker, B.S., Shoemaker, C.A.: Differentiating a finite element biodegradation simulation model for optimal control. Water Resour. Res.
**32**(1), 187–192 (1996)CrossRefGoogle Scholar - 28.Minsker, B.S., Shoemaker, C.A.: Dynamic optimal control of in-situ bioremediation of ground water. Water Resourc. Plan. Manag.
**124**(3), 149–161 (1998)CrossRefGoogle Scholar - 29.Montemayor-garcía, G., Toscano-pulido, G.: A Study of Surrogate models for their use in multiobjective evolutionary algorithms. In: 8th International Conference on Electrical Engineering Computing Science and Automatic Control (CCE) (2011)Google Scholar
- 30.Mueller, J., Shoemaker, C.A., Piche, R.: SO-I: a surrogate model algorithm for expensive nonlinear integer programming problems including global optimization applications. J. Global Optim.
**59**(4), 865–889 (2014)Google Scholar - 31.Mueller, J., Shoemaker, C.A., Piche, R.: SO-MI: a surrogate model algorithm for computationally expensive nonlinear mixed-integer, black-box global optimization problems. Comput. Oper. Res.
**40**(5), 1383–1400 (2013)Google Scholar - 32.di Pierro, F., Khu, S.T., Savić, D., Berardi, L.: Efficient multi-objective optimal design of water distribution networks on a budget of simulations using hybrid algorithms. Environ. Model. Softw.
**24**(2), 202–213 (2009)CrossRefGoogle Scholar - 33.Ponweiser, W., Wagner, T., Biermann, D., Vincze, M.: Multiobjective optimization on a limited budget of evaluations using model-assisted s-metric selection. In: Proceedings of the 10th International Conference on Parallel Problem Solving from Nature: PPSN X. Springer, Berlin, Heidelberg, pp. 784–794 (2008)Google Scholar
- 34.Powell, M.J.D.: The Theory of Radial Basis Function Approximation in 1990, pp. 105–210. Oxford University Press, USA (1992)Google Scholar
- 35.Priddy, K.L., Keller, P.E.: Artificial Neural Networks: An Introduction. SPIE Press, New York (2005)CrossRefGoogle Scholar
- 36.Regis, R., Shoemaker, C.: Local function approximation in evolutionary algorithms for the optimization of costly functions. IEEE Trans. Evol. Comput.
**8**(5), 490–505 (2004)CrossRefGoogle Scholar - 37.Regis, R., Shoemaker, C.: A stochastic radial basis function method for the global optimization of expensive functions. INFORMS J. Comput.
**19**, 497–509 (2007)MATHMathSciNetCrossRefGoogle Scholar - 38.Regis, R.G., Shoemaker, C.A.: Constrained global optimization of expensive black box functions using radial basis functions. J. Global Optim.
**31**, 153–171 (2005)MATHMathSciNetCrossRefGoogle Scholar - 39.Regis, R.G., Shoemaker, C.A.: Improved strategies for radial basis function methods for global optimization. J. Global Optim.
**37**(1), 113–135 (2007)MATHMathSciNetCrossRefGoogle Scholar - 40.Regis, R.G., Shoemaker, C.A.: Parallel stochastic global optimization using radial basis functions. INFORMS J. Comput.
**21**(3), 411–426 (2009)MATHMathSciNetCrossRefGoogle Scholar - 41.Regis, R.G., Shoemaker, C.A.: A quasi-multistart framework for global optimization of expensive functions using response surface models. J. Global Optim.
**56**(4), 1719–1753 (2013)Google Scholar - 42.Regis, R.G., Shoemaker, C.A.: Combining radial basis function surrogates dynamic coordinate search in high dimensional expensive black-box optimization. Eng. Optim.
**45**(5), 529–555 (2013)MathSciNetCrossRefGoogle Scholar - 43.Rosales-p, A., Coello, C.A.C., Gonzalez, J.A., Reyes-garcia, C.A., Escalante, H.J.: A hybrid surrogate-based approach for evolutionary multi-objective optimization. In: IEEE Congress on Evolutionary Computation (CEC), pp. 2548–2555 (2013)Google Scholar
- 44.Sacks, J., Welch, W.J., Mitchell, T.J., Wynn, H.P.: Design and analysis of computer experiments. Stat. Sci.
**4**(4), 409–423 (1989)MATHMathSciNetCrossRefGoogle Scholar - 45.Santana-Quintero, L., Serrano-Hernandez, V., Coello Coello, C.A., Hernandez-Diaz, A., Molina, J.: Use of radial basis functions and rough sets for evolutionary multi-objective optimization. In: IEEE Symposium on Computational Intelligence in Multicriteria Decision Making, pp. 107–114 (2007)Google Scholar
- 46.Steuer, R., Choo, E.U.: An interactive weighted tchebycheff procedure for multiple objective programming. Math. Program.
**26**, 326–344 (1983)MATHMathSciNetCrossRefGoogle Scholar - 47.Vapnik, V.N.: Statistical Learning Theory, 1st edn. Wiley, New York (1998)MATHGoogle Scholar
- 48.Vrugt, J.A., Robinson, B.A.: Improved evolutionary optimization from genetically adaptive multimethod search. Proc. Natl. Acad. Sci. USA
**104**(3), 708–711 (2007)CrossRefGoogle Scholar - 49.Wild, S.M., Regis, R.G., Shoemaker, C.A.: ORBIT: optimization by Radial Basis Function Interpolation in Trust-Regions. SIAM J. Sci. Comput.
**30**(6), 3197–3219 (2007)MathSciNetCrossRefGoogle Scholar - 50.Wild, S.M., Shoemaker, C.A.: Global convergence of radial basis function trust-region algorithms for derivative-free optimization. SIGEST article. SIAM Rev.
**55**(2), 349–371 (2013)MATHMathSciNetCrossRefGoogle Scholar - 51.Martínez, Zapotecas: S., Coello Coello, C.A.: Moea/d assisted by rbf networks for expensive multi-objective optimization problems. In: Proceeding of the Fifteenth Annual Conference on Genetic and Evolutionary Computation Conference. GECCO ’13, pp. 1405–1412. ACM, New York, NY, USA (2013)Google Scholar
- 52.Zhang, Q., Li, H.: MOEA/D: a multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput.
**11**(6), 712–731 (2007)CrossRefGoogle Scholar - 53.Zhang, Q., Liu, W., Tsang, E., Virginas, B.: Expensive multiobjective optimization by MOEA/D with Gaussian process model. IEEE Trans. Evol. Comput.
**14**(3), 456–474 (2010)CrossRefGoogle Scholar

## Copyright information

**Open Access**This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.