Introduction

Nowadays, there are many practical engineer problems that can be modeled as noisy and expensive multi-objective optimization problems (MOPs), such as the gait optimization of autonomous snake robot. It needs to maximize the speed and head stability of the sidewinding gait of a snake robot [1]. Because the snake robot achieves locomotion by twisting its body [2], its speed and head stability are a pair of conflicting objectives. The objective values can only be observed through experiments, so the cost of observations is expensive. Due to the influence of environment and robot gait, it is difficult to get accurate objective information from the speed and vision sensors. Therefore, this optimization problem can be modeled as an expensive MOP with independent noises on each objective.

MOPs usually have multiple conflicting objectives to be optimized at the same time. Evolutionary algorithms (EAs) [3] are widely used to solve MOPs because of some advantages. Those advantages include that obtaining a set of non-dominated solutions in one run [4], simultaneously handling problems with multiple local and non-convex Pareto fronts [5]. Therefore, a large number of multi-objective evolutionary algorithms (MOEAs) have proposed for MOPs in recent decades [6]. These MOEAs usually are divided into three types: algorithms based on Pareto domination, indictors and decomposition. The first type of MOEAs select candidate solutions by the Pareto dominance. For example, nondominated sorting genetic algorithm-II (NSGA-II) [7] uses the dominance relationship and crowding distance as criteria for its selection. The second type of MOEAs assign a fitness to each solution by a suitable indicator. Indicator-based selection evolutionary algorithm (IBEA) [8] and Hypervolume (HV) estimation algorithm (HypE) [9] use \(I_{\epsilon +}\) [10] and HV [10] as their indicators, respectively. The third type of MOEAs apply a series of vectors to decompose an MOP into multiply single objective optimization problems (SOPs). For instance, multi-objective evolutionary algorithm based on decomposition (MOEA/D) [11]. It is noteworthy that existing MOEAs need more than ten thousands of function evaluations [12]; however, there are only a few hundred function evaluations can be used due to the limitation of money and(or) time cost(s) in expensive MOPs [13].

Recently, one popular method to solve expensive MOPs is to use surrogate models [6, 14] (also called meta-models [15]) to replace the real function evaluations in EAs. Specifically, modeling methods are applied to the existing data set to build a computationally cheap surrogate model to replace the expensive function evaluation, so as to reduce the computation budget. These MOEAs assisted by the surrogate are called as multi-objective surrogate-assisted evolutionary algorithms (SA-MOEAs) [16]. To improve the reliability of the surrogates and improve the optimization performance, SAEAs usually need corresponding sampling criteria which aim to select suitable solutions for evaluating. They usually need to take convergence, diversity and approximation uncertainty for expensive MOPs into consideration [13, 17]. The methods to construct surrogates include response surface method (RSM) [18], Gaussian process (GP) [19], artificial neural networks (ANN) [20], support vector machines (SVM) [21], radial basis function networks (RBFN) [22]. According to the methods of constructing surrogates, these SAEAs can be divided into two classes:

  1. 1.

    Regression-based SAEAs: These SAEAs use surrogates to approximate objective functions, indictors, or scalar functions, etc. The examples of the first category include Kriging-assisted two-archive evolutionary algorithm (KTA2) [17], a surrogate-assisted particle swarm optimization algorithm with an adaptive dropout mechanism (ADSAPSO) [23] Kriging assisted reference vector guided evolutionary algorithm (KRVEA) [5]. KRVEA uses angle penalized distance (APD) [5] to select the solutions to keep the convergence and diversity of population. However, when the number of inactive fixed reference vectors, i.e., the reference vectors without any associated solution, the increasing number of those reference vectors between generations reaches a threshold, it would select the solutions with uncertainty. KTA2 considers convergence, diversity, and approximation uncertainty in its sampling criterion, which analyzes the optimization process need on convergence, diversity and approximation uncertainty then adaptively selects a sampling strategy based one aspect. ADSAPSO adopts the dominance relationship and crowding distance as the criteria sampling, which takes convergence and diversity into consideration. The examples of the second category include neural networks assisted steady state indicator-based evolutionary algorithm (NN-SS-IBEA) [24]. In the sampling criteria of NN-SS-IBEA, it chooses individual with a high HV contribution which considers convergence and diversity. The examples of the third category include Pareto based efficient global optimization (ParEGO) [25], efficient global optimization-assisted multi-objective evolutionary algorithm based on decomposition (MOEA/D-EGO) [26]. ParEGO and MOEA/D-EGO select the solution which has the best expected improvement (EI) [27] of the aggregation function to sample, which takes both convergence and uncertainty into account.

  2. 2.

    Classification-based SAEAs: These SAEAs make use of the existing data to train a classifier. It is used as the surrogate model to select candidate solutions as their function evaluations, such as classification and Pareto-based MOEA (CPS-MOEA) [28], classification based on surrogate-assisted evolutionary algorithm (CSEA) [29]. CPS-MOEA selects the solutions with good fitness which mainly considers convergence. As there is no solution with good fitness, CPS-MOEA selects the solutions for function evaluations randomly which takes the approximation uncertainty into account. CSEA selects nondominated solutions with a high credibility level and dominated solutions with a low credibility level for function evaluations which takes convergence and approximation uncertainty into account. The main differences between CPS-MOEA and CSEA include their surrogates and sampling criteria. CPS-MOEA applies a classification tree as its surrogate but CSEA uses feed-forward neural network (FNN) [29] as its surrogate. CPS-MOEA selects solutions with good fitness by surrogate but CSEA selects solutions by comprehensively considering the dominance relationship and credibility.

Noises would lead that the objective function cannot be accurately evaluated. Thus, the optimization direction may be misled [30]. To improve the performance of the MOEAs, we need to take some extra methods for dealing with noises for noisy MOPs. At present, the approaches for handling noises in MOEAs can be compartmentalized into three categories [31] as follow:

  1. 1.

    Averaging: The average value of all solutions after independently sampling several times is taken as the objective values after denoising. With this approach, resampling \(n \) times, then the standard deviation of the objectives would decrease \(\sqrt{n }\) times [32]. This is a common method for decreasing noises but it requires many function evaluations [33,34,35]. For example, noise-tolerant strength Pareto evolutionary algorithm (NTSPEA) [35], each solution is assigned resampling times according to the number of dominated solutions in the population.

  2. 2.

    Ranking: The ranking methods are used to decrease noises include two types: ranking based on probability [36,37,38,39] and clustering [40]. For the first one, constructing a model with confidence level which is applied to predict the dominated probability between solutions with noises. For example, nondominated sorting genetic algorithm-II-A (NSGA-II-A) [37] constructs an SVM model with a confidence level by \(\alpha \) dominance to predict the relationship between solutions with noises. The other one, the average and standard deviation of each solution are calculated by resampling, and the clustering radius is obtained by the calculation results, so as to select more reliable solutions. For instance, modified nondominated sorting genetic algorithm-II (M-NSGA-II) [40] uses this method to solve noisy MOPs.

  3. 3.

    Modeling: This approach considers that noises may has seriously influence on one solution but on a model can be minimal. Therefore, it uses the noisy data to construct a regression model for denoising. For example, regularity model [31], restrict Boltzmann machine [41, 42] and univariate marginal distribution algorithm [43]. They all handle noises through the idea of modeling which brings a new direction for denoising.

At present, there are some MOEAs that combine the above methods for denoising, such as Filters-based NSGA-II (FNSGA-II) [44]. It uses resampling, mean and Wiener filters for denoising, where resampling mean filter belong to averaging and Wiener filters belong to modeling. However, there are many MOEAs to solve either expensive MOPs or noisy MOPs. For noisy and expensive MOPs, these MOEAs cannot achieve satisfactory performance. According to the above analysis, we find that the MOEAs for expensive MOPs have poor performance for noisy MOPs due to the influence of noise, and the MOEAs for noisy MOPs which need a large number of evaluations cannot meet the computation budget limitation [31, 33]. To solve this type of MOPs, we need to face three main challenges. First, because of the influence of noises, the reliability of surrogates would be reduced. Second, due to the limited number of function evaluations, some existing denoising methods are invalid. Third, most sampling criteria cannot make full use of limited function evaluations. Therefore, we propose an adaptive model switch-based surrogate-assisted evolutionary algorithm (MS-MOEA) in this paper. We adopt IBEA as the basic optimizer, where \(I_{\epsilon +}\) and \(L_p\)-norm distance [45] are adopted as indicators to select sampling solutions and the GP and RBFN models are used as the surrogates to approximate objective functions. The main contributions of this work as follow:

  • A noise treatment method is proposed for decreasing the influence of noises. The approach divides all solutions into two sets: one of which all solutions greatly affected by noises, the other contains all remaining solutions. When the number of solutions in the first set reaches a certain scale, then an RBFN model is constructed by all data, and the data is smoothened by the model for denoising.

  • An adaptive model switch approach is designed for the proposed algorithm. Either the RBFN or GP model is chosen as the surrogate according to whether the data set is denoised or not.

  • We design an improved environmental selection for the MS-MOEA. It combines different selection methods to select individuals according to the impact of noise on the population.

  • We propose an adaptive sampling strategy in this paper. This strategy includes three sampling methods and an approach to adaptively switch between the sampling methods.

The rest of the paper is organized as follows. An introduction to noisy MOPs, IBEA, GP and RBFN is given in the next section. The details of the proposed algorithm for noisy and expensive MOPs are described in the following section. The experimental settings, results and analysis are presented in the next section. The conclusions and future work are concluded in the last section.

Preliminaries

Noisy multi-objective optimization problems

We consider the mathematical expression of MOPs is as follows:

(1)

where x is decision vector and d is the dimension of decision vector. \(F({{\varvec{x}}})\) is the objective function which has m optimization objectives (\(m\) \(\ge \)2). Since the objectives usually are conflicting, many solutions cannot be compared by objective values. The concept of the Pareto dominance [46] is to compare the solutions of MOPs. For example, there are two different solutions a and b, only \(\forall \) \(i=1,2,\ldots ,m\), \(f_i({{\varvec{a}}})\) \(\le \) \(f_i({{\varvec{b}}})\) and \(F({{\varvec{a}}})\) \(\ne \) \(F({{\varvec{b}}})\), a Pareto dominates b.

The MOPs that we need to solve have noises in the objective space, so the expression of the optimization objectives is shown in Eq. (2).

$$\begin{aligned} {\begin{matrix} T({{\varvec{x}}}) &{}= F({{\varvec{x}}}) + \varvec{\beta }, \\ \varvec{\beta } &{}\sim N(0,\sigma ^2I), \end{matrix}} \end{aligned}$$
(2)

where \(F({\varvec{x}})\) is our optimization objectives, but we can only observe \(T({{\varvec{x}}})\). In this paper, we assume the noises are independent additive noises. \(\varvec{\beta }\) meets the Gaussian distribution. I is an identity matrix.

Indicator-based selection evolutionary algorithm

The evolutionary search of expensive MOPs need the assistance of surrogate models, which means that the approximation error of the model would have a great impact on the search process. IBEA\(_{\epsilon +}\) uses \(I_{\epsilon +}\) as selection criterion rather than the Pareto dominance relationship. \(I_{\epsilon +}\) is the minimal difference that the dominance relationship between two solution sets become one weakly dominating the other:

$$\begin{aligned} \begin{aligned} I_{\epsilon +}(A,B) = \min _{\epsilon +}\{\forall \varvec{x_2}\in B,\exists \varvec{x_1}\in A: f_i(\varvec{x_1})- \epsilon \\ \le f_i(\varvec{x_2}),\, \mathrm{for} \quad i=1,\ldots ,m\}, \end{aligned} \end{aligned}$$
(3)

where A and B are solution sets, respectively, and m is the number of objectives. \(I_{\epsilon +}(A,B)\) represents the minimum distance of each objective dimension when the dominance relationship between A and B is to be transformed into weak dominance.

Compared with the Pareto dominance, there are two advantages of \(I_{\epsilon +}\) are as follows: it does not only represent the natural extension of the Pareto dominance but also transforms multiple objectives into one indictor which reduces the approximation errors of surrogate models to a certain extent. When the dominance relationship is adopted for ranking, it is necessary to ensure that the prediction objective error does not affect the size relation of each two individuals in every objective. However, \(I_{\epsilon +}\) can reduce the approximation errors of surrogate models to a certain extent and the reason as follow. If A (or B) only contains one solution \(x_1\) (or \(x_2\)), the Eq. (3) can be written as

$$\begin{aligned} I\left( x_{1}, x_{2}\right) =\max \left( F\left( x_{1}\right) -F\left( x_{2}\right) \right) \end{aligned}$$
(4)

Thus, when \(I_{\epsilon +}\) is adpoted as the indicator, it is only necessary to ensure that the prediction objective error of the selected surrogate model does not affect the size relation of each two individuals in one objective. Therefore, the adaptive IBEA\(_{\epsilon +}\) is adopted as basic optimizer in this work. IBEA\(_{\epsilon +}\) uses \(I_{\epsilon +}\) as the indicator to assign the fitness for each individual in the population. The binary tournament selection is used to select the parents. It makes use of crossover and mutation to generate offsprings. The pseudo code of adaptive IBEA\(_{\epsilon +}\) is shown in Algorithm 1.

figure a

According to the above analysis, it is proved that \(I_{\epsilon +}\) can be directly used in fitness calculation. IBEA\(_{\epsilon +}\) adopts Eq. (5) to assign fitness for individuals:

$$\begin{aligned} \begin{aligned} F(\varvec{x})&=\sum _{\varvec{x}^{'}\in P \setminus \{\varvec{x}\} } -e^{I(\{\varvec{x}^{'}\},\{\varvec{x}\})/(k*c)}, \\ c&= \max _{\varvec{x}^{'}\in P}|I(\varvec{x},\varvec{x}^{'})|, \end{aligned} \end{aligned}$$
(5)

where k represents a scaling factor (\(k > \) 0), c is defined to improve the robustness of the algorithm. Due to the dominance maintaining of \(I_{\epsilon +}\). Therefore, the better solution would be allocated higher fitness, and them are easier to be selected in the population iteration. Because the size of population is fixed, we need to delete some individuals by the fitness. Therefore, when the composition of the population changes, the fitness of every individual need to be updated. The Eq. (6) is used to update the

$$\begin{aligned} \begin{aligned} F(\varvec{x})&= F(\varvec{x}) + e^{-I({\varvec{W}},{\varvec{x}})/(k*c)}, \\ c&= \max _{\varvec{x}^{'}\in P}|I(\varvec{x},\varvec{x}^{'})|, \end{aligned} \end{aligned}$$
(6)

where W is the individual that needs to be deleted, other parameters same like Eq. (5).

Radial basis function network

RBFN [47] is a type of artificial neural network which usually has three layers: an input layer, a hidden layer and an output layer. The structure of RBFN, as shown in Fig. 1. In the RBFN, each neuron of the hidden layer is a radial basis function that takes the distance from the input point to the corresponding center point as the input variable. The calculation formula of the RBFN output is shown in Eq. (7):

$$\begin{aligned} \begin{aligned} \mathrm{Output}&=\sum _{i=1}^N w_i\phi _i + \mathrm{Bias}, \\ \phi _i(r)&= \phi _i(|{{\varvec{x}}}-{{\varvec{c}}}|), \end{aligned} \end{aligned}$$
(7)

where N represents the number of neurons in hidden layer, \(w_i\) is the weight of ith the hidden layer neuron and the output layer, \(\phi _i\) represents the output of ith the hidden layer and its input (\(|{{\varvec{x}}}-{{\varvec{c}}}|\)) is the distance of input point and ith center point. Bias is the bias of the whole network.

There are four types of kernel functions commonly used in RBFN: Gaussian, reflected sigmoid, multiquadric and inverse multiquadric kernels [48]. In this work, we use the most commonly used Gaussian kernel function as the radial basis function, as shown in Eq. (8):

$$\begin{aligned} \begin{aligned} \phi (r)&= e^{\left( -\frac{r^2}{2\sigma ^2}\right) }, \\ \end{aligned} \end{aligned}$$
(8)

where \(r = |{{\varvec{x}}}-{{\varvec{c}}}|\), \(\sigma \) is the spread rate used to control the smoothness of RBFN which decides the range of central point.

Fig. 1
figure 1

Structure of RBFN

In this paper, generalized least square method [49] is used for calculating the weight of radial basis function network, as shown in Eq. (9). It is assumed that there are K training points X = [\({{\varvec{x}}}^1,{{\varvec{x}}}^2,\ldots ,{{\varvec{x}}}^K\)] and Y = [\({y} ^1,{y} ^2,\ldots ,{y} ^K\)]. Because the decision space dimension of the MOPs is d and the objective space dimension is m, it is necessary to construct m RBFNs to replace all objective functions:

$$\begin{aligned} \begin{aligned} W = (\varPhi ^T\cdot \varPhi )^{-1}\cdot \varPhi ^T\cdot Y, \end{aligned} \end{aligned}$$
$$\begin{aligned} \begin{aligned} W = \left[ \begin{array}{cccc} w_1\\ w_2\\ \vdots \\ w_n \end{array} \right] ,\quad \varPhi =\left[ \begin{array}{cllcl} \phi _{1}^{1} &{} \phi _{2}^{1} &{} \cdots &{} \phi _{Q}^{1} &{} 1 \\ \vdots &{} \vdots &{} \ddots &{} \vdots &{} \vdots \\ \phi _{1}^{K} &{} \phi _{2}^{K} &{} \cdots &{} \phi _{Q}^{K} &{} 1 \end{array}\right] , {Y}=\left[ \begin{array}{c} y^{1} \\ y^{2} \\ \vdots \\ y^{K} \end{array}\right] , \end{aligned} \end{aligned}$$
(9)

where Q is the number of neuron in hidden layer, K represents the number of training data, \(\phi ^i_j\) respresents the output of the jth hidden layer neuron at sample \(\varvec{x}^i\). Q and the central point are determined by K-means clustering algorithm [50].

Taking the Gaussian kernel as an example, the farther the input vector from the center of RBF, the lower the activation degree of the node is, so the RBFN has the characteristics of local fitting [47]. In addition, the training speed of RBFN is fast and the amount of calculation is small, which is one of the advantages of using RBFN as a surrogate model.

Gaussian process

Gaussian process is a excellent performance modeling method. It cannot only provide the predicted value of the points needed to be predicted, but also provide the uncertainty of points. Therefore, GP has been widely used in SAEAs. In this paper, we adopt DACE [51] to construct the GP model.

It is assumed that there are K training points X = [\({{\varvec{x}}}^1,{{\varvec{x}}}^2,\ldots ,{{\varvec{x}}}^K\)] and Y = [\({y} ^1,{y} ^2,\ldots ,{y} ^K\)] in data set S. Because the decision space dimension of MOPs is d and the objective space dimension is m, it is necessary to construct m GPs to replace all objective functions. The Gaussian process model based on data set S satisfies Eq. (10):

$$\begin{aligned} \begin{aligned} y(\varvec{x})&=\mu (\varvec{x})+\alpha (\varvec{x}), \\ \alpha (\varvec{x})&\sim N\left( 0, \sigma ^{2}\right) , \end{aligned} \end{aligned}$$
(10)

where \(\mu (\varvec{x})\) is the predicted value of \(\varvec{x}\) by the model, \(\alpha (\varvec{x})\) represents an independently Gaussian distribution with 0 as the mean value, \(\sigma \) as the standard deviation.

Fig. 2
figure 2

Framework of MS-MOEA

All training samples need to be trained to construct GP. For any two training samples \(\varvec{x}^i\) and \(\varvec{x}^j\), the two random processes \(\alpha (\varvec{x}^i)\) and \(\alpha (\varvec{x}^j)\) are defined as Eq. (11):

$$\begin{aligned} \mathrm{cov}\left( \alpha \left( \varvec{x}^{i}\right) , \alpha \left( \varvec{x}^{j}\right) \right) =\sigma ^{2} {{\varvec{R}}}\left( \left[ R\left( \varvec{x}^{i}, \varvec{x}^{j}\right) \right] \right) , \end{aligned}$$
(11)

where \({{\varvec{R}}}\) represents the correlation matrix consisting of the correlation function values of all samples. For any \(\varvec{x}^i,\varvec{x}^j\), \(R(\varvec{x}^i,\varvec{x}^j) = e^{-d(\varvec{x}^i,\varvec{x}^j)}\), \(d\left( \varvec{x}^{i}, \varvec{x}^{j}\right) =\sum _{i=1}^{K} \theta _{i}\left| \varvec{x}_{k}^{i}-\varvec{x}_{k}^{j}\right| ^{2}\). All \(\theta _i\) represent the hyperparameters.

To improve the performance of the surrogate model on the data set S, we need to maximize the likehood function as shown in Eq. (12) by optimizing the hyperparameters:

$$\begin{aligned} \frac{1}{\left( 2 \pi \sigma ^{2}\right) ^{K / 2} \!\sqrt{{\text {det}}(\varvec{R})}} e^{ \left[ -\frac{(\varvec{y}-\mu \mathbf {1})^{T} \varvec{R}^{-1}(\varvec{y}-\mu \mathbf {1})}{2 \sigma ^{2}}\right] } \end{aligned}$$
(12)

where \(\mathbf {1}\) respresents a identity matrix of \(K*1\). \(\varvec{y} = \varvec{Y}^T\). To maximize the likehood function, the following conditions of \(\mu \) and \(\sigma \) should be satisfied, as shown in Eq. (13):

$$\begin{aligned} \begin{aligned} \hat{\mu }&=\frac{\mathbf {1}^{T} \varvec{R}^{-1} \varvec{y}}{\mathbf {1}^{T} R^{-1} \mathbf {1}} \\ \hat{\sigma }^{2}&=\frac{(\varvec{y}-\mathbf {1} \mu )^{T} \varvec{R}^{-1}(\varvec{y}-\mathbf {1} \mu )}{K} \end{aligned} \end{aligned}$$
(13)

Finally, the calculation formula of predicted values and variance is shown in Eq. (14):

$$\begin{aligned} \begin{aligned} \hat{y}(\bar{\varvec{x}})&=\hat{\mu }+r^{T}(\bar{\varvec{x}}) \varvec{R}^{-1}(y-\mathbf {1} \hat{\mu }), \\ s^{2}\!(\bar{\varvec{x}})&=\hat{\!}{\sigma }^{2}\!\!\left[ \!1\!\!-\!r^{T}\!(\bar{\varvec{x}}) \varvec{R}^{-1} r(\bar{\varvec{x}})+\frac{\left( 1-\mathbf {1}^{T} R^{-1} r(\bar{\varvec{x}})\right) \!^{2}}{\mathbf {1}^{T} R^{-1} r(\bar{\varvec{x}})}\!\right] , \end{aligned} \end{aligned}$$
(14)

where \(r(\bar{\varvec{x}})=\left[ R\left( \bar{\varvec{x}},\varvec{x}^{1}\right) , R\left( \bar{\varvec{x}}, \varvec{x}^{2}\right) , \ldots , R\left( \bar{\varvec{x}}, \varvec{x}^{K}\right) \right] \) represents the correlation matrix consisting of the forecast point and all sample points.

It is noteworthy that the computational complexity of training a GP is \(O(N^3)\) [52] and the algorithms for optimizing the hyperparameters \(\theta \) would increase it. Therefore, GP is not suitable for the high-dimensional problems.

Proposed algorithm

In this work, an adaptive model switch-based SAEA (MS-MOEA) is proposed for noisy expensive multi-objective optimization. In this SAEA, we adopt IBEA\(_{\epsilon +}\) as the basic optimizer, an adaptive model switching strategy between RBFN and GP to approximate objective functions, an infill sampling criterion to select solutions for real function evaluations. The framework of the proposed MS-MOEA is shown in Fig. 2. From the figure, MS-MOEA can be divided into six main steps.

  1. 1.

    Initialization: An initial population \(P_0\) in the size N is generated by Latin hypercube sampling [53]. All individuals in the \(P_0\) have been evaluated by the real objective functions T(x) which are expensive and noisy, then the individuals of \(P_0\) would be added to the archive A.

  2. 2.

    Noise treatment: A solution with the best fitness is sampled several times, and the average value is calculated as the actual objective value of the solution each objective. The absolute value of maximum noise \({E_{\max }}\) can be calculated by the sampling value and mean value. To estimate the influence of noises, we need to calculate the ratio of \({E_{\max }}\) to objective values and record it as \(\lambda \) and then compare \(\lambda \) and the threshold of noises \(\delta \), when \(\lambda > \delta \), a denoising method is applied for objective values. Conversely, it is unnecessary. The denoising method is to take the data with noises to construct RBFN and the regression value as the denoised objective value.

  3. 3.

    Model switch: According to the result of noises estimation to select the suitable surrogate model from GP and RBFN, where \(\lambda > \delta \), the algorithm selects RBFN and sets model flag to 0, conversely, the algorithm selects GP and sets model flag to 1.

  4. 4.

    Model building/rebuilding: All original data and the model flag are used to construct a suitable regression model which is used to replace the real function evaluations for each objective.

  5. 5.

    Offspring generation: The offspring solutions \(O_t\) can be produced by reproduction operators such as crossover and mutation. Then, the real function evaluations are replaced by the constructed models to predict the objective values of all individuals. Parental selection chooses individuals from A based on \(I_{\epsilon +}\) and newly sampled individuals, plus individuals in A for randomly migrating to commonly form parental population. To ensure the convergence of population, we set the number of generations assisted by surrogates to 20. This parameter is set according to KRVEA [5].

  6. 6.

    Adaptive sampling: By analyzing the improvement of population convergence and diversity, then we adopt an adaptive sampling approach to switch in three strategies based on convergence, diversity, and uncertainty. The sampled solutions would be evaluated by real objective functions, then the results of evaluations would be collected into archive A.

  7. 7.

    Repeat Steps 2–6 until the number of function evaluations is met.

In the following sections, we will discuss the details of Preliminaries, Proposed Algorithm, Empirical studies, and Conclusions.

Noise treatment

As mentioned in “Noisy multi-objective optimization problems”, we need to handle the noises in MOPs. For the minimization MOPs, noises have slight influence to the population in the early stage of optimization, because the population has not converged and the objective values are relatively large. In addition, taking the view into consideration, this view holds that when the impact of noises is slight that is helpful to generate better solutions, but when it is too big that evolutionary search degenerates into random search [30]. Therefore, in order to reduce the influence of noises, we propose a method to estimate the impact of noises and a method to denoise.

According to the above analysis, a method is designed to estimate the influence of noise which divides the whole stage of optimization into early and late stages. When the algorithm is in the early stage of optimization, it means that the noise has little influence on the population, so the noise is not handled. Conversely, the algorithm would construct RBFNs with original data for all objectives and use the regression value as denoised objective value. The method of estimating the influence of the noise is to use the fitness of \(IBEA_{\epsilon +}\) to sort the individuals in the initial population and selects the individual with the best fitness to resample 5 times by real objective functions. Then, we calculate the average value and find out the absolute value of maximum noise \({E_{\max }}\) by the average value and all sample values for each objective. Next, the N-th individual sorted is for the estimation. Then the algorithm calculates the ratio of \({E_{\max }}\) to objective values and records it as \(\lambda \) in each objective, where N is the population size. Finally, we need to compare \(\lambda \) with the threshold of noises \(\delta \). As long as there is an objective \(\lambda \) < \(\delta \), it means that the algorithm is in the early stage of optimization, and the algorithm sets the noise flag to 0. Conversely, the algorithm would set the noise flag to 1, and once the noise flag is 1, the noise estimation method is no longer executed.

The output of RBFN consists of the product of all hidden layer neurons and corresponding weights, plus network bias. The output of all hidden layer neurons is controlled by the distance between input point and hidden layer center point. It is beneficial for denoising, because the correlation of objective values are affected by the distance in the decision space and the noises only exists in the objective space. In addition, The effect of denoising method depends on the number of center points and the value of the spread rate.

Model switch strategy

For the minimization MOPs, the noises in the early stage of optimization has little influences. Therefore, we can ignore the influence of noises for evolutionary process in the early stage of optimization. GP is a surrogate model with excellent performance which cannot only provide the prediction, but also provide the prediction error as the approximation uncertainty. It is extremely important in the early stage of evolution especially when the population is trapped in the local optimum. However, when the noises have non-ignorable influence, GP cannot handle noises well due to the fitting of GP to the objective is overfitting. Thus, It is not appropriate to use GP as the surrogate model in the late stage of evolution. RBFN has the characteristics of local fitting can smooth the noisy data, so RBFN as the surrogate model in the late stage of optimization can achieve better performance. Based on the above analysis, we propose a adaptive model switch strategy between GP and RBFN based on the influence of noises. If the noise flag is 0, it means that algorithm need to concentrate on the convergence of population and approximation uncertainty so the GP surrogate model would be constructed for MS-MOEA. If the noise flag is 1, it means that algorithm needs to concentrate on the influence of noises so the RBFN surrogate model would be constructed for MS-MOEA.

Fig. 3
figure 3

Procedure of adaptive sampling

Improved environmental selection

In the offspring generation, we need to select next parent population by the environmental selection. It is very important to consider the convergence and diversity of population in environmental selection. However, the basic optimizer IBEA\(_{\epsilon +}\) cannot balance both aspects which is not very effective in maintaining diversity [17]. Therefore, we proposed an improved environmental selection.

figure b

It can be seen from the above analysis that the evolutionary stage of the population can be divided into the early and late stages according to the noise flag. When the population is in the early stage, the population has not converged, it is important for environmental selection to pay attention to convergence. At the same time, considering that the population may fall into the local optimum, environmental selection needs to take uncertainty into account. Therefore, we adopt environmental selection based on convergence and uncertainty, respectively. Where convergence selection depends on the \(I_{\epsilon +}\) indictor and uncertainty selection depends on the prediction error of GP. All individuals selected by environmental selection are regarded as new parental populations. When the population is in the late stage of evolution, the population has converged well. To get better solutions, we need to improve the diversity of population. Thus, we adopt environmental selection based on convergence and diversity, respectively. Where diversity selection depends on dominance relationship and the \(L_p\)-norm-based distance [54]. All individuals selected by environmental selection are regarded as new parental populations. For details of the improved environment selection are shown in Algorithm 2.

Adaptive sampling approach

Different sampling approaches can bring different improvements to the population or surrogate model. For example, the convergence sampling and diversity sampling approaches are helpful to improve the performance of population in convergence and diversity, while the uncertainty sampling approach is helpful to improve the global approximation accuracy of surrogate models. Because the MOPs studied in this work are expensive, we need to make good use of limited function evaluations and different sampling approaches. Therefore, an adaptive sampling approach is proposed for MS-MOEA. The procedure of adaptive sampling as shown in Fig. 3.

The idea of adaptive sampling approach is through preserve the populations that obtained by the convergence selection and diversity selection last time in the adaptive sampling are recorded as \(P^{g-1}_c\) and \(P^{g-1}_d\). At the same time, two selection approaches are used to select populations which are regarded as \(P^{g}_c\) and \(P^{g}_d\). \(P^{g-1}_c\) and \(P^{g}_c\) are used to compare the improvement of convergence. If there is an improvement from \(P^{g-1}_c\) to \(P^{g}_c\), the convergence sampling would be adopted to sample. When there is not an improvement from \(P^{g-1}_c\) to \(P^{g}_c\), the algorithm would use the noise flag to judge the evolutionary stage of the population to make a choice from adopting the diversity sampling directly and comparing the improvement of diversity. If there is an improvement from \(P^{g-1}_d\) to \(P^{g}_d\), the diversity sampling would be adopted. When there is not an improvement from \(P^{g-1}_d\) to \(P^{g}_d\), the approach would choose uncertainty sampling.

As for the methods to judge the improvement of convergence and diversity, we adopt two indicators to measure. To measure the improvement of convergence, we calculate the distances from all solutions to the estimated ideal point. To measure the improvement of diversity, we calculate the pure diversity (PD) [55] for each population. The calculation formula of PD is shown in Eq. (15):

$$\begin{aligned} \mathrm{P D}(P)= & {} \max _{s_{i} \in X}\left( \mathrm{P D}\left( P-s_{i}\right) +d\left( s_{i}, P-s_{i}\right) \right) , \nonumber \\ d(s, P)= & {} \min _{s_{i} \in P}\left( {\text {dis}}\left( s, s_{i}\right) \right) , \end{aligned}$$
(15)

where \(dis(s,s_i)\) is the \(L_p\) norm between individual s and individual \(s_i\). d(sP) is used to obtain the minimum \(L_p\) norm. PD is an indicator to measure the diversity of population which does not require any reference, nor any parameters.

For the judgement results, the algorithm uses a specific sampling approach to select \(\eta \) individuals for expensive function evaluations:

  1. 1.

    Convergence sampling approach deletes the individual with minimum fitness calculated by \(I_{\epsilon +}\) in \(P^{g}_c\) until remaining \(\eta \) individuals in \(P^{g}_c\). The details can be found in Algorithm 3.

  2. 2.

    Diversity sampling approach selects the nondominated individual with maximum contribution to the PD of \(P^{g}_d\) until the number of the selected individuals equals to \(\eta \). The details can be found in Algorithm 4.

  3. 3.

    Uncertainty sampling approach randomly selects \(\chi \) solutions from \(P^{g}_d\). Next, the solution with the maximum uncertainty would be selected in \(\chi \) solutions. The approach would be repeated \(\eta \) times and \(\eta \) samples are used to evaluate. The details can be found in Algorithm 5.

figure c
figure d
figure e

Empirical studies

To verify the performance of the proposed algorithm, we compare it with five SAEAs. The five algorithms include KRVEA [5], ParEGO [25], MOEA/D-EGO [26], CSEA [29], and KTA2 [17]. The characteristics of these algorithms are summarized as follows:

  • KRVEA: The reference vector guided evolutionary algorithm (RVEA) [56] is adopted as the basic optimizer. The Kriging model is used to approximate each objective in KRVEA. It uses APD to select solutions for evaluating, but when the diversity of population decreased, the solutions with maximum uncertainty would be selected.

  • ParEGO: It uses an aggregation function to scalarize the MOP into an SOP using a randomly chosen weight vector for each generation.The Kriging model is used to approximate the scalar function. It selects the solution with the best EI for evaluating.

  • MOEA/D-EGO: MOEA/D [11] is used as the basic optimizer. It uses a fuzzy clustering method to construct multiple GPs as local models. For each SOP, the solution with the best EI would be selected.

  • CSEA: It belongs to the SAEAs based on classification. The existing data is used to train a feed-forward neural networks (FNN) [57] to classify solutions with more potentiality to evaluate.

  • KTA2: Two_Arch2 [54] is adopted as the basic optimizer. It uses an influential point-insensitive Kriging model [17] to approximate each objective function. In addition, it adopts an adaptive sampling strategy to select solutions to evaluate.

This set of the experiments are conducted on DTLZ problems with two, three objectives. In the experiments, the number of decision variables of DTLZ1 and DTLZ3 is set to seven. The number of decision variables of other test problems is set to twelve. All algorithms use the Wilcoxon rank-sum (WRS) test to compare the results at a significance level of 0.05 over 30 independent runs. “+”, “−” and “\(\approx \)” are used to compare the proposed algorithm and compared SAEAs. Symbols “+”, “−” denote that MS-MOEA performs significantly better and significantly worse than the compared SAEAs. Symbol “\(\approx \)” indicates that there is no statistically significant difference between MS-MOEA and the compared SAEAs.

The inverted generational distance (IGD) is adopted for all experiments as performance indicator. IGD is an comprehensive indicator that can consider both convergence and diversity of the nondominated solutions. IGD is defined as follows:

$$\begin{aligned} \mathrm{IGD}(\mathrm{P F}, P)=\frac{\sum _{x \in \mathrm{P F}} d_{\min }(x, P)}{|\mathrm{P F}|}, \end{aligned}$$
(16)

where PF is a set of points uniformly distributed on the real Pareto front and \(|\mathrm{PF}|\) represents the number of elements in PF. For all test problems, the number of reference points for calculating IGD is set to 10,000. P is the nondominated solution set obtained by the algorithm. \(d_{\min }(x,P)\) is the minimum Euclidean distance from solution x in the PF to the solution in the P.

Parameter settings

To ensure the fairness of the experiments, the common parameter setting of all compared SAEAs are listed below:

  1. 1.

    The initial population size N and the number of initial data are set to 100.

  2. 2.

    The number of expensive function evaluations is set to 300.

  3. 3.

    The variance of additive noise is set to 0.2.

In addition, the other parameter settings of the five compared SAEAs are set according to the recommended settings in the original literature. The key parameters of the proposed algorithm are set as follows:

  1. 1.

    The fitness scaling factor of IBEA\(_{\epsilon +}\) k is set to 0.05.

  2. 2.

    The number of individuals for sampling \(\eta \) is set to 5.

  3. 3.

    The hyperparameter range of Matlab Toolbox DACE for GP model is set to \([10^{-5},10^{5}]\). We adopt the algorithm raised by Hookes and Jeeves [58] to optimize the hyperparameters.

  4. 4.

    The number of center points of RBFN hidden layer is set to ceil\(\sqrt{\mathrm{size}(D)}\), D is the number of the training data.

  5. 5.

    The spread rate \(\sigma \) is set to 2*Dis, Dis is the maximum distance between all center points.

  6. 6.

    The number of model assisted iteration \(w_{\max }\) is set to 20.

  7. 7.

    The threshold of noise \(\delta \) is set to 0.03.

  8. 8.

    The parameters of crossover and mutation are set to \(p_c = 1\), \(\eta _c = 20\), \(p_m = 1\), \(\eta _m = 20\).

MS-MOEA has a parameter \(\delta \) which directly affects the performance of noise treatment. It is used for estimating the influence of noise. Therefore, we set an experiment on DTLZ1 and DTLZ2 to select a proper value for the threshold of noise \(\delta \). DTLZ1 is a hard-to-converge test function. It is usually difficult to converge to the late stage of evolution when the noise has non-ignorable impact. DTLZ2 is an easy-to-converge test function. It is easy to converge to the late stage of evolution. That is the reason why we select them as the test functions. To better compare the effects of different noise thresholds on the performance of the proposed algorithm, we take a set of \(\delta = \{0.0003,0.003,0.03,0.13,0.23,0.33,0.43,0.53\}\) for experiments. For the convenience of observation, we set the x-axis coordinate to lg(x). Figure 4 is the result of DTLZ1 and DTLZ2 with two objectives which is executed by 30 runs independently.

Fig. 4
figure 4

a Is the performance profile plots of DTLZ1, b is the performance profile plots of DTLZ2 Average IGD profile plots over the different parameter \(\delta \) in MS-MOEA on the two objectives DTLZ1 and DTLZ2 based on 30 independent runs

Table 1 Results of N-(MS-MOEA) and MS-MOEA on the DTLZ2 and DTLZ5 problems

From Fig. 4, we can discover that \(\delta \) determines the sensitivity of noise estimation method. The smaller this parameter \(\delta \) is set, the more sensitive it is to noises. From Fig. 4a, we can discover this conclusion. When \(\delta \) \(\ge \) 0.03, it gets the familiar results in DTLZ1. When \(\delta \) < 0.03, the algorithm performance is significantly reduced. The larger this parameter \(\delta \) is set, the more non-sensitive it is to noises. From Fig. 4b, we can discover the conclusion. When \(\delta \) = 0.03, it gets the best results in DTLZ2. When \(\delta \) is set be larger, the noise needs to cause greater impact on the population before denoising. For keeping the suitable sensitivity of noise estimation method, \(\delta \) is set to 0.03.

Effect of noise treatment

To test the performance of the proposed noise treatment, we design an experiment that selects DTLZ2 and DTLZ5 as the test problems. Because DTLZ2 and DTLZ5 belong to the test problems that are sensitive to noises. Every test problem would be tested 30 runs independently. MS-MOEA and the version without noise treatment, termed N-(MS-MOEA), are compared. To ensure the fairness of the experiment, the objective values without noises are used in the calculation of IGD, and the real objective values do not participate in population iteration. The results are shown in Table 1.

The experimental results are as expected, MS-MOEA have achieved the best results on all test problems. It proves the effectiveness of our proposed noise treatment method. Noises in the objective space have a direct impact on the dominance relationship between individuals in the population. It would has an adverse impact on the evolutionary search of the population, and even degenerate into random search. Therefore, when the noises has a large influence on the objective values, it is very necessary to deal with the noises.

Effect of model switch strategy

To discuss the role of model switch in MS-MOEA, we set up the following experiment. This experiment adopts DTLZ1, DTLZ2 and DTLZ5 as the test problems. Each problem would be tested 30 runs independently. According to the different models used in the experiment, the experimental variables are recorded as (MS-MOEA)-GP, (MS-MOEA)-RBFN and MS-MOEA. To ensure the fairness of the experiment, the objective values without noises are used in the calculation of IGD, and the objective values do not participate in the optimization process. In addition, the frequency of usage of the two surrogate models in MS-MOEA is used to further explain the experimental results. The meaning of FG and FR is shown below. The results are shown in Table 2.

  • FG: It is the frequency of the usage of GP model.

  • FR: It is the frequency of the usage of RBFN model.

Table 2 Results of (MS-MOEA)-GP, (MS-MOEA)-RBFN and MS-MOEA on the DTLZ2 and DTLZ5 problems

According to the results of experiment, we discover that the model switch strategy can find better surrogate model between GP and RBFN. It also proves that selecting an appropriate model according to the influence of noise on the objective values is conducive to evolutionary optimization. In addition, we can find that the performance of MS-MOEA and (MS-MOEA)-GP is similar in DTLZ1, it is because DTLZ1 is a hard-converging test function. Therefore, it is always in the early stage of the evolution process, when the impact of the noise is little. That is why the RBFN is never used in DTLZ1. In addition, we can discover the performance of using GP as the surrogate model is better than using RBFN as the surrogate model from the experimental results. The performance of MS-MOEA and (MS-MOEA)-RBFN is similar in DTLZ2 and DTLZ5, because both of them are sensitive to the noise. Therefore, they are always in the late stage of the evolution process, when the impact of the noise is non-ignorable. That is why the GP is never used in DTLZ2 and DTLZ5. In addition, we can discover the performance of using RBFN as the surrogate model is better than using GP as the surrogate model from the experimental results. The results can prove that our model switch strategy can skillfully use the influence of noise to select the appropriate surrogate model for the evolutionary optimization.

Effect of improved environment selection

Table 3 Results of (MS-MOEA)-C+C, (MS-MOEA)-C+D, (MS-MOEA)-U+C, (MS-MOEA)-U+D and MS-MOEA on the DTLZ2 and DTLZ5 problems

To verify the role of the improved environment selection, we design an experiment. The conventional environment selection methods used in different evolutionary stages in the algorithm are compared. This experiment adopts DTLZ1, DTLZ2, DTLZ5 and DTLZ7 as the test problems. Each problem would be tested 30 runs independently, too. To ensure the fairness of the experiment, the objective values without noises are used in the calculation of IGD, and the real objective values does not participate in all the compared algorithms. The results are shown in Table 3.

  • (MS-MOEA)-C+C: It only adopts convergence selection in all evolutionary stages.

  • (MS-MOEA)-C+D: It adopts convergence selection in the early stage and diversity selection in the late stage.

  • (MS-MOEA)-U+C: It adopts uncertainty selection in the early stage and convergence selection in the late stage.

  • (MS-MOEA)-U+D: It adopts uncertainty selection in the early stage and diversity selection in the late stage.

As expected, MS-MOEA almost makes the best results in all the variants. From the results of experiments, (MS-MOEA)-C+C and MS-MOEA have similar performance on DTLZ1, because DTLZ1 is a difficult test problem, it is better to adopt a convergence selection in all evolutionary stages. We can also find that (MS-MOEA)-C+C and (MS-MOEA)-U+C have similar performance in DTLZ2 and DTLZ5. This is mainly because DTLZ2 and DTLZ5 are sensitive to noises. Therefore, the environmental selection in the late stage plays an important role. As for the reason why MS-MOEA is not significantly better than other algorithms on DTLZ2 and DTLZ5 is that the noise treatment can only deal with noises to a certain extent. Some objectives of DTLZ7 are sensitive to noise and some objectives are insensitive. Our improved environment selection uses different environmental selections for different evolutionary stages. Therefore, it has achieved satisfactory results. In summary, the selection considering convergence, diversity and uncertainty are the key to achieve the most satisfactory performance of MS-MOEA.

Effect of adaptive sampling approach

To test the performance of the adaptive sampling approach in MS-MOEA, MS-MOEA is compared with MS-MOEA without adaptive sampling approach that is regarded as (MS-MOEA)-N. This experiment adopts DTLZ1, DTLZ2 and DTLZ7 as the test problems. Each problem would be tested 30 runs independently. To ensure the fairness of the experiment, the objective value without noises is used in the calculation of IGD, and the objective value does not participate in the optimization process. The results are shown in Table 4.

Table 4 Results of (MS-MOEA)-N and MS-MOEA on the DTLZ1, DTLZ2 and DTLZ7 problems

By analyzing the experimental results, we discover that the algorithm with adaptive sampling approach have better performance than the other one. Because the adaptive sampling can find the improvement of convergence and diversity and use targeted sampling approach to select solutions to evaluate. It is beneficial to make full use of the limited function evaluations. It is the adaptive sampling approach that makes MS-MOEA get the best results on all test problems.

To study the usage of different sampling approach in every test function, we summarize the results of 30 independent runs about them and these results are displayed in Fig. 5. It is easy to be discovered that the convergence sampling approach is most frequently used in DTLZ1, the uncertainty sampling approach is next and the diversity sampling approach is the last. DTLZ1 is a multi-modal test function, so it is difficult to converge and easy to fall into a local optimum. In DTLZ2, the diversity sampling approach is also more adopted. Because DTLZ2 is an easy to converge test function, it needs to take both convergence and diversity into account. The most frequently adopted sampling approach in DTLZ7 is the diversity sampling approach. The real PF of DTLZ7 is irregular. Therefore, the algorithm needs to pay attention to the maintenance of diversity. According to Fig. 5 and Table 4, our adaptive sampling approach can select appropriate sampling approach from the convergence approach, the diversity approach and the uncertainty approach based on the characteristic of the test functions.

Comparative experiments

To test the overall performance of the proposed algorithm, we employ KRVEA, ParEGO, MOEA/D-EGO, CSEA and KTA2 as compared algorithms and carry out experiments on the test problems DTLZ1\(\sim \)7. This paper uses the compared algorithms implemented in PlatEMO [59] to test. The number of objectives includes two and three in this experiment. IGD is used as performance indicator for 14 test problems, and each problem would be tested 30 runs independently. For ensuring the fairness of the experiment, the objective values without noises is used in the calculation of IGD, and the objective values does not participate in process of all the compared algorithms. The results are shown in Table 5.

Fig. 5
figure 5

Average times of different sampling approach in DTLZ1, DTLZ2 and DTLZ7

Table 5 Results of KRVEA, ParEGO, MOEA/D-EGO, CSEA, KTA2 AND MS-MOEA on the DTLZ1\(\sim \)7 problems
Fig. 6
figure 6

Non-dominated front obtained by each algorithm on DTLZ2 with two objectives in the run associated with the median IGD value, respectively

Fig. 7
figure 7

Non-dominated front obtained by each algorithm on DTLZ1 with three objectives in the run associated with the median IGD value, respectively

From the experimental results, we can observe that MS-MOEA obtains the best results of 12 test problems; ParEGO obtains the best results of 1 test problems; KTA2 obtains the best result on 1 test problems; KRVEA, MOEA/D-EGO and CSEA do not get the best results on any test problem. Therefore, MS-MOEA has achieved the best performance between all compared algorithms on the DTLZ problems. The final non-dominated solutions obtained by six algorithms on two objectives DTLZ2 in the run associated with the median IGD values are plotted in Fig. 6. In addition, Fig. 7 shows the final non-dominated solutions obtained by six algorithms on three objectives DTLZ1 in the run associated with the median IGD values.

According to the characteristics of DTLZ series test problems, all the above test problems can be divided into three categories:

  1. 1.

    DTLZ1 and DTLZ3 are the first category. This kind of test problems are multimodal and have a large number of local optimal solutions, so it is difficult to converge to the true PF. DTLZ3 is more difficult to converge than DTLZ1.

  2. 2.

    DTLZ2 and DTLZ4 are the second type of test problems, which are uni-modal and easy to converge, but it is difficult to maintain the diversity of the population, and DTLZ4 is sensitive to the initial data.

  3. 3.

    DTLZ5\(\sim \)7 are the third type. The PF of these test problems are irregular. The real PF of DTLZ5 and DTLZ6 are degenerated curves, while the real PF of DTLZ7 is discontinuous. These test problems require that the algorithm can better maintain the diversity of population.

For the first type of test problems, MS-MOEA almost achieves the best result. Through the analysis of the results, it is found that the reason for this result may be that the model switch strategy can select a suited surrogate model from GP and RBFN. At the same time, the adaptive sampling approach can make use of approximation uncertainty to get rid of the local optimum. In addition, KTA2 has the similar results with MS-MOEA when the number of objectives is three. Because both of DTLZ1 and DTLZ3 are difficult to converge, noise have little influence on them. However, when the number of objectives is two, MS-MOEA is significantly better than KTA2. Because KTA2 is designed for many-objectives problems.

For the second type of test problems, MS-MOEA achieves the best results which show that the proposed noise treatment is very effective. The adaptive sampling approach can also find the most improved aspect for sampling by analyzing the population of candidate solutions. However, other compared algorithms are greatly affected by noise, because they do not deal with the noise, resulting in the degradation of algorithms performance.

For the third type of test problems, MS-MOEA obtains the best results on DTLZ5 and DTLZ6, which benefits from the noise treatment, the improved environmental selection and adaptive sampling approach. However, the performance of DTLZ7 is inferior to ParEGO. Through the analysis of the experimental results, it is found that the reason is that the noise treatment designed in this paper is lack when some objectives are easy to converge and other objectives are difficult to converge in the face of the test problem, which may be the reason why MS-MOEA is inferior to ParEGO.

Conclusions

In this work, we propose an adaptive model switch-based surrogate-assisted evolutionary algorithm called MS-MOEA for noisy expensive multi-objective optimization problems. we propose a method to estimate the influence of noise and a noise reduction method based on RBFN, and the noise experiment shows that this noise treatment is effective. This algorithm adopts a model switch strategy based on noise influence to select the surrogate model that meets the needs from GP and RBFN. We find that this strategy is helpful to improve the performance of the algorithm by the experiment. The proposed improved environmental selection and adaptive sampling approach are used to obtain candidates and samples. Through the comparison between MS-MOEA and the other five SAEAs, it is found that MS-MOEA achieves better performance than the compared SAEAs on noisy DTLZ problems.

However, there are still two deficiencies that can be improved in this paper. We need to improve the adaptive sampling approach for multimodal MOPs, these problems require the algorithm to use approximation uncertainty to search better solutions. For the same MOP, there are both noise sensitive and noise insensitive objectives, the noise treatment method performs poorly. In addition, the noise studied in this paper belongs to additive noises. Compared with the additive noise, the multiplicative noise is a more challenging task to be addressed.