Introduction

In real-world applications, such as parallel machine scheduling [1], hybrid electric vehicle optimization [2], workflow scheduling in clouds [3], and chemical refinery process optimization [4], many problems are required to simultaneously optimize a number of conflicting objectives, which are called multi-objective optimization problems (MOPs) when the number of objectives is not more than three and many-objective optimization problems (MaOPs) when that is not less than four. Generally, the mathematical model of a multi-/many-objective optimization problem can be given as follows:

$$\begin{aligned} \min&\mathbf {f}(\mathbf {x}) = (f_1(\mathbf {x}),f_2(\mathbf {x}),\ldots ,f_m(\mathbf {x})) \nonumber \\ \text {s.t. }&\mathbf {x} \in \mathbb {R}^D \end{aligned}$$
(1)

where m is the number of objectives, \(\mathbb {R}^D\) represents D-dimensional decision space, and \(\mathbf {x} = (x_{1},x_{2},\ldots ,x_{D})\) is the decision vector.

Due to the conflicts between the objectives, it is impossible to find a solution that has the best performance on all objectives [5]. Instead, a set of optimal solutions will be found for a multi-/many-objective optimization problem, which is called the Pareto set (PS) in the decision space and the Pareto front (PF) in the objective space. The evolutionary algorithms have been widely applied to solve multi-objective optimization problems because they can provide a number of optimal solutions in a single run. However, it is difficult to find a good set of solutions for the problem when the number of objectives increases due to the loss of selection pressure. A number of evolutionary algorithms have been proposed for solving many-objective optimization problems [6, 7], denoted as MaOEAs, which can mainly be classified into three categories: (1) Pareto dominance-based MaOEAs, (2) indicator-based MaOEAs, and (3) decomposition-based MaOEAs [6].

The Pareto dominance-based MaOEAs mainly focus on convergence performance. However, the selection pressure of most evolutionary algorithms will vanish when the number of objectives increases. Many approaches have been proposed to modify the dominance relationship to push the population toward the true Pareto front [8]. In [9], the fuzzification of the Pareto-dominance relation is used to support a new ranking method to select individuals in the environmental selection. \(\epsilon \)-dominance [10] employed a relaxed factor \(\epsilon \) to control the dominance relation between solutions. Yang et al. [11] proposed a Pareto dominance based on grid-dominance to select the solution having the higher priority of dominance. In \(\alpha \)-dominance [12], an \(\alpha \)-domination strategy that relaxes the domination introduced a weak trade-off among optimization objectives.

The indicator-based MaOEAs adopt the performance indicators for environmental selection. Normally, the number of performance indicators is much less than the number of objectives, thus increasing the selection pressure. Villalobos and Coello [13] proposed a \(\Delta P\) indicator in conjunction with a differential evolution [14] (called \(\Delta P\)-MOEA) for many-objective optimization problems. Ishibuchi et al. [15] proposed an inverse generational distance plus (IGD\(^{+}\)) indicator-based evolutionary algorithm (IGD\(^{+}\)-EMOA) for MaOPs with no more than eight objectives. Tian et al. [16] proposed a new evolutionary algorithm based on an improved inverted generational distance indicator for solving many-objective optimization problems with different shapes of the Pareto front. Recently, Li et al. proposed to apply the stochastic ranking technique in SRA [17] to balance the search biases on different indicators. In [18], Wang et al. suggested sorting on convergence and diversity performances for environmental selection.

The decomposition-based MaOEAs normally solve a many-objective optimization problem by converting it into a set of single-objective optimization problems or dividing the Pareto front into a group of segments to be a number of many-objective optimization sub-problems [19]. Yuan et al. [19] proposed to preserve the diversity of solutions in high-dimensional objective space by exploiting the perpendicular distance from the solution to the reference vector in the objective space. In [20], Cheng et al. presented an adaptive strategy to adjust the distribution of the reference vectors dynamically and adopted the angle-penalized distance to balance the convergence and diversity of solutions in the high-dimensional objective space. Based on MOEA/D-M2M, a MOEA/D variant, Liu et al. [21] tried to detect the importance of different objectives for adaptively adjusting the subregions of its subproblems. In [22], Chen et al. proposed to adaptively partition an objective space based on the contributions of subspaces to the population convergence to improve the population convergence while keeping the population diversity.

Note that some MaOEAs are not able to fall into the above three categories as multiple strategies are simultaneously utilized. For example, in MOEA/DD [23] and NSGA-III [24], both the dominance and decomposition strategies are utilized. In [25], the indicator and dominance strategies are used.

The literature review shows that the balance between the convergence and uniformity of the optimal solutions found for MaOPs is significantly important for the optimization algorithms. The decomposition-based MaOEAs decompose a many-objective optimization problem into a set of sub-problems, which are optimized simultaneously by using scalar functions [20, 26]. Thus, the reference vectors and the aggregation functions play significant roles in finding a good set of solutions uniformly distributed on the front and approaching the true Pareto front as much as possible. Although the reference vectors are able to assist in maintaining the uniform of a solution set, the performance of the method for finding optimal solutions mainly relies on the search direction and the diversity of the population. Correct search directions are used to ensure the quick convergence to the PF, and the diversity of the population can assist in preventing falling into local optima.

Generally, the same performance indicator will be used for all solutions in the current population for environmental selection. The performance indicator used focuses on the convergence only, such as the Tchebycheff function, or the convergence and diversity simultaneously in a criterion, such as PBI and APD. When only the convergence performance is used for environmental selection, the diversity will not be considered, resulting in poor exploration and incapable of covering the Pareto front. On the other hand, the indicator simultaneously considering the convergence and diversity performance of the solution in a criterion for environmental selection may slow down the speed to find the optimal solutions.

Thus, in our method, convergence and diversity performances of solutions are used simultaneously at each generation but are applied for different reference vectors for selecting solutions for the next generation. Some solutions in the population are selected based on the convergence performance, which is expected to guide the population to approach the Pareto front quickly. In contrast, others are selected based on the diversity performance, which is expected to explore different regions for finding optimal solutions as much as possible. Thus, the trade of balance between exploitation and exploration is met.

The main contributions of this paper are summarized as follows:

  1. 1.

    Each solution plays a role in exploitation or exploration. Thus, the balance between convergence and diversity can be ensured for the population.

  2. 2.

    A mating pool is used for mating selection. The solutions in the mating pool are selected based on the closeness to each reference vector.

  3. 3.

    The computational complexity of the proposed method is \(O(N \times |A|)\) if \(|A|>N\), where |A| is the number of solutions in the archive A and N represents the population size. However, when \(|A| < N\), the computational complexity of MaOEA/D-OPI is \(O(N^2)\).

The remainder of this paper is organized as follows: Section “Reference vectors” briefly reviews the reference vector used in this work. Section “A decomposition-based many-objective evolutionary algorithm with optional performance indicators (MaOEA/D-OPI)” gives a detailed description of the proposed algorithm MaOEA/D-OPI. An empirical analysis of the experimental results is provided in Section “Experimental studies”. Finally, the conclusions and future work are summarized in Section “Conclusion”.

Reference vectors

The reference vectors are unit vectors starting from the origin and distributing uniformly in the objective space. In our proposed method, the two-layer weight vector generation method proposed by Li et al. [27] is adopted to generate a set of reference vectors. Two sets of uniformly distributed points, denoted as \(BD = \{\mathbf {bd}_1, \mathbf {bd}_2, \ldots , \mathbf {bd}_{N_1}\}\) and \(IN = \{\mathbf {in}_1, \mathbf {in}_2, \ldots , \mathbf {in}_{N_2}\}\), respectively, are firstly generated on a unit hyperplane using the canonical simplex-lattice design method, in which

$$\begin{aligned} {\left\{ \begin{array}{ll} \mathbf {bd}_i = (bd^{1}_i, bd^{2}_i, \ldots , bd^{m}_i), \\ \mathbf {in}_j = (in^{1}_j, in^{2}_j, \ldots , in^{m}_j), \\ db^{k}_i \in \{\frac{0}{H},\frac{1}{H},\ldots ,\frac{H}{H}\}, \sum _{k=1}^m bd^{k}_i = 1, \\ in^{k}_j \in \{\frac{0}{H},\frac{1}{H},\ldots ,\frac{H}{H}\}, \sum _{k=1}^m in^{k}_j = 1,\\ i=1,2,\ldots ,N_1,\\ j=1,2,\ldots ,N_2 \end{array}\right. } \end{aligned}$$
(2)

where \(N_1\) and \(N_2\) are the number of uniformly distributed points, \(N_1 + N_2 = N\) and H is a positive integer for the simplex-lattice design [24]. Then the coordinates of solutions in IN are shrunk by a coordinate transformation, which will be changed to

$$\begin{aligned} in^k_j = \frac{1-\tau }{m}+\tau \times in^k_j \end{aligned}$$
(3)

where \(\tau \in [0,1]\) is called a shrinkage factor, which is set to 0.5 in our method as that in [27]. After that, the corresponding unit reference vectors \(\mathbf {w}\) can be obtained by mapping the reference points, including those in both \(\mathbf {bd}\) and \(\mathbf {in}\), from a hyperplane to a hypersphere.

A decomposition-based many-objective evolutionary algorithm with optional performance indicators (MaOEA/D-OPI)

In one category of the decomposition-based many-objective evolutionary algorithms, a multi-/many-objective optimization problem will be converted to a number of single-objective optimization problems using a scalarization method, such as weighted sum, weighted Tchebycheff [28], penalty-based boundary intersection (PBI) [26], and angle penalized distance (APD) [20]. Normally, the same performance indicator is used for the environmental selection of the many-objective evolutionary algorithms from beginning to end. Thus, when only the convergence performance is used for environmental selection in the decomposition-based algorithms, the exploration of the method will be poor, and the non-dominated solutions will be incapable of covering the Pareto front due to the lack of diversity. On the other hand, when both convergence and diversity are considered for environmental selection, the speed of finding the optimal solutions will be slow down. Thus, in this paper, we propose to use convergence and diversity performance simultaneously at each generation, which is applied to different reference vectors to select the solutions for the next generation. Some solutions in the population are chosen based on the convergence performance, which is expected to guide the population to approach the Pareto front quickly. The remainders are selected based on the diversity performance, which is expected to explore different regions to find optimal solutions as much as possible. The details of the proposed method are given in this section.

The general framework of MaOEA/D-OPI

Algorithm 1 presents the pseudocode of the proposed MaOEA/D-OPI algorithm. An initial population P with N solutions will be randomly generated in the decision space and evaluated using the objective functions. The non-dominated solutions in P will be selected by the fast sorting technique proposed by Deb et al. [29] and saved to an archive Arc. Then a set of uniformly distributed reference vectors \(\mathbf {W}\) will be generated using the method given in [27]. The following procedure will be repeated when the stopping criterion is not met. Two solutions will be selected from the current parent population P for each reference vector. The angle between the reference vector and the selected two solutions will be calculated, respectively, and the solution closest to this reference vector will be kept for the reproduction of the offspring. The solutions in the offspring population will be used to update the archive Arc. Simultaneously, each solution in the parent and offspring populations will be associated with its closest reference vector, and the environmental selection strategy proposed in this paper is adopted to choose the solutions that survive to the next generation. N solutions in the archive Arc closest to each reference vector will be output.

figure a

Mating selection

The simulated binary crossover (SBX) [30] and the polynomial mutation [31] are used for the reproduction of the population. In order that the solutions participating in the crossover are uniformly distributed in the objective space, in our proposed method, a pair of solutions in the parent population will be randomly selected for each reference vector, and the solution with the smaller angle to this reference vector will be kept for reproduction. The method to calculate the angle between the objective vector of the solution \(\mathbf {x}_i\), denoted as \(\mathbf {f}(\mathbf {x}_i)\), and the reference vector \(\mathbf {w}_j\) is given in the following:

$$\begin{aligned} \theta _{ij} = \arccos \frac{\mathbf {f}(\mathbf {x}_i)\cdot \mathbf {w}_j}{\Vert \mathbf {f}(\mathbf {x}_i)\Vert \Vert \mathbf {w}_j\Vert } \end{aligned}$$
(4)

where \(\theta _{ij}\) represents the angle between the objective vector \(\mathbf {f}(\mathbf {x}_i)\) and the reference vector \(\mathbf {w}_j\).

Algorithm 2 gives the pseudocode of the mating selection.

figure b

Environmental selection

Both convergence and diversity of a population play a significantly important role in finding a set of good solutions that approach close to the true Pareto front and are distributed uniformly as much as possible on the Pareto front. Thus, either the exploitation or the exploration capability of a solution is critical in the optimization. In most existing evolutionary algorithms, a unified criterion, normally focusing on convergence only or convergence and diversity simultaneously, is applied to measure the performance of each solution in the population. Then a new population will be selected. However, in the case that only the convergence performance is considered in environmental selection, poor exploration may happen due to the lack of diversity of the population. Thus, it will result in the incapability of covering the Pareto front. On the other hand, when the convergence and diversity performances are ensembled for the environmental selection to measure the performance of a solution, the moment for finding the optimal solutions may be put off. Thus, to find a set of solutions with good convergence and uniformly distribution, we propose to provide two scalarization functions for each reference vector, and one of them will be used for choosing one of all solutions associated with this reference vector to survive to next generation. Thus, some solutions with good convergence will lead the population to exploit the region with good objective values, and others will take on the duty to explore the search space to find more solutions as much as possible. Equations (5) and (6) give the scalarization functions focusing on exploitation and exploration, respectively, of the objective vector of solution \(\mathbf {x}_i\) associated for each reference vector \(\mathbf {w}_k\).

$$\begin{aligned} C_{ik} = \Vert \mathbf {f}(\mathbf {x}_i)\Vert \times \cos \theta _{ik} \end{aligned}$$
(5)
$$\begin{aligned} D_{ik} = \Vert \mathbf {f}(\mathbf {x}_i)\Vert \times \sin \theta _{ik} \end{aligned}$$
(6)

\(C_{ik}\) and \(D_{ik}\) represent two criteria for environmental selection. Note that only one of Eqs. (5) and (6) will be used for each reference vector to select a solution to be kept to the next generation.

Algorithm 2 gives the pseudocode of the mating selection used in our method. In Algorithm 3, the angle of each solution \(\mathbf {x}_i\) of the populations P and O to each reference vector \(\mathbf {w}_k\), denoted as \(\theta _{ik}\), will be calculated using Eq. (4), and \(\mathbf {x}_i\) will be associated with the reference vector \(\mathbf {w}_k\) that satisfies \(k = \{k|argmin_{k=1,2,\ldots ,N}\theta _{ik}\}\). After that, for each reference vector \(\mathbf {w}_k\) that has solutions associated with, a scalarization function will be selected randomly, and the solution with the best value of the scalarization function will be kept to the next generation.

figure c

Figure 1 gives a simple example to show the environmental selection strategy used in our method. Suppose there are three solutions associated with the reference vector \(\mathbf {w}_k\). Clearly, solution \(\mathbf {x}_1\) will be kept to the next generation if Eq. (5) is adopted as the scalarization function because \(\mathbf {x}_1\) has the smallest distance to the ideal point along this reference vector. Otherwise, if Eq. (6) is utilized as the scalarization function, solution \(\mathbf {x}_3\) will be selected to be survived to the next generation. It is because solution \(\mathbf {x}_3\) is closest to the reference vector, which is expected to contribute to the exploration of good solutions.

Fig. 1
figure 1

An example of environmental selection

Experimental studies

In order to evaluate the performance of the proposed method, we conduct a number of experiments on DTLZ [32], WFG [33] and MaF [34] benchmark problems, and compare the results to some state-of-the-art algorithms for many-objective optimization problems.

Test problems and performance metrics

Three widely used test suits, DTLZ [32], WFG [33] and MaF [34], are adopted for the empirical studies in this paper. The number of objectives is set to 3, 5, 8, 10, 15 and 20 for DTLZ and WFG, and is set to 10, 15 and 20 for MaF test problems, respectively. The characteristics of DTLZ, WFG and MaF are summarized in Table 1. The number of decision variables is set to \(D = m+L-1\) as recommended in [35], where m is the number of objectives, and \(L = 5\) for DTLZ1, \(L = 10\) for DTLZ2-6, WFG1-9, MaF1-7 and MaF10-12, while \(L = 20\) for DTLZ7. For other MaF test problems, the number of decision variables is set to 2 for MaF8 and MaF9, and \(m \times 20\) for MaF14 and MaF15, respectively.

Table 1 The characteristics of three benchmark test suits

Two performance criteria, including the inverted generational distance (IGD) [36] and hypervolume (HV) [37], are used to evaluate the performance of different algorithms for DTLZ, MaF and WFG, respectively, in this paper. Both of them can simultaneously measure the convergence and diversity of the optimized solutions.

Let \(P^{*}\) represents a set of solutions sampled uniformly on the true Pareto front, and Q is the set of optimal solutions found by an algorithm. Then the IGD value of the set Q can be calculated using the following equation:

$$\begin{aligned} IGD(Q,P^{*}) = \frac{{\sum \nolimits _{\mathbf {x}^{*} \in P^{*}} {dist(\mathbf {x}^{*},Q)} }}{{|P^{*}|}} \end{aligned}$$
(7)

where \(dist(\mathbf {x}^{*},S)\) represents the Euclidean distance between a point \(\mathbf {x}^{*}\) in P and its closest neighbor point in S. \(|P^{*}|\) denotes the number of solutions in \(P^{*}\). From Eq. (7), we can see that if the number of solutions in \(P^{*}\) is large enough, the solutions sampled will uniformly cover the true Pareto front. Thus, the smaller the IGD value is, the better the performance of the algorithm is. In our experiments, the number of solutions in \(P^{*}\) is set to 10000.

The hypercube column (HV) metric [37] can provide a comprehensive assessment of the convergence and diversity of the optimized solutions by calculating the volume of the objective space enclosed by the optimized solution set and a reference point. Let \(Z^{r}=(z^{r}_{1},z^{r}_{1},...,z^{r}_{m})^{T}\) be the reference point in the objective space that is dominated by all solutions in the current Pareto set. The HV value of the optimization solutions can be calculated using the following equation:

$$\begin{aligned} HV(Q) = VOL(\mathop \cup \limits _{x \in Q} [{f_1}(\mathbf {x}),z_1^r] \times ...[{f_m}(\mathbf {x}),z_m^r]) \end{aligned}$$
(8)

where \(VOL(\cdot )\) represents the Lebesgue measure. The larger the HV value is, the better the quality of Q approaching the true PF.

The parameter settings

All algorithms taking part in the comparisons are run on the PlatEMO2.0 [38] platform. The crossover probability is set to \(P_{c}=1.0\) with distribution index \(\eta _{c}=20\), and the mutation probability is set to \(P_{m}=1.0\) with distribution \(\eta _{m}=20\).

For making a fair comparison, the population size is set to be the same number based on the number of objectives for all algorithms, which are listed in Table 2. \(p_1\) and \(p_2\) represent the number of reference points on the boundary and inside layers, respectively. All algorithms perform 20 independent runs for each problem and will be terminated when the maximum number of objective evaluations reaches 50, 000. The Wilcoxon rank-sum test [39] with a significance level of 0.05 is applied to assess whether the performance of a solution obtained by one of the two methods is expected to be better than the other [40]. The symbols ‘\(+\)’, ‘\(\approx \)’ and ‘−’ represent the compared algorithms that are significantly better, equivalent to, and worse than MaOEA/D-OPI, respectively, in terms of the median values.

Table 2 The setting of the population size N

Comparison with MaOEA/D-OPI variants

Unlike the existing methods, two optional criteria are provided for any reference vector to select the solution that survived to the next generation along this reference vector. In order to evaluate the performance of the proposed strategy, we substitute the environmental selection criterion used in this paper with the existing environmental selection criteria, including APD, Tchebycheff and PBI. These variants are denoted as MaOEA/D-APD, MaOEA/D-Tchebycheff and MaOEA/D-PBI, respectively. Tables 3 and 4 give the statistical results on DTLZ1-DTLZ7 with 3, 5, 8, 10, 15, 20 objectives and WFG1-WFG9 with 10, 15, and 20 objectives, respectively. From Table 3, we can see that MaOEA/D-OPI can obtain 18/42, 15/42 and 26/42 better results than MaOEA/D-APD, MaOEA/D-Tchebycheff and MaOEA/D-PBI, respectively, and from Table 4, we can see that MaOEA/D-OPI can obtain 23, 27 and 27 better results out of 27 problems, showing that our proposed environmental selection strategy is much better than existing approaches. In order to see the efficiency of providing two performance indicators for environmental selection, we further modify our proposed MaOEA/D-OPI method by using only one of the performance indicators, denoted MaOEA/D-convergence and MaOEA/D-diversity, respectively. The statistical results obtained by MaOEA/D-convergence and MaOEA/D-diversity on DTLZ and WFG are also given in Tables 3 and 4, respectively. From Table 3, we can see that MaOEA/D-OPI can obtain 16/42 and 20/42 better results than MaOEA/D-convergence and MaOEA/D-diversity, respectively. However, we can find that the MaOEA/D-diversity method can obtain better results on DTLZ2, DTLZ4 and DTLZ7. It can be understood that the DTLZ2 and DTLZ4 test problems are normally used to validate the capability of an approach for finding the uniform distributed solutions on the Pareto front, and the DTLZ7 problem is used for testing the ability of a method to find solutions on the disconnected PF as much as possible. Thus, A good population diversity of a method is prefered for optimization compared to a good convergence performance. From Table 4, we can see that MaOEA/D-OPI loses on none WFG problems than MaOEA/D-convergence, but is defeated by MaOEA/D-diversity. The reason is that the optimization population at each generation is close to the true Pareto front. Thus, the solutions focusing on convergence do not play an important role in speeding up the search. Hence, the MaOEA/D-diversity method can obtain better results than MaOEA/D-OPI. However, from the results of 69 test problems in total, we can find that our proposed MaOEA/D-OPI obtains 25/69 better results than MaOEA/D-diversity and loses on 19 out of 69 problems, showing the efficiency of our proposed method.

Table 3 Performance comparison with three variants on seven DTLZ benchmark problems
Table 4 Performance comparison with three variants on nine WFG benchmark problems
Table 5 Median and median absolute deviation of the IGD values obtained by PREA, MaOEAIGD, MaOEAIT, NMPSO, GrEA, KnEA, Two_Arch2 and MaOEA/D-OPI
Fig. 2
figure 2

Parallel coordinates of the final solutions obtained by seven compared algorithms for the 20-objective DTLZ3 instance. a PREA. (b) MaOEAIGD. c MaOEAIT. d NMPSO. e GrEA. f KnEA. g Two_Arch2. h MaOEA/D-OPI

Comparison with seven state-of-the-art algorithms for many-objective optimization problems

To evaluate the efficiency of MaOEA/D-OPI, seven state-of-the-art MaOEAs are adopted for performance comparison, including PREA [41], MaOEAIGD [21], MaOEAIT [42], NMPSO [43], GrEA[11], KnEA [44] and Two_Arch2 [25] on DTLZ and MaF benchmark problems. In PREA [41], there are two stages to choose the next population. A promising region will be constructed by a ratio based indicator in the first stage to abandon some solutions with poor quality, then a strategy according to the parallel distance is introduced in the second stage to choose solutions for surviving to the next generation. The MaOEAIGD [45] utilized the IGD indicator in each generation to select solutions with good convergence and diversity. Furthermore, an efficient decomposition-based nadir point estimation method for constructing the utopian Pareto front was proposed, which is regarded as the best approximated Pareto front during the iteration. In MaOEAIT [42], the convergence and diversity performances are addressed in two independent and sequential stages. More specifically, a non-dominated dynamic weight aggregation method based on the genetic algorithm is utilized to locate the Pareto-optimal solutions for MaOPs with different PF shapes, which are then employed to learn the Pareto-optimal subspace for convergence. Afterwards, the diversity of the population is addressed by a set of single-objective optimization problems with reference lines within the learned Pareto-optimal subspace. In NMPSO [43], a balanceable fitness estimation method (BFE) was proposed to overcome the limitations of both Pareto dominance and decomposition approaches. GrEA [11] proposed a grid-based approach to improve the selection pressure toward the optimal direction. KnEA [44] is an excellent algorithm for solving many-objective optimization problems in which a new strategy was proposed to select knee solutions to accelerate the convergence. In Two_Arch2 [25], solutions with good convergence are selected by Pareto dominance and saved in archive CA, and solutions with excellent diversity are selected by \(I_{\epsilon }\) [46] and saved in archive DA. The selection of offspring to CA and DA are independent of each other by different methodologies.

Table 5 presents the results of MaOEA/D-OPI and seven state-of-the-art algorithms in terms of the median and median absolute deviation on the 42 DTLZ instances. The best median result and the statistically comparable results are shown in the gray background. From Table 5, we can see that MaOEA/D-OPI obtained the best or competitive results on 24 problems out of the 42 instances than the other seven approaches. The proportions of the test instances on which MaOEA/D-OPI significantly outperforms PREA, MaOEAIGD, MaOEAIT, NMPSO, GrEA, KnEA and Two_Arch2 are 29/42, 32/42, 41/42, 28/42, 32/42, 29/42 and 29/42, respectively. However, from Table 5, MaOEA/D-OPI is not good for solving problems with degenerate or disconnected Pareto fronts, such as DTLZ5, DTLZ6 and DTLZ7. The reason is probably that the reference vectors have not been updated in the optimization, resulting in misguiding the direction of the search. From Table 5, we can further notice that MaOEA/D-OPI can obtain better results than other algorithms on problems with multimodal PF landscapes such as DTLZ1 and DTLZ3. For a more intuitive understanding of the behaviour of MaOEA/D-OPI on the multimodal problems, we graphically plot the parallel coordinates of the final solutions obtained by the compared algorithms for 20-objective DTLZ3 in Fig. 2. The horizontal and vertical axes represent each objective and the final optimal value of each individual on the corresponding objective, respectively. From Fig. 2, we can see that MaOEA/D-OPI has much better convergence and diversity performances than other seven methods for solving high-dimensional many-objective optimization problems with multimodal PF landscapes. From Fig. 2, we can see that MaOEA/D-OPI are much better at convergence and diversity performances than the other seven methods to solve high-dimensional many-objective optimization problems with multimodal PF landscapes.

Table 6 Median and median absolute deviation of the IGD values obtained by PREA MaOEAIGD, MaOEAIT, NMPSO, GrEA, KnEA, Two_Arch2 and MaOEA/D-OPI on MaF1-15
Table 7 Summary of the Results on DTLZ and MaF test problems, where seven compared algorithms is better than (\(+\)), worst than (−), and comparable (\(\approx \)) to than MaOEA/D-OPI according to the Wilcoxon rank sum test with Bonferroni correction

In order to further evaluate the performance of MaOEA/D-OPI on many-objective optimization problems, a number of experiments are conducted on the MaF test suit with 10, 15 and 20 objectives. Table 6 shows the statistical results on MaF test problems. From Table 6, the proportion of problems on which MaOEA/D-OPI performs better than PREA, MaOEAIGD, MaOEAIT, NMPSO, GrEA, KnEA and Two_Arch2 are 18/45, 32/45, 39/45, 20/45, 23/45, 22/45 and 19/45. The method loses to win than RPEA, NMPSO, and Two-Arch2 on MaF test problems. However, from Table 7, we can see that when all results obtained on DTLZ and MaF test problems are considered, our proposed MaOEA/D-OPI can achieve better results on 47, 48, and 48 out of 87 test problems than PREA, NMPSO, and Two_Arch2 respectively. The overall results show that MOEA/D-OPI has better performance than RPEA, NMPSO and Two-Arch2 for solving many-objective problems. From Table 7, we can further find that our proposed MaOEA/D-OPI method achieves better results on 64, 80, 55, and 51 out of 87 test problems than MaOEAIGD, MaOEAIT, GrEA, and KnEA, respectively. Therefore, the overall performance of MOEA/D-OPI is efficient for many-objective problems.

Computational complexity analysis

This subsection gives an upper bound of the computation complexity within one generation of MOEA/D-OPI. There are four main parts in one generation:

  1. 1.

    The angle between each solution and reference vector will be calculated before the solution is associated with its closest reference vector. Thus, its computational complexity is \(O(N^2)\), where N is the size of the population. On the other hand, the computational complexity of convergence evaluation is \(O(N^2)\). Thus, The computational complexity to calculate two performance indicators is \(O(N^2)\).

  2. 2.

    N solutions will be kept to the mating pool, each of which is selected from two solutions according to the closeness to a reference vector. Thus, the computational complexity will be O(N) for mating selection.

  3. 3.

    In the environmental selection, the population is selected according to sort the diversity or convergence. Thus, the time complexity of the environmental selection is \(O(N^{2})\).

  4. 4.

    The computational complexity of archive updating costs \(O(m \times |A|)\) for each solution of the population because the non-dominated sorting is applied. m is the number of objectives, and |A| represents the number of solutions in the archive. Thus, the total computational complexity to update the archive is \(O(m \times N \times |A|)\) for all solutions in the population to update the archive.

As \(m<< N\), the overall computational complexity of MaOEA/D-OPI for one generation is \(O(N \times |A|)\) if \(|A| > N\). Otherwise, the computational complexity of MaOEA/D-OPI will be \(O(N^{2})\).

Conclusion

Optional performance indicators were proposed to be utilized in this paper to balance the convergence and diversity of a population. A many-objective optimization problem is converted to a number of single problems based on reference vectors. The criterion for selecting a solution to survive to the next generation along each reference vector is randomly set to convergence or diversity. The experimental results on DTLZ, WFG and MaF test problems showed good performance of our proposed method for many-objective optimization problems. Nevertheless, when MaOEA/D-OPI is used for handling problems with irregular Pareto fronts, the performance of MaOEA/D-OPI is poor. The main reason is that the reference vectors are fixed in the optimization, which is not suitable for problems with irregular fronts. Thus, in future work, an effective strategy will be studied for enhancing the performance of MaOEA/D-OPI for solving MaOPs with irregular Pareto fronts.