Introduction

Multi-objective optimization problems (MOPs) that have multiple conflicting objectives are widely used in real-world applications in the areas of production management, economic planning, engineering design, and system control [1,2,3]. To solve these MOPs, researchers have proposed a number of multi-objective swarm intelligence optimizations and multi-objective evolutionary algorithms (MOEAs). As the complexity of an MOP increased, the concept of many-objective optimization problems (MaOPs), which have more than three objectives, is put forward.

Over the past few years, research on MOEAs for solving MaOPs has made significant advances. Some of this research is based on the Pareto-dominance relationship, such as the NSGA-III [4] and SPEA2 [5], which have good performance in low-dimensional search spaces. Some of this research is based on performance indicators, such as HypE [6], assigning corresponding fitness value to each solution according to the hypervolume indicator value. Some algorithms are based on a decomposition approach, such as MOEA/D [7] and MOEA/DD [8], decomposing the multi-objective optimization problems into multiple optimization subproblems and solving them separately. Others are based on a preference approach, such as in [9, 10]. Although MOEAs have been widely studied, their insufficient performance in solving MaOPs has prompted researchers to devote their attention to the field of swarm intelligence algorithms [11].

Swarm intelligence algorithms are optimization algorithms that simulate the intelligent behavior of certain insects and animals in nature, and that display superior performance in solving complex problems. Swarm intelligence optimization algorithms include the classical ant colony algorithm (ACO) [12, 13], particle swarm algorithm (PSO) [14, 15], cuckoo search algorithm (CS) [16], bat algorithm (BA) [17, 18], and the pigeon-inspired algorithm (PIO) [19], the last of which is studied in this paper. In addition, several hybrid bio-inspired optimization approaches have been proposed, such as in the literature [20,21,22].

The PIO algorithm, which simulates the spontaneous homing behavior of pigeons, was proposed by Duan et al. [23] in 2014, to solve complex problems, such as the air robot path planning problem.

The PIO algorithm has been the subject of much research due to its advantages, such as its simple principle, its robustness and its requiring few parameters to be adjusted. Researchers have turned to the multi-objective pigeon-inspired optimization (MPIO) algorithm to improve the algorithm’s performance in solving complex problems and to apply it in a variety of fields [24,25,26]. However, with the advent of MaOPs, the performance of MPIO algorithms was deemed insufficient, leading to the proposal of the many-objective pigeon-inspired algorithm (MAPIO), taking into consideration four, six, eight, and ten objectives [27].

Although some measures have been taken to improve the performance of the Pareto-dominance mechanism, the selection of individuals and the convergence are not guaranteed, and the limitations of non-dominance solutions produced by MAPIO become more prominent in MaOPs. Therefore, this paper proposes the knee point-driven many-objective pigeon-inspired optimization algorithm (KnMAPIO). The knee point is one of the Pareto optima solutions, where a slight improvement of one objective will lead to serious degradation of at least one other objective [28] in multi-objective optimization problem. It provides a set of solutions closer to the true Pareto front for individual selection when problem-specific preferences are unavailable [29].

The two main contributions of this article are as follows:

(1) A novel many-objective pigeon-inspired optimization algorithm based on knee point is proposed. An environmental selection strategy based on knee-oriented dominance is proposed for the individual selection, and the selection pressure and the diversity of the population are improved through full use of extreme points and boundary points. The proposed algorithm provides a new scheme for the swarm intelligence algorithm to solve high-dimensional problems.

(2) Also a new velocity updating equation has been proposed with Gaussian distribution [30, 31], Cauchy distribution [32], and Levy distribution [33] to reduce the possibility of falling into local optimality. With the change in the iteration stage, the distribution strategies used were adjusted to improve the search performance of the algorithm.

The rest of this paper is organized as follows. In the next section, the related work of the pigeon-inspired optimization algorithm and knee-oriented algorithms is described. The details of the proposed KnMAPIO algorithm are given in The proposed algorithm for many-objective optimization”. In the next section, the performance of the proposed KnMAPIO algorithm and four other algorithms is tested on the knee-oriented benchmark PMOPs, and the comparison results of the eight algorithms on the WFG and DTLZ test functions are discussed. The conclusions and directions for future work are summarized in the last section.

Related work

The PIO algorithm [23] simulates the behavior of pigeons' spontaneous homing. It is composed of two independent loops. In the first loop, the map and compass operator are used to search the target space globally, and in the secend loop, the landmark operator is used to search the solution space locally. Many scholars have modified the PIO and applied it to different fields [34, 35]. Li et al. [36] proposed a novel pigeon-inspired optimization algorithm with edge potential function and simulated annealing for the task of detecting targets in unmanned aerial vehicles (UAVs). Duan et al. [37] proposed a collaborative control method with a predation and escape pigeon-inspired algorithm for UAVs, which uses the inner and outer ring controller to solve the problem of tight formation cooperative control. Hai et al. [38] introduced a double strategic evolutionary game into PIO to improve the coordination and the search efficiency.

As the scale of the problem increases, the MPIO is proposed accordingly. Qiu et al. [39] proposed a variant of PIO, called multi-objective PIO, which is used in the parameter design of brushless direct current motors, and it uses the Pareto sorting scheme [40] and consolidation operator to enhance the selection pressure of the individuals. Liang et al. [41] proposed an MPIO algorithm with self-organizing multimodal properties and improved the space division of the solution set by the special crowding distance, which makes solving multimodal problems more efficiently. Duan et al. [42] proposed a novel MPIO with a limit cycle-based mutant mechanism to produce new solutions and search directions, and it has good performance in terms of the diversity and accuracy of solutions. Shang et al. [43] proposed a multi-objective pigeon-inspired optimization algorithm for community detection with objective functions of negative ratio association and ratio cut and adjusted the representation and update of pigeons to an adaptive process by introducing genetic operators.

With the emergence of MaOPs, the existing MPIO algorithm proved unable to provide sufficient selection pressure and solving accuracy. Cui et al. [27] proposed a pigeon-inspired optimization algorithm for many-objective optimization problems (MAPIO), which uses the BFE approach [15] and the external archive approach [15] to improve the selection capacity of the individuals.

However, the performance of MAPIO proved insufficient when user or problem-specific preferences were not available. To solve this problem, this paper introduces a knee point mechanism to supply Pareto dominance, and proposes a novel many-objective pigeon-inspired optimization algorithm based on knee point. In many-objective optimization problems, the knee points are a subset of the Pareto optimal solutions, in which the improvement of one objective will result in serious degradation of at least one other objective [44]. Therefore, the solutions in the knee area of the Pareto front are selected preferentially. Zhang et al. [29] proposed an adaptive strategy to identify knee points in a small neighborhood for solving MaOPs. Their research proved that a large hypervolume will be obtained if the knee solutions are selected preferentially in many-objective optimization problems. This means that the convergence and the diversity of the algorithm are guaranteed by locating the local knee point of the non-dominant front. Yue et al. [45] proposed a knee point-driven multi-objective particle swarm optimization algorithm. Their research indicated that the knee point mechanism can select the local and global optimal particles effectively, and the proposed algorithm was used to solve sparse reconstruction in compressed sensing. Zou et al. [46] proposed a new prediction strategy with center point and knee points to solve dynamic MOPs, and the location and distribution of the Pareto front after environmental change can be predicted accurately by introducing the knee set into the predicted population. Yu et al. [47] put forward a priori knee identification multi-objective evolutionary algorithm with α-dominance, which can speed up the convergence rate and reduce misleading search processes by eliminating the dominance-resistant solutions (DRSs) in the search for knee solutions. Yu et al. [48] proposed a new MOEA for locating the knee point area by using two local dominance relationships, α-dominance and knee-oriented dominance. The α-dominance guided the search of different potential knee regions and eliminated DRSs. The knee-oriented dominance provided the precise identification of the concave knee region and identified as many knee points as possible. In this paper, according to the knee-oriented dominance strategy, the final solution set is obtained by selecting the non-dominant individuals in the critical layer.

Inspired by these studies, we introduce the environment selection based on knee-oriented dominance into the many-objective pigeon-inspired algorithm and modify the individual velocity update equation to improve the convergence of the proposed algorithm.

The proposed algorithm for many-objective optimization

In this section, the proposed KnMAPIO algorithm is described in detail. First, the general framework of KnMAPIO is presented. Next, the knee point-driven environment selection, the novel velocity update equation, and the archive update are introduced. Finally, the computational complexity of the proposed algorithm is discussed.

The general framework of the proposed algorithm

The general framework of the proposed KnMAPIO is presented in Algorithm 1. First, the initialization process of the population is performed, where population \(P\) is randomly initialized, and the external archive set \(A\) is set to null. For each individual \(p_{i}\) in the population \(P\), the position \(X_{i}\) is randomly initialized, the velocity \(V_{i}\) is set to 0, and the fitness value of each individual \(p_{i}\) is calculated. At the same time, the local center point \(P_{center}\) of the population is calculated. Then, as shown in lines 8–9, the non-dominated individuals in population \(P\) are placed as the elite individuals in the external archive set \(A\), and the corresponding fitness values are calculated. Next, the main evolutionary process is carried out. The position \(X_{i}\) and the velocity \(V_{i}\) of individual \(p_{i}\) are updated by formula (3) and formula (4) in “Novel velocity update equation”. In so doing, a new population is obtained, and the fitness values of all individuals in the new population are calculated. Next, the local best individual \(pbest_{i}\) and local center points \(P_{center}\) of the population are obtained through the individual dominance relationship and the corresponding formula. In line 20, the knee-oriented dominance environment selection strategy is implemented for the parent population \(\left( {P,P_{center} ,P_{best} ,Archive,R} \right)\) to generate good performance solutions. Next, a new swarm \(S\) is obtained by executing two evolutionary strategies—simulated binary crossover (SBX) and polynomial-based mutation (PM) on external archive \(A\). The elite individuals are retained in the archive \(A\) again through the update mechanism in “Archive update”. This entire evolutionary process is repeated until the maximum iteration is reached. Finally, the final population \(P\) and the external archive \(A\) are obtained.

The knee point-driven environmental selection

In this section, the proposed knee point-driven environmental selection strategy is introduced. More specifically, this solution adopts the knee-oriented dominance strategy in the critical layer of Pareto dominance. Figure 1 depicts the process of the environmental selection with knee-oriented dominance. The entire solution space is divided into two separate subspaces, \(S_{1}\) and \(S_{2}\), and the reference vectors in the subspaces are \(R_{1}\) and \(R_{2}\), respectively. \(L_{1}\), \(L_{2}\), and \(L_{3}\) are the stratifications of individuals in the population after the Pareto dominance sorting. The aim of this strategy is to select the individuals in the critical layer \(L_{3}\) into offspring. Assume that \(A\), \(B\), \(C\), and \(D\) are the knee points on layer \(L_{3}\) (for the definition of knee point, refer to “Knee point”), knee point \(A\) dominates knee point \(B\) in subspace \(S_{1}\), while knee point \(C\) and knee point \(D\) do not dominate each other in subspace \(S_{2}\) according to the knee-oriented dominance described in “Knee-oriented dominance relationship”. Thus, knee points \(A\), \(C\), and \(D\) as non-dominant individuals are selected into the offspring. A new population is generated by the entire above-mentioned evolutionary process. The strategy also avoids premature convergence and improves the insufficient selection pressure caused by Pareto dominance strategy. Algorithms 2 and 3 in Sect. 3.2.3 introduce the procedure of environment selection in detail.

figure a

Knee point

The knee point is one point in the convex hull of individual minima far from the hyperplane [48]. Figure 2 illustrates the knee point in the Pareto front. \(A\) and \(C\) are the extreme points, and they construct the hyperplane \(S\). \(d\) is the maximum distance from point \(K\) (the knee points) on the Pareto front to the hyperplane \(S\).

Knee-oriented dominance relationship

Pareto dominance: For any two different solutions \(x,y \in S\), it holds that \(x\) dominates \(y\), denoted by \(x \prec y\), if and only if

$$ \begin{aligned}&\forall i \in \left\{ {1,2, \ldots ,m} \right\},f_{i} \left( x \right) \le f_{i} \left( y \right) \wedge \exists \\ &\quad j \in \left\{ {1,2, \ldots ,m} \right\}f_{j} \left( x \right) < f_{j} \left( y \right).\end{aligned} $$
(1)
Fig. 1
figure 1

Environment selection proposed in KnMAPIO

Fig. 2
figure 2

Example of the knee point

With the increase in the number of objectives, the Pareto dominance mechanism will gradually lose its effect and all the solutions will become non-dominant to each other. At this time, the selection pressure of the algorithm will drop sharply. The knee points in knee-oriented dominance are considered as the points with better properties. The knee points and their surrounding points were taken as the first selection criterion for individuals, and the distance between the point and the hyperplane formed by the extreme point was used to measure the performance of individuals, which was converted into angle information. The appropriate dominance relationship can improve the selection pressure in the potential knee point region. As a result, the knee-oriented dominance relationship proposed in [48] is used in the environment selection of KnMAPIO. The dominance relationship is defined in formula 1.

Suppose there are two solutions \(M\) and \(N\) in the region of the convex hull of individual minima, and \(M\) dominates \(N\) when the following conditions are met:

$$ \left\{ \begin{aligned} & \mu \left( {M,N} \right) = \left\langle {\overrightarrow {{N_{{id}} M}} ,\overrightarrow {{MN}} } \right\rangle \qquad \\ & \quad - \tau \left( {\mathop {\max \left\{ {\delta _{i} \left( M \right)} \right\} + }\limits_{{i = 1.....m}} \mathop {\min \left\{ {\delta _{i} \left( M \right)} \right\}}\limits_{{i = 1.....m}} } \right) \\ & \delta _{i} \left( M \right) = \arctan \left( {\frac{{\sqrt {\sum\nolimits_{{j = 1,j \ne i}}^{m} {\left( {f_{j} \left( M \right) - f_{j} \left( {N_{{id}} } \right)} \right)^{2} } } }}{{\left| {f_{i} \left( M \right) - \max f_{i} \left( E \right) - \varepsilon } \right|}}} \right), \\ \end{aligned} \right. $$
(2)

where \(\mu \left( {M,N} \right) < 0\) means that solution \(M\) knee-oriented dominates solution \(N\). \(\left\langle {\overrightarrow {{N_{id} M}} ,\overrightarrow {MN} } \right\rangle\) represents the acute angle between the two vectors \(\overrightarrow {{N_{id} M}}\) and \(\overrightarrow {MN}\). \(\tau\) controls the size of the knee region, and \(\tau \in \left[ {1/2,1} \right]^{2}\). \(\mathop {\max \left\{ {\delta_{i} \left( M \right)} \right\}{ + }}\nolimits_{i = 1.....m} \mathop {\min \left\{ {\delta_{i} \left( M \right)} \right\}}\nolimits_{i = 1.....m}\) means that the region size is dominated by solution \(M\) with the help of extreme points. \(\delta_{i} \left( M \right)\) is an acute angle determined by the \(ith\) objective value of solution \(M\). \(f_{i} \left( M \right)\) is the \(ith\) objective fitness value of individual \(M\). \(\varepsilon\) is a positive constant to guarantee that the denominator is not 0. \(N_{id}\) is the ideal point and can be calculated as follows:

$$ f_{j} \left( {N_{id} } \right) = \min f_{j} \left( E \right) - \varepsilon , $$
(3)

where \(E = \left\{ {E_{i} \left| {i = 1,2,...m} \right.} \right\}\) is the set of extreme points, and \(\varepsilon\) is a positive constant.

An example of the knee-oriented dominance relationship can be seen in Fig. 3. \(M\) and \(N\) are two individuals, \(N_{id}\) is the ideal point, \(\phi\) is the acute angle between the two vectors \(\overrightarrow {{N_{id} M}}\) and \(\overrightarrow {MN}\), and \(\delta_{2} = \mathop {\max \left\{ {\delta_{i} \left( M \right)} \right\}}\nolimits_{i = 1.....m}\) and \(\delta_{3} = \mathop {\min \left\{ {\delta_{i} \left( M \right)} \right\}}\nolimits_{i = 1.....m}\) represent that the solution \(M\) knee-oriented dominates solution \(N\) when \(\phi\) is smaller than \(\tau \left( {\delta_{2} + \delta_{3} } \right)\). In addition, the dominance regions of individuals in a two-dimensional space can be seen in Fig. 4, where individual \(M\) dominates individual \(N\), and shadows and dashed lines represent the corresponding dominance region of different individuals.

Fig. 3
figure 3

Example of the knee-oriented dominance relationship

Processs of environmental selection

The process of knee point-driven environmental selection is shown in Algorithm 2. First, the offspring \(P\) is initialized as an empty set, and the parent population \(Q\) is sorted as \(L_{1} ,L_{2} , \cdots L_{l}\) according to the Pareto domination relationship. The individuals in all layers \(L_{1} ,L_{2} , \cdots L_{{l{ - }1}}\) before the critical layer \(L_{l}\) are placed into offspring \(P\) according to the layer number in ascending order. Next, only a portion of the individuals in the critical layer \(L_{l}\) can be selected into the offspring \(P\) according to the knee-oriented dominance selection strategy (refer to Algorithm 3), and the number of individuals in this part is \(n - \left| P \right|\), where \(n\) is the size of the output offspring.

Fig. 4
figure 4

Dominance region of individuals M and N

figure b

Algorithm 3 introduces the knee-oriented dominance selection strategy used in Algorithm 2 in detail. First, the initialization procedure is performed. Next, the solution space is divided into different subspaces \(\ell_{i}\), and each subspace is initialized to empty. Next, the solutions in the critical layer are associated with the reference vector \(R_{r}\) in the nearest subspace according to the grouping function, where \(C_{1} ,C_{2} \cdots C_{k}\) stand for \(C_{1}\) subspace, \(C_{2}\) subspace, and \(C_{k}\) subspace. In lines 5–8, the individuals in subspace \(C_{i}\) are ordered according to the knee-oriented dominance relationship, and the front number is assigned to each individual. In line 7, the solutions in different subpopulations are recombined in a set \(u\). Next, several sub-layers are grouped according to the knee-oriented dominance front number. The selection procedure in lines 10–14 is similar to the non-dominated sorting process in NSGA-II [40]. All solutions in the critical layer are sorted in ascending order according to the front number, and the solutions with the same front number are sorted in descending order by the crowding distance.

figure c

Novel velocity update equation

In this section, a novel velocity update equation is introduced, and Gaussian distribution [30, 31], Cauchy distribution [32], and Levy distribution [33] are used to improve search ability. Specifically, the Cauchy distribution tends to global search in the early iteration. The Levy distribution is used to update the individuals of the population in the late iteration, so that a few individuals can carry out local search and others can carry out global search to reduce the possibility of falling into local optimum to a certain extent. The introduction of Gaussian distribution improves the centralized search ability, which improves the convergence speed of the algorithm. These distributions are used at different stages to affect the updating of velocity, in order to provide new search directions and reduce the possibility of the algorithm falling into the local optimal.

The proposed update equation is as follows:

$$ \begin{aligned} V_{i} \left( {t_{{now}} } \right) & = e^{{ - Rt_{{now}} }} \cdot V_{i} \left( {t_{{now}} - 1} \right) \\ & \quad + r_{1} \cdot Cauchy \cdot tr \cdot \left( {1 - \log _{T}^{{t_{{now}} }} } \right) \\ & \quad \cdot \left( {X_{{glo}} - X_{i} \left( {t_{{now}} } \right)} \right) \\ & \quad + r_{3} \cdot Levy \cdot tr \cdot \log _{T}^{{t_{{now}} }} \cdot \left( {X_{{cen}} - X_{i} \left( {t_{{now}} } \right)} \right) \\ & \quad + r_{5} \cdot G \cdot \left( {X_{{glo}} - X_{{cen}} } \right), \\ \end{aligned} $$
(4)
$$ X_{i} \left( {t_{now} } \right) = X_{i} \left( {t_{now} - 1} \right) + V_{i} \left( {t_{now} } \right), $$
(5)

where \(t_{now}\) represents the current iteration number, \(T\) represents the maximum iteration number, \(R\) denotes the map and compass operator, and \(tr\) is the migration factor to ensure a smooth transition between map and compass operator and landmark operator. \(Cauchy\), \(Levy\), and \(G\) represent the Cauchy distribution, Levy distribution, and Gaussian distribution, respectively.

The search combined with the Levy distribution has a relatively high probability of large stride in the process of random walk, thereby ensuring that the walk is not limited to a small local area, which can increase the diversity of the population and expand the search range. This provides a high probability of other search directions in the later iterations of the population. Gaussian distribution can search for large probability and small range variation in individual local area. Therefore, Gaussian distribution is introduced into the search process to improve the local search ability of the algorithm, and the velocity updating strategy of this part also affects the entire algorithm iteration process. Cauchy distribution has thicker and longer tails, the offspring generated in the algorithm will be more dispersed, and the diversity of offspring population will be better, which is more suitable for global optimization. Therefore, we introduce Cauchy distribution into the search process to improve the global search ability of the algorithm in the early iteration process.

\(X_{glo}\) is the position information about the global best in all individuals. The meaning of \(X_{cen}\) is the center position information of some individuals in the current iteration, and it can be calculated according to the formula (6).

$$ X_{cen} = \frac{{\sum\nolimits_{j = 1}^{{n_{1}^{x} }} {S_{1j}^{X} } }}{{n_{1}^{x} }}S_{1}^{X} n_{1}^{X} , $$
(6)

where \(\sum\nolimits_{j = 1}^{{n_{1}^{x} }} {S_{1j}^{X} }\) represents the sum of all solutions of individuals in the non-dominated set \(S_{1}^{X}\), and \(n_{1}^{X}\) is the number of the solutions in the set \(S_{1}^{X}\).

The \(r_{1}\), \(r_{3}\) and \(r_{5}\) represent three learning factors and are defined as follows:

$$ r_{i} = \left\{ {\begin{array}{*{20}c} {0,0 < rand() \le \frac{1}{M}} \\ {1,\frac{1}{M} < rand() \le 1} \\ \end{array} } \right., $$
(7)

where \(rand()\) is a random number between [0,1], and \(M\) presents the number of objectives. The part \(r_{i}\) located in formula (4) will not influence the \(V_{i} \left( {t_{now} } \right)\) if \(r_{i}\) is 0. With the change of parameter \(M\), \(V_{i} \left( {t_{now} } \right)\) will update dynamically when \(r_{i}\) is 1.

Archive update

Simulated binary crossover (SBX) [49] and polynomial-based mutation (PM) [49] as another search pattern are used to provide an additional search direction in the process of the archive update, and then, the new solutions \(S\) are produced. In KnMAPIO, the archive update strategy [15, 50] is also adopted to retain the elite solutions. Meanwhile, elite individuals are selected and eventually retained in the external archive by using the BFE method [15], which not only ensures the convergence of the population but also guarantees the diversity. These two methods can guide the solutions approximate to the true PF. By comparing the Pareto-dominance relationships between \(S\) and the external archive \(A\), archive \(A\) is updated until it reaches the terminal condition.

Computational complexity

The computational complexity of the proposed algorithm is mainly derived from the knee point-driven environment selection and the archive update. The knee point-driven environment selection includes non-dominance sorting and the knee-oriented dominance sorting. The computational complexity of the non-dominance sorting is \(O\left( {N^{2} M} \right)\) when the population size is \(N\) and the objective vector dimension is \(M\). The knee-oriented dominance sorting is applied to the critical layer of non-dominance sorting. For any two solutions, the computational complexity is \(O\left( M \right)\) when the angle of two solutions is calculated. The worst case is that all the solutions are at the same layer, and the computational complexity of knee-oriented dominance sorting is \(O\left( {N^{2} M} \right)\) at this time. The computational complexity of the crowding distance is \(O\left( {M \times N\log N} \right)\), which is the same as that of NSGA-II [40].

Table 1 Parameter setting for the PMOPs test functions
Table 2 The KGD value of five algorithms for different objectives in the PMOP test problems
Table 3 The KIGD value of five algorithms for different objectives in the PMOP test problems

The computational complexity of the archive update is mainly obtained by comparing the individuals in the population with the elite individuals in the archive set. Since the archive size is usually selected in proportion to the size of the population, the complexity is \(O\left( {N^{2} M} \right)\) in the worst case. Overall, the computational complexity of the proposed KnMAPIO is \(O\left( {N^{2} M} \right)\).

Experimental results

The proposed algorithm KnMAPIO and other four algorithms were tested on the knee−oriented benchmark PMOPs to verify the performance of detecting the knee points and the knee regions. KnMAPIO was also compared with five state−of−the−art many−objective algorithms and three algorithms for the last 3 years on standard benchmarks DTLZ and WFG to measure the degree to wihch the solutions obtained by these algorithms dominate the true Pareto Front. The experimental results were analyzed numerically, a process that has been applied in many other fields [51,52,53].

All the comparative experiments in this paper were run on PlatEMO proposed by Tian et al. [54] in MATLAB 2016b, using the Intel Xeon Gold 5222CPU @ 3.80 GHz and 64 GB RAM. The operating system used in this paper is 64−bit Microsoft 10 and based on the × 64 processor.

Table 4 Parameter setting for the DTLZ and WFG test functions

Verification of the performance of KnMAPIO to identify knee points

Benchmark problems and experimental settings

PMOPs are a classic test suite with complex knee regions for assessing the accurate identification capability of the knee point [55]. The PMOPs PMOP1−PMOP3, PMOP5−PMOP6, PMOP8−PMOP9, and PMOP11−PMOP14 were selected as the test examples to measure the performance of the algorithm. All the parameters in these test problems are listed in Table 1\(M\)  denotes the number of objectives, which is set as 3, 5, 8, and 10 in this experiment, and the corresponding population number is 105, 126, 156, and 275, respectively. \(N\) represents the number of decision variables. \(k\left( x \right)\) is the function of knee regions, and  \(\left( {A,B,s,p,l} \right)\) is a set of parameters controlling the shape and the number of knee regions. The algorithms were run 30 times independently for each test function.

Table 5 The coverage value of six algorithms for different objectives in the DTLZ test problems
Table 6 The coverage value of four algorithms for the last 3 years on the DTLZ test problems

In this experiment, five algorithms including KnEA [29], LA-MOEA [47], LBD-MOEA [48], MAPIO [27], and the proposed KnMAPIO were compared to verify the performance in identifying knee points. All parameters were set as recommended in the original papers. Among them, the rate of knee points was 0.5 in KnEA [29]. In LA-MOEA, the parameter about the localized \(\alpha\)-dominance was set to \(\alpha = 0.75\) [34]. In LBD-MOEA, the \(\left( {H_{1} ,H_{2} } \right)\) was set to (1, 5), (1, 3), (1, 2), and (1, 3) for the generation of the reference vector with 3, 5, 8, and 10 objectives, and for other parameters, refer to [48]. For KnMAPIO, the setting of parameter \(\tau\) in knee-oriented dominance was similar to that of LBD-MOEA. For a fair comparison, we set the same parameters as MAPIO [23], including the transition factor \(tr\) as 1, and the map and compass factor \(R\) as 0.3.

Performance measures

To measure the performance of the proposed algorithm KnMAPIO and other algorithms on benchmark PMOPs, two indicators were used: knee-driven generational distance (KGD) [55] and knee- driven inverted generational distance (KIGD) [55].

(1) KGD

KGD indicates the proximity of the solutions to the reference points in the Pareto Front knee region, which can evaluate the convergence performance of the proposed algorithm. The smaller the KGD value, the better is the convergence performance. KGD can be calculated as follows:

$$ KGD = \frac{1}{{\left| {\mathbb{Q}} \right|}}\sum\limits_{i = 1}^{{\left| {\mathbb{Q}} \right|}} {d\left( {v_{i} ,{\mathbb{R}}} \right)} , $$
(8)

where \({\mathbb{Q}}\) is the approximation solution set obtained by the algorithm, \({\mathbb{R}}\) represents a reference points set that is distributed in the knee region uniformly, and \(d\left( {v_{i} ,{\mathbb{R}}} \right)\) denotes the Euclidean distance between point \(v_{i}\) belongs to approximation solution set \({\mathbb{Q}}\) and the closest reference point in set \({\mathbb{R}}\).

(2) KIGD

KIGD indicates the extent to which the solution obtained by the algorithm covers the knee region. The smaller the KIGD value, the better is the diversity performance. This means that the solutions obtained by the algorithm can cover the knee region extensively. KIGD can be calculated by the formula (9):

$$ KIGD = \frac{1}{{\left| {\mathbb{R}} \right|}}\sum\limits_{i = 1}^{{\left| {\mathbb{R}} \right|}} {d\left( {v_{i} ,{\mathbb{Q}}} \right)} , $$
(9)

where \({\mathbb{Q}}\) is the approximation solution set obtained by the algorithm. \({\mathbb{R}}\) represents a reference points set that is uniformly distributed in the knee region, and \(d\left( {v_{i} ,{\mathbb{Q}}} \right)\) denotes the Euclidean distance between point \(v_{i}\) belongs to approximation solution set \({\mathbb{R}}\) and the closest reference point in set \({\mathbb{Q}}\).

Experimental results and analysis

In this section, Tables 2 and 3 show the comparison results on KnEA [29], LA-MOEA [47], LBD-MOEA [48], MAPIO [27], and the proposed KnMAPIO. Table 2 presents the KGD values of five algorithms on the PMOP test suite, allowing for an evaluation of the convergence performance of the algorithms. Table 3 presents the KIGD values of five algorithms, allowing for an evaluation of the diversity performance of algorithms. In these tables, the highlight values represent the best results, and ‘ + ,’ ‘−’, and ‘ = ’ denote the results from other algorithms that are respectively higher than, lower than, or equal to the results from the proposed KnMAPIO.

From Table 2, the number of best results—0, 9, 7, 8, 20—is produced by KnEA, LA_MOEA, LBD_MOEA, MAPIO, and KnMAPIO, respectively. As observed from the last row of Table 2, compared with KnEA, KnMAPIO has a significant advantage on the 37 test functions. The performance of LA_MOEA is similar to that of LBD_MOEA, in which 15 items are better than the proposed KnMAPIO. However, LA_MOEA has 21 items that are worse than KnMAPIO, and LBD_MOEA has 23 items that are worse than the proposed algorithm. In the comparison results with MAPIO, seven items of MAPIO are better than KnMAPIO, while 18 items are worse than the proposed algorithm. These comparisons show that our algorithm has certain advantage in KGD.

From the different PMOP problems presented in Table 2, it can be seen that the proposed algorithm has good performance on PMOP 2, 3, 5, 6, 9, 11, 13, 14 test functions, indicating the superior performance of the proposed algorithm in solving these problems, such as the concave basic shape, multimodal, and non-separable problems. For PMOP3, our algorithm achieves the good preference in all four objectives. For PMOP1, although the proposed algorithm is not optimal in each objective, it also preforms better than some algorithms in some objectives. Further research will be carried out in solving linear problems, fundamental problems and shape problems in the future. In the comparison results with MAPIO, the proposed algorithm has a slight advantage. This is because the proposed algorithm not only uses BFE strategy but also uses knee point-driven environment selection to increase the probability of knee region identification.

As shown in Table 3, KnMAPIO obtained 20 of the best results, LBD_MOEA obtained 10 of the KnEA obtained eight, LA_MOEA obtained six, and MAPIO failed to obtain any. As observed from the last row of Table 3, compared with KnEA, KnMAPIO has a significant advantage on 30 test functions. The performance of LA_MOEA is similar to that of LBD_MOEA, in which 26 items are worse than the proposed KnMAPIO, while the performance of MAPIO has 31 items worse than KnMAPIO. These comparison results show that the proposed algorithm has significant advantages in performance.

From different PMOP problems in Table 3, the proposed algorithm has superior performance on PMOP 2, 3, 5, 8, and 11 test functions. For PMOP2 and PMOP11, our algorithm achieves the good performance in all four objectives, indicating its robust diversity in solving the problems with the characteristics of concave, basic, shape and complex PoF. However, our algorithm does not perform satisfactorily in solving the PMOP13 problem. The comparison results make it obvious that the knee-driven environment selection strategy can significantly improve the identifying effect of the proposed algorithm. Overall, Tables 2 and Table 3 show that the proposed algorithm has excellent overall performance, which is attributed to the improvement of individual selection pressure by the knee point-driven environmental selection.

Overall, Tables 2 and 3 show that the proposed algorithm has excellent overall performance, which is attributed to the improvement of individual selection pressure by the knee point-driven environmental selection.

Approximation degree of KnMAPIO to the true Pareto front

Benchmark problems and experimental settings

DTLZ [56] and WFG [57] are standard benchmarks for assessing the capability of the many-objective optimization algorithm, and these functions have obvious attributes for solving linear, multi-modal, disconnected, mixed convex/concave, and deceptive problems. DTLZ1-DTLZ7 were selected as the test examples for DTLZ problems, and WFG1–WFG9 were chosen for the WFG problems. The settings of parameters for the DTLZ and WFG are shown in Table 4. The population size in the DTLZ and WFG test suites are 120, 132, 156, and 275 with the number of objectives at four, six, eight, and ten. The algorithms were run 30 times independently for each test function.

The proposed algorithm was tested on the standard benchmarks DTLZ and WFG with five other state-of-the-art algorithms: NSGA-III [4], GrEA [58], MOEA/D [7], RVEA [59], and VaEA [60]. The proposed algorithm was also compared with three algorithms for the last 3 years on the standard benchmark DTLZ and WFG test functions to verify the performance of our proposed algorithm, including CSEA [61], hpaEA [62], and MaPSO-MC [63]. CSEA [61] was proposed in 2019 by Pan et al., who used an artificial neural network to predict the dominance relationship between candidate solutions and reference solutions instead of approximating the objective values separately. In 2020, Tian et al. proposed hpaEA [62], which was first algorithm to differentiate the non-dominated solutions by exhibiting tendencies toward the Pareto-optimal front as prominent solutions and using the hyperplane formed by their neighboring solutions. MaPSO-MC [63] was proposed by Hu et al. in 2021 to solve the hybrid recommendation model by using a generation-based fitness evaluation strategy coupled with diversity enhancement (GBFE–DE) and ISDE+to evaluate individual performance.

All parameters were set as recommended in the original papers. Simulated binary crossover [4] and polynomial mutation [4] were used in some algorithms [59,60,61] to produce offspring, the distribution indices of crossover and mutation were set to \(n_{c} = 20\), and \(n_{m} = 20\). The probabilities of crossover and mutation were set to \(p_{c} = 1.0\) and \(p_{m} = 1/D\), where \(D\) represents the number of decision variables. For GrEA, the setting of the parameter \(div\) was taken from [58], which denotes the number of divisions in each dimension. The parameters about the range of neighborhood in MOEA/D were set to \(N/10\) for all test problems. Other relevant parameters were set by referring to [7]. In [59], the change rate of the penalty function \(\alpha\) was set to 2, and the frequency of reference vector adaptation was set to \(f_{r} = 0.1\). For CSEA, the setting of maximum epochs for training the FNN \(T\) was 500, the hidden neurons \(H\) was 10, and the number of parameters for reference solutions was 6. The number of prominent solutions \(K\) in hpaEA [62] was set to 6, and other parameters were set to the same as NSGA-III. In [63], the setting of learning parameters \(c_{1}\), \(c_{2}\), \(c_{3}\) belonged to \(\left[ {1.5,2.5} \right]\), and the number of parameters of generation influence \(\theta\) in MaPSO-MC was 2.

Performance measures

To measure the approximation degree of the algorithm to the true Pareto front, the coverage [7] was used in this experiment. C-metric indicates the dominant relationship between the solution set and the Pareto front, which can measure the coverage performance of a solution set. The C-metric can be calculated as follows:

$$ C\left( {Pop,PF} \right) = \frac{{\left| {\left\{ {u \in PF\left| {\exists v \in Pop:v\,\,{\text{dominance}}} \right.u} \right\}} \right|}}{{\left| {PF} \right|}}, $$
(10)

where the \(PF\) is the Pareto front, and \(Pop\) represents the objective values of all the individuals in the population. The numerator is the number of solutions in \(PF\) that are dominated by at least one solution in \(Pop\), and the denominator represents the total number of solutions in \(PF\).

Experimental results and analysis

In this section, Tables 5 and 6 show the comparison results on GrEA [58], MOEA/D [7], NSGA-III [4], RVEA [59], VaEA [60], CSEA [61], hpaEA [62], MaPSO-MC [63], and the proposed KnMAPIO, and the performance of the coverage is demonstrated in the DTLZ test functions. Tables 7 and 8 show the comparison results of the algorithms on the WFG test functions, and the performance of the coverage is demonstrated in these test functions. In these tables, the highlighted numbers represent the best results, and ‘ + ,’ ‘-,’ ‘ = ’ denote the results from the other algorithms that are respectively higher than, lower than, or equal to the results from the proposed KnMAPIO.

Table 7 The coverage value of six algorithms for different objectives in the WFG test problems
Table 8 The coverage value of four algorithms for the last 3 years on the WFG test problems

As can be seen from Table 5, the number of best results—5, 5, 1, 0, 0, and 14—is produced by GrEA, MOEA/D, NSGA-III, RVEA, VaEA, and the proposed KnMAPIO, respectively. In the six, eight, and ten objectives of DTLZ4, our algorithm achieves the same result as MOEA/D. Looking at the last row in Table 5, compared with RVEA, KnMAPIO has a significant advantage on 28 test functions. NSGA-III has only 3 items better than the proposed algorithm, GrEA and VaEA have four items better, and MOEA/D has six items better than the KnMAPIO. As can be seen from Table 6, the number of best results—5, 1, 4, and 18—is produced by CSEA, hpaEA, MaPSO-MC, and the proposed KnMAPIO, respectively. Looking at the last row of Table 6, KnMAPIO has 11 items better than CSEA, while hpaEA has only 1 item better than KnMAPIO, and MaPSO-MC has 3 items better. KnMAPIO has 16 items better than MaPSO-MC and has significant advantages over the other three algorithms on DTLZ 5 and DTLZ 6.

Table 5 shows that the proposed algorithm has good performance on DTLZ 2, 5, and 6 test functions, and it again proves the superior performance of the proposed algorithm in solving concave and multimodal functions. However, the table also shows that KnMAPIO does not perform well on DTLZ7, because of the multiple modal characteristics of the test functions and the difficulty in obtaining the convergence solutions. Table 6 shows that the performance of the proposed algorithm is better than that of the other three algorithms. The coverage values of some functions are 0.00e+0 in this table, indicating that no solutions in PF are dominated by the solutions in the final population. The coverage values of VaEA on DTLZ 7 are all 1.0000e+0, which indicates that all solutions in PF are dominated by some solutions in the final population.

Table 7 reveals the comparison results of KnMAPIO with five advanced algorithms on WFG test functions, with four to ten objectives using the indicator of coverage. The number of best results—9, 4, 0, 0, 0, and 19—is produced by GrEA, MOEA/D, NSGA-III, RVEA, VaEA, and the proposed KnMAPIO, respectively.

Our algorithm achieves the same result as other algorithms in all objectives of WFG1, and the coverage values are all 1.0000e+0, which indicates that represents all solutions in PF are dominated by some solutions in the population. It also shows that these algorithms have the same performance in solving the bias functions. Compared with NSGA-III, RVEA and VaEA, KnMAPIO has an obvious advantage over the 36 test functions. Among the comparison results, GrEA has only seven items better than KnMAPIO, but 20 items worse than KnMAPIO. MOEA/D has 5 items better than KnMAPIO, but almost two thirds of the items are worse than the proposed algorithm. Table 7 shows that the proposed algorithm has good performance on WFG 4, 5, 6, and 9 test functions, especially on WFG 5. The good performance of the proposed algorithm on WFG 5 indicates the best coverage on the concave–basic–shape functions.

Table 8 reveals the comparison results of KnMAPIO with three advanced algorithms on WFG test functions with –four to ten objectives using the indicator of coverage. The number of best results—4, 1, 14, and 17—was respectively produced by CSEA, hpaEA, MaPSO-MC, and the proposed KnMAPIO. Our algorithm achieves the same result as other algorithms in all objectives of WFG1 and the coverage values are all 1.0000e+0. KnMAPIO has 19 items better than CSEA and 23 items better than hpaEa, with only three items worse. KnMAPIO has similar performance to MaPSO-MC on the WFG test problem. Although MaPSO-MC performs better on WFG 2 and WFG 4, the KnMAPIO has much better performance on WFG 3, WFG 5, and WFG 9.

The results in the tables show that the proposed algorithm has superior performance to that of the other algorithms, which can be attributed to the knee-oriented dominance and the distribution-based individual update strategy. These strategies also improve the selection pressure of the solution set toward the true PF.

In addition, the performance of using only one distribution, two distributions, and three distributions in our proposed algorithm was compared on DTLZ test functions. The experimental results are shown in Table 9. MAPIO algorithm does not uses the velocity position update formula proposed by us. MAPIOC is the algorithm only uses the Cauchy distribution. MAPIOG algorithm uses the Gaussian distribution only. MAPIOLC is the algorithm uses Levy and Cauchy distributions simultaneously, and MAPIOGC algorithm uses Gaussian and Cauchy distributions at the same time.

Table 9 The coverage value of algorithms for different objectives in the DTLZ test problems

The Coverage value of algorithms for different objectives in the DTLZ test problems has shown in Table 9. KnMAPIO has 13 best results, MAPIOC has 3 best results. MAPIOG, MAPIOLC, and MAPIOGC have only 2 best results. As can be seen from the last row, MAPIO has nine results worse than our algorithm, but zero result better than KnMAPIO. MAPIOC has only one item better than the proposed method, but five items worse than ours. MAPIOLC and MAPIOGC have neither result better than ours, but worse in three. Although the proposed algorithm is slightly inadequate in some test problems, its overall performance is better than other comparison algorithms. The performance of the algorithm using only one distribution is slightly worse than the performance of the algorithm using two distributions, and the performance of the algorithm using two distributions is worse than the performance of the algorithm combining the three distributions. So the strategy we proposed is effective.

Conclusion

This paper proposes a knee point-driven many-objective pigeon-inspired algorithm (KnMAPIO) to solve the problem of selection of individuals in the case of non-decision maker preference. The proposed algorithm uses an environmental selection strategy based on knee-oriented dominance to enhance the selection pressure of the individuals, and the diversity of the population and the identification ability of the knee point region have been promoted through the full use of extreme points, boundary points, and knee sorting mechanisms. Next, a new velocity update equation with Gaussian distribution, Cauchy distribution, and Levy distribution is proposed, and the use of these three distributions at the appropriate stage effectively improves the search performance. In most test problems, our algorithm has excellent performance in identifying knee points and their regions in comparison experiments with four other algorithms on the knee-oriented benchmark PMOPs. In addition, compared with five state-of-the-art many-objective algorithms and three algorithms for the last 3 years on the standard benchmark DTLZ and WFG suites, our algorithm shows significantly better performance.

In future work, we will apply the proposed KnMAPIO algorithm to solve practical engineering problems without decision-maker preference and provide a set of efficient solutions.