Introduction

Nature-inspired optimization methods have been used effectively to solve a wide variety of complex problems that consist of both single and multiple objective search domains. Among these methods, swarm intelligence is a promising research area. Introduced to solve single objective problems, Particle Swarm Optimization (PSO) [15] has attracted many researchers in metaheuristic optimization area, and started to gain prominence at solving multiple objective problems not more than 5 years after its introduction (see [27] for the first attempt on multi-objective optimization). This is because of the relative simplicity and the success as a single-objective optimizer, as well as high speed of convergence [4, 22]. Furthermore, due to its population based nature, it enables to obtain a set of trade-off solutions in a single run, unlike the traditional techniques which employ a series of separate runs [36].

However, there still exist three main issues to be considered in Multi-objective Particle Swarm Optimization (MOPSO): (1) archive maintenance, (2) process to update global best and individual best, and (3) solutions for local optima and premature convergence problems [11, 13].

Maintaining an external archive, which is used to keep a historical record of non-dominated solutions in accordance with a quality measure, serves the main purpose of multi-objective optimization. Computational cost and memory size considerations cause keeping the size of external archive fixed seems more efficient [13, 29]. While maintaining the external archive, to obtain a fairly distributed set of non-dominated solutions, employing a density measure in objective space is a straightforward approach. Strategies such as crowding distance [34, 37], adaptive grid [5], clustering [33], maximin fitness [21], parallel cell coordinate system [13], and hypersurface contribution [35] can be used for maintaining the archive.

Regarding the update issue, the movement of a particle in MOPSO is affected by personal and global best selection (i.e., the selection of leaders). The selection of leaders is a crucial issue [41], and this selection directly affects the convergence and diversity attitudes, and effectiveness of the algorithm [48]. In other words, the balance between exploitation and exploration capabilities depends on the leader selection process. The trade-off between exploration and exploitation is critical to the performance of an evolutionary algorithm [31, 47].

Although the fast convergence is an advantage for PSO or MOPSO, it becomes a drawback, if it is not controlled effectively. Fast convergence (premature convergence), especially for the earlier stages of the run, may cause particles to be “trapped in a similar local topology” [9]. As a consequence, it may not be possible to achieve a precise approximation to the true Pareto front [26]. Perturbation operator (mutation, disturbance) is a common practice to compensate for premature convergence, and maintaining diversity of the swarm along the optimization process [9, 13].

This paper contributes to the literature by proposing the use of “random immigrants” approach, an effective method to promote diversity for MOPSO. Random immigrant approach has been developed to address maintaining diversity for genetic algorithms and proved to be beneficial [25, 43]. It is based on a simple philosophy of replacing the worst or randomly selected particles from the swarm with randomly created particles. To the best of authors’ knowledge, ours is the first study to use random immigrants approach for MOPSO.

Literature review

This section aims to give basic concepts and definitions on multi-objective optimization, and a short survey related to the fast convergence problem mentioned above.

Multi-objective optimization

Multiple objective optimization problems deal with at least two objective functions to be optimized. These objective functions are non-commensurable and competing. It means that they may be represented in different units, and they may have same level of importance comparatively. Assuming all the objective functions to be minimized, a multi-objective optimization problem can be defined as in [36]:

$$\begin{aligned}&\text {Minimize} \ f(\mathbf {x}):=[f_{1}(\mathbf {x}),f_{2}(\mathbf {x}), \dotsc , f_{k}(\mathbf {x})] \end{aligned}$$
(1)
$$\begin{aligned}&\text {Subject to} \nonumber \\&\qquad \qquad g_{i}(\mathbf {x}) \le 0 \quad i=1,2,\dotsc ,m,\end{aligned}$$
(2)
$$\begin{aligned}&\qquad \qquad h_{j}(\mathbf {x}) = 0 \quad j=1,2,\dotsc ,p, \end{aligned}$$
(3)

where \(\mathbf {x}=[x_{1},x_{2},\dotsc ,x_{n}]^{{\text {T}}}\) represents decision variables vector, \(f_{i} : \mathbb {R} ^{n}\rightarrow \mathbb {R}, i=1,2,\dotsc , k\) gives objective functions, and \(g_{i}, h_{j}: \mathbb {R} ^{n}\rightarrow \mathbb {R}, i=1,2,\dotsc , m, j=1,2,\dotsc , p\) represents inequality and equality constraints, respectively. The desired solution is in the form of “trade-off” solutions between objective functions [34]. In other words, an improvement in one of the objective functions causes worsening for at least one of the others.

The most common two approaches to a multi-objective optimization problem are: (1) transforming the problem into a single objective one; (2) obtaining a set of trade-off solutions (preserving the problem as is) [17]. For the first case, simple additive weighting can be used, or all but one objective function can be moved to the constraint set. However, with the weighting method, not all Pareto-optimal solutions can be found for the problems that have non-convex objective or search spaces [3]. Additionally, many different weights may result the same single solution [3], and it can be very difficult to precisely and accurately select the weights [17]. When moving objective functions to the constraint set, it may be difficult to set right-hand side values for objective functions as constraints.

The main goal of a multi-objective optimization algorithm is to identify solutions in Pareto-optimal set [17, 19]. Yet, all the elements of a Pareto optimal set may not be desirable or achievable [6, 36], and the Pareto-optimal set can be infinite, while we have some space and time limitations. Therefore, it is desirable to obtain a set of solutions that represents the Pareto-optimal set as well as possible [17].

Real-world multi-objective optimization problems may be too complex to be solved by exact methods, such as linear programming and gradient search [50]. Population-based metaheuristic algorithms are accepted as effective computational solvers for multi-objective optimization problems. Due to their search capabilities through large spaces using populations, they are able to get some Pareto-optimal solutions in a single optimization run. Additionally, they are not effected by the shape of the Pareto-optimal front. Of these algorithms, MOPSO is a competitive one. Interested reader should refer [36] for a detailed explanation of a general MOPSO algorithm. Our approach omits the mutation phase of general MOPSO algorithm and adds random immigrants step after updating the leaders in an external archive.

Related work

Fieldsend and Singh [10] introduced a multi-objective algorithm based on the PSO and demonstrated the inclusion of stochastic turbulence variable. Using this new variable in MOPSO, they showed significant performance increases.

Mostaghim and Teich [28] proposed a method, called Sigma method for selecting the best local guide for each particle. They added a turbulence factor to the updated position of each particle in the swarm.

Xiao-hua et al. [42] proposed a modified PSO named Intelligent Particle Swarm Optimization (IPSO). They used a “clonal selection operator” to accelerate the approximation to optimum. One of the elements in this operator is called “clonal mutation”, which helps to produce a solution set around Pareto optimal solutions.

Sierra and Coello [37] proposed a multi-objective particle swarm optimizer, which is based on Pareto dominance and uses of a crowding factor. They used uniform and non-uniform mutation schemes. In uniform mutation, the variability range allowed for each decision variable is kept constant over iterations, whilst in non-uniform mutation, this variability decreases over time.

Raquel and Naval [34] proposed an MOPSO algorithm which is called MOPSO-CD. They used crowding distance mechanism and mutation operator to maintain diversity. They performed mutation on the entire swarm initially and then rapidly decreased its coverage over time.

Peng and Zhang [32] proposed a decomposition based MOPSO. They applied the polynomia mutation on positions after they are calculated.

Izui et al. [14] proposed a multi-objective optimization method for structural problems based on MOPSO. They applied a mutation operator. In this operator, the probability of mutation decreases as the number of iterations increases, while mutation rate is fixed.

Agrawal et al. [1] proposed an interactive particle swarm metaheuristic for multi-objective optimization. They employed a mutation operator which is defined “self-adaptive mutation”. This operator has some variation in probability according to the number of particles in the repository.

Padhye et al. [30] reviewed some proposals for guide selection in MOPSO and compared them with each other in terms of convergence, diversity, and computational times. They made a proposal named “velocity trigger” as a substitute for turbulence operator coupled with a boundary handling method. They reported that the new proposals were found to be effective for higher objective and higher parameter space problems.

Yen and Leong [44] proposed an MOPSO algorithm with dynamic multiple swarms. In the swarm growing strategy, they use uniform mutation operator with the mutation rate equal to one/number of dimensions in decision space.

Al Moubayed et al. [2] proposed an MOPSO algorithm that employs decomposition. Instead of mutation, they used an information exchange method that helps avoiding local optima without a need for applying any genetic operator.

Yen and Leong [45] proposed a constraint MOPSO which adopts a multi-objective constraint handling technique. They applied uniform and Gaussian operators. Uniform mutation encourage exploration and Gaussian mutation promotes exploitation. The frequency of applying the mutation operators depends on the feasibility ratio of the particles’ personal best.

Daneshyari and Yen [7] proposed a cultural MOPSO which adapts the parameters of the MOPSO using the knowledge stored in various parts of the belief space. They applied a time-decaying mutation operator. The number of particles that undergo mutation, the range of mutation for each mutated particle, and the dimensions selected for mutation are regulated accordingly.

Mahmoodabadi et al. [23] modified the MOPSO in two stages. The first stage involves combining PSO with convergence and divergence operators. The second stage involves new leader selection method and adaptive elimination method which aims to limit the number of non-dominated solutions in the archive. They used the divergence operator as a simple controlled mutation.

Hu and Yen [12] proposed a method for density estimation for selecting leaders and maintaining the archive in MOPSO. They used “Parallel Cell Distance” between a solution and all other solutions in an archive after the archive is mapped from Cartesian Coordinate System into Parallel Cell Coordinate System. To perturb an article, they used Gaussian Mutation.

Leung et al. [20] presented a new algorithm that extends PSO to deal with multi-objective problems. Their first contribution is that the square root distance computation among particles for local best selection, and second is the procedure to update the archive members. They used mutation operator to enhance the exploratory ability of the algorithm.

Fan, Chang, and Chuang [9] proposed a multi-objective particle swarm optimizer which is constructed based on the concept of Pareto dominance taking both the diversified search and empirical movement strategies into account. They used polynomial mutation to maintaining the diversity of the particles along the optimization process.

Hu and Yen [13] proposed an integrated and adaptive MOPSO based on Parallel Cell Coordinate System (pccsAMOPSO). Their proposal includes a leader group, self-adaptive parameters, and perturbing operator for balancing convergence and diversity. They employed an elitism learning strategy with a Gaussian mutation as the perturbation operator.

Zhu et al. [49] presented a novel archive-guided MOPSO algorithm (AgMOPSO) where the leaders for velocity update are selected from an external archive. They also used an immune-based evolutionary strategy to evolve the external archive. They stated that this kind of updating scheme was verified to promote the convergence speed and keep the diversity.

Han et al. [11] proposed a variant of MOPSO, named Adaptive Gradient Multi-objective Particle Swarm Optimization (AGMOPSO). They used self-adaptive flight parameters mechanism to balance the convergence and diversity. They claimed that the proposed algorithm can find better spread solutions and has faster convergence to the true Pareto-optimal front.

Xiang et al. [41] proposed a many objective PSO (MaPSO). They suggested a new leader selection strategy. They kept multiple historical solutions from which the leader is selected for each particle. They also use linearly decreased parameter which promotes convergence initially and diversity later. It was shown that their proposed MaPSO is highly competitive or significantly superior to other algorithms. In another study on many objective PSO, Luo et al. [22] proposed an algorithm called IDMOPSO. They used a selection strategy for personal best to enhance the capability of local exploration. They also developed a multi-global best selection mechanism to balance convergence and diversity.

Pan et al. [31] proposed a diversity enhanced multi-objective particle swarm optimization called DEMPSO. In that study, analysis of particles’ velocities is developed to assist variable clustering and elite selection. A diversity enhancing process based on the velocity analysis is carried out during the particles’ evolution.

Current study intents to make a slight extension to [5]. Coello et al. [5] presented an approach in which Pareto-dominance is incorporated into standard PSO to handle problems with several objective functions. In that study, the movement of a particle is based on its own previous movements (personal best) and the movements of particles in a repository (i.e., leader is selected from an external archive of non-dominated solutions), as well. If the current position of a particle is better than the previous movements, current position is located as personal best. On the other hand, leader is selected randomly from the repository with respect to locations of non-dominated solutions from a hypercube. A more crowded hypercube has less chance to be selected. External archive and the positions of the solutions included in this archive are updated regularly at each iteration. A dynamic mutation operator is also employed in that study. Both the numbers of the particles which are subject to be applied mutation operator and the positions of a particle to be mutated decrease through iterations.

In the current study, we applied the same procedures like [5] except the mutation operator. Instead of mutation operator, random immigrants method is used for diversity preservation.

Random immigrants method

In the field of evolutionary algorithms, random immigrants method is known as an effective tool for diversity, especially for dynamic optimization [38, 43]. This method is based on replacing some of the individuals (particles) with new ones. New individuals may be included in a complete random manner, or inclusion may be based on a memory scheme. The predecessors, or omitted individuals, can be selected randomly, or they can be determined with respect to a quality measure (e.g., fitness value or dominance relation).

Coello et al. [5] used mutation operator, such that the number of mutant individuals and the range of effect on decision variables decrease in a nonlinear fashion. In other words, all the particles are affected in the beginning with a high range. As the iteration number gets closer to the end, the number of mutants decreases to almost zero. Doing so, the algorithm gains a highly explorative and leveled (between exploration and exploitation) search capacity.

figure a

In a previous study [46], Martinez and Coello employed a re-initialization procedure to increase the diversity. In that study, a particle increases its age when it does not improve its personal position, and after exceeding a pre-defined age threshold, the particle is re-initialized. Our approach is totally different by means of re-initialization scheme.

Experimentation

Test problems

The first test problem was used by [8], and it is given in Eq. (4):

$$\begin{aligned} \text {Minimize} \ f_{1}(x_{1},x_{2})&= x_{1} \nonumber \\ \text {Minimize} \ f_{2}(x_{1},x_{2})&= \frac{g(x_{2})}{x_{1}} \nonumber \\ \text {Subject to}&\nonumber \\ g(x_{2})&=2.0-e^{-(\frac{x_{2}-0.2}{0.004})^{2}}-0.8 e^{-(\frac{x_{2}-0.6}{0.4})^2} \nonumber \\&0.1 \le x_{1},x_{2} \le 1. \end{aligned}$$
(4)

The second test problem was used by [16], and given in Eq. (5):

$$\begin{aligned} \text {Maximize} \ f_{1}(x_{1},x_{2})&= -x_{1}^{2}+x_{2} \nonumber \\ \text {Maximize} \ f_{2}(x_{1},x_{2})&= \frac{1}{2} x_{1}+x_{2}+1 \nonumber \\ \text {Subject to}&\nonumber \\&\frac{1}{6} x_{1}+x_{2}- \frac{13}{2} \le 0 \nonumber \\&\frac{1}{2} x_{1}+x_{2}- \frac{15}{2} \le 0 \nonumber \\&\frac{5}{x_{1}}+x_{2}- 30 \le 0 \nonumber \\&0 \le x_{1},x_{2} \le 7 \end{aligned}$$
(5)

The third test problem was used by [18], and it is given in Eq. (6):

$$\begin{aligned} \text {Minimize} \ f_{1}(\mathbf {x})&=\sum _{i=1}^{2} -10 e^{-0.2 \sqrt{x_{i}^2 + x_{i+1}^2}} \nonumber \\ \text {Minimize} \ f_{2}(\mathbf {x})&= \sum _{i=1}^{3}\left( \vert x_{i}\vert ^{0.8}+5 \sin (x_{i}^3) \right) \nonumber \\&-5 \le x_{i} \le 5 \nonumber \\&i=1,2,3 \end{aligned}$$
(6)
Table 1 Values of parameters
Fig. 1
figure 1

Pareto-fronts for the first test problem

Fig. 2
figure 2

Pareto-fronts for the second test problem

Fig. 3
figure 3

Pareto-fronts for the third test problem

Fig. 4
figure 4

Pareto-fronts for the fourth test problem

Similar motivation is employed in the current study for random immigrants method. Immigration probability and the number of immigrants decrease with a nonlinear fashion. Random immigrants procedure is shown in Algorithm 1.

The fourth test problem is called “portfolio optimization problem” [24], and it is given in Eq. (7):

$$\begin{aligned} \text {Minimize}&\sum \limits ^{N}_{i=1} \sum \limits ^{N}_{j=1} x _{i}x_{j} \sigma _{ij} \nonumber \\ \text {Maximize}&\sum \limits ^{N}_{i=1} x_{i} \mu _{i} \nonumber \\ \text {Subject to}&\nonumber \\&\sum \limits ^{N}_{i=1} x_{i}=1, \nonumber \\&0 \le x_{i}\le 1, \ i=1,\ldots , N. \end{aligned}$$
(7)

The data set for the first three test problems can be obtained from [39], while for the fourth one, it is available in [40].

Performance metrics

In this study, inspired by the analysis described in [5], three performance metrics are used for comparisons. These metrics are generational distance , spacing, and error ratio, and they are given in Eqs. (8)–(10).

$$\begin{aligned} D&=\frac{\sqrt{\sum \nolimits ^{n}_{i=1} d^2_{i}}}{n}, \end{aligned}$$
(8)

where n is the number of vectors in the set of nondominated solutions found so far, and \(d_{i}\) is the Euclidean distance between each of these and the nearest member of the Pareto optimal set:

$$\begin{aligned} S&=\sqrt{\frac{1}{n-1} \sum \nolimits ^{n}_{i=1} (\bar{d}-d_{i})^2}, \end{aligned}$$
(9)

where \(d_{i}=\min _{j}( \vert f^{i}_{1}(\mathbf {x})-f^{j}_{1}(\mathbf {x})\vert +\vert f^{i}_{2}(\mathbf {x})-f^{j}_{2}(\mathbf {x})\vert ), i,j=1,\ldots , n, \ i\ne j\), \(\bar{d}\) is the mean of all \(d_{i}\), and n is the number of nondominated vectors so far:

$$\begin{aligned} E&=\frac{\sum \nolimits ^{n}_{i=1} e_{i}}{n}, \end{aligned}$$
(10)

where n is the number of vectors in the current set of nondominated vectors available, \(e_{i}=0\) if the vector i is a member of the Pareto optimal set, and \(e_{i}=1\), otherwise. In addition to the above-mentioned metrics, run times are also evaluated to compare the speed of the two competing algorithms.

Hypothesis testing and parameters

To test how effective the proposed approach is, we compared two algorithms, namely MOPSO with mutation and MOPSO with random immigrants. We run both algorithms 30 times for each of the four problems. We compared both algorithms with respect to four performance metrics, both statistically and graphically. Statistical hypothesis testing (z test) is performed for comparison. The null hypothesis and the alternative hypothesis are stated as:

\(H_{0}\): \(\mu _{1}=\mu _{2}\)There is no significant difference between two approaches with respect to the related performance metric.

\(H_{1}\): \(\mu _{1}\ne \mu _{2}\)There is a significant difference between two approaches with respect to the related performance metric.

Here, \(\mu _{1}\) represents a mean value of 30 runs for respective performance metric for mutation MOPSO, and \(\mu _{2}\) represents a mean value of 30 runs for respective performance metric for random immigrants MOPSO. The parameters for both approach are given in Table 1.

Results

Graphical comparison for all four test problems is given in Figs. 1, 2, 3 and 4. The results shown in these figures represent the best solutions with respect to the D performance metric. It can be seen that both mutation and random immigrants MOPSO produce a good approximation for the real Pareto-front for given test problems.

For each test problem, results for each run with respect to each performance metric are given in Tables 2, 4, 6, and 8, respectively. Based on these results, z-test is performed, and z values are given in Tables 3, 5, 7, and 9.

Results for the first test problem

According to the results given in Tables 2 and 3 for S performance metric null hypothesis can be rejected. It means that there is a significant difference between two approaches, with respect to the mean values of S. It seems that the mutation MOPSO outperforms random immigrants MOPSO for this performance metric.

Regarding the D and E performance metrics, there is not enough evidence to reject null hypothesis. It means that there is no significant difference between two approaches, with respect to the mean values of D and E. For the time performance metric, the null hypothesis can be rejected. It means that there is a significant difference between the approaches, with respect to the run times of the algorithms. Random immigrants MOPSO seems to outperform mutation MOPSO method.

Table 2 Results for the first test problem
Table 3 Statistical hypothesis testing for the first test problem
Table 4 Results for the second test problem
Table 5 Statistical hypothesis testing for the second test problem
Table 6 Results for the third test problem
Table 7 Statistical hypothesis testing for the third test problem
Table 8 Results for the fourth test problem

Results for the second test problem

According to the results given in Tables 4 and 5, for SD,  and E performance metrics, null hypothesis cannot be rejected. In other words, there is no significant difference between two approaches. Regarding the time performance metric, null hypothesis can be rejected. There is a significant difference between two approaches, and random immigrants MOPSO outperforms mutation MOPSO algorithm.

Table 9 Statistical hypothesis testing for the fourth test problem

Results for the third test problem

According to the results given in Tables 6 and 7, null hypothesis cannot be rejected. It means that there is no significant difference between two approaches with respect to these performance metrics. Regarding the time performance metric, null hypothesis can be rejected, and the mutation MOPSO outperforms the random immigrants MOPSO.

Results for the fourth test problem

According to the results given in Tables 8 and 9, for S and D performance metrics, null hypothesis cannot be rejected. It means that there is no significant difference between two approaches with respect to these performance metrics. Nevertheless, for E performance metric, null hypothesis can be rejected. It means that there is a significant difference between two approaches with respect to E performance metric, and mutation MOPSO approach outperforms random immigrants MOPSO approach. Regarding the time performance metric, null hypothesis cannot be rejected. In other words, there is no significant difference between two approaches with respect to the running times.

Conclusion

Computational Intelligence is developed to create solutions for the real-life complex problems for which linear, non-linear, or stochastic models could not propose a remedy. Nature-inspired metaheuristic algorithms take an improving importance in computational intelligence. The most important issue in metaheuristics is to balance the exploitation and the exploration depending on the problem studied. When the problem has a single objective, the hybridized algorithms provide solutions, but it is more difficult when a multi objective analysis is made.

This paper is original to develop MOPSO using Random Immigrants, which has enlarged the diversity of solutions remarkably. Application of the proposed algorithm on four well-accepted data sets has shown the benefits of the solution for four performance measures, process timing, Generational Distance, Spacing, and Error Ratio. Statistical experiments are constructed to compare the Mutation Approach (that gives better than other existing solutions) and the Random Immigrants approach for diversity. Results show that Random Immigrants approach is faster in providing the solution in most of the cases and as good as mutation approach in the others.

Implementing the proposed algorithm in a real-life case and to compare the results (for both real-life cases and test suits) with the state-of-the-art variants of the MOPSO will be the extension of this study. Archive maintenance and local optima problems faced in using MOPSO worth new studies, which will be further analysis in our team. Having observed the timing performances of MOPSO with Random Immigrants, we would also like to recommend the application of this method in Big Data handling.

This paper will open a new dimension for the MOPSO researchers and provide a new tool for computational intelligence application.