Introduction

In practical applications, multi-objective optimization problems (MOPs) are widespread, such as design optimization of high-frequency transformers1, route planning2, and model training in federated learning3. So far, scholars have developed several effective multi-objective evolutionary algorithms (MOEAs)4,5,6,7 to tackle these problems. However, the search space significantly enlarges with increasing dimensions in both objective space and decision variables, leading to a severe degradation in the performance of most low-dimensional MOEAs when dealing with such problems. Therefore, LSMOPs have shown promising research prospects in recent years, and the approaches to solving LSMOPs are roughly classified into three categories.

The first type of method is optimizing the original variables of LSMOPs. These methods mainly include two types of strategies. The first strategy involves segmenting the decision variables through distinct methods, followed by separate optimization. Cooperative Coevolutionary (CC) Differential Evolution algorithm (CCGDE3)8 randomly partitions decision variables into equally sized subgroups, with each subgroup’s decision variables being optimized uniquely. MOEA based on random-based dynamic grouping strategy (MOEA/D-RDG)9 selects subgroup sizes with probability, and then randomly groups the decision variables. CC large-scale algorithm (CCLSM)10 groups decision variables based on a fast interdependency identification algorithm. However, the performance of these methods is closely related to the accuracy of grouping, and when the accuracy of grouping is low, it can cause a decline in the performance of the algorithm. The second strategy involves analyzing the decision variables and subsequently conducting optimization on them using different strategies for the variables of various types identified in the analysis. MOEA based on decision variable analysis (MOEA/DVA)11 separates the variables into mixed, distance and position variables. Taking into account population convergence, the distance variables are further subdivided into several sub-components and optimized independently. To address the difficulty in accurately classifying decision variables in MOEA/DVA, a clustering method is employed in12 to partition variables associated with diversity and convergence. Later, a parallel framework derived from large-scale MOEA (LMEA) is proposed in13, which eliminates the dependencies between sub-processes. Nonetheless, the analysis of decision variables using these methods often demands a substantial amount of computational resources.

The second type of method is the problem transformation scheme, whose main idea is using some transformation strategies to convert high-dimensional problems into low-dimensional ones. Subsequently, low-dimensional optimization algorithms are adopted along with other operators to solve LSMOPs. The characteristic of these algorithms is that they do not directly optimize the original variables of the LSMOPs. WOF14 achieves the transformation from the original to a weight optimization problem by assigning weights to the grouped decision variables. Since the grouping mechanism in WOF cannot provide heuristic information, a stochastic dynamic grouping approach is employed in15. This approach adaptively determines the group size, allowing individuals to discover more appropriate groupings during different stages of the optimization process. An adaptive variable grouping method based on Pearson correlation is developed in16, which adaptively divides variables as the solution evolves. To further enhance the efficiency of WOF, the two distinct optimization problems are merged into one stage in17. Additionally, this approach utilizes the advantages of the competitive swarm optimizer (CSO) algorithm to improve the search capability of the optimization process. However, these methods update the individuals only according to the optimal weights generated by some fixed individuals and lack guidance of the overall weight information, and the performance of these algorithms is affected to a certain extent. Large-scale multiobjective optimization framework (LSMOF)18 proposed a bidirectional search approach that associates the constructed motion direction vectors of partial individuals with weights, increasing the chance of intersecting the motion direction of individuals with the Pareto optimal set (PS) to improve search efficiency. However, this search direction covers a limited search space and may lead to search failure when the search direction does not point to the PS.

The third type of method is based on some novel search strategy. This method mainly generates offspring by designing a special search strategy. An efficient large-scale CSO algorithm (LMOCSO)19 adopts a two-pair competition mechanism to realize the learning of poor performance particles to good performance particles, and proposes an update strategy that simultaneously considers the velocity and acceleration of the particles, which effectively improves the performance of the poorer particles. However, only the poor particles are updated during each competition, the good particles are not reasonably guided, which could hinder the algorithm’s ability to break free from local optima, thereby impeding its capacity to converge towards better solutions. Inverse modeling MOEA (IM-MOEA)20 first transforms solutions using an inverse model that utilizes the Gaussian process as the basis from the decision space into the objective space. Subsequently, potential solutions are sampled to generate offspring through the inverse model from the objective space. However, some features of its reference vectors may pose certain challenges in handling LSMOPs. IM-MOEA based on decomposition algorithm (IM-MOEA/D)21 adopts a novel spatial grouping approach that employs k-means algorithm. Additionally, in order to select optimal reference vectors, it implements a selection approach using decomposition and global replacement as the basis. As this method introduces multiple parameters, adjusting different parameter values may affect its performance. Scalable small subpopulations based covariance matrix adaptation evolution strategy (\(\mathrm {S^3}\)-CMA-ES)22 adopts the variable analysis method described in the paper11 to divide the decision variables. Then, it applies an adaptive evolution strategy based on the covariance matrix to further divide them into smaller subgroups, and offspring are generated by iteratively updating the Gaussian model. However, for some complex problems with strong correlation between variables, the decision variables are often difficult to divide, leading to a decrease in the performance of the algorithm.

Due to the fact that WOF does not directly optimize the original variables, the algorithm’s convergence accuracy is limited to some extent. Nevertheless, it still plays a significant role in LSMOPs due to its advantages in optimizing low-dimensional variables and achieving fast convergence speeds. However, it also has some disadvantages, and there is still room for improvement in its performance. The traditional WOF employs a single-population optimisation approach, which may result in the loss of population diversity when the algorithm converges rapidly. The weight population is optimised based on selected reference solutions, and a set of weights selected from the weight population is used for population updating, which may result in the population becoming concentrated in local regions. The research found that incorporating a set of weights containing more information for population updating represents an effective method for the full utilisation of the low computational cost advantage of WOF and the avoidance of population concentration in local regions. Building upon this, a multi-population multi-stage adaptive weighted optimization algorithm framework (MPSOF) is proposed. The main contributions of MPSOF are summarized as follows.

  • A three-stage algorithm framework for LSMOPs is developed. In stage I, the multiple populations strategy is implemented to maintain population diversity. In stage II, an optimization process is devised for balancing both diversity and convergence in subpopulations. In stage III, a global optimization is executed, thereby improving the distribution of solutions and algorithm convergence quality.

  • An adaptive mixed-weight individual updating strategy is introduced to diversify the individual updating methods. Addressing the issues in traditional WOF where selecting optimal weights through reference directions leads to the duplication of individual update weights, reducing population diversity, and wasting computational resources, a new set of weight vectors is derived through processing the optimal weight vectors of each subpopulation during the optimization’s second stage. Based on the frequency of weight vector repetition and the evolutionary status of each subpopulation, a novel individual generation strategy is designed, which adaptively adjusts the individual updating method, balancing the exploration and exploitation capabilities of subpopulation individuals and conserving computational resources.

  • To validate the efficacy of MPSOF, three types of experiments are conducted using some typical large-scale multi-objective datasets. By adopting different MOEAs within the new algorithm framework, it is proved that this framework can enhance the effectiveness of traditional MOEAs in addressing LSMOPs. Moreover, by embedding the same MOEA into different frameworks for LSMOPs, it is substantiated that the new framework has superior performance. By embedding a suboptimal algorithm into the new algorithm framework and conducting experiments on these test sets, and comparing the experimental results with those obtained from several advanced large-scale MOEAs, it is validated that MPSOF outperforms other MOEAs.

Related work

MOPs should simultaneously consider multiple conflicting objectives during the optimization process. It can be described mathematically as:

$$\begin{aligned}&\min f(X)=({{f}_{1}}(X),{{f}_{2}}(X),\ldots ,{{f}_{M}}(X)) \nonumber \\&\text {subject to }X\in {{\mathbb {R}}^{d}},\text { and }{{h}_{k}}\left( X \right) \le 0,k\in \left\{ 1,2,\ldots,I \right\} \end{aligned}$$
(1)

where f(X) is an objective function with a number of M conflicting objectives, X=(\({{x}_{1}}\),\({{x}_{2}}\),\ldots,\({{x}_{d}}\)) donates the d-dimensional decision variable vector. h and I are constraint functions and number of constraints, respectively. \(\mathbb {R}^{d}\) is feasible spaces. Due to the conflicting nature of the objective, the Pareto dominance relationship is used to distinguish the quality of two solutions. If \({{X}_{1}}\), \({{X}_{2}}\) \(\in\) \(\mathbb {R}^{d}\), \({{X}_{1}}\) is known as Pareto dominated \({{X}_{2}}\) (represented as \({{X}_{1}}\) \(\prec\) \({{X}_{2}}\)), when and only when the following conditions are met :

$$\begin{aligned}&\forall i\in 1,2,\ldots,M,{{f}_{i}}\left( {{X}_{1}} \right) \le {{f}_{i}}\left( {{X}_{2}} \right) \nonumber \\&\exists j\in 1,2,\ldots,M,{{f}_{j}}\left( {{X}_{1}} \right) <{{f}_{j}}\left( {{X}_{2}} \right) \end{aligned}$$
(2)

If X is not dominated by other solutions in \(\mathbb {R}^{d}\), then X is called the Pareto optimal solution, the set of all Pareto optimal solutions {X \(\in\) \(\mathbb {R}^{d}\) | \(\forall\) Y \(\in\) \(\mathbb {R}^{d}\),Y \(\nprec\) X} is called the PS, {f(X) | X\(\in\) PS} is called the Pareto front (PF). When d is no less than 100 and M is no less than 2, the optimization problems represented by Eq. (1) are often called LSMOPs.

Basic weighted optimization framework

WOF23,24 contains alternating optimization and normal optimization. In alternating optimization, a transformation function is employed to convert the decision space from high-dimensional to low-dimensional based on the selected reference solution. After that, the existing MOEA searches for the optimal solution in the reduced subspace. In the normal optimization, the original variables are optimized by a selected MOEA to obtain higher quality solutions.

Figure 1 portrays the flowchart of the WOF. Where \(\varepsilon\) is a control parameter for dividing the two stages, FE and maxFE represent the number of function evaluations at present and in total, respectively. At the alternating optimization stage, the original variables are first optimized for a predetermined number of evaluations, after which q reference solutions are selected using the reference direction selection strategy. For each selected solution, its decision variables are grouped and the weight population W is initialized. The weight population is optimized using a selected MOEA by associating with the solution. Specifically, the variables of the reference solution are grouped into \(\gamma\) groups using an ordered grouping method. By associating a weight \({{w}_{j}}\) with each group, an optimization problem with a d-dimensional decision variable is transformed into a \(\gamma\)-dimensional weight optimization problem. The above transformation process can be formulated as follows.

$$\begin{aligned} \psi \left( \textbf{w},\textbf{x} \right) =\left( \underbrace{{{w}_{1}}{{x}_{1}},\ldots,{{w}_{1}}{{x}_{l}}}_{{{w}_{1}}},\ldots,\underbrace{{{w}_{\gamma }}{{x}_{d-l+1}},\ldots,{{w}_{\gamma }}{{x}_{d}}}_{{{w}_{\gamma }}} \right) \end{aligned}$$
(3)

The transformation function is as shown in Eq. (4)

$$\begin{aligned} \psi \left( {{w}_{j}},{{x}_{i}}^{\text { }\!\!'\!\!\text { }} \right) ={{x}_{i}}=\left\{ \begin{array}{*{35}{l}} {{x}_{i}}^{'}+\left( {{w}_{j}}-1.0 \right) \cdot \left( {{x}_{i,\max }}-{{x}_{i}}^{'} \right) \text { }if\text { }{{w}_{j}}>1.0 \\ {{x}_{i,\min }}+{{w}_{j}}\cdot \left( {{x}_{i}}^{'}-{{x}_{i,\min }} \right) \text { }if\text { }{{w}_{j}}\le 1.0 \\ {{w}_{j}}\in \left[ 0,2 \right] \\ \end{array} \right. \end{aligned}$$
(4)

where \({{w}_{j}}\) (j = 1,\ldots,\(\gamma\)) represents the j-th weight, \(x_{i}^{'}\) represents the selected reference solution for optimizing the weight \({{w}_{j}}\), \({{x}_{i,max}}\) and \({{x}_{i,min}}\) are the upper and lower bounds of variable \({{x}_{i}}\) (i = 1,\ldots,d).

For the optimized weight population W, q weight individuals are selected from it and then applied to the original population to produce some new individuals. Subsequently, the newly generated individuals are merged with those obtained during the optimization process of W. Any duplicates are removed, and the initial population for the second stage is obtained through non-dominated selection. Finally, further optimization is performed using a suitable algorithm.

Figure 1
figure 1

Flowchart of the WOF.

Competitive swarm optimizer

The CSO algorithm25 takes inspiration from the collective behavior observed in biological populations, such as the particle swarm optimization (PSO) algorithm. Unlike PSO, which uses global and individual best positions to drive population evolution, CSO incorporates competition among particles to guide population evolution. The competition process is depicted in Fig. 2. In the current population, competition takes place between two randomly selected particles. The particle demonstrating exhibits superior performance is called the winner, while the other is labeled as the loser. The winner advances to the subsequent iteration, and the loser updates its position with Eq. (5).

$$\begin{aligned} {{v}_{L}}\left( t+1 \right)&= {{c}_{1}}{{v}_{L}}\left( t \right) +{{c}_{2}}\left( {{x}_{W}}\left( t \right) -{{x}_{L}}\left( t \right) \right) +\varphi {{c}_{3}}\left( \bar{x}\left( t \right) -{{x}_{L}}\left( t \right) \right) \nonumber \\ {{x}_{L}}\left( t+1 \right)&= {{x}_{L}}\left( t \right) +{{v}_{L}}\left( t+1 \right) \end{aligned}$$
(5)

where \({{c}_{1}}\), \({{c}_{2}}\), \({{c}_{3}}\) are three random vectors in the domain [0,1], \({{v}_{L}}\) (t) is the velocity of the loser, \({{x}_{W}}\) (t) and \({{x}_{L}}\) (t) are the positions of the winner and loser, \(\bar{x}\) (t) is the global average position and \(\varphi\) is the control parameter.

In recent years, due to CSO’s efficient search capabilities in addressing large-scale single objective problems26, Tian et al. extended CSO to solve LSMOPs and proposed an efficient optimization algorithm based on CSO19. They designed a two-stage approach for updating individual position, greatly enhancing the search efficiency of traditional CSO algorithm. The specific update strategy for the particles is as follows.

Figure 2
figure 2

The competitive process of CSO.

The fitness of each particle is calculated based on27. Afterward, among the competing particles, those with poor fitness learn from those with good fitness and update their own positions according to Eq. (6).

$$\begin{gathered} \vec{v}_{L} \left( {t + 1} \right) = c_{0} \vec{v}_{L} \left( t \right) + c_{1} \left( {\vec{x}_{W} \left( t \right) - \vec{x}_{L} \left( t \right)} \right) \hfill \\ \vec{x}_{L} \left( {t + 1} \right) = \vec{x}_{L} \left( t \right) + \vec{v}_{L} \left( {t + 1} \right) + c_{0} \left( {\vec{v}_{L} \left( {t + 1} \right) - \vec{v}_{L} \left( t \right)} \right) \hfill \\ \end{gathered}$$
(6)

where \({x}_{W}\) and \({x}_{L}\) are the positions of the winner and loser, \({c}_{0}\) and \({c}_{1}\) are random numbers between 0 and 1.

Research motivations

Reducing the dimensionality of decision space is an effective method in addressing LSMOPs. As a classical dimensionality reduction method for LSMOPs, WOF adopts the variable grouping weighted transformation method to decrease the dimensionality of the optimization variables, thereby reducing the algorithm’s computational complexity and enhancing its convergence speed. However, the traditional WOF uses the same weights for each group of variables, and scales the variables in the same group to the same degree, which leads to difficulties in maintaining population diversity. By dynamically adjusting the evaluation numbers of the weights to change their diversity17, the population convergence and diversity are well balanced. However, within the same evolutionary generation, individuals still use a single set of weights to update their positions, thus limiting their ability to maintain diversity. How to design a better algorithmic framework that effectively balances population convergence and diversity is essential for enhancing algorithmic performance in LSMOPs. Under this motivation, a three-stage algorithm framework is developed. It aims to capitalize on the dimensionality reduction capability of WOF and effectively leverage its relatively small computation cost in the initial stage. Additionally, it adopts a multi-population method to sustain the population diversity and accelerate convergence speed. In the second stage, an adaptive multi-population hybrid weight position update method is designed for each subpopulation, which uses global weight information to enrich the individual weight updating method, and the convergence and diversity of subpopulations is better balanced. Since the optimization target during the optimization process is not the original variables, which exerts a definite impact on the algorithm’s convergence accuracy. Therefore, in the third stage, the population with original variables is optimized with an existing MOEA. Furthermore, the algorithmic framework of this paper also can be seamlessly integrated with different MOEAs.

Figure 3
figure 3

Flowchart of the MPSOF.

The proposed MPSOF

The framework of MPSOF

Figure 3 illustrates the basic flowchart of MPSOF. First, three populations \({{P}_{1}}\), \({{P}_{2}}\), \({{P}_{3}}\) are initialized. Then, each population adopts the same alternating optimization method as WOF in Fig. 1 for the first phase update. These updated populations are immediately used as the initial populations for the second phase, named \({{P}_{11}}\), \({{P}_{22}}\), \({{P}_{33}}\). After preliminary optimization, each of them has q reference solution ({\(\textbf{x}_{1}^{i}\),...,\(\textbf{x}_{q}^{i}\)}, i = 1, 2, 3) selected. Then, the weighted populations are initialized and further optimized, as shown in the weighted optimization area Fig. 1. Each optimized weight population has q weight vectors selected. New individuals are then adaptively updated based on these weight vectors and the mixed weight vectors obtained through the fusion of these weight vectors. Finally, the initial population for the third phase is selected from the new populations generated in the first two stages. MPSOF utilizes an existing MOEA to optimize the overall variables and generate the final solution set.

Algorithm 1 presents the pseudocode of MPSOF. There are three stages in the entire algorithm. First, the initial population is divided into three subpopulations. When FE \(\le\) \(\theta\) \(\times\) maxFE, each subpopulation executes Algorithm 2 to complete the position updating of the individual in the first stage. At this stage, the evaluations number for each optimized subpopulation is \({{t}_{1}}\), and the evaluations number for each optimized weight is \({{t}_{2}}\). The main goal of this part is to leverage the multi-population method to enhance population diversity, while fully capitalizing on the advantages of WOF, including its low computational complexity and fast execution speed. When FE \(\le\) \(\delta\) \(\times\) maxFE, it enters the second stage of optimization. The reference solutions for the three subpopulations are obtained using the same method as in Algorithm 2. Subsequently, the weights are optimized with these reference solutions. In each weighted optimization phase, the weight variables of three subpopulation are generated. Following this, an adaptive multi-population hybrid weight updating strategy is designed to produce new offspring individuals. The concrete optimization steps of this stage are illustrated in Algorithm 3, having the primary purpose of balancing convergence and diversity using the designed updating strategy. In the third stage, a large-scale MOEA is chosen to optimize the new individuals generated from the first and second stages on a global scale, further enhancing the convergence accuracy.

Algorithm 1
figure a

MPSOF

Algorithm 2 provides detailed steps for the alternating optimization part of MPSOF. The initial population is optimized in Line 1, and q solutions are chosen from each of the three population in Line 2. In the main loop (Lines 4-10), the decision variables of each selected solution are grouped and weight populations are initialized (Line 5). The weights are optimized using MOEA by combining the selected solution with the weights in the population via a transformation function (Line 6), and q weight vectors are selected from each of the three optimized weight populations (Line 7). These are then applied to the corresponding subpopulations to generate new individuals (Line 8). The individuals are merged in Line 9. Finally, the new population of the first stage is obtained by non-dominated sorting (Line 11).

Algorithm 2
figure b

Alternating Optimization Operations

Algorithm 3 gives the specific steps for hybrid weighted updating the positions of individuals. Among them, the front 6 lines are similar to the execution process in Algorithm 2. In the main loop (Lines 4–10), the mixed weighted updating proportion factor xr is calculated (Line 7). Subsequently, the adaptive updating strategy is used to generate the positions of the individuals in different subpopulations (Line 8), balancing their convergence and diversity. Detailed descriptions of the hybrid proportionality factor and the adaptive multi-population hybrid weighted updating strategy will be presented in Sections “Design the mixed proportion factor” and “Adaptive multi-population mixed weights individual updating strategy”.

Design the mixed proportion factor

In WOF, the optimal weights are obtained by reference directions24, the excellent weights might be selected multiple times. When the same weights are applied to the same individual, more duplicate individuals will be generated, and the algorithm will waste more computation resources on evaluating these duplicate individuals. To avoid generating many duplicate weights, a deletion strategy is adopted for multiple duplicate weights during the second stage of the algorithm. Only the weight vectors that are not duplicated among them are retained to act on the individuals. In addition, according to the proportion of repeated weights, the mixed proportion factor is designed. After deleting the duplicate weights, a specific portion of individuals are chosen to update their positions using the mixed weights, while another part of individuals update their positions using the weights of the subpopulation itself. This operator enables a trade-off in convergence and diversity. Concrete design of the mixing scale factors is presented below.

Supposing there are three weight populations, denoted as \({{W}_{1}}\), \({{W}_{2}}\), \({{W}_{3}}\), and the size of them is k. These weight populations are then applied separately to the corresponding reference solutions for optimisation, with the objective of obtaining the optimised weight populations. Subsequently, q optimal weight vectors are selected from the k weight vectors respectively. The number of duplicate weight vectors, represented by num1, num2, num3, is calculated for each set. Subsequently, the overall number of duplicate weight vectors is computed using Eq. (7). The weight repetition proportion of the current population is obtained with Eq. (8). The mixed weight contains global weight information rather than the only optimal weight of the subpopulation. Each subpopulation uses two types of weights to generate new individuals according to a designed proportion. When a large number of individuals utilize mixed weights to produce new individuals, the diversity of the subpopulation experiences significant enhancement, while its convergence is weakened. On the contrary, the convergence of the subpopulation undergoes substantial improvement, but its diversity is weakened. To balance the effect of the two types of weights, a mixed proportion factor xr based on r is designed and the value of it is limited between approximately 0.3 and 0.8 as shown in Eq. (9).

$$\begin{aligned}&num=num1+num2+num3 \end{aligned}$$
(7)
$$\begin{aligned}&r=\frac{num}{3\times q} \end{aligned}$$
(8)
$$\begin{aligned}&xr={{e}^{r-1.25}} \end{aligned}$$
(9)
Algorithm 3
figure c

Mixed weighted updating the positions of individuals

The mixed proportion factor xr reflects the evolutionary state of the current population. A higher xr value corresponds to a larger number of the current duplicate weights, resulting in fewer non-duplicate weights being utilized to update the population. This means that the new individuals will be inclined towards local optimality, leading the population into an exploitation phase. On the contrary, the more non-repetitive weights are utilized to obtain new individuals, the individual diversity is improved, making it more likely for the population to enter the exploratory phase. When the population tends to the exploitation stage, more individuals are updated using the mixing weights to elevate the diversity and bolster the exploratory capabilities of the individuals. Conversely, when the population tends to the exploration stage, more individuals are updated using their optimal weights to enhance the population convergence. Through this operation, the exploitation and exploration capabilities of the subpopulations are balanced well. The process of determining the mixing proportion factor is shown in Algorithm 4.

Algorithm 4
figure d

Mixing proportion factor

Adaptive multi-population mixed weights individual updating strategy

The WOF provides an effective optimization method for LSMOPs, and also offers a new approach for using low-dimensional MOEAs to solve LSMOPs. Yet, traditional weighted optimization algorithms rely solely on a set of optimized weight vectors to update individuals. Although this approach does enhance the algorithm’s convergence speed to a certain extent, its lack of diversity may result in local optima. To fully leverage the advantages of this approach, this paper develops an adaptive multi-population hybrid weight individual updating method. The method addresses the dual challenges of enhancing population diversity and balancing subpopulation convergence with diversity. The hybrid weight individual updating method is described as follows.

For the three subpopulations in the second stage, the optimal weight vectors \({{G}_{w1}}\), \({{G}_{w2}}\), \({{G}_{w3}}\) of each subpopulation can be determined using the same method as in the WOF. Then, the weights \({{W}_{s1}}\), \({{W}_{s2}}\), \({{W}_{s3}}\) of these three subpopulations are combined to obtain weight vectors with global information. As shown in Eqs. (10) and (11), the optimal weights of three subpopulations can be different. For simplicity in this paper, the weight of three subpopulations is 1/3. Subsequently, according to the mixed ratio factor, different numbers of individuals are selected and updated using the unique optimal weight vectors of the subpopulations themselves and the mixed weight vectors, generating some initial individuals for the third stage, as shown in Eqs. (13), (14), and (15).

$$\begin{aligned}&G{{m}_{w}}={{W}_{s1}}\times {{G}_{w1}}+{{W}_{s2}}\times {{G}_{w2}}+{{W}_{s3}}\times {{G}_{w3}} \end{aligned}$$
(10)
$$\begin{aligned}&{{W}_{s1}}+{{W}_{s2}}+{{W}_{s3}}=1 \end{aligned}$$
(11)
$$\begin{aligned}&G{{m}_{w}}=1/3\times ({{G}_{w1}}+{{G}_{w2}}+{{G}_{w3}}) \end{aligned}$$
(12)
$$\begin{aligned} {{n}_{{{P}_{11}}}}&= \psi (G{{m}_{w}},xr\times {{P}_{11}}) \nonumber \\ {{n}_{{{P}_{21}}}}&= \psi (G{{m}_{w}},xr\times {{P}_{22}}) \nonumber \\ {{n}_{{{P}_{31}}}}&= \psi (G{{m}_{w}},xr\times {{P}_{33}}) \end{aligned}$$
(13)
$$\begin{aligned}&{{n}_{{{P}_{12}}}}=\psi ({{G}_{w1}},\left( 1-xr \right) \times {{P}_{11}}) \nonumber \\&{{n}_{{{P}_{22}}}}=\psi ({{G}_{w2}},\left( 1-xr \right) \times {{P}_{22}}) \nonumber \\&{{n}_{{{P}_{32}}}}=\psi ({{G}_{w3}},\left( 1-xr \right) \times {{P}_{33}}) \end{aligned}$$
(14)
$$\begin{aligned} {{n}_{{{P}_{1}}}}&= {{n}_{{{P}_{11}}}}\cup {{n}_{{{P}_{12}}}} \nonumber \\ {{n}_{{{P}_{2}}}}&= {{n}_{{{P}_{21}}}}\cup {{n}_{{{P}_{22}}}} \nonumber \\ {{n}_{{{P}_{3}}}}&= {{n}_{{{P}_{31}}}}\cup {{n}_{{{P}_{32}}}} \end{aligned}$$
(15)

where \({{G}_{w1}}\), \({{G}_{w2}}\), \({{G}_{w3}}\) denote the three sets of optimal weights selected from the three optimized weight populations, \({{Gm}_{w}}\) is a set of weights obtained by fusing the weights of multiple subpopulations and \(\psi\) is the transformation function. Algorithm 5 shows the pseudocode for the individual updating process.

Algorithm 5
figure e

Adaptive update strategy

Experimental studies

In this section, three types of experiments are designed to validate the effectiveness of the developed algorithm framework (MPSOF) from various perspectives. The first type of experiment involves using different optimization algorithms in the new framework to assess its impact on the original algorithm performance. The second type of experiments involves applying the fast and elitist genetic algorithm (NSGAII)28 to three different optimization frameworks, WOF24, LSMOF18 and MPSOF for evaluating the performance of these frameworks. The third category of experiment involves embedding the general performance algorithm LMOCSO into MPSOF, and then comparing it with seven advanced large-scale MOEAs using the LSMOP, WFG, and UF test function problems.

Quality metrics

In experiments, the evaluation of the algorithms’ performance relies on commonly used performance indicators, Inverse Generation Distance (IGD)29, Hypervolume (HV)30 and Spacing (SP)31.

IGD is defined by Eq. (16). Where S symbolizes the true PS, P represents the approximate PS, N denotes the population size, and dis(\({{p}_{i}}\), S) signifies the minimum Euclidean distance between points \({{p}_{i}}\) and S. IGD quantifies the algorithm diversity and convergence by computing the mean distances from the closest generated solutions to the true Pareto front (PF). A lower IGD value signifies that the algorithm has superior comprehensive performance.

$$\begin{aligned} IGD(S,P)=\frac{\sum \limits _{i=1}^{N}{{{p}_{i}}\in P\text { dis(}{{p}_{i}},S\text {)}}}{N} \end{aligned}$$
(16)

HV is defined as shown in Eq. (17), which measures the algorithm performance by quantifying the hypervolume formed between the generated solutions and the reference points. It is an effective one-dimensional quality metric that is strictly monotonic in terms of Pareto domination. A higher value of HV suggests that the algorithm possesses a greater capability to cover the real PF. Here, P represents the approximate optimal solutions achieved by the algorithm in objective space, \({{a}^{i}}\) = \({{\text {(}a_{1}^{i},a_{2}^{i},\ldots ,a_{m}^{i})}^{T}}\) is the reference point for selection, \({{D}_{0}}\)(.) denotes the Lebesgue measure.

$$\begin{aligned} HV={{D}_{0}}\left( \bigcup \limits _{{{p}_{j}}\in P}{\left[ {{f}_{1}}(x),a_{1}^{i} \right] }\times \cdots \times \left[ {{f}_{m}}(x),a_{m}^{i} \right] \right) \end{aligned}$$
(17)

SP is defined by Eq. (18), which represents the distribution of individuals within the approximate PS in target space. It is a diversity evaluation indicator. A lower SP value corresponds to a more evenly distributed solution set. Here, N is the population size, \({{Dis}_{q}}\) is the Euclidean distance between the q-th and the nearest individuals, \(\bar{Dis}\) is the average distance.

$$\begin{aligned} SP = \frac{{{{\left[ {\frac{1}{N}\sum \limits _{q = 1}^N {{{\left( {Di{s_q} - \bar{Dis} } \right) }^2}} } \right] }^{\frac{1}{2}}}}}{{\bar{Dis} }} \end{aligned}$$
(18)

Experimental parameters

Experimental settings

In this paper, each algorithm is performed independently 20 times on the PlatEMO32 platform, with maxFE set to 100000 as the termination criterion and N set to 105. The two main parameters in MPSOF are obtained by parameter sensitivity analysis: \(\theta\) is 0.4, \(\delta\) is 0.6, while the other parameters remain consistent with those of WOF. Among them, \({{t}_{1}}\) is 1000, \({{t}_{2}}\) is 500, \(\gamma\) is 4, the initialized weight population size k equals 10, the number of selected weight vectors and reference solutions q equals M+1. Other comparison algorithms adopt the advised parameter settings provided in the relevant papers. Statistical results are obtained through the utilization of Wilcoxon rank-sum test with a significance level of 0.05. The bold font indicates the optimal solutions, and the symbols “+/−/=” represent whether the comparison algorithms are considerably superior, inferior, or similar to MPSOF.

Table 1 Average IGD values for LSMOP1, LSMOP5 and LSMOP8 with \(\delta\) change from 0.1 to 0.9 under the interval 0.1 when \(\theta\) is different.

Parameter sensitivity analysis

\(\theta\) and \(\delta\) are two important parameters of MPSOF, which divide the whole algorithm into three stages. Different parameters will have different impacts on the algorithm’s performance. To determine these two parameters, parameter sensitivity analysis is used in this paper. Three functions with different properties were selected from LSMOP for the experiments, where the decision variable dimension and the number of objectives are respectively 500 and 3. \(\theta\) and \(\delta\) change from 0.1 to 0.9 respectively with an interval of 0.1. The first experiment is the average IGD value obtained by the three functions as \(\delta\) takes different values and \(\theta\) changes from 0.1 to 0.9 (\(\theta\) \(\le\) \(\delta\)). The results are shown in Table 1. For example, “0.1681” in Table 1 indicates that the average IGD is 0.1681 with \(\delta\) change from 0.4 to 0.9 under the interval 0.1 when \(\theta\) is 0.4 for LSMOP1. The second experiment is the average IGD value obtained by the three functions as \(\theta\) takes different values and \(\delta\) changes from 0.1 to 0.9 (\(\theta\) \(\le\) \(\delta\)). The results are shown in Table 2. For example, “0.0704” in Table 2 indicates that the average IGD is 0.0704 with \(\theta\) change from 0.1 to 0.6 under the interval 0.1 when \(\delta\) is 0.6. The last lines of the two tables are the average IGD of LSMOP1, LSMOP5 and LSMOP8.

From Table 1, the test results for the three typical functions show that the average IGD value is minimized when \(\theta\) is 0.4, so \(\theta\) is set to 0.4 in this paper. From Table 2, the test results for these three functions demonstrate that the average IGD value is minimized when \(\delta\) is 0.6. Taking into account the average characteristics impact of parameters on the algorithm, this paper has chosen to set \(\delta\) as 0.6.

Table 2 Average IGD values for LSMOP1, LSMOP5 and LSMOP8 with \(\theta\) change from 0.1 to 0.9 under the interval 0.1 when \(\delta\) is different.

Experimental results and analysis

In experiments, the algorithms’ performance in solving LSMOPs is evaluated by testing 2-objective and 3-objective problems with variables of 500 and 1000 dimensions.

General performance of MPSOF

The main intention of the experiments in this part is to validate that MPSOF has an enhancing effect on the performance of different embedded algorithms. Three different classical algorithms NSGAII28, a competitive PSO for multi objective (CMOPSO)33 and LMOCSO19 are applied in MPSOF. Tables 3 and 4 display the statistical data of IGD and HV values obtained from LSMOP by different algorithms.

Table 3 shows that MPSNSGAII is significantly superior to NSGAII in 32 test instances out of 36 test instances. Similarly, MPSCMOPSO outperforms CMOPSO in 34 test instances. Compared to the original algorithm, MPSNSGAII and MPSCMOPSO exhibit slightly lower performance on LSMOP2 with 3-objective, but the difference between them is minimal. This is mainly because the classical algorithm itself has good performance when dealing with simpler linear problems, and maintaining population diversity does not significantly increase the accuracy of the performance of the algorithm. MPSLMOCSO demonstrates superior performance over LMOCSO in 33 out of 36 test instances. For the LSMOP9 with three objectives of discontinuous PF, MPSLMOCSO performs worse than LMOCSO. According to the results from the final row of tests, NSGAII, CMOPSO, and LMOCSO outperform the algorithm in this paper on 3, 2, and 3 test instances, are inferior to the algorithm in this paper on 32, 34, and 33 test instances, and are similar in performance to the algorithm in this paper on 1, 0, and 0 test instances. This indicates that MPSOF is capable of effectively utilizing existing MOEAs to tackle LSMOPs, as the complexity of the optimization problem has been significantly reduced.

Table 3 The statistical IGD results with NSGAII, MPSNSGAII, CMOPSO, MPSCMOPSO, LMOCSO, and MPSLMOCSO on 36 LSMOP test problems.
Table 4 The statistical HV results with NSGAII, MPSNSGAII, CMOPSO, MPSCMOPSO, LMOCSO, and MPSLMOCSO on 36 LSMOP test problems.

Table 4 displays that the MPSOF has demonstrated strong performance across the majority of the test problems (LSMOP1, LSMOP5, LSMOP8, and LSMOP9). Additionally, MPSLMOCSO only performs slightly lower than LMOCSO on the LSMOP3 with 2-objective and LSMOP7 with 3-objective. A zero value of HV means that the compared algorithm cannot generate any solutions that meet the objective requirements. The table also indicates that among 36 test instances, NSGAII, CMOPSO, and LMOCSO are superior to the MPSOF on 3, 2, and 1 instances respectively, inferior to the MPSOF on 25, 26, and 30 instances, and show similar performance to MPSOF on 8, 8, and 5 instances.

Table 5 The statistical IGD results with WOF-NSGAII, LS-NSGAII, and MPSNSGAII on 36 LSMOP test problems.

Performance comparison of different optimization frameworks

To verify the efficacy of various large-scale multi-objective algorithm frameworks, NSGAII28 is incorporated into WOF24, LSMOF18 and MPSOF as the main optimizer. Subsequently, experiments are conducted on LSMOP, and the obtained results are compared for analysis.

Tables 5 and 6 display the average IGD and HV values achieved by NSGAII in the different algorithm frameworks.

Table 6 The statistical HV results with WOF-NSGAII, LS-NSGAII, and MPSNSGAII on 36 LSMOP test problems.

Table 5 indicates that the three frameworks respectively yield the optimal values in 2, 9, and 25 out of 36 test instances. On the discontinuous LSMOP9 problem with 3-objective, WOF-NSGAII achieves the best results, but the differences between the three algorithms are not significant, indicating that none of the three algorithms has an absolute advantage in this class of problems. Additionally, MPSNSGAII performs slightly lower than LS-NSGAII on the LSMOP6 and LSMOP7 problems with 2-objective. From the last row of the test results, WOF-NSGAII and LS-NSGAII outperform MPSNSGAII on 1 and 8 test problems, are inferior to MPSNSGAII on 26 and 23 test problems, and show comparable performance to MPSNSGAII on 9 and 5 test problems, respectively. The comprehensive performance of MPSOF is the best. This is due to the introduction of an additional set of global weight updating populations, which effectively balances diversity and convergence.

Table 6 demonstrates that the overall performance of MPSNSGAII outperforms the other two comparative frameworks. Specifically, WOF-NSGAII, LS-NSGAII and MPSNSGAII achieve 5, 8, and 19 best results respectively. Among them, LS-NSGAII performed well on the 3-objective LSMOP3 and LSMOP5 problems, However, the similar HV values between MPSNSGAII and LS-NSGAII indicate comparable performance in handling these two types of problems. For the 2-objective LSMOP3 problem, all three algorithms result in HV values of zero, indicating that none of the compared algorithms demonstrate ideal performance. For 36 test instances, WOF-NSGAII outperforms MPSNSGAII on 3 instances, is inferior to MPSNSGAII on 19 instances, and shows similar performance to the algorithms in this paper on 14 instances. LS-NSGAII outperforms MPSNSGAII on 4 instances, is inferior to MPSNSGAII on 15 instances, and shows similar performance to MPSNSGAII on 17 instances.

Regardless of the IGD or HV values, MPSNSGAII achieves optimal results in most instances, indicating a clear advantage of the proposed framework over the other two algorithm frameworks.

Performance comparison of cutting-edge large-scale MOEAs

To compare the performance of MPSOF with several advanced algorithms, LMOCSO is integrated into MPSOF and named MPSLMOCSO. Subsequently, MPSLMOCSO is compared with seven other algorithms (LSTPA34, DGEA35, SSCEA36, IM-MOEA/D21, FDV37, ATLMOEA38 and LERD39) on three function sets of LSMOP, WFG, and UF.

These algorithms represent the three types of methods currently used to solve LSMOPs. Specifically, LSTPA is a CSO that accelerates convergence using fitness values. DGEA is a direction-guided adaptive reproduction algorithm. SSCEA is a CC algorithm that performs segmentation processing on subspaces. IM-MOEA/D is a distribution estimation algorithm that can effectively solve LSMOPs. FDV belongs to the problem transformation algorithms, which processes decision variables through fuzzy and precise evolution. ATLMOEA is an adaptive two-stage algorithm that uses neural networks to accelerate optimization. LERD is an improved algorithm for DVA that greatly reduces the cost of function evaluation required for DVA.

Figure 4
figure 4

The performance statistics of eight comparison algorithms with three function sets.

To validate the efficacy of MPSLMOCSO, LMOCSO is first compared with other seven algorithms. The statistical results of the algorithms that achieve optimal average IGD values on LSMOP1-LSMOP9, WFG1-WFG9, and UF1-UF10 are presented in Fig. 4.

Table 7 The statistical IGD results with LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV ,ATLMOEA, LERD and MPSLMOCSO on 36 LSMOP test problems.
Table 8 The statistical HV results with LATPA, DGEA, SSCEA, IM-OEA/D, FDV, ATLMOEA, LERD and MPSLMOCSO on 36 LSMOP test problems.
Table 9 The statistical IGD results with LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV, ATLMOEA, LERD and MPSLMOCSO on 36 WFG test problems.
Table 10 The statistical HV results with LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV, ATLMOEA, LERD and MPSLMOCSO on 36 WFG test problems.
Table 11 The statistical IGD results with LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV ATLMOEA, LERD and MPSLMOCSO on 20 UF test problems.
Table 12 The statistical HV results with LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV, ATLMOEA, LERD and MPSLMOCSO on 20 UF test problems.
Figure 5
figure 5

Pareto optimal solutions obtained from the best IGD values of eight algorithms running on the LSMOP2, WFG7 and UF3 with 2 objectives and 1000 decision variables.

Figure 6
figure 6

Pareto optimal solutions obtained from the best IGD values of eight algorithms running on the LSMOP4, WFG8 and UF9 with 3 objectives and 500 decision variables.

The lower triangle, circle, upper triangle, hexagon, square, plus, asterisk and diamond symbols respectively denote the optimal average IGD of LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV, ATLMOEA, LERD and LMOCSO achieved for the corresponding test problems, while the horizontal axis denotes the corresponding test problem. For example, in Fig. 4b , the “\(\bigcirc\)” indicated that DGEA gets the optimal average IGD for WFG9 among the eight algorithms. The “\(\Box\)” indicates that FDV derives the optimal average IGDs for WFG2, WFG4, WFG5 and WFG8 compared to other seven algorithms. For LSMOP1-LSMOP9, LMOCSO achieves the best results only on LSMOP9, while FDV and ATLMOEA achieve 4 and 4 best results. For the other two test suites, LMOCSO fails to achieve the optimal results, which suggests that LMOCSO does not exhibit significant advantages in performance.

The following experiments are conducted on three function sets using LMOCSO as the optimizer in MPSOF. The resulting average IGD and HV values are depicted in Tables 7 and 8.

Table 7 displays that MPSLMOCSO has the best average performance for large part of instances among the compared algorithms. Notably, it achieves the optimal outcomes for LSMOP1, LSMOP4, LSMOP6, and LSMOP7. This is attributed to MPSLMOCSO combining problem transformation mechanism with a multi-population method, enabling multiple populations to search within a small-scale space and achieve a balance between diversity and convergence. The performance of LSTPA, DGEA, SSCEA, IM-MOEA/D, ARMOEA and LERD is subpar compared to MPSLMOCSO, probably because LSTPA is fundamentally accelerated by CSO, which may not be sufficiently effective in searching within the limited evaluation numbers in the context of complex and high-dimensional problems. DGEA is not able to accurately determine the parental populations when searching for decision spaces with similar information, resulting in the algorithm being trapped in a local optimum. MPSLMOCSO performs less well than FDV on the LSMOP9 problem. From the results in the last row, LSTPA, SSCEA and LERD perform worse than MPSLMOCSO on 36 test instances. DGEA performs worse than MPSLMOCSO in 31 test instances, and its performance is similar to the algorithm in this paper on 5 test instances. While IM-MOEA/D performs worse than and similar to MPSLMOCSO in 35 and 1 test instances. FDV performs better than MPSLMOCSO in 6 test instances, but falls short of MPSLMOCSO in 27 test instances, and exhibits comparable performance to MPSLMOCSO in 3 test instances. ATLMOEA shows similar performance to MPSLMOCSO across 32 test instances, but is inferior to MPSLMOCSO in 4 instances.

Figure 7
figure 7

The convergence curves of IGD results for LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV, ATLMOEA, LERD and MPSLMOCSO.

Table 8 displays that MPSLMOCSO excels in achieving superior performance in 26 out of the 36 test instances. All algorithms yield an HV value of zero for LSMOP3 with 2 objectives and decision variables of 500 and 1000 dimensions. The same is true for the 2-objective LSMOP7. This indicates that none of them could generate any satisfactory solutions in the final generation. Table 8 also indicates that LSTPA, SSCEA and LERD perform similarly to MPSLMOCSO on 4 test instances, both IM-MOEA/D and ATLMOEA perform similarly to MPSLMOCSO on 5 test instances, while DGEA and FDV algorithms perform similarly to MPSLMOCSO 6 test problems. LSTPA, SSCEA and LERD are less effective than MPSLMOCSO on 32 test instances, IM-MOEA/D is less effective than MPSLMOCSO on 31 test instances, both DGEA and ATLMOEA are inferior to MPSLMOCSO on 30 test instances. FDV is inferior to MPSLMOCSO on 27 test problems. FDV and ATLMOEA are superior to MPSLMOCSO on 3 and 1 test instances, respectively.

Tables 9 and 10 depict the statistical outcomes of IGD and HV values of different algorithms on WFG.

Table 9 indicates that MPSLMOCSO is more competitive on WFG compared to other algorithms, achieving 29 best solutions in terms of average IGD out of 36 test cases. LSTPA performs better on WFG9 with 2 objectives and 1000-dimensional decision variables. It is evident that the IGD values of MPSLMOCSO and LSTPA do not differ significantly. FDV mainly excels on WFG4. The last row of the table shows that SSCEA and ATLMOEA are inferior to MPSLMOCSO on 36 test instances. This is primarily due to the fact that SSCEA relies on DVA for variable classification, which may result in imprecise classification in complex high-dimensional search spaces when using perturbation methods. ATLMOEA uses neural network training for information retrieval, but the accuracy of information from limited training sessions cannot be guaranteed. DGEA outperforms MPSLMOCSO on 1 test instance, lags behind on 30 test instances, and performs similarly on 5 instances. IM-MOEA/D outperforms MPSLMOCSO on only 1 instance and is inferior to MPSLMOCSO on 35 test instances. FDV and LERD outperform MPSLMOCSO on 6 and 1 test instances respectively, but fall short on 25 and 34 test instances, and perform similarly on 5 and 1 instances respectively.

Table 10 shows that MPSLMOCSO achieves 23 best solutions in terms of average HV out of 36 test cases. Additionally, MPSLMOCSO performs well on WFG2, WFG3, WFG7 and WFG8. SSCEA does not achieve optimal values on all test cases, as the adopted method of variable analysis makes it difficult to correctly classify variables, especially when tackling intricate problems involving high-dimensional variables. According to the test results, the performance of LSTPA, DGEA, SSCEA, IM-MOEA/D FDV, ATLMOEA and LERD is lower than that of MPSLMOCSO in 32, 27, 36, 35, 27, 35 and 32 test cases respectively. DGEA, IM-MOEA/D, FDV and LERD exhibit better performance than MPSLMOCSO in 2, 1, 7 and 1 cases, respectively. Furthermore, DGEA, FVD and LERD perform similarly to MPSLMOCSO on 7, 2 and 3 instances, while both IM-MOEA/D and ATLMOEA perform similarly on 1 test instance.

Tables 11 and 12 show the statistical outcomes of IGD and HV values of different algorithms on UF.

Table 11 demonstrates that among the several comparative algorithms, MPSLMOCSO exhibits better comprehensive performance and excels in achieving superior performance in 15 out of 20 cases. FDV exhibits outstanding performance on UF5 and UF6. The test results show that DGEA, SSCEA, IM-MOEA/D and ATLMOEA are all inferior to MPSLMOCSO on 20 test instances. This is attributed to MPSOF enhances population diversity by employing a multi-population approach, and through the optimization design of the second stage, effectively balances convergence and diversity while reducing the likelihood of encountering local optima. LSTPA and FDV outperforms MPSLMOCSO on 2 and 4 test instances, are inferior to MPSLMOCSO on 15 and 12 instances, and are similar to MPSLMOCSO on 3 and 4 instances respectively. LERD is inferior to MPSLMOCSO on 19 instances and similar to MPSLMOCSO on 1 instance.

Table 12 demonstrates that the HV of the eight algorithms is all zero on UF5, indicating that they failed to find any valid solutions for UF5. Out of 20 test problems, MPSLMOCSO achieves optimal values in 14 cases, while FDV achieves optimal values in 3. Across a large portion of the test instances, the performance of LSTPA, DGEA, SSCEA, IM-MOEA/D and ATLMOEA is consistently inferior to the proposed algorithm. Additionally, FDV and LERD outperform MPSLMOCSO on 2 and 1 test cases, underperform MPSLMOCSO on 11 and 16 test cases, and show comparable performance to MPSLMOCSO on 7 and 3 test cases respectively.

In order to visually present the algorithms’ performance more intuitively, this paper chooses two functions (containing different PF shapes) in each test suite and shows the optimized PF for each algorithm in Figs. 5 and 6. The red line denotes the true PF used for comparison, while the blue line denotes the approximate PF generated by algorithms. As depicted in Fig. 5, on the 2-objective problems, MPSLMOCSO exhibits excellent performance both in convergence and distribution. LSTPA, DGEA, FDV, and ATLMOEA produce uniformly distributed but non-convergent solution sets. SSCEA, IM-MOEA/D, and LERD perform poorly on UF3 as they fall into local optima. Figure 6 shows that only MPSLMOCSO obtains solution sets that more uniformly cover the true PFs on LSMOP4 and WFG8. DGEA and FDV exhibit marginally lower performance compared to MPSLMOCSO, while the other algorithms exhibit poorer distribution and convergence. The optimal PFs achieved by MPSLMOCSO and FDV on UF9 do not fully encompass the true Pareto front, yet they still possess a notable advantage over the other algorithms, which are entirely trapped in local optima.

Figure 8
figure 8

The convergence curves of SP results for LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV, ATLMOEA, LERD and MPSLMOCSO.

Furthermore, the convergence curves of IGD results for eight large-scale MOEAs on different test sets are plotted in Fig. 7. It is obvious that the convergence speed of IM-MOEA/D is slower compared to other algorithms, and it fails to reach a satisfactory level at the end of evolution. FDV and LERD, which are respectively based on problem transformation and reconstruction-based DVA, exhibit faster convergence than MPSLMOCSO in the early stages of evolution. However, their convergence slows down significantly compared to MPSLMOCSO in the later stages, resulting in MPSLMOCSO outperforming them at the end of evolution. MPSLMOCSO exhibits rapid convergence in the early stages, with a slower convergence rate in the mid-term, emphasizing a balance between diversity and convergence, and further converges in the later stages. Moreover, MPSLMOCSO demonstrates superiority across these problems, thereby validating the effectiveness of the proposed three-stage algorithm framework. The convergence curves of SP results of eight algorithms on three different test sets are plotted in Fig. 8, including representative problems of different types of 2-objective and 3-objective in each test set. It illustrates that at the outset of evolution, the diversity of MPSLMOCSO is inferior to that of IM-MOEA/D, which is based on inverse model decomposition, FDV, and LERD. However, in the mid-term of evolution, the SP value declines rapidly, indicating an increase in population diversity. This also demonstrates the effectiveness of the proposed multi-population approach and the hybrid weight updating strategy. At the end of evolution, MPSLMOCSO outperforms other algorithms.

Real-life application

The precision of voltage transformers’ (VTs) ratio error (RE) is essential for detecting, controlling, and protecting power systems. Therefore, VT manufacturers and users typically conduct regular calibrations and checks to ensure that their performance meets the necessary requirements. The time−varying renewable energy (TREE) estimation problem is transformed into LSMOPs in the paper40, enabling real-world issues to be extracted as bi-objective benchmark problems. The effectiveness of MPSOF in practical applications is verified by testing the TREE problem. The test results of LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV, ATLMOEA, LERD, and MPSLMOCSO are shown in Table 13. Where “–” indicates that the value is meaningless. It can be observed that the proposed algorithm performs exceptionally well across the five problems, demonstrating competitiveness when facing the TREE problem.

Table 13 The statistical HV results with LSTPA, DGEA, SSCEA, IM-MOEA/D, FDV, ATLMOEA, LERD and MPSLMOCSO on 2-objective TREE test problems.

Computational complexity

The complexity of initializing the population in MPSOF is O(ND). In the first and second stages, the complexity is mainly determined by the chosen optimization algorithm, variable grouping method and weighted optimization. Taking the LMOCSO algorithm as an example, its algorithmic complexity is O((13/4+9/2M)N2 + (3M/2 + D)N). MPSOF uses an ordered grouping method, which has a complexity of O(Dlog(D)). The complexity of weighted optimization depends on the chosen optimization algorithm. The population is updated using optimized weights in both stages, with complexities of O((N+D)q) and O((1+N+3D)q), respectively. In the third stage, the population is updated as a whole using the chosen optimization algorithm, and the complexity is consistent with that of the selected algorithm. Therefore, the overall complexity of MPSOF can be translated as O((13/4+9/2M)N2 + (3M/2 + D)N + Dlog(D) + (N+D)q).

The running times of the three algorithmic frameworks on NSGAII on the second type of experiments are plotted in Fig. 9. It can be observed that LS-NSGAII exhibits some advantages over MPSNSGAII due to its optimization acceleration by combining weights with reference directions. However, the test results in Tables 3 and 4 show that this bidirectional search capability is weak and struggles to comprehensively search the huge solution space. The running time of MPSNSGAII is lower than that of WOF-NSGAII for all problems except for the LSMOP2 and LSMOP4 problems. This is because MPSNSGAII removes duplicate weights during optimization, reducing the duplication of updating individuals and the time spent evaluating duplicate individuals. Overall, MPSOF achieves a good trade-off between efficiency and performance compared to the two algorithmic frameworks.

Figure 9
figure 9

The running time of WOF, LSMOF and MPAOF on NSGAII.

Conclusion

This paper introduces a multi-population multi-stage adaptive weighting framework designed specifically for solving LSMOPs. Based on the traditional weighted optimization framework, MPSOF divides the entire optimization process into three stages. Initially, the first stage adopts the traditional weighted update method to maintain a satisfactory convergence rate for the algorithm. Subsequently, the second stage designs an adaptive mixed individual update strategy, enhancing individual diversity and convergence. Lastly, the third stage incorporates a reasonable optimization algorithm to optimize the global variables, thereby improving the accuracy of solutions.

By conducting comparative analysis on the results of three types of experiments, it is verified that MPSOF effectively upgrades the capabilities of the original algorithms and outperforms the traditional algorithmic frameworks. Furthermore, when compared to some advanced optimization algorithms, it still shows good properties on most LSMOPs without using the optimal optimization algorithm as the optimizer.

The experimental findings also demonstrate that MPSOF does not obtain optimal solutions for all problems, which is in line with the “no free lunch principle”. Compared to WOF, the algorithm in this paper adds two parameters \(\theta\) and \(\delta\) for three-stage switching. This paper uses parameter sensitivity analysis to obtain a compromise solution. In practical problem, the chosen parameters are not optimal and lack adaptivity. Finding a solution for designing adaptive parameters according to problem characteristics and population dynamics in the evolutionary process remains an unresolved challenge for future research.