Introduction

Several real-world applications, e.g., neural architecture search [64] and aerodynamic design [17], require optimization of expensive-to-evaluate objectives, where the objective values are calculated using computationally expensive simulations [44]. Considering vehicle structure optimization [39] as an example, the evaluation of a single design using a crashworthiness simulation takes 20 h [39]. For such expensive optimization problems (EOPs), the number of function evaluations (FEs) is restricted due to limited budgets of computational and financial resources. Consequently, the main challenge of EOPs is to obtain acceptable solutions under the restricted number of FEs. For this challenge, sample-efficient approaches, such as Bayesian optimization, that reduce the number of FEs are prevalent. Although EOPs are encountered for both single-objective and multi-objective optimization domains, this paper focuses on single-objective EOPs and denotes them as EOPs for simplicity.

Over the last two decades, various sample-efficient approaches have been developed based on evolutionary algorithms (EAs) [13]. The main motivation behind this pursuit is that typical EAs often assume hundreds of thousands of FEs and thereby become impractical in terms of solving EOPs. Several sample-efficient approaches consider very restricted FE budgets, usually with a few thousand or even hundreds of FEs. In addition to such extreme cases, there exist many EOP instances using low-cost simulations, where budgets are moderately restricted to, e.g., several thousands of FEs. For example, in the automatic calibration of watershed models [54], a maximum of 10,000 FEs are used, and the corresponding evaluation of one solution using the Soil and Water Assessment Tool takes at least 2 min. However, such EOPs with moderately restricted budgets, referred to as moderately EOPs in this paper, have not undergone adequate systematic research compared to ones with very restricted budgets.

Surrogate-assisted EAs (SAEAs) [8, 16] are a popular sample-efficient approach for solving EOPs. Surrogate models of the objective function are constructed using machine learning (ML) techniques, and SAEAs utilize them to identify expected-to-improve solutions. For example, several works have proposed SAEAs using Bayesian optimization, where EAs are used to optimize the Expected Improvement (EI) metric [18]. SAEAs have been proven to be effective when the number of FEs is a few thousand or in the order of hundreds [13]. However, application of SAEAs to moderately EOPs in practice experiences the following difficulties.

  • Other than their excellent performance corresponding to a few thousand of FEs, SAEAs tend to struggle with premature convergence when a higher number of FEs are considered [47, 60]. This problem has been highlighted more clearly in complex function landscapes [24].

  • Owing to the aforementioned difficulty, tuning the parameter configurations of SAEAs, which govern their performance, is important [26, 32, 40]. However, advance fine-tuning under restricted FE budgets is usually hindered in EOPs.

  • Most SAEAs are time-consuming as they repeatedly construct and reuse ML models during a run. Their runtimes are often not ascribed much importance, relying on an assumption that they are negligibly smaller than the computational times required by simulations [5, 10]. However, reducing the runtime becomes crucial when using low-cost simulations [5].

Note that the second difficulty is not only applicable to SAEAs but also EA-based algorithms; it is known that EA performances depend significantly on their parameter configurations [27]. Thus, despite the great success of SAEAs, development of high-performance, auto-tunable, and computationally efficient algorithms is essential.

Incorporating automatic parameter-tuning mechanisms into EAs is an effective approach to improve performance while avoiding manual parameter-tuning [20]. Adaptive EAs [19] are a popular paradigm in this regard—their parameter configurations are controlled during a run. For example, jDE [3] controls two parameters used in the differential evolution (DE) [46] algorithm—the scaling factor and the mutation rate. In addition to these two parameters, SaDE [42] controls mutation and crossover strategies of DE. Several works have proposed various adaptive EAs and established that they outperform standard EAs, with comparable runtimes [3, 23, 61].

Accordingly, adaptive EAs can be used to develop computationally efficient and auto-tunable methods for solving moderately EOPs. However, most adaptive EAs are not designed for this purpose as they usually assume hundreds of thousands of FEs [23, 41]. In particular, the adaptive EA, in its basic form, is subject to the following limitations when it is extended to an EOP. Most adaptive EAs are designed to update new parameter configurations based on ones that generate good solutions in past generations, and updated parameter settings are employed without validating their effectiveness in advance. This may require extensive trial-and-error to identify good parameter configurations and thereby requires a high number of FEs [29]. Further, most adaptive EAs are designed to assign different parameter configurations to each solution in a population, i.e., so-called individual-based adaptation. However, this adaptation style may be less effective in reliably identifying a good parameter configuration, because the effectiveness of each configuration is usually validated with respect to only one sample.

This paper presents a novel adaptive EA as a computationally efficient and auto-tunable approach for solving moderately EOPs. The presented algorithm, an emulation-based adaptive DE (EBADE), is based on the DE framework and does not utilize surrogate models. However, to improve sampling efficiency, EBADE is designed to emulate the principle of sample-efficient approaches, such as those in SAEAs, by controlling the parameter configurations. In particular, the following two strategies are involved in EBADE. First, a prior validation process is introduced to pre-screen candidate parameter configurations before use. In this process, as in EI-based sampling, candidates that likely generate “expected-to-improve” solutions are selected without using FEs. This intends to prevent the use of less-effective parameter configurations and thereby reduce the number of FEs. Second, EBADE employs a subpopulation-based adaptation, in which a parameter configuration is assigned to each subpopulation, rather than to each solution. Thus, each parameter configuration is used to produce multiple solutions to update its corresponding subpopulation. This intends to validate the effectiveness of parameter configurations using multiple samples carefully. That is, in EOP, parameter configurations should be evaluated by multiple samples. The main contributions of this work are as follows:

  • To the best of our knowledge, this is the first attempt to develop an adaptive EA for moderately EOPs. This paper contributes to the development of computationally efficient and auto-tunable approaches for EOPs.

  • An adaptation mechanism for parameter configurations effective under restricted FE budgets is introduced. Its effectiveness is validated by comparing EBADE with not only popular adaptive DEs but also state-of-the-art SAEAs.

Note that our previous work [37] presented an early investigation on the prior validation mechanism and integrated it into two popular adaptive DEs—jDE and SaDE. However, the prior validation process was designed to pre-screen candidate parameter configurations based on their ability to reproduce the current best solution, which suffers from premature convergence. Further, it was applicable only to individual-based adaptation frameworks like jDE and SaDE. Consequently, the prior validation-based jDE and SaDE underperform in the case of moderately EOPs compared to state-of-the-art SAEAs. EBADE extends the prior validation process to emulate EI-based sampling on the subpopulation-based adaptation framework.

The remainder of this paper is organized as follows. Section 2 describes the standard DE framework, as well as possible parameters and genetic operators considered as primary options in the present study. The latter half of Sect. 2 presents a literature review. Section 3 presents the detailed mechanism of EBADE. Section 4 reports experiments conducted using the CEC 2013 real-parameter single-objective benchmark function suite [25]. We compare the performances of EBADE and popular adaptive DEs as well as state-of-the-art SAEAs to investigate their effectiveness for moderately EOPs. Section 5 discusses the EBADE algorithm. Finally, Sect. 6 presents our conclusions and prospective directions of future research.

Background

This section describes the DE algorithm as background information. Subsequently, related works are summarized.

Differential Evolution

DE is a population-based evolutionary algorithm for solving a real-parameter optimization problem using a single-objective function, \(f:\mathbb {R}^D\rightarrow \mathbb {R}\), where D denotes the problem dimension. In this paper, we consider the minimization of f, where the search space \(\mathcal {X}\) is bounded by \(\mathcal {X} \in [l_j,u_j]_{j=1}^{D}\).

During initialization, DE produces and then evaluates N initial solutions, forming an initial population \(\mathcal {P}=\{\varvec{x}_{i}\}^{N}_{i=1}\). Each initial solution is uniformly sampled from the search space \(\mathcal {X}\). Subsequently, as the main loop, all solutions in \(\mathcal {P}\) are updated via the following procedures. For each solution \(\varvec{x}_{i}\), a mutant solution \(\varvec{v}_{i}\) is produced using a defined mutation strategy with a scaling factor \(F \in [0,1]\). Note that the original paper considers \(F \in [0,2]\) [46], but \(F \in [0,1]\) is typically assumed [52]. Table 1 summarizes popular mutation strategies, where \(\varvec{x}_{r_{1}}\), \(\varvec{x}_{r_{2}}\), \(\varvec{x}_{r_{3}}\), \(\varvec{x}_{r_{4}}\), and \(\varvec{x}_{r_{5}}\) denote mutually exclusive solutions randomly selected from \(\mathcal {P}\) and different from \(\varvec{x}_{i}\); \(\varvec{x}_{ best }\) denotes the current best solution in \(\mathcal {P}\); and \(\varvec{x}_{p best }\) denotes a randomly selected solution from the top \(\lfloor N \times p \rfloor \) solutions in \(\mathcal {P}\) where \(p \in [0,1]\) is a hyperparameter to define greediness [63]. Next, a trial solution \(\varvec{u}_{i}\) is generated by applying a defined crossover strategy to \(\varvec{x}_{i}\) and \(\varvec{v}_{i}\) with a crossover rate \( CR \in [0,1]\). Typically, binomial or exponential crossover strategies are used, which are described in Algorithms 1 and 2, respectively, where rand[0, 1] denotes a random value in [0, 1] sampled from a unified distribution. After generating trial solutions for all the solutions, they are evaluated using f. If \(\varvec{u}_{i}\) is not worse than \(\varvec{x}_{i}\), i.e., \(f(\varvec{u}_{i}) \le f(\varvec{x}_{i})\), \(\varvec{x}_{i}\) is updated using \(\varvec{u}_{i}\). These procedures are repeated until the termination criteria are satisfied.

Table 1 Popular mutation strategies for DE
Algorithm 1
figure a

Binomial crossover

Algorithm 2
figure b

Exponential crossover

Table 2 List of adaptive and surrogate-assisted DEs for single-objective optimization

Related works

This subsection first reviews popular adaptive DEs. Then, surrogate-assisted DEs are introduced. Finally, the position of EBADE in this context is discussed.

Table 2 summarizes the related works discussed below. In the “Adaptation style” column, related works are categorized into two classes—Indiv. and Subpop. corresponding to individual-based and subpopulation-based adaptation, respectively. For algorithms without adaptation of any parameter configuration, the entry in this column is set to “–”. The problem dimension and maximum number of FEs adopted in the experiments are listed in the columns “D” and “\( FE _{\max }\)”, respectively, to indicate the type of problem addressed.

Adaptive DEs

Individual-based adaptation

In the jDE [3] framework, each solution is paired with specific values of F and \( CR \), i.e., individual-based adaptation. This may potentially provide suitable parameter configurations for a particular solution. The hyperparameters F and \( CR \) are randomly sampled once again with predefined probabilities \(\tau _F\) and \(\tau _{ CR }\). They can then be updated based on comparisons between a solution and a trial solution in terms of fitness. This comparison-based adaptation utilizes the algorithm characteristic of the DE framework [52]. Several branches of jDE have been proposed, including FDSADE [53] and ISADE [15], which adaptively control both \(\tau _F\) and \(\tau _{ CR }\). JADE [63] explored another paradigm of the sampling method of the hyperparameters utilizing probabilistic distributions determined using previous information concerning superior solutions, which is different from uniform random sampling used in jDE. JADE revealed the impact of the sampling method on adaptive DEs, and several subsequent variants have incorporated the concept of JADE, e.g., MDE_pBX [14] and SHADE [50]. SHADE is a highly popular variant among adaptive DEs that updates probabilistic distributions using success-history memories. This mechanism improves the robustness of JADE by maintaining a diverse set of parameters of probabilistic distributions. Modern approaches, including L-SHADE [51] and jSO [4], utilize the SHADE framework to control F and \( CR \). They also control N using linear population size reduction, which promotes a transition from exploration to exploitation with the progression of the search phase. Additionally, jSO employs fine-tuned scheduling-based adaptation.

Some adaptive DEs control both genetic operators and hyperparameters as parameter configurations to improve the capacity to specialize the framework to a given problem. However, this suffers from the drawback of increased complexity due to the increase in the number of parameter configurations requiring adaptation. One possible approach to address the bottleneck involves predefined sets of parameter configurations, e.g., SaDE [42] defines a limited number of pairs of mutations and crossovers, e.g., rand/1/bin and rand-to-best/2/bin. Similarly, CoDE [55] prepares predefined parameter sets \(\{F, CR \}\) as well as pairs of genetic operators. EPSDE [35] adapts parameter configurations selected from two pools of hyperparameters and genetic operators. Recently, the selection of the best mutation strategy in adaptive DEs from all candidates has been studied. CSDE [48] and AL-SHADE [23] are state-of-the-art adaptive DEs that demonstrate that adaptation of mutation strategies remains important while proposing new mutation strategies. DE-DDQN [45] and FLDE [49] primarily adapt to mutation strategies via reinforcement learning and random forest, respectively, but these ML-based methods require long computational times.

Subpopulation-based adaptation

Recent studies have integrated a multi-population scheme into the DE framework to utilize the divide-and-conquer strategy. They divide the population into multiple subpopulations, with each searching a different area. When the adaptation mechanism is included in a multi-population scheme, different parameter configurations of DE are assigned to subpopulations. For example, two populations are used for exploitation and exploration in the study [31]. In MPEDE [58], three defined parameter configurations are paired with corresponding subpopulations, where each subpopulation size is adaptively controlled depending on the current search dynamics. Moreover, some works allocate different adaptation mechanisms of existing adaptive DEs to subpopulations to further improve performance by combining various adaptation mechanisms. For instance, a hybrid amalgamation of CoDE and JADE was proposed as HMJCDE [22], and EDEV [59] incorporates EPSDE, CoDE, and JADE.

Surrogate-assisted DEs

Surrogate-assisted DEs without adaptation of DE parameter configurations

In EOPs, especially under severely restricted FE budgets, SAEAs comprise one of the most popular approaches. We now introduce some surrogate-assisted DEs. CADE [30] uses a support vector machine (SVM) as the classification model to screen solutions, reducing the number of FEs. CRADE [28] combines two SVMs to approximate the objective function and classify solutions, compensating for the weaknesses of both models. GPEME [26] is a standard SAEA, which reduces dimension using the Sammon mapping when the problem dimension exceeds 30. Then, it generates offspring solutions and evaluates only one solution that is “expected-to-improve” using the lower confidence bound metric obtained by the Kriging model constructed using recently evaluated solutions. ESAO [57] combines the global radial basis function network (RBFN) and the local Kriging model. RBFN is used to roughly select a search region and the elaborate search is conducted using the Kriging model. SAHO [40] adaptively selects optimizers from DE and TLBO [43] while screening solutions using RBFN to design diverse searches. Recently, SAEAs that adaptively select RBFNs with different configurations have been proposed, e.g., DSS-DE [32] and SADE-ATDSC [38], since the performances of SAEAs depend on parameter configurations of both EAs and MLs.

Surrogate-assisted DEs with individual-based adaptation

Some surrogate-assisted DEs incorporate adaptive DE mechanisms into SAEAs to inherit the advantages of both EAs and SAEAs. However, most of these algorithms use individual-based adaptation—SAEAs with subpopulation-based adaptation have not been developed. For example, DE-AEC [62] uses RBFN to screen solutions and adapts F and \( CR \), as in jDE. S-JADE [7] enhances the performance of JADE using RBFN. Like ESAO, S-JADE [7] utilizes two types of RBFNs; global and local RBFNs are constructed using all evaluated solutions and neighborhood solutions of the current population, respectively.

Alternatively, DE parameter configurations have also been adapted while screening solutions. This produces various solutions to be screened by varying DE parameter configurations. SMDE [21] uses four mutation strategies and generates the same number of offspring solutions via mutation strategies for each base solution. Then, the solution with the best predicted fitness is adopted. Thus, SBSM-DE considers four candidate solutions generated in different ways, although FE is conducted only once per base solution. Other examples are as follows. DESSA [29] is used with CoDE or SaDE. For example, DESSA-CoDE has three sets of parameters and three mutation strategies, i.e., nine pairs of parameter configurations are considered as adaptation candidates. After generating the nine corresponding trial solutions, rank-SVM is used as a surrogate model for screening. SMA-EPSDE [33] and ESMDE [34] are derivatives of EPSDE and use two mutation strategies (rand/1 and current-to-rand/1), two crossover strategies (binomial and exponential), \(F\in [0.5,1.0]\), and \( CR \in [0.0,1.0]\) as candidate parameter configurations. They randomly sample parameter configurations and generate solutions until the approximated fitness exceeds the fitness of base solutions. The difference between SMA-EPSDE and ESMDE comprises the dataset selection criteria for the Kriging model. Sa-DE-DPS [11], SAPDE-ANN [1], and SAPDE-RSM [1] conduct a search on the approximation function for a certain number of generations while adapting parameter configurations to accelerate the search.

Position of EBADE in this context

According to Table 2, most adaptive DEs are tested with \(D \times \) 10,000 FEs, while many surrogate-assisted DEs are designed for FEs not exceeding 1000. Some surrogate-assisted DEs which screen solutions and adapt parameter configurations simultaneously use \(D \times \) 10,000 FEs, but real-world EOPs with more than 10,000 FEs are rarely encountered [2, 9, 12, 17, 21, 32, 36, 56, 57]. Thus, we investigate if EBADE improves performance within 10,000 FEs, i.e., for moderately EOPs.

As noted in the “Adaptation style” column in Table 2, most existing adaptive DEs use individual-based adaptation, while subpopulation-based adaptation is only being studied recently. In this study, EBADE adapts parameter configurations using subpopulation-based adaptation, without using surrogate models. In the experiments presented in Sect. 4, we compare EBADE with adaptive DEs and SAEAs. For adaptive DEs, we choose comparison algorithms based on both individual-based and subpopulation-based adaptation to investigate the effectiveness of EBADE compared to both typical and modern categories.

The prior validation mechanism of EBADE introduced in the next section has never been discussed in the existing literature. In contrast, most adaptive EAs update parameter configurations by referring to those that have generated good solutions in past generations.

Proposed algorithm

The proposed algorithm, EBADE, is an adaptive DE for solving EOPs. In contrast to typical adaptive EAs, a prior validation process is utilized to identify good DE parameter configurations before using them. This accelerates the evolutionary search while reducing the number of FEs by avoiding the use of less effective parameter configurations. Here, we first discuss the idea underlying our prior validation process, i.e., the process of estimating the effectiveness of parameter configurations in advance, and subpopulation-based adaptation. Subsequently, the detailed algorithms of EBADE are explained.

Concept

Usually, the effectiveness of algorithmic parameter configurations cannot be estimated without using them, because it depends on the current search dynamics and random factors. However, the idea of EBADE is to produce good parameter configurations by emulating the sampling mechanism of SAEA without consuming FEs. Specifically, EBADE comprises the prior validation mechanism and subpopulation-based adaptation. Our idea is based on two insights summarized below.

  • Generally, SAEAs first generate multiple candidate solutions and estimate the quality of candidate solutions, and then remove undesired ones to reduce FEs. When estimating quality, SAEAs often utilize the improvability of candidate solutions, e.g., the EI metric. However, EBADE cannot estimate the quality of candidate solutions directly due to the lack of surrogates. Instead, EBADE utilizes parameter configurations that directly affect solution generation. By emulating the SAEA mechanism, EBADE generates multiple parameter configurations and estimates their quality. Then, without consuming FEs, EBADE removes less effective parameter configurations before generating solutions for FE. Emulating the idea of the EI metric, EBADE selects a candidate parameter configuration that is likely to generate “expected-to-improve” solutions. Specifically, the superior solution mentioned in the preceding paragraph is set to the solution with the best fitness improvement ratio (FIR). Since FIR represents the degree of improvement in the fitness value, the solution with the best FIR may not necessarily coincide with the solution with the best fitness value, preventing the latter from being chosen even if it is stuck in a local optimum. Also, by selecting the solution with the best FIR, the various solutions that are progressing will continue to be selected even if some solutions in the population stagnate. Thus, our prior validation mechanism is expected to enable EBADE to guide solutions to the “expected-to-improve” area, transcending premature convergence to the local optima.

  • Although testing or validation data consist of multiple samples in the ML domain, most existing adaptive DEs utilize individual-based adaptation. In other words, a parameter configuration is validated using only one solution in the existing adaptive DEs. Thus, individual-based adaptation may be inefficient in EOPs, because a large number of FEs are consumed over multiple generations to gain a large number of samples for a parameter configuration during its validation. Accordingly, EBADE employs subpopulation-based adaptation, where a parameter configuration is paired with multiple solutions, i.e., all solutions in a subpopulation. This improves the validation accuracy of the effectiveness of parameter configurations, reflecting the results of FE by narrowing down the number of parameter configurations and validating each with respect to multiple samples. Additionally, subpopulation-based adaptation is also vital for the prior validation mechanism as one candidate parameter configuration is validated with respect to multiple samples in the prior validation phase. The subpopulation-based scheme can mitigate the possibility of a sample moving in an unintended direction due to random numbers, resulting in an unjustified evaluation of parameter configurations.

Consequently, EBADE is designed to improve the efficiency of adaptation using the prior validation mechanism and subpopulation-based adaptation, emulating sample-efficient approaches such as SAEAs without using any surrogate.

Parameter configuration vector

In EBADE, four algorithmic parameter configurations, the mutation strategy, the crossover strategy, the scaling factor, and the crossover rate are controlled during a run. Let \(\varvec{\theta } = [\theta _{v}, \theta _{u}, \theta _{F}, \theta _{ CR }]\) be a parameter configuration vector that defines the DE parameter configuration, where each variable in \(\varvec{\theta }\) is defined as follows:

  • \(\theta _{v} \in \{1, 2, 3, 4\}\) specifies the index of the mutation strategy to be used. EBADE uses four mutation strategies, best/1, current-to-best/1, current-to-pbest/1, and rand-to-best/1, indexed by 1, 2, 3, and 4, respectively.

  • \(\theta _{u} \in \{1, 2\}\) specifies the index of the crossover strategy to be used. Binomial and exponential crossover strategies are used and indexed by 1 and 2, respectively.

  • \(\theta _{F} \in [0,1]\) indicates the specific value of the scaling factor F used in the mutation strategy specified by \(\theta _{v}\).

  • \(\theta _{ CR } \in [0,1]\) indicates the specific value of the crossover rate \( CR \) used in the crossover strategy specified by \(\theta _{u}\).

For example, given \(\varvec{\theta } = [1, 2, 0.5, 0.8]\), a DE algorithm uses best/1 mutation with \(F=0.5\) and exponential crossover with \( CR =0.8\). As the mutation strategy, we choose those that accelerate convergence of the DE population the most [6], since the rapid convergence of the population is prioritized during the optimization of EOPs [16].

In summary, the solution-generation procedure of DE specified with \(\varvec{\theta }\), i.e., the generation of trial solutions, is described in Algorithm 3.

Algorithm 3
figure c

Get trial solution

Overall framework

EBADE uses and subsequently optimizes M subpopulations, \(\mathcal {P}_{1}, \mathcal {P}_{2}, \dots , \mathcal {P}_{M}\), simultaneously, where each subpopulation size is set to a common value N. Each subpopulation \(\mathcal {P}_{m}\) has its own parameter configuration vector \(\varvec{\theta }_{m}\).

Algorithm 4 describes the overall procedure of EBADE, which consists of the following four components—initialization, a search phase, post hoc validation, and prior validation. To begin with, each subpopulation \(\mathcal {P}_{m}\) and its parameter configuration vector \(\varvec{\theta }_{m}\) are initialized. Next, in the search phase, DE is executed for each subpopulation with its corresponding parameter configuration. Subsequently, post hoc validation is performed to identify ineffective parameter configurations that have failed to produce good solutions during the search phase. Finally, ineffective parameter configurations are modified in the prior validation phase into plausibly good parameter configurations without consuming FEs. Subsequently, EBADE returns to the search phase with the modified parameter configurations and the three latter phases are repeated until the termination criteria are satisfied.

The rest of this section describes the detailed procedure of each phase.

Algorithm 4
figure d

EBADE 

Initialization

Initially, each subpopulation \(\mathcal {P}_{m}\) is initialized with N initial solutions. The initial solutions are produced using the same method as the standard DE (see Sect. 2.1). Accordingly, similar to existing DEs with multi-population mechanisms [22, 31, 58, 59], EBADE produces different initial subpopulations to boost the performance while preventing premature convergence. Thereafter, all initial solutions are evaluated using the objective function. Subsequently, the parameter configuration vector \(\varvec{\theta }_{m}\) for \(\mathcal {P}_{m}\) is initialized as follows. The values of \(\theta _v\) and \(\theta _u\) are set to random integer values sampled from \(\{1,2,3,4\}\) and \(\{1,2\}\), respectively. For \(\theta _{F}\) and \(\theta _{ CR }\), \(\theta _{F}=0.5\) and \(\theta _{ CR }=0.9\) are used as the default configurations of standard DE [46].

Search phase

This phase conducts evolutionary search to collect multiple validation samples and estimate the effectiveness of parameter configurations during the following post hoc validation phase. To this end, each subpopulation is updated by executing the DE solution-generation process for one generation.

Lines 6-16 in Algorithm 4 describe the procedure of this phase. A DE algorithm is executed on each subpopulation for one generation, corresponding to the update of all N solutions in \(\mathcal {P}_{m}\). First, EBADE constructs the whole population \(\mathcal {P}_{ all }\) by concatenating all subpopulations, \(\mathcal {P}_{1}, \mathcal {P}_{2}, \dots , \mathcal {P}_{M}\). This population \(\mathcal {P}_{ all }\) is used as the pool of candidate solutions in the mutation strategy. Since this process also influences the discovery of superior solutions, we employ an information-sharing strategy that utilizes all possible solutions. In other words, all possible variables in \(\mathcal {P}_{ all }\), i.e., \(\varvec{x}_{r_{1, 2, 3}}\), \(\varvec{x}_{ best }\), and \(\varvec{x}_{p best }\), are used while generating a mutant solution \(\varvec{v}\) for \(\varvec{x}\in \mathcal {P}_{m}\), even if they belong to other subpopulations.

Next, all solutions in all subpopulations are updated following the original DE methodology. Specifically, for each subpopulation \(\mathcal {P}_{m}\), each trial solution \(\varvec{u}\) of \(\varvec{x}\) is generated using \(\varvec{\theta }_{m}\) and \(\mathcal {P}_{ all }\). Then, each \(\varvec{u}\) is evaluated using the objective function. If the fitness value \(f(\varvec{u})\) of \(\varvec{u}\) does not exceed that of \(\varvec{x}\), \(\varvec{x}\) is replaced with \(\varvec{u}\). During the search phase, \(N \times M\) solutions can be sampled for each \(\varvec{\theta }\); thus, exactly \(N \times M\) FEs are consumed.

Post hoc validation phase

This phase identifies good/bad parameter configurations by estimating their effectiveness based on the results of the previous search phase. Good parameter configurations are reused during the next search phase without any modification, whereas bad ones are modified during the next prior validation phase.

A straightforward way to define a good parameter configuration is in terms of its success to generate the current best solution in terms of the objective values. Although this definition was used in our previous work [37], it may easily induce premature convergence, as it does not consider the improvement of objective values. Here, we define good parameter configurations as “worth-to-continue” ones, which are identified as those generating solutions with high FIR values. Specifically, suppose a solution \(\varvec{x}_{g}\) is generated during the gth generation from its corresponding parent solution \(\varvec{x}_{g-1}\), and the FIR value for \(\varvec{x}_{g}\), denoted as \(\delta _f(\varvec{x}_g)\), is calculated as

$$\begin{aligned} \delta _f(\varvec{x}_g) = 1 - \frac{f(\varvec{x}_g)}{f(\varvec{x}_{g-1}) + \delta _C}, \end{aligned}$$
(1)

where \(\delta _C \ge 0\) denotes a constant value to avoid division by zero. Corresponding to a large value of \(\delta _f(\varvec{x}_g)\), it is expected that the parameter configuration vector \(\varvec{\theta }\) used to generate \(\varvec{x}_g\) from \(\varvec{x}_{g-1}\) can be effective to identify further good solutions during the next search phase. In contrast, corresponding to a small value of \(\delta _f(\varvec{x}_g)\), \(\varvec{\theta }\) may be ineffective as it does not contribute to discovering a good search region.

Accordingly, EBADE determines good/bad parameter configurations using the following procedures. They are described by the lines 17-21 in Algorithm 4. First, EBADE  reconstructs the whole population \(\mathcal {P}_{ all }\) by combining all subpopulations updated during the search phase. Subsequently, an FIR value for each solution \(\varvec{x}\) in \(\mathcal {P}_{ all }\) is calculated, and the top M solutions having the M highest FIR values, \(\varvec{x}^{*}_{1}, \varvec{x}^{*}_{2}, \dots , \varvec{x}^{*}_{M}\), are determined. Then, a parameter configuration vector \(\varvec{\theta }_{m}\) is identified as a good one if there is at least one top solution \(\varvec{x}^*\) generated by \(\varvec{\theta }_{m}\); otherwise, it is identified as a bad one. Technically, EBADE stores indices of the bad parameter configurations in an index set of ineffective parameter configurations \(\Theta \), which is mathematically defined as follows:

$$\begin{aligned} \Theta = \{i \in \{1, 2, \dots , M\} \mid \not \exists \varvec{x}^*_m \in \mathcal {P}_i, m=1,2,\dots , M\}. \end{aligned}$$
(2)

Note that EBADE utilizes FIR to determine good parameter configurations, but this may be hindered until solutions begin to improve to some extent after several generations. Other possible indicators may be required to detect good configurations even with a small improvement, but we will leave this for future work.

Prior validation phase

This process modifies bad parameter configurations to likely good ones before using them in the next search phase. As defined during post hoc validation, we consider good parameter configurations to be those that have generating good solutions with the top M FIR values. Using this definition, we modify bad parameter configurations to enable them to generate such top solutions.

Lines 22-25 in Algorithm 4 correspond to this phase and Algorithm 5 presents the detailed procedure of our prior validation phase. First, EBADE selects the target solution \(\varvec{x}_{ target }\) from the solutions discovered so far. This is commonly set for all bad parameter configurations \(\varvec{\theta }_{i}\) to improve the probability of generating effective parameter configurations using \(\varvec{x}_{ target }\) repeatedly. The solution with the best FIR in \(\mathcal {P}_ all \) is selected as the target solution. Note that the search direction is also guided towards the best solution as mutation strategies with high exploitation ability are utilized. Thus, the modified parameter configuration is expected to consider the directions towards the best solution and the area that remains to be searched, i.e., the best FIR solution.

For each bad parameter configuration indexed by \(i \in \Theta \), EBADE randomly generates multiple parameter configuration candidates and then removes less effective ones, only retaining one. Then, the remaining candidate is replaced with \(\varvec{\theta }_{i}\). Technically, K candidate parameter configurations \(\varvec{\theta }_{k}\), \(k=1,2,\dots ,K\) are generated following the procedure used during initialization, except for \(\theta _{F}\) and \(\theta _{ CR }\), where \(K\in \mathbb {N}\) is a hyperparameter. For each \(\varvec{\theta }_{k}\), its \(\theta _{F}\) and \(\theta _{ CR }\) are randomly selected from [0, 1]. Then, EBADE selects one candidate parameter configuration that produces a solution closest to a target solution \(\varvec{x}_{ target }\). In particular, for each candidate parameter configuration \(\varvec{\theta }_{k}\), EBADE tests the ability of \(\varvec{\theta }_{k}\) to generate solutions close to a target solution \(\varvec{x}_{ target }\). More specifically, for each solution \(\varvec{x}\) in its corresponding subpopulation \(\mathcal {P}_{i}\), its trial solution \(\varvec{u}\) is generated by the DE solution-generation process specified using \(\varvec{\theta }_{k}\), i.e., \( GetTrialSolution (\varvec{\theta }_{k}, \varvec{x}, \mathcal {P}_{ all })\). These sample solutions are not evaluated using the objective function. Then, the minimum Euclidean distance between each \(\varvec{u}\) and \(\varvec{x}_{ target }\) is recorded as d(k). Finally, the index having the minimum d(k), \(k=1,2,\dots ,K\), is set as \(k^*\), and \(\varvec{\theta }_{i}\) is replaced with the candidate \(\varvec{\theta }_{k^*}\).

Algorithm 5
figure e

Prior validation

Experiments

In this section, the performance of EBADE is evaluated on single-objective benchmark functions with restricted FE budgets. All experiments are conducted using a Intel(R) Core(TM) i7-10700 4.8 GHz CPU and 16 GB RAM.

Experimental configurations

Test problems

We used 28 bound-constrained benchmark functions, F1, F2, \(\dots \), and F28, used in the competition on Real-Parameter Single Objective Optimization at the IEEE Congress on Evolutionary Computation 2013 [25]. Note that F1–F5, F6–F20, and F21–28 are categorized as unimodal, multimodal, and composition functions, respectively (see [25] for detailed definitions). The problem dimension was set to \(D\in \{10, 20, 30\}\). The search space of all functions was commonly set to \(\varvec{x} \in [-100, 100]^D\).

Comparison algorithms

We compared the performances of EBADE and four adaptive DEs, SHADE, jSO, CSDE, and EDEV, and four SAEAs, GPEME, S-JADE, SAHO, and ESMDE. The brief descriptions and parameter configurations of comparison algorithms are summarized below.

  • SHADE, which has been the basis of modern adaptive DEs, employs the current-to-pbest/1 mutation strategy and adapts F and CR based on success history memories, \(M_{F,r}\) and \(M_{CR,r}\), respectively, using individual-based adaptation. Via comparison with SHADE, the impacts of our prior validation and subpopulation-based adaptation mechanisms are evaluated. We used \(H=100, M_{F,h,init }=0.5, M_{CR,h,init }=0.5, |Archive |=100, p_{\min }=2/N\), and \(p_{\max }=0.2\) [50].

  • jSO, which is an extension of L-SHADE [51]. Based on a specific schedule for parameter control, jSO adapts F, CR, and N for individual-based adaptation. In particular, N decreases as the number of FEs increases to evolve more generations by the end of the search. We verify if EBADE outperforms local search conducted by jSO at the end of the search in moderately EOPs. We used \(H=5, M_{F,h,init }=0.3, M_{CR,h,init }=0.8, |Archive |=N, N_{init}=25\log {D^{3/2}}, N_{\min }=4, p_{\min }=0.125\), and \(p_{\max }=0.25\) [4].

  • CSDE, which is one of the state-of-the-art adaptive DEs. CSDE adapts F, CR, and the mutation strategy for individual-based adaptation. CSDE shifts between two mutation strategies (current-to-pbest/1 and pbest-to-rand/1) depending on the degree of stagnation in the search. Comparison with CSDE reveals if EBADE outperforms state-of-the-art adaptive DEs in moderately EOPs. We used \(F_{init }=0.5\), \(CR_{init }=0.5\), \(N=100\), \(FP=200\), \(\mu =0.5\), and \(\sigma =0.1\) [48].

  • EDEV, which is a state-of-the-art subpopulation-based adaptive DE. EDEV adaptively assigns JADE, CoDE, and EPSDE to subpopulations. Thus, EDEV adapts F, CR, the mutation strategy, and the crossover strategy. Comparison with EDEV is important, because both EDEV and EBADE are subpopulation-based adaptive DEs that adapt four parameter configurations of DE. We used \(\lambda _1=\lambda _2=\lambda _3=0.1\), \(\lambda _4=0.7\), \(ng=20\), (JADE: \(\mu _{F,init}=0.5\), \(\mu _{CR,init}=0.5\), \(c=0.1\), \(p_{\min }=0.05\), \(p_{\max }=0.2\)), (CoDE:\(\{F,CR\}\in \{\{1.0,0.1\}\), \(\{1.0,0.9\}\), \(\{0.8,0.2\}\}\), \(\{\)rand/1/bin, rand/2/bin, current-to-rand/1\(\}\)), (EPSDE:\(P_{F}\) \(=\{0.4,0.5,\dots ,0.9\}\), \(P_{CR}\) \(=\{0.1,0.2,\dots ,0.9\}\), \(P_{v}\) \(=\{\)rand/1, best/2, current-to-rand/1\(\}\)) [59].

  • GPEME, which is one of the most popular and frequently compared SAEAs, employs DE and the Kriging model. Thus, GPEME is the standard of SAEA and comparison with it is essential. We used \(N=100, F=0.8, CR=0.8, \tau =100, \lambda =50, l=4, \omega =2\), regression \(=\) zero-order, correlation \(=\) Gaussian, \(\theta \in [10^{-5}, 10^2]\), and \(\theta _{init}=10^{-2}\) [26].

  • S-JADE, which is a state-of-the-art SAEA, consists of modified JADE with multiple RBFN models. By comparing with S-JADE, we compare the performances of EBADE and state-of-the-art SAEAs with DE parameter configuration adaptation in moderately EOPs. We used \(N=100\), \(F_{out}=0.5\), \(CR_{out}=0.75\), \(p_{pbest_{out}}=0.05\), \(F_{in}=0.5\), \(CR_{out}=0.5\), \(p_{pbest_{in}}=0.1\), \(std_F=0.1\), \(std_{CR}=0.1\), \(L=10\), \(\epsilon =0.01\), \(c=0.1\), \(FE_{max_{in}}=2,000\), kernel \(=\) cubic, and \(r=rand(0,\ 1.25)\) [7].

  • SAHO, which is also a state-of-the-art SAEA, employs DE and TLBO as optimizers and the RBFN model as a surrogate. In addition to SAHO being one of the state-of-the-art SAEAs, comparison with GPEME and SAHO provides a relative performance comparison of EBADE for different ML models used in these SAEAs. We used \(N=100\), \(F=0.9\), \(CR=0.5\), \(K=30\), \(neighbor=5D\), and kernel \(=\) cubic [40].

  • ESMDE, which adapts DE configuration during screening of the solution by the Kriging model. The candidate configurations are two mutation strategies (rand/1 and current-to-rand/1), two crossover strategies (binomial and exponential), \(F\in [0.5,1.0]\), and \(CR\in [0.0,1.0]\). By comparing with ESMDE, we verify if the adaptation mechanism of EBADE outperforms SAEAs with DE parameter adaptation in moderately EOPs. We used \(c=10\), regression \(=\) zero-order, correlation \(=\) Gaussian, \(\theta \in [10^{-5}, 10^2]\), and \(\theta _{init}=10^{-2}\) [34].

Parameter configuration of EBADE

The size of each subpopulation was taken to be \(N=4\), the number of subpopulations was \(M=25\), the number of candidate configurations was \(K=6\), and the parameter in the current-to-pbest/1 is \(p=0.5\). Thus, the size of the whole population \(\mathcal {P}_{ all }\) is \(N \times M = 100\), and these configurations are identical to those in the compared adaptive DEs above. Note that \(\delta _C\) in Eq. (1) was not needed, because the objective values except at the global optimum were non-negative. The objective value at the global optimum of each function was 0, but no trial reached the global optimum.

Evaluation scheme

All algorithms were forcibly terminated when the number of FEs reached its maximum budget \( FE _{\max }\), including FEs used for the initialization phase. The performance of algorithms was evaluated in terms of the best objective value discovered at \( FE _{\max }\), and their mean values over 21 independent trials were reported. The Wilcoxon signed-rank test was applied to identify significant differences with a significance level of \(p<0.05\). Additionally, the average runtimes of all algorithms to complete one trial were compared.

We set \( FE _{\max }=6000\) as a default value as moderately restricted budgets of FEs were assumed. However, additional results corresponding to 2000–10,000 FEs are presented in Sect. 4.3 to investigate the scalability of EBADE with respect to the number of FEs.

Results

Tables 3, 4, and 5 summarize the performances recorded corresponding to 6000 FEs for problems with \(D\in \{10, 20, 30\}\), respectively. The best and worst performances are highlighted in bold and italic, respectively. In the tables, “\(+\)”, “−”, and “\(\sim \)” indicate that the performance of a compared algorithm is statistically better than, statistically worse than, and comparable to that of EBADE, respectively. Further, the average rank and the overall statistical results, i.e., the counts of \(+/-/\sim \), are summarized at the bottom of each table.

Table 3 Comparison of optimization results obtained at 6000 FEs (\(D=10\), 21 trials)
Table 4 Comparison of optimization results obtained at 6000 FEs (\(D=20\), 21 trials)
Table 5 Comparison of optimization results obtained at 6000 FEs (\(D=30\), 21 trials)

As is evident from Table 3, EBADE outperformed adaptive DEs (SHADE, jSO, CSDE, and EDEV) and one SAEA (ESMDE) significantly on multiple problem instances with \(D=10\). In particular, the performance of EBADE was statistically better than those of adaptive DEs on at least 14 problem instances. In addition, EBADE exhibited comparable performance with respect to three SAEAs (GPEME, S-JADE, and SAHO). In particular, the performance of EBADE was statistically better than SAEAs on at least nine problem instances, and there were only a maximum of six statistically worse cases. Consequently, EBADE was assigned the best rank, outperforming the considered SAEAs. This indicates a benefit of auto-tunable approaches in solving EOPs—the performance of EBADE is less problem-dependent owing to adaptive control of parameter configurations during a run. Although the performance of SAEAs could be improved via fine-tuning, it is usually hindered in EOPs.

Even when D is increased to 20 and 30, EBADE remained effective, as evidenced by Tables 4 and 5. The performance of EBADE was better than those of adaptive DEs and highly competitive with SAEAs on several problem instances. For \(D=30\), the effectiveness of EBADE was slightly poorer, as GPEME outperformed EBADE on certain problems. However, EBADE was assigned the best rank, indicating that it performed well on average. All four adaptive EAs considered for comparison were assigned worse ranks than the SAEAs, except for ESMDE, as they are not designed for restricted FEs. These observations empirically corroborate the effectiveness of our emulation-based adaptation mechanism.

Table 6 presents the average algorithmic runtimes required to complete one trial corresponding to each problem dimension. EBADE required the second-highest runtime among the five adaptive DEs, but the differences were only a few seconds at most. On the other hand, EBADE outperformed the SAEAs in terms of speed; the runtime of EBADE was smaller than that of the SAEAs by at least an order of two.

In summary, overall results suggest that, without any help of surrogate models, EBADE performs comparably with SAEAs while operating much faster than them. This observation empirically suggests that the proposed emulation-based adaptation enables the adaptive DE mechanism to accelerate its evolutionary search under a restricted number of FEs, demonstrating the possibility to realize computationally efficient and auto-tunable optimizers for solving moderately EOPs.

Table 6 Average runtime [sec] required to complete one trial (\( FE _{\max }=6000\))

Additional results

To investigate the scalability of EBADE with respect to the number of FEs, we conducted additional experiments with different values of \( FE _{\max }\) under the same experimental environment as in Sect. 4.1, except for \( FE _{\max }\). Specifically, \( FE _{\max }\) was set to \(\{2000, 4000, 8000, 10{,}000\}\) in addition to its default value of 6000. Here, we report the average ranks and the overall statistical results.

Table 7 reports the statistical results summarized as the counts of \(+/-/\sim \). Figure 1 depicts the average ranks of all nine algorithms. Overall, the performance of EBADE was statistically better than those of all the four adaptive DEs even when \( FE _{\max }\) was decreased/increased from its default value 6000. The superiority of EBADE over the four adaptive DEs is also corroborated by Fig. 1, since EBADE was assigned the best rank among five adaptive DEs corresponding to all problem dimensions and all \( FE _{\max }\) cases. These results prove that EBADE was successfully adapted to EOPs.

When \( FE _{\max }\) was 2000, three SAEAs—GPEME, S-JADE, and SAHO—outperformed EBADE corresponding to all problem dimensions, as the number of “\(+\)” was sufficiently larger than that of “−”. This result clearly demonstrated the effectiveness of surrogate-assisted search on highly restricted FE budgets. However, when \( FE _{\max }\) was increased to 4000, the performance of EBADE became competitive to those of the SAEAs, except in the 30-dimensional cases. When \( FE _{\max }\) was further increased to 8000 and 10,000, the effectiveness of EBADE was highlighted even more. These tendencies can be also observed in Fig. 1—EBADE was always assigned the second or third rank when \( FE _{\max }\le 4000\), and its rank improved to first for \( FE _{\max } \ge 6000\). On the other hand, the ranks of SAEAs gradually decreased with the increase of \( FE _{\max }\); this indicates stagnation in SAEA performances. This is because SAEAs tend to struggle with premature convergence with a higher number of FEs. As observed in [47, 60], generally, surrogate-assisted searches provide a strong exploitation bias to promote evolutionary search under very restricted budgets of FEs. However, this degrades the diversity of solutions with a higher number of FEs, degrading the search performance and the quality of (global) surrogates. Consequently, as shown in Fig. 1, SAEAs performed very well for very restricted budgets of FEs; however, their performances gradually degraded with the increase of FEs. EBADE was assigned to the best rank for all dimensions when \( FE _{\max } \ge 6000\), as shown in Fig. 1. Thus, this observation indicates that EBADE is well scaled to the increase of the problem dimensions D on moderately EOPs. A possible reason for this is that EBADE can possess a better diversity of solutions than that of SAEAs; after the adaptation of parameter configurations, EBADE generates solutions similar to the standard DEs without any pre-screening process using surrogates.

The following differences were identified between the different algorithmic mechanisms. The relatively simple algorithm of SHADE ensured its scalability with respect to D, since the average rank increased slightly as D increased, as in Fig. 1. jSO also exhibited improved performance—it ranked second among adaptive DEs and its average rank was slightly higher than those of all SAEAs when \( FE _{\max } \ge 8000\) \((D=10)\) and \( FE _{\max } =\) 10,000 \((D=20)\), as in Fig. 1. This was attributed to the mechanism by which jSO reduces the population size towards each set \( FE _{\max }\) to gain the number of solution evolution. If this mechanism were incorporated into EBADE, it may further improve its performance. However, EBADE already outperformed SHADE and jSO based solely on the prior validation mechanism and subpopulation-based adaptation. As depicted in Fig. 1, the average rank of CSDE increased with an increase in \( FE _{\max }\) for all dimensions, but the statistical test results obtained in comparison with EBADE did not improve, as presented in Table 7. Thus, the state-of-the-art CSDE is specialized for cases with plentiful FE budgets. EDEV seemed to consume a large number of FEs to improve performance, because it has a huge search space of parameter configurations. In comparison with SAEAs, EBADE was highly competitive beyond the type of ML used in compared SAEAs, since its rank was higher than those of GPEME and SAHO, at least for \( FE _{\max } \ge 6000\), as in Fig. 1. Although S-JADE and ESMDE adaptively control DE parameter configurations, further consideration may be required in the combined implementation of SAEAs and adaptation methods to solve moderately EOPs effectively, especially owing to the low rank of ESMDE in Fig. 1.

In summary, the compared SAEAs exhibited excellent performances when \( FE _{\max } = 2000\), but EBADE derived good performances when \( FE _{\max } \ge 6000\). This observation shows the effectiveness of EBADE on moderately EOPs.

Table 7 Statistical results (the count of \(+/-/\sim \)) for 2000, 4000, 6000, 8000, and 10,000 FEs
Fig. 1
figure 1

Average rank of all nine algorithms over the number of FEs

Discussion

In this section, the results of ablation studies are reported to confirm the effectiveness of the main components of EBADE. First, we discuss the effect of controlling the DE parameter configurations in moderately EOPs. Subsequently, the effectiveness of the prior validation and subpopulation-based adaptation is discussed. Finally, the adaptation results of the DE parameter configurations are investigated. Experimental environments identical to those reported in Sect. 4.1 are used with \( FE _{\max }\in \{2000, 4000, 6000, 8000, 10,000\}\) unless stated otherwise.

Impact of parameter adaptation in moderately EOPs

We empirically demonstrate that controlling the DE parameter configuration is crucial to improve the performance of adaptive DEs even for moderately EOPs. To this end, we compared the performances of EBADE and various DEs with fixed parameter configurations. In particular, we introduced eight DE variants with different combinations of mutation and crossover strategies; best/1/bin (b/1/b), current-to-best/1/bin (cb/1/b), current-to-pbest/1/bin (cpb/1/b), rand-to-best/1/bin (rb/1/b), best/1/exp (b/1/e), current-to-best/1/exp (cb/1/e), current-to-pbest/1/exp (cpb/1/e), and rand-to-best/1/exp (rb/1/e), where bin and exp denote binomial and exponential crossover, respectively; and \(F=0.5\) and \(CR=0.9\) were used for all variants. For cpb/1/b and cpb/1/e, we used \(p=0.5\).

Table 8 Statistical results (the count of \(+/-/\sim \)) for 2000, 4000, 6000, 8000, and 10,000 FEs in comparison with eight DE variants with fixed parameter configurations

Table 8 summarizes the statistical results in terms of the counts of \(+/-/\sim \), where “\(+\)”, “−”, and “\(\sim \)” indicate that the performance of a DE variant is statistically better than, statistically worse than, and comparable with that of EBADE, respectively. Further, Fig. 2 depicts the average ranks. In Table 8, two DE variants using cb/1/b and rb/1/b were observed to outperform EBADE on several problem instances, with \( FE _{\max } \le 4000\) as a sufficiently large number of “\(+\)”. However, when \( FE _{\max }\) was increased to 6000, the performance of EBADE was gradually improved. When \( FE _{\max }\) was further increased to 8000 and 10,000, EBADE statistically outperformed all DE variants on at least 13 problem instances. Almost an identical trend was observed in Fig. 2. Specifically, although the average rank of EBADE was sometimes below those of cb/1/b and rb/1/b for \( FE _{\max } \le 4000\), EBADE always ranked first for \( FE _{\max } \ge 6000\).

These results indicate that the configurations of cb/1/b and rb/1/b, i.e., providing a strong bias for exploitation, enhanced DE performance when \( FE _{\max }\) was highly restricted; however, this benefit may be less important as SAEAs performed well under such restricted budgets. An important drawback of using these settings was observed to be premature convergence when \( FE _{\max }\) was increased to, for example, 6000 or even 10,000. Actually, the average ranks of cb/1/b and rb/1/b were rapidly degraded as \( FE _{\max }\) was increased, as depicted in Fig. 2. In contrast, EBADE performed well by controlling the DE parameter configurations, especially for \( FE _{\max } \ge 6000\), i.e., under moderately restricted FE budgets.

General trends of fixed DE variants are as follows. When \( FE _{\max }\) was increased from 2000 to 10,000, the average ranks of DE variants using the binomial crossover strategy, except for cpb/1/b, i.e., b/1/b, cb/1/b, and rb/1/b, continued to decrease, while those of cpb/1/b or DE variants using the exponential crossover strategy, i.e., b/1/e, cb/1/e, cpb/1/e, and rb/1/e, continued to improve. EBADE was assigned consistently to good ranks. This result shows that the adaptation of the DE parameter configurations was effective in solving moderately EOPs.

Fig. 2
figure 2

Average ranks of fixed DE variants and EBADE over the number of FEs

Impact of adaptation of DE parameters

EBADE is designed to simultaneously adapt the DE parameter configurations (\(F, CR \), and the mutation and crossover strategies) to determine good parameter configurations from a variety of candidates. We further validate the effectiveness of this strategy by comparing EBADE with its variant using fixed values of \({\theta }_{F}\) and \({\theta }_{ CR }\), denoted as EBADE-fixFCR.

Specifically, EBADE-fixFCR is designed to always use \({\theta }_{F}=0.5\) and \({\theta }_{ CR }=0.9\) [46]; it tunes only the mutation and crossover strategies. Table 9 presents the statistical results summarized as the counts of \(+/-/\sim \) with \( FE _{\max } \in \{2000, 4000, 6000, 8000, 10,\!000\}\). In the table, “\(+\)”, “−”, and “\(\sim \)” indicate that the performance of EBADE-fixFCR is statistically better than, statistically worse than, and comparable to that of EBADE, respectively.

When \( FE _{\max }\) was 2000, EBADE-fixFCR statistically outperformed EBADE on more than seven problems for all problem dimensions. This indicates that the recommended values of \({\theta }_{F}\) and \({\theta }_{ CR }\) boost the convergence speed of EBADE-fixFCR under the restricted budgets of FEs, by reducing the search space of parameter configurations. However, with \( FE _{\max } \ge 6000\), the number of “−” was larger than that of “\(+\)” for all D. This is highlighted particularly for \(D=20\) and 30. For example, EBADE outperformed EBADE-fixFCR on 12 problems with \(D=20, FE _{\max } = 8000\) while reducing the number of “\(+\)”. Consequently, the adaptation of DE parameter configurations including \({\theta }_{F}\) and \({\theta }_{ CR }\) is important to enhance the performance of EBADE in moderately EOPs.

Table 9 Statistical results (the count of \(+/-/\sim \)) for 2000, 4000, 6000, 8000, and 10,000 FEs in comparison with EBADE and a variant where \({\theta }_{F}\) and \({\theta }_{ CR }\) are fixed to 0.5 and 0.9, respectively

Parameter analysis for the number of candidate configurations

Now, we evaluate the effectiveness of the prior validation process. To this end, we introduce an EBADE variant without the prior validation process, which can be implemented by setting K to 1. EBADE with \(K=1\) always generates only one candidate parameter configuration, and thus randomly selects a parameter configuration. Further, a sensitivity analysis is performed for K. In particular, we evaluate the performance of EBADE with \(K\in \{1, 2, 4, 8, 10\}\) in addition to its default value, six.

Table 10 summarizes the statistical results of the counts of \(+/-/\sim \), where “\(+\)”, “−”, and “\(\sim \)” indicate that the performance of an EBADE variant is statistically better than, statistically worse than, and comparable with that of EBADE with the default value \(K=6\), respectively. To denote the variant without the prior validation process, “(w/o PV)” is appended to \(K=1\) in the table. As shown in the table, the performance of EBADE was sensitive to the parameter K, because the statistical results changed dramatically as K shifted. In particular, when prior validation was excluded (\(K=1\)), the performance of EBADE degraded significantly on multiple problem instances compared to EBADE using the default value \(K=6\), confirming the effectiveness of the prior validation process. Moreover, \(K=2\) was inappropriate, since EBADE with the default setting (\(K=6\)) statistically outperformed that with \(K=2\). Thus, two candidates of parameter configurations are insufficient to conduct prior validation efficiently. When the value of K increased to 10, the performance of EBADE slightly improved for \( FE _{\max } \le 4000\); however, it degraded for \( FE _{\max } \ge 8000\). This is because a large value of K tends to increase the exploitation bias in the search, and thereby EBADE may suffer from premature convergence. Specifically, when the value of K increases, a greater variety of parameter configurations is sampled and then EBADE tends to select DE parameter configurations that generate overly similar solutions to the target solution.

Table 10 Statistical results (the count of \(+/-/\sim \)) for 2000, 4000, 6000, 8000, and 10,000 FEs in comparison with EBADE with \(K\in \{1, 2, 4, 8, 10\}\)

Parameter analysis for the number of subpopulations

We also validate the effectiveness of subpopulation-based adaptation. For this purpose, an EBADE variant that adapts DEs for individual-based adaptation is prepared. This variant can be implemented by setting \(M=100\) and \(N=1\). The size of the whole population of this variant is the same as the original EBADE; \(N \times M = 4 \times 25 = 100\) and \(N \times M = 1 \times 100 = 100\) for EBADE and this variant, respectively. Thus, they can be fairly compared. The number of top solutions in the post hoc validation phase, described in line 20 in Algorithm 4, is set to 25 for this variant, as all parameter configurations remain constant throughout the search stage otherwise. Additionally, we conduct a sensitivity analysis for M. Specifically, we evaluate the performance of EBADE with \(M\in \{2, 4, 5, 10, 20, 50, 100\}\) in addition to its default value 25. For these M, N was set to \(N\in \{50, 25, 20, 10, 5, 2, 1\}\), respectively, to ensure that the population size remains 100.

Table 11 presents the statistical results in terms of the count of \(+/-/\sim \). The symbols are defined as in the previous subsection, where the default value of M is \(M=25\). To highlight the variant with individual-based adaptation, “(Indiv.)” is appended to \(M=100\) in the table. When individual-based adaptation was conducted (\(M=100\)), the performance of EBADE degraded significantly on multiple problem instances compared with EBADE using the default value \(M=25\), confirming the effectiveness of the proposed subpopulation-based adaptation. This is because the proposed subpopulation-based adaptation can improve the validation accuracy of the effectiveness of parameter configuration by validating those with multiple samples. Specifically, the subpopulation-based adaptation can mitigate the risk of a sample moving in an unintended direction due to random numbers, resulting in an unjustified evaluation of parameter configurations. However, the individual-based adaptation is designed to validate a parameter configuration using a single sample, and thereby, this risk should occur frequently. Accordingly, EBADE performed well using good parameter configurations, which were well detected by the subpopulation-based adaptation.

Moreover, EBADE with \(M\in \{2, 4, 5\}\) should be avoided as they statistically underperformed compared to EBADE with the default setting (\(M=25\)). These results suggest that the number of parameter configurations tested in parallel should be greater than five and the number of solutions validating one parameter configuration should be less than 20. Further, EBADE with \(M=50\) is also inappropriate, because two solutions are insufficient to validate one parameter configuration. Increasing the value of N while decreasing M can improve this validation accuracy, because the number of validation samples, N, increases; however, the number of parameter configurations decreases, leading to a poor diversity of solutions. On the other hand, increasing M (and thus decreasing N) can improve the diversity of solutions; however, less effective configurations may be used owing to low validation accuracy. Thus, it is important to use a parameter setting of \(\{N, M\}\) that balances this trade-off. Our experimental result suggests that \(\{N, M\} = \{10, 10\}, \{5, 20\}, \{4, 25\}\) produces a good balance, as EBADE with these settings performed well.

Table 11 Statistical results (the count of \(+/-/\sim \)) for 2000, 4000, 6000, 8000, and 10,000 FEs in comparison with EBADE with \(M\in \{2, 4, 5, 10, 20, 50, 100\}\)

Adaptation results

Finally, we verify whether the high performance of EBADE  was achieved after EBADE performed adaptation without biasing the use of only certain DE parameter configurations. Specifically, for each element of the parameter configurations, i.e., the scaling factor F, the crossover rate \(\textit{CR}\), the mutation strategy, and the crossover strategy, we enumerated the number of times the subpopulations of EBADE used each candidate within 10,000 FEs, and reported it as a ratio. For F and \(\textit{CR}\), the domains of definition \(F,\textit{CR}\in [0, 1]\) were divided into five ranges with widths of 0.2, and the number of samples was examined for each range.

Figure 3 summarizes the ratio of each candidate used by problem function and the dimension. Because the selected ratio did not depend on D significantly, the results are reported with \(D\in \{10, 30\}\) and those with \(D=20\) are omitted. As illustrated in this figure, EBADE selected various parameter configurations of F, \( CR \), and the mutation and crossover strategies over the search process. As in Fig. 3a), values of the scaling factor F in [0.0, 0.2] were most frequently used, while higher values like \(F \in (0.8, 1.0]\) were less frequently used. This may be attributed to premature convergence to the local optima induced by larger values or slow convergence as the solutions continue to move widely through the search space. On the other hand, larger crossover rate values \(\textit{CR}\) were often selected, as illustrated in Fig. 3b), especially in unimodal functions, i.e., F1–F5. In other words, mutant solutions generated with mutation strategies to accelerate convergence should be actively utilized in EOPs. The best/1 mutation strategy was the most selected, as depicted in Fig. 3c). This tendency is natural, since the best/1 strategy exhibits the strongest exploitation ability, which is suitable for EOPs. However, the other strategies were used adequately to maintain the diversity of solutions and the combined use of these four strategies contributed to the high performance of EBADE, which can be also confirmed in the comparison between EBADE and DE with best/1/bin or best/1/exp in Sect. 5.1. In Fig. 3 d), the binomial crossover strategy was used slightly more. This tendency was reinforced as D was increased. Since the importance of solution diversity became more notable as D was increased, the binomial strategy, which conducts crossover more uniformly, was more suitable. In summary, EBADE exhibited high performance by selecting more candidates appropriate for EOPs. However, it avoided heavy bias by adaptively selecting all of them.

Fig. 3
figure 3

Adaptation results obtained from EBADE

Conclusions

In this paper, we introduced a new adaptive DE variant named emulation-based adaptive DE (EBADE). An adaptive EA for EOPs with moderately restricted budgets was developed for the first time by emulating the principle of sample-efficient approaches, e.g., SAEAs. Specifically, EBADE is characterized by prior validation and subpopulation-based adaptation; emulating the EI-based solution screening with surrogates in SAEAs. EBADE pre-screens for expected improvements from candidate DE parameter configurations before using them without any surrogate. It adopts a multi-population mechanism and each parameter configuration is assigned to a subpopulation to determine the effectiveness of parameter configurations accurately by validating one configuration with respect to multiple solutions. The experimental results obtained in moderately expensive cases demonstrated the statistically significant superiority of EBADE among popular and modern adaptive DEs. Additionally, EBADE was also highly competitive with state-of-the-art SAEAs in moderately EOPs; while SAEAs induced premature convergence, EBADE continued to improve in performance with the shortest runtime. Consequently, this paper contributes to solving moderately EOPs with high-performance, computationally efficient, and auto-tunable approaches.

Despite the effectiveness of EBADE, some possible drawbacks can be pointed out. First, EBADE employs FIR to detect good parameter configurations, but this may be hindered until EBADE discovers good solutions. Accordingly, we will explore alternative indicators instead of FIR to detect good parameter configurations even with a small improvement in fitness values. Second, the Euclidean distance is utilized in the prior validation phase, but it may not be adequate when the problem dimension increases. Thus, we will consider other metrics such as the Mahalanobis distance and cosine similarity to improve the scalability of EBADE to problem dimensions. Finally, we plan to extend this framework for multi-objective optimization problems and compare the performances of the extended EBADE and multi-objective SAEAs.