Sequential improvement for robust optimization using an uncertainty measure for radial basis functions
- 1.2k Downloads
Abstract
The performance of the sequential metamodel based optimization procedure depends strongly on the chosen building blocks for the algorithm, such as the used metamodeling method and sequential improvement criterion. In this study, the effect of these choices on the efficiency of the robust optimization procedure is investigated. A novel sequential improvement criterion for robust optimization is proposed, as well as an improved implementation of radial basis function interpolation suitable for sequential optimization. The leave-one-out cross-validation measure is used to estimate the uncertainty of the radial basis function metamodel. The metamodeling methods and sequential improvement criteria are compared, based on a test with Gaussian random fields as well as on the optimization of a strip bending process with five design variables and two noise variables. For this process, better results are obtained in the runs with the novel sequential improvement criterion as well as with the novel radial basis function implementation, compared to the runs with conventional sequential improvement criteria and kriging interpolation.
Keywords
Sequential improvement Metamodeling Radial basis function Kriging Metamodel uncertainty Sheet bending1 Introduction
Engineering can be seen as a sequence of making choices, from fundamental design choices to defining design details. In the last century, many empirical and theoretical models have been developed to help with these decisions. Nowadays engineers frequently use computational models to assess the effects of design choices and determine optimal designs.
In many engineering problems, it is infeasible to find an optimal design through trial and error due to a large number of interacting parameters. Therefore, many optimization methods have been developed to automate this process, each method being suitable to solve a specific type of problem in an efficient way. With these methods an optimal configuration of the design may be found. However, such a design may malfunction in real-life due to variability of the model parameters. A specific branch of optimization research focuses on these problems: within robust optimization the goal is to find a design which fulfills the requirements even under the influence of parameter variations.
When dealing with expensive computational models, it is desired to acquire insight into the problem at hand without performing an excessive amount of model evaluations. One approach is to replace the model with an easy to evaluate metamodel, also known as surrogate or response surface model. The metamodel is built based on a moderate number of evaluations of the computational model, and thereafter used for optimization. Some researchers developed methods to iteratively enrich the metamodels with new evaluations of the computational model. This is known as sequential optimization or sequential improvement.
Sequential robust optimization flowchart. The building blocks of the optimization procedure are written in bold letters. The gray boxes are investigated in this work
The robust optimization procedure consists of several steps (Fig. 1). For some steps a choice has to be made between different methods to perform a prescribed task. For example, latin hypercube sampling, fractional factorial sampling, random sampling or a combination of these may be chosen as sampling strategy for the initial Design Of Experiments (DOE). To fit a set of data points, many methods have been developed, such as polynomial regression, neural networks, kriging, Radial Basis Functions (RBF), moving least squares and many others. Even within these methods many choices can be made on the exact configuration of the method. Hence, these methods can be seen as building blocks of the optimization algorithm. Every building block may be replaced with any other building block which is capable of performing the prescribed task. Clearly, the choice for a certain building block may severely affect the efficiency of the optimization algorithm.
In the present work, the efficiency of the robust optimization algorithm is studied. More specifically, the effects of the metamodeling method and the sequential improvement criterion on the optimization are examined. A widely used metamodeling method is kriging, as observed by several researchers (Wang and Shan 2007; Kleijnen 2009; Forrester and Keane 2009). One advantage of kriging is that it includes a prediction of the metamodel uncertainty. However, clustering of DOE points adversely affects the predictive capability of kriging methods (Zimmerman et al. 1999; Havinga et al. 2013). This is of interest because point clusters are formed during optimization due to the sequential improvement procedure. An interpolating metamodeling method which may be adapted to account for clustering of DOE points is RBF interpolation. Good fitting capabilities for nonlinear functions have been reported for RBF interpolation in several studies (Franke 1982; Jin et al. 2001). However, the uncertainty measures which have been developed for RBF are not suitable for sequential optimization. Therefore, an improved configuration of the RBF metamodel is proposed, taking into account the local DOE density and including a prediction of the metamodel uncertainty based on the Leave-One-Out Cross-Validation (LOOCV) measure.
Sequential improvement in robust optimization has been scarcely investigated up to now. For deterministic optimization, the Expected Improvement (EI) criterion was proposed by Jones et al. (1998). This criterion can be used for robust optimization, but does not take into account the uncertainty of the objective function value at the sampled design points. Jurecka (2007) adapted the expected improvement criterion of Jones for robust optimization. However, this criterion has some disadvantages which will be shown in this work. Therefore, a new expected improvement criterion for robust optimization is proposed.
The sequential robust optimization procedure is generic and, therefore, not limited to specific research fields. However, our interest goes to robust optimization of metal forming processes. The aim is to design metal forming processes with a low sensitivity to variations of material properties and process parameters such as friction and tool wear. Our background may be noticed in the examples given throughout this work, without violating the genericity of the approach.
This paper is structured as follows: in Section 2 a selective overview of the state of the art of robust optimization is given: the metamodeling methods including error prediction (Section 2.1) and the sequential improvement criterion (Section 2.2) are discussed. The RBF metamodel with uncertainty measure is proposed in Section 3. The novel robust expected improvement formulation is introduced in Section 4. The efficiency of the expected improvement criteria and metamodeling methods are compared in Section 5 based on mathematical functions. In Section 6 a demonstration problem is presented: a V-bending process with two noise variables, five design variables and two nonlinear constraints. The robust optimization results are given in Section 7 and the conclusion is given in Section 8.
2 State of the art in robust optimization
The famous paper of Jones et al. (1998) in which the expected improvement criterion was introduced, is titled Efficient Global Optimization of Expensive Black-Box Functions. The term black-box function denotes that it is not of interest how the function works. The only requirement is that the function is able to process a set of input variables and return one or multiple output values. Therefore, the optimization procedure is suitable for optimization with any kind of numerical model. Unlike experiments, trials with the same input variable values always give the same output value. Therefore, there is no need to perform multiple tests of the same variable set. The adjective expensive emphasizes the computational cost of the function and stresses the need to keep the number of function evaluations as low as possible.
In the case of deterministic optimization, both the black-box function as well as the optimization problem are defined in the design variable space x. In contrary, in robust optimization the black-box function is defined in the combined space of the design variable x and the noise variable z, while the optimization problem is stated in the design variable space x. At a certain point in the design space x ′, the objective function is defined by the behavior of the black-box function throughout subspace (x ′, z). Given an assumption on the probability distribution of the noise variable z, the probability distribution for the output of the black-box function may be determined at x ′, and subsequently this probability is used in the objective function.
The objective function Obj(Y(x)) determines which property of the probability distribution Y(x) should be minimized. The robustness of a production process can be defined in several ways, such as the average error of the production process or the percentage of the products that meet the product requirements. The most straightforward way to quantify these measures is to determine the statistical parameters of the process output Y(x), such as the mean μ Y , standard deviation σ Y or median. Two different ways to improve the accuracy of the process are by shifting the mean μ Y or decreasing the standard deviation σ Y (Koch et al. 2004). In some cases, these may be conflicting objectives (Beyer and Sendhoff 2007). Some researchers choose to combine these measures in a single objective, for example by minimizing μ Y + kσ Y , as is done in this study. Many other robust optimization objectives can be found in literature, such as Pareto front determination (Coelho and Bouillard 2011; Nishida et al. 2013), quantile optimization (Rhein et al. 2014), minimization of the worst case scenario (Marzat et al. 2013) and minimization of the signal-to-noise ratio (Taguchi and Phadke 1984; Leon et al. 1987).
The lower bound and upper bound of the design variables x are defined by L B and U B. The constraint functions C i (G i (x)) and \(\text {C}_{i}^{\text {eq}} (H_{i} (\mathbf {x} ) )\) determine the constraints for the distributions G i (x) and H i (x).
Usually, every simulation yields the values of the functions f, g i and h i for one variable set (x, z). However, the whole subspace (x, Z) must be known to exactly evaluate the objective function Obj(Y(x)) and the constraint functions C i (G i (x)) and \(\text {C}_{i}^{\text {eq}} (H_{i} (\mathbf {x} ) )\) for one value of the design variable x. Clearly, it is not desired to spend too many simulations for just one point in the design variable space. Therefore, the optimization algorithm should be able to handle uncertainty in the objective and constraint function values, and efficiently select whether to sample multiple points in the noise space or explore other regions of the design space.
When defining a robust optimization problem, one of the first problems is to determine the distribution p Z (z) of the noise variable z. In the optimization of a production process, these may be properties such as material variation, friction variation or sheet thickness variation. These variations can be characterized by min-max values or by statistical properties such as mean and standard deviation. In some cases, knowledge of the correlations between the different noise variables is required for a correct estimation of the robustness of the process (Wiebenga et al. 2014).
In the following sections the different components used to solve the optimization problem of (6) are discussed (see Fig. 1).
2.1 Metamodeling methods
A major component of the optimization algorithm is the metamodeling method which is used to fit the objective function f and the constraint functions g i and h i using a set of observations D = {(x (i), y (i))|i = 1…n}. The sequential improvement strategies discussed in this study (Section 2.2) do not only need the estimate of the functions \(\hat {f}\), \(\hat {g}_{i}\) and \(\hat {h}_{i}\) at untried points x but also require an estimate of the prediction uncertainty at these points. Therefore, only methods that provide this estimate will be discussed.
As a basis for interpolating simulation data, it is good practice to normalize the inputs and to use a regression model as underlying model. Often, a simple polynomial model is used to capture the major trends of the observations y. Hence, the metamodeling methods are used to fit the residue \(\mathbf {y} - \hat {f}_{\text {regr}}(\mathbf {x})\) of the polynomial fit. In the following text, fitting of the residue is assumed. For clarity, y is used in the equations instead of \(\mathbf {y} - \hat {f}_{\text {regr}}(\mathbf {x})\).
2.1.1 Kriging
The exponents p are related to the smoothness of the function (Jones et al. 1998). Under assumption that the underlying function f is smooth, the values of p can be set to 2 and only the 𝜃 parameters have to be optimized. The ability to fit complex nonlinear models and to estimate the prediction uncertainty attracted many researchers in the field of sequential optimization to the use of kriging interpolation.
2.1.2 Radial basis functions
Shape of the Gaussian (a) and multiquadric (b) basis functions with varying shape parameter value
As stated before, an estimate of the prediction uncertainty is required for sequential sampling. Some researchers derived formulations for the prediction uncertainty for RBF. Li et al. (2010) fit a RBF with Gaussian basis functions with a single parameter for the kernel width using LOOCV. They assume the prediction \(\hat {f}(\mathbf {x})\) to be a realization of a stochastic process and assume the variance of the stochastic process to be ψ(0)−ψ T Ψ −1 ψ (Gibbs 1997; Sóbester et al. 2004; Ji and Kim 2013). This uncertainty measure is applicable for basis functions which are decaying from the center point. A more general uncertainty measure is given by Yao et al. (2014), which derive the relation between the unknown weight factor w at an untried point x with the error of the prediction \(f(\mathbf {x}) - \hat {f}(\mathbf {x})\). They predict the weight factor \(\hat {w}(\mathbf {x})\) by interpolating the weight factors with a RBF model with the same model parameters as the underlying model for the prediction \(\hat {f}(\mathbf {x})\). The disadvantage of this measure is that the prediction uncertainty reduces to zero for points which are far away from all DOE points. Nikitin et al. (2012) propose to directly interpolate the LOOCV values at all observation points. However, it is not clear how this uncertainty measure should be used for sequential optimization, since the uncertainty value at the DOE points is not equal to zero. Therefore, we propose a new uncertainty measure in Section 3 with two parameters which is determined with the LOOCV values and a likelihood function.
2.2 Sequential improvement
Due to the computational cost of the simulations, efficient exploration of the design space is required. This can be achieved with sequential improvement strategies. The concepts of sequential improvement will be first explained for the deterministic case with the objective to minimize f and thereafter the robust optimization case will be discussed.
An important question is whether to search in the neighbourhood of the optimum of the predictor \(\hat {f}\) or to search in a region with high prediction uncertainty. The former is denoted as local search and can be achieved by sampling \(\min _{\mathbf {x}} \hat {f}(\mathbf {x})\) whereas the latter is denoted as global search and can be achieved by sampling \(\max _{\mathbf {x}} \hat {s}^{2}(\mathbf {x})\).
The sequential improvement procedure for robust optimization is split in two steps: first a location in the design space x is selected based on expected improvement, thereafter the location in the noise space z is selected. At any point in the design space x an estimate of the objective function value \(\text {Obj} (\hat {Y} (\mathbf {x}) )\) and an estimate of its uncertainty \(\hat {s}_{\text {Obj}} (\mathbf {x}) \) can be determined. In contrast to the deterministic optimization case, there is not a clearly defined best point y min in the design space, since the prediction error \(\hat {s}_{\text {Obj}} (\mathbf {x})\) is larger than zero for any x. Some authors propose to define the best point x b as the point {x b ∈x (i):i = 1…n} with the lowest value for the objective function prediction (Jurecka 2007; Wiebenga et al. 2012). In this work a different criterion is used. The criterion and the reasoning behind it are discussed in Section 6.2.
Now a criterion is needed to quantify whether a candidate infill point x can potentially yield an improvement over the current best solution x b , given \(\text {Obj}(\hat {Y}(\mathbf {x}))\), \(\hat {s}_{\text {Obj}}(\mathbf {x})\), \(\text {Obj}(\hat {Y}(\mathbf {x}_{b}))\) and \(\hat {s}_{\text {Obj}}(\mathbf {x}_{b})\). One approach is to ignore the prediction error at the best point \(\hat {s}_{\text {Obj}}(\mathbf {x}_{b})\) and use the criterion of (15) (Jurecka et al. 2007). Note that selecting the current best point as candidate infill point x = x b yields a nonzero value of the expected improvement. Clearly it is desired to sample more points at x b to decrease the prediction error of the current best estimate.
Probability of the objective function value for x b and x. The expected improvement criterion of Jurecka (2007) is calculated with the probability given by the hatched area
The last step of the sequential improvement procedure is to select a location in the noise space z at the selected location in the design space x ′. The purpose of the new simulation is to improve the prediction of \(\text {Obj}(\hat {Y}(\mathbf {x}^{\prime }))\). Hence criteria such as \(\arg \!\max _{\mathbf {z}} \hat {s}^{2} (\mathbf {x}^{\prime },\mathbf {z})\) (Jurecka et al. 2007) or \(\arg \!\max _{\mathbf {z}} (\hat {s}^{2} (\mathbf {x}^{\prime },\mathbf {z}) p_{\mathbf {Z}}(\mathbf {z}))\) (Jurecka 2007; Wiebenga et al. 2012) can be used for this purpose.
The abovementioned sequential improvement criteria can be regarded as global improvement criteria, in the sense that a criterion which is defined in the complete function domain is to be minimized to obtain a new infill point. A different class of sequential improvement methods is given by trust region strategies. Trust region strategies work with metamodels (also referred to as approximation models) which can be trusted within a certain trust region. The size and position of the trust region and the metamodel itself are updated based on evaluations of a high-fidelity model in a local search procedure. Hence, the choice of a new infill point is determined by the search path of the algorithm. The algorithm has been extended from local quadratic approximation models to any kind of approximation models by Alexandrov et al. (1998). One specific application of trust region strategies is to solve optimization problems using variable fidelity models which are coupled through mapping functions. Gano et al. (2006) used kriging interpolation functions to map variable fidelity models in a trust region optimization procedure. For further reading we refer to the application of trust region strategies to different engineering problems, such as the design of an autonomous hovercraft (Rodríguez et al. 2001) and the weight minimization of a fiber-reinforced cantilever beam (Zadeh et al. 2009).
This concludes the overview of the building blocks of the robust sequential optimization algorithm. In Section 3 an implementation of a RBF model is proposed, including estimation of the prediction error. In Section 4 a novel expected improvement criterion for robust optimization is proposed, including a discussion on the several criteria available for robust optimization. The effectiveness of these building blocks will be compared with other methods in Section 7.
3 RBF with uncertainty measure
As discussed in Section 2.1.2, there is a lot of freedom for the user on the exact implementation of the RBF method. Therefore, a proposal is made for an implementation for efficient sequential optimization. The procedure is elaborated in this section and the implementation choices are discussed.
First, the scaled distance to the nearest neighbor d i is determined with (21). Thereafter, the distance to the nearest neighbor is normalized with (22), such that c i ≤1∀c i holds. The latter operation is needed, to ensure that each value for the scaling parameter 𝜃 results in a unique metamodel. Note that the scaling parameter 𝜃 has not yet been selected and can be freely chosen. When the scaling parameter 𝜃 is multiplied with a constant C, the scaled distance to the nearest neighbors d i changes accordingly (d i (C 𝜃) = Cd i (𝜃)). If the local shape parameters are set to the distance to the nearest neighbor (c i = d i ), the relative effect of basis function ψ (i) throughout the scaled model space (𝜃∘x) will remain the same when multiplying the scaling parameter 𝜃 with a constant. Hence, all models with parameter sets C 𝜃 ∀C > 0 will be equal. To avoid this effect, the normalization from (22) is used.
Interpolation of dataset with varying point density using Gaussian RBF, with constant shape parameters and 𝜃 = 0.1 ( Open image in new window
) or 𝜃 = 0.3 ( Open image in new window
) and with varying shape parameters and 𝜃 = 1 ( Open image in new window
)
The optimal scaling parameters 𝜃 are determined by minimization of the L 2 norm of the error ||𝜖||2. The number of components of 𝜃 equals the number of dimensions of the metamodel.
4 Robust expected improvement
The full derivation can be found in the Appendix. With zero uncertainty \(\hat {s}_{b}\) of the objective at x b the equation reduces to (15).
Expected improvement criteria for robust optimization: Jones (a), Jurecka (b) and novel criterion (c) with \(\hat {s}_{b} = 0.4\). The expected improvement value of x = x b is marked with a circle ∘
When observing the criterion of Jurecka (2007) (Fig. 5b) some remarkable characteristics are seen. First observation is that the expected improvement value at x b equals zero. Therefore, the current best point will never be selected as the new infill point. Also a large region has an EI value close to zero, which makes it difficult for the algorithm to distinct between regions with high and low probability of improvement. Furthermore it is seen that in some regions the EI value decreases with increasing uncertainty \(\hat {s}\), such as around \((\hat {\mu }_{b}-\hat {\mu } = 0.2, \hat {s}=0)\). However if two points would have the same expected value \(\hat {\mu }\), the point with the highest uncertainty \(\hat {s}\) should be preferred. In the Sections 5 and 7 it is examined how these properties of the Jurecka criterion influence its efficiency.
5 Demonstration: Gaussian random fields
The efficiency of the sequential improvement criteria and metamodeling methods for robust optimization have been compared based on a robust optimization test with a large set of Gaussian random fields. The approach and results are presented in Section 5.1 for the sequential improvement criterion and in Section 5.2 for the metamodeling methods.
5.1 Sequential improvement criterion
The efficiency of the three sequential improvement criteria for robust optimization have been compared based on an optimization test with mathematical functions. To eliminate the influence of the metamodeling technique a set of 15000 Gaussian random fields f(x, z) with predefined statistical parameters has been generated. Therefore, the kriging predictor can be used to obtain the best prediction and uncertainty measure given any set of observations D = {((x,z)(i), f(x,z)(i))|i = 1…n}. The random fields have one design parameter x and one noise parameter z in the range [0,1]. The correlation function of the random fields is Gaussian (7) with model parameters p x = 2 and p z = 2. The model parameters 𝜃 x and 𝜃 z are selected randomly from the respective sets {30,40,50…300} and {30,40,50…100}. The probability distribution of the random field f(x,z) is \(p_{f} \sim \mathcal {N}(0,1^{2})\). The probability of the noise parameter z is given by a truncated normal distribution with mean 0.5, standard deviation 0.1, truncated outside the range [0,1]. During the optimization the selection of new infill points is restricted to an equidistant grid of 25 points in x-direction and 21 points in z-direction. As an initial DOE the center point of the field (x,z)=(0.5,0.5) is taken and a total of 50 sequential improvement iterations are performed.
RMSE evolution at robust optimum during optimization averaged over 15000 runs: Jones ( Open image in new window
) (a,d), Jurecka ( Open image in new window
) (b,d) and novel criterion ( Open image in new window
) (c,d). RMSE of separate runs is shown with thin gray lines
Contingency table for results of Jones’ criterion and the novel criterion, stating whether the correct or wrong solution has been found after 50 iterations
| Novel criterion | |||
|---|---|---|---|
| correct | wrong | ||
| Jones | correct | 13450 | 367 |
| wrong | 513 | 670 | |
The convergence of the Jurecka criterion is shown in Fig. 6b. During the first 10 iterations the average convergence is similar to the average convergence of the other methods but thereafter the convergence rate drops. Furthermore it is observed that the Jurecka criterion fails to select the correct robust optimum after 50 iterations in 38.7 % of the runs. One cause for the poor convergence is that the current best optimum x b is never selected as an infill point since the expected improvement value at x b equals zero (Fig. 5b). Therefore, it may happen that an underestimated value of the current best optimum x b is not verified by additional samples, leading to erroneous results.
The mathematical test reveals performance differences between the three expected improvement criteria. However limited conclusions should be drawn since only one type of function (2D Gaussian random field) with one metamodeling method (kriging) have been studied. Therefore, all expected improvement criteria are included in the study on the efficiency of the robust optimization procedure of a V-bending process with a Finite Element (FE) model with seven variables. The study is presented in Section 6.
5.2 Metamodeling method
Overview of Gaussian random field parameters for the metamodel test
| set nr. | 1 | 2 | 3 | 4 |
|---|---|---|---|---|
| \(\theta _{\text {x}}\) | 40 | 40 | 40 | 40 |
| \(\theta _{\text {z}}\) | 40 | 100 | 40 | 100 |
| \(\sigma _{\text {n}}\) | 0 | 0 | 0.2 | 0.2 |
For each Gaussian random field, the same optimization procedure has been followed as in Section 5.1, except for the details mentioned in this paragraph. The novel criterion for sequential improvement has been used. A small initial DOE is used, with a full factorial design (four points) and six additional latin hypercube design points. Thereafter, 50 sequential optimization iterations have been performed. During the optimization the selection of new infill points is restricted to an equidistant grid of 101 points in x-direction and 21 points in z-direction. As metamodeling methods, kriging, RBF with Gaussian basis functions (RBFG) and RBF with multiquadric basis functions (RBFMQ) have been used. A constant value has been used to detrend the data for all metamodeling methods. Note that fitting of the kriging model requires determination of the statistical parameters (11), whereas the kriging model in Section 5.1 was fit with known statistical parameters.
Potential improvement PI (30) averaged over 2000 runs for kriging ( Open image in new window
), RBFG ( Open image in new window
) and RBFMQ ( Open image in new window
): set 1 (a), set 2 (b), set 3 (c) and set 4 (d)
Comparison between metamodeling methods based on the real objective function values at the predicted optima after 50 iterations. Statistical significant results (p-value smaller than 0.001) have a gray background
The comparison of the effect of metamodeling methods on the efficiency of the robust optimization procedure shows that the choice for the best metamodel depends on the problem at hand. To assess the effect of the discussed metamodeling methods in a real engineering problem, the robust optimization problem of a V-bending metal forming process is presented in the following section.
6 Demonstration process: V-bending
Impression of the V-bending process
6.1 V-bending model
Main angle 𝜃 M and transition angle 𝜃 T
Optimization variables of the V-bending process. The noise variables are shown in red
A 2D FE model has been built for the optimization. The implicit MSC.Marc solver has been used. Plane strain condition is assumed in the model and the die and punch have been modeled with elastic quadrilateral elements. A half sheet has been modeled with 500 quadrilateral elements due to symmetry. The elasticity of the tooling and the sheet is set to 210 GPa. Isotropic hardening and Von Mises yield criterion are used for the sheet. An experimentally obtained hardening curve is implemented in the model with a tabular input.
The numerical stability of the model is of utmost importance for the optimization procedure. The effect of numerical noise on metamodel based optimization has been studied by Wiebenga and van den Boogaard (2014). The order of variation due to numerical noise should be lower than the order of variation caused by variation of noise variables. To ensure this, it is highly recommended to perform a study on the numerical noise of the model before running the optimization. In our experience, optimizing contact and convergence settings and using fixed time steps in the simulation helped to decrease the numerical noise of the model.
The average simulation time of the model is around 11 minutes. Even though the complexity of the model is limited and the computational cost is reasonable, it is too expensive to perform simulations in the full parameter space. Therefore an efficient optimization scheme is required.
6.2 Optimization algorithm
In this section an overview of the full optimization algorithm is given. The full procedure is implemented in MATLAB. For the initial DOE a resolution IV 27−3 fractional factorial design is used to have some DOE points on the bounds of the domain, combined with a latin hypercube design of 24 points (constructed with default MATLAB settings, meaning that it is optimized with maximin criterion in five iterations) to get a good distribution of the points throughout the full domain. The combined DOE has 40 points, which is sparse for a seven-dimensional nonlinear function but sufficient for starting a sequential optimization procedure. Out of the 40 initial simulations, two simulations failed. Hence, all runs start with the same set of 38 points.
Overview of used metamodels
| Name | Metamodel type | Basis function | Polynomial regression |
|---|---|---|---|
| kriging0 | kriging | Gaussian | 0th order |
| kriging1 | kriging | Gaussian | 1st order |
| RBFG0 | RBF | Gaussian | 0th order |
| RBFG1 | RBF | Gaussian | 1st order |
| RBFMQ0 | RBF | multiquadric | 0th order |
| RBFMQ1 | RBF | multiquadric | 1st order |
A latin hypercube sample of n mc = 50 points with the distribution p Z (z) is used. It is found by experience that this is sufficient to estimate \(\hat {\mu }_{f}\), \(\hat {\sigma }_{f}\) and \(\hat {s}_{\text {Obj}}\) with affordable computational cost and reasonable accuracy.
The three sequential improvement criteria discussed in this work (Jones, Jurecka and the novel criterion) have been used for the V-bending optimization. At each iteration of sequential improvement the expected improvement is determined with respect to a point x b . Other researchers in sequential robust optimization chose {x b ∈x (i):i = 1…n} with the lowest value for the objective function prediction (Jurecka 2007; Wiebenga et al. 2012). However, at the end of the robust sequential optimization procedure, the choice of an optimal design is based on a metamodel with predictive uncertainty \(\hat {s}_{\text {Obj}} (\mathbf {x})\). One could select the design point with the lowest objective function value, but that would probably be reconsidered if the predictive uncertainty \(\hat {s}_{\text {Obj}} (\mathbf {x}_{\text {opt}})\) would be too high. Therefore, selecting an optimal design can be seen as a multi-objective optimization problem on a higher level. Both a low objective function value as well as a low objective function uncertainty are desired. We propose to select the best design by minimizing \(\text {Obj} (\hat {Y} (\mathbf {x}) ) + 6 \hat {s}_{\text {Obj}} (\mathbf {x})\). During sequential improvement, this criterion is used to select the best point x b from the set of already sampled design points {x b ∈x (i):i = 1…n}.
During the sequential optimization procedure some simulations fail to converge. Therefore, a constraint is applied for the sequential improvement criterion that the normalized distance to failed simulations should be higher than 0.1. To examine the convergence behavior of the algorithm, no termination criterion is selected and the runs are continued until a total of 400 successful sequential improvement simulations is reached. Therefore, the final size of the DOE is 38+400=438, which is reasonably sparse for a strongly nonlinear 7-dimensional space. A total of 18 runs are performed: six metamodeling methods times three sequential improvement criteria.
7 Results
Results of all runs after 400 iterations
| Optimum | Objective | Constraint | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| \(\alpha \) | D | L | \(R_{1}\) | \(R_{2}\) | \(\mu _{\theta _{M}}\) | \( \sigma _{\theta _{M}}\) | obj | RMSE | \(\mu _{\theta _{T}} - 2 \sigma _{\theta _{T}}\) | \(\mu _{\theta _{T}} + 2 \sigma _{\theta _{T}}\) | RMSE | ||
| Jones | krig0 | 93.6 | 0.53 | 4.38 | 2.27 | 1.15 | 90.0 (90.1) | 0.13 (0.12) | 0.40 (0.42) | 0.13 (0.10) | 94.0 (94.0) | 95.4 (95.4) | 0.07 (0.03) |
| krig1 | 91.3 | 0.55 | 5.00 | 1.86 | 1.28 | 90.0 (90.0) | 0.06 (0.13) | 0.18 (0.40) | 0.05 (0.12) | 92.3 (92.4) | 95.6 (95.4) | 0.07 (0.12) | |
| RBFG0 | 91.7 | 0.52 | 5.00 | 2.14 | 1.49 | 90.0 (90.0) | 0.04 (0.06) | 0.12 (0.21) | 0.01 (0.06) | 92.9 (92.8) | 95.3 (95.5) | 0.00 (0.09) | |
| RBFG1 | 91.3 | 0.55 | 5.00 | 1.87 | 1.28 | 90.0 (90.0) | 0.05 (0.07) | 0.14 (0.24) | 0.02 (0.07) | 92.1 (92.2) | 95.3 (95.3) | 0.04 (0.07) | |
| RBFMQ0 | 93.8 | 0.55 | 4.46 | 2.03 | 1.11 | 90.0 (90.1) | 0.11 (0.14) | 0.34 (0.47) | 0.04 (0.07) | 92.4 (92.4) | 94.4 (94.5) | 0.03 (0.06) | |
| RBFMQ1 | 93.6 | 0.54 | 4.98 | 2.03 | 1.22 | 90.0 (90.0) | 0.08 (0.16) | 0.24 (0.48) | 0.05 (0.10) | 92.2 (92.2) | 94.9 (95.0) | 0.05 (0.07) | |
| Jurecka | krig0 | 93.2 | 0.53 | 4.32 | 2.30 | 1.27 | 90.0 (90.0) | 0.10 (0.17) | 0.31 (0.54) | 0.04 (0.10) | 94.0 (94.0) | 95.7 (95.8) | 0.04 (0.07) |
| krig1 | 92.4 | 0.57 | 5.00 | 1.77 | 1.07 | 90.0 (90.0) | 0.06 (0.08) | 0.18 (0.29) | 0.04 (0.06) | 92.3 (92.3) | 95.1 (95.1) | 0.04 (0.04) | |
| RBFG0 | 93.9 | 0.55 | 4.96 | 2.00 | 1.15 | 90.0 (89.8) | 0.09 (0.17) | 0.26 (0.72) | 0.06 (0.26) | 92.4 (92.3) | 94.8 (94.6) | 0.11 (0.22) | |
| RBFG1 | 91.2 | 0.55 | 5.00 | 1.63 | 1.30 | 90.0 (90.0) | 0.13 (0.14) | 0.39 (0.46) | 0.03 (0.06) | 92.6 (92.6) | 95.8 (95.7) | 0.01 (0.05) | |
| RBFMQ0 | 91.4 | 0.55 | 5.00 | 1.83 | 1.24 | 90.0 (90.0) | 0.08 (0.10) | 0.24 (0.31) | 0.02 (0.08) | 92.1 (92.2) | 95.2 (95.2) | 0.04 (0.08) | |
| RBFMQ1 | 90.9 | 0.54 | 4.98 | 1.91 | 1.36 | 90.0 (90.0) | 0.06 (0.06) | 0.19 (0.18) | 0.02 (0.07) | 92.1 (92.2) | 95.5 (95.4) | 0.03 (0.07) | |
| Novel | krig0 | 92.6 | 0.58 | 5.00 | 1.77 | 1.03 | 90.0 (90.0) | 0.10 (0.14) | 0.30 (0.45) | 0.04 (0.14) | 92.5 (92.5) | 95.3 (95.1) | 0.01 (0.12) |
| criterion | krig1 | 90.4 | 0.53 | 4.74 | 2.02 | 1.50 | 90.0 (90.1) | 0.06 (0.08) | 0.17 (0.30) | 0.04 (0.10) | 92.1 (92.0) | 95.9 (96.1) | 0.03 (0.12) |
| RBFG0 | 91.6 | 0.52 | 4.80 | 2.12 | 1.48 | 90.0 (90.0) | 0.06 (0.04) | 0.19 (0.14) | 0.02 (0.04) | 92.7 (92.7) | 95.2 (95.2) | 0.05 (0.06) | |
| RBFG1 | 92.2 | 0.53 | 5.00 | 2.00 | 1.31 | 90.0 (90.0) | 0.06 (0.05) | 0.19 (0.19) | 0.01 (0.06) | 92.1 (92.1) | 94.6 (94.6) | 0.04 (0.05) | |
| RBFMQ0 | 91.4 | 0.52 | 4.54 | 2.13 | 1.49 | 90.0 (90.0) | 0.04 (0.05) | 0.13 (0.15) | 0.01 (0.01) | 92.6 (92.6) | 95.1 (95.1) | 0.01 (0.02) | |
| RBFMQ1 | 91.9 | 0.52 | 4.97 | 2.08 | 1.43 | 90.0 (90.1) | 0.07 (0.05) | 0.21 (0.21) | 0.05 (0.08) | 92.6 (92.8) | 95.0 (95.1) | 0.05 (0.15) | |
Metamodel prediction of the main angle 𝜃 M at optimum found after 400 iterations for Jones criterion - RBFG0 (a), Jurecka criterion - kriging0 (b), Jurecka criterion - RBFMQ0 (c) and novel criterion - RBFMQ0 (d). Figures (e–f) show the checked surface corresponding to figures (a–d) respectively. The black dots represent the 7 by 7 grid where the simulations were performed
Obviously, it is not possible to know the global optimum for sure without performing an excessive amount of FE model runs. Therefore, it is not feasible to compare the convergence of the global optimum for all runs. However, the goal of a robust optimization run is to find a good solution of the optimization problem with moderate computational effort. Looking at the results in Table 5, it can be seen that the found optima have a checked objective function value in between 0.14° (novel criterion - RBFG0) and 0.72° (Jurecka - RBFG0). The objective function value is underestimated for 15 out of 18 runs. Only one run has constraint violations. Furthermore the RMSE (\(\hat {s}_{\text {Obj}}\)) is underestimated in 17 out of 18 runs for both the 𝜃 M metamodel as well as for the 𝜃 T metamodel. This is expected since the optimum is selected based on \(\text {Obj} (\hat {Y} (\mathbf {x}) ) + 6 \hat {s}_{\text {Obj}} (\mathbf {x})\), which includes both objective function value as well as RMSE value. Therefore, it is more probable for the optimum to be selected from regions where these values are underestimated. In general it can be said that the robust sequential optimization algorithm yields good results for the V-bending problem. Nearly all results fulfill the constraints and satisfying objective function values have been found. The reference for the objective function value is the best result found with the previous study performed with the same model, where an objective function value of 0.85° was found using four design variables (Wiebenga et al. 2012).
The number of sequential optimization runs is not sufficient to assess the statistical relevance of the results. Hence, it cannot be stated which metamodeling method or sequential improvement criterion is most efficient for robust optimization of the V-bending problem. For statistical relevant results, we refer to the test with mathematical functions in Section 5. However, the V-bending problem results give one example for the performance of the building blocks of the procedure. More general results may be obtained through new studies on the efficiency of the building blocks for other engineering problems. Furthermore, several other conclusions can be drawn based on detailed observation of the results.
Checked objective function values averaged over all runs with the same expected improvement criterion (markers). The averaged metamodel prediction of the objective function value is bounded by \(\text {Obj} (\hat {Y} (\mathbf {x}) ) \pm \hat {s}_{\text {Obj}} (\mathbf {x})\), for the Jones ( Open image in new window
), Jurecka ( Open image in new window
) and novel criterion ( Open image in new window
)
Checked objective function values averaged over all runs with the same metamodeling method (markers). The averaged metamodel prediction of the objective function value is bounded by \(\text {Obj} (\hat {Y} (\mathbf {x}) ) \pm \hat {s}_{\text {Obj}} (\mathbf {x})\), for kriging ( Open image in new window
), RBFG ( Open image in new window
) and RBFMQ ( Open image in new window
)
Evolution of R 2 values of metamodels, estimated based on 5000 simulations performed throughout the whole domain, for the main angle 𝜃 M (black line) and the transition angle 𝜃 T (gray line)
Objective function value predictions and checked values (diamonds) for Jones ( Open image in new window
) (a,d,g,j,m,p), Jurecka ( Open image in new window
) (b,e,h,k,n,q) and novel criterion ( Open image in new window
) (c,f,i,l,o,r) at optimum determined with \(\text {Obj} (\hat {Y} (\mathbf {x}) ) + 6 \hat {s}_{\text {Obj}} (\mathbf {x})\), with constraint that the probability of constraint violation should be less than 1 %. This constraint could not be fulfilled for the Jones - kriging0 run at 50 iterations and for the Jurecka - kriging0 run at 50, 150 and 200 iterations. The error bars represent the \(\pm \hat {s}_{\text {Obj}} (\mathbf {x})\) range
8 Conclusion
In this work the effect of the metamodeling method and the sequential improvement criterion on the efficiency of the robust optimization procedure is investigated.
The only available criterion for sequential improvement in robust optimization was the Jurecka criterion. However, this criterion has some artifacts, such as that, when two points have the same prediction for the objective function value, the point with the largest prediction uncertainty is not always preferred as new sampling point. Furthermore, the current estimate of the best design point has an expected improvement value of zero by definition. Therefore, an erroneously underpredicted objective function value will not be verified with additional simulations, even when the prediction uncertainty is large. Therefore, we propose a new criterion for sequential improvement in robust optimization, which is derived from the deterministic expected improvement criterion of Jones et al. (1998). This criterion shows better performance than the Jurecka criterion and than the deterministic Jones criterion, both in a test with mathematical test functions (Section 5) as well as in a V-bending optimization problem (Section 7). Whereas the validity of this conclusion is shown to be statistically significant for the test with mathematical test functions, no statistical test could be performed based on the small number of optimization runs in the V-bending test. It is recommended to verify the presented observations with new studies on robust optimization in engineering.
Regarding the metamodeling method, kriging has been used by many researchers for sequential optimization. We propose an improved implementation of RBF suitable for sequential optimization, including an estimation of the metamodel uncertainty. In a robust optimization test with mathematical test functions it was shown with statistical significance that kriging metamodels perform best when no noise is present on the black-box function, whereas the proposed implementation of the RBF interpolation function with multiquadric basis functions performs best when noise is present. Hence, it is shown that the choice for the best metamodeling method for robust optimization should depend on the investigated optimization problem. In the test with the optimization of a strongly nonlinear V-bending process the RBF metamodels outperformed the kriging metamodels with a small margin. Furthermore the RBF metamodels proved to suffer less from fluctuations in their global predictive capability throughout the optimization procedure.
Given these observations, we believe that the novel robust sequential improvement criterion as well as the proposed implementation of RBF may be successfully used for other applications in the field of robust optimization and metamodel based optimization respectively. Obviously, the effect on the optimization efficiency may be strongly dependent on the optimization problem at hand. Therefore, we believe that there is no general solution for an optimal sequential optimization algorithm. However, knowing the pros and cons of the available building blocks may significantly improve the chances of success in robust optimization.
References
- Alexandrov N, Dennis J, Lewis R, Torczon V (1998) A trust-region framework for managing the use of approximation models in optimization. Struct Optim 15(1):16–23CrossRefGoogle Scholar
- Beyer H G, Sendhoff B (2007) Robust optimization - a comprehensive survey. Comput Methods Appl Mech Eng 196(33–34):3190–3218MathSciNetCrossRefMATHGoogle Scholar
- Coelho R, Bouillard P (2011) Multi-objective reliability-based optimization with stochastic metamodels. Evol Comput 19(4):525–560CrossRefGoogle Scholar
- Fornberg B, Zuev J (2007) The runge phenomenon and spatially variable shape parameters in rbf interpolation. Comput Math Appl 54(3):379–398MathSciNetCrossRefMATHGoogle Scholar
- Forrester A, Keane A (2009) Recent advances in surrogate-based optimization. Progress Aerospace Sci 45 (1–3):50–79CrossRefGoogle Scholar
- Forrester A, Sobester A, Keane A (2008) Engineering design via surrogate modelling: a practical guide. WileyGoogle Scholar
- Franke R (1982) Scattered data interpolation: tests of some method. Math Comput 38(157):181–200MathSciNetMATHGoogle Scholar
- Gano S, Renaud J, Martin J, Simpson T (2006) Update strategies for kriging models used in variable fidelity optimization. Struct Multidiscip Optim 32(4):287–298CrossRefGoogle Scholar
- Gibbs M (1997) Bayesian gaussian processes for regression and classification. University of Cambridge, PhD thesisGoogle Scholar
- Hardy R L (1971) Multiquadric equations of topography and other irregular surfaces. J Geophys Res 76:1905–1915CrossRefGoogle Scholar
- Havinga J, Van Den Boogaard T, Klaseboer G (2013) Sequential optimization of strip bending process using multiquadric radial basis function surrogate models. Key Eng Mater 554–557:911–918CrossRefGoogle Scholar
- Huang D, Allen T, Notz W, Zeng N (2006) Global optimization of stochastic black-box systems via sequential kriging meta-models. J Global Optim 34(3):441–466MathSciNetCrossRefMATHGoogle Scholar
- Janusevskis J, Le Riche R (2013) Simultaneous kriging-based estimation and optimization of mean response. J Global Optim 55(2):313–336MathSciNetCrossRefMATHGoogle Scholar
- Ji Y, Kim S (2013) An adaptive radial basis function method using weighted improvement. pp 957–968Google Scholar
- Jin R, Chen W, Simpson T (2001) Comparative studies of metamodelling techniques under multiple modelling criteria. Struct Multidiscip Optim 23(1):1–13CrossRefGoogle Scholar
- Jones D, Schonlau M, Welch W (1998) Efficient global optimization of expensive black-box functions. J Global Optim 13(4):455–492MathSciNetCrossRefMATHGoogle Scholar
- Jurecka F (2007) Robust design optimization based on metamodeling techniques. PhD thesisGoogle Scholar
- Jurecka F, Ganser M, Bletzinger K U (2007) Update scheme for sequential spatial correlation approximations in robust design optimisation. Comput Struct 85(10):606–614CrossRefGoogle Scholar
- Kitayama S, Yamazaki K (2014) Sequential approximate robust design optimization using radial basis function network. Int J Mech Mater Des 10(3):313–328CrossRefGoogle Scholar
- Kleijnen J (2009) Kriging metamodeling in simulation: a review. Eur J Oper Res 192(3):707–716MathSciNetCrossRefMATHGoogle Scholar
- Koch P, Yang R J, Gu L (2004) Design for six sigma through robust optimization. Struct Multidiscip Optim 26(3–4):235–248CrossRefGoogle Scholar
- Krige D (1951) A statistical approach to some basic mine valuation problems on the witwatersrand. J Chem Metallur Mining Soc South AfricaGoogle Scholar
- Lehman J, Santner T, Notz W (2004) Designing computer experiments to determine robust control variables. Statistica Sinica 14(2):571–590MathSciNetMATHGoogle Scholar
- Leon R V, Shoemaker A C, Kacker R N (1987) Performance measures independent of adjustment: An explanation and extension of taguchi’s signal-to-noise ratios. Technometrics 29(3):253–265MathSciNetCrossRefMATHGoogle Scholar
- Li C, Wang F L, Chang Y Q, Liu Y (2010) A modified global optimization method based on surrogate model and its application in packing profile optimization of injection molding process. Int J Adv Manuf Technol 48 (5–8):505–511CrossRefGoogle Scholar
- Lophaven S N, Nielsen HB, Søndergaard J (2002) DACE, A Matlab Kriging ToolboxGoogle Scholar
- Marzat J, Walter E, Piet-Lahanier H (2013) Worst-case global optimization of black-box functions through kriging and relaxation. J Global Optim 55(4):707–727MathSciNetCrossRefMATHGoogle Scholar
- McNemar Q (1947) Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika 12(2):153–157CrossRefGoogle Scholar
- Ng E, Geller M (1969) A table of integrals of the error functions. U S Bur Standards-J Res Math Sci 73 B (1):1–20MathSciNetMATHGoogle Scholar
- Nikitin I, Nikitina L, Clees T (2012) Nonlinear metamodeling of bulky data and applications in automotive design. Springer, Berlin, pp 295–301MATHGoogle Scholar
- Nishida Y, Kobayashi H, Nishida H, Sugimura K (2013) Performance improvement of a return channel in a multistage centrifugal compressor using multiobjective optimization. J Turbomach 135(3)Google Scholar
- Rhein B, Clees T, Ruschitzka M (2014) Robustness measures and numerical approximation of the cumulative density function of response surfaces. Commun Stat Simul Comput 43(1):1–17MathSciNetCrossRefMATHGoogle Scholar
- Rippa S (1999) An algorithm for selecting a good value for the parameter c in radial basis function interpolation. Adv Comput Math 11(2–3):193–210MathSciNetCrossRefMATHGoogle Scholar
- Rodríguez J, Pérez V, Padmanabhan D, Renaud J (2001) Sequential approximate optimization using variable fidelity response surface approximations. Struct Multidiscip Optim 22(1):24–34CrossRefGoogle Scholar
- Sacks J, Welch W J, Mitchell T J, Wynn H P (1989) Design and analysis of computer experiments. Stat Sci 4(4):409–423MathSciNetCrossRefMATHGoogle Scholar
- Sóbester A, Leary S, Keane A (2004) A parallel updating scheme for approximating and optimizing high fidelity computer simulations. Struct Multidiscip Optim 27(5):371–383CrossRefGoogle Scholar
- Sun G, Song X, Baek S, Li Q (2014) Robust optimization of foam-filled thin-walled structure based on sequential kriging metamodel. Struct Multidiscip Optim 49(6):897–913CrossRefGoogle Scholar
- Taguchi G, Phadke M (1984) Quality engineering through design optimization, pp 1106–1113Google Scholar
- Ur Rehman S, Langelaar M, van Keulen F (2014) Efficient kriging-based robust optimization of unconstrained problems. J Comput Sci 5(6):872–881MathSciNetCrossRefGoogle Scholar
- Wang G, Shan S (2007) Review of metamodeling techniques in support of engineering design optimization. J Mech Des Trans ASME 129(4):370–380CrossRefGoogle Scholar
- Wiebenga J, van den Boogaard A (2014) On the effect of numerical noise in approximate optimization of forming processes using numerical simulations. Int J Mater Forming 7(3):317–335Google Scholar
- Wiebenga J, Van Den Boogaard A, Klaseboer G (2012) Sequential robust optimization of a v-bending process using numerical simulations. Struct Multidiscip Optim 46(1):137–153CrossRefGoogle Scholar
- Wiebenga J, Atzema E, An Y, Vegter H, Van Den Boogaard A (2014) Effect of material scatter on the plastic behavior and stretchability in sheet metal forming. J Mater Process Technol 214(2):238– 252CrossRefGoogle Scholar
- Wiebenga J H (2014) Robust design and optimization of forming processes. PhD thesis, Enschede. http://doc.utwente.nl/91096/ Google Scholar
- Yao W, Chen X, Huang Y, Van Tooren M (2014) A surrogate-based optimization method with rbf neural network enhanced by linear interpolation and hybrid infill strategy. Optim Methods Softw 29(2):406–429MathSciNetCrossRefMATHGoogle Scholar
- Zadeh P, Toropov V, Wood A (2009) Metamodel-based collaborative optimization framework. Struct Multidiscip Optim 38(2):103– 115CrossRefGoogle Scholar
- Zimmerman D, Pavlik C, Ruggles A, Armstrong M (1999) An experimental comparison of ordinary and universal kriging and inverse distance weighting. Mathem Geol 31(4):375–390CrossRefGoogle Scholar
Copyright information
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.















