1 Introduction

Optimization has long been a fundamental tool for applications in mathematics, medicine, engineering, the military, and many other fields. In general, optimization is an approach used to determine the best performance of a system that can be achieved based on optimized parameters to make the final solution as perfect and efficient as possible [1]. In optimization problems, the final goal is usually considered the maximization or minimization of some objective function (the function used to evaluate the nature or quality of the resulting solution). In this case, several steps are involved, such as defining a mathematical model for the system, identifying variable, defining the type of system needed, specifying constraints, and then finding the optimization response [2].

Traditional optimization techniques include the dynamic programming approach (DPA), stochastic search, the steepest descent method, and Newton’s method. Most of these methods are nonpolynomial and can find optimal solutions smoothly. Unfortunately, they are usually limited by the specific problem and the size of the given dataset. Furthermore, they are often highly complex and specialized in nature. In addition, the execution times and running costs are very high [3]. Therefore, such approaches are only suitable for solving small-scale problems. Currently, most real-world optimization problems are complex, highly nonlinear, and large-scale spatial, and the information metrics used are essentially multifaceted. On this basis, traditional methods are not suitable for such situations. The popular trend is to combine the multifaceted nature of the subject system with various proposed heuristics to solve difficult problems. Heuristics methods include nearest neighbor algorithms, evolutionary algorithms, insertion algorithms, memetic algorithms, dynamic relaxation, etc. By introducing greedy mechanisms, the number of search steps is fixed, and the number of iterations is reduced. With these advantages, heuristic methods are able to find near-optimal solutions with less time costs than those of other methods [4]. However, most of them are designed for specific problems. Due to the introduction of greedy mechanisms, these methods often fall into local optima when solving complex problems. Soon after the development of heuristic methods, metaheuristics, which are problem-independent adaptive strategies, emerged to provide a set of guidelines or strategies for developing heuristics. In this term, “meta” means using a higher level, usually one level higher than that of the given heuristic, to solve a variety of intractable and complex hard problems [5]. Specifically, metaheuristics perform the search process by using multiple agents, which are essentially systems that find the best solution based on constraints and rules over multiple iterations. The iterative process is stopped when a predefined criterion is satisfied, at which point the solution found is determined to be the best solution and the system reaches a fused state. Metaheuristics are not dependent on gradient information, are easy to execute and are highly flexible because they treat the optimization process as a black box [6]. As research has progressed, metaheuristics have been developed as powerful tools and are now widely used in various fields.

In general, metaheuristic algorithms are classified into four categories, namely, evolutionary-based algorithms, physics-based algorithms, population-based algorithms, and human-based algorithms. Inspired by Darwinian evolutionary theory, evolutionary-based algorithms were introduced in the early 1970s and usually draw from natural rules to find optimal solutions. By setting one or more operators (crossover, variation, or selection operators), such algorithms evaluate search agents during optimization [7]. Evolutionary-based algorithms include genetic algorithms (GA) [8], biogeography-based optimization algorithms (BBO) [9], differential evolutionary algorithms (DE) [10], etc. Physics-based algorithms are modelled according to physical rules. For example, the gravitational search algorithm (GSA) [11] and the equilibrium optimizer (EO) [12] are all physics-based approaches. Population-based algorithms simulate the group behaviour of animals, performing hunting based on cooperation. In population-based algorithms, multiple agents perform the search process together and share useful information with each other, thus improving operational efficiency. Representative algorithms include the artificial bee colony (ABC) algorithm [13], the sailfish optimizer (SFO) [14], whale optimization algorithm (WOA) [15], Harris hawks algorithm (HHO) [16], etc. Human-based algorithms generally simulate the special behaviour of humans. Such algorithms include the harmony search algorithm (HS) [17], teaching-learning-based optimization (TLBO) [18].

Elephant herding optimization (EHO) is a population-based algorithm proposed by Wang et al. [19] . Inspired by the nomadic herding behaviour of elephant populations, EHO is designed to solve global optimization problems. In EHO, an entire population is divided into different clans. Each clan is led by a female matriarch, while the male elephants live independently when they reach adulthood. EHO sets up two operators, a clan updating operator and a separation operator, to update the positions of the agents. A clan updating operator is implemented for each individual, and the next position of each individual is updated according to the current position in relation to the position of the clan leader. Subsequently, the agent with the worst position is replaced by the separation operator to speed up convergence and enhance the diversity of the population. In terms of advantages, EHO has fewer controlled parameters, is easier to implement, and has better performance than other optimization approaches, so it attracts much attention from mathematicians and engineers [20].

In response to the shortcomings of EHO, researchers have proposed many improvement ideas and optimization methods. For instance, Muthusamy et al. [21] introduced the sine cosine algorithm (SCA) and opposition-based learning (OBL) to update the positions of agents in the original EHO algorithm. Through this operation, the convergence speed and the searchability of the algorithm were enhanced. Li and Wang [20] proposed a new improved EHO approach based on dynamic topology and BBO methods. The operators of this EHO variant were updated, and a new separation operator was set, thus ensuring that the evolutionary process of the examined population moved in a better direction. Ismaeel et al. [22] proposed three versions of an unbiased algorithm, EEHO15, EEHO20, and EEHO25, to overcome the defect of unreasonable convergence in the original EHO algorithm. Three similar variants were proposed by Elhosseini et al. [23], namely cultural-based, alpha-tuning, and biased initialization EHO, with the aim of strengthening the global optimization capability of EHO. Balamurugan et al. [24] proposed an opposition-based EHO (OEHO) algorithm. In their work, opposition-based learning was adopted to enhance the performance of EHO. Xu et al. [25] proposed a novel algorithm, LFEHO, which combined Levy flight with EHO to overcome the problems of the original EHO algorithm falling into local optima and exhibiting poor convergence performance. Improved EHO algorithms such as IMEHO were presented by Li et al. [26] The robustness and diversity of a given population are optimized by using a new global learning-based evolution strategy to update the velocities and positions of agents. Reference [27] extended and improved the original EHO algorithm by designing six strategies to update the agents. The information obtained from the previous iterations of individuals was adopted for weighing operations. Therefore, the obtained results were better than those output by the original EHO algorithm. In addition, engineers extended the application of EHO to solve practical engineering problems in real production cycles.

In addition, MRFO is a novel optimization algorithm proposed by Zhao et al. [28]. It is a promising and powerful optimizer inspired by the predatory behaviour of manta rays. MRFO has three foraging operators, namely, chain foraging, cyclone foraging, and tumble foraging operators. Recently, researchers have introduced MRFO to other fields. For example, Ekinci et al. [29] combined the opposition-based (OBL) with MRFO and hybrid simulated annealing algorithm to provides MRFO a better exploration capability. Shaheen et al. [30] optimized the MRFO algorithm by using adaptive penalty parameters and introduced it to handle a thermal scheduling problem. In addition, MRFO has been extended to many fields, such as feature selection (FS) [31], fuel power generation [32], system identification [33], photovoltaic system operation [34], and other areas.

According to the no-free lunch (NFL) theorem [35], any algorithm that is applicable for solving a given problem cannot provide the best solution for all problems. In other words, there is a need to improve the metaheuristic algorithms that exist in reality. Maintaining a balance between exploration and exploitation is the common and fundamental feature of any optimization algorithm and the most challenging problem in metaheuristics. Among them, exploration is the process of exploring optimal solutions over a wide range of unknown regions, while the exploitation of explored regions allows an algorithm to effectively find high-quality solutions. Therefore, it is necessary to fine-tune exploration and exploitation, as they are the two main factors used to obtain satisfactory results. In addition, it has been shown that although some algorithms are very effective in solving certain types of problems or large-scale problems, more significant optimization results can be achieved by combining the advantages of different algorithms in a hybrid manner [36]. By combining them, the convergence performance of the resulting hybrid algorithm is improved and the balance between exploration and exploitation is better maintained, thus allowing the algorithm to avoid falling into local optima. The above facts have motivated researchers to continuously develop hybrid metaheuristic algorithms with good exploration and exploitation capabilities and have made our research highly relevant.

Like most metaheuristic algorithms, EHO is prone to falling into local optima when dealing with complex multipeaked problems, which stems from the stochasticity of such problems and the imbalance between exploration and exploitation. Therefore, these shortcomings become major issues that affect the performance of EHO. At the same time, EHO also has drawbacks such as a poor exploitation capability and a low convergence rate. To summarize, there are three defects in the EHO algorithm that need to be solved.

  1. (a)

    The position of the female matriarch is determined by all members of the clan. Under this circumstance, once the algorithm falls into a local optimum, the matriarch will lose her ability to explore.

  2. (b)

    The location update strategy for individual elephants is fixed and lacks a certain control mechanism for adjustments according to the iterative process. Therefore, it is difficult to strike a balance between exploration and exploitation when dealing with complex problems.

  3. (c)

    In addition, the separation operator does not guarantee that the worst individual in the clan is updated to a better position. To address the shortcomings of EHO, a manta ray foraging and Gaussian mutation-based EHO algorithm for global optimization called MGEHO is proposed. The contributions of this paper are summarized as follows:

    1. 1.

      The position of search agents in EHO is optimized by hybridizing the tumbling foraging strategy in MRFO, which improves the ability of the original EHO algorithm to jump out of local optima. Additionally, the global optimization ability of the original algorithm is enhanced.

    2. 2.

      To maintain a balance between exploration and exploitation, a nonlinear dynamic convergence factor is set to improve the clan updating operator in EHO and the Somersault foraging in MRFO. On this basis, the global exploration and local exploitation processes can be coordinated.

    3. 3.

      Furthermore, Gaussian mutation is introduced to adjust the position of the worst individual of each clan. By using random numbers conforming to a Gaussian distribution instead of the original model parameters, new solutions in the vicinity of the candidate solutions are generated, and the diversity of the population is enhanced.

    4. 4.

      To fully validate its effectiveness, MGEHO is tested on 33 benchmark functions and compared with 9 other well-known metaheuristics. Furthermore, performance tests are conducted on the algorithm in different dimensions. Moreover, nonparametric statistical tests are set up to highlight the differences between MGEHO and other competing approaches. A paired Wilcoxon symbolic rank test and a Friedman test at the 5\(\%\) significance level are invoked to test the compared algorithms. Besides, diversity tests of MGEHO and EHO are performed to verify the improved performance on exploration and exploitation.

    5. 5.

      The performance of the proposed MGEHO is also evaluated and compared with selected advanced algorithms by using 32 modern CEC2014 and CEC2017 benchmark functions. Statistical tests are invoked to analyze the results of means and standard deviations. The Holm test is used as a post hoc test. In addition, the running time analysis is conducted to highlight the fundamental differences between the proposed and existing methods.

    6. 6.

      In addition, the proposed MGEHO algorithm is used to solve three practical engineering problems: (a) a design problem regarding a gear train; (b) a design problem with a three-bar truss; and (c) a design problem with respect to a pressure vessel. The experimental results show that MGEHO has superior adaptability and the ability to handle optimization problems.

The rest of the paper is organized as follows: Sect. 2 reviews the original EHO and MRFO algorithms. Section 3 provides a detail of the proposed MGEHO method. Experiments and the analysis of the results are given in Sect. 4. Tests on practical engineering problems are conducted in Sect. 5. Section 6 summarizes the conclusions of this paper and plans for future work.

2 Preliminary

2.1 EHO

In general, elephants are gregarious creatures. Within a population, they are divided into different clans. Individual elephants from the same clan live together under the leadership of a female matriarch. After reaching adulthood, a fixed number of males leave the clan. EHO simulates these characteristics of elephant herd life by dividing the behaviours of searching agents into two parts, namely, a clan updating operator and separation operator.

2.1.1 Clan updating operator

The position of an elephant individual j in clan \({C_i}\) is updated in the following way:

$$\begin{aligned} {x_{\text {new},{c_i},j}} = {x_{{c_i},j}} + \alpha \times ({x_{\text {best},{c_i}}} - {x_{{c_i},j}}) \times r, \end{aligned}$$
(1)

where \({x_{\text {new},{c_i},j}}\) and \({x_{{c_i},j}}\) are the new and original positions of elephant individual j in clan \({C_i}\), respectively. \(\alpha\) and r are both random numbers within [0,1]. In addition, \({x_{\text {best},{c_i}}}\) is the best individual of clan \({C_i}\), which is updated by the following equation:

$$\begin{aligned} {x_{\text {new},{c_i},j}}= & {} \beta \times {x_{\text {center},{c_i}}},\end{aligned}$$
(2)
$$\begin{aligned} {x_{\text {center},{c_i}}}= & {} \frac{1}{{{n_{{c_i}}}}} \times \sum \limits _{j = 1}^{{n_{{c_i}}}} {{x_{{c_i},j}}}, \end{aligned}$$
(3)

where \({x_{\text {center},{c_i}}}\) represents the central positions of all individuals in clan \({C_i}\) and \({n_{{c_i}}}\) stands for the number of elephants in clan \({C_i}\). \(\beta\) affects the extent to which \({x_{\text {center},{c_i}}}\) acts on \({x_{\text {new},{c_i},j}}\), \(\beta \in [0,1]\).

2.1.2 Separation operator

A separation operator is proposed to update the worst individual in each clan by the following equation:

$$\begin{aligned} {x_{\text {worst},{c_i}}} = {x_{\min }} + ({x_{\max }} - {x_{\min }} + 1) \times r, \end{aligned}$$
(4)

where \({x_{\max }}\) and \({x_{\min }}\) are the upper and lower bounds of the individual positions, respectively. r is a random number within [0,1].

2.1.3 Pseudocode of EHO

In general, the pseudocode of EHO is shown in Algorithm 1.

figure a

2.2 MRFO

In MRFO, three manta ray foraging strategies are simulated, namely chain foraging, cyclone foraging, and somersault foraging. Similar to other metaheuristics, MRFO also performs initialization first and then updates each agent sequentially through iteration to find the best solution.

2.2.1 Chain foraging

In the first stage, manta rays form a foraging chain sequentially from front to back, and the position of each search agent is updated by the current optimal solution and the solutions generated during previous iterations. For a set of individuals \({x_i}\), the process of position updating is represented as follows:

$$\begin{aligned} x_i^d(t + 1) = \left\{ \begin{array}{ll} x_i^d(t) + (r + \alpha ) \times [x_{\text {best}}^d(t) - x_i^d(t)] & i = 1\\ x_i^d(t) + r \times [x_{i - 1}^d(t) - x_i^d(t)] + \alpha \times [x_{\text {best}}^d(t) - x_i^d(t)]& i = 2, \ldots \ldots N \end{array} \right. , \end{aligned}$$
(5)

where \(x_i^d(t)\) represents the position of ith individual at time t in the dth dimension. r represents a random number within [0,1]. N is the size of the population. \(x_{\text {best}}^d\) is the \(\alpha\) stands for a weight coefficient determined in the following way:

$$\begin{aligned} \alpha = 2 \times r \times \sqrt{\left| {\log (r)} \right| }. \end{aligned}$$
(6)

2.2.2 Cyclone foraging

During this stage, manta rays swim forward in a spiral pattern. Unlike the spiral foraging approach in the WOA, each individual not only approaches its prey along the spiral path but also adjusts to the positions of the current and previous best agents and keeps swimming forward. The mathematical model for this process is represented as follows:

$$\begin{aligned} x_i^d(t + 1) = \left\{ \begin{array}{ll} x_{\text {best}}^d + (r + \beta ) \times [x_{\text {best}}^d(t) - x_i^d(t)] &{}\quad i = 1\\ x_{\text {best}}^d + r \times [x_{i - 1}^d(t) - x_i^d(t)] + \beta \times [x_{\text {best}}^d(t) - x_i^d(t)]&{}\quad i = 2, \ldots \ldots N \end{array} \right. . \end{aligned}$$
(7)

The weight coefficient \(\beta\) is expressed as follows:

$$\begin{aligned} \beta = 2{e^{{r_1}\left( \frac{{T - t + 1}}{T}\right) }} \times \sin (2\pi {r_1}), \end{aligned}$$
(8)

where \({r_1}\) is a random number, \({r_1} \in [0,1]\). T denotes the maximum number of iterations. In addition, agents at this stage update their positions by searching for random positions in the space to enhance their exploration abilities in the following manner:

$$\begin{aligned} x_i^d(t + 1) = \left\{ \begin{array}{ll} x_{\text {rand}}^d + (r + \beta ) \times [x_{\text {rand}}^d - x_i^d(t)] & i = 1\\ x_{\text {rand}}^d + r \times [x_{i - 1}^d(t) - x_i^d(t)] + \beta \times [x_{\text {rand}}^d - x_i^d(t)]& i = 2, \ldots \ldots N \end{array} \right. , \end{aligned}$$
(9)

where \(x_{\text {rand}}^d\) denotes a random position in dimension d and \(x_{\text {rand}}^d = x_{\min }^d + r \times (x_{\max }^d - x_{\min }^d)\). \({x_{\max }}\) and \({x_{\min }}\) represent the upper and lower bounds of the search space, respectively.

2.2.3 Somersault foraging

In somersault foraging, the MRFO algorithm treats food as a hub, and each agent makes a flipping motion along the hub. On this basis, agents can update their positions around the current optimal solution to enhance their exploitation capabilities. The mathematical model of this process is shown as follows:

$$\begin{aligned}&x_i^d(t + 1) = x_i^d(t) + S \times [{r_2} \times x_{\text {best}}^d(t) - {r_3} \times x_i^d(t)]\nonumber \\&\quad i = 1,2, \ldots \ldots N, \end{aligned}$$
(10)

where S is the somersault factor, which is used to indicate the range of each flip. In MRFO, the parameter is set to 2. \({r_2}\) and \({r_3}\) are both random numbers within [0,1].

2.2.4 Pseudocode of MRFO

The pseudocode of MRFO is shown in Algorithm 2.

figure b

3 The proposed method

In EHO, the position update strategies of the clan leader and other members are different. Specifically, the position of the patriarch is determined by the action factor \(\beta\) and the central positions of all agents in the same clan during each iteration. When \(\beta\) is small, the patriarch moves near the origin. When \(\beta\) is large, the patriarch suddenly moves to the center of the clan. This kind of strategy lacks guidance towards the global optimal solution, so the algorithm may fall into a local optimum, which affects its convergence efficiency and reduces its exploration ability. Second, the search process of EHO is fixed and lacks an effective way to control the balance between exploration and exploitation, and this largely affects the performance of the algorithm. Third, the worst individuals in the clan are randomly replaced by the separation operator, which makes it difficult to ensure the improvement of population diversity.

Therefore, MGEHO is proposed for the purpose of improving the above shortcomings of EHO. In this section, three aspects of the newly developed strategy are explained in detail.

3.1 Hybrid MRFO with EHO

First, we update the clan updating operator in EHO. Specifically, the somersault foraging strategy is introduced to update the position of the patriarch in each clan, as shown in Fig. 1. According to this novel behaviour, the patriarch can be updated to a position symmetric to the pivot point with the global optimal solution as the pivot point. As the distance between the position of each agent and that of the optimal solution decreases, the fluctuation of the current position decreases, and the range of somersault foraging decreases. Unlike opposition-based learning, somersault foraging is centered around the global optimal solution, so it is more directional, and the convergence rate of the algorithm is further enhanced. An agent that falls into a local optimum can be compared with its updated agent so that it can jump out of the local optimum.

Fig. 1
figure 1

Somersault foraging strategy

Second, after finishing the clan updating and separation operations, chain foraging and cyclone foraging are executed. Later, the elite agents are selected for retention by comparing the fitness values of the agents generated in the ith iteration.

3.2 Nonlinear convergence factor

In the original EHO algorithm, there is a lack of effective control over diversity and convergence, which limits the ability of the algorithm to find the best convergence rate. For this reason, a nonlinear convergence factor is introduced that can change dynamically with the iterative process. In the early iterations of the algorithm, the weights are relatively large, which can enhance the global search ability and diversity of the population. In later iterations, the weights decrease so that the algorithm can focus on the explored areas, improve the local search ability, and accelerate the convergence rate. The convergence factor is expressed as follows:

$$\begin{aligned} \omega = {\omega _l} + {\omega _u} \times {e^{ - 2k\sin \left( \frac{t}{{{t_{\max }}}}\right) }}, \end{aligned}$$
(11)

where \({\omega _u}\) and \({\omega _l}\) are the upper and lower bounds of the convergence factor, respectively. k is the impact factor. When k takes different values, the corresponding changes in the convergence factor are shown in Fig. 2. In the early iterations, the convergence factor is larger and scales down more, which will allow the population to perform a broad exploration. In later iterations, the convergence factor decreases and slows, which allows MGEHO to perform a fine local search. As observed in Fig. 2, when the impact factor k changes, \(\omega\) decreases in a nonlinear manner, and the rate of decrease gradually accelerates as the impact factor increases. The choice of the influencing factor depends on the choices and requirements of the specific problem. On this basis, the position of elephant individual j in clan \({C_i}\) is updated in the following way:

$$\begin{aligned} {x_{\text {new},{c_i},j}} = \omega \times {x_{{c_i},j}} + \alpha \times ({x_{\text {best},{c_i}}} - {x_{{c_i},j}}) \times r. \end{aligned}$$
(12)

In addition, the position of the patriarch in each clan is refreshed in the following manner:

$$\begin{aligned} x_{_{\text {new},{c_i}}}^d(t + 1)&= {} \omega \times x_{_{\text {new},{c_i}}}^d(t) + S\nonumber \\&\times [{r_2} \times x_{\text {best}}^d(t) - {r_3} \times x_{\text {new},{c_i}}^d(t)]. \end{aligned}$$
(13)
Fig. 2
figure 2

Variation curves of the convergence factor

3.3 Gaussian mutation strategy

Gaussian mutation [37] was proposed by Back and Schwefe to improve the search efficiency of heuristic algorithms. Generally, Gaussian mutation generates new solutions in the vicinity of the candidate solutions. It can explore every corner of the search space in small steps during the execution of the search, thus enhancing the population diversity. As a successful strategy, it has been widely used in algorithms such as the GA and PSO [38]. When performing the mutation operation, the original values are replaced with eligible random numbers as the mean and variance of a Gaussian distribution [39]. The expression of the Gaussian distribution function is as follows:

$$\begin{aligned} {f_{(0,{\sigma ^2})}}(\delta ) = \frac{1}{{\sqrt{2\pi {\sigma ^2}} }}{e^{\frac{{ - {\delta ^2}}}{{2{\sigma ^2}}}}}, \end{aligned}$$
(14)

where \({\sigma ^2}\) is the variance value of the candidate solutions. \(\delta\) represents the mean value. After applying a Gaussian distribution to MGEHO, the worst individual position of each clan is updated to explore a larger spatial region and increase the diversity of the candidate solutions. The equation for applying the Gaussian distribution is shown as follows:

$$\begin{aligned} {{{\hat{x}}}_i} = {x_i} \times [1 + \text {Gaussian}(\delta )], \end{aligned}$$
(15)

where \(\text {Gaussian}(\delta )\) is a Gaussian distributed random variable and \({x_i}\) represents the ith candidate solution.

3.4 Pseudocode of MGEHO

The pseudocode of MGEHO is shown in Algorithm 3.

figure c

3.5 Computational complexity of MGEHO

One of the main measures of the optimization algorithm is the execution time. The modification of MGEHO consists of the following three main parts: mixture of MRFO’s operators, adaptive inertia weights, and Gaussian variants. When the population size N, dimension of the optimization problem and maximum number of iterations are set to N, D and T, respectively, the complexity of MGEHO can be calculated by the following way.

In MGEHO, the time used to initialize N search agents is \(O(N*D)\). The time to execute the clan-updating operator with adaptive inertia weights is \(O(N*D)\). Next, the time to execute the separating operator with Gaussian variation is O(N). Finally, the number of iterations is reduced because the MRFO operator is embedded within each iteration and no additional iterations are needed to be performed again. The time required to execute the MRFO operator is \(O(N*D)\). When T iterations are performed, the total time complexity of MGEHO is \(O(N*D+T(N*D+N+N*D)) =O (N*(D+T+ 2*T*D))\), which is higher than the \(O(T*N*D)\) of the basic EHO.

4 Experimental results

In this section, experimental results are presented. To verify the effectiveness and applicability of MGEHO, experiments are conducted on 33 classical benchmark functions, 32 CEC2014 and CEC2017 benchkmark functions. To ensure the fairness of the experiments, each algorithm is run 30 times independently on the benchmark functions to maintain stability. In addition, the population size N for all algorithms is set to 30, and the maximum number of iterations \({t_{\max }}\) is set to 500. Statistical tests are conducted to examine the differences between the different algorithms. The environmental conditions are an Intel (R) Core (TM) i5-9300H CPU @ 2.40 GHz, 16 GB of RAM, and the Windows 10 operating system, and the simulation experiments are set up on the MATLAB R2019b platform.

4.1 Tests on 33 benchmark functions

4.1.1 Benchmark functions

In this subsection, 33 classical benchmark functions from reference [40] are selected, and the basic information regarding these functions is shown in Tables 1, 2 and 3. Among them, F1–F10 are single-peaked functions that have only one global optimal solution in the global range and can test the convergence efficiency and exploration capability of each algorithm. F11–F23 are high-dimensional multipeaked functions that have multiple local extremes in the search space and are used to test the ability of each algorithm to search globally and jump out of local optima. F24–F33 are fixed-dimensional multipeak functions that are used to test the ability of each algorithm to maintain a balance between exploration and exploitation.

Table 1 Details of the unimodal benchmark functions
Table 2 Details of the multimodal benchmark functions
Table 3 Details of the fixed-dimension multimodal benchmark functions

4.1.2 Vaule analysis of the critical parameters

The selection of the parameters can make an impact on the algorithm. \(\omega _l,\omega _u\) and k are three controlled parameters in MGEHO. In this subsection, value tests are performed to verify the effects of different parameters. During the designment of MGEHO, various values of these parameters are set for this experiment. It is worth noting that the interval of \(\omega _l,\omega _u\) and k is continuous and has numerous values. Here, \(\left[ {{\omega _l},{\omega _u}} \right]\) is defined as \(\left\{ {\left[ {0.1,0.7} \right] ,\left[ {0.2,0.8} \right] ,\left[ {0.3,0.9} \right] } \right\}\) and \(k={1,2,3}\). Since there are 3 values for each of the 2 sets of parameters, the full-order factor will have \(3*3=9\) design combinations. The 9 MGEHO variants are named after MGEHO1 to MGEHO9, and the settings of variants are shown in Table 4. Two functions, namely F5, F9, F11, F17, F29, F30, are employed from each category of unimodal, multimodal and fixed dimension multimodal functions The original EHO is selected for comparison with the other 9 different variants. Each algorithm is averaged over 30 independent runs of the fitness of the functions.

Table 4 Combination of MGEHO’s parameters
Table 5 Comparison results of EHO with 9 variants of MGEHO

According to the results in Table 5, the different combinations of the proposed MGEHO outperform the original EHO, and the obtained superiority is proved by Friedman’s test results and the associated p values. Considering the overall ranking, the strongest contender is MGEHO3, followed by MGEHO2 and MGEHO6. In other words, MGEHO with \(\left[ {{\omega _l},{\omega _u}} \right] = \left[ {0.1,0.7} \right]\) and \(k=3\) has the best performance among the nine combinations. Based on that, MGEHO with \(\left[ {{\omega _l},{\omega _u}} \right] = \left[ {0.1,0.7} \right]\) and \(k=3\) is adopted for the subsequent experiments.

4.1.3 Comparison with other algorithms

MGEHO is compared with 9 other metaheuristics. These include EHO [19], enhanced EHO based on the \(\gamma\) value (EEHO15) [22], OEHO [24], LFEHO [25], the IGWO [41], the EO [12], HHO [16], the WOA [15], and the SFO [14]. To make the experimental process fairer and more reliable, the parameters are set within each of the selected algorithms, as shown in Table 6. The other parameters are kept consistent except for those in Table 6.

Table 6 Parameter settings of different algorithms

4.1.4 Scalability test for MGEHO

Aiming to verify the adaptability of MGEHO under different conditions, the dimensions of F1–F23 are expanded from 30, 100 to 200, and the rest of the conditions are kept constant. After the experiments are completed, the mean value (Ave) and standard deviation (Std) are chosen as the evaluation indexes. The mean value reflects the accuracy and quality of a given algorithm, and the standard deviation reflects its stability. When solving a minimization problem, the smaller the mean value is, the better the performance of the algorithm. Similarity, the smaller the standard deviation is, the more stable the operation of the algorithm. The results are presented in Tables 789, and 10. In addition, the best solution in each table is shown in bold.

When the dimensionality of F1–F23 is 30, the means and standard deviations recorded after 30 independent operations are shown in Table 7, where boldface indicates the relative optimal value of the corresponding comparison algorithm. It is clear that MGEHO performs better than the other approaches in general. In the single-peak function test, MGEHO obtains global optimal solutions on F1, F2, F4, F8, and F10 and the best solution among the 10 algorithms on F5 and F9 with the best performance stability. HHO achieves better solutions on F6, and the EO obtains a better solution on F7. This indicates that the exploratory nature of MGEHO is enhanced by incorporating somersault foraging from MRFO. Additionally, by combining the convergence factor and Gaussian perturbation, the population diversity is improved, enabling the algorithm to further develop the exploration space to obtain the best solutions. Regarding the multipeak functions, MGEHO achieves the global optimal solutions on F12, F14, F19, F20, F22, and F23 and outperforms the other algorithms on F13, F17, and F18. This indicates that MGEHO has a strong ability to jump out of local optima; this stems from the clan updating operator and the convergence factor increasing the diversity and randomness of the population.

Table 7 Comparison of results obtained on the F1–F23 benchmark functions (\(d = 30\))
Table 8 Comparison of results obtained on the F24–F33 benchmark functions
Table 9 Comparison of results obtained on the F1–F23 benchmark functions (\(d = 100\))
Table 10 Comparison of results obtained on the F1–F23 benchmark functions (\(d = 200\))

The experimental results on the fixed-dimension multimodal benchmark functions are shown in Table 8. Through the data shown in the table, MGEHO achieves superior solutions on F24, F26, F31, F32, F33 and the best average result on F30. The IGWO performs best on F27 and F28 and runs most stably on F29 and F30. HHO achieves the best result on F25, and the solution obtained on F29 is the closest to the global solution. This indicates that by introducing the nonlinear convergence factor, an algorithm can effectively control the balance between exploration and exploitation by maintaining diversification and intensification throughout the search process.

Table 9 shows the test results of different algorithms on the F1–F23 benchmark functions when the dimensionality is 100. As the dimensionality increases, algorithms tend to need more information to search for optimal solutions; thus, increasing the dimensionality tests the ability of each algorithm to adapt to problems with different variants. The comparison results illustrate that the proposed MGEHO algorithm outperforms the other nine algorithms. Among the 23 benchmark functions, MGEHO surpasses the other algorithms on 17 of them. For F11, EEHO15 can perform better. LFEHO obtains the best solutions on F10 and F23, the IGWO has the best performance on F21, and the EO achieves the global optimal solutions on F12, F14, and F21. HHO has better results than those of the other algorithms on F6, F7, F12, F14, and F15. The SFO is able to provide the best solution on F16, and the WOA has the best results on F12 and F21. In summary, MGEHO has certain reproducibility and stability advantages with respect to solving complex high-dimensional problems.

Table 10 shows the mean and standard deviation results obtained by different algorithms when the dimensionality of F1–F23 is increased to 200. The performances of the algorithms fluctuate differently as the dimensionality increases. Regarding the high-dimensional single-peaked problem, MGEHO finds the best solutions for eight of the ten benchmark functions. In the high-dimensional multipeaked problem, MGEHO outperforms the other algorithms on 9 of the 13 benchmark functions. Therefore, the robustness and accuracy of the original algorithm are improved by mixing MRFO with EHO. It can be concluded that MGEHO is still the best-performing one among the nine other algorithms.

4.1.5 Convergence curve analysis

Convergence curves can reflect the abilities and convergence speeds of algorithms when dealing with optimization problems. The curves obtained with 30 dimensions are shown in Fig. 3. From these plots, it can be observed that MGEHO achieves the best solutions, which are more accurate than those of any other algorithm on both the single-peak and multipeak functions, thereby exhibiting superior performance. The results on the fixed-size functions indicate that MGEHO also ranks among the top algorithms in terms of its ability to balance exploration and exploitation.

Fig. 3
figure 3figure 3

Convergence curves

4.1.6 Statistical analysis

García et al. [42] pointed out that when evaluating the performances of metaheuristic algorithms, it is not sufficient to make comparisons based only on their mean and standard deviation values. At the same time, there are inevitable chance factors that affect the experimental results during the experimental process. To examine the differences between different algorithms, the nonparametric Wilcoxon signed-rank test [43] and Friedman test [44] at the \(5\%\) significance level are invoked in the experiments.

Table 11 Nonparametric statistical comparison based on the Wilcoxon signed-rank test for F1–F33
Table 12 Nonparametric statistical comparison based on the Friedman rank test for F1–F23

The results of the Wilcoxon signed-rank test are shown in Table 11, where the symbols “+/=/–” represent that MGEHO is better than, equal to, or worse than the currently compared algorithm. p values less than 0.05 indicate that there are statistically significant differences in the results. In the table, NaN indicates that the algorithm achieved the theoretically best solution for the corresponding problem; therefore, the comparison is not meaningful. In addition, the bold values in the table indicate values greater than 5\(\%\). According to the experimental results, for the comparison between MGEHO and EHO and its variants, there are no statistically significant differences between MGEHO and EHO on F7 and F24; between MGEHO and EEHO15 on F15 and F16; between MGEHO and OEHO on F15, F16, F24, and F31; and between MGEHO and LFEHO on F10, F30, and F31. According to the comparison with other metaheuristics, there are no statistically significant differences between MGEHO and the IGWO on F15; between MGEHO and the EO on F16 and F25; between MGEHO and HHO on F25; between MGEHO and the SFO on F15 and F27; and between MGEHO and the WOA on F14, F18, and F30. Overall, MGEHO is significantly different from most of the algorithms in terms of the obtained results. In addition, MGEHO outperforms the other algorithms on most of the benchmark functions. To make the results more convincing, the Friedman test at a significance level of \(p = 0.05\) is performed. The differences between the proposed MGEHO algorithm and other algorithms are tested by comparing the average rankings of the different algorithms on all benchmark functions; these rankings are shown in Tables 12 and 13. According to the literature, the algorithm with the lowest ranking is considered the most efficient algorithm. By comparison, MGEHO ranks first with different dimensions (\(d = 30,100,200\)), proving that the proposed algorithm has the largest competitive advantage.

Table 13 Nonparametric statistical comparison based on the Friedman rank test for F24–F33

4.1.7 Diversity analysis of MGEHO

To measure the performance of exploration and exploitation in metaheuristics, some researchers have tried indirect research methods, such as diversity analysis. In terms of the exploration and exploitation process, high diversity means that the solution is traversing and can spread across locations in the search space, thus allowing for extensive exploration. Reduced diversity then means that the agents search within a certain neighborhood and focuses on the process of exploitation. Diversity is calculated by the following equation:

$$\begin{aligned} {\text {Div}_j}= & {} \frac{1}{n}\sum \limits _{i = 1}^n {\left| {\text {median}\left( {{x^j}} \right) - x_i^j} \right| },\end{aligned}$$
(16)
$$\begin{aligned} Div= & {} \frac{1}{m}\sum \limits _{j = 1}^m {Di{v_j}}, \end{aligned}$$
(17)

where Div represents the overall diversity of the population, \({\text {Div}_j}\) is the diversity of the data set consisting of the data in j-th dimension, m is the dimension of the search agent, and n is the population size, \(\text {median} \left( {{x^j}} \right)\) stands for the median of the data in j-th dimension, and \(x_i^j\) is the j-th dimension of i-th individual.

Figure 4 reflects the difference in diversity between the original EHO and MGEHO. The horizontal axis indicates the number of iterations and the vertical axis indicates the average distance between particles. It can be seen that both algorithms exhibit high diversity in the beginning phase due to the random initialization operation. Then, the diversity changes under the influence of different algorithmic mechanisms As the iterations run, the search agents traverse the whole search space and the average distance gradually becomes smaller, making the diversity decrease, and the algorithms gradually turn to the exploitation, which is consistent with the above analysis. In general, the diversity of MGEHO is significantly enhanced based on EHO. The reason lies in the additional exploration and variation strategies applied in MGEHO. These strategies lead to an enhanced level of diversity in the population, and therefore the probability of the algorithm falling into a locally optimal solution during the iterations is reduced.

Fig. 4
figure 4

Diversity curves of MGEHO and EHO

4.1.8 Sensitivity analysis on high-dimensional problems

To further evaluate the effectiveness of MGEHO when optimizing complex high-dimensional problems, 23 benchmarks of F1–F23 are expanded to two higher dimension types, 500, 1000. The varying dimensions are set to evaluate the dimension impact on the algorithm. Under the consideration of fairness, other conditions remain the same as in the previous experiments. Tables 14 and 15 shows the results of MGEHO with other algorithms of 23 benchmark functions on 500 and 1000 dimensions. The best solutions for each function are in bold. Moreover, based on the Friedman test in Table 16, the average ranking of MGEHO is 2.11 and 2.25 on 500 and 1000 dimensions, which are lower than the ranking results of all involved variants of the algorithm. From this aspect, it can be confirmed that MGEHO is superior than the above algorithms in optimizing complex high-dimensional problems.

Table 14 Comparison of results obtained on the F1–F23 benchmark functions (\(d=500\))
Table 15 Comparison of results obtained on the F1–F23 benchmark functions (\(d=1000\))
Table 16 Results of Friedman rank test of high-dimensional problems

4.2 Experiments on CEC2014 and CEC2017 benchmark functions

To further verify the performance of MGEHO when dealing with complex problems, experiments are conducted on the CEC2014 and CEC2017 composite benchmark functions. Besides, seven advanced metaheuristic algorithms were selected for comparison, they are the improved elephant herding optimization (EHOI) [21], the RUNge Kutta optimizer (RUN) [45], the gradient-based optimizer (GBO) [46], the Hunger Games Search (HGS) [47], the Slime mould algorithm (SMA) [48], the an elite evolutionary strategy based Harris hawks optimization (EESHHO) [49], the multi-trial vector-based differential evolution (MTDE) [50]. It should be noted that the parameter settings of these algorithms are selected from the recent reported literature. Moreover, the population size N for all algorithms is set to 30, the spatial dimension is 30, and the maximum number of iterations \({t_{\max }}\) is 1000. The results are based on 30 runs.

4.2.1 Benchmark functions

In this subsection, 32 benchmark functions from CEC2014 and CEC2017 [51] are employed in this section to demonstrate the efficacy of MGEHO. The descriptions of these functions are presented in Table 17, where D represents the dimension of the function and “search range” is the parameter that constrains the frontier of the search space of the corresponding function. In general, these functions cover four categories: F1– F3 and F16–F17 are unimodal functions, F4–F10 and F18 are multimodal functions, F11–F14 and F19–F28 are mixed-state functions, and F29–F32 are compositional functions. It is well known that unimodal functions are suitable for confirming the exploratory ability of the method because there exists only one optimal solution. Unlike unimodal functions, there are a large number of locally optimal solutions in the multimodal case. The number of these solutions increases with dimensionality, which is used to test the capability of the exploitation and jumping out of the local optimum.

Table 17 Details of 32 CEC2014 and CEC2017 benchmark functions

4.2.2 Experimental results and analysis

The comparison results are shown in Table 18. The table reports the mean, standard deviation obtained for 30 runs. The best solutions for each function are in bold. According to the results, it is clear that MGEHO outperforms the comparison algorithm on most of the functions. Compared to the other metaheuristics, MGEHO obtained the first ranking on 18 functions, which is better than any of the comparison algorithms. Although both MGEHO and EHOI are hybrid GWOs, the former is significantly superior to the latter. The main reason is that MGEHO not only uses a mixture of EHO and MRFO as the information sharing method but also utilizes two other information enhancement methods, which contributes to the performance of the algorithm.

Table 18 Results on CEC2014 and CEC2017 benchmark functions

Figure 5 shows the convergence curves of the algorithm on the three functions. Where the horizontal axis represents the number of function iterations and the vertical axis represents the average results obtained by the algorithm in 30 independent runs. It is noted that MGEHO has better accuracy in finding the best results than any of the algorithms, while the convergence rate can be maintained at a faster level. The experiments show that MGEHO has the advantage of handling complex problems.

Fig. 5
figure 5

Convergence curves of MGEHO and advanced MAs

4.2.3 Statistical analysis

The Wilcoxon signed-rank test with a significance level of \(5\%\) was adopted to analyze the comparison results of different algorithms, as shown in Table 20. Where “\(+/=/-\)” refers to the number of MGEHO wins on the function, ties on the function, and loses on the function. From observing the results in the table, it can be seen that the number of times MGEHO outperforms other metaheuristics is 31, 27, 22, 29, 29, 27, 32, respectively. Most of the data in the table are less than the significance level of 0.05, so it can be determined that MGEHO significantly outperforms one of the other seven algorithms.

Another statistical procedure of Holms was also invoked in this research [52]. Holm’s test is a post hoc testing procedure that is typically used to correct for experimental precision in multiple comparisons. MGEHO was chosen as the control algorithm. p value provides information regarding whether the statistical hypothesis test is significant or not, and it also indicates the significance of the results. A smaller p value indicates stronger evidence against the null hypothesis. All algorithms are ranked according to the p value and compared to \(a/(k-i)\), where the significance level is 0.05, k is the degree of freedom, and i is the algorithm number. The test starts with the most significant p value and rejects the null hypothesis sequentially, and whenever a null hypothesis cannot be rejected, the test stops and considers all remaining hypotheses accepted. As can be seen in Table 19, the p values of MTDE, EESHHO, HGS, EHOI, SMA, RUN, and GBO are smaller than their \(a/(k-i)\) values, which means that the equality hypothesis is rejected and there are significant differences between these five algorithms and the control algorithm.

Table 19 Consequence of Holm’s test (MGEHO is the control algorithm)
Table 20 The p values of the Wilcoxon signed rank test obtained on 32 CEC benchmark functions

4.2.4 Running time analysis

To analyze the differences between MGEHO and other metaheuristics, the running time of the algorithm on 32 functions is presented in Table 21. The results of time in the table for each algorithm are based on 30 independent runs. As can be seen from the table, SMA takes the most computation time. While MTDE is the fastest in runtime, and this advantage is absolute. The proposed MGEHO ranks six among the compared algorithms but is faster than RUN and SMA. The comparison reveals that the embedding of different operators implements MGEHO require more computation time in exchange for performance improvement. But overall, this trade-off is acceptable because their time complexity is in the same order of magnitude.

Table 21 Time comparison between MGEHO and other selected algorithms on 32 benchmark functions

5 Applications of MGEHO on engineering problems

There are many cases in real-world engineering that are essentially optimization problems [40]. In this subsection, MGEHO is used to determine solutions to real engineering problems and compared with other algorithms to check the applicability of the proposed algorithm. Three practical engineering problems are selected, namely, a design problem regarding a gear train, a design problem with a three-bar truss, and a design problem with respect to a pressure vessel. To handle optimization constraints, constraint methods need to be integrated into the optimizer. Hybrid methods, repair algorithms, special operators, and penalty functions are applicable [53]. Among them, penalty functions are the simplest and easiest to implement. The death penalty is the most suitable type of penalty function. It can automatically eliminate feasible solutions by constructing the main objective value of a mathematical model and applying heuristic algorithms [54]. Therefore, the three abovementioned mathematical model problems are treated in this way. The population size N is uniformly set to 30, and the maximum number of iterations \({t_{\max }}\) is 500.

5.1 Design problem of a gear train

Proposed by Sandgren [55], this problem is a discrete unconstrained optimization problem. Four variables \({x_1},{x_2},{x_3},{x_4}\) are involved, which represent the number of teeth in each of the four gears of the train. All variables are positive integers and lie in the interval [12, 60]. The main objective of the problem is to minimize the gear ratio cost, and the mathematical model is expressed as follows:

$$\begin{aligned} \begin{array}{l} \min {\qquad \ \ }{F_1}({x_1},{x_2},{x_3},{x_4}) = {\left(\frac{1}{{6.931}} - \frac{{{x_2}{x_3}}}{{{x_1}{x_4}}}\right)^2}\\ \text{where}{\qquad }12 \le {x_1},{x_2},{x_3},{x_4} \le 60 \end{array} \end{aligned}$$
Table 22 The comparison results for the gear train design problem

MGEHO is applied to the problem and compared with EHO, EEHO15, OEHO, LFEHO, the IGWO, the EO, the HHO, the SFO, and the WOA. The results are presented in Table 22. The best solution is in bold. Compared with other algorithms, MGEHO achieves the best solution. The best cost obtained is 2.3694E–16, which is achieved when \({x_1},{x_2},{x_3},{x_4}\) are set to 36.83, 12.04, 12.26, and 27.76, respectively.

5.2 Design problem with a three-bar truss

This problem is a common minimization constraint problem. The objective is to minimize the total weight of three diffeomorphic frames, where the constraints are mainly stress, deflection, and buckling constraints. The structure of the combined diffeomorphic frames is shown in Fig. 6. The mathematical model of this problem is represented as follows:

$$\begin{aligned} \begin{aligned} \min&{\quad }{F_2}({x_1},{x_2}) = (2\sqrt{2} {x_1} + {x_2}) \times L\\ \text {where}&\quad 0 \le {x_1},{x_2} \le 1\\ \text {s.t.}&\quad \left( \frac{{\sqrt{2} {x_1} + {x_2}}}{{\sqrt{2} x_1^2 + 2{x_1}{x_2}}}\right) \times P - \sigma \le 0\\&\quad \left( \frac{{{x_2}}}{{\sqrt{2} x_1^2 + 2{x_1}{x_2}}}\right) \times P - \sigma \le 0\\&\quad \left( \frac{1}{{{x_1} + \sqrt{2} {x_2}}}\right) \times P - \sigma \le 0\mathrm{{ }}\\ L = 100\text{cm}&\quad \sigma ,P = 2\text{kN/cm}^2 \end{aligned}. \end{aligned}$$
Fig. 6
figure 6

Design of a three-bar truss

Table 23 The comparison results for the three-bar truss design problem

After running MGEHO and EHO, EEHO15, OEHO, LFEHO, the IGWO, the EO, the HHO, the SFO, and the WOA independently 30 times each, the best results are obtained, as shown in Table 23. MGEHO obtains the best solution to the problem in bold (263.89578) when \({x_1},{x_2}\) are 0.78157 and 0.42853, respectively. The results show that MGEHO can handle the problem effectively and has superior performance to that of other algorithms.

5.3 Design problem with respect to a pressure vessel

This is a common engineering design problem proposed by Kannan and Kramer [56]. Four variables are involved, namely, the thickness of the shell \({T_s}\), the thickness of the head \({T_h}\), the inner radius of the cylinder R, and the length of the cylinder L. The objective of the problem is to minimize the manufacturing cost of the pressure vessel, as shown in Fig. 7. The problem contains four constraints, three of which are linear, and one is nonlinear. The mathematical form of the problem is as follows:

$$\begin{aligned} \begin{aligned} \min \quad {F_4}(x) &= 0.6224{x_1}{x_3}{x_4} + 1.7781{x_2}x_3^2 + 19.84x_1^2{x_3} + 3.1661x_1^2{x_4}\\ \text {where}&\quad x = [{x_1},{x_2},{x_3},{x_4}] = [{T_s},{T_h},R,L]\\ \text {s.t.}&\quad {g_1}(x) = - {x_1} + 0.0193{x_3} \le 0\\&\quad {g_2}(x) = - {x_2} + 0.00954{x_3} \le 0\\&\quad {g_3}(x) = - \frac{4}{3}\pi x_3^3 - \pi x_3^2{x_4} + 1296000 \le 0\\&\quad {g_4}(x) = {x_4} - 240 \le 0\\&\quad 1 \times 0.0625 \le {x_1},{x_2} \le 99 \times 0.0625\\&\quad 10 \le {x_3},{x_4} \le 200 \end{aligned}. \end{aligned}$$
Fig. 7
figure 7

Design of a pressure vessel

Table 24 The comparison results for the pressure vessel design problem

MGEHO is applied to optimize the problem and compared with nine other algorithms. They are EHO, EEHO15, OEHO, LFEHO, the IGWO, the EO, the HHO, the SFO, and the WOA, respectively. The obtained experimental results are shown in Table 24. The best solution is in bold.  According to the data, MGEHO achieves the best solution among those output by the 10 algorithms, with 0.79386, 0.38794, 40.63038, 195.71868 as the four variable values and 5966.37963 as the best cost.

6 Conclusion

In this paper, a new hybrid metaheuristic algorithm is proposed, named MGEHO. MGEHO is adapted to address the weaknesses of the original EHO algorithm, such as its low accuracy and poor convergence rate. The clan updating operator in EHO is optimized by incorporating the somersault foraging strategy from MRFO, and the patriarch position updating method is optimized. On this basis, the ability of MGEHO to jump out of local optima and achieve global optimization is effectively improved. Furthermore, a nonlinear convergence factor, which can vary with the operating state, is set. This strategy contributes to coordinating global searches and achieving a balance between exploration and exploitation. To improve the diversity of the given population, Gaussian mutation is also introduced to EHO to adjust the position of the worst individual in each clan. In addition, MGEHO is tested on 33 classical benchmark functions and compared with EHO, EEHO15, OEHO, LFEHO, the IGWO, the EO, HHO, the SFO, and the WOA. The experiment on 32 modern CEC2014 and CEC2017 benchmark functions is also performed among MGEHO, EHOI, RUN, GBO, HGS, SMA, EESHHO, and MTDE. The results show that MGEHO is more effective and stable than the competing algorithms. Additionally, MGEHO is applied to three specific engineering practice problems. According to the results, MGEHO obtains better solutions than those of the compared algorithms, verifying the ability of MGEHO to solve engineering optimization problems effectively.

In future work, MGEHO will be applied to multilevel threshold image segmentation and will transformed into a binary version of a feature selection method. Moreover, MGEHO is also intended to be combined with additional models and methods in machine learning, such as support vector machines, unsupervised learning, artificial neural networks, and convolutional neural networks.