Skip to main content
Log in

A Hierarchical Method of Parameter Setting for Population-Based Metaheuristic Optimization Algorithms

  • Published:
Journal of Applied and Industrial Mathematics Aims and scope Submit manuscript

Abstract

Metaheuristic algorithms for a global optimization problem have unbound strategy parameters that affect solution accuracy and algorithm efficiency. The task of determining the optimal values of unbound parameters is called the parameter setting problem and can be solved by static parameter setting methods (performed before the algorithm run) and dynamic parameter control methods (performed during the run). The paper introduces a novel hierarchical parameter setting method for the class of population-based metaheuristic optimization algorithms. A distinctive feature of this method is the use of a hierarchical algorithm model. The lower level represents a sequential algorithm from this class, and the upper level represents an algorithm with the parallel island model. Parameter setting is performed by the hierarchical method, which combines parameter tuning for the sequential algorithm and adaptive parameter control for the parallel algorithm. Parameter control is based on vector fitness criteria which consist of a convergence rate and a solution value. An approach for estimating the convergence rate for a multistep optimization method is proposed. Experimental results for CEC benchmark problems are presented and discussed.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1.
Fig. 2.

REFERENCES

  1. K. A. De Jong, Evolutionary Computation: A Unified Approach (Springer, Cambridge, 2006).

    MATH  Google Scholar 

  2. J. Kennedy and R. Eberhart, Swarm Intelligence (Morgan Kaufmann, San Francisco, 2001).

    Google Scholar 

  3. M. Clerc, Particle Swarm Optimization (Wiley-ISTE, London, 2006).

    Book  MATH  Google Scholar 

  4. T. Dokeroglu, E. Sevinc, T. Kucukyilmaz, and E. Cosar, “A survey on new generation metaheuristic algorithms,” Comput. Ind. Eng. 137, (2019). https://doi.org/10.1016/j.cie.2019.106040

  5. T. Bäck, D. Fogel, and Z. Michalewicz, Evolutionary Computation 2: Advanced Algorithms and Operations (Inst. Phys. Publ., Philadelphia, 2000).

    Book  MATH  Google Scholar 

  6. Y. Lorion, T. Bogon, I. J. Timm, and O. Drobnik, “An agent based parallel particle swarm optimization—APPSO,” in IEEE Swarm Intell. Symp. (2009), 52–59. https://doi.org/10.1109/SIS.2009.4937844

  7. B. Niu, Y. Zhu, X. He, and H. Wu, “MCPSO: A multi-swarm cooperative particle swarm optimizer,” Appl. Math. Comput. Spec. Iss. Intell. Comput. Theory Methodol. 185 (2), 1050–1062 (2008). https://doi.org/10.1016/j.amc.2006.07.026

  8. K. De Jong, “Parameter setting in EAs: a 30 year perspective,” Parameter Setting Evol. Algorithms. Stud. Comput. Intell. 54, 1–18 (2007). https://doi.org/10.1007/978-3-540-69432-8_1

  9. S. Smit and A. Eiben, “Comparing parameter tuning methods for evolutionary algorithms,” Proc. 2009 IEEE Congr. Evol. Comput. (2009), 399–406. https://doi.org/10.1109/CEC.2009.4982974

  10. A. E. Eiben and S. K. Smit, “Parameter tuning for configuring and analyzing evolutionary algorithms,” Swarm Evol. Comput. 1, 19–31 (2011). https://doi.org/10.1016/j.swevo.2011.02.001

  11. A. Eiben, Z. Michalewicz, M. Schoenauer, and J. Smith, “Parameter control in evolutionary algorithms,” Parameter Setting Evol. Algorithms. Stud. Comput. Intell. 54, 19–46 (2007). https://doi.org/10.1109/4235.771166

  12. Z. Zhan, J. Zhang, Y. Li, and H. Chung, “Adaptive particle swarm optimization,” IEEE Trans. Syst. Man Cybern. Part B: Cybern. 39 (6), 1362–1381 (2009). https://doi.org/10.1109/TSMCB.2009.2015956

  13. E. Seliverstov and A. Karpenko, “Hierarchical model of parallel metaheuristic optimization algorithms,” Proc. Comput. Sci. Proc. 13 Int. Symp. “Intelligent Systems 2018” (INTELS’18) 150, 441–449 (2019). https://doi.org/10.1016/j.procs.2019.02.075

  14. J. Nocedal and S. Wright, Numerical Optimization (Springer Sci. & Bus. Media, New York, 2006).

    MATH  Google Scholar 

  15. R. Mendes, J. Kennedy, and J. Neves, “The fully informed particle swarm: Simpler, maybe better,” IEEE Trans. Evol. Comput. 8, 204–210 (2004). https://doi.org/10.1109/TEVC.2004.826074

  16. I. Griva, S. Nash, and A. Sofer, Linear and Nonlinear Optimization (Soc. Ind. Appl. Math., Philadelphia, 2008).

    MATH  Google Scholar 

  17. X. Li, K. Tang, M. N. Omidvar, Z. Yang, K. Qin, and H. China, “Benchmark functions for the CEC 2013 special session and competition on large-scale global optimization,” Tech. Rep. Evol. Comput. Mach. Learn. Group, RMIT (2013).

  18. M. R. Bonyadi and Z. Michalewicz, “Analysis of stability, local convergence, and transformation sensitivity of a variant of the particle swarm optimization algorithm,” IEEE Trans. Evol. Comput. 20 (3), 370–385 (2016).

    Article  Google Scholar 

  19. F. V. Bergh and A. P. Engelbrecht, “A convergence proof for the particle swarm optimiser,” Fundam. Inf. 105 (4), 341–374 (2010).

    MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to E. Yu. Seliverstov.

Additional information

Translated by V. Potapchouck

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Seliverstov, E.Y. A Hierarchical Method of Parameter Setting for Population-Based Metaheuristic Optimization Algorithms. J. Appl. Ind. Math. 16, 776–788 (2022). https://doi.org/10.1134/S1990478922040172

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1134/S1990478922040172

Keywords

Navigation