Skip to main content

A Novel Method for Generating Benchmark Functions Using Recurrent Neural Network

  • Conference paper
  • First Online:
Intelligent Computing Theories and Application (ICIC 2017)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 10361))

Included in the following conference series:

  • 3066 Accesses

Abstract

In recent years numerous evolutionary algorithms have been proposed to optimize multi–modal problems. These algorithms test the performance by benchmark functions for simulating real-world problems. However, the benchmark functions don’t have enough similarity and complexity compared to real world. Thus, Recurrent Benchmark Generator (RBG) is proposed in this paper to generate complex and different benchmark functions. This generator obtains a mass of modals by recurrent neural network, which are added various fluctuations of normal benchmark functions to keep a balance between complexity and gradient. The experimental results indicate that the novel approach produces more complex benchmark functions which are more conformed to real world problems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Holland, J.H.: Adaptation in Natural and Artificial Systems. University of Michigan Press, Ann Arbor (1975)

    Google Scholar 

  2. Eberhart, R., Kennedy, J.: A new optimizer using particle swarm theory. In: Proceedings of the Sixth International Symposium on Micro Machine and Human Science, pp. 39–43. IEEE (1995)

    Google Scholar 

  3. Wang, L., Yang, B., Orchard, J.: Particle swarm optimization using dynamic tournament topology. Appl. Soft Comput. 48, 584–596 (2016)

    Article  Google Scholar 

  4. Wang, L., Yang, B., Abraham, A.: Distilling middle-age cement hydration kinetics from observed data using phased hybrid evolution. Soft. Comput. 20(9), 3637–3656 (2016)

    Article  Google Scholar 

  5. Wang, L., Yang, B., Chen, Y., Zhang, X., Orchard, J.: Improving neural-network classifiers using nearest neighbor partitioning. IEEE Trans. Neural Netw. Learn. Syst. (2016, in Press). doi:10.1109/TNNLS.2016.2580570

  6. Qu, B.Y., Liang, J.J., Wang, Z.Y., Chen, Q., Suganthan, P.N.: Novel benchmark functions for continuous multimodal optimization with comparative results. Swarm Evol. Comput. 26, 23–34 (2016)

    Article  Google Scholar 

  7. Suganthan, P.N., Hansen, N., Liang, J.J., Deb, K., Chen, Y.P., Auger, A., Tiwari, S.: Problem definitions and evaluation criteria for the CEC 2005 special session on real-parameter optimization. KanGAL report, 2005005 (2005)

    Google Scholar 

  8. Li, T., Rogovchenko, Y.V.: Oscillation criteria for even-order neutral differential equations. Appl. Math. Lett. 61, 35–41 (2016)

    Article  MathSciNet  MATH  Google Scholar 

  9. Li, T., Rogovchenko, Y.V.: Oscillation of second-order neutral differential equations. Math. Nachr. 288(10), 1150–1162 (2015)

    Article  MathSciNet  MATH  Google Scholar 

  10. Mirjalili, S., Lewis, A.: Obstacles and difficulties for robust benchmark problems: a novel penalty-based robust optimisation method. Inf. Sci. 328, 485–509 (2016)

    Article  Google Scholar 

  11. Pineda, F.J.: Generalization of back-propagation to recurrent neural networks. Phys. Rev. Lett. 59(19), 2229 (1987)

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgements

This work was supported by National Natural Science Foundation of China under Grant No. 61572230, No. 61573166, No. 61373054, No. 61472164, No. 61472163, No. 61672262, No. 61640218, Shandong Provincial Natural Science Foundation, China, under Grant ZR2015JL025, ZR2014JL042. Science and technology project of Shandong Province under Grant No. 2015GGX101025, Project of Shandong Province Higher Educational Science and Technology Program under Grant no. J16LN07. Shandong Provincial Key R&D Program under Grant No. 2016ZDJS01A12, No. 2016GGX101001.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lin Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Sun, F., Wang, L., Yang, B., Zhou, J., Chen, Z. (2017). A Novel Method for Generating Benchmark Functions Using Recurrent Neural Network. In: Huang, DS., Bevilacqua, V., Premaratne, P., Gupta, P. (eds) Intelligent Computing Theories and Application. ICIC 2017. Lecture Notes in Computer Science(), vol 10361. Springer, Cham. https://doi.org/10.1007/978-3-319-63309-1_68

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-63309-1_68

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-63308-4

  • Online ISBN: 978-3-319-63309-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics