Skip to main content
Log in

Fortified Cuckoo Search Algorithm on training multi-layer perceptron for solving classification problems

  • Original Article
  • Published:
Personal and Ubiquitous Computing Aims and scope Submit manuscript

Abstract

Multi-layer perceptron (MLP) in artificial neural networks (ANN) is one among the trained neural models which can hold several layers as a hidden layer for intensive training to obtain optimal results. On the other hand, the classification problem has a high level of attraction towards researchers to increase the accuracy in classification. In ANN, feedforward neural network (FNN) is one model that possesses the art of solving classification and regression problems. When input data is given to FNN, it will apply the sum of product rule and the activation function to map the input with its appropriate output. In the sum of product rule, a term called weights is to be chosen appropriately to map between the input and output. In standard FNN, the weights are chosen in a random way which may lead to slower convergence towards the optimal choice of weight values. In this paper, an effective optimization model is proposed to optimize the weights of MLP of FNN for effective classification problems. Four different datasets were chosen, and the results are interpreted with statistical performance measures.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Algorithm 1
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Mangasarian OL, Wolberg WH (1990) Cancer diagnosis via linear programming. University of Wisconsin-Madison, Computer Sciences Department

    MATH  Google Scholar 

  2. Rosenblatt F (1957) The perceptron, a perceiving and recognizing automaton project para. Cornell Aeronautical Laboratory

    Google Scholar 

  3. McCulloch WS, Pitts W (1943) A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biol. 5:115 133

    MathSciNet  MATH  Google Scholar 

  4. Thirugnanasambandam K, Prakash S, Subramanian V et al (2019) Appl Intell 49:2059. https://doi.org/10.1007/s10489-018-1355-3

    Article  Google Scholar 

  5. Werbos P (1974) Beyond regression: new tools for prediction and analysis in the behavioral sciences, Ph.D. thesis,. Harvard University

    Google Scholar 

  6. Patricia Melin P, Sanchez D, Castillo O (2012) Genetic optimization of modular neural networks with fuzzy response integration for human recognition. Inform Sci 197:1–19. https://doi.org/10.1016/j.ins.2012.02.027

    Article  Google Scholar 

  7. Guo ZX, Wong WK, Li M (2012) Sparsely connected neural network-based time series forecasting. Inform. Sci 193:54–71. https://doi.org/10.1016/j.ins.2012.01.011(15.06.12)

    Article  Google Scholar 

  8. Gardner MW, Dorling SR (1998) Artificial neural networks (the multilayer perceptron)—a review of applications in the atmospheric sciences. Atmos Environ 32:2627–2636

    Article  Google Scholar 

  9. Barakat M, Lefebvre D, Khalil M, Druaux F, Mustapha O (2013) Parameter selection algorithm with self adaptive growing neural network classifier for diagnosis issues. Int J Mach Learn Cybern 4(3):217–233

    Article  Google Scholar 

  10. Csáji BC (2001) Approximation with artificial neural networks. Etvs Lornd University, Hungary, Faculty of Sciences

    Google Scholar 

  11. Hornik K, Stinchcombe M, White H (1989) Multilayer feedforward networks are universal approximators. Neural Networks 2:359–366

    Article  MATH  Google Scholar 

  12. Auer P, Burgsteiner H, Maass W (2008) A learning rule for very simple universal approximators consisting of a single layer of perceptrons. Neural Netw 21:786–795

    Article  MATH  Google Scholar 

  13. Reed RD, Marks RJ (1998) Neural smithing: supervised learning in feedforward artificial neural networks. MIT Press

    Google Scholar 

  14. Werbos PJ (1992) Neurocontrol and supervised learning: an overview and evaluation. Handbook Intell Control 65:89

    Google Scholar 

  15. Oja E (2002) Unsupervised learning in neural computation. Theoret Comput Sci 287:187–207

    Article  MathSciNet  MATH  Google Scholar 

  16. Sejnowski TJ (1999) Unsupervised learning: foundations of neural computation. The MIT Press

    Google Scholar 

  17. Hush DR, Horne BG (1993) Progress in supervised neural networks. IEEE Signal Process Mag 10:8–39

    Article  Google Scholar 

  18. Adeli H, Hung S (1994) An adaptive conjugate gradient learning algorithm for efficient training of neural networks. Appl Math Comput 62:81–102

    MATH  Google Scholar 

  19. Charalambous C (1992) Conjugate gradient algorithm for efficient training of artificial neural networks. IET, pp 301–310

    Google Scholar 

  20. Hagan MT, Menhaj MB (1994) Training feedforward networks with the Marquardt algorithm. IEEE Trans Neural Netw 5:989–993

    Article  Google Scholar 

  21. Zhang N (2009) An online gradient method with momentum for two-layer feedforward neural networks. Appl. Math. Comput. 212:488–498

    MathSciNet  MATH  Google Scholar 

  22. S.E. Fahlman, An empirical study of learning speed in back-propagation networks, Technical report, 1988, <http://repository.cmu.edu/cgi/viewcontent.cgi?article=2799&context=compsci>.

  23. Vogl TP, Mangis J, Rigler A, Zink W, Alkon D (1988) Accelerating the convergence of the back-propagation method. Biol Cybern 59:257–263

    Article  Google Scholar 

  24. Ng S, Cheung C, Leung S, Luk A (2003) Fast convergence for backpropagation network with magnified gradient function, vol 3. IEEE, pp 1903–1908

    Google Scholar 

  25. Gori M, Tesi A (1992) On the problem of local minima in backpropagation. IEEE Trans Pattern Anal Mach Intell 14:76 86

    Article  Google Scholar 

  26. Lee Y, Oh SH, Kim MW (1993) An analysis of premature saturation in back propagation learning. Neural Netw 6:719 728

    Article  Google Scholar 

  27. Magoulas G, Vrahatis M, Androulakis G (1997) On the alleviation of the problem of local minima in back-propagation. Nonlinear Anal 30:4545–4550

    Article  MathSciNet  MATH  Google Scholar 

  28. Jacobs RA (1988) Increased rates of convergence through learning rate adaptation. Neural Netw 1:295–307

    Article  Google Scholar 

  29. van Ooyen A, Nienhuis B (1992) Improving the convergence of the back-propagation algorithm. Neural Netw 5:465–471

    Article  Google Scholar 

  30. Weir MK (1991) A method for self-determination of adaptive learning rates in back propagation. Neural Netw 4:371 379

    Article  Google Scholar 

  31. Gudise VG, Venayagamoorthy GK (2003) Comparison of particle swarm optimization and backpropagation as training algorithms for neural networks. IEEE, pp 110–117

    Google Scholar 

  32. Alweshah M, Alkhalaileh S, Albashish D, Mafarja M, Bsoul Q,Dorgham O (2020) A hybrid mine blast algorithm for feature selection problems. Soft Comput:1–18

  33. Alweshah M (2020) Solving feature selection problems by combining mutation and crossover operations with the monarch butterfly optimization algorithm. Appl Intell:1–24

  34. Agrawal P, Ganesh T, Mohamed AW (2020 Oct) A novel binary gaining–sharing knowledge-based optimization algorithm for feature selection. Neural Comput Appl. 26:1–20

    Google Scholar 

  35. Zhang J-R, Zhang J, Lok T-M, Lyu MR (2007) A hybrid particle swarm optimization–back-propagation algorithm for feedforward neural network training. Appl Math Comput 185(2):1026–1037

    MATH  Google Scholar 

  36. Blake C (1998) UCI repository of machine learning databases. http://www.ics.uci.edu/~mlearn/MLRepository.html

  37. Wdaa A, Sttar I, Sttar A (2008) Differential evolution for neural networks learning enhancement. PhD diss., Universiti Teknologi Malaysia

  38. Mirjalili S (2011) Hybrid particle swarm optimization and gravitational search algorithm for multilayer perceptron learning. Diss Universiti Teknologi Malaysia

  39. Fine TL (1999) Feedforward neural network methodology. Springer Verlag

    MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to R.S. Raghav.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Thirugnanasambandam, K., Prabu, U., Saravanan, D. et al. Fortified Cuckoo Search Algorithm on training multi-layer perceptron for solving classification problems. Pers Ubiquit Comput 27, 1039–1049 (2023). https://doi.org/10.1007/s00779-023-01716-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00779-023-01716-1

Keywords

Navigation