Skip to main content
Log in

Optimization using lion algorithm: a biological inspiration from lion’s social behavior

  • Special Issue
  • Published:
Evolutionary Intelligence Aims and scope Submit manuscript

Abstract

Nature-inspired optimization algorithms, especially evolutionary computation-based and swarm intelligence-based algorithms are being used to solve a variety of optimization problems. Motivated by the obligation of having optimization algorithms, a novel optimization algorithm based on a lion’s unique social behavior had been presented in our previous work. Territorial defense and territorial takeover were the two most popular lion’s social behaviors. This paper takes the algorithm forward on rigorous and diverse performance tests to demonstrate the versatility of the algorithm. Four different test suites are presented in this paper. The first two test suites are benchmark optimization problems. The first suite had comparison with published results of evolutionary and few renowned optimization algorithms, while the second suite leads to a comparative study with state-of-the-art optimization algorithms. The test suite 3 takes the large-scale optimization problems, whereas test suite 4 considers benchmark engineering problems. The performance statistics demonstrate that the lion algorithm is equivalent to certain optimization algorithms, while outperforming majority of the optimization algorithms. The results also demonstrate the trade-off maintainability of the lion algorithm over the traditional algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2

Similar content being viewed by others

References

  1. Rozenberg G, Bck T, Kok JN (2011) Handbook of natural computing, 1st edn. Springer Publishing Company, New York

    Google Scholar 

  2. Yang XS, Deb S, Fong S, He X, Zhao YX (2016) From swarm intelligence to metaheuristics: nature-inspired optimization algorithms. Computer 49(9):52–59

    Article  Google Scholar 

  3. Shadbolt N (2004) Nature-inspired computing. IEEE J Intell Syst 19(1):2–3

    Article  Google Scholar 

  4. Neumann F, Witt C (2010) Bioinspired computation in combinatorial optimization algorithms and their computational complexity. Natural computing series, XII, p 216

    Chapter  Google Scholar 

  5. Corne D, Deb K, Knowles J, Yao X (2010) Selected applications of natural computing. In: Rozenberg G, Back T, Kok JN (eds) Handbook of natural computing. Springer, Berlin

    Google Scholar 

  6. Holland JH (1975) Adaptation in natural and artificial systems. University of Michigan Press, Ann Arbor

    Google Scholar 

  7. Bongard J (2009) Biologically inspired computing. IEEE Comput J 42(4):95–98

    Article  Google Scholar 

  8. Forbes N (2000) Biologically inspired computing. Comput Sci Eng 2(6):83–87

    Article  Google Scholar 

  9. Mahdiani HR, Ahmadi A, Fakhraie SM, Lucas C (2010) Bio-inspired imprecise computational blocks for efficient VLSI implementation of soft-computing applications. IEEE Trans Circuits Syst I Regul Pap 57(4):1549–8328

    Article  MathSciNet  Google Scholar 

  10. Rajakumar BR (2012) The lion’s algorithm: a new nature-inspired search algorithm. In: Second international conference on communication, computing and security, vol 6, pp 126–135. https://doi.org/10.1016/j.protcy.2012.10.016

    Article  Google Scholar 

  11. Rajakumar BR (2014) Lion algorithm for standard and large scale bilinear system identification: a global optimization based on lion’s social behavior. In: 2014 IEEE congress on evolutionary computation (CEC), pp 2116–2123

  12. Chander S, Vijaya P, Dhyani P (2016) ADOFL: multi-kernel-based adaptive directive operative fractional lion optimisation algorithm for data clustering. J Intell Syst 27:317

    Article  Google Scholar 

  13. Babers R, Hassanien AE, Ghali NI (2015) A nature-inspired metaheuristic lion optimization algorithm for community detection. In: 2015 11th IEEE international computer engineering conference (ICENCO)

  14. Chander S, Vijaya P, Dhyani P (2017) Multi kernel and dynamic fractional lion optimization algorithm for data clustering. Alex Eng J 57:267

    Article  Google Scholar 

  15. Bauer H, Iongh de HH, Silvestre I (2003) Lion social behaviour in the West and Central African Savanna belt. Mamm Biol 68(1):239–243

    Article  Google Scholar 

  16. Fogel LJ, Owens AJ, Walsh MJ (1966) Artificial intelligence through simulated evolution. Wiley Publishing, New York

    MATH  Google Scholar 

  17. Doerr B, Happ E, Klein C (2012) Crossover can probably be useful in evolutionary computation. Theor Comput Sci 425:17–33

    Article  Google Scholar 

  18. Back T, Hoffmeister F, Schwefel HP (1993) An overview of evolutionary algorithms for parameter optimization. J Evol Comput 1(1):1–24 (De Jong K (ed), Cambridge: MIT Press)

    Article  Google Scholar 

  19. Jong De KA (1975) An analysis of the behavior of a class of genetic adaptive systems. Doctoral thesis, Dept. Computer and Communication Sciences, University of Michigan, Ann Arbor

  20. Packer C, Pusey AE (2016) Divided we fall: cooperation among lions. Sci Am 276:52–59

    Article  Google Scholar 

  21. Packer C, Pusey AE (1982) Cooperation and competition within coalitions of male lions: Kin selection or game theory? Nature 296(5859):740–742

    Article  Google Scholar 

  22. Grinnell J, Packer C, Pusey AE (1995) Cooperation in male lions: Kinship, reciprocity or mutualism? Anim Behav 49(1):95–105

    Article  Google Scholar 

  23. Packer C, Pusey AE (1982) Cooperation and competition within coalition of male lions: Kin selection or game theory. Macmillan J 296(5859):740–742

    Google Scholar 

  24. Lotfi E, Akbarzadeh-T MR (2016) A winner-take-all approach to emotional neural networks with universal approximation property. Inform Sci 346:369–388

    Article  Google Scholar 

  25. He S, Wu QH, Saunders JR (2009) Group search optimizer: an optimization algorithm inspired by animal searching behavior. IEEE Trans Evol Comput 13(5):973–990

    Article  Google Scholar 

  26. Karaboga D, Akay B (2009) A comparative study of artificial bee colony algorithm. Appl Math Comput 214(1):108–132

    MathSciNet  MATH  Google Scholar 

  27. Yao X, Liu Y, Lin G (1999) Evolutionary programming made faster. IEEE Trans Evol Comput 3(2):82–102

    Article  Google Scholar 

  28. Fogel DB (1995) Evolutionary computation: toward a new philosophy of machine intelligence. IEEE Press, New York

    MATH  Google Scholar 

  29. Fogel LJ, Owens AJ, Walsh MJ (1965) Artificial intelligence through a simulation of evolution. In: Proc. 2nd cybern. sci. symp. biophysics cybern. syst., Washington: Spartan Books, pp 131–155

  30. Schwefel HP (1995) Evolution and optimum seeking. Wiley, New York

    MATH  Google Scholar 

  31. Yao X, Liu Y (1997) Fast evolution strategies. Control Cybern 26(3):467–496

    MathSciNet  MATH  Google Scholar 

  32. Hansen N, Ostermeier A (1996) Adapting arbitrary normal mutation distributions in evolution strategies: the covariance matrix adaptation. In: IEEE int. conf. evolution. comput. (ICEC) proc, pp 312–317

  33. Hansen N (2006) The CMA evolution strategy: a comparing review. In: Lozano JA, Larraaga P, Inza I, Bengoetxea E (eds) Towards a new evolutionary computation. Springer, Berlin

    Google Scholar 

  34. Hedar A, Fukushima M (2006) Evolution strategies learned with automatic termination criteria. In: Proceedings of SCIS-ISIS 2006, Tokyo, Japan

  35. Goldberg DE (1989) Genetic algorithms in search optimization and machine learning. Addison Wesley, Reading, p 41

    Google Scholar 

  36. Kennedy J, Eberhart R (1995) Particle swarm optimization. In: Proceedings of IEEE international conference on neural networks, IV, pp 1942–1948. https://doi.org/10.1109/ICNN.1995.488968

  37. Karaboga D (2005) An idea based on honey bee swarm for numerical optimization. Technical Report-TR06, Erciyes University, Engineering Faculty, Computer Engineering Department

  38. Wolpert DH, Macready WG (1997) No free lunch theorems for optimization. IEEE Trans Evol Comput 1(1):67–82

    Article  Google Scholar 

  39. Cheng YP, Li Y, Wang G, Zheng Y-F, Cui XT (2017) A novel bacterial foraging optimization algorithm for feature selection. Expert Syst Appl 83:1–17

    Article  Google Scholar 

  40. Yang X-S, Deb S (2009) Cuckoo search via Lévy flights. In: World congress on nature and biologically inspired computing (NaBIC 2009), IEEE Publications, pp 210–214

  41. Yang X-S (2010) Nature inspired metaheuristic algorithms, 2nd edn. Luniver Press, London

    Google Scholar 

  42. Mirjalili S (2015) Moth-flame optimization algorithm: a novel nature-inspired heuristic paradigm. Knowl Based Syst 89:228–249

    Article  Google Scholar 

  43. Mirjalili S (2016) Dragonfly algorithm: a new meta-heuristic optimization technique for solving single-objective, discrete, and multi-objective problems. J Neural Comput Appl 27(4):1053–1073

    Article  Google Scholar 

  44. Mirjalili SM, Mirjalili A, Lewis (2014) Grey wolf optimizer. Adv Eng Softw 69:46–61

    Article  Google Scholar 

  45. Mirjalili S, Lewis A (2016) The whale optimization algorithm. Adv Eng Softw 95:51–67

    Article  Google Scholar 

  46. Askarzadeh A (2016) A novel metaheuristic method for solving constrained engineering optimization problems: crow search algorithm. Comput Struct 169:1–12

    Article  Google Scholar 

  47. Li LL, Yang YF, Wang C-H, Lin K-P (2018) Biogeography-based optimization based on population competition strategy for solving the substation location problem. Expert Syst Appl 97:290–302

    Article  Google Scholar 

  48. Price VK, Storn MR (1997) Differential evolution: a simple evolution strategy for fast optimization. Dr Dobb’s J 22:18–24

    MATH  Google Scholar 

  49. Rashedi E, Nezamabadi-Pour H, Saryazdi S (2009) GSA: a gravitational search algorithm. Inf Sci 179(13):2232–2248

    Article  Google Scholar 

  50. Geem ZW, Kim JH, Loganathan GV (2001) A new heuristic optimization algorithm: harmony search. Simulation 76:60–68

    Article  Google Scholar 

  51. Ingber L (1993) Simulated annealing: practice versus theory. Math Comput Model 18(11):29–57

    Article  MathSciNet  Google Scholar 

Download references

Funding

None.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Rajakumar Boothalingam.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

This article does not contain any studies with human participants or animals performed by any of the authors.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

Definition 1

An \({X^{cub}}\) of an \({X^{male}}\) and an \({X^{female}}\) is the sum of Hadamard product of crossover mask and \({X^{male}}\) and Hadamard product of complement of the same crossover mask and \({X^{female}}\), provided the crossover mask is essentially to be a binary vector with \({C_r} \cdot L\) number of binary ones.

Assumption 1

Within a pride, male lions are always stronger than female lions, applicable to cubs also.

Lemma 1

If the product of \({C_r}\) and \(L\) is equal to zero, then \({X^{cubs}}\) are equal to \({X^{female}}\).

Proof of Lemma 1

According to Definition 1, \(B\), which is a vector with \(L\) elements, has \({C_r} \cdot L\)number of ones and\(L\left( {1 - {C_r}} \right)\) number of zeros. If \({C_r} \cdot L=0\), then \(B\) has no ones and \(L\) number of zeros, which means \(B\) is a vector of zeros. By applying this \(B\) in the mathematical representation of crossover operation given in [11], the first term of RHS becomes \(B\) and the second term of RHS becomes \({X^{female}}\) as \(\bar {B}=1 - B\). Hence, \({X^{cubs}}\) becomes \({X^{female}}\) when \({C_r} \cdot L=0\).

Lemma 2

If the product of \({C_r}\) and \(L\) is equal to one, then \({X^{cubs}}\) are equal to \({X^{male}}\).

Proof of Lemma 2

Similar to the proof of Lemma 1, \(B\) is a vector of ones, when \({C_r} \cdot L=L\) and \(\bar {B}\) becomes a vector of zeros. This in turn, makes the first RHS term of Eq. (11) as \({X^{male}}\) and the second RHS term as a vector of zeros. Hence, \({X^{cubs}}\) become \({X^{male}}\) when \({C_r} \cdot L=L\).

Lemma 3

\(E_{1}^{{nomad}}\) is greater than \(E_{2}^{{nomad}}\), if \({d_1}\) and \(f\left( {X_{1}^{{nomad}}} \right)\) are greater than \({d_2}\) and \(f\left( {X_{2}^{{nomad}}} \right)\), respectively.

Proof of Lemma 3

According to the lemma,

$$\text{max} \left( {{d_1},{d_2}} \right)={d_1}$$
(16)
$$\text{max} \left( {f\left( {X_{1}^{{nomad}}} \right),f\left( {X_{2}^{{nomad}}} \right)} \right)=f\left( {X_{1}^{{nomad}}} \right)$$
(17)

Hence, Eqs. (24) and (25) (from Theorem 2) takes the form,

$$E_{1}^{{nomad}}=\exp \left( 1 \right)=2.71$$
(18)
$$E_{2}^{{nomad}}=\exp \left( {\frac{{{d_2}}}{{{d_1}}}} \right)\frac{{f\left( {X_{1}^{{nomad}}} \right)}}{{f\left( {X_{2}^{{nomad}}} \right)}}.$$
(19)

\(E_{2}^{{nomad}}\) can be greater than \(\exp \left( 1 \right)\)only if \(f\left( {X_{1}^{{nomad}}} \right)>>f\left( {X_{2}^{{nomad}}} \right)\) as the first term produces exponential decay from \(\exp \left( 1 \right)\), when \({d_1}>{d_2}\). But, if \(f\left( {X_{1}^{{nomad}}} \right)>>f\left( {X_{2}^{{nomad}}} \right)\), then probably \({d_1}>>{d_2}\)due to its linear relationship with \(f\left( {X_{1}^{{nomad}}} \right)\), which leads \(\exp \left( \bullet \right)\) towards zero, and hence \(E_{2}^{{nomad}}\) becomes lesser than \(\exp \left( 1 \right)\). Thus it is proved that \(E_{1}^{{nomad}}>E_{2}^{{nomad}}\), when \({d_1}>{d_2}\) and\(f\left( {X_{1}^{{nomad}}} \right)>f\left( {X_{2}^{{nomad}}} \right)\).

Lemma 4

\(E_{1}^{{nomad}}\) is greater than \(E_{2}^{{nomad}}\), if \({d_1}\) and \(f\left( {X_{2}^{{nomad}}} \right)\) are greater than \({d_2}\) and \(f\left( {X_{1}^{{nomad}}} \right)\), respectively.

Proof of Lemma 4

According to the lemma,

$$\text{max} \left( {{d_1},{d_2}} \right)={d_1}$$
(20)
$$\text{max} \left( {f\left( {X_{1}^{{nomad}}} \right),f\left( {X_{2}^{{nomad}}} \right)} \right)=f\left( {X_{2}^{{nomad}}} \right)$$
(21)

As \(\frac{{f\left( {X_{2}^{{nomad}}} \right)}}{{f\left( {X_{1}^{{nomad}}} \right)}}>1\) from Eq. (21)

$$E_{1}^{{nomad}}=\exp (1)\frac{{f\left( {X_{2}^{{nomad}}} \right)}}{{f\left( {X_{1}^{{nomad}}} \right)}}>\exp (1)$$
(22)
$$E_{2}^{{nomad}}=\exp \left( {\frac{{{d_2}}}{{{d_1}}}} \right)<\exp (1)$$
(23)

Hence, it is proved that \(E_{1}^{{nomad}}>E_{2}^{{nomad}}\), when \({d_1}>{d_2}\) and \(f\left( {X_{2}^{{nomad}}} \right)>f\left( {X_{1}^{{nomad}}} \right)\).

Lemma 5

\(E_{2}^{{nomad}}\) is greater than \(E_{1}^{{nomad}}\), if \({d_2}\) and \(f\left( {X_{1}^{{nomad}}} \right)\) are greater than \({d_1}\) and \(f\left( {X_{2}^{{nomad}}} \right)\), respectively.

Lemma 6

\(E_{2}^{{nomad}}\) is greater than \(E_{1}^{{nomad}}\), if \({d_2}\) and \(f\left( {X_{2}^{{nomad}}} \right)\) are greater than \({d_1}\) and \(f\left( {X_{1}^{{nomad}}} \right)\), respectively.

Proof of Lemma 5 and 6

Lemma 5 is the vice versa of lemma 3 as \(E_{2}^{{nomad}}>\exp \left( 1 \right)\) and \(E_{1}^{{nomad}}<\exp \left( 1 \right)\). Lemma 6 is the vice versa of lemma 4 as \(E_{2}^{{nomad}}=\exp \left( 1 \right)\) and probably \(E_{1}^{{nomad}}<\exp \left( 1 \right)\) proved through Axiom 3.

Axiom 1

\({C_r} \cdot L=L\) is an integer possibly between \(1\) and \(L - 1\), i.e.,\({C_r} \cdot L \in \left( {1,L - 1} \right)\).

Axiom 2

\(X_{{^{1}}}^{{nomad}}\) and \(X_{2}^{{nomad}}\) are essentially different and hence \({d_1}\) is not always equal to \({d_2}\).

Axiom 3

\(f\left( {X_{1}^{{nomad}}} \right)\) and \(f\left( {X_{2}^{{nomad}}} \right)\) exhibit linear variation with respect to \({d_1}\) and \({d_2}\), respectively.

Theorem 1

\({X^{cubs}}\) can be a subset of both \({X^{male}}\) and \({X^{female}}\) only if \({C_r}\) is selected in such a way that \({C_r} \cdot L=L\).

Proof of Theorem 1

It is known that \(B\) has \({C_r} \cdot L\) number of ones in arbitrary vector positions and zeros in the remaining positions. Hence, \({X^{male}} \circ B\) have elements of \({X^{male}}\) and zeros from the positions where \(B\) has ones and zeros, respectively. In contrast, \({X^{female}} \circ \bar {B}\) has elements of \({X^{female}}\) and zeros from the positions where \(B\) has zeros and ones, respectively, as \(\bar {B}\) is the one’s complement of \(B\). Hence, it can be said that \({X^{male}} \circ B \subset {X^{male}}\) and \({X^{female}} \circ \bar {B} \subset {X^{female}}\). As ‘+’ operator in the mathematical representation of crossover operation given in [11] is equivalent to set union operation, the resultant \({X^{cubs}}\) are subset of both and only \({X^{male}}\) and \({X^{female}}\), i.e., \({X^{cubs}} \subset {X^{male}},{X^{female}}\).

Theorem 2

In a nomad coalition of only two lions, the evaluation score \(E_{1}^{{nomad}}\) will be always greater than \(E_{2}^{{nomad}}\) when \(E_{1}^{{nomad}}\) is greater than or equal to exponential function of unity and vice versa.

Proof of Theorem 2

Let \(E_{1}^{{nomad}}\) and \(E_{2}^{{nomad}}\) be the evaluation scores of \(X_{{^{1}}}^{{nomad}}\) and \(X_{2}^{{nomad}}\), respectively. The evaluation scores can be calculated as

$$E_{1}^{{nomad}}=\exp \left( {\frac{{{d_1}}}{{\text{max} \left( {{d_1},{d_2}} \right)}}} \right)\frac{{\text{max} \left( {f\left( {X_{1}^{{nomad}}} \right),f\left( {X_{2}^{{nomad}}} \right)} \right)}}{{f\left( {X_{1}^{{nomad}}} \right)}}$$
(24)
$$E_{2}^{{nomad}}=\exp \left( {\frac{{{d_2}}}{{\text{max} \left( {{d_1},{d_2}} \right)}}} \right)\frac{{\text{max} \left( {f\left( {X_{1}^{{nomad}}} \right),f\left( {X_{2}^{{nomad}}} \right)} \right)}}{{f\left( {X_{2}^{{nomad}}} \right)}}$$
(25)

where \({d_1}\) is the Euclidean distance between \(X_{{^{1}}}^{{nomad}}\) and \({X^{male}}\), \({d_2}\) is the Euclidean distance between \(X_{2}^{{nomad}}\) and \({X^{male}}\).

From lemmas 3 and 4, it can be said that if \(E_{1}^{{nomad}}=\exp \left( 1 \right)\) (according to lemma 3) and \(E_{1}^{{nomad}}>\exp \left( 1 \right)\) (according to lemma 4), then \(E_{2}^{{nomad}}<\exp \left( 1 \right)\). Similarly, lemmas 5 and 6 asserts \(E_{2}^{{nomad}}>E_{1}^{{nomad}}\), if \(E_{2}^{{nomad}} \geq \exp \left( 1 \right)\). Hence, the theorem states that it is not necessary to calculate both \(E_{1}^{{nomad}}\) and \(E_{2}^{{nomad}}\) to evaluate \(X_{{^{1}}}^{{nomad}}\) and \(X_{2}^{{nomad}}\), respectively. It is sufficient to calculate either of them and can be concluded by comparing it with \(\exp \left( 1 \right)\).

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Boothalingam, R. Optimization using lion algorithm: a biological inspiration from lion’s social behavior. Evol. Intel. 11, 31–52 (2018). https://doi.org/10.1007/s12065-018-0168-y

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12065-018-0168-y

Keywords

Navigation