Skip to main content

Multi-Agent Evolutionary Model for Global Numerical Optimization

  • Chapter
Agent-Based Evolutionary Search

Part of the book series: Adaptation, Learning, and Optimization ((ALO,volume 5))

Abstract

Agent-based computation has been studied for several years in the field of distributed artificial intelligence and has been widely used in other branches of computer science. In this chapter, first, multi-agent systems and genetic algorithms are integrated to form a new algorithm, namely Multi-Agent Genetic Algorithm (MAGA), for solving the global numerical optimization problem. An agent in MAGA represents a candidate solution to the optimization problem in hand. All agents live in a latticelike environment, with each agent fixed on a lattice-point. In order to increase energies, they compete or cooperate with their neighbors, and they can also use knowledge. Making use of these agent-agent interactions, MAGA realizes the purpose of minimizing the objective function value. Theoretical analyses show that MAGA converges to the global optimum. Second, Macro-Agent Evolutionary Model (MacroAEM) is proposed with the intrinsic properties of decomposable functions in mind. In this model, a sub-function forms a macro-agent, and 3 new behaviors, namely competition, cooperation, and selfishness, are developed for macro-agents to optimizing objective functions. Finally, the MacroAEM model is integrated with MAGA, which results a new algorithm, namely Hierarchical Multi-Agent Genetic Algorithm (HMAGA), especially for optimizing decomposable functions. The convergence of HMAGA is also analyzed theoretically and the results show that HMAGA also converges to the global optima. To validate the performance of MAGA, MacroAEM, and HMAGA, benchmark functions are used. The scalability of MAGA along the problem dimension is studied with great care. The results show that MAGA achieves a good performance when the dimensions are increased from 20 to 10,000. Moreover, even when the dimensions are increased to as high as 10,000, MAGA still can find high quality solutions at a low computational cost. Therefore, MAGA has a good scalability and is a competent algorithm for solving high dimensional optimization problems. The experimental results about HMAGA show that HMAGA achieves a good performance, too, especially for high-dimensional functions. Apart from this, the analyses on time complexity demonstrate that HMAGA has a good scalability.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Russell, S., Norvig, P.: Artificial intelligence: a modern approach. Prentice- Hall, New York (1995)

    MATH  Google Scholar 

  2. Jennings, N.R., Sycara, K., Wooldridge, M.: A roadmap of agent research and development. In: Jennings, N.R., Sycara, K., Georgeff, M. (eds.) Autonomous Agents and Multi-Agent Systems Journal, vol. 1(1), pp. 7–38. Kluwer Academic Publishers, Boston (1998)

    Google Scholar 

  3. Ferber, J.: Multi-Agent Systems: An Introduction to Distributed Artificial Intelligence. Addison-Wesley, New York (1999)

    Google Scholar 

  4. Liu, J.: Autonomous Agents and Multi-Agent Systems: Explorations in Learning, Self-Organization, and Adaptive Computation. World Scientific, Singapore (2001)

    Book  Google Scholar 

  5. Liu, J., Tang, Y.Y., Cao, Y.C.: An evolutionary autonomous agents approach to image feature extraction. IEEE Trans. Evol. Comput. 1(2), 141–158 (1997)

    Article  Google Scholar 

  6. Liu, J., Jing, H., Tang, Y.Y.: Multi-agent oriented constraint satisfaction. Artificial Intelligence 136(1), 101–144 (2002)

    Article  MathSciNet  MATH  Google Scholar 

  7. Whitley, D.: Cellular genetic algorithms. In: Belew, R.K., Booker, L.B. (eds.) Proc. Fifth Int. Conference on Genetic Algorithms, p. 658. Morgan Kaufmann, San Mateo (1993)

    Google Scholar 

  8. Nakashima, T., Ariyama, T., Yoshida, T., Ishibuchi, H.: Performance evaluation of combined cellular genetic algorithms for function optimization problems. In: Proc. 2003 IEEE Int. Symposium on Computational Intelligence in Robotics and Automation, Kobe, Japan, vol. 1, pp. 295–299 (2003)

    Google Scholar 

  9. Folino, G., Pizzuti, C., Spezzano, G.: Parallel hybrid method for SAT that couples genetic algorithms and local search. IEEE Trans. Evol. Comput. 5, 323–334 (2001)

    Article  Google Scholar 

  10. Holland, J.H.: Adaptation in nature and artificial system. MIT Press, Cambridge (1992)

    Google Scholar 

  11. Goldberg, E.: Genetic Algorithms in Search, Optimization & Machine Learning. Addison-Wesley, Reading (1989)

    MATH  Google Scholar 

  12. Mitchell, M.: An Introduction to Genetic Algorithms. MIT Press, Cambridge (1996)

    Google Scholar 

  13. Pan, Z.J., Kang, L.S.: An adaptive evolutionary algorithms for numerical optimization. In: Yao, X., Kim, J.H., Furuhashi, T. (eds.) SEAL 1996. LNCS, vol. 1285, pp. 27–34. Springer, Heidelberg (1997)

    Chapter  Google Scholar 

  14. Leung, Y.W., Wang, Y.: An orthogonal genetic algorithm with quantization for global numerical optimization. IEEE Trans. Evol. Comput. 5(1), 41–53 (2001)

    Article  Google Scholar 

  15. Kazarlis, S.A., Papadakis, S.E., Theocharis, J.B., Petridis, V.: Microgenetic algorithms as generalized hill-climbing operators for GA optimization. IEEE Trans. Evol. Comput. 5, 204–217 (2001)

    Article  Google Scholar 

  16. Rudolph, G.: Convergence analysis of canonical genetic algorithms. IEEE Trans. Neural Networks, Special Issue on Evolutional Computing 5(1), 96–101 (1994)

    Article  Google Scholar 

  17. Dinabandhu, B., Murthy, C.A., Sankar, K.P.: Genetic algorithm with elitist model and its convergence. International Journal of Pattern Recognition and Artificial Intelligence 10(6), 731–747 (1996)

    Article  Google Scholar 

  18. Iosifescu, M.: Finite Markov Processes and Their Applications. Wiley, Chichester (1980)

    MATH  Google Scholar 

  19. Yao, X., Liu, Y., Lin, G.: Evolutionary programming made faster. IEEE Trans. Evol. Comput. 3, 82–102 (1999)

    Article  Google Scholar 

  20. Mühlenbein, H., Schlierkamp-Vose, D.: Predictive models for the breeder genetic algorithm. Evolutionary Computation 1(1), 25–49 (1993)

    Article  Google Scholar 

  21. Liang, Y., Zhou, C., Li, S.: Optimization of Rosenbrock’s function based on genetic algorithms. Journal of Software 8(9), 701–708 (1997)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Liu, J., Zhong, W., Jiao, L. (2010). Multi-Agent Evolutionary Model for Global Numerical Optimization. In: Sarker, R.A., Ray, T. (eds) Agent-Based Evolutionary Search. Adaptation, Learning, and Optimization, vol 5. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-13425-8_2

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-13425-8_2

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-13424-1

  • Online ISBN: 978-3-642-13425-8

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics