Skip to main content

Heuristic Hill-Climbing as a Markov Process

  • Conference paper
Artificial Intelligence: Methodology, Systems, and Applications (AIMSA 2008)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 5253))

Abstract

The purpose of this paper is twofold: on one hand, modelling the hill-climbing heuristic search algorithm as a stochastic process serves for deriving interesting properties about its expected performance; on the other hand, the probability that a hill-climbing search algorithm ever fails when approaching the target node (i.e., it does not find a descendant with a heuristic value strictly lower than the current one) can be considered as a pesimistic measure of the accuracy of the heuristic function guiding it. Thus, in this work, it is suggested to model heuristic hill-climbing search algorithms with Markov chains in order to fulfill these goals. Empirical results obtained in various sizes of the (n,m)-Puzzle domain prove that this model leads to very accurate predictions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Pearl, J.: Heuristics. Addison-Wesley, Reading (1984)

    Google Scholar 

  2. Ghallab, M., Nau, D., Traverso, P.: Automated Planning: Theory and Practice. Morgan Kaufmann, San Francisco (2004)

    MATH  Google Scholar 

  3. Bonet, B., Geffner, H.: Planning as heuristic search. Artificial Intelligence 129(1–2), 5–33 (2001)

    Article  MATH  MathSciNet  Google Scholar 

  4. Bonet, B., Geffner, H.: Planning as Heuristic Search: New Results. In: Biundo, S., Fox, M. (eds.) ECP 1999. LNCS, vol. 1809, pp. 359–371. Springer, Heidelberg (2000)

    Chapter  Google Scholar 

  5. Hoffmann, J., Nebel, B.: The FF planning system: Fast plan generation through heuristic search. Journal of Artificial Intelligence Research 14, 253–302 (2001)

    MATH  Google Scholar 

  6. Refanidis, I., Vlahavas, I.: The grt planning system: Backward heuristic construction in forward state-space planning. Journal of the Artificial Intelligence Research 15, 115–161 (2001)

    Article  MATH  Google Scholar 

  7. Korf, R.E.: Search. In: Encyclopedia of Artificial Intelligence, vol. 2, pp. 994–998 (1987)

    Google Scholar 

  8. Stewart, W.J.: Introduction to the Numerical Solution of Markov Chains. Princeton University Press, Princeton (1994)

    MATH  Google Scholar 

  9. Korf, R.E., Reid, M., Edelkamp, S.: Time complexity of iterative-deepening-A*. Artificial Intelligence 129(1–2), 199–218 (2001)

    Article  MATH  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Danail Dochev Marco Pistore Paolo Traverso

Rights and permissions

Reprints and permissions

Copyright information

© 2008 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

López, C.L. (2008). Heuristic Hill-Climbing as a Markov Process. In: Dochev, D., Pistore, M., Traverso, P. (eds) Artificial Intelligence: Methodology, Systems, and Applications. AIMSA 2008. Lecture Notes in Computer Science(), vol 5253. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-85776-1_23

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-85776-1_23

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-85775-4

  • Online ISBN: 978-3-540-85776-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics