Skip to main content
Log in

Semi-infinite discounted Markov decision processes: Policy improvement and singular perturbations

  • Published:
Mathematical Methods of Operations Research Aims and scope Submit manuscript

Abstract.

In this paper, Discounted Markov Decision Processes with finite state and countable action set (semi-infinite DMDP for short) are considered. A policy improvement finite algorithm which finds a nearly optimal deterministic strategy is presented. The steps of the algorithm are based on the classical policy improvement algorithm for finite DMDPs. Singularly perturbed semi-infinite DMDPs are investigated. In case of perturbations, some sufficient condition is given to guarantee that there exists a nearly optimal deterministic strategy which can approximate nearly optimal strategies for a whole family of singularly perturbed semi-infinite DMDP.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Author information

Authors and Affiliations

Authors

Additional information

Manuscript received: December 2000/Final version received: March 2001

Rights and permissions

Reprints and permissions

About this article

Cite this article

Abbad, M., Rahhali, K. Semi-infinite discounted Markov decision processes: Policy improvement and singular perturbations. Mathematical Methods of OR 54, 279–290 (2001). https://doi.org/10.1007/s001860100143

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1007/s001860100143

Navigation