Abstract
In Part II we treated two basic models of Stochastic Dynamic Programming: Control Models with independent disturbances and Markovian Decision Processes. In the first of these models the state ζ t = ζ t π in period t and the disturbance η t+1 in the same period are stochastically independent of each other. However, there are other important models in which η t+1 depends on ζ t . This holds in particular in some of the Markov renewal programs treated in Chap. 22 below, where a random time η t+1 elapses between the t-th and the (t + 1)-st decision; another relevant model is the Markovian control model MCM introduced below. We now present a model, called an MDP with disturbances, which due to a very flexible transition law comprises all these models. As a further important generalization, necessary for the Markov renewal programs, we allow the discount factor in each period to be a function of the disturbance.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bertsekas, D. P., & Shreve, S. E. (1978). Stochastic optimal control. New York: Academic Press.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this chapter
Cite this chapter
Hinderer, K., Rieder, U., Stieglitz, M. (2016). Markovian Decision Processes with Disturbances. In: Dynamic Optimization. Universitext. Springer, Cham. https://doi.org/10.1007/978-3-319-48814-1_21
Download citation
DOI: https://doi.org/10.1007/978-3-319-48814-1_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-48813-4
Online ISBN: 978-3-319-48814-1
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)