Abstract
This paper deals with discrete time infinite horizon stochastic decision processes with various reward criteria. Sufficient conditions are obtained for the value of a class of strategies to be equal to the value of the subclass of non-randomized strategies from this class.
Similar content being viewed by others
References
R.J. Aumann, Mixed and behavior strategies in infinite extensive games, Ann. Math. Studies 53 (1964) 627–650.
D.P. Bertsekas and S.E. Shreve,Stochastic Optimal Control: The Discrete Time Case (Academic Press, New York, 1978).
D. Blackwell, D. Freedman and M. Orkin, The optimal reward operator in dynamic programming, Ann. Statist. 1 (1974) 926–941.
E.B. Dynkin and A.A. Yushkevich,Controlled Markov Processes (Springer, New York, 1979).
E.A. Feinberg, Non-randomized Markov and semi-Markov strategies in dynamic programming, Theory Probab. Appl. 27 (1982) 116–126.
E.A. Feinberg, Controlled Markov processes with arbitrary numerical criteria, Theory Probab. Appl. 27 (1982) 486–503.
E.A. Feinberg, Sufficient classes of strategies in discrete dynamic programming I: Decomposition of randomized strategies and embedded models, Theory Probab. Appl. 31 (1986) 658–668.
E.A. Feinberg, On stationary strategies in Borel dynamic programming, submitted to Math. Oper. Res. (1989).
E.A. Feinberg and I.M. Sonin, Persistently nearly optimal strategies in stochastic dynamic programming, in:Statistics and Control of Stochastic Processes (Steklov Seminar, 1984), Optimization Software, New York (1985) pp. 69–101.
I.I. Gikhman and A.V. Skorokhod,Controlled Random Processes (Springer, New York, 1979).
K.M. van Hee, Markov strategies in dynamic programming, Math. Oper. Res. 3 (1978) 37–41.
T.P. Hill, On the existence of good Markov strategies, Trans. Amer. Math. Soc. 247 (1979) 157–176.
T.P. Hill and V.C. Pestien, The existence of good Markov strategies for decision processes with general payoffs, Stochastic Process. Appl. 24 (1987) 61–76.
G. Kallianpur,Stochastic Filtering Theory (Springer, New York, 1980).
N.V. Krylov, The construction of an optimal strategy for a finite controlled chain, Theory Probab. Appl. 10 (1965) 45–54.
P.A. Meyer,Probability and Potentials (Blaisdell, Waltham, MA, 1966).
J. Neveu,Mathematical Foundations of the Calculus of Probability (Holden-Day, San Francisco, 1965).
M. Schäl, Stationary policies in dynamic programming models under compactness assumptions, Math. Oper. Res. 8 (1983) 366–372.
A.N. Shiryaev,Optimal Stopping Rules (Springer, New York, 1978).
I.M. Sonin and E.A. Feinberg, Sufficient classes of strategies in controllable countable Markov chains with total criterion, Sov. Math. Dokl. 29 (1984) 308–311.
R. Strauch, Negative dynamic programming, Ann. Math. Statist. 37 (1966) 871–890.
J. van der Wal,Stochastic Dynamic Programming (Mathematisch Centrum, Amsterdam, 1981).
A.A. Yushkevich and R.J. Chitashvili, Controlled random sequences, Russian Math. Surveys 37 (1982) 239–274.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Feinberg, E.A. Non-randomized strategies in stochastic decision processes. Ann Oper Res 29, 315–332 (1991). https://doi.org/10.1007/BF02283603
Issue Date:
DOI: https://doi.org/10.1007/BF02283603