Abstract
The precise definition of the term “Markov chain” as used in this monograph will be given below. However, the following remarks will help clarify our usage for the benefit of those readers who have had previous contact with the terminology. A Markov process is a special type of stochastic process distinguished by a certain Markov property; a Markov chain is a Markov process with a denumerable (namely, finite or denumerably infinite) number of states. The time parameter may be taken to be the set of nonnegative integers or the set of nonnegative real numbers; accordingly we have the discrete parameter case or the continuous parameter case. The adjective “simple” is sometimes used to qualify our Markov chain, but since we do not discuss “multiple” chains we shall not make the distinction. Furthermore, we shall discuss only Markov chains “with stationary (or temporally homogeneous) transition probabilities” so that the qualifying phrase in quotes will be understood. Finally, our discussion does not differentiate between a finite or a denumerably infinite number of states so that no special treatment is given to the former case.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1960 Springer-Verlag OHG. Berlin · Göttingen · Heidelberg
About this chapter
Cite this chapter
Chung, K.L. (1960). Fundamental defintions. In: Markov Chains with Stationary Transition Probabilities. Die Grundlehren der Mathematischen Wissenschaften, vol 104. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-49686-8_1
Download citation
DOI: https://doi.org/10.1007/978-3-642-49686-8_1
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-49408-6
Online ISBN: 978-3-642-49686-8
eBook Packages: Springer Book Archive