Abstract
Markov chains with a countably infinite state space exhibit some types of behavior not possible for chains with a finite state space. Figure 5.1 helps explain how these new types of behavior arise. If p > 1/2, then transitions to the right occur with higher frequency than transitions to the left. Thus, reasoning heuristically, we expect Xn to be large for large n. This means that, given X0 = 0, the probability P n0j should go to zero for any fixed j with increasing n. If one tried to define the steady state probability of state j as limn→∞P n0j ,then this limit would be 0 for all j. These probabilities would not sum to 1, and thus would not correspond to a limiting distribution. Thus we say that a steady state does not exist. In more heuristic terms, the state keeps increasing forever. The truncation of figure 5.1 to k states is analyzed in exercise 4.3. The solution there defines ρ=p/q and shows that π1=(1−ρ)ρ1/(1−ρk) for ρ≠1 and π1 = 1/k for ρ=1. For ρ<1 the limiting behavior as k → ∞ is π1= (1−ρ)ρi for ρ<1 and πi=0 otherwise. In section 5.3 we analyze birth death Markov chains, of which figure 5.1 is an example, without first truncating the chain.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1996 Springer Science+Business Media New York
About this chapter
Cite this chapter
Gallager, R.G. (1996). Markov Chains with Countably Infinite State Spaces. In: Discrete Stochastic Processes. The Springer International Series in Engineering and Computer Science, vol 321. Springer, Boston, MA. https://doi.org/10.1007/978-1-4615-2329-1_5
Download citation
DOI: https://doi.org/10.1007/978-1-4615-2329-1_5
Publisher Name: Springer, Boston, MA
Print ISBN: 978-1-4613-5986-9
Online ISBN: 978-1-4615-2329-1
eBook Packages: Springer Book Archive