Difference between revisions of "Markov chain, recurrent"
(→References: Feller: internal link) |
(TeX) |
||
(One intermediate revision by one other user not shown) | |||
Line 1: | Line 1: | ||
{{MSC|60J10}} | {{MSC|60J10}} | ||
+ | {{TEX|done}} | ||
[[Category:Markov chains]] | [[Category:Markov chains]] | ||
− | A [[Markov chain|Markov chain]] in which a random trajectory | + | A [[Markov chain|Markov chain]] in which a random trajectory $\xi(t)$, starting at any state $\xi(0)=i$, returns to that state with probability 1. In terms of the [[Transition probabilities|transition probabilities]] $p_{ij}(t)$, recurrence of a discrete-time Markov chain is equivalent to the divergence for any $i$ of the series |
− | + | $$\sum_{t=0}^\infty p_{ij}(t).$$ | |
− | In a recurrent Markov chain a trajectory | + | In a recurrent Markov chain a trajectory $\xi(t)$, $0\leq t<\infty$, $\xi(0)=i$, returns infinitely often to the state $i$ with probability 1. In a recurrent Markov chain there are no inessential states and the essential states decompose into recurrent classes. An example of a recurrent Markov chain is the symmetric random walk on the integer lattice on the line or plane. In the symmetric walk on the line a particle moves from position $x$ to $x\pm1$ with probabilities $1/2$; in the symmetric walk on the plane a particle moves from $(x,y)$ to one of the four points $(x\pm1,y)$, $(x,y\pm1)$ with probabilities $1/4$. In these examples a particle, starting the walk at an arbitrary point, returns to that point with probability 1. The symmetric walk on the integer lattice in the three-dimensional space, when the probability of transition from $(x,y,z)$ to a neighbouring point $(x\pm1,y,z)$, $(x,y\pm1,z)$, $(x,y,z\pm1)$ is equal to $1/6$, is not recurrent. In this case the probability of return of the particle to its initial point is approximately 0.35. |
====References==== | ====References==== | ||
− | + | {| | |
+ | |valign="top"|{{Ref|F}}|| W. Feller, [[Feller, "An introduction to probability theory and its applications"|"An introduction to probability theory and its applications"]], '''1''' , Wiley (1966) | ||
+ | |} | ||
====Comments==== | ====Comments==== | ||
Line 16: | Line 19: | ||
====References==== | ====References==== | ||
− | + | {| | |
+ | |valign="top"|{{Ref|Fr}}|| D. Freeman, "Markov chains" , Holden-Day (1975) | ||
+ | |- | ||
+ | |valign="top"|{{Ref|I}}|| M. Iosifescu, "Finite Markov processes and their applications" , Wiley (1980) {{MR|0587116}} {{ZBL|0436.60001}} | ||
+ | |- | ||
+ | |valign="top"|{{Ref|KS}}|| J.G. Kemeny, J.L. Snell, "Finite Markov chains" , v. Nostrand (1960) {{MR|1531032}} {{MR|0115196}} {{ZBL|0089.13704}} | ||
+ | |- | ||
+ | |valign="top"|{{Ref|KSK}}|| J.G. Kemeny, J.L. Snell, A.W. Knapp, "Denumerable Markov chains" , Springer (1976) {{MR|0407981}} {{ZBL|0348.60090}} | ||
+ | |- | ||
+ | |valign="top"|{{Ref|Re}}|| D. Revuz, "Markov chains" , North-Holland (1975) {{MR|0415773}} {{ZBL|0332.60045}} | ||
+ | |- | ||
+ | |valign="top"|{{Ref|Ro}}|| V.I. Romanovsky, "Discrete Markov chains" , Wolters-Noordhoff (1970) (Translated from Russian) {{MR|0266312}} {{ZBL|0201.20002}} | ||
+ | |- | ||
+ | |valign="top"|{{Ref|Se}}|| E. Seneta, "Non-negative matrices and Markov chains" , Springer (1981) {{MR|2209438}} {{ZBL|0471.60001}} | ||
+ | |- | ||
+ | |valign="top"|{{Ref|Sp}}|| V. Spitzer, "Principles of random walk" , v. Nostrand (1964) {{MR|0171290}} {{ZBL|0119.34304}} | ||
+ | |} |
Latest revision as of 16:12, 30 July 2014
2020 Mathematics Subject Classification: Primary: 60J10 [MSN][ZBL]
A Markov chain in which a random trajectory $\xi(t)$, starting at any state $\xi(0)=i$, returns to that state with probability 1. In terms of the transition probabilities $p_{ij}(t)$, recurrence of a discrete-time Markov chain is equivalent to the divergence for any $i$ of the series
$$\sum_{t=0}^\infty p_{ij}(t).$$
In a recurrent Markov chain a trajectory $\xi(t)$, $0\leq t<\infty$, $\xi(0)=i$, returns infinitely often to the state $i$ with probability 1. In a recurrent Markov chain there are no inessential states and the essential states decompose into recurrent classes. An example of a recurrent Markov chain is the symmetric random walk on the integer lattice on the line or plane. In the symmetric walk on the line a particle moves from position $x$ to $x\pm1$ with probabilities $1/2$; in the symmetric walk on the plane a particle moves from $(x,y)$ to one of the four points $(x\pm1,y)$, $(x,y\pm1)$ with probabilities $1/4$. In these examples a particle, starting the walk at an arbitrary point, returns to that point with probability 1. The symmetric walk on the integer lattice in the three-dimensional space, when the probability of transition from $(x,y,z)$ to a neighbouring point $(x\pm1,y,z)$, $(x,y\pm1,z)$, $(x,y,z\pm1)$ is equal to $1/6$, is not recurrent. In this case the probability of return of the particle to its initial point is approximately 0.35.
References
[F] | W. Feller, "An introduction to probability theory and its applications", 1 , Wiley (1966) |
Comments
References
[Fr] | D. Freeman, "Markov chains" , Holden-Day (1975) |
[I] | M. Iosifescu, "Finite Markov processes and their applications" , Wiley (1980) MR0587116 Zbl 0436.60001 |
[KS] | J.G. Kemeny, J.L. Snell, "Finite Markov chains" , v. Nostrand (1960) MR1531032 MR0115196 Zbl 0089.13704 |
[KSK] | J.G. Kemeny, J.L. Snell, A.W. Knapp, "Denumerable Markov chains" , Springer (1976) MR0407981 Zbl 0348.60090 |
[Re] | D. Revuz, "Markov chains" , North-Holland (1975) MR0415773 Zbl 0332.60045 |
[Ro] | V.I. Romanovsky, "Discrete Markov chains" , Wolters-Noordhoff (1970) (Translated from Russian) MR0266312 Zbl 0201.20002 |
[Se] | E. Seneta, "Non-negative matrices and Markov chains" , Springer (1981) MR2209438 Zbl 0471.60001 |
[Sp] | V. Spitzer, "Principles of random walk" , v. Nostrand (1964) MR0171290 Zbl 0119.34304 |
Markov chain, recurrent. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Markov_chain,_recurrent&oldid=25534