Professional Documents
Culture Documents
AARON PLAVNICK
Abstract. This paper provides some background for and proves the Funda-
mental Theorem of Markov Chains. It provides some basic denitions and
notation for recursion, periodicity, and stationary distributions.
Contents
1. Denitions and Background 1
2. Fundamental Theorem of Markov Chains 3
References 4
1. Definitions and Background
So what is a Markov Chain, lets dene it.
Denition 1.1. Let {X
0
, X
1
, . . .} be a sequence of random variables and Z =
0, 1, 2, . . . be the union of the sets of their realizations.Then {X
0
, X
1
, . . .} is
called a discrete-time Markov Chain with state space Z if:
P(X
n+1
= i
n+1
|X
n
= i
n
, . . . , X
1
= i
1
) = P(X
n+1
= i
n+1
|X
n
= i
n
)
Now lets set up some notation for the one-step transition probabilities of the
Markov Chain Let:
p
ij
(n) = P(X
n+1
= j|X
n
= i); n = 0, 1, . . .
We will limit ourselves to homogeneous Markov Chains. Or Markov Chains that
do not evolve in time.
Denition 1.2. We say that a Markov Chain is homogeneous if its one-step tran-
sition probabilities do not depend on n ie.
n, m N and i, j Z p
ij
(n) = p
ij
(m)
We then dene the n-step transition probabilities of a homogeneous Markov
Chain by
p
(m)
ij
= P(X
n+m
= j|X
n
= i)
Where by convention we dene
p
(0)
ij
=
(
1 if i = j
0 if i 6= j
Date: August 22, 2008.
1
2 AARON PLAVNICK
With these denitions we are now ready for a quick theorem: the Chapman-
Kolmogorov equation. This serves to relate the transition probabilities of a Markov
Chain.
Theorem 1.3.
p
(m)
ij
=
X
kZ
p
(r)
ik
p
(mr)
kj
; r N {0}
Proof. Making use of the total probability rule as well as the Markovian property
in this proof yields:
By total probability rule:
p
(m)
ij
= P(X
m
= j|X
0
= i) =
X
kZ
P(X
m
= j, X
r
= k|X
0
= i)
=
X
kZ
P(X
m
= j|X
r
= k, X
0
= i)P(X
r
= k|X
0
= i)
and now by the Markovian property:
=
X
kZ
P(X
m
= j|X
r
= k)P(X
r
= k|X
0
= i) =
X
kZ
p
(r)
ik
p
(mr)
kj
i
= 1
Such a distribution is said to be stationary if it satises
j
=
X
iZ
i
p
ij
^^independence
matrix
multipli
cation
THE FUNDAMENTAL THEOREM OF MARKOV CHAINS 3
Denition 1.6. We say that a subset space state C Z is closed if
X
jC
p
ij
= 1 i C
If Z itself has no proper closed subsets then the Markov Chain is said to be
irreducible.
Denition 1.7. We dene the period of a state i by
d
i
= gcd(m Z|p
(m)
ii
> 0)
State i is aperiodic if d
i
= 1
Denition 1.8. A state i is said to be accessible from from a state j if there is an
m 1 s.t
p
(m)
ij
> 0
If i is accessible from j and j is accessible from i then i and j are said to commu-
nicate.
Denition 1.9. We dene the rst-passage time probabilities by
f
(m)
ij
= P(X
m
= j; X
k
6= j, 0 < k < m1|X
0
= i) ; i, j Z
And we will denote the expected value, the expected return time, of this distri-
bution by
ij
=
X
m=1
mf
(m)
ij
Denition 1.10. We say that a state i is recurrent if
X
m=1
f
(m)
ij
= 1
and transient if
X
m=1
f
(m
ij
< 1
A recurrent state i is positive-recurrent if
ii
< and null-recurrent if
ii
=
2. Fundamental Theorem of Markov Chains
Theorem 2.1. For any irreducible, aperiodic, positive-recurrent Markov Chain
there exist a unique stationary distribution {
j
, j Z} s.t i Z
Proof. Because our chain is irreducible, aperiodic, and positive-recurrent we know
that for all i Z
j
= lim
n
p
(n)
ij
> 0. Likewise as
j
is a probability distribution
X
jZ
j
. We know that for any m
m
X
i=0
p
(m)
ij
X
i=0
p
(m)
ij
1
4 AARON PLAVNICK
taking the limit
lim
m
m
X
i=0
p
(m)
ij
=
X
i=0
j
1
which implies that for any M
M
X
i=0
j
1
Now we can use the Chapman-Kolmogorov equation analogously
p
(m+1)
ij
=
X
i=0
p
(m)
ik
p
kj
M
X
i=0
p
(m)
ik
p
kj
as before we take the limit as m, M yielding
X
i=0
k
p
kj
Now we search for contradiction and assume strict inequality holds for at least one
state j. Summing over these inequalities yields
X
j=0
j
>
X
j=0
X
k=0
k
p
kj
=
X
k=0
X
j=0
p
kj
=
X
k=0
k
But this is a contradiction, hence equality must hold
j
=
X
k=0
k
p
kj
Thus a unique stationary distribution exists.
Futher it can be shown that this unique distribution is related to the expected
return value of the Markov chain by
j
=
1
jj
References
[1] F. E. Beichelt. L. P. Fatti. Stochastic Processes And Their Applications. Taylor and Francis.
2002.