Next: Metropolis-Hastings algorithm
Up: MCMC and EM Algorithms
Previous: Bayesian Inference
Monte Carlo methods can solve these general problems:
- draw random variables from a probability density function
- estimate integral
Monte Carlo methods can be used to evaluate the integrations for
in Baysian inference, by drawing samples
from
and then approximating
This is referred to as Monte Carlo integration.
A Markov Chain is a sequence of random variables
generated by a Markov process. At each time
, the next state
is sampled from a distribution called transition kernel
.
By definition, the next state
of a Markov process only depends
on the current state
, but is independent of all previous states
(limited horizon):
Let
be the
possible states of the
Markov chain. We define
Then the following Chapman-Kolomogrov equation gives the
probability of the next state taking a specific value
:
We represent both distributions at stage
and
column vectors:
then the Chapman-Kolomogrov equation can be expressed in matrix
form:
This state transition can be recursively extended until reaching the
initial state
:
If the Markov chain is irreducible and aperiodic, the distribution
may reach a stationary distribution
after a large number of transitions, independent of the initial
distribution
. Any further transition will not change
the distribution any more:
In other words, the stationary distribution
is the
eigenvector of the transition matrix
corresonding to the
eigenvalue
.
A sufficient condition for a unique stationary distribution is
that the detailed balance equation holds:
under this condition the chain is reversible, and the ith element of
vector
is
(last equal sign due to
), i.e.,
is stationary. This condition guarantees stationary
distribution from the beginning of the chain, and is not a necessary
condition as the chain can also become stationary gradually.
Next: Metropolis-Hastings algorithm
Up: MCMC and EM Algorithms
Previous: Bayesian Inference
Ruye Wang
2006-10-11