Transcript ppt-2
Gibbs sampler - simple properties
It’s not hard to show that this MC chain is
aperiodic.
Often is reversible distribution.
If in addition the chain is irreducible
(which depends on which elements of
have non-zero probability), then this MC
is a correct MCMC algorithm for
simulating random variables with
distribution .
Systematic Gibbs sampler
A commonly used variant of the Gibbs
sampler is the MC obtained by systematic
cycle trough the vertex set.
For instance, for
we may decide
to update vertex:
Back to the Boolean cube
In this case:
,
.
In each step we pick
and flip the
appropriate coordinate of
with
probability .
Stationary distribution for the Boolean Cube
We will show that the uniform distribution
is reversible (and therefore stationary).
Denote by
the transition probability from
state to .
We need to show that:
for any
two states.
Denote by
the number of indexes in
which and differ:
◦ Case no.1:
◦ Case no.2:
◦ Case no.3:
Reminder - Coupling
For a given MC
with state space
and transitions matrix , a coupling for the
MC is a MC
on the state space
with
transition probabilities defined by:
Reminder (cont.) - The coupling lemma
For a coupling
, based on a ground
MC
on . Suppose
is a
function satisfying the condition: for
all
and all
:
then the mixing time
for
is bounded by
.
Fast convergence of the Gibbs
sampler for the Boolean cube
Coupling: we run two Markov chains
and
simultaneously.
We go threw the coordinates
systematically.
If at some integer time it holds that:
We flip both of them with probability
Otherwise we flip each one separately with
same probability.
Fast convergence of the Gibbs
sampler for the Boolean cube
For any coordinate : after one cycle
threw the coordinates:
After cycles threw the chain:
The event
implies that for at
least one coordinate :
meaning that:
Fast convergence of the Gibbs
sampler for the Boolean cube
Setting
and solving for
gives:
Now, using the coupling lemma, it means
that the mixing time is bounded by:
Gibbs sampler for q-colorings
For a vertex and an assignment of colors
to the vertices other than , the conditional
distribution of the color at is uniform over
the set of all colors that are not attained in
at some neighbor of .
A Gibbs Sampler for random q-colorings is
therefore an -valued MC where at each
time
, transitions take place as follows:
◦ Pick a vertex uniformly at random.
◦ Pick
according to the uniform distribution
over the set of colors that are not attained at any
neighbor of .
◦ Leave the color unchanged at all other vertices,
i.e. let
for all vertices
except .
Gibbs sampler for q-colorings
It has as a stationary distribution, the
proof is similar to the hard-core model
case, based on the fact that it is
reversible.
Whether or not the chain is irreducible
depends on and , and it is a nontrivial
question to generally determine this. But
for big enough (
is always enough) it
is.
Fast convergence of Gibbs sampler
for q-coloring
Theorem [HAG 8.1]: Let
be a graph.
Let be the number of vertices in , and
suppose any vertex has at most
neighbors. Suppose furthermore that
Then, for any fixed , the number of
iterations needed for the systematic sweep
Gibbs sampler described above (starting
from any fixed -coloring ) to come within
total variation distance of the target
distribution is at most:
Fast convergence of Gibbs sampler
for q-coloring
Comments:
◦ Can be proved for
(rather than
◦ The important bound is:
some
.
).
for
Fast convergence of Gibbs sampler
for q-coloring
Coupling: we run two
-valued Markov
chains
and
simultaneously.
When a vertex is chosen to be updated, we
pick a new color for it uniformly at random
among the set of colors that are not attained
by any neighbor of . Concretely:
Let
be an i.i.d sequence of random
permutations of
, each of them
uniformly distributed on the set the of
such permutations.
Fast convergence of Gibbs sampler
for q-coloring (cont.)
At each time , the updates of the 2
chains use the permutation
and
the vertex to be updated is assigned the
new value
where:
in the first chain.
In the second chain, we set
where:
Successful and not-successful updates
If
at some (random, hopefully not too
large) time , we will also have
for all
First consider the probability that:
for a given vertex .
We call an update of the 2 chains at time
successful if it results in having:
Otherwise – we say that the update failed.
Fast convergence of Gibbs sampler
for q-coloring
Notations:
The event of having successful update has
probability:
Respectively:
Using that
and that
we
get:
(we only used
and
some algebra).
Fast convergence of Gibbs sampler
for q-coloring
Hence, we have, after steps of the MC-s,
that for each vertex :
Now, consider updates during the 2nd sweep
of the Gibbs sampler, i.e. between times
and .
For an update at time during the 2nd sweep
to fail the configurations and need to
differ in at least one neighbor of .
Each neighbor has
with
probability at most
By summing over at most neighbors we get
that:
Fast convergence of Gibbs sampler
for q-coloring
“discrepancy” means that there exists a
neighbor of with
.
By Repeating the arguments above we get:
Hence, after steps of the Markov chains,
each vertex has probability at most:
By arguing the same for the third sweep:
Fast convergence of Gibbs sampler
for q-coloring
and for any :
The event
implies that
for at least one vertex , meaning that:
Fast convergence of Gibbs sampler
for q-coloring
Summary:
By setting
and solving for ,
we get by using the coupling lemma that
for:
It holds that the mixing time after
is bounded by .
scans
Fast convergence of Gibbs sampler
for q-coloring
To go from the number of scans to the
number of steps of the MC, we have to
multiply by , giving that:
should be enough.
In order to make sure that
gets an
integer value we should take to be: