presentation source
Download
Report
Transcript presentation source
1
CE 530 Molecular Simulation
Lecture 8
Markov Processes
David A. Kofke
Department of Chemical Engineering
SUNY Buffalo
[email protected]
Monte Carlo Integration: Review
Stochastic evaluation of integrals
• sum integrand evaluated at randomly generated points
• most appropriate for high-dimensional integrals
error vanishes more quickly (1/n1/2)
better suited for complex-shaped domains of integration
Monte Carlo simulation
• Monte Carlo integration for ensemble averages
U
Importance Sampling
•
•
•
•
1
N!
dr
N
N e U ( r
U (r ) Z
N
N)
p(rN)
emphasizes sampling in domain where integrand is largest
it is easy to generate points according to a simple distribution
stat mech p distributions are too complex for direct sampling
need an approach to generate random multidimensional points
according to a complex probability distribution
f
• then integral is given by I p p
2
3
Markov Processes
Stochastic process
• movement through a series of well-defined states in a way that
involves some element of randomness
• for our purposes,“states” are microstates in the governing
ensemble
Markov process
• stochastic process that has no memory
• selection of next state depends only on current state, and not on
prior states
• process is fully defined by a set of transition probabilities pij
pij = probability of selecting state j next, given that presently in state i.
Transition-probability matrix P collects all pij
4
Transition-Probability Matrix
Example
If in state 1, will stay in state 1
with probability 0.1
• system with three states
p 11 p 12 p 13 0.1 0.5 0.4
P p 21 p 22 p 23 0.9 0.1 0.0
p
31 p 32 p 33 0.3 0.3 0.4
If in state 1, will move to state 3
with probability 0.4
Never go to state 3 from state 2
Requirements of transition-probability matrix
• all probabilities non-negative, and no greater than unity
• sum of each row is unity
• probability of staying in present state may be non-zero
5
Distribution of State Occupancies
Consider process of repeatedly moving from one state to the
next, choosing each subsequent state according to P
• 1 2 2 1 3 2 2 3 3 1 2 3 etc.
Histogram the occupancy number for each state
• n1 = 3
• n2 = 5
• n3 = 4
p1 = 0.33
p2 = 0.42
p3 = 0.25
1 2 3
After very many steps, a limiting distribution emerges
Click here for an applet that demonstrates a Markov process
and its approach to a limiting distribution
6
The Limiting Distribution 1.
Consider the product of P with itself
p 11 p 12 p 13 p 11 p 12 p 13
All ways of going from state
P 2 p 21 p 22 p 23 p 21 p 22 p 23
1 to state 2 in two steps
p
p
p
p
p
p
31
32
33 31
32
33
p 11p 11 p 12p 21 p 13p 31 p 11p 12 p 12p 22 p 13p 32 etc.
p 21p 11 p 22p 21 p 23p 31 p 21p 12 p 22p 22 p 23p 32 etc.
p p p p p p
31 11
32 21
33 31 p 31p 12 p 32p 22 p 33p 32 etc.
Probability of going from
state 3 to state 2 in two steps
In general Pn is the n-step transition probability matrix
• probabilities of going from state i to j in exactly n steps
( n)
( n)
( n)
p 11
p 12
p 13
( n)
( n)
( n)
n
P p 21 p 22 p 23 defines p ( n )
ij
( n)
( n)
( n)
p 31 p 32
p
33
7
The Limiting Distribution 2.
Define p i(0) as a unit state vector
p1(0) 1 0 0 p 2(0) 0 1 0 p 3(0) 0 0 1
Then p i(n) p i(0)P n is a vector of probabilities for ending at
each state after n steps if beginning at state i
p 1( n)
( n)
( n)
( n)
p 11
p12
p13
(0) n
( n)
( n)
( n)
( n)
( n)
( n)
p 1 P 1 0 0 p 21 p 22 p 23 p 11
p 12
p 13
( n)
( n)
( n)
p 31 p 32
p
33
The limiting distribution corresponds to n
()
()
()
• independent of initial state p1 p 2 p 3 p
8
The Limiting Distribution 3.
Stationary property of p
p lim p i(0)P n
n
lim p i(0)P n 1 P
n
pP
p is a left eigenvector of P with unit eigenvalue
• such an eigenvector is guaranteed to exist for matrices with rows
that each sum to unity
Equation for elements of limiting distribution p
p i p jp ji
j
0.1 0.5 0.4
e.g. P 0.9 0.1 0.0
0.3 0.3 0.4
not independent
p1 0.1p1 0.9p 2 0.3p 3
p 2 0.5p1 .1p 2 0.3p 3
p 3 0.4p1 0.0p 2 0.4p 3
p1 p 2 p 3 p1 p 2 p 3
9
Detailed Balance
Eigenvector equation for limiting distribution
• p i p jp ji
j
A sufficient (but not necessary) condition for solution is
• p ip ij p jp ji
• “detailed balance” or “microscopic reversibility”
Thus
• p i p jp ji
j
p ip ij
j
p i p ij p i
j
0.1
P 0.9
0.3
0.5
0.1
0.3
0.4
0.0
0.4
For a given P, it is not always
possible to satisfy detailed
balance; e.g. for this P
p 3p 32 p 2p 23
zero
10
Deriving Transition Probabilities
Turn problem around...
…given a desired p, what transition probabilities will yield
this as a limiting distribution?
Construct transition probabilities to satisfy detailed balance
Many choices are possible
• e.g. p 0.25 0.5 0.25
• try them out
0 1 0
P 0.5 0 0.5
0 1 0
Most efficient
0.97 0.02 0.01
P 0.01 0.98 0.01
0.01 0.02 0.97
Least efficient
0.0 0.5 0.5
0.42 0.33 0.25
P 0.17 0.66 0.17 P 0.25 0.5 0.25
0.5 0.5 0.0
0.25 0.33 0.42
Barker
Metropolis
11
Metropolis Algorithm 1.
Prescribes transition probabilities to satisfy detailed balance,
given desired limiting distribution
Recipe:
From a state i…
• with probability tij, choose a trial state j for the move (note: tij = tji)
• if pj > pi, accept j as the new state
• otherwise, accept state j with probability pj/pi
generate a random number R on (0,1); accept if R < pj/pi
• if not accepting j as the new state, take the present state as the next
one in the Markov chain p ii 0
Metropolis, Rosenbluth, Rosenbluth, Teller and Teller,
J. Chem. Phys., 21 1087 (1953)
12
Metropolis Algorithm 2.
What are the transition probabilities for this algorithm?
• Without loss of generality, define i as the state of greater probability
pj
p ij t ij
pi
p ji t ji
p ii 1 p ij
p j
in
general:
p
t
min
ij
ij
p ,1
i
pi p j
j i
Do they obey detailed balance?
?
p ip ij p jp ji
pj ?
p it ij p jt ji
pi
t ij t ji
Yes, as long as the underlying matrix T of the Markov chain is
symmetric
• this can be violated, but acceptance probabilities must be modified
13
Markov Chains and Importance Sampling 1.
Importance sampling specifies the desired limiting distribution
We can use a Markov chain to generate quadrature points
according to this distribution
1 inside R
Example
s
0
0.5
r2
0.5
dx
0.5
dy ( x 2 y 2 ) s ( x, y )
0.5
0.5
0.5
0.5
dx
0.5
outside R
dys ( x, y )
r 2s
s
V
V
q = normalization constant
V
• Method 1: let p1( x, y) s( x, y) / q1
• then r 2 s
2
2
r
2
p1 p
1
s
p1 p
1
q1r
q1
p1
p1
q1 r
q1
p1
r
2
p1
Simply sum r2 with points
given by Metropolis sampling
14
Markov Chains and Importance Sampling 2.
Example (cont’d)
p ( x, y ) r 2 s / q2
• Method 2: let
r2s
• then
r
2
p2 p
2
s
p2 p
2
q2
p2
q2 / r 2
p2
q2
q2 1/ r 2
p2
1
r 2
p2
Algorithm and transition probabilities
• given a point in the region R
• generate a new point in the vicinity of
given point
xnew = x + r(-1,+1)dx ynew = y + r(-1,+1)dy
• accept with probability min(1,p new / p old )
p 1new s new / q1 s new
• note
p 1old
s old / q1
s old
Normalization constants cancel!
• Method 1: accept all moves that stay in R
2 new r 2 old
r
• Method 2: if in R, accept with probability
15
Markov Chains and Importance Sampling 3.
Subtle but important point
• Underlying matrix T is set by the trial-move
algorithm (select new point uniformly in
vicinity of present point)
• It is important that new points are selected
in a volume that is independent of the Different-sized
trial sampling
present position
regions
• If we reject configurations outside R,
without taking the original point as the
“new” one, then the underlying matrix
becomes asymmetric
16
Evaluating Areas with Metropolis Sampling
What if we want the absolute area
of the region R, not an average over
it?
0.5
0.5
A
0.5
dx
0.5
dys ( x, y ) s
V
• Let p1( x, y) s( x, y) / q1
• then A s
q1 p q1
p
1
p1
1
• We need to know the normalization
constant q1
• but this is exactly the integral that we
are trying to solve!
Absolute integrals difficult by MC
• relates to free-energy evaluation
17
Summary
Markov process is a stochastic process with no memory
Full specification of process is given by a matrix of
transition probabilities P
A distribution of states are generated by repeatedly stepping
from one state to another according to P
A desired limiting distribution can be used to construct
transition probabilities using detailed balance
• Many different P matrices can be constructed to satisfy detailed
balance
• Metropolis algorithm is one such choice, widely used in MC
simulation
Markov Monte Carlo is good for evaluating averages, but not
absolute integrals
Next up: Monte Carlo simulation