Probability and Stochastic Processes
Download
Report
Transcript Probability and Stochastic Processes
Probability and Stochastic
Processes
References:
Wolff, Stochastic Modeling and the Theory of
Queues, Chapter 1
Altiok, Performance Analysis of Manufacturing
Systems, Chapter 2
Chapter 0
1
Random Variables
•
•
•
•
•
•
•
•
•
Discrete vs. Continuous
Cumulative distribution function
Density function
Probability distribution (mass) function
Joint distributions
Conditional distributions
Functions of random variables
Moments of random variables
Transforms and generating functions
Chapter 0
2
Functions of Random Variables
• Often we’re interested in some combination of r.v.’s
– Sum of the first k interarrival times = time of the kth arrival
– Minimum of service times for parallel servers = time until next
departure
• If X = min(Y, Z) then X x if and only if Y x and Z x
– therefore, Pr X x 1 Pr Y x, Z x
– and if Y and Z are independent, Pr X x 1 Pr Y x Pr Z x
• If X = max(Y, Z) then Pr X x Pr Y x, Z x
• If X = Y + Z , its distribution is the convolution of the
distributions of Y and Z. Find it by conditioning.
Chapter 0
3
Conditioning (Wolff)
• Frequently, the conditional distribution of Y given X is
easier to find than the distribution of Y alone. If so,
evaluate probabilities about Y using the conditional
distribution along with the marginal distribution of X:
Pr Y A
Pr Y A X x f x dx
X
– Example: Draw 2 balls simultaneously from urn containing four
balls numbered 1, 2, 3 and 4. X = number on the first ball, Y =
number on the second ball, Z = XY. What is Pr(Z > 5)?
– Key: Maybe easier to evaluate Z if X is known
4
Pr Z 5 Pr Z 5 X x Pr X x
x 1
Chapter 0
4
Convolution
• Let X = Y+Z.
Pr X x Z z Pr Y Z x Z z Pr Y x z Z z
FX x Pr Y x z Z z f Z z dz
x z
fY Z y Z z f Z z dydz
• If Y and Z are independent,
FX x
x z
fY y f Z z dydz
– Example: Poisson
Chapter 0
5
Moments of Random Variables
• Expectation = “average” E X xf X x dx or x Pr X x
E g X g x f X x dx or g x Pr X x
2
2
2
Var
X
E
X
E
X
E
X
E
X
“volatility”
• Variance =
• Standard Deviation Var X
• Coefficient of Variation
CvX Var X E X
(s.c.v.) CvX2 Var X E 2 X
Chapter 0
6
Linear Functions of Random Variables
• Covariance Cov X , Y E X E X Y E Y E XY E X E Y
• Correlation Cov X , Y
XY
Var X Var Y
If X and Y are independent then Cov X , Y XY 0
E X Y E X E Y
E aX aE X
Var aX a 2 Var X
Var X Y Var X Var Y 2Cov X , Y
Chapter 0
7
Transforms and Generating Functions
• Moment-generating function
M * E e X e x f X x dx
E X k 1
• Laplace-Stieltjes transform
k
d k E e X
d k
s 0
E e sX e sx f X x dx, s 0
E X k 1
• Generating function (z – transform)
k
d k E e sX
ds k
s 0
Let N be a nonnegative integer random variable; Pn Pr N n , n 0,1, 2,...
G z n 0 Pn z n ,
z 1.
dG z
d 2G z
2
EN
, EN EN
dz z 1
dz 2 z 1
Chapter 0
8
Special Distributions
• Discrete
–
–
–
–
Bernoulli
Binomial
Geometric
Poisson
• Continuous
–
–
–
–
Uniform
Exponential
Gamma
Normal
Chapter 0
9
Bernoulli Distribution
“Single coin flip” p = Pr(success)
n 1
p,
N = 1 if success, 0 otherwise Pr N n
1 p, n 0
EN p
Var N p 1 p
CvN2
1 p
p
M * 1 p pe
Chapter 0
10
Binomial Distribution
“n independent coin flips” p = Pr(success)
N = # of successes
n k
nk
Pr N k p 1 p , k 0,1,..., n
k
E N np
Var N np 1 p
1 p
Cv
np
2
N
M 1 p pe
*
n
Chapter 0
11
Geometric Distribution
“independent coin flips” p = Pr(success)
N = # of flips until (including) first success
Pr N k 1 p
k 1
p, k 1, 2,...
EN 1 p
Var N 1 p p 2
CvN2 1 p
Memoryless property: Have flipped k times without success;
Pr N k n N k 1 p
n 1
Chapter 0
p (still geometric)
12
z-Transform for Geometric Distribution
(1-p)n-1p,
Given Pn =
G z n 1 1 p
n 1
n = 1, 2, …., find G z n1 Pn z n
pz n pz n 1 1 p z
pz
, using geometric series
1 1 p z
Then,
n 1
n
a
n 0
dG z
p
EN
dz z 1 1 p pz 2
pz n 0 1 p z
n
1
for a 1
1 a
z 1
1
p
2
d
G z
2 1 p
2 p
2
EN2 EN
,
so
E
N
and
2
2
2
dz
p
p
z 1
Var N E N
2
E N
2
Chapter 0
1 p
2
p
13
Poisson Distribution
“Occurrence of rare events” = average rate of occurrence
per period;
N = # of events in an arbitrary period
k e
Pr N k
, k 0,1, 2,...
k!
EN
Var N
CvN2 1
Chapter 0
14
Uniform Distribution
X is equally likely to fall anywhere within interval (a,b)
1
fX x
, a xb
ba
EX
Var X
ab
2
b a
2
12
b a
2
3b a
2
Cv X2
a
Chapter 0
b
15
Exponential Distribution
X is nonnegative and it is most likely to fall near 0
f X x e x , x 0
FX x 1 e x , x 0
EX
Var X
1
1
2
Cv X2 1
Also memoryless; more on this later…
Chapter 0
16
Gamma Distribution
X is nonnegative, by varying parameter b get a variety of shapes
b xb 1e x
b 1 x
fX x
b
EX
Var X
, x 0, where b x e dx for b 0
0
b
b
2
1
Cv
b
2
X
When b is an integer, k, this is called the Erlang-k distribution,
and k k 1! Erlang-1 is same as exponential.
Chapter 0
17
Normal Distribution
X follows a “bell-shaped” density function
2
1
x
fX x
e
2
2 2
, x
EX
Var X 2
From the central limit theorem, the distribution of the sum of
independent and identically distributed random variables
approaches a normal distribution as the number of summed
random variables goes to infinity.
Chapter 0
18
m.g.f.’s of Exponential and Erlang
If X is exponential and Y is Erlang-k,
M X*
and M Y*
k
Fact: The mgf of a sum of independent r.v.’s equals the
product of the individual mgf’s.
Therefore, the sum of k independent exponential r.v.’s (with
the same rate ) follows an Erlang-k distribution.
Chapter 0
19
Stochastic Processes
•
•
•
•
•
Poissson process
Markov chains
Regenerative processes
Residual life
Applications
– Machine repair model
– M/M/1 queue
– Inventory
Chapter 0
20
Stochastic Processes
Set of random variables, or observations of the same random
variable over time: X t , t 0 (continuous-parameter) or
X n , n 0,1,... (discrete-parameter)
Xt may be either discrete-valued or continuous-valued.
A counting process is a discrete-valued, continuous-parameter
stochastic process that increases by one each time some
event occurs. The value of the process at time t is the
number of events that have occurred up to (and including)
time t.
Chapter 0
21
Poisson Process
Let X t , t 0 be a stochastic process where X(t) is the
number of events (arrivals) up to time t. Assume X(0)=0 and
(i) Pr(arrival occurs between t and t+t) = t o t ,
where o(t) is some quantity such that limt 0 o t / t 0
(ii) Pr(more than one arrival between t and t+t) = o(t)
(iii) If t < u < v < w, then X(w) – X(v) is independent of X(u) –
X(t).
Let pn(t) = P(n arrivals occur during the interval (0,t). Then …
n
t
e t
pn t
,n 0
n!
Chapter 0
22
Poisson Process and Exponential Dist’n
Let T be the time between arrivals. Pr(T > t) = Pr(there are no
t
arrivals in (0,t) = p0(t) = e
Therefore, FT t Pr T t 1 e t , t 0, and
fT t e t , t 0
that is, the time between arrivals follows an exponential
distribution with parameter = the arrival rate.
The converse is also true; if interarrival times are exponential,
then the number of arrivals up to time t follows a Poisson
distribution with mean and variance equal to t.
Chapter 0
23
When are Poisson arrivals reasonable?
1.
The Poisson distribution can be seen as a limit of the
binomial distribution, as n , p0 with constant =np.
-
-
2.
3.
many potential customers deciding independently about arriving
(arrival = “success”),
each has small probability of arriving in any particular time interval
Conditions given above: probability of arrival in a small
interval is approximately proportional to the length of the
interval – no bulk arrivals
Amount of time since last arrival gives no indication of
amount of time until the next arrival (exponential –
memoryless)
Chapter 0
24
More Exponential Distribution Facts
1.
Suppose T1 and T2 are independent with T1 exp 1 , T2 exp 2
Then Pr T1 T2
1
1 2
Suppose (T1, T2, …, Tn ) are independent with Ti exp i
Let Y = min(T1, T2, …, Tn ) . Then Y exp 1 2 ... n
3. Suppose (T1, T2, …, Tk ) are independent with Ti exp i
2.
Let W= T1 + T2 + … + Tk . Then W has an Erlang-k distribution with
k 1
w
density function
fW w
e w , w 0 with
k 1!
E W
Var W
k
and
k
2
Chapter 0
25
Continuous Time Markov Chains
A stochastic process X t , t 0 with possible values
(state space) S = {0, 1, 2, …} is a CTMC if
Pr X u t j X s , s u Pr X u t j X u
“The future is independent of the past given the present”
Define p t Pr X u t j X u i (note: indep. of u )
ij
Then
0 pij t 1,
p t 1
ij
j
Chapter 0
26
CTMC Another Way
1.
2.
Each time X(t) enters state j, the sojourn time is
exponentially distributed with mean 1/qj
When the process leaves state i, it goes to state j i with
probability pij, where pii 0, 0 pij 1, pij 1
Let P t pij t , where P 0 I
Then
j
j t Pr X t j pij t i 0
i
Chapter 0
27
CTMC Infinitesimal Generator
The time it takes the process to go from state i to state j
Tij exp qij
Then qij is the rate of transition from state i to state j, qi qij
j
The infinitesimal generator is
q0
q
Q 10
q20
q01
q1
q21
q02
q12
q2
q0
q p
1 10
q2 p20
Chapter 0
qo p01 q0 p02
q1 q1 p12
q2 p21 q2
28
Long Run (Steady State) Probabilities
Let limt pij t j
•
Under certain conditions these limiting probabilities can be
shown to exist and are independent of the starting state;
•
They represent the long run proportions of time that the
process spends in each state,
•
Also the steady-state probabilities that the process will be
found in each state.
Then Q 0 with
1
i i
or, equivalently, q j j qi pij j for all j 0,1, 2,...
i j
rate out of j = rate into j
Chapter 0
29
Phase-Type Distributions
• Erlang distribution
• Hyperexponential distribution
• Coxian (mixture of generalized Erlang) distributions
Chapter 0
30