Probability Theory and Random Processes
Download
Report
Transcript Probability Theory and Random Processes
Probability Theory and Random Processes
Communication Systems, 5ed., S. Haykin and M.
Moher, John Wiley & Sons, Inc., 2006.
Probability
• Probability theory is based on the phenomena that
can be modeled by an experiment with an outcome
that is subject to chance.
• Definition: A random experiment is repeated n time
(n trials) and the event A is observed m times (m
occurrences). The probability is the relative
frequency of occurrence m/n.
Probability Based on Set Theory
• Definition: An experiment has K possible outcomes
where each outcome is represented as the kth sample sk.
The set of all outcomes forms the sample space S. The
probability measure P satisfies the
• Axioms:
– 0 ≤ P[A] ≤ 1
– P[S] = 1
– If A and B are two mutually exclusive events (the two events
cannot occur in the same experiment), P[AUB]=P [A] + P[B],
otherwise P[AUB] = P[A] + P[B] – P[A∩B]
– The complement is P[Ā] = 1 – P[A]
– If A1, A2,…, Am are mutually exclusive events, then P[A1] + P[A2]
+ … + P[Am] = 1
Venn Diagrams
sk
Sample can only come
from A, B, or neither.
S
A
B
Events A and B that are mutually
exclusive events in the sample space S.
Sample can only come
from both A and B.
sk
S
A
Events A and B are not mutually
exclusive events in the sample space S.
B
Conditional Probability
• Definition: An experiment involves a pair of events A
and B where the probability of one is conditioned on
the occurrence of the other. Example: P[A|B] is the
probability of event A given the occurrence of event
B
• In terms of the sets and subsets
– P[A|B] = P[A∩B] / P[A]
– P[A∩B] = P[A|B]P[B] = P[B|A]P[A]
• Definition: If events A and B are independent, then
the conditional probability is simply the elementary
probability, e.g. P[A|B] = P[A], P[B|A] = P[B].
Random Variables
• Definition: A random variable is the assignment of a
variable to represent a random experiment. X(s)
denotes a numerical value for the event s.
• When the sample space is a number line, x = s.
• Definition: The cumulative distribution function (cdf)
assigns a probability value for the occurrence of x
within a specified range such that FX(x) = P[X ≤ x].
• Properties:
– 0 ≤ FX(x) ≤ 1
– FX(x1) ≤ FX(x2), if x1 ≤ x2
Random Variables
• Definition: The probability density function (pdf) is
an alternative description of the probability of the
random variable X: fX(x) = d/dx FX(x)
• P[x1 ≤ X ≤ x2] = P[X ≤ x2] - P[X ≤ x1]
= FX(x2) - FX(x1)
= fX(x)dx over the interval [x1,x2]
Example Distributions
• Uniform distribution
xa
0,
1
f X ( x)
, a xb
b
a
xb
0,
xa
0,
x a
FX ( x)
, a xb
b
a
xb
1,
Several Random Variables
• CDF:
FX ,Y ( x, y ) PX x, Y y
• Marginal cdf:
• PDF:
y
x
FX ( x)
f
X ,Y
• Conditional pdf:
f
X ,Y
(u , v) du dv
f
X ,Y
(u , v) du dv 1
f X ( x)
f
2
f X ,Y ( x, y)
FX ,Y ( x, y)
xy
• Marginal pdf:
FY ( y )
(u, v) du dv
X ,Y
( x, v) dv
f Y ( y | x)
fY ( y )
f
f X ,Y ( x, y)
f X ( x)
X ,Y
(u, y ) du
Statistical Averages
• Expected value:
X EX xf X x dx
• Function of a random variable:
EY
Y g( X )
yf y dy Eg X g X f x dx
Y
• Text Example 5.4
X
Statistical Averages
• nth moments:
x
EX
n
n
f X x dx
2
f X x dx
x
EX
2
Mean-square value of X
• Central moments:
x
E X X
n
X
n f X x dx
x
E X X
2
2
2
f
x
dx
X
X
X
Variance of X
Joint Moments
• Correlation:
i
E X ,Y
k
x y
i
k
f X ,Y x, y dx dy
Expected value of the product
- Also seen as a weighted inner product
• Covariance:
covXY E X EX Y EY
EXY X Y
• Correlation coefficient:
Correlation of the central moment
uncorrelat ed
covXY 0,
X Y 1, strongly correlated
Random Processes
• Definition: a random process is described as a timevarying random variable X t
X t EX t xf X t x dx
• Mean of the random process:
• Definition: a random process is first-order stationary if its
pdf is constant
f X t x f X t x X t X X2 t X2 Constant mean, variance
1
2
• Definition: the autocorrelation is the expected value of
the product of two random variables at different times
RX t1 , t2 EX t1 X t2
R t , t R t t Stationary to
X
1
2
X
1
2
second order
Random Processes
• Definition: the autocorrelation is the expected value
of the product of two random variables at different
times
RX t1 , t2 EX t1 X t2
RX t1 , t2 RX t1 t2
Stationary to
second order
• Definition: the autocovariance of a stationary
random process is
C X t1 , t 2 E X t1 X X t2 X
RX t1 t2 X2
Properties of Autocorrelation
• Definition: autocorrelation of a stationary process
only depends on the time differences
RX EX t X t
• Mean-square value:
RX 0 E X 2 t
• Autocorrelation is an even function:
RX RX
• Autocorrelation has maximum at zero:
RX RX 0
Example
• Sinusoidal signal with random phase
X t A cos2f c t ,
1
, -
f t 2
otherwise
0,
• Autocorrelation
RX EX t X
A2
cos2f c
2
As X(t) is compared to
itself at another time, we
see there is a periodic
behavior it in correlation
Cross-correlation
• Two random processes have the cross-correlation
X t , Y t
RXY t , u EX t Y u
• Wide-sense stationary cross-correlation
RX t , RX t , RY u, RY u
RXY , u RXY
Example
• Output of an LTI system when the input is a RP
• Text 5.7
Power Spectral Density
• Definition: Fourier transform of autocorrelation
function is called power spectral density
SX f
j 2f
R
e
dτ
X
RX τ
j 2f
S
f
e
df
X
• Consider the units of X(t) Volts or Amperes
• Autocorrelation is the projection of X(t) onto itself
• Resulting units of Watts (normalized to 1 Ohm)
Properties of PSD
• Zero-frequency of PSD
S X 0
R dτ
X
• Mean-square value
S
E X 2 t
Which theorem does this property resemble?
• PSD is non-negative
• PSD of a real-valued RP
SX f 0
S X f S X f
X
f df
Example
• Text Example 5.12
– Mixing of a random process with a sinusoidal process
Y t X t cos2f ct
Wide-sense stationary RP
(to make it easier)
Uniformly distributed, but
not time-varying
– Autocorrelation
RY EY t Y t
– PSD
SY f
1
RX cos2f c
2
1
SY f f c SY f f c
4
PSD of LTI System
• Start with what you know and work the math
Y t ht * X t
SY f
R e
Y
j 2f
dτ
SY f EY t Y t e j 2f d
Eht * X t ht * X t e j 2f d
j 2f
d
E h 1 X t 1 d 1 h 2 X t 2 d 2 e
j 2f
d 1 d 2 d
e
t
X
t
X
E
h
h
2
1
2
1
PSD of LTI System
• The PSD reduces to
SY f
j 2f
h
h
R
e
d 1 d 2 d
1
2
1 2 X
Change of variables 1 2 0 0 1 2
SY f
j 2f 0 1 2
h
h
R
e
d 1 d 2 d 0
0
2
0
1
X
0
SY f H f S X f
2
System shapes power spectrum of input as
expected from a filtering like operation
Gaussian Process
• The Gaussian probability density function for a single
variable is
y Y 2
1
fY y
exp
2
2 Y
2 Y
• When the distribution has zero mean and unit variance
y2
1
fY y
exp
2
2
• The random variable Y is said to be normally distributed
as N(0,1)
Properties of a Gaussian Process
• The output of a LTI is Gaussian if the input is Gaussian
• The joint pdf is completely determined by the set of
means and autocovariance functions of the samples of
the Gaussian process
• If a Gaussian process is wide-sense stationary, then the
output of the LTI system is strictly stationary
• A Gaussian process that has uncorrelated samples is
statistically independent
Noise
• Shot noise
• Thermal noise
• White noise
• Narrow