Transcript Q-Layout
Random Variables
A random variable X is a real valued function defined
on the sample space, X: S R. The set {s S : X(s) [a,
b] is an event}.
Example 1
Let S be the sample space of an experiment consisting of
tossing two fair dice. Then, S = {(1, 1}, (1, 2),..., (6, 6)}. Let X be
a random variable defined over S that assigns to each
outcome the sum of the dice. Then, X ((1,1)) = 2, X ((1,2)) = 3,
... X ((6,6)) = 12.
Usually, we specify the distribution of a random
variable without reference to the probability space.
If X denotes the random variable that is defined as the sum of two
fair dice, then
If X denotes the random variable that is defined as the sum of two
fair dice, then
P{X = 2} = P{(1, 1)} = 1/36
If X denotes the random variable that is defined as the sum of two
fair dice, then
P{X = 2} = P{(1, 1)} = 1/36
P{X = 3} = P{(1, 2), (2, 1)} = 2/36
If X denotes the random variable that is defined as the sum of two
fair dice, then
P{X = 2} = P{(1, 1)} = 1/36
P{X = 3} = P{(1, 2), (2, 1)} = 2/36
P{X = 4} = P{(1, 3), (3, 1), (2, 2)} = 3/36
If X denotes the random variable that is defined as the sum of two
fair dice, then
P{X = 2} = P{(1, 1)} = 1/36
P{X = 3} = P{(1, 2), (2, 1)} = 2/36
P{X = 4} = P{(1, 3), (3, 1), (2, 2)} = 3/36
P{X = 5} = P{(1, 4), (4, 1), (2, 3), (3, 2)} = 4/36
If X denotes the random variable that is defined as the sum of two
fair dice, then
P{X = 2} = P{(1, 1)} = 1/36
P{X = 3} = P{(1, 2), (2, 1)} = 2/36
P{X = 4} = P{(1, 3), (3, 1), (2, 2)} = 3/36
P{X = 5} = P{(1, 4), (4, 1), (2, 3), (3, 2)} = 4/36
P{X = 6} = P{(1, 5), (5, 1), (2, 4), (4, 2), (3, 3)} = 5/36
P{X = 7} = P{(1, 6), (6, 1), (3, 4), (4, 3), (5, 2), (2, 5)} = 6/36
P{X = 8} = P{(2, 6), (6, 2), (3, 5), (5, 3), (4, 4)} = 5/36
P{X = 9} = P{(3, 6), (6, 3), (5, 4), (4, 5)} = 4/36
P{X = 10} = P{(4, 6), (6, 4), (5, 5)} = 3/36
P{X = 11} = P{(5, 6), (6, 5)} = 2/36
P{X = 12} = P{(6, 6)} = 1/36
The random variable X takes on values X = n, where n = 2,
..., 12. Since the events corresponding to each value are
mutually exclusive, then:
P
12
n2
{ X n} n 2 P( X n) 1.
12
Example 2
Let S be the sample space of an experiment consisting of
tossing two fair coins. Then, S = {(H, H}, (H, T), (T, H), (T, T)}.
Let Y be a random variable defined over S that assigns to
each outcome the number of heads
Example 2
Let S be the sample space of an experiment consisting of
tossing two fair coins. Then, S = {(H, H}, (H, T), (T, H), (T, T)}.
Let Y be a random variable defined over S that assigns to
each outcome the number of heads. Then, Y is a random
variable that takes on values 0, 1, 2:
P(Y =0) = 1/4
P(Y =1) = 2/4
P(Y=2) = 1/4.
P(Y =0) + P(Y =1) + P(Y=2) = 1.
Example 3
A die is repeatedly tossed until a six appears. Let X denote
the number of tosses required, assuming successive tosses are
independent.
Example 3
A die is repeatedly tossed until a six appears. Let X denote
the number of tosses required, assuming successive tosses are
independent. The random variables X takes on values 1, 2, ...,
with respective probabilities:
Example 3
A die is repeatedly tossed until a six appears. Let X denote
the number of tosses required, assuming successive tosses are
independent. The random variables X takes on values 1, 2, ...,
with respective probabilities:
P(X=1) = 1/6
P(X=2) = (5/6)(1/6)
P(X=3) = (5/6)2(1/6)
P(X=4) = (5/6)3(1/6)
.
.
.
P(X=n) = (5/6)n-1(1/6)
P
{ X n} i 1 P ( X n)
n 1
1
1 n 1
= i 1 (1 )
6
6
1/6
=
1-(1-1/6)
=1
Distribution functions
The distribution function F (also called cumulative
distribution function (cdf)) of a random variable is
defined by F(x) = P(X ≤ x), where x is a real number.
Distribution functions
The distribution function F (also called cumulative
distribution function (cdf)) of a random variable is
defined by F(x) = P(X ≤ x), where x is a real number.
upper case
lower case
Properties of distribution functions
(i) F ( x) is a nondecreasing function of x.
.
Properties of distribution functions
(i) F ( x) is a nondecreasing function of x.
(ii) lim F ( x) F () =1.
x
Properties of distribution functions
(i ) F ( x) is a nondecreasing function of x.
(ii ) lim F ( x) F () =1.
x
(iii ) lim F ( x) F ( ) =0.
x
Properties of distribution functions
(i) F ( x) is a nondecreasing function of x.
(ii) lim F ( x) F () =1.
x
(iii) lim F ( x) F ( ) =0.
x
(iv) P( X x) 1 P( X x) 1 F ( x).
.
Properties of distribution functions
(i ) F ( x) is a nondecreasing function of x.
(ii ) lim F ( x) F () =1.
x
(iii ) lim F ( x ) F ( ) =0.
x
(iv ) P ( X x) 1 P ( X x) 1 F ( x).
(iv ) P ( a X b) P ( X b) P ( X b) F (b) F (a ).
Note that P(X < x) does not necessarily equal F(x) since F(x)
includes the probability that X equals x.
Discrete random variables
A discrete random variable is a random variable that
takes on either a finite or a countable number of
states.
Probability mass function
The probability mass function (pmf) of a discrete random
variable is defined by p(x) = P(X=x).
If x takes on values x1, x2, ..., then
i 1
p( xi ) 1.
Probability mass function
The probability mass function (pmf) of a discrete random
variable is defined by p(x) = P(X=x).
If x takes on values x1, x2, ..., then
i 1
p( xi ) 1.
The cumulative distribution function F is given by
F ( x) all x x p( xi ).
i
Example
Let X be a random with pmf p(2) = 0.25, p(4) = 0.6, and p(6) =
0.15. Then, the cdf F of X is given by
Example
Let X be a random with pmf p(2) = 0.25, p(4) = 0.6, and p(6) =
0.15. Then, the cdf F of X is given by
0
0.25
F ( x)
0.85
1
if
if
if
if
x2
2 x4
4 x6
6x
The Bernoulli random variable
Let X be a random variable that takes on values 1 (success) or
0 (failure), then the pmf of X is given by
p(0) = P(X=0) = 1 - p and
p(1) = P(X=1) = p
where 0 ≤ p ≤ 1 is the probability of “success.”
A random variable that has the above pmf is said to be a
Bernoulli random variable.
The Geometric random variable
A random variable that has the following pmf is said to be a
geometric random variable with parameter p.
p(n) = P(X=n) = (1 – p)n-1p,
for n = 1, 2, ... .
The Geometric random variable
A random variable that has the following pmf is said to be a
geometric random variable with parameter p.
p(n) = P(X=n) = (1 – p)n-1p,
for n = 1, 2, ... .
Example: A series of independent trials, each having a
probability p of being a success, are performed until a success
occurs. Let X be the number of trials required until the first
success.
The Binomial random variable
A random variable that has the following pmf is said to be a
geometric random variable with parameters (n, p)
n i
p(i) p (1 p) n i
i
n
n!
where n Ai
,
(n i )!i !
i
n is an integer 1 and 0 p 1.
Example: A series of n independent trials, each having a
probability p of being a success and 1 – p of being a failure,
are performed until a success occurs. Let X be the number of
successes in the n trials.
The Poisson random variable
A random variable that has the following pmf is said to be a
Poisson random variable with parameter
l (l 0)
i
p (i ) P{ X i} e
l
l
i!
, i 0,1,....
The number of cars sold per day by a dealer is Poisson with
parameter l = 2. What is the probability of selling no cars
today? What is the probability of receiving 100?
Solution:
P(X=0) = e-2 0.135
P(X = 2)= e-2(22 /2!) 0.270
Example: The number of cars sold per day by a dealer is
Poisson with parameter l = 2. What is the probability of
selling no cars today? What is the probability of receiving
100?
Solution:
P(X=0) = e-2 0.135
P(X = 2)= e-2(22 /2!) 0.270
Continuous random variables
A continuous random variable is a random variable
whose set of possible values is uncountable.
In particular, we say that