Math Camp 2: Probability Theory

Download Report

Transcript Math Camp 2: Probability Theory

Math Camp 2: Probability
Theory
Sasha Rakhlin
Introduction









-algebra
Measure
Lebesgue measure
Probability measure
Expectation and variance
Convergence
Convergence in probability and almost surely
Law of Large Numbers. Central Limit Theorem
Useful Probability Inequalities





Jensen’s inequality
Markov’s inequality
Chebyshev’s inequality
Cauchy-Schwarz inequality
Hoeffeding’s inequality
-algebra

Let  be a set. Then a -algebra  is a
nonempty collection of subsets of 
such that the following hold:




If E  ,  - E  
If Fi   i, i Fi  
Measure

A measure  is a function defined on a
-algebra  over a set  with values in
[0, ] s.t.



() = 0
(E) = i (Ei) if E = i Ei
(, , ) is called a measure space
Lebesgue measure

The Lebesgue measure  is the unique
complete translation-invariant measure
on a -algebra s.t. ([0,1]) = 1
Probability measure



Probability measure is a positive
measure  over (, ) s.t. () = 1
(, , ) is called a probability space
A random variable is a measurable function
X: R
Expectation and variance

If X is a random variable over a
probability space (, , ), the
expectation of X is defined as
E ( X )   Xd


The variance of X is
var( X )  E (( X  E ( X )) 2 )
Convergence


xn  x if   > 0  N s.t. |xn – x| <  for n
>N
P
X n 
X (Xn converges to X in
probability) if P( X n  X   )  0   > 0
Convergence in probability
and almost surely

Any event with probability 1 is said to
happen almost surely. A sequence of
real random variables Xn converges
almost surely to a random variable X iff
P ( lim X n  X )  1
n 

Convergence almost surely implies
convergence in probability
Law of Large Numbers.
Central Limit Theorem

Weak LLN: if X1, X2, … is an infinite
sequence of i.i.d. random variables with
P
 = E(X1) = E(X2) =…, X n 
 , that
is, lim P( X     )  0
Xn  
CLT: lim
P(
 z )  ( z ) where  is
n 
n

n
/ n
the cdf of N(0,1)
Jensen’s inequality

If  is a convex function, then
 ( E ( X ))  E ( ( X ))
Markov’s inequality

E( X )
If X  0 and t  0, Pr( X  t ) 
t
Chebyshev’s inequality


If X is random variable and t > 0,
var( X )
Pr(| X  E ( X ) | t ) 
t2
e.g.

Pr(| X  E ( X ) | 2 ) 
1
4
Cauchy-Schwarz inequality

If E(X2) and E(Y2) are finite,
E ( XY )  E ( X 2 ) E (Y 2 )
Hoeffding’s inequality

Let ai  Xi  bi for i = 1, …, n. Let Sn = 
Xi, then for any t > 0,
2 t 2
Pr( Sn  E ( Sn )  t )  2 i 1
n
( bi ai ) 2