Parameter, Statistic and Random Samples
Download
Report
Transcript Parameter, Statistic and Random Samples
Parameter, Statistic and Random Samples
• A parameter is a number that describes the population. It is a fixed
number, but in practice we do not know its value.
• A statistic is a function of the sample data, i.e., it is a quantity
whose value can be calculated from the sample data. It is a random
variable with a distribution function. Statistics are used to make
inference about unknown population parameters.
• The random variables X1, X2,…, Xn are said to form a (simple)
random sample of size n if the Xi’s are independent random
variables and each Xi has the sample probability distribution. We say
that the Xi’s are iid.
week1
1
Example – Sample Mean and Variance
• Suppose X1, X2,…, Xn is a random sample of size n from a population
with mean μ and variance σ2.
•
The sample mean is defined as
1 n
X Xi.
n i 1
• The sample variance is defined as
1 n
2
S
X
X
.
i
n 1 i 1
2
week1
2
Goals of Statistics
• Estimate unknown parameters μ and σ2.
• Measure errors of these estimates.
• Test whether sample gives evidence that parameters are (or are
not) equal to a certain value.
week1
3
Sampling Distribution of a Statistic
• The sampling distribution of a statistic is the distribution of values
taken by the statistic in all possible samples of the same size from
the same population.
• The distribution function of a statistic is NOT the same as the
distribution of the original population that generated the original
sample.
• The form of the theoretical sampling distribution of a statistic will
depend upon the distribution of the observable random variables in
the sample.
week1
4
Sampling from Normal population
• Often we assume the random sample X1, X2,…Xn is from a normal
population with unknown mean μ and variance σ2.
• Suppose we are interested in estimating μ and testing whether it is
equal to a certain value. For this we need to know the probability
distribution of the estimator of μ.
week1
5
Claim
• Suppose X1, X2,…Xn are i.i.d normal random variables with
unknown mean μ and variance σ2 then
2
X ~ N ,
n
• Proof:
week1
6
Recall - The Chi Square distribution
• If Z ~ N(0,1) then, X = Z2 has a Chi-Square distribution with
parameter 1, i.e., X ~ 21 .
• Can proof this using change of variable theorem for univariate
random variables.
• The moment generating function of X is
1/ 2
1
mX t
1
2
t
• If X1 ~ 2v1 , X 2 ~ 2v2 , , X k ~ 2vk , all independent then
k
T X i ~ 2k v
i 1
1 i
• Proof…
week1
7
Claim
• Suppose X1, X2,…Xn are i.i.d normal random variables with mean μ
and variance σ2. Then, Z i X i are independent standard normal
variables, where i = 1, 2, …, n and
Xi
2
Z
~
n
i
i 1
i 1
n
2
n
• Proof: …
week1
8
t distribution
• Suppose Z ~ N(0,1) independent of X ~ χ2(n). Then, T
Z
X /v
~ t v .
• Proof:
week1
9
Claim
• Suppose X1, X2,…Xn are i.i.d normal random variables with mean μ
and variance σ2. Then,
X
~ tn1
S/ n
• Proof:
week1
10
F distribution
• Suppose X ~ χ2(n) independent of Y ~ χ2(m). Then,
X /n
~ Fn ,m
Y /m
week1
11
Properties of the F distribution
• The F-distribution is a right skewed distribution.
• Fm,n
1
Fn,m
i.e. PFn,m
1
1
1
a P
P Fm,n
F
a
n ,m a
• Can use Table 7 on page 796 to find percentile of the F- distribution.
• Example…
week1
12
The Central Limit Theorem
• Let X1, X2,…be a sequence of i.i.d
random variables with E(Xi) = μ < ∞
n
and Var(Xi) = σ2 < ∞. Let S n X i
i 1
S n n
lim
P
z
PZ z z for - ∞ < x < ∞
Then, n
n
where Z is a standard normal random variable and Ф(z)is the cdf for the
standard normal distribution.
•
This is equivalent to saying that Z n
Z ~ N(0,1).
•
S n n
converges in distribution to
n
Xn
x
lim
P
x
Also, n
n
i.e. Z n
Xn
converges in distribution to Z ~ N(0,1).
n
week1
13
Example
• Suppose X1, X2,…are i.i.d random variables and each has the Poisson(3)
distribution. So E(Xi) = V(Xi) = 3.
• The CLT says that P X1 X n 3n x 3n x as n ∞.
week1
14
Examples
• A very common application of the CLT is the Normal approximation to the
Binomial distribution.
• Suppose X1, X2,…are i.i.d random variables and each has the Bernoulli(p)
distribution. So E(Xi) = p and V(Xi) = p(1- p).
• The CLT says that P X 1 X n np x np1 p x as n ∞.
• Let Yn = X1 + … + Xn then Yn has a Binomial(n, p) distribution.
So for large n, PYn y P Yn np
np1 p
y np
y np
np1 p
np1 p
• Suppose we flip a biased coin 1000 times and the probability of heads on
any one toss is 0.6. Find the probability of getting at least 550 heads.
• Suppose we toss a coin 100 times and observed 60 heads. Is the coin fair?
week1
15