Estimation/Confidence Intervals for Popn Mean

Download Report

Transcript Estimation/Confidence Intervals for Popn Mean

One Sample Means Test: What if  is
unknown? (sampling from normal population)
We know that:
What is the
distribution of:
y  0
~ N(0,1)

n
y  0
t
s
n
If sample came from
a normal distribution,
t has a
t-distribution with
n-1 degrees of
freedom.
1) Symmetric about 0.
2) Looks like a standard normal density, only more spread out.
3) The spread of the distribution is indexed to a parameter called the
degrees of freedom (df).
4) As the degrees of freedom increase, the t-distribution gets closer to the
standard normal distribution. (Safe to use z instead of t when n>30.)
One Sample Inf-1
See Table 3 Ott & Longnecker
Tail probabilities of the
t-distribution
Df
.1
.05
.025
.01
.005
.001
1
3.078
6.314
12.706
31.821
63.657
318.309
5
1.476
2.015
2.571
3.365
4.032
5.893
10
1.372
1.812
2.228
2.764
3.169
4.144
15
1.341
1.753
2.131
2.602
2.947
3.733
20
1.325
1.725
2.086
2.528
2.845
3.552
25
1.316
1.708
2.060
2.485
2.787
3.450
30
1.310
1.697
2.042
2.457
2.750
3.385
40
1.303
1.684
2.021
2.423
2.704
3.307
NORMAL
(0,1)
1.282
1.645
1.960
2.326
2.576
3.090
One Sample Inf-2
Rejection Regions for hypothesis tests using
t-distribution critical values
H0:
HA:
 = 0
1.  > 0
For Pr(Type I error) = , df = n - 1
y  0
T .S.: t 
s n
Reject H0 if t > t,n-1
2.  < 0
t < -t,n-1
3.   0
| t | > t/2,n-1
One Sample Inf-3
Degrees of Freedom
Why are the degrees of freedom only n - 1 and not n?
We start with n independent pieces of information with which we
n
estimate the sample mean. y  1 y
n
Now consider the sample variance:

i 1
i
s
n
1
n 1
2
(y

y)
 i
i 1
Because the sum of the deviations yi  y are equal to zero, if we
know n-1 of these deviations, we can figure out the nth deviation.
Hence there are only n-1 independent deviations that are available to
estimate the variance (and standard deviation). That is, there are only
n-1 pieces of information available to estimate the standard deviation
after we “spend” one to estimate the sample mean. The t-distribution is
a normal distribution adjusted for unknown standard deviation hence it
is logical that it would have to accommodate the fact that only n-1
pieces of information are available.
One Sample Inf-4
Confidence Interval for  when  unknown (samples
are assumed to come from a normal population)
y  t / 2, n 1
s
n
with df = n - 1 and confidence
coefficient (1 - ). (Can use
z/2 if n>30.)
Example: Compute 95% CI for  given
y  40.1
df  n  1  8
y  t .025 ,8
s
n
s  5. 6
n9
t .025 ,8  2.306
40.1  2.306
  .05
z.025  1.960
40.1  3.659
40.1  4.304
 2  .025
5.6
9
One Sample Inf-5
The Level of Significance of a Statistical Test (p-value)
• Suppose the result of a statistical test you carry out is to reject the Null.
• Someone reading your conclusions might ask: “How close were you to not
rejecting?”
• Solution: Report a value that summarizes the weight of evidence in favor of Ho,
on a scale of 0 to 1. This the p-value. The larger the p-value, the more evidence
in favor of Ho.
Formal Definition: The p-value of a test is the probability of observing a
value of the test statistic that is as extreme or more extreme (toward Ha)
than the actually observed value of the test statistic, under the assumption
that Ho is true. (This is just the probability of a Type I error for the observed
test statistic.)
Rejection Rule: Having decided upon a Type I error
probability , reject Ho if p-value < .
One Sample Inf-6
Equivalence between confidence intervals and
hypothesis tests
Rejecting the two-sided null Ho:  = 0
is equivalent to
0 falling outside a (1-)100% C.I. for .
Rejecting the one-sided null Ho:   0
is equivalent to
0 being greater than the upper endpoint of a (1-2)100% C.I. for ,
or 0 falling outside a one-sided (1-)100% C.I. for  with –infinity as
lower bound.
Rejecting the one-sided null Ho:   0
is equivalent to
0 being smaller than the lower endpoint of a (1-2)100% C.I. for ,
or 0 falling outside a one-sided (1-)100% C.I. for  with +infinity as
upper bound.
One Sample Inf-7
Example: Practical Significance vs. Statistical Significance
Dr. Quick and Dr. Quack are both in the business of selling diets, and they
have claims that appear contradictory. Dr. Quack studied 500 dieters and
claims,
A statistical analysis of my dieters shows a significant weight loss for my
Quack diet.
The Quick diet, by contrast, shows no significant weight loss by its dieters. Dr.
Quick followed the progress of 20 dieters and claims,
A study shows that on average my dieters lose 3 times as much weight
on the Quick diet as on the Quack diet.
So which claim is right? To decide which diets achieve a significant weight loss
we should test:
Ho:   0
vs.
Ha:  < 0
where  is the mean weight change (after minus before) achieved by dieters on
each of the two diets. (Note: since we don’t know  we should do a t-test.)
One Sample Inf-8
MTB output for Quack diet analysis (Stat  Basic Stats  1 - Sample t)
One-Sample T: Quack
Test of mu = 0 vs mu < 0
Variable
Quack
N
500
Mean
-0.913
Variable
Quack
95.0% Upper Bound
-0.194
Difference
1
Size
500
StDev
9.744
T
-2.09
SE Mean
0.436
P
0.018
Power
0.6295
R output for Quick diet analysis (Read 20 values into vector “quack”)
> t.test(quick,alternative=c("less"),mu=0,conf.level=0.95)
One Sample t-test, data: quick
t = -1.0915, df = 19, p-value = 0.1443
alternative hypothesis: true mean is less than 0
95 percent confidence interval: -Inf 1.594617
sample estimate of mean of x: -2.73
> power.t.test(n=20,delta=1,sd=11.185,type="one.sample",
alternative="one.sided")
n = 20, delta = 1, power = 0.104
One Sample Inf-9
Summary
1.
Quick’s average weight loss of 2.73 is over 3 times as much as the 0.91
weight loss reported by Quack.
2.
However, Quack’s small weight loss was significant, whereas Quick’s larger
weight loss was not! So Quack might not have a better diet, but he has more
evidence, 500 cases compared to 20.
Remarks
1.
Significance is about evidence, and having a large sample size can make
up for having a small effect.
2.
If you have a large enough sample size, even a small difference can be
significant. If your sample size is small, even a large difference may not be
significant.
3.
Quick needs to collect more cases, and then he can easily dominate the
Quack diet (though it seems like even a 2.7 pound loss may not be enough
of a practical difference to a dieter).
4.
Both the Quick & Quack statements are somewhat empty. It’s not enough to
report an estimate without a measure of its variability. Its not enough to
report a significance without an estimate of the difference. A confidence
interval solves these problems.
One Sample Inf-10
A confidence interval shows both statistical and practical significance.
Quack two & one-sided 95% CIs
y  z.025
s
 .913 1.96(.436)  (1.768,0.058)
n
(, y  z.05 s
n )  (,0.194)
One-sided CI says
mean is sig. less than
zero.
Quick two & one-sided 95% CIs
y  t.025,19
s
11.185
 2.73  2.093
 (7.965,2.505)
n
20
(, y  t.05,19 s
n )  (,1.595)
One-sided CI says
mean is NOT sig. less
than zero.
One Sample Inf-11