No Slide Title

Download Report

Transcript No Slide Title

STAT 651
Lecture 7
Copyright (c) Bani Mallick
1
Topics in Lecture #7

Sample size for fixed power

Never, ever, accept a null hypothesis

Paired comparisons in SPSS

Student’s t-distributions

Confidence intervals when s is unknown

SPSS output on confidence intervals,
without formulae
Copyright (c) Bani Mallick
2
Book Sections Covered in Lecture #7

Chapter 5.5 (sample size)

Chapter 6.4 (paired data)

Chapter 5.7 (t-distribution)

My own screed (never, ever, accept a null
hypothesis)
Copyright (c) Bani Mallick
3
Lecture 6 Review: Hypothesis
Testing

Suppose you want to know whether the
population mean change in reported caloric
intake equals zero

We have already done this!!!!!

Confidence intervals tell you where the
population mean m is, with specified
probability

If zero is not in the confidence interval, then
you can reject the hypothesis
Copyright (c) Bani Mallick
4
Lecture 6 Review: Type I Error
(False Reject)

A Type I error occurs when you say that the
null hypothesis is false when in fact it is true

You can never know for certain whether or
not you have made such an error

You can only control the probability that
you make such an error

It is convention to make the probability of a
Type I error 5%, although 1% and 10% are
also used
Copyright (c) Bani Mallick
5
Lecture 6 Review: Type I Error Rates

Choose a confidence level, call it 1 - a

The Type I error rate is a

90% confidence interval: a = 10%

95% confidence interval: a = 5%

99% confidence interval: a = 1%
Copyright (c) Bani Mallick
6
Lecture 6 Review: Type II: The Other
Kind of Error

The other type of error occurs when you do
NOT reject H 0 even though it is false

This often occurs because you study sample
size is too small to detect meaningful
departures from H 0

Statisticians spend a lot of time trying to
figure out a priori if a study is large enough
to detect meaningful departures from a null
hypothesis
Copyright (c) Bani Mallick
7
Lecture 6 Review: P-values

Small p-values indicate that you have rejected
the null hypothesis

If p < 0.05, this means that you have
rejected the null hypothesis with a confidence
interval of 95% or a Type I error rate of 0.05

If p > 0.05, you did not reject the null
hypothesis at these levels
Copyright (c) Bani Mallick
8
Lecture 6 Review: Statistical Power

Statistical power is defined as the probability
that you will reject the null hypothesis when
you should reject it.

If b is the Type II error, power = 1 - b

The Type I error (test level) does NOT
depend on the sample size: you chose it
(5%?)

The power depends crucially on the sample
size
Copyright (c) Bani Mallick
9
Sample Size Calculations

You want to test at level (Type I error) a the
null hypothesis that the mean = 0
• You want power 1 - b to detect a change of
from the hypothesized mean by the amount
D or more, i.e., the mean is greater than
D or the mean is less than -D
• There is a formula for this!!
Copyright (c) Bani Mallick
10
Sample Size Calculations

a, b, D, s

Look up za/2 and zb

Remember what they are?

Find the values in Table 1 which give you
readings of 1-a/2 and 1-b

Required sample size is
n
Copyright (c) Bani Mallick
s2
D
2
z
a 2
 zb 
2
11
Sample Size Calculations

a=0.01, 1-b =0.90, D =180, s =600

Look up za/2 =2.58 and zb =1.28 (Check this)

n
s
2
D2
z
a 2  zb 
= 166
2

a=0.01, 1-b =0.80, D =180, s =600, zb
=0.84 (Check this)

n = 130: the less power you want, the
smaller the sample size
Copyright (c) Bani Mallick
12
More on Sample Size Calculations

Most often, sample sizes are done by
convention or convenience:

Your professor has used 5 rats/group before
successfully

You have time only to interview 50 subjects in
total
Copyright (c) Bani Mallick
13
More on Sample Size Calculations

More often, sample sizes are done by
convention or convenience:

In this case, the sample size calculations can
be used after a study if you find no
statistically significant effect

You can then guess how large a study you
would have needed to detect the effect you
have just seen but which was not statistically
significant
Copyright (c) Bani Mallick
14
Never Accept a Null Hypothesis

Suppose we use a 95% confidence interval, it
includes zero. Why do I say: with 95%
confidence, I cannot reject that the
population mean is zero.

I never, ever say: I can therefore conclude
that the population mean is zero.

Why is this? Are statisticians just weird?
(maybe so, but not in this case)
Copyright (c) Bani Mallick
15
Never Accept a Null Hypothesis:
Reason 1

Suppose we use a 95% confidence interval, it
includes zero: [-3,6]. Why do I say: with
95% confidence, I cannot reject that the
population mean is zero.

Remember the definition of a confidence
interval: the chance is 95% that the true
population mean is between -3 and 6: hence,
the true population mean could be 5, and is
not necessarily = 0.
Copyright (c) Bani Mallick
16
Never Accept a Null Hypothesis:
Reason 2

Suppose we use a 95% confidence interval, it
includes zero: [-3,6]. Why do I say: with
95% confidence, I cannot reject that the
population mean is zero.

Potential for chicanery: if you want to
accept the null hypothesis, how can you best
insure it?
Copyright (c) Bani Mallick
17
Never Accept a Null Hypothesis:
Reason 2

An example of chicanery: generic drugs

In the pharmaceutical industry, all the
expense involves getting a drug approved by
the FDA

After a drug goes off-patent, generic drugs
can be marketed

The main regulation is that the generic must
be shown to be “bioeqiuvalent” to the patent
drug
Copyright (c) Bani Mallick
18
Never Accept a Null Hypothesis:
Reason 2

The generic must be shown to be
“bioeqiuvalent” to the patent drug

One way would be to run a study and do a
statistical test to see whether the drugs have
the same effects/actions: the null hypothesis
is that the patent and generic are the same

The alternative is that they are not

If the null is rejected, the generic is rejected,
and $$$ issues arise
Copyright (c) Bani Mallick
19
Never Accept a Null Hypothesis:
Reason 2

Test to see whether the drugs have the same
effects/actions: the null hypothesis is that the
patent and generic are the same

If the null is rejected, the generic is rejected,
and $$$ issues arise

If you pick a tiny sample size, there is no
statistical power to reject the null
hypothesis
Copyright (c) Bani Mallick
20
Never Accept a Null Hypothesis:
Reason 2

If you pick a tiny sample size, there is no
statistical power to reject the null
hypothesis

The FDA is not stupid: they insist that the
sample size be large enough that any
medically important differences can be
detected with 80% (1 - b) statistical power
Copyright (c) Bani Mallick
21
Never Accept a Null Hypothesis





p-values are not the probability that the null
hypothesis is true.
For example, suppose you have a vested
interest in not rejecting the null hypothesis.
Small sample sizes have the least power for
detecting effects.
Small sample sizes imply large p-values.
Large p-values can be due to a lack of power,
or a lack of an effect.
Copyright (c) Bani Mallick
22
Paired Comparisons: Count you
Number of Populations!




The hormone assay data illustrate an
important point.
Sometimes, we measure 2 variables on the
same individuals
Reference Method and Test Method
There is only 1 population. How do we
compare the two variables to see if they have
the same mean?
Copyright (c) Bani Mallick
23
Paired Comparisons: Count you
Number of Populations!



There is only 1 population. How do we
compare the two variables to see if they have
the same mean?
Answer (Ott & Longnecker, Chapter 6.4): do
what we did and first compute the difference
of the variables and make inference on this
difference: now have 1 variable
In making inference, match the number of
variables to the number of populations!
Copyright (c) Bani Mallick
24
Paired Comparisons in SPSS

SPSS has a nice routine way of doing a paired
comparison analysis, providing confidence
intervals and p-values

“Analyze”

“Compare Means”

“Paired Samples t-test”

Highlight the variables that are paired and
select: use “options” to get other than 95%
CI
Copyright (c) Bani Mallick
25
Paired Comparisons in SPSS

Demo using computer comes next
Copyright (c) Bani Mallick
26
Boxplots and Histograms for Paired
data



For paired data, SPSS makes it easy to
automatically get confidence intervals: it
takes the difference of the paired variables
for you
However, for boxplots, qq-plots, etc., you
have to do this manually.
Here is how you can define a new variable,
called “differen”, in the armspan data for
males.
Copyright (c) Bani Mallick
27
Computing the Difference in Paired
Comparisons

Click on “Transform”

Click on “Compute”



New window shows up, in “Target Variable”
type in differen
Click on “Type & Label” and type in your label
(Height - Armspan in Inches)
click on “Continue”
Copyright (c) Bani Mallick
28
Computing the Difference in Paired
Comparisons


Highlight height and move over by clicking
the mover button
In “Numeric Expression”, type in the minus
sign -

Highlight armspan and move over

Click on “OK”

You are done!
Copyright (c) Bani Mallick
29
Selecting Cases in SPSS

“Data”

“Select Cases”

Push button of “If condition is satisfied”

Select “If”

Select “Gender” and move over

Then type = ‘Female’ and “Continue”

“OK” --> all analyses will be on Females
Copyright (c) Bani Mallick
30
Student’s t-Distribution




In real life, the population standard deviation
s is never known
We estimate it by the sample standard
deviation s
To account for this estimation, we have to
make our confidence intervals (make a
guess): longer or shorter?
Stump the experts!
Copyright (c) Bani Mallick
31
Student’s t-distribution



Of course: you have to make the confidence
interval longer!
This fact was discovered by W. Gossett, the
brewmaster of Guinness in Dublin.
He wrote it up anonymously under the name
“Student”, and his discovery is hence called
Students t-distribution because he used
the letter t in his paper.
Copyright (c) Bani Mallick
32
Student’s t-Distribution




Effectively, if you want a (1-a100%
confidence interval, what you do is to replace
za/2 (1.645, 1.96, 2.58) by ta/2(n-1) found
in Table 2 of the book.
n-1 is called the degrees of freedom
The increase in length of the confidence
interval depends on n.
If n gets larger, does the CI get larger or
smaller?
Copyright (c) Bani Mallick
33
Student’s t-Distribution

The (1-a100% CI when s was known was
X  za /2s / n

The (1-a100% CI when is s unknown is

X  ta /2 (n-1)s / n
You replace


s by s and
za /2 by ta/2(n-1)
Copyright (c) Bani Mallick
34
Student’s t-Distribution

Take 95% confidence, a = 0.05

za/2 = 1.96

n = 3, n-1 = 2,

n = 10, n-1 = 9,

n = 30, n-1 = 29,

n = 121, n-1 = 120,
ta/2(n-1) = 4.303
ta/2(n-1) = 2.262
ta/2(n-1) = 2.045
ta/2(n-1) = 1.98
Copyright (c) Bani Mallick
35
Student’s t-Distribution



Luckily, SPSS is smart.
It automatically uses Student’s tdistribution in constructing confidence
intervals and p-values!
So, all the output you will see in SPSS has
this correction built in
Copyright (c) Bani Mallick
36
Student’s t-Distribution


In the old days, people used the t-test to
decide whether the hypothesize value is in
the CI.
If your hypothesis is that m= 0, then you
reject the hypothesis if
X
t=
 ta /2 (n-1) or < -ta /2 (n-1)
s/ n

You learn nothing from this not available in a
CI, but its value is in SPSS
Copyright (c) Bani Mallick
37
WISH Numerical Illustration

s = 613, Xbar = -180

n = 3, s.e. = 613 / 31/2 = 354, ta/2(n-
1) = 4.303, CI is -180 plus and minus 1523,
hence the interval is [-1703, 1343]


n = 121, s.e. = 613 / 1211/2 = 59,
ta/2(n-1) = 1.98, CI is -180 plus and minus
118, hence the interval is [-298,-62]
Note change in conclusions!
Copyright (c) Bani Mallick
38
Armspan Data for Males



Outcome is height – armspan in inches
In SPSS, “Analyze”, “Descriptives”, “Explore”
will get you to the right analysis
Illustrate how to do this in SPSS
Copyright (c) Bani Mallick
39
Armspan Data for Males

Sample mean = -0.26

Sample standard error = 0.2391

Lower bound of 95% CI = -0.7406

Upper bound of 95% CI = 0.2206

Is there evidence with 95% confidence that
armspans for males differ systematically from
heights?
Copyright (c) Bani Mallick
40
Armspan Data for Males

Might ask: what about with 90% confidence

Illustrate how to do this in SPSS
Copyright (c) Bani Mallick
41
Armspan Data for Males

Sample mean = -0.26

Sample standard error = 0.2391

Lower bound of 90% CI = -0.6609

Upper bound of 90% CI = 0.1409

Is there evidence with 90% confidence that
armspans for males differ systematically from
heights?
Copyright (c) Bani Mallick
42
Armspan Data for Males





SPSS will compute the p-value for you as well
as confidence intervals.
For paired comparisons, “Analyze”, “Compare
Means”, “Paired Sample”.
Highlight the paired variables.
It computes the difference of the first named
variable in the list minus the second
Illustration in SPSS
Copyright (c) Bani Mallick
43
Armspan Data for Males

t = -1.087

p-value (significance level) = 0.282

SPSS also automatically does a 95%
confidence interval for the population mean
difference between heights and armspans
Copyright (c) Bani Mallick
44