Calculate Sn, Sp, LRs

Download Report

Transcript Calculate Sn, Sp, LRs

PTP 560
• Research Methods
Week 9
Thomas Ruediger, PT
Calculate Sn, Sp, LRs
Confidence Intervals
with small samples
 What is small sample size?
 Less than 30 (one of those special numbers in stats)
 Sampling distribution tends to spread out
 Standard normal curve not adequate
 Use the t-distributions





Theoretical sampling distributions
Flatter peak, wider at the tails
Approaches normal curve as sample size increases
Use values of t instead of z
Described by degrees of freedom (n-1 for confidence
intervals)
Hypothesis testing
• Are the differences
– Representative of “real” effects?
– Just by chance?
• Null hypothesis ( HO)
– Means are not different
– Stated in terms of population parameter
– μA=μB
• Alternative hypothesis (H1)
– Difference too large to be by chance
– μA≠μB
– May be directional or non-directional
Truth
Decision
Ho is True Ho is False
Reject Ho
Do Not Reject Ho
Type I error
α
Correct
Correct
Type II error
β
Type I Error
• Significance Level (Alpha level , α level)
• Your choice of how much risk you are willing to
take of saying there is a difference when there
really is no difference
• Set this before the study
– Conventionally is 0.05
• This is arbitrary, but almost always what we choose
– Choose the level based on the Type I error concern
Type I Error
• Probability Values (Evaluated after the study)
– Probability of finding this big a difference by chance
• p = .07 of this big a difference by chance
– You are not stating the probability of the inverse
• p = .93 that it is real difference is not appropriate
– Compare p-value to alpha level if greater
– Compare your p-value (calculated after the study) with
your α level (set before the study)
– If the p-value is less than α, reject the null
– If the p-value is greater than α, fail to reject the null
Type II Error
Statistical Power
• Beta (β)
– Probability of failing to reject a false HO(null hypotheses)
– Β of 0.20 is 20% chance we will make a Type II error
• Statistical Power
– Complement of β (not compliment)
– In this example 0.80 (1.00 – 0.20 = 0.80)
– 80% probability of correctly rejecting the null
• Before: a priori - power used to determine sample
size
• After: post hoc – power reported if HO not rejected
“If there was a difference, could we have found it?”
Determinants of
Statistical Power
• Significance Criterion
– As α increases, power increases
(As α increases from 0.05 to 0.10, power increases)
• Variance
– As variance decreases, power increases
• Sample size
– As sample size increases, power increases
• Effect size (difference b/w the group means)
– As effect size increases, power increases
z• z - score represents the distance between:
– A sample score and
– Sample mean
– Divided by the standard deviation
You will see this in osteoporosis scores
(+2 for z- score is 2 SD away from a healthily woman mean)
• z - ratio represents the distance between:
– A sample mean and
– Population mean
– Divided by the standard error of the mean
Critical Region
 That portion of the curve above and below z
 If calculated z > critical z, reject HO
 One or two tailed test?
 Non directional hypothesis– two tailed
 z of 2.00 encompasses 95.44 % (non-critical)
 4.66% is the critical area
 z of 1.96 encompass 95%,
 Critical region is 5%, 2.5% in each tail of a non-directional test
 Directional hypothesis– one tailed
 z of 1.645 encompasses 95%
 Critical region is 5%, all in one tail of a directional test, while NON-direction will be 2.5%
 Practically, you are disregarding everything in the other tail
 Table A.1 back of P & W
Figure A: Intervention 1 is different than Intervention 2
Figure B: Intervention 1 is less effective than Intervention 2
Parametric Statistics
• Used to estimate population parameters
• Based on assumptions
– Randomly drawn from a normally distributed population
– Variances in the samples equal (at least roughly)
– Interval or ratio scale
• Classically, if assumptions violated, use nonparametric tests
• Many view parametric stats as Robust enough to
withstand even major violation
t-test
 Examines two means
 Two groups
 Two conditions/two performances
 Statistical significance based on
 Difference in the means
 Between the groups
 The effect size
 Variance
 Within the groups
 How variable are the scores
Fig 19.1
t-test
 Based on a ratio
 Difference between group means/Variability within groups
 Difference between means
 Treatment effect and error variance
 One mean- second mean and variability between the groups. In both the numerator and
denominator of t-test ratio, so holds it to zero.
 Variability within groups
 Error variance alone
 Equal and unequal variances affect t-ratio
 SPSS and most other packages automatically test for this. Where?

Ratio can be written: Treatment effect and error variance/Error variance
NOTE:
Error variance
 Not mistakes
 Is anything that is not due to the independent variable
t-test
 If the null is true
 Ratio reduces to: Error/Error
 The bigger the difference - The bigger the ratio
How does the ratio get bigger?
 This ratio is compared to the critical
 Determines significance
 Is the ratio significant?
 Based on critical value (but now t instead of z)
 Entering arguments (Table A.2)
 Alpha level (almost always 0.05
 Degrees of freedom ( one or a few less than n )
 CI can be constructed for where the true mean difference lies
t-test
 Independent t-test




Usually random assignment
Can be convenience assignment
No inherent relationship between the groups
Degrees of freedom = total sample size – 2
 Paired t-test
 An inherent relationship between the groups
 Self (repeated measure test)
 Twins
 Difference scores for each pair compared
 Degrees of freedom = number of paired scores – 1
 Find this in P & W or in an SPSS output table – don’t calculate for this class!
ANOVA
 Examines three (or more) means
 Three (or more) groups
 Three (or more) conditions/ three (or more) performances
 Statistical significance based on
 Difference in the means
 Between the groups
 The effect size
 Variance
 Within the groups
 How variable are the scores
 This should sound familiar
ANOVA
 Based on ratio
Treatment effect and error variance/Error variance

 For ANOVA it is the F-ratio
 Derived from the Sum of Squares (SS)
 Larger the SS the larger the ______________?
 Calculate SS
(each score minus sample mean, square each result, sum them)
 Then determine the Mean Square (MS)
 MSb = SSb/dfb (dfb = one less than the number of groups)
 MSe = SSe/df e (df e = total N – number of groups)
 F statistic is the ratio = MSb/Mse
 Ratio of the between groups variance to error variance

Find this in P & W or in an SPSS output table – don’t calculate for this class!