Inferential Statistics

Download Report

Transcript Inferential Statistics

SAA 2023
COMPUTATIONALTECHNIQUE FOR
BIOSTATISTICS
Inferential Statistics
Inferential Statistics
Computational Statistics:
- All you need to do is choose
statistic
- Computer does all other
steps for you!!!!!!
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
For instance, if I am asked by a layperson what is meant
exactly by statistics, I will refer to the following Old
Persian saying: “Mosht Nemouneyeh Kharvar Ast
(translated: a handful represents the heap).
Inferential Statistics
This brief statement will describe, in one sentence, the
general concept of inferential statistics. In other words,
learning about a population by studying a randomly
chosen sample from that particular population can be
explained by using an analogy similar to this one.
Inferential Statistics
Involves using obtained sample
statistics to estimate the corresponding
population parameters
Most common inference is using a
sample mean to estimate a population
mean (surveys, opinion polls)
Drawing conclusions from sample to
population
Sample should be representative
Inferential Statistics
Inferential statistics allow us to make
determinations about whether groups
are significantly different from each
other.
Inferential Statistics
Inferential statistics is not a system of
magic and trick mirrors. Inferential
statistics are based on the concepts of
probability (what is likely to occur) and
the idea that data distribute normally
Inferential Statistics
Sample of
observations
Entire population of
observations
Random selection
Statistic
Parameter
µ=?
X
Statistical inference
Inferential Statistics
Inferential statistics is based on a strange
and mystical concept called falsification.
Although you might think the process is
simple
Write a hypothesis, test it, hope to prove it
Inferential statistics works this way:
Inferential Statistics
Write a hypothesis you believe to be true.
Write the OPPOSITE of this hypothesis,
which is called the null hypothesis.
Test the null, hoping to reject it.
- If the null is rejected, you have evidence
that the hypothesis you believe to be
true may be true.
- If the null is failed to reject, reach no
conclusion.
Inferential Statistics
Two major sources of error in research
Inferential statistics are used to make generalizations
from a sample to a population. There are two sources of
error (described in the Sampling module) that may result
in a sample's being different from (not representative of)
the population from which it is drawn.
These are:
Sampling error - chance, random error
Sample bias - constant error, due to
inadequate design
Inferential Statistics
Inferential statistics take into account sampling
error. These statistics do not correct for
sample bias. That is a research design issue.
Inferential statistics only address random
error (chance).
Sampling error - chance, random error
Sample bias - constant error, due to
inadequate design
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
p-value
The reason for calculating an inferential statistic is
to get a p-value (p = probability). The p value is
the probability that the samples are from the
same population with regard to the dependent
variable (outcome).
Usually, the hypothesis we are testing is that the
samples (groups) differ on the outcome. The pvalue is directly related to the null hypothesis.
Inferential Statistics
The p-value determines whether or not we
reject the null hypothesis. We use it to
estimate whether or not we think the null
hypothesis is true.
The p-value provides an estimate of how
often we would get the obtained result by
chance, if in fact the null hypothesis were
true.
Inferential Statistics
If the p-value (< α- value) is small, reject
the null hypothesis and accept that the
samples are truly different with regard to
the outcome.
If the p-value (> α- value) is large, fail to
reject the null hypothesis and conclude
that the treatment or the predictor variable
had no effect on the outcome.
Inferential Statistics
Steps for testing hypotheses





Calculate descriptive statistics
Calculate an inferential statistic
Find its probability (p-value)
Based on p-value, accept or reject the null hypothesis (H0)
Draw conclusion
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
One-Sample T: GPAs
Variable N Mean StDev SE Mean
90% CI
GPAs
50 2,9580 0,4101 0,0580 (2,8608; 3,0552)
Variable N Mean StDev SE Mean
95% CI
GPAs
50 2,9580 0,4101 0,0580 (2,8414; 3,0746)
Sample mean = 2,96
- 90% confident that μ (population mean) is between 2,86 and 3,06
- 95% confident that μ (population mean) is between 2,84 and 3,07
95% CI width > 90% CI width
The larger confidence coefficient, the greater the CI width
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
Inferential Statistics
One-Sample T: GPAs
Test of μ = 3 vs < 3 (directional, one-tail to the right)
95% Upper
Variable N Mean StDev SE Mean
Bound
T
P
GPAs
50 2,9580 0,4101 0,0580
3,0552 -0,72 0,236
Test of μ = 3 vs ≠ 3 (non-directional, two-tail)
Variable N Mean StDev SE Mean
95% CI
T
P
GPAs
50 2,9580 0,4101 0,0580 (2,8414; 3,0746) -0,72 0,472
Test of μ = 3 vs > 3 (directional, one-tail to the left)
95% Lower
Variable N Mean StDev SE Mean
Bound
T
P
GPAs
50 2,9580 0,4101 0,0580
2,8608 -0,72 0,764
Inferential Statistics
One-Sample T: GPAs
Test of μ = 3 vs < 3
95% Upper
Variable N Mean StDev SE Mean
Bound
T
P
GPAs
50 2,9580 0,4101 0,0580
3,0552 -0,72 0,236
The test statistic, t = -0,72, is the number of std deviations that the sample mean
2,96 is from the hypothesized mean, μ = 3.
p-value is the probability that a random sample mean is less than or equal to
2,96 when Ho: μ = 3 is true.
The rejection region consists of all p-values less than α.
p-value = 0,236 > α = 0,05
Then, we fail to reject Ho
Conclusion: There is not sufficient evidence in the sample to conclude that the
true mean GPA, μ, is less than 3.00.
Inferential Statistics
Two sample Ttest
One-sample t-test
One-sample t-test
Inferential Statistics
Non-directional (2-tailed test): In this form of the test, departure can
be observed from either end of the distribution. Thus, no direction
for expected results are specified. The null and alternative
hypotheses are as follows:
Directional (1-tailed test): In this form of the test, the rejection region
lies at only one end of the distribution. The direction is specified
before any analysis begins.
One-sample t-test
Cholesterol Example: Suppose population mean μ is 211
x = 220 mg/ml
s = 38.6 mg/ml
n = 25 (town)
H0 : m = 211 mg/ml
HA : m ¹ 211 mg/ml
t0.05,24  2.064
X  0 220  211
t

 1.17
s / n 38.6 / 25
For an a = 0.05 test we use the critical value determined
from the t(24) distribution.
Since |t| = 1.17 < 2.064 (table t) at the a = 0.05 level
We fail to reject H0.
The difference is not statistically significant
Hypothesis Testing
One-sample t-test
 If the observation is
outside the standard, we
reject the hypothesis that
the sample is
representative of the
population
6
4
2
10
/9
0
20
/8
0
30
/7
0
40
/6
0
50
/5
0
60
/4
0
70
/3
0
80
/2
0
90
/1
0
99
/0
1
0
10
8
6
4
2
0
10
/9
0
20
/8
0
30
/7
0
40
/6
0
50
/5
0
60
/4
0
70
/3
0
80
/2
0
90
/1
0
99
/0
1
 We observe a sample and
infer information about
the population
8
10
8
6
4
2
0
10
/9
0
20
/8
0
30
/7
0
40
/6
0
50
/5
0
60
/4
0
70
/3
0
80
/2
0
90
/1
0
99
/0
1
 We set a standard beyond
which results would be
rare (outside the
expected sampling
error)
10
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Degree of Freedom (df)
Two-sample t-test
Two-sample t-test
Inferential Statistics
Non-directional (2-tailed test): In this form of the test, departure can
be observed from either end of the distribution. Thus, no direction
for expected results are specified. The null and alternative
hypotheses are as follows:
Directional (1-tailed test): In this form of the test, the rejection region
lies at only one end of the distribution. The direction is specified
before any analysis begins.
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
Two-sample t-test
N
Healthy 9
Cystic Fibrosis 13
Mean Std. Dev.
18.9
11.9
5.9
6.3
We test the hypothesis of equal means for the two
populations, assuming a common variance.
H0 : 1 = 2, HA : 1  2
( x1  x2 )  ( 1   2 )
t
~ t (df  n1  n2  2)
s x1  x2
Two-sample t-test
s
2
n

1
 1s12  n2  1s22
n1  n2  2 
9  1 5.9


 13  1  6.32
9  13  2 

sx1  x2
t
2

754.76   37.73
20 
1 1
1 1
 s (  )  37.7 * (  )  2.66
n1 n2
9 13
2
x1  x 2
sx1 x2

18.9  11.9

 2.63
2.66
t0.0520  2.086
| t | t0.0520 , reject H 0 , accept H A .
Two-sample t-test
MINITAB PROCEDURE – Comparing Two Population Means (RealEstate)
Stat ► Basic Statistics ► 2-Sample t
Choose Samples in one column
Samples: Select …………(SalePrice)
Subscripts: Select ………..(Location)
Click Graphs Check Boxplots of data OK
Click Options
Confidence level: Enter 90 OK OK
Boxplot of SalePrice
1000000
900000
Ho : μ1 = μ2 or μ1 - μ2 = 0
800000
HA : μ1 ≠ μ2 or μ1 - μ2 ≠ 0
SalePrice
700000
600000
500000
400000
300000
200000
100000
East
West
Location
Two-sample t-test
Result: Two-sample T for SalePrice
Location N Mean StDev SE Mean
East
34 312702 188163 32270
West
26 339143 255571 50122
Difference = μ (East) - μ (West)
Estimate for difference: -26442
90% CI for difference: (-126602; 73719)
T-Test of difference = 0 (vs not =): T-Value = -0,44 p-Value = 0,660 DF = 44
The difference in mean sale prices = $26,441
90% Confidence that the true mean difference in sale prices is in the interval
(-$126,602 < μ (East) - μ (West) < $73,719)
p-Value = 0,660 > α = 0,05/0,10
Fail to reject Ho
We do not have sufficient evidence in the sample to conclude that there are
significant differences in mean sale prices in markets east and west.
Paired t-test
 Test to deal with two observations with strong
comparability (e.g. two treatments on the
same individuals, or one individual Before vs.
After treatment, very close plots)
 Sample 1. X11, X12, …, X1n
 Sample 2. X21, X22, …, X2n
 Method:
Calculate differences between two measurements
for each individual di = Xi1 – Xi2
Calculate
2
d

d 
n
j
, sd 
 (d
j
d)
n 1
, sd 
sd
n
Paired t-test
H 0 :  d  0, H A  0.
d  d
d
t

~ t ( n  1)
sd
sd
n
Paired t-test
Test infection of virus on tobacco leaves. Number of
death pots on leaves.
Virus 1
X1j
Plant
1
2
3
4
5
6
7
8
Total
Mean
9
17
31
18
7
8
20
10
120
15
H 0 : d  0, H A  0.
Virus 2
X2j
dj
10
11
18
14
6
7
17
5
88
11
-1
6
13
4
1
1
3
5
32
4
  0.05, t0.05( 7 )  2.365
d 4
(1) 2  6 2  ...  52  (32) 2 / 8
sd 
8 1
 4.31
4.31
sd 
 1.52
8
4
t
 2.632  t0.05
1.52
Paired t-test
Advantages:
2
2
1). Usually sd  sx1  x2 , it is easy to find
a true small difference.
2). Do not need to consider if the
variances of two populations are same
or not.
Paired t-test
Hypothesis Testing
Hypothesis testing is always a fivestep procedure:
Formulation of the null and the
alternative hypotheses
Specification of the level of significance
Calculation of the test statistic
Definition of the region of rejection
Selection of the appropriate hypothesis
Hypothesis Testing
3 Steps to carry out a Hypothesis Test
1. Specify the 5 required elements of a
Hypothesis test listed above.
2. Using the sample data, compute either the
value of the test statistic or the p-value
associated with the calculated test
statistic.
Hypothesis Testing
3 Steps to carry out a Hypothesis Test
3.
Use one of three possible decision rules (don't mix
them up!):
1. Table method: Compare the calculated value with a
table of the critical values of the test statistic
 If the absolute (calculated) value of the test statistic 
to the critical value from the table, reject the null
hypothesis (HO) and accept the alternative hypothesis
(HA).
 If the calculated value of the test statistic < the critical
value from the table, fail to reject the null hypothesis
(H0).
Hypothesis Testing
Hypothesis Testing
Hypothesis Testing
3 Steps to carry out a Hypothesis Test
3. Use one of three possible decision rules
(don't mix them up!):
2. Graph method: Compare the calculated value
with the t-distribution graph of the test statistic
 Reject the NULL hypothesis if the test statistic
falls in the critical region. Fail to reject the
NULL if the test statistic does not fall in the
critical region.
Hypothesis Testing
Rejection region at the α=10% significance level.
Hypothesis Testing
Rejection region at the α=5% significance level.
Hypothesis Testing
Rejection region at the α=10% significance level.
Hypothesis Testing
3 Steps to carry out a Hypothesis Test
3. Use one of three possible decision rules
(don't mix them up!):
3. p-value method: Reject the NULL hypothesis
if the p-value is less that α. Fail to reject the
NULL if the p-value is greater than α.
 The exact p-value can be computed, and if p <
0.05, then H0 is rejected and the results are
declared statistically significant . Otherwise, if p
 0.05 then H0 is failed to reject and the results
are declared not statistically significant .
Hypothesis Testing
Hypothesis Testing
Example of hypothesis testing can be found
in the jury system. There are three party
involved in a court case. i.e. plaintiff
(prosecutors), defendant (accuse) and the
judges.
The judge will form a hypothesis as below
before hearing a case;
Hypothesis
Ho : The evidences are not significantly
strong enough to proof the defendant guilty
H1 : The evidences are significantly strong to
proof defendant guilty
Hypothesis Testing
The main function of the plaintiff play in a
case is to continuously supply strong
evidence to proof the defendant guilty.
Whereas the function of the defendant is to
defend himself by rejecting the evident
provide by the plaintiff.
The judges role is to collect information
supplied by the plaintiff and defendant and
make decision about the hypothesis validity.
Hypothesis Testing
To perform a hypothesis test, we start with two
mutually exclusive hypotheses. Here’s an
example: when someone is accused of a crime,
we put them on trial to determine their innocence
or guilt. In this classic case, the two possibilities
are the defendant is not guilty (innocent of the
crime) or the defendant is guilty. This is
classically written as:
H0: Defendant is Innocent ← Null Hypothesis
HA: Defendant is Guilty ← Alternate Hypothesis
Hypothesis Testing
Hypothesis Testing
Unfortunately, our justice systems are not
perfect. At times, we let the guilty go free and
put the innocent in jail. The conclusion drawn
can be different from the truth, and in these
cases we have made an error. The table below
has all four possibilities. Note that the
columns represent the “True State of Nature”
and reflect if the person is truly innocent or
guilty. The rows represent the conclusion
drawn by the judge or jury.
Type I and Type II Errors
Type I and Type II Errors
Alpha and Beta
Type I and Type II Errors
Type I and Type II Errors
NULL HYPOTHESIS
TRUE
FALSE
Reject the null Type I error
α
CORRECT
hypothesis
Rejecting a true
null hypothesis
Fail to reject the
CORRECT
null hypothesis
Type II error
β
Failing to reject a
false null hypothesis
Type I and Type II Errors
Two of the four possible outcomes are
correct. If the truth is they are innocent and
the conclusion drawn is innocent, then no
error has been made. If the truth is they are
guilty and we conclude they are guilty,
again no error. However, the other two
possibilities result in an error.
Type I and Type II Errors
A Type I (read “Type one”) error is when
the person is truly innocent but the jury
finds them guilty. A Type II (read “Type
two”) error is when a person is truly guilty
but the jury finds him/her innocent. Many
people find the distinction between the
types of errors as unnecessary at first;
perhaps we should just label them both as
errors and get on with it.
Type I and Type II Errors
However, the distinction between the two
types is extremely important. When we
commit a Type I error, we put an innocent
person in jail. When we commit a Type II
error we let a guilty person go free. Which
error is worse?
Type I and Type II Errors
The generally accepted position of society
is that a Type I Error or putting an innocent
person in jail is far worse than a Type II
error or letting a guilty person go free. In
fact, the burden of proof in criminal cases
is established as “Beyond reasonable
doubt”.
Another way to look at Type I vs. Type II
errors is that a Type I error is the
probability of overreacting and a Type II
error is the probability of under reacting.
Type I and Type II Errors
In statistics, we want to quantify the
probability of a Type I and Type II error.
The probability of a Type I Error is α (Greek
letter “alpha”) and the probability of a Type
II error is β (Greek letter “beta”). Without
slipping too far into the world of theoretical
statistics and Greek letters, let’s simplify
this a bit. What if the probability of
committing a Type I error was 20%?
Type I and Type II Errors
A more common way to express this would
be that we stand a 20% chance of putting
an innocent man in jail. Would this meet
your requirement for “beyond reasonable
doubt”? At 20% we stand a 1 in 5 chance
of committing an error. This is not sufficient
evidence and so cannot conclude that
he/she is guilty.
Type I and Type II Errors
The formal calculation of the probability of
Type I error is critical in the field of
probability and statistics. However, the
term "Probability of Type I Error" is not
reader-friendly. For this reason, the phrase
"Chances of Getting it Wrong" is used
instead of "Probability of Type I Error".
Type I and Type II Errors
Most people would agree that putting an
innocent person in jail is "Getting it Wrong"
as well as being easier for us to relate to.
To help you get a better understanding of
what this means, the table below shows
some possible values for getting it wrong.
Type I and Type II Errors
Chances of Getting it Wrong
(Probability of Type I Error)
Percentage
Chances of sending an innocent man to jail
20% Chance
1 in 5
5% Chance
1 in 20
1% Chance
1 in 100
.01% Chance
1 in 10,000
Controlling Type I and Type II Errors
 , , and n are interrelated. If one is kept constant,
then an increase in one of the remaining two will
cause a decrease in the other.
 For any fixed , an increase in the sample size n will
cause a ??????? in 
 For any fixed sample size n , a decrease in  will
cause a ??????? in .
 Conversely, an increase in  will cause a ???????
in  .
 To decrease both  and , ??????? the sample
size n.
Planning a study
Suppose
you
were
interested
in
determining whether treatment X has an
effect on outcome Y—there are several
issues that need to be addressed so that a
sound inference can be made from the
study result
Planning a study
What is the population?
How will you select a sample that is
representative of that population?
There are many ways to produce a sample, but
not all of them will lead to sound inference
Sampling strategies
Probability samples—result when subjects
have a known probability of entering the
sample
Simple random sampling
Stratified sampling
Cluster sampling
Sampling strategies
Probability samples can be made to be
representative of a population
Non-probability samples may or may not
be representative of a population—it may
be difficult to convince someone that the
sample results apply to any larger
population
Planning a study
Clinical trials are generally designed to be
efficacy trials—highly controlled situations
that maximize internal validity
We want to design a study to test the
effect of treatment X on outcome Y, and try
to make sure that any difference in Y is
due to X
Planning a study
 At the end of this study you observe a difference
in outcome Y between the experimental group
and the control group.
 All of the effort in designing the study with strict
control is for one reason—at the end of the study
you want only two plausible explanations for the
observed outcome
Chance
Real effect of treatment X
Planning a study
 The reason you want only these two
explanations is because if you can rule out
chance, you can conclude that treatment X must
have been the reason for the difference in
outcome Y
 All inferential statistical tests are used to
estimate the probability of the observed outcome
assuming chance alone is the reason for the
difference.
 If there are multiple competing explanations for
the observed result, then ruling out chance
offers little information about the effectiveness of
treatment X
Inferential statistics
Hypothesis testing—answering the
question of whether or not treatment X
may have no effect on outcome Y
Point estimation—determining what the
likely effect of treatment X is on outcome Y
Hypothesis testing
The goal of hypothesis testing is
somewhat twisted — it is to disprove
something you don’t believe
In this case you are trying to disprove that
treatment X has no effect on outcome Y
You start out with two hypotheses
Hypothesis testing
Null Hypothesis (HO)
Treatment X has no effect on outcome Y
Alternative Hypothesis (HA)
Treatment X has an effect on outcome Y
Hypothesis testing
 If the trial has been carefully controlled, there
are only two explanations for a difference
between treatment groups—efficacy of X,
and chance
 Assuming that the null hypothesis is correct,
we can use a statistical test to calculate that
the observed difference would have occurred.
This is known as the significance level, or pvalue of the test.
Hypothesis testing
P-value
The probability of the observed outcome,
assuming that chance alone was involved in
creating the outcome.
In other words,
assuming the null hypothesis is correct, what is
the probability that we would have seen the
observed outcome.
This is only meaningful if chance is the only
competing plausible explanation.
Hypothesis testing
If the p-value is small, meaning the
observed outcome would have been
unlikely, we will reject that chance played
the only role in the observed difference
between groups and conclude that
treatment X does in fact have an effect on
outcome Y
How small is small?
Hypothesis testing
Reality ->
Decision
Retain HO
Reject HO
HO is true
HO is false
Correct
Decision
Type II Error
()
(.2, .1)
Correct
Decision
Type I Error
()
(.05, .01)
Hypothesis testing
 Power analysis is used to try to minimize Type II
errors.
 Power (1-) is the probability of rejecting the null
hypothesis when the effect of X is some
specified value other than zero.
 Usually one specifies an expected effect and
uses power analysis to calculate the sample size
needed to keep  below some value (.2 is
common)
Point estimation
 Hypothesis testing can only tell you whether or
not the effect of X is zero, it does not tell you
how large or small the effect is.
 Important — a p-value is not an indication of the
size of an effect, it depends greatly on sample
size
 If you want an estimate of the actual effect, you
need confidence intervals
Point estimation
 Confidence intervals give you an idea of what
the actual effect is likely to be in the population
of interest
 The most common confidence interval is 95%
and gives an upper and lower bound on what
the effect is likely to be.
 The size of the interval depends on the sample
size, variability of the measure, and the degree
of confidence you want that the interval contains
the true effect.
Point estimation
 Many people prefer confidence intervals to
hypothesis testing, because confidence
intervals contain more information
 Not only can you tell whether the effect could
be zero (is zero contained in the interval of
possible effect values?) but you also have the
entire range of possible values the effect
could be
 So, a confidence interval gives you all the
information of a hypothesis test and a whole
lot more.
Choosing the right test
Typically one is interested in comparing
group means.
If the outcome is continuous, and one
independent variable:
Two groups — t-test
Three or more groups -- ANOVA
Choosing the right test
If the outcome is continuous and there is
more than one independent variable:
ANOVA, if all independent variables are
categorical
ANCOVA or multiple linear regression, if some
independent variables are continuous
Logic of Hypothesis testing
The further the observed value is from the
mean of the expected distribution, the
more significant the difference
Steps in test of Hypothesis
1. Determine the appropriate test
2. Establish the level of significance:α
3. Determine whether to use a one tail or
two tail test
4. Calculate the test statistic
5. Determine the degree of freedom
6. Compare computed test statistic against
a tabled value
3. Determine Whether to Use One
or Two Tailed Test
If the alternative hypothesis specifies
direction of the test, then one tailed
Otherwise, two tailed
Most cases
5. Determine Degrees of Freedom
Number of components that are free to
vary about a parameter
Df = Sample size – Number of parameters
estimated
Df is n-1 for one sample test of mean
Common Inferential Stats
T-Test -compare means of two groups
interval/ratio level of measurement
independent samples t-test
dependent or paired samples
Common Inferential Stats
ANOVA - analysis of variance􀂄
more than 2 means to compare or more
than 2 testing of means􀂄
interval/ratio level of measurement
Chi-square (x2) -testing hypothesis about
number of cases that fall into various
categories
nominal/ordinal level of measurement
Summary
 Descriptive stats summarize measures of central tendency and
variability􀂄
 Inferential determine how likely it is that results based on sample are
the same in population􀂄
 Must know level of measurement of variables to choose correct
 Parametric and non-parametric two types of statistics requiring
analysis of assumptions
 􀂄Pearson r, t tests and ANOVA examples of parametric
 􀂄Pearson r measure relationship or association between 2 variables
 T test determines if there is a significant difference between 2 group
means
 ANOVA determines if there is a significant difference between 3 or
more means
 X2 non-parametric statistic to assess relationship between 2
categorical variables
Two types of ANOVA
1.Independent groups - two different sets
of individuals In the graphic below, college
students are randomly assigned to Groups
1 and 2.
Two types of ANOVA
Example: Researchers are interested in exam anxiety. They administer
an anxiety inventory to students just before the final exam in a
Sociology class. They also adminster it before the final exam in a
Political Science class. To compare the two sets of scores, they use
either ANOVA or t-test for independent samples (hand
calculation).
Two types of ANOVA
2.Paired samples (sometimes referred to as Repeated
Measures or With Replication) - either the same
individuals or from matched groups (i.e., matched on
everything but the treatment (level of the Independent
variable).
Two types of ANOVA
Example: Researchers are interested in exam anxiety. They administer
an anxiety inventory on the second day of class. Then they give it
again on the day of the midterm. To compare the two sets of scores,
they use either ANOVA with replication or t-test for paired samples
(hand calculation).