Statistical Inference

Download Report

Transcript Statistical Inference

Statistical Inference
Making decisions regarding the
population base on a sample
Decision Types
• Estimation
– Deciding on the value of an unknown parameter
• Hypothesis Testing
– Deciding a statement regarding an unknown parameter
is true of false
• Prediction
– Deciding the future value of a random variable
• All decisions will be based on the values of
statistics
Estimation
• Definitions
– An estimator of an unknown parameter is a sample
statistic used for this purpose
– An estimate is the value of the estimator after the data
is collected
• The performance of an estimator is assessed by determining
its sampling distribution and measuring its closeness to the
parameter being estimated
Examples of Estimators
The Sample Proportion
Let p = population proportion of interest
or binomial probability of success.
Let
X
no. of succeses
pˆ 

n no. of bimomial trials
= sample proportion or proportion of successes.
Then the sampling distributi on of p̂
is a normal distribution with
mean  pˆ  p
 pˆ 
p (1  p )
n
Sampling distributi on of p̂
30
25
20
15
c
10
5
0
0
0.1
0.2
0.3
 pˆ  p
0.4
0.5
0.6
0.7
0.8
0.9
1
The Sample Mean
Let x1, x2, x3, …, xn denote a sample of size n from a normal
distribution with mean  and standard deviation .
n
Let
x
x
i
i 1
n
 sample mean
Then the sampling distributi on of x
is a normal distribution with
mean  x  
x 

n
0.3
Sampling distributi on of x
0.25
population
0.2
n =5
n = 10
0.15
n = 15
c
n = 20
0.1
0.05
0
80
90
100
x  
110
120
Confidence Intervals
Estimation by Confidence Intervals
• Definition
– An (100) P% confidence interval of an unknown
parameter is a pair of sample statistics (t1 and t2)
having the following properties:
1. P[t1 < t2] = 1. That is t1 is always smaller than t2.
2. P[the unknown parameter lies between t1 and t2] = P.
•
the statistics t1 and t2 are random variables
•
Property 2. states that the probability that the
unknown parameter is bounded by the two
statistics t1 and t2 is P.
Critical values for a distribution
• The a upper critical value for a any distribution
is the point xa underneath the distribution such
that P[X > xa] = a
a
xa
Critical values for the standard Normal
distribution
P[Z > za] = a
a
za
Critical values for the standard Normal
distribution
P[Z > za] = a
Confidence Intervals for a proportion p
Let t1  pˆ  za / 2 pˆ  pˆ  za / 2
 pˆ  za / 2
pˆ 1  pˆ 
n
and t2  pˆ  za / 2 pˆ  pˆ  za / 2
 pˆ  za / 2
pˆ 1  pˆ 
n
p1  p 
n
p1  p 
n
Then t1 to t2 is a (1 – a)100% = P100%
confidence interval for p
z
Logic:
pˆ  p
 pˆ
has a Standard Normal distribution
Then P za  z  za   1  a  P


pˆ  p
and P  za / 2 
 za / 2   1  a
 pˆ


P   za / 2 pˆ  pˆ  p  za / 2 pˆ   1  a
P   za / 2 pˆ  p  pˆ  za / 2 pˆ   1  a
Hence


P pˆ  za / 2 pˆ  p  pˆ  za / 2 pˆ  1  a
Pt1  p  t2   1  a
Thus t1 to t2 is a (1 – a)100% = P100% confidence
interval for p
Example
• Suppose we are interested in determining the success rate
of a new drug for reducing Blood Pressure
• The new drug is given to n = 70 patients with abnormally
high Blood Pressure
• Of these patients to X = 63 were able to reduce the
abnormally high level of Blood Pressure
• The proportion of patients able to reduce the abnormally
high level of Blood Pressure was
X 63
pˆ 

 0.900
n 70
This is an
estimate of p.
If P = 1 – a = 0.95 then a/2 = .025
Then
and
and za = 1.960
pˆ 1  pˆ 
t1  pˆ  za / 2
n
0.900.10
 (0.90)  (1.960)
70
 (0.90)  .0703  0.8297
This comes from
the Table
pˆ 1  pˆ 
t2  pˆ  za / 2
n
0.900.10
 (0.90)  (1.960)
70
 (0.90)  .0703  0.9703
Thus a 95% confidence interval for p is 0.8297 to
0.9703
What is the probability that p is beween 0.8297 and
0.9703?
Is it 95% ?
Answer:
p (unknown) , 0.8297 and 0.9703 are numbers.
Either p is between 0.8297 and 0.9703 or it is not.
The 95% refers to success of confidence interval
procedure prior to the collection of the data.
After the data is collected it was either successful in
capturing p or it was not.
Statistical Inference
Making decisions regarding the
population base on a sample
Two Areas of Statistical
Inference
Estimation
Hypothesis Testing
Estimation
• Definitions
– An estimator of an unknown parameter is a sample
statistic used for this purpose
– An estimate is the value of the estimator after the data
is collected
• The performance of an estimator is assessed by determining
its sampling distribution and measuring its closeness to the
parameter being estimated
Confidence Intervals
Estimation of a parameter by a range
of values (an interval)
Estimation by Confidence Intervals
• Definition
– An (100) P% confidence interval of an unknown
parameter is a pair of sample statistics (t1 and t2)
having the following properties:
1.
P[t1 < t2] = 1. That is t1 is always smaller than t2.
2. P[the unknown parameter lies between t1 and t2] = P.
•
the statistics t1 and t2 are random variables
•
Property 2. states that the probability that the
unknown parameter is bounded by the two
statistics t1 and t2 is P.
Confidence Interval for a Proportion
100(1 – a)% Confidence Interval for the population proportion:
pˆ  za / 2 pˆ
 pˆ 
p1  p 

n
pˆ 1  pˆ 
n
za / 2  upper a / 2 critical point
of the standard normal distribtio n
Interpretation: For about 100(1 – a)P% of all randomly selected
samples from the population, the confidence interval computed in
this manner captures the population proportion.
Comment
The usual choices of a are 0.05 and 0.01
In this case the level of confidence, 100(1 - a)%,
is 95% and 99% respectively
Also the tabled value za/2 is:
z0.025 = 1.960 and
z0.005 = 2.576 respectively
Example
• Suppose we are interested in determining the success rate
of a new drug for reducing Blood Pressure
• The new drug is given to n = 70 patients with abnormally
high Blood Pressure
• Of these patients to X = 63 were able to reduce the
abnormally high level of Blood Pressure
• The proportion of patients able to reduce the abnormally
high level of Blood Pressure was
X 63
pˆ 

 0.900
n 70
This is an
estimate of p.
If P = 1 – a = 0.95 then a/2 = .025
Then
and
and za = 1.960
pˆ 1  pˆ 
t1  pˆ  za / 2
n
0.900.10
 (0.90)  (1.960)
70
 (0.90)  .0703  0.8297
This comes from
the Table
pˆ 1  pˆ 
t2  pˆ  za / 2
n
0.900.10
 (0.90)  (1.960)
70
 (0.90)  .0703  0.9703
Thus a 95% confidence interval for p is 0.8297 to
0.9703
What is the probability that p is beween 0.8297 and
0.9703?
Is it 95% ?
Answer:
p (unknown) , 0.8297 and 0.9703 are numbers.
Either p is between 0.8297 and 0.9703 or it is not.
The 95% refers to success of confidence interval
procedure prior to the collection of the data.
After the data is collected it was either successful in
capturing p or it was not.
Error Bound
For a (1 – a)% confidence level, the
approximate margin of error in a sample
proportion is
Error Bound  za
pˆ 1  pˆ 
n
Factors that Determine the Error Bound
1. The sample size, n.
When sample size increases, margin of error decreases.
2. The sample proportion, p̂ .
If the proportion is close to either 1 or 0 most individuals
have the same trait or opinion, so there
is little natural variability and the margin of error
is smaller than if the proportion is near 0.5.
3. The “multiplier” za/2.
Connected to the “(1 – a)%” level of confindence of the
Error Bound. The value of za/2 for a 95% level of
confidence is 1.96 This value is changed to change the
level of confidence.
Determination of Sample Size
In almost all research situations the
researcher is interested in the question:
How large should the sample be?
Answer:
Depends on:
• How accurate you want the answer.
Accuracy is specified by:
• Specifying the magnitude of
the error bound
• Level of confidence
Error Bound:
B  za / 2
p1  p 
 za / 2
n
pˆ 1  pˆ 
n
• If we have specified the level of confidence then the
value of za/2 will be known.
• If we have specified the magnitude of B, it will also
be known
Solving for n we get:
za2/ 2 p1  p  za2/ 2 p * 1  p *
n

2
2
B
B
Summarizing:
The sample size that will estimate p with an Error
Bound B and level of confidence P = 1 – a is:
za2/ 2 p * 1  p *
n
2
B
where:
• B is the desired Error Bound
• za is the a/2 critical value for the standard normal
distribution
• p* is some preliminary estimate of p.
If you do not have a preliminary estimate of p, use p* = 0.50
Reason
2
za / 2 p * 1  p *
n
For p* = 0.50
B2
n will take on the largest value.
3000
2500
n
2000
1500
1000
500
0
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
p*
Thus using p* = 0.50, n may be larger than required
if p is not 0.50. but will give the desired accuracy or
better for all values of p.
Example
• Suppose that I want to conduct a survey and want to
estimate p = proportion of voters who favour a downtown
location for a casino:
I know that the approximate value of p is
• p* = 0.50. This is also a good choice for p if one has no
preliminary estimate of its value.
• I want the survey to estimate p with an error bound B = 0.01
(1 percentage point)
• I want the level of confidence to be 95% (i.e. a = 0.05 and
za = z0.05 = 1.960
Then
2

1.960 0.50 0.50
n
 9604
2
0.01
Confidence Intervals
for the mean , ,
of a Normal Population
Confidence Intervals for the mean
of a Normal Population, 
Let
and
t1  x  za / 2 x  x  za / 2
t2  x  za / 2 x  x  za / 2

n

n
Then t1 to t2 is a (1 – a)100% = P100%
confidence interval for 
Logic:
z
x
has a Standard Normal distribution
x
Then P za  z  za   1  a  P


x
and P  za / 2 
 za / 2   1  a
x


Hence
Px  za / 2 x    x  za / 2 x   1  a
Pt1    t2   1  a
Thus t1 to t2 is a (1 – a)100% = P100% confidence
interval for 
Example
• Suppose we are interested average Bone Mass Density
(BMD) for women aged 70-75
• A sample n = 100 women aged 70-75 are selected and
BMD is measured for eahc individual in the sample.
• The average BMD for these individuals is:
x  25.63
• The standard deviation (s) of BMD for these individuals
is:
s  7.82  
If P = 1 – a = 0.95 then a/2 = .025
Then
t1  x  za / 2

 x  za / 2
and za = 1.960
s
n
n
7.82
 25.63  1.960
 25.63  1.53  24.10
100
and
t 2  x  za / 2

 x  za / 2
s
n
n
7.82
 25.63  1.960
 25.63  1.53  27.16
100
Thus a 95% confidence interval for  is 24.10 to 27.16
Determination of Sample Size
Again a question to be asked:
How large should the sample be?
Answer:
Depends on:
• How accurate you want the answer.
Accuracy is specified by:
• Specifying the magnitude of
the error bound
• Level of confidence
Error Bound:
B  za / 2

n
• If we have specified the level of confidence then the
value of za/2 will be known.
• If we have specified the magnitude of B, it will also
be known
Solving for n we get:
z 
z s *
n

2
B
B2
2
a/2
2
2
a/2
2
Summarizing:
The sample size that will estimate  with an Error
Bound B and level of confidence P = 1 – a is:
z 
z s *
n

2
2
B
B
2
a/2
2
2
a/2
2
where:
• B is the desired Error Bound
• za is the a/2 critical value for the standard normal
distribution
• s* is some preliminary estimate of .
Notes:
z 
z s *
n

2
B
B2
2
a/2
2
2
a/2
2
• n increases as B, the desired Error Bound, decreases
– Larger sample size required for higher level of accuracy
• n increases as the level of confidence, (1 – a), increases
– za increases as a/2 becomes closer to zero.
– Larger sample size required for higher level of
confidence
• n increases as the standard deviation, , of the population
increases.
– If the population is more variable then a larger sample
size required
Summary:
The sample size n depends on:
• Desired level of accuracy
• Desired level of confidence
• Variability of the population
Example
• Suppose that one is interested in estimating the average
number of grams of fat (m) in one kilogram of lean beef
hamburger :
This will be estimated by:
• randomly selecting one kilogram samples, then
• Measuring the fat content for each sample.
• Preliminary estimates of  and  indicate:
– that  and  are approximately 220 and 40 respectively.
• I want the study to estimate  with an error bound 5
and
• a level of confidence to be 95% (i.e. a = 0.05 and za =
z0.05 = 1.960)
Solution

1.960  40 
n
2
5
2
2
 245.9  246
Hence n = 246 one kilogram samples are
required to estimate  within B = 5 gms with a
95% level of confidence.
Statistical Inference
Making decisions regarding the
population base on a sample
Decision Types
• Estimation
– Deciding on the value of an unknown parameter
• Hypothesis Testing
– Deciding a statement regarding an unknown parameter
is true of false
• Prediction
– Deciding the future value of a random variable
• All decisions will be based on the values of
statistics
Estimation
• Definitions
– An estimator of an unknown parameter is a sample
statistic used for this purpose
– An estimate is the value of the estimator after the data
is collected
• The performance of an estimator is assessed by determining
its sampling distribution and measuring its closeness to the
parameter being estimated
Comments
• When you use a single statistic to estimate a
parameter it is called a point estimator
• The estimate is a single value
• The accuracy of this estimate cannot be
determined from this value
• A better way to estimate is with a confidence
interval.
• The width of this interval gives information on
its accuracy
Estimation by Confidence Intervals
• Definition
– An (100) P% confidence interval of an unknown
parameter is a pair of sample statistics (t1 and t2)
having the following properties:
1. P[t1 < t2] = 1. That is t1 is always smaller than t2.
2. P[the unknown parameter lies between t1 and t2] = P.
•
the statistics t1 and t2 are random variables
•
Property 2. states that the probability that the
unknown parameter is bounded by the two
statistics t1 and t2 is P.
Confidence Intervals
Summary
Confidence Interval for a Proportion
pˆ  za / 2 pˆ
 pˆ 
p1  p 

n
pˆ 1  pˆ 
n
za / 2  upper a / 2 critical point
of the standard normal distribtio n
B  za / 2 pˆ  za / 2
p 1  p 
n
 za / 2
 Error Bound
pˆ 1  pˆ 
n
Determination of Sample Size
The sample size that will estimate p with an Error Bound B
and level of confidence P = 1 – a is:
za2/ 2 p * 1  p *
n
2
B
where:
• B is the desired Error Bound
• za is the a/2 critical value for the standard normal
distribution
• p* is some preliminary estimate of p.
Confidence Intervals for the mean
of a Normal Population, 
x  za / 2 x
or x  za / 2
or x  za / 2

n
s
n
x  sample mean
za / 2  upper a / 2 critical point
of the standard normal distribtio n
s  sample standard deviation  
Determination of Sample Size
The sample size that will estimate  with an Error Bound B
and level of confidence P = 1 – a is:
z 
z s *
n

2
2
B
B
2
a/2
2
2
a/2
2
where:
• B is the desired Error Bound
• za is the a/2 critical value for the standard normal
distribution
• s* is some preliminary estimate of s.
Hypothesis Testing
An important area of statistical
inference
Definition
Hypothesis (H)
– Statement about the parameters of the population
• In hypothesis testing there are two hypotheses
of interest.
– The null hypothesis (H0)
– The alternative hypothesis (HA)
Either
– null hypothesis (H0) is true or
– the alternative hypothesis (HA) is true.
But not both
We say that are mutually exclusive and
exhaustive.
One has to make a decision
– to either to accept null hypothesis
(equivalent to rejecting HA)
or
– to reject null hypothesis (equivalent to
accepting HA)
There are two possible errors that can be
made.
1. Rejecting the null hypothesis when it is
true. (type I error)
2. accepting the null hypothesis when it is
false (type II error)
An analogy – a jury trial
The two possible decisions are
– Declare the accused innocent.
– Declare the accused guilty.
The null hypothesis (H0) – the accused is
innocent
The alternative hypothesis (HA) – the accused
is guilty
The two possible errors that can be made:
– Declaring an innocent person guilty.
(type I error)
– Declaring a guilty person innocent.
(type II error)
Note: in this case one type of error may be
considered more serious
Decision Table showing types of Error
H0 is True
H0 is False
Accept H0
Correct
Decision
Type II
Error
Reject H0
Type I
Error
Correct
Decision
To define a statistical Test we
1. Choose a statistic (called the test statistic)
2. Divide the range of possible values for the
test statistic into two parts
• The Acceptance Region
• The Critical Region
To perform a statistical Test we
1. Collect the data.
2. Compute the value of the test statistic.
3. Make the Decision:
• If the value of the test statistic is in
the Acceptance Region we decide to
accept H0 .
• If the value of the test statistic is in
the Critical Region we decide to
reject H0 .
Example
We are interested in determining if a coin is fair.
i.e.
H0 : p = probability of tossing a head = ½.
To test this we will toss the coin n = 10 times.
The test statistic is x = the number of heads.
This statistic will have a binomial distribution with
p = ½ and n = 10 if the null hypothesis is true.
Sampling distribution of x when H0 is true
0.3
0.25
0.2
0.15
0.1
0.05
0
0
1
2
3
4
5
6
7
8
9
10
Note
We would expect the test statistic x to be around 5 if
H0 : p = ½ is true.
Acceptance Region = {3, 4, 5, 6, 7}.
Critical Region = {0, 1, 2, 8, 9, 10}.
The reason for the choice of the Acceptance region:
Contains the values that we would expect for x if the
null hypothesis is true.
Definitions: For any statistical testing
procedure define
1. a = P[Rejecting the null hypothesis when it
is true] = P[ type I error]
. b = P[accepting the null hypothesis when it
is false] = P[ type II error]
In the last example
1. a = P[ type I error] = p(0) + p(1) + p(2) + p(8)
+ p(9) + p(10) = 0.109, where p(x) are
binomial probabilities with p = ½ and n = 10 .
. b = P[ type II error] = p(3) + p(4) + p(5) + p(6)
+ p(7), where p(x) are binomial probabilities
with p (not equal to ½) and n = 10. Note: these
will depend on the value of p.
Table: Probability of a Type II
error, b vs. p
p
0.1
0.2
0.3
0.4
0.6
0.7
0.8
0.9
b
0.070
0.322
0.616
0.820
0.820
0.616
0.322
0.070
Note: the magnitude of b
increases as p gets
closer to ½.
Comments:
1. You can control a = P[ type I error] and b = P[
type II error] by widening or narrowing the
acceptance region. .
2. Widening the acceptance region decreases a
= P[ type I error] but increases b = P[ type II
error].
3. Narrowing the acceptance region increases a
= P[ type I error] but decreases b = P[ type II
error].
Example – Widening the Acceptance
Region
1. Suppose the Acceptance Region includes
in addition to its previous values 2 and 8
then a = P[ type I error] = p(0) + p(1) +
p(9) + p(10) = 0.021, where again p(x)
are binomial probabilities with p = ½ and
n = 10 .
. b = P[ type II error] = p(2) + p(3) + p(4) +
p(5) + p(6) + p(7) + p(8). Tabled values
of are given on the next page.
Table: Probability of a Type II
error, b vs. p
p
0.1
0.2
0.3
0.4
0.6
0.7
0.8
0.9
b
0.264
0.624
0.851
0.952
0.952
0.851
0.624
0.264
Note: Compare these
values with the
previous definition
of the Acceptance
Region. They have
increased,
Example – Narrowing the Acceptance
Region
1. Suppose the original Acceptance Region
excludes the values 3 and 7. That is the
Acceptance Region is {4,5,6}. Then a = P[ type
I error] = p(0) + p(1) + p(2) + p(3) + p(7) +
p(8) +p(9) + p(10) = 0.344.
. b = P[ type II error] = p(4) + p(5) + p(6) .
Tabled values of are given on the next page.
Table: Probability of a Type II
error, b vs. p
p
0.1
0.2
0.3
0.4
0.6
0.7
0.8
0.9
b
0.013
0.120
0.340
0.563
0.563
0.340
0.120
0.013
Note: Compare these
values with the
otiginal definition of
the Acceptance
Region. They have
decreased,
Acceptance
Region
Acceptance
Region
Acceptance
Region
{2,3,4,5,6,7,8}.
{3,4,5,6,7}.
{4,5,6}.
a = 0.021
p
0.1
0.2
0.3
0.4
0.6
0.7
0.8
0.9
b
0.264
0.624
0.851
0.952
0.952
0.851
0.624
0.264
a = 0.109
p
0.1
0.2
0.3
0.4
0.6
0.7
0.8
0.9
b
0.070
0.322
0.616
0.820
0.820
0.616
0.322
0.070
a = 0.344
p
0.1
0.2
0.3
0.4
0.6
0.7
0.8
0.9
b
0.013
0.120
0.340
0.563
0.563
0.340
0.120
0.013
Hypothesis Testing
An important area of statistical
inference
Definition
Hypothesis (H)
– Statement about the parameters of the population
• In hypothesis testing there are two hypotheses
of interest.
– The null hypothesis (H0)
– The alternative hypothesis (HA)
Either
– null hypothesis (H0) is true or
– the alternative hypothesis (HA) is true.
But not both
We say that are mutually exclusive and
exhaustive.
Decision Table showing types of Error
H0 is True
H0 is False
Accept H0
Correct
Decision
Type II
Error
Reject H0
Type I
Error
Correct
Decision
The Approach in Statistical Testing is:
• Set up the Acceptance Region so that a is
close to some predetermine value (the usual
values are 0.05 or 0.01)
• The predetermine value of a (0.05 or 0.01)
is called the significance level of the test.
• The significance level of the test is a =
P[test makes a type I error]
Determining the Critical Region
1. The Critical Region should consist of values of
the test statistic that indicate that HA is true.
(hence H0 should be rejected).
2. The size of the Critical Region is determined so
that the probability of making a type I error, a,
is at some pre-determined level. (usually 0.05 or
0.01). This value is called the significance level
of the test.
Significance level = P[test makes type I error]
To find the Critical Region
1. Find the sampling distribution of the test statistic
when is H0 true.
2. Locate the Critical Region in the tails (either
left or right or both) of the sampling distribution
of the test statistic when is H0 true.
Whether you locate the critical region in the left
tail or right tail or both tails depends on which
values indicate HA is true.
The tails chosen = values indicating HA.
3. the size of the Critical Region is chosen so that
the area over the critical region and under the
sampling distribution of the test statistic when is
H0 true is the desired level of a =P[type I error]
Sampling distribution
of test statistic when H0
is true
Critical Region - Area = a
The z-test for Proportions
Testing the probability of success in a
binomial experiment
Situation
• A success-failure experiment has been
repeated n times
• The probability of success p is unknown. We
want to test
– H0: p = p0 (some specified value of p)
Against
– HA:
p  p0
The Data
• The success-failure experiment has been repeated
n times
• The number of successes x is observed.
x
pˆ   the poportion of successes
n
• Obviously if this proportion is close to p0 the Null
Hypothesis should be accepted otherwise the null
Hypothesis should be rejected.
The Test Statistic
• To decide to accept or reject the Null Hypothesis
(H0) we will use the test statistic
z
pˆ  p0
 pˆ

pˆ  p0
p0 1  p0 
n
• If H0 is true we should expect the test statistic z to
be close to zero.
• If H0 is true we should expect the test statistic z to
have a standard normal distribution.
• If HA is true we should expect the test statistic z to
be different from zero.
The sampling distribution of z when H0 is true:
The Standard Normal distribution
0
Reject H0
Accept H0
z
Reject H0
The Acceptance region:
a/2
a/2
Reject H0
 za / 2
0
za / 2
Accept H0
z
Reject H0
PAccept H 0 when true   P za / 2  z  za / 2   1  a
PReject H 0 when true   Pz   za / 2 or z  za / 2   a
• Acceptance Region
 za / 2  z  za / 2
– Accept H0 if:
• Critical Region
– Reject H0 if:
z   za / 2 or z  za / 2
• With this Choice
PType I Error   PReject H 0 when true 
 Pz   za / 2 or z  za / 2   a
Summary
To Test for a binomial probability p
H0: p = p0 (some specified value of p)
Against
HA : p  p0
we
1. Decide on a = P[Type I Error] = the
significance level of the test (usual choices
0.05 or 0.01)
2. Collect the data
3. Compute the test statistic
z
pˆ  p0
 pˆ

pˆ  p0
p0 1  p0 
n
4. Make the Decision
• Accept H0 if:  za / 2  z  za / 2
• Reject H0 if: z   za / 2 or z  za / 2
Example
• In the last provincial election the proportion
of the voters who voted for the Liberal party
was 0.08 (8 %)
• The party is interested in determining if that
percentage has changed
• A sample of n = 800 voters are surveyed
We want to test
– H0: p = 0.08 (8%)
Against
– HA: p  0.08 (8%)
Summary
1. Decide on a = P[Type I Error] = the
significance level of the test
Choose (a = 0.05)
2. Collect the data
• The number in the sample that support the
liberal party is x = 92
x 92
pˆ  
 0.115 (11.5%)
n 800
3. Compute the test statistic
pˆ  p0
pˆ  p0
z

 pˆ
p0 1  p0 
n
0.115  0.80

 3.649
0.801  0.80 
800
4. Make the Decision za / 2  z0.025  1.960
• Accept H0 if: 1.960  z  1.960
• Reject H0 if: z  1.960 or z  1.960
Since the test statistic is in the Critical
region we decide to Reject H0
Conclude that H0: p = 0.08 (8%) is false
There is a significant difference (a = 5%)
in the proportion of the voters supporting
the liberal party in this election than in the
last election
The two-tailed z-test for
Proportions
Testing the probability of success in a
binomial experiment
Situation
• A success-failure experiment has been
repeated n times
• The probability of success p is unknown. We
want to test
– H0: p = p0 (some specified value of p)
Against
– HA:
p  p0
The Test Statistic
• To decide to accept or reject the Null Hypothesis
(H0) we will use the test statistic
z
pˆ  p0
 pˆ

pˆ  p0
p0 1  p0 
n
• Acceptance Region
 za / 2  z  za / 2
– Accept H0 if:
• Critical Region
– Reject H0 if:
z   za / 2 or z  za / 2
• With this Choice
PType I Error   PReject H 0 when true 
 Pz   za / 2 or z  za / 2   a
The Acceptance region:
a/2
a/2
Reject H0
 za / 2
0
za / 2
Accept H0
z
Reject H0
PAccept H 0 when true   P za / 2  z  za / 2   1  a
PReject H 0 when true   Pz   za / 2 or z  za / 2   a
The one tailed z-test
• A success-failure experiment has been
repeated n times
• The probability of success p is unknown.
We want to test
– H0: p  p0 (some specified value of p)
Against
– HA: p  p0
• The alternative hypothesis is in this case
called a one-sided alternative
The Test Statistic
• To decide to accept or reject the Null Hypothesis
(H0) we will use the test statistic
z
pˆ  p0
 pˆ

pˆ  p0
p0 1  p0 
n
• If H0 is true we should expect the test statistic z to
be close to zero or negative
• If p = p0 we should expect the test statistic z to
have a standard normal distribution.
• If HA is true we should expect the test statistic z to
be a positive number.
The sampling distribution of z when p = p0 :
The Standard Normal distribution
0
Accept H0
z
Reject H0
The Acceptance and Critical region:
a
0
Accept H0
za
z
Reject H0
PAccept H 0 when true   Pz  za   1  a
PReject H 0 when true   P z  za   a
• Acceptance Region
– Accept H0 if:
z  za
• Critical Region
– Reject H0 if:
z  za
• The Critical Region is called one-tailed
• With this Choice
PType I Error   PReject H 0 when true 
 Pz  za   a
Example
• A new surgical procedure is developed for
correcting heart defects infants before the
age of one month.
• Previously the procedure was used on
infants that were older than one month and
the success rate was 91%
• A study is conducted to determine if the
success rate of the new procedure is greater
than 91% (n = 200)
We want to test
– H0: p  0.91 (91%)
Against
– HA: p  0.91 (91%)
p  the success rate of the new procedure
Summary
1. Decide on a = P[Type I Error] = the
significance level of the test
Choose (a = 0.05)
2. Collect the data
• The number of successful operations in the
sample of 200 cases is x = 187
x 187
pˆ  
 0.935 (93.5%)
n 200
3. Compute the test statistic
pˆ  p0
pˆ  p0
z

 pˆ
p0 1  p0 
n
0.935  0.91

 1.235
0.911  0.91
200
4. Make the Decision za  z0.05  1.645
• Accept H0 if: z  1.645
• Reject H0 if:
z  1.645
Since the test statistic is in the Acceptance
region we decide to Accept H0
Conclude that H0: p  0.91 (91%) is true
More precisely H0 can’t be rejected
There is a no significant (a = 5%) increase
in the success rate of the new procedure
over the older procedure
Comments
• When the decision is made to accept H0 is
made one should not conclude that we have
proven H0.
• This is because when setting up the test we
have not controlled b = P[type II error] =
P[accepting H0 when H0 is FALSE]
• Whenever H0 is accepted there is a
possibility that a type II error has been
made.
In the last example
The conclusion that there is a no significant
(a = 5%) increase in the success rate of the
new procedure over the older procedure
should be interpreted:
We have been unable to proof that the new
procedure is better than the old procedure
Some other comments:
When does one use a two-tailed test?
When does one use a one tailed test?
Answer: This depends on the alternative
hypothesis HA.
Critical Region = values that indicate HA
Thus if only the upper tail indicates HA, the
test is one tailed.
If both tails indicate HA, the test is two tailed.
Also:
The alternative hypothesis HA usually
corresponds to the research hypothesis (the
hypothesis that the researcher is trying to prove)
1. The new procedure is better
2. The drug is effective in reducing levels of
cholesterol.
3. There has a change in political opinion from
the time the survey was taken till the present
time (time of current survey).
The z-test for the Mean of a
Normal Population
We want to test, , denote the mean
of a normal population
Situation
• A sample of n observations are collected from
a Normal distribution
• The mean of the Normal distribution, , is
unknown. We want to test
– H0:  = 0 (some specified value of )
Against
– HA :    0
The Data
• Let x1, x2, x3 , … , xn denote a sample from a
normal population with mean  and standard
deviation .
• Let
n
x
x
i 1
n
i
 the sample mean
• we want to test if the mean, , is equal to some
given value 0.
• Obviously if the sample mean is close to 0 the
Null Hypothesis should be accepted otherwise the
null Hypothesis should be rejected.
The Test Statistic
• To decide to accept or reject the Null Hypothesis
(H0) we will use the test statistic
z
x  0
x

x  0

 n
x  0

x  0
 n
s
n
• If H0 is true we should expect the test statistic z to
be close to zero.
• If H0 is true we should expect the test statistic z to
have a standard normal distribution.
• If HA is true we should expect the test statistic z to
be different from zero.
The sampling distribution of z when H0 is true:
The Standard Normal distribution
0
Reject H0
Accept H0
z
Reject H0
The Acceptance region:
a/2
a/2
Reject H0
 za / 2
0
za / 2
Accept H0
z
Reject H0
PAccept H 0 when true   P za / 2  z  za / 2   1  a
PReject H 0 when true   Pz   za / 2 or z  za / 2   a
• Acceptance Region
 za / 2  z  za / 2
– Accept H0 if:
• Critical Region
– Reject H0 if:
z   za / 2 or z  za / 2
• With this Choice
PType I Error   PReject H 0 when true 
 Pz   za / 2 or z  za / 2   a
Summary
To Test for mean , of a normal population
H0:  = 0 (some specified value of )
Against
HA :    0
1. Decide on a = P[Type I Error] = the
significance level of the test (usual choices
0.05 or 0.01)
2. Collect the data
3. Compute the test statistic
z n
x  0

x  0
 n
s
4. Make the Decision
• Accept H0 if:  za / 2  z  za / 2
• Reject H0 if: z   za / 2 or z  za / 2
Example
A manufacturer Glucosamine capsules claims
that each capsule contains on the average:
• 500 mg of glucosamine
To test this claim n = 40 capsules were
selected and amount of glucosamine (X)
measured in each capsule.
Summary statistics:
x  496.3 and s  8.5
We want to test:
H 0 :   500
Manufacturers claim is correct
against
H A :   500
Manufacturers claim is not
correct
The Test Statistic
z
x  0
x

x  0

 n
x  0

x  0
 n
s
n
496.3  500
 40
8.5
 2.75
The Critical Region and
Acceptance Region
Using a = 0.05
za/2 = z0.025 = 1.960
We accept H0 if
-1.960 ≤ z ≤ 1.960
reject H0 if
z < -1.960 or z > 1.960
The Decision
Since
z= -2.75 < -1.960
We reject H0
Conclude: the manufacturers’s claim is
incorrect:
Hypothesis Testing
A review of the concepts
In hypotheses testing there are two hypotheses
1. The Null Hypothesis (H0)
2. The Alternative Hypothesis (HA)
• The alternative hypothesis is usually the
research hypothesis - the hypothesis that the
researcher is trying to prove.
• The null hypothesis is the hypothesis that the
research hypothesis is not true.
A statistical Test is defined by
1. Choosing a statistic (called the test
statistic)
2. Dividing the range of possible values for
the test statistic into two parts
• The Acceptance Region
• The Critical Region
To perform a statistical Test we
1. Collect the data.
2. Compute the value of the test statistic.
3. Make the Decision:
• If the value of the test statistic is in
the Acceptance Region we decide to
accept H0 .
• If the value of the test statistic is in
the Critical Region we decide to
reject H0 .
• You can compare a statistical test to a meter
Value of test
statistic
Acceptance
Region
Critical
Critical
Region
Region
Critical Region is the red zone of the meter
Value of test
statistic
Acceptance
Region
Critical
Critical
Region
Region
Accept H0
Acceptance
Region
Value of test
statistic
Critical
Critical
Region
Region
Reject H0
Acceptance
Region
Critical
Region
Sometimes the critical region is
located on one side. These tests are
called one tailed tests.
Whether you use a one tailed test or a two tailed
test depends on:
1. The hypotheses being tested (H0 and HA).
2. The test statistic.
If only large positive values of the test statistic
indicate HA then the critical region should be
located in the positive tail. (1 tailed test)
If only large negative values of the test statistic
indicate HA then the critical region should be
located in the negative tail. (1 tailed test)
If both large positive and large negative values
of the test statistic indicate HA then the critical
region should be located both the positive and
negative tail. (2 tailed test)
Usually 1 tailed tests are appropriate if HA is
one-sided.
Two tailed tests are appropriate if HA is two sided.
But not always
Once the test statistic is determined, to set up the
critical region we have to find the sampling
distribution of the test statistic when H0 is true
This describes the behaviour of the test
statistic when H0 is true
We then locate the critical region in the tails of
the sampling distribution of the test statistic when
H0 is true
a /2
a /2
The size of the critical region is chosen so that
the area over the critical region is a.
This ensures that the P[type I error] = P[rejecting
H0 when true] = a
a /2
a /2
To find P[type II error] = P[ accepting H0 when
false] = b, we need to find the sampling
distribution of the test statistic when H0 is false
sampling distribution
of the test statistic
when H0 is false
sampling distribution
of the test statistic
when H0 is true
b
a /2
a /2
The p-value approach to
Hypothesis Testing
In hypothesis testing we need
1. A test statistic
2. A Critical and Acceptance region
for the test statistic
The Critical Region is set up under the
sampling distribution of the test statistic.
Area = a (0.05 or 0.01) above the critical
region. The critical region may be one tailed or
two tailed
The Critical region:
a/2
a/2
Reject H0
 za / 2
0
za / 2
Accept H0
z
Reject H0
PAccept H 0 when true   P za / 2  z  za / 2   1  a
PReject H 0 when true   Pz   za / 2 or z  za / 2   a
In test is carried out by
1. Computing the value of the test
statistic
2. Making the decision
a. Reject if the value is in the Critical
region and
b. Accept if the value is in the
Acceptance region.
The value of the test statistic may be in the
Acceptance region but close to being in the
Critical region, or
The it may be in the Critical region but close to
being in the Acceptance region.
To measure this we compute the p-value.
Definition – Once the test statistic has been
computed form the data the p-value is defined
to be:
p-value = P[the test statistic is as or more
extreme than the observed value of
the test statistic]
more extreme means giving stronger evidence to
rejecting H0
Example – Suppose we are using the z –test for the
mean  of a normal population and a = 0.05.
Z0.025 = 1.960
Thus the critical region is to reject H0 if
Z < -1.960 or Z > 1.960 .
Suppose the z = 2.3, then we reject H0
p-value = P[the test statistic is as or more extreme than
the observed value of the test statistic]
= P [ z > 2.3] + P[z < -2.3]
= 0.0107 + 0.0107 = 0.0214
Graph
p - value
-2.3
2.3
If the value of z = 1.2, then we accept H0
p-value = P[the test statistic is as or more extreme than
the observed value of the test statistic]
= P [ z > 1.2] + P[z < -1.2]
= 0.1151 + 0.1151 = 0.2302
23.02% chance that the test statistic is as or more
extreme than 1.2. Fairly high, hence 1.2 is not very
extreme
Graph
p - value
-1.2
1.2
Properties of the p -value
1. If the p-value is small (<0.05 or 0.01) H0 should be
rejected.
2. The p-value measures the plausibility of H0.
3. If the test is two tailed the p-value should be two
tailed.
4. If the test is one tailed the p-value should be one
tailed.
5. It is customary to report p-values when reporting
the results. This gives the reader some idea of the
strength of the evidence for rejecting H0
Summary
• A common way to report statistical tests is to
compute the p-value.
• If the p-value is small ( < 0.05 or < 0.01) then
H0 is rejected.
• If the p-value is extremely small this gives a
strong indication that HA is true.
• If the p-value is marginally above the
threshold 0.05 then we cannot reject H0 but
there would be a suspicion that H0 is false.
Next topic: Student’s t - test