Transcript Chapter 12
Week 11
November 10-14
Four Mini-Lectures
QMM 510
Fall 2014
Chapter Learning Objectives
Too
much?
LO12-1: Calculate and test a correlation coefficient for significance.
LO12-2: Interpret the slope and intercept of a regression equation.
LO12-3: Make a prediction for a given x value using a regression equation.
LO12-4: Fit a simple regression on an Excel scatter plot.
LO12-5: Calculate and interpret confidence intervals for regression coefficients.
LO12-6: Test hypotheses about the slope and intercept by using t tests.
LO12-7: Perform regression analysis with Excel or other software.
LO12-8: Interpret the standard error, R2, ANOVA table, and F test.
LO12-9: Distinguish between confidence and prediction intervals for Y.
LO12-10: Test residuals for violations of regression assumptions.
LO12-11: Identify unusual residuals and high-leverage observations.
12-2
Chapter 12
Chapter 12: Correlation and Regression
ML 11.1
Visual Displays
• Begin the analysis of bivariate data (i.e., two variables) with a scatter plot.
• A scatter plot:
- displays each observed data pair (xi, yi) as a dot on an X / Y grid.
- indicates visually the strength of the relationship between X and Y
Sample Scatter Plot
12-3
Chapter 12
Correlation Analysis
Chapter 12
Correlation Analysis
Strong Positive Correlation
Strong Negative Correlation
12-4
Weak Positive Correlation
No Correlation
Note: r is an estimate of the population
correlation coefficient r (rho).
Weak Negative Correlation
Nonlinear Relation
Chapter 12
Correlation Analysis
Steps in Testing if r = 0 (population correlation = 0)
•
Step 1: State the Hypotheses
H0: r = 0
H1: r ≠ 0
•
Step 2: Specify the Decision Rule
For degrees of freedom d.f. = n 2, look up the critical value ta in
Appendix D or Excel =T.INV.2T(α,df). for a 2-tailed test
•
Step 3: Calculate the Test Statistic
1 ≤ r ≤ +1
r = 0 indicates no
linear relationship
•
Step 4: Make the Decision
If using the t statistic method, reject H0 if t > ta or if the p-value a.
12-5
Chapter 12
Correlation Analysis
Alternative Method to Test for r = 0
•
Equivalently, you can calculate the critical value for the correlation
coefficient using
Critical values of r for various
sample sizes
•
This method gives a benchmark for the correlation coefficient.
•
However, there is no p-value and is inflexible if you change your
mind about a.
•
MegaStat uses this method, giving two-tail critical values for a =
0.05 and a = 0.01.
12-6
Chapter 12
Simple Regression
ML 11.2
What is Simple Regression?
•
•
•
Simple regression analyzes the relationship between two
variables.
It specifies one dependent (response) variable and one
independent (predictor) variable.
This hypothesized relationship (in this chapter) will be linear.
12-7
Chapter 12
Simple Regression
Interpreting an Estimated Regression Equation
12-8
Chapter 12
Simple Regression
Prediction Using Regression: Examples
12-9
Chapter 12
Simple Regression
Cause-and-Effect?
Can We Make Predictions?
12-10
Chapter 12
Regression Terminology
Model and Parameters
•
The assumed model for a linear relationship is
y = b0 + b1x + e .
•
The relationship holds for all pairs (xi, yi).
•
The error term e is not observable; it is assumed to be independently
normally distributed with mean of 0 and standard deviation s.
• The unknown parameters are:
b0
b1
Intercept
Slope
12-11
Chapter 12
Regression Terminology
Model and Parameters
•
The fitted model or regression model used to predict the expected value
of Y for a given value of X is
•
The fitted coefficients are
b0
Estimated intercept
b1
Estimated slope
12-12
Chapter 12
Regression Terminology
A more precise method is to let
Excel calculate the estimates. Enter
observations on the independent
variable x1, x2, . . ., xn and the
dependent variable y1, y2, . . ., yn
into separate columns, and let
Excel fit the regression equation.
Excel will choose the regression
coefficients so as to produce a
good fit.
12-13
Chapter 12
Regression Terminology
Scatter plot shows a sample of
miles per gallon and horsepower
for 15 vehicles.
•
Slope Interpretation: The slope of 0.0785 says that for each additional unit of
engine horsepower, the miles per gallon decreases by 0.0785 mile. This estimated
slope is a statistic because a different sample might yield a different estimate of the
slope.
•
Intercept Interpretation: The intercept value of 49.216 suggests that when the
engine has no horsepower, the fuel efficiency would be quite high. However, the
intercept has little meaning in this case, not only because zero horsepower makes no
logical sense, but also because extrapolating to x = 0 is beyond the range of the
observed data.
12-14
Chapter 12
Ordinary Least Squares (OLS) Formulas
OLS Method
•
The ordinary least squares method (OLS) estimates the slope
and intercept of the regression line so that the sum of squared
residuals is minimized.
•
The sum of the residuals = 0.
•
The sum of the squared residuals is SSE.
12-15
Chapter 12
Ordinary Least Squares (OLS) Formulas
Slope and Intercept
•
The OLS estimator for the slope is:
Excel function:
=SLOPE(YData, XData)
•
The OLS estimator for the intercept is:
Excel function:
=INTERCEPT(YData, XData)
These formulas
are built into
Excel.
12-16
Example: Achievement Test Scores
20 high school students’ achievement exam scores.
Note that verbal scores average higher than quant scores (slope
exceeds 1 and intercept shifts the line up almost 20 points).
12-17
Quant
Verbal
Obs
X
Y
1
520
398
2
329
505
3
225
183
4
424
332
5
650
737
6
491
578
7
384
344
8
311
367
9
236
298
10
344
600
11
541
643
12
324
328
13
515
556
14
528
527
15
380
504
16
629
695
17
228
133
18
454
478
19
514
413
20
677
742
Chapter 12
Ordinary Least Squares (OLS) Formulas
Chapter 12
Ordinary Least Squares (OLS) Formulas
Slope and Intercept
12-18
Chapter 12
Assessing Fit
Assessing Fit
•
We want to explain the total variation in Y around its mean (SST for total
sums of squares).
•
The regression sum of squares (SSR) is the explained variation in Y.
12-19
Chapter 12
Assessing Fit
Assessing Fit
•
The error sum of squares (SSE) is the unexplained variation in Y.
•
If the fit is good, SSE will be relatively small compared to SST.
•
A perfect fit is indicated by an SSE = 0.
•
The magnitude of SSE depends on n and on the units of measurement.
12-20
Chapter 12
Assessing Fit
Coefficient of Determination
•
R2 is a measure of relative fit based on a comparison of SSR
(explained variation) and SST (total variation).
0 R2 1
•
Often expressed as a percent, an R2 = 1 (i.e., 100%) indicates
perfect fit.
•
In simple regression, R2 = r2 where r2 is the squared correlation
coefficient).
12-21
Chapter 12
Assessing Fit
Example: Achievement Test Scores
Strong relationship
between quant score
and verbal score (68
percent of variation
explained)
R2 = SSR / SST = 387771 / 567053 = .6838
Regression Statistics
Multiple R
0.82694384
R Square
0.68383612
Adjusted R Square
0.66627146
Standard Error
99.8002618
Observations
20
SSR
ANOVA Table
Source
Regression
Residual
Total
df
SS
MS
1 387771.3 387771.3
18 179281.7 9960.092
19 567053
SST
12-22
F
38.9325
Excel shows the sums
needed to calculate R2.
Chapter 12
Tests for Significance
Standard Error of Regression
•
The standard error (se) is an overall measure of model fit.
Excel’s Data
Analysis >
Regression
calculates se.
•
If the fitted model’s predictions are perfect (SSE = 0), then se = 0. Thus,
a small se indicates a better fit.
•
Used to construct confidence intervals.
•
Magnitude of se depends on the units of measurement of Y and on data
magnitude.
12-23
Chapter 12
Tests for Significance
Confidence Intervals for Slope and Intercept
•
Standard error of the slope and intercept:
Excel’s Data Analysis >
Regression constructs
confidence intervals
for the slope and
intercept.
•
Confidence interval for the true slope and intercept:
12-24
Chapter 12
Tests for Significance
Hypothesis Tests
•
•
If b1 = 0, then the regression model collapses to a
constant b0 plus random error.
Excel ‘s Data
Analysis >
Regression performs
these tests.
The hypotheses to be tested are:
d.f. = n 2
Reject H0 if tcalc > ta/2
or if p-value α.
12-25
Example: Achievement Test Scores
20 high school students’
achievement exam
scores.
Chapter 12
Analysis of Variance: Overall Fit
Excel shows 95%
confidence intervals
and t test statistics
Intercept
Slope
Coefficients Standard Error
t Stat
P-value
19.5924736
75.25768037 0.260339 0.797557
1.03046307
0.165149128 6.239591 6.93E-06
Regression output
variables
coefficients
Intercept
19.5925
Slope
1.0305
12-26
std. error
75.2577
0.1651
t (df=18)
0.260
6.240
p-value
.7976
6.93E-06
Lower 95%
Upper 95%
-138.5180458
177.702993
0.683497623
1.37742851
confidence interval
95% lower 95% upper
-138.5180
177.7030
0.6835
1.3774
MegaStat is similar
but rounds off and
highlights p-values
to show significance
(light yellow .05,
bright yellow .01)
F Test for Overall Fit
• To test a regression for overall significance, we use an F test to
compare the explained (SSR) and unexplained (SSE) sums of
squares.
12-27
Chapter 12
Analysis of Variance: Overall Fit
Example: Achievement Test Scores
20 high school students’
achievement exam
scores.
ANOVA Table
Source
Regression
Residual
Total
ANOVA table
Source
Regression
Residual
Total
12-28
df
SS
MS
1 387771.2894 387771.3
18 179281.6606 9960.092
19
567052.95
SS
387,771.2894
179,281.6606
567,052.9500
df
1
18
19
F Significance F
38.9325
6.92538E-06
MS
387,771.2894
9,960.0923
F
38.93
p-value
6.93E-06
Excel shows the
ANOVA sums, the F
test statistic , and its
p-value.
MegaStat is similar,
but also highlights
p-values to indicate
significance (light
yellow .05, bright
yellow .01)
Chapter 12
Analysis of Variance: Overall Fit
How to Construct an Interval Estimate for Y
•
Confidence interval for the conditional mean of Y is shown below.
•
Prediction intervals are wider than confidence intervals for the mean
because individual Y values vary more than the mean of Y.
Chapter 12
Confidence and Prediction Intervals for Y
Excel does not do
these CIs!
12-29
11.3
Three Important Assumptions
1.
The errors are normally distributed.
2.
The errors have constant variance (i.e., they are homoscedastic).
3.
The errors are independent (i.e., they are nonautocorrelated).
Non-normal Errors
•
Non-normality of errors is a mild violation since the regression
parameter estimates b0 and b1 and their variances remain unbiased
and consistent.
•
Confidence intervals for the parameters may be untrustworthy
because the normality assumption is used to justify using Student’s
t distribution.
12-30
Chapter 12
Tests of Assumptions
Non-normal Errors
•
A large sample size would compensate.
•
Outliers could pose serious problems.
Normal Probability Plot
•
•
12-31
The normal probability plot tests the assumption
H0: Errors are normally distributed
H1: Errors are not normally distributed
If H0 is true, the
residual probability
plot should be linear,
as shown
in the example.
Chapter 12
Residual Tests
What to Do about Non-normality?
1.
Trim outliers only if they clearly are mistakes.
2.
Increase the sample size if possible.
3.
If data are totals, try a logarithmic transformation of both X and Y.
12-32
Chapter 12
Residual Tests
Heteroscedastic Errors (Nonconstant Variance)
•
The ideal condition is if the error magnitude is constant (i.e.,
errors are homoscedastic).
•
Heteroscedastic errors increase or decrease with X.
•
In the most common form of heteroscedasticity, the
variances of the estimators are likely to be understated.
•
This results in overstated t statistics and artificially narrow
confidence intervals.
Tests for Heteroscedasticity
•
12-33
Plot the residuals against X.
Ideally, there is no pattern in the
residuals moving from left to
right.
Chapter 12
Residual Tests
Tests for Heteroscedasticity
•
12-34
Chapter 12
Residual Tests
The “fan-out” pattern of increasing residual variance is the most common
pattern indicating heteroscedasticity.
What to Do about Heteroscedasticity?
•
Transform both X and Y, for example, by taking logs.
•
Although it can widen the confidence intervals for the coefficients,
heteroscedasticity does not bias the estimates.
Autocorrelated Errors
•
Autocorrelation is a pattern of non-independent errors.
•
In a first-order autocorrelation, et is correlated with et 1.
•
The estimated variances of the OLS estimators are biased, resulting in
confidence intervals that are too narrow, overstating the model’s fit.
12-35
Chapter 12
Residual Tests
Runs Test for Autocorrelation
•
In the runs test, count the number of the residuals’ sign reversals (i.e., how often
does the residual cross the zero centerline?).
•
If the pattern is random, the number of sign changes should be n/2.
•
Fewer than n/2 would suggest positive autocorrelation.
•
More than n/2 would suggest negative autocorrelation.
Durbin-Watson (DW) Test
•
Tests for autocorrelation under the hypotheses
H0: Errors are nonautocorrelated
H1: Errors are autocorrelated
•
The DW statistic will range from 0 to 4.
DW < 2 suggests positive autocorrelation
DW = 2 suggests no autocorrelation (ideal)
DW > 2 suggests negative autocorrelation
12-36
Chapter 12
Residual Tests
What to Do about Autocorrelation?
•
Transform both variables using the method of first differences in which
both variables are redefined as changes. Then we regress Y against X.
•
Although it can widen the confidence interval for the coefficients,
autocorrelation does not bias the estimates.
•
Don’t worry about it at this stage of your training. Just learn to
detect whether it exists.
12-37
Chapter 12
Residual Tests
Example: Excel’s Tests of Assumptions
Excel’s Data Analysis >
Regression does
residual plots and gives
the DW test statistic. Its
standardized residuals
are done in a strange
way, but usually they are
not misleading.
Warning: Excel offers
normal probability plots
for residuals, but they
are done incorrectly.
12-38
Chapter 12
Residual Tests
Chapter 12
Residual Tests
Example: MegaStat’s Tests of Assumptions
MegaStat will do all
three tests (if you
check the boxes). Its
runs plot (residuals
by observation) is a
visual test for
autocorrelation,
which Excel does not
offer.
12-39
Example: MegaStat’s Tests of Assumptions
near-linear plot - indicates normal errors
no pattern - suggests homoscedastic errors
12-40
no pattern - suggests homoscedastic errors
DW near 2 - suggests no autocorrelation
Chapter 12
Residual Tests
Standardized Residuals
• Use Excel, MINITAB, MegaStat or other software to compute
standardized residuals.
• If the absolute value of any standardized residual is at least 2,
then it is classified as unusual.
Leverage and Influence
• A high leverage statistic indicates the observation is far from the
mean of X.
• These observations are influential because they are at the “end
of the lever.”
• The leverage for observation i is denoted hi.
12-41
Chapter 12
Unusual Observations
Leverage
• A leverage that exceeds 4/n is unusual.
12-42
Chapter 12
Unusual Observations
Example: Achievement Test Scores
•
•
12-43
If the absolute value of any
standardized residual is at
least 2, then it is classified
as unusual.
Leverage that exceeds 4/n
indicates an influential X
value (far from mean of X).
Quant
Verbal
Obs
X
Y
Predicted Y
1
520
398
555.4
-157.4
0.070
-1.636
2
329
505
358.6
146.4
0.081
1.530
3
225
183
251.4
-68.4
0.171
-0.753
4
424
332
456.5
-124.5
0.050
-1.280
5
650
737
689.4
47.6
0.176
0.526
6
491
578
525.5
52.5
0.059
0.542
7
384
344
415.3
-71.3
0.057
-0.736
8
311
367
340.1
26.9
0.092
0.283
9
236
298
262.8
35.2
0.159
0.385
10
344
600
374.1
225.9
0.073
2.351
11
541
643
577.1
65.9
0.081
0.689
12
324
328
353.5
-25.5
0.084
-0.267
13
515
556
550.3
5.7
0.067
0.059
14
528
527
563.7
-36.7
0.074
-0.382
15
380
504
411.2
92.8
0.058
0.959
16
629
695
667.8
27.2
0.153
0.297
17
228
133
254.5
-121.5
0.168
-1.335
18
454
478
487.4
-9.4
0.051
-0.097
19
514
413
549.3
-136.3
0.067
-1.413
20
677
742
717.2
24.8
0.210
0.279
Residual Leverage Std Residual
Chapter 12
Unusual Observations
Outliers
Outliers may be caused by
• an error in recording data.
• impossible data (can be omitted).
• an observation that has been
influenced by an unspecified
“lurking” variable that should have
been controlled but wasn’t.
12-44
12B-44
Chapter 12
Other Regression Problems
To fix the problem
• delete the observation(s) if you are
sure they are actually wrong.
• formulate a multiple regression
model that includes the lurking
variable.
Model Misspecification
•
If a relevant predictor has been omitted, then the model
is misspecified.
•
For example, Height depends on Gender as well as Age.
•
Use multiple regression instead of bivariate regression.
Ill-Conditioned Data
• Well-conditioned data values are of the same general order
of magnitude.
• Ill-conditioned data have unusually large or small data
values and can cause loss of regression accuracy or
awkward estimates.
12-45
Chapter 12
Other Regression Problems
Ill-Conditioned Data
•
Avoid mixing magnitudes by adjusting the magnitude of your
data before running the regression.
•
For example, Revenue= 139,405,377 mixed with ROI = .037.
Spurious Correlation
•
In a spurious correlation two variables appear related because of
the way they are defined.
•
This problem is called the size effect or problem of totals.
•
Expressing variables as per capita or per cent may be helpful.
12-46
Chapter 12
Other Regression Problems
Model Form and Variable Transforms
• Sometimes a nonlinear model is a better fit than a linear
model. Excel offers other model forms for simple regression
(one X and one Y)
• Variables may be transformed (e.g., logarithmic or exponential
functions) in order to provide a better fit.
• Log transformations reduce heteroscedasticity.
• Nonlinear models may be difficult to interpret.
12-47
Chapter 12
Other Regression Problems
Assignments
ML 11.4
• Connect C-8 (covers chapter 12)
•
•
•
•
You get three attempts
Feedback is given if requested
Printable if you wish
Deadline is midnight each Monday
• Project P-3 (data, tasks, questions)
•
•
•
•
0-48
Review instructions
Look at the data
Your task is to write a nice, readable report (not a spreadsheet)
Length is up to you
Projects: General Instructions
General Instructions
For each team project, submit a short (5-10 page) report (using Microsoft Word
or equivalent) that answers the questions posed. Strive for effective writing (see
textbook Appendix I). Creativity and initiative will be rewarded. Avoid careless
spelling and grammar. Paste graphs and computer tables or output into your
written report (it may be easier to format tables in Excel and then use Paste
Special > Picture to avoid weird formatting and permit sizing within Word).
Allocate tasks among team members as you see fit, but all should review and
proofread the report (submit only one report).
0-49
Project P-3
You will be assigned team members and a dependent variable (see Moodle) from the 2010
state database Big Dataset 09 - US States. The team may change the assigned dependent
variable (instructor assigned one just to give you a quick start). Delegate tasks and
collaborate as seems appropriate, based on your various skills. Submit one report. Data:
Choose an interesting dependent variable (non-binary) from the 2010 state database posted
on Moodle. Analysis: (a). Propose a reasonable model of the form Y = f(X1, X2, ... , Xk) using
not more than 12 predictors. (b) Use regression to investigate the hypothesized relationship.
(c) Try deleting poor predictors until you feel that you have a parsimonious model, based on
the t-values, p-values, standard error, and R2adj. (d) For the preferred model only, obtain a list
of residuals and request residual tests and VIFs. (e) List the states with high leverage and/or
unusual residuals. (f) Make a histogram and/or probability plot of the residuals. Are the
residuals normal? (g) For the predictors that were retained, analyze the correlation matrix
and/or VIFs. Is multicollinearity a problem? If so, what could be done? (h) If you had more
time, what might you do?
Watch the instructor’s video walkthrough
using Assault as an example (posted on
Moodle)
0-50
Project P-3 (preview, data, tasks)
• Example using the 2005 state database:
• 170 variables on n = 50 states
• Choose one variable as Y ( the response).
• Goal: to explain why Y varies from state to state.
• Start choosing X1, X2, … , Xk (the predictors).
• Copy Y and X1, X2, … , Xk to a new spreadsheet.
• Study the definitions for each variable (e.g., Burglary is the
burglary rate per 100,000 population.
12-51
Project P-3 (preview, data, tasks)
• Why multiple predictors?
• One predictor usually is an incorrect specification.
• Fit can usually be improved.
• How many predictors: Evans’ Rule (k n/10)
• Up to one predictor per 10 observations
• For example, n = 50 suggests k = 5 predictors.
• Evans’ Rule is conservative. It’s OK to start with more (you
will end up with fewer after deleting weak predictors).
12-52
Project P-3 (preview, data, tasks)
• Work with partners? Absolutely – it will be more fun.
• Post questions for peers or instructor on Moodle.
• Get started. But don’t run a bunch of regressions until you
have studied Chapter 13.
• It’s a good idea to have the instructor look over your list of
intended Y and X1, X2, … , Xk in order to avoid unnecessary rework if there are obvious problems.
• Look at all the categories of variables – don’t just grab the first
one you see (there are 170 variables). Or just use the one your
instructor assigned.
12-53