Self-BLAME - NewCL.org
Download
Report
Transcript Self-BLAME - NewCL.org
Logistic Regression
Part I - Introduction
Logistic Regression
• Regression where the response variable is
dichotomous (not continuous)
• Examples
– effect of concentration of drug on whether
symptoms go away
– effect of age on whether or not a patient
survived treatment
– effect of negative cognitions about SELF,
WORLD, or Self-BLAME on whether a
participant has PTSD
Simple Linear Regression
• Relationship between continuous
response variable and continuous
explanatory variable
• Example
– Effect of concentration of drug on reaction
time
– Effect of age of patient on number of years of
post-operation survival
Simple Linear Regression
• RT (ms) = β0 + β1 x concentration (mg)
• β0 is value of RT when concentration is 0
• β1 is change in RT caused by a change in
concentration of 1mg.
• E.g. RT = 400 + 50 x concentration
Logistic Regression
• What do we do when we have a response
variable which is not continuous, but is
dichotomous
Probability of Disease
Concentration
Odds of Disease
Concentration
Log(Odds) of Disease
Concentration
Odds
• Odds are simply the ratio of the
proportions for the two possible
outcomes.
• If p is the proportion for one outcome,
then 1- p is the proportion for the
second outcome.
Odds (Example)
• At concentration level 16 we observe 75
participants out of 100 showing no disease
(healthy)
• If p is the probability of healthy is then p = 0.75.
• Then 1 – p is the probability of not healthy, and
equals 0.25
• Odds of showing healthy over not healthy
given concentration level 16
• p / (1 – p) = 0.75/0.25 = 3
• Means that it is 3 times more likely that person is
healthy at concentration level 16
Logarithms
• Logarithms are a way of expressing
numbers as powers of a base
• Example
– 102 = 100
– 10 is called the “base”
– The power, 2 in this case, is called the
“exponent”
• Therefore 102 = 100 means that log10100
=2
Log Odds
• Odds of being healthy after 16mg of drug
is 3
• Log odds is log(3) = 1.1
• Lets say that odds of being healthy after
2mg of drug is 0.25
• Means that it is four times more likely to
not be healthy after 2mg of drug
• Log odds is log(0.25) = -1.39
Logistic Regression
• With Log-odds we can now look at the linear
relationship between dichotomous response and
continuous explanatory
pˆ
0 1 X
log
1 pˆ
Where, for example, p is the probability of being
healthy at different levels of drug concentration,
X
Example: Simple Logistic
Regression
• Look at the effect of drug concentration on
probability of NOT having disease (i.e.
being healthy)
• Use SPSS to do the regression (we’ll all do
this soon)
• Get
pˆ nodisease
2.92 0.106 Concentration
log
1 pˆ nodisease
Looks Like
pˆ nodisease
2.92 0.106 Concentration
log
1 pˆ nodisease
• Interpreting parameters (b0 and b1) in logistic
regression is a little tricky
• An increase of 1mg of concentration increases
the log(odds) of being healthy by 0.106
• An increase of 1mg of concentration increases
the odds of being healthy by
eb1 e.106 1.111
• Increasing concentration by 1mg increases
odds of being healthy by a factor of 1.11
Slope Parameter
• Parameter β1 in general:
– if positive then increasing X increases the
odds of p
– if negative then increasing X decreases the
odds of p
– the larger (in magnitude) the larger the effect
of X on p
• Like simple linear regression, can test
whether or not β1 is significantly different
from 0.
Let’s break to do simple Logistic
Regression
• Open XYZ.sav in SPSS
• Fit logistic regression with
– PTSD (Y/N) as response variable
– Self-BLAME as explanatory variable
•
•
•
•
Is the effect of Self-BLAME significant?
Get parameter estimates
Write equation of model
What is the odds of having PTSD given SelfBLAME score of 3?
• Use the interpretation of the regression
coefficient to work out odds given Self-BLAME
of 4.
Logistic Regression
Part II – Multiple Logistic
Regression
Multiple Linear Regression
• Simple Linear Regression extended out to
more than one explanatory variable
• Example
– Effect of both concentration and age on
reaction time
– Effect of age, number of previous operations,
time in anaesthesia, cholesterol level, etc. on
number of years of post-operation survival
Multiple Linear Regression
RT (ms) = β0 + β1 x concentration (mg) + β2 x
age + β3 x gender (0=male,1=female)
β0 is value of RT when concentration is 0.
β1 is change in RT caused by a change in
concentration of 1mg.
β2 is change in RT caused by a change in age of 1
year.
β3 is change in RT caused by going from male to
female in gender.
Multiple Logistic Regression
• Look at the effect of drug concentration, age and
gender on probability of NOT having disease
pˆ
0 1 X 1 2 X 2 3 X 3
log
1 pˆ
Where p is the probability of not having the
disease, X1 is the concentration of drug (mg),
X2 is age (years), and X3 is gender (0 for males,
1 for females)
pˆ nodisease
2.92 0.106 Concentration 0.0532 Age 0.001 Gender
log
1 pˆ nodisease
• Again, use SPSS to fit logistic model
• Increasing concentration increases odds of not
having the disease (again, being healthy)
• Increasing age decreases odds of being healthy
• “Increasing” gender (from male to female)
increases odds of being healthy
• In particular, increasing age decreases the odds
of being healthy by a factor of 0.95
• M to F increases odds by factor of 1.001
Was it worth adding the factors?
• When we add parameters we make our
model more complicated.
• We really want this addition to be “worth it”
• In other words, adding age and gender
should improve our explanation of disease
• But what constitutes an improvement
Was it worth adding the factors?
• Quality (badness) of model fit is given by
-2logL
• If we fit want to see if it was worth adding
parameters we can compare the quality of the
fit of the simple and the more complex model
• Quality of model fit follows a chi-square (χ2)
distribution with degrees-of-freedom (df) equal to
the number of parameters in the model
• The difference between quality of fit also
follows a χ2 distribution with df equal to the
difference in the number of parameters between
the two models
Was it worth adding these factors?
• Simple logistic regression model has
overall χ2 of 45.7
• This multiple logistic regression model with
2 extra parameters has χ2 of 40.02
• Test whether χ2 = 45.7 - 40.02 = 5.68 is a
significant improvement
• Critical χ2 for 2 df is 5.99
• Our χ2 is smaller and so NO, not worth it
BUT…
• It doesn’t look like gender is having much of an
effect
• Check SPSS output and see that Wald χ2 for
Gender is 0.527, which has p = .47
• Perhaps it wasn’t worth adding both parameters,
but it will be worth just adding Age
• Age has Wald-χ2 = 4.33, p = .03
• When we only add Age, change in χ2 = 5.5 and
we test against χ2 with df of 1, which has p = .02
Logistic Regression Model Building
• What if we have a whole host of possible
explanatory variables
• We want to build a model which predicts whether
a person will have a disease given a set of
explanatory variables
• SAME as multiple linear regression
–
–
–
–
–
Forward selection
Backward elimination
Stepwise
All subsets
Hierarchical
How to know if a model is good
• All about having a model which does a good job of
appropriately classifying participants as having disease
or not
• In particular, model predicts how many people have
disease and how many people don’t have the disease
• The model can be
– Correct in two ways
• Correctly categorise a person who has a disease as having a
disease
• Correctly say no disease when no disease
– Incorrect in two ways
• Incorrectly categorise a person who has a disease as not having a
disease
• Incorrectly say no disease when disease
Accuracy of model
• Proportion of correct classifications
– Number of correct disease participants plus
number of correct no disease participants
divided by number of participants in total
nCD nC ND
nCD nC ND nICD nIC ND
Sensitivity of model
• Proportion of ‘successes’ correctly
identified
– Number of correct no disease participants
divided by total number of no disease
participants
nC ND
nC ND nIC ND
Specificity of model
• Proportion of ‘failures’ correctly identified
– Number of correct disease participants
divided by total number of disease
participants
nC D
nC D nIC D
Now…a real example
• Startup, Makgekgenene and Webster
(2007) looked at whether or not the
subscales of the Posttraumatic Cognitions
Inventory (PTCI) are good predictors of
Posttraumatic Stress Disorder (PTSD)
• Subscales are
– Negative Cognitions About SELF
– Negative Cognitions about the WORLD
– Self-BLAME
Descriptive Results
• PTSD participants showed higher scores
than non-PTSD in all three subscales
variables
Multiple Logistic Regression
• Response variable:
– whether or not the participant has PTSD
• Explanatory variables:
– Negative Cognitions About SELF
– Negative Cognitions about the WORLD
– Self-BLAME
Let’s do the Logistic Regression
• Open XYZ.sav in SPSS
• Run the appropriate regression
• What are the parameter estimates for our
three explanatory variables?
• Which of these are significant (at α = .05)?
• What are the odds ratios for those that are
significant?
• Anything unusual?
Self-BLAME
• Self-BLAME has a negative odds ratio.
• This means that increasing self-blame
decreases the chance of having PTSD
• This is surprising, especially since
participants with PTSD showed higher
Self-BLAME scores
• What’s going on?
Self-BLAME and SELF scales
• Startup et al. (2007) explain this by stating
that Self-BLAME is made up of both
behavioural and characterological
questions
• SELF, however, may also tap into
characterological aspects of self-blame
• Behavioural self-blame can be considered
adaptive. It may help avoid PTSD
• Characterological self-blame, however,
may be detrimental, and lead to PTSD
Suppressor Effect
• The relationship between SELF and PTSD is
strong, and accounts for the negative
relationship. This includes the effect of
characterological self-blame.
• The variation in PTSD that is left for Self-BLAME
to account for is the positive aspect of the
relationship between the Self-BLAME scores
and PTSD.
• The negative aspect of Self-BLAME scores has
been suppressed (already accounted for by
SELF). The positive aspect of Self-BLAME can
now come out.
Homework (haha)
• Evaluate the model by looking at
– Accuracy of model’s predictions
– Sensitivity of model’s predictions
– Specificity of model’s predictions