Binary Logistic Regression with SPSS

advertisement
Binary Logistic Regression
with SPSS
Karl L. Wuensch
Dept of Psychology
East Carolina University
Download the Instructional
Document
• http://core.ecu.edu/psyc/wuenschk/SPSS/
SPSS-MV.htm .
• Click on Binary Logistic Regression .
• Save to desktop.
• Open the document.
When to Use Binary Logistic Regression
• The criterion variable is dichotomous.
• Predictor variables may be categorical or
continuous.
• If predictors are all continuous and nicely
distributed, may use discriminant function
analysis.
• If predictors are all categorical, may use
logit analysis.
Wuensch & Poteat, 1998
• Cats being used as research subjects.
• Stereotaxic surgery.
• Subjects pretend they are on university
research committee.
• Complaint filed by animal rights group.
• Vote to stop or continue the research.
Purpose of the Research
•
•
•
•
•
Cosmetic
Theory Testing
Meat Production
Veterinary
Medical
Predictor Variables
•
•
•
•
Gender
Ethical Idealism (9-point Likert)
Ethical Relativism (9-point Likert)
Purpose of the Research
Model 1: Decision = Gender
• Decision 0 = stop, 1 = continue
• Gender 0 = female, 1 = male
• Model is ….. logit =
 Yˆ 
  a  bX
ln ODDS   ln 
 1  Yˆ 


• Yˆ is the predicted probability of the event
which is coded with 1 (continue the research)
rather than with 0 (stop the research).
Iterative Maximum Likelihood
Procedure
• SPSS starts with arbitrary regression
coefficents.
• Tinkers with the regression coefficients to
find those which best reduce error.
• Converges on final model.
SPSS
• Bring the data into SPSS
• http://core.ecu.edu/psyc/wuenschk/SPSS/
Logistic.sav
• Analyze, Regression, Binary Logistic
• Decision  Dependent
• Gender  Covariate(s), OK
Look at the Output
Case Processing Summary
Unweighted Cases
Selected Cases
a
N
Included in Analysis
Mis sing Cases
Total
Unselected Cases
Total
a. If weight is in effect, see classification table for the total
number of cases.
• We have 315 cases.
Percent
315
100.0
0
.0
315
100.0
0
.0
315
100.0
Block 0 Model, Odds
• Look at Variables in the Equation.
• The model contains only the intercept
(constant, B0), a function of the marginal
distribution of the decisions.
Variables in the Equation
B
Step 0
Constant
-.379
S.E.
.115
Wald
10.919
 Yˆ
ln ODDS   ln 
 1  Yˆ

df
Sig.
1

   . 379


.001
Exp(B)
.684
Exponentiate Both Sides
• Exponentiate both sides of the equation:
• e-.379 = .684 = Exp(B0) = odds of deciding to
continue the research.
Yˆ
1  Yˆ
 Exp (  . 379 )  . 684 
128
187
• 128 voted to continue the research, 187 to stop
it.
Probabilities
•
•
•
•
•
Randomly select one participant.
P(votes continue) = 128/315 = 40.6%
P(votes stop) = 187/315 = 59.4%
Odds = 40.6/59.4 = .684
Repeatedly sample one participant and
guess how e will vote.
Humans vs. Goldfish
• Humans Match Probabilities
–
–
(suppose p = .7, q = .3)
.7(.7) + .3(.3) = .49 + .09 = .58
• Goldfish Maximize Probabilities
–
.7(1) = .70
• The goldfish win!
SPSS Model 0 vs. Goldfish
• Look at the Classification Table for Block 0.
Classification Table
a,b
Predicted
decision
Step 0
Observed
decision
stop
Percentage
Correct
continue
stop
187
0
100.0
continue
128
0
.0
Overall Percentage
59.4
a. Constant is included in the model.
b. The cut value is .500
• SPSS Predicts “STOP” for every participant.
• SPSS is as smart as a Goldfish here.
Block 1 Model
• Gender has now been added to the model.
• Model Summary: -2 Log Likelihood = how
poorly model fits the data.
Model Summary
Step
1
-2 Log
likelihood
399.913a
Cox & Snell
R Square
.078
Nagelkerke
R Square
.106
a. Estimation terminated at iteration number 3 because
parameter estimates changed by less than .001.
Block 1 Model
• For intercept only, -2LL = 425.666.
• Add gender and -2LL = 399.913.
• Omnibus Tests: Drop in -2LL = 25.653 =
Model 2.
• df = 1, p < .001.
Omnibus Tests of Model Coefficients
Chi-square
Step 1
df
Sig.
Step
25.653
1
.000
Block
25.653
1
.000
Model
25.653
1
.000
Variables in the Equation
• ln(odds) = -.847 + 1.217Gender
ODDS  e
a  b Gender
Variables in the Equation
B
Step
a
1
S.E.
Wald
df
Sig.
Exp(B)
gender
1.217
.245
24.757
1
.000
3.376
Constant
-.847
.154
30.152
1
.000
.429
a. Variable(s) entered on s tep 1: gender.
Odds, Women
ODDS
 e
 . 847  1 . 217 ( 0 )
 e
 . 847
 0 . 429
• A woman is only .429 as likely to decide to
continue the research as she is to decide
to stop it.
Odds, Men
ODDS  e
 . 847  1 . 217 ( 1)
e
. 37
 1 . 448
• A man is 1.448 times more likely to vote to
continue the research than to stop the research.
Odds Ratio
male _ odds
female _ odds

1 . 448
 3 . 376  e
1 . 217
. 429
• 1.217 was the B (slope) for Gender, 3.376 is the
Exp(B), that is, the exponentiated slope, the
odds ratio.
• Men are 3.376 times more likely to vote to
continue the research than are women.
Convert Odds to Probabilities
• For our women,
Yˆ 
ODDS
1  ODDS

0 . 429
 0 . 30
1 . 429
• For our men,
Yˆ 
ODDS
1  ODDS

1 . 448
2 . 448
 0 . 59
Classification
• Decision Rule: If Prob (event)  Cutoff,
then predict event will take place.
• By default, SPSS uses .5 as Cutoff.
• For every man, Prob(continue) = .59,
predict he will vote to continue.
• For every woman Prob(continue) = .30,
predict she will vote to stop it.
Overall Success Rate
• Look at the Classification Table
Classification Table
a
Predicted
decision
Step 1
Observed
decision
stop
stop
continue
continue
140
47
74.9
60
68
53.1
Overall Percentage
66.0
a. The cut value is .500
140  68
315
Percentage
Correct

208
315
• SPSS beat the Goldfish!
 66 %
Sensitivity
• P (correct prediction | event did occur)
• P (predict Continue | subject voted to Continue)
• Of all those who voted to continue the research,
for how many did we correctly predict that.
68
68  60

68
128
 53 %
Specificity
• P (correct prediction | event did not occur)
• P (predict Stop | subject voted to Stop)
• Of all those who voted to stop the research, for
how many did we correctly predict that.
140
140  47

140
187
 75 %
False Positive Rate
• P (incorrect prediction | predicted occurrence)
• P (subject voted to Stop | we predicted Continue)
• Of all those for whom we predicted a vote to Continue
the research, how often were we wrong.
47
47  68

47
115
 41 %
False Negative Rate
• P (incorrect prediction | predicted nonoccurrence)
• P (subject voted to Continue | we predicted Stop)
• Of all those for whom we predicted a vote to Stop the
research, how often were we wrong.
60
140  60

60
200
 30 %
Pearson 2
• Analyze, Descriptive Statistics, Crosstabs
• Gender  Rows; Decision  Columns
Crosstabs Statistics
• Statistics, Chi-Square, Continue
Crosstabs Cells
• Cells, Observed Counts, Row
Percentages
Crosstabs Output
• Continue, OK
• 59% & 30% match logistic’s predictions.
gender * decision Crosstabulation
decision
stop
gender
Female
Count
% within gender
Male
Count
% within gender
Total
Count
% within gender
continue
Total
140
60
200
70.0%
30.0%
100.0%
47
68
115
40.9%
59.1%
100.0%
187
128
315
59.4%
40.6%
100.0%
Crosstabs Output
• Likelihood Ratio 2 = 25.653, as with
logistic.
Chi-Square Tests
Value
Asymp. Sig.
(2-sided)
df
Pearson Chi-Square
25.685b
1
.000
Likelihood Ratio
25.653
1
.000
N of Valid Cases
315
a. Computed only for a 2x2 table
b. 0 cells (.0%) have expected count less than 5. The
minimum expected count is 46.73.
Model 2: Decision =
Idealism, Relativism, Gender
• Analyze, Regression, Binary Logistic
• Decision  Dependent
• Gender, Idealism, Relatvsm
Covariate(s)
• Click Options and check “HosmerLemeshow goodness of fit” and “CI for
exp(B) 95%.”
• Continue, OK.
Comparing Nested Models
• With only intercept and gender,
-2LL = 399.913.
• Adding idealism and relativism dropped
-2LL to 346.503, a drop of 53.41.
• 2(2) = 399.913 – 346.503 = 53.41, p = ?
Model Summary
Step
1
-2 Log
likelihood
346.503a
Cox & Snell
R Square
.222
Nagelkerke
R Square
.300
a. Estimation terminated at iteration number 4 because
parameter estimates changed by less than .001.
Obtain p
• Transform, Compute
• Target Variable = p
• Numeric Expression =
1 - CDF.CHISQ(53.41,2)
p=?
• OK
• Data Editor, Variable View
• Set Decimal Points to 5 for p
p < .0001
• Data Editor, Data View
• p = .00000
• Adding the ethical ideology variables
significantly improved the model.
Hosmer-Lemeshow
• Hø: predictions made by the model fit
perfectly with observed group
memberships
• Cases are arranged in order by their
predicted probability on the criterion.
• Then divided into ten bins with
approximately equal n.
• This gives ten rows in the table.
For each bin and each event, we have
number of observed cases and expected
number predicted from the model.
Contingency Table for Hosmer and Lemeshow Test
decision = s top
Observed
Step
1
decision = continue
Expected
Observed
Expected
Total
1
29
29.331
3
2.669
32
2
30
27.673
2
4.327
32
3
28
25.669
4
6.331
32
4
20
23.265
12
8.735
32
5
22
20.693
10
11.307
32
6
15
18.058
17
13.942
32
7
15
15.830
17
16.170
32
8
10
12.920
22
19.080
32
9
12
9.319
20
22.681
32
10
6
4.241
21
22.759
27
• Note expected freqs decline in first
column, rise in second.
• The nonsignificant chi-square is indicative
of good fit of data with linear model.
Hosmer and Lemeshow Test
Step
1
Chi-square
8.810
df
Sig.
8
.359
Hosmer-Lemeshow
• There are problems with this procedure.
• Not even Hosmer and Lemeshow
recommend it these days.
• Even with good fit the test may be
significant if sample sizes are large
• Even with poor fit the test may not be
significant if sample sizes are small.
Linearity of the Logit
• We have assumed that the log odds are
related to the predictors in a linear fashion.
• Use the Box-Tidwell test to evaluate this
assumption.
• For each continuous predictor, compute
the natural log.
• Include in the model interactions between
each predictor and its natural log.
Box-Tidwell
• If an interaction is significant, there is a
problem.
• For the troublesome predictor, try
including the square of that predictor.
• That is, add a polynomial component to
the model.
• See T-Test versus Binary Logistic
Regression
Variables in the Equation
B
S.E.
gender
1.147
idealism
1.130 1.921
Wald
.269 18.129
.346
df
Sig.
Exp(B)
1
.000 3.148
1
.556 3.097
relatvsm
1.656 2.637
.394
1
.530 5.240
idealism by
Step 1a
-.652
.690
.893
1
.345
.521
idealism_LN
relatvsm by
-.479
.949
.254
1
.614
.620
relatvsm_LN
Constant
-5.015 5.877
.728
1
.393
.007
a. Variable(s) entered on step 1: gender, idealism, relatvsm, idealism * idealism_LN
, relatvsm * relatvsm_LN .
No Problem Here.
Model 3: Decision =
Idealism, Relativism, Gender, Purpose
• Need 4 dummy variables to code the five
purposes.
• Consider the Medical group a reference
group.
• Dummy variables are: Cosmetic, Theory,
Meat, Veterin.
• 0 = not in this group, 1 = in this group.
Add the Dummy Variables
• Analyze, Regression, Binary Logistic
• Add to the Covariates: Cosmetic, Theory,
Meat, Veterin.
• OK
Block 0
• Look at “Variables not in the Equation.”
• “Score” is how much -2LL would drop if a
single variable were added to the model
with intercept only.
Variables not in the Equation
Score
Step
0
Variables
Overall Statistics
df
Sig.
gender
25.685
1
.000
idealism
47.679
1
.000
relatvsm
7.239
1
.007
cosmetic
.003
1
.955
theory
2.933
1
.087
meat
.556
1
.456
veterin
.013
1
.909
77.665
7
.000
Effect of Adding Purpose
• Our previous model had -2LL = 346.503.
• Adding Purpose dropped -2LL to 338.060.
Model Summary
Step
1
-2 Log
likelihood
338.060a
Cox & Snell
R Square
.243
Nagelkerke
R Square
.327
a. Estimation terminated at iteration number 5 because
parameter estimates changed by less than .001.
• 2(4) = 8.443, p = .0766.
• But I make planned comparisons (with medical
reference group) anyhow!
Classification Table
• YOU calculate the sensitivity, specificity,
false positive rate, and false negative rate.
Classification Table
a
Predicted
decision
Step 1
Observed
decision
stop
stop
continue
Overall Percentage
a. The cut value is .500
continue
Percentage
Correct
152
35
81.3
54
74
57.8
71.7
Answer Key
•
•
•
•
Sensitivity = 74/128 = 58%
Specificity = 152/187 = 81%
False Positive Rate = 35/109 = 32%
False Negative Rate = 54/206 = 26%
Wald Chi-Square
• A conservative test of the unique
contribution of each predictor.
• Presented in Variables in the Equation.
• Alternative: drop one predictor from the
model, observe the increase in -2LL, test
via 2.
Variables in the Equation
95.0% C.I.for EXP(B)
B
Step
a
1
Wald
df
Sig.
Exp(B)
Lower
Upper
gender
1.255
20.586
1
.000
3.508
2.040
6.033
idealism
-.701
37.891
1
.000
.496
.397
.620
relatvs m
.326
6.634
1
.010
1.386
1.081
1.777
cos metic
-.709
2.850
1
.091
.492
.216
1.121
theory
-1.160
7.346
1
.007
.314
.136
.725
meat
-.866
4.164
1
.041
.421
.183
.966
veterin
-.542
1.751
1
.186
.581
.260
1.298
Constant
2.279
4.867
1
.027
9.766
a. Variable(s) entered on s tep 1: gender, idealism, relatvs m, cosmetic, theory, meat, veterin.
Odds Ratios – Exp(B)
• Odds of approval more than cut in half (.496) for
each one point increase in Idealism.
• Odds of approval multiplied by 1.39 for each one
point increase in Relativism.
• Odds of approval if purpose is Theory Testing
are only .314 what they are for Medical
Research.
• Odds of approval if purpose is Agricultural
Research are only .421 what they are for
Medical research
Inverted Odds Ratios
• Some folks have problems with odds
ratios less than 1.
• Just invert the odds ratio.
• For example, 1/.421 = 2.38.
• That is, respondents were more than two
times more likely to approve the medical
research than the research designed to
feed to poor in the third world.
Classification Decision Rule
• Consider a screening test for Cancer.
• Which is the more serious error
– False Positive – test says you have cancer,
but you do not
– False Negative – test says you do not have
cancer but you do
• Want to reduce the False Negative rate?
Classification Decision Rule
• Analyze, Regression, Binary Logistic
• Options
• Classification Cutoff = .4, Continue, OK
Effect of Lowering Cutoff
• YOU calculate the Sensitivity, Specificity,
False Positive Rate, and False Negative
Rate for the model with the cutoff at .4.
• Fill in the table on page 15 of the handout.
Answer Key
Value When Cutoff =
.5
.4
Sensitivity
58%
75%
Specificity
81%
72%
False Positive Rate
32%
36%
False Negative Rate
26%
19%
Overall % Correct
72%
73%
SAS Rules
• See, on page 16 of the handout, how easy
SAS makes it to see the effect of changing
the cutoff.
• SAS classification tables remove bias
(using a jackknifed classification
procedure), SPSS does not have this
feature.
Presenting the Results
• See the handout.
Interaction Terms
• Center continuous variables
• Compute the interaction terms or
• Let Logistic compute them.
Deliberation and Physical
Attractiveness in a Mock Trial
• Subjects are mock jurors in a criminal trial.
• For half the defendant is plain, for the
other half physically attractive.
• Half recommend a verdict with no
deliberation, half deliberate first.
Get the Data
• Bring Logistic2x2x2.sav into SPSS.
• Each row is one cell in 2x2x2 contingency
table.
• Could do a logit analysis, but will do
logistic regression instead.
• Tell SPSS to weight cases by Freq. Data,
Weight Cases:
• Dependent = Guilty.
• Covariates = Delib, Plain.
• In left pane highlight Delib and Plain.
• Then click >a*b> to create the interaction
term.
• Under Options, ask for the HosmerLemeshow test and confidence intervals
on the odds ratios.
Significant Interaction
• The interaction is large and significant
(odds ratio of .030), so we shall ignore the
main effects.
Variables in the Equation
95.0% C.I.for EXP(B)
Wald
Step
a
1
df
Sig.
Exp(B)
Lower
Upper
Delib
3.697
1
.054
.338
.112
1.021
Plain
4.204
1
.040
3.134
1.052
9.339
Delib by Plain
8.075
1
.004
.030
.003
.338
.037
1
.847
1.077
Constant
a. Variable(s) entered on step 1: Delib, Plain, Delib * Plain .
• Use Crosstabs to test the conditional
effects of Plain at each level of Delib.
• Split file by Delib.
•
•
•
•
Analyze, Crosstabs.
Rows = Plain, Columns = Guilty.
Statistics, Chi-square, Continue.
Cells, Observed Counts and Column
Percentages.
• Continue, OK.
Rows = Plain, Columns = Guilty
• For those who did deliberate, the odds of a
guilty verdict are 1/29 when the defendant
was plain and 8/22 when she was
attractive, yielding a conditional odds ratio
of 0.09483 .
Plain * Guilty Crosstabulation
a
Guilty
No
Plain
Attrractive
Count
% within Plain
Plain
Count
% within Plain
Total
Count
% within Plain
a. Delib = Yes
Yes
Total
22
8
30
73.3%
26.7%
100.0%
29
1
30
96.7%
3.3%
100.0%
51
9
60
85.0%
15.0%
100.0%
• For those who did not deliberate, the odds
of a guilty verdict are 27/8 when the
defendant was plain and 14/13 when she
was attractive, yielding a conditional odds
ratio of 3.1339.
Plain * Guilty Crosstabulation
a
Guilty
No
Plain
Attrractive
Count
% within Plain
Plain
Count
% within Plain
Total
Count
% within Plain
a. Delib = No
Yes
Total
13
14
27
48.1%
51.9%
100.0%
8
27
35
22.9%
77.1%
100.0%
21
41
62
33.9%
66.1%
100.0%
Interaction Odds Ratio
• The interaction odds ratio is simply the ratio of
these conditional odds ratios – that is,
.09483/3.1339 = 0.030.
• Among those who did not deliberate, the plain
defendant was found guilty significantly more
often than the attractive defendant, 2(1, N = 62)
= 4.353, p = .037.
• Among those who did deliberate, the attractive
defendant was found guilty significantly more
often than the plain defendant, 2(1, N = 60) =
6.405, p = .011.
Interaction Between Continuous
and Dichotomous Predictor
Interaction Falls Short of
Significance
Standardizing Predictors
• Most helpful with continuous predictors.
• Especially when want to compare the
relative contributions of predictors in the
model.
• Also useful when the predictor is
measured in units that are not intrinsically
meaningful.
Predicting Retention in ECU’s
Engineering Program
Practice Your New Skills
• Try the exercises in the handout.
Download