• Multivariate F-Tests
• Multicolinearity
• The EVILS of Stepwise
Regression
• Intercept Dummies
• Interaction Effects
– Interaction Dummies
– Slope Dummies
Spring 2007 Lecture 9 Slide #1
F-tests are used to test for the statistical significance of the overall model fit. Normally, that’s redundant given that you already have the t-stats for the individual b’s.
F
ESS /( K 1)
RSS /( n K )
This amounts to a ratio of the explained variance by the residual variance, correcting for the number of observations and parameters. The F value is compared to the F-distribution, just like a t-distribution, to obtain a p-value.
F-tests become useful when we need to assess nested models.
Spring 2007 Lecture 9 Slide #2
The question is whether a more complex model adds to the explanatory power over a simpler model. E.g., using the Guns data, does including a set of gender/age (pm1024) and race (pb1064) variables improve the prediction of the violent crime rate (vio) variable after we’ve included average income (avginc) and population density (density)?
To find out, we calculate an F-statistic for the model improvement:
F H n K
( RSS { K H } RSS { K })/ H
( RSS { K })/( n K )
Where the complex model has K parameters, and the simpler model has K-H parameters.
•RSS{K-H} is the residual sum of squares for the simpler model
•RSS{K} is the residual sum of squares for the more complex model
Spring 2007 Lecture 9 Slide #3
Simpler Model: vio = b
0
+ b
1
(avginc) + b
2
(density)
More Complex Model: vio = b
0
+ b
1
(avginc) + b
2
(density) + b
3
(pm1024) + b
4
(pb1064)
Given the models, K = 5, K - H = 3, and n = 1173. Calculating the RSS’s involves running the two models, obtaining the RSS from each. For these models, RSS
K
= 60,820,479.3 and RSS
K-H
= 68,295,254.4. So:
F H n K
(68295254.4
60820497.3) /2
60820479.3/(1173 5)
3737378.5
52072.3
71.8
Given that df
1
= H (2) and df
2
= n-K (1168), the p-value of the model improvement shown in Appendix Table 5C (p. 762) is <0.01.
Spring 2007 Lecture 9 Slide #4
• Test for the additional explanatory power of the age/gender and race varables ( pm1024 and pb1064 ) when modeling the murder and robbery rates (mur and rob). Use the simple model from the prior lecture exercise ( avginc and density ) as your base of comparison.
• Discuss the theoretical meaning of your results.
Spring 2007 Lecture 9 Slide #5
• If any X i is a linear combination of other
X’s in the model, b i cannot be estimated
– Remember that the partial regression coefficient strips both the X’s and Y of the overlapping covariation: e
X i
| X j
X i
X i
X i
b
0
b
1
X j
– If an X is perfectly predicted by the other X’s, then: e
X i
| X j
0
Spring 2007 Lecture 9 Slide #6
• We rarely find perfect multicolinearity in practice, but high multicolinearity results in loss of statistical resolution:
– Large standard errors
– Low t-stats, high p-values
• Erodes hypothesis tests
– Enormous sensitivity to small changes in:
• Data
• Model specification
Spring 2007 Lecture 9 Slide #7
• Use of VIF measures:
– Variance Inflation Factor
• Degree to which other coefficients’ variance is increased due to the inclusion of the specified variable
• 1/VIF is equivalent to:
(AKA “Tolerance”)
1 R e
2
Xi | X j
• Acceptable tolerance? Partly a function of n-size:
• If n < 50, tolerance should exceed 0.7
• If n < 300, tolerance should exceed 0.5
• If n < 600, tolerance should exceed 0.3
• If n < 1000, tolerance should exceed 0.1
• Calculating Tolerance in Stata (using VIF)
Spring 2007 Lecture 9 Slide #8
• Drop one or other of co-linear variables
• Caution: May result in model mis-specification
• Use theory as a guide
• Add new data
• Special samples may maximize independent variation
– E.g.: Elite samples may disentangle income, education
• Data Scaling
• Highly correlated variables may represent multiple indicators of an underlying scale
• Data reduction through factor analysis
Spring 2007 Lecture 9 Slide #9
• Run a VIF on the “more complex” model from the nested
F-test exercise (from lecture slide #5)
• Now add the variable pw1064 (percent of the population white, ages 10-64). Run VIF again.
• Using the pm1024 and pb1064 variables
– F-Test will show significant improvement (over the reduced model from earlier)
– But not all individual coefficients were statistically significant
• What gives?
– Correlate pb1064, pw1064 and pm1029
– Strategy?
• Drop either pw1064 or pb1064 (use theory to decide which)
• Re-run the model
Spring 2007 Lecture 9 Slide #10
• Stata permits a number of mechanical “search strategies”
– Stepwise regression (forward, backward, upside down)
• How they work: use of sequential F-tests
• These methods pose serious problems:
– If X’s are strongly related, one or the other will tend to be excluded
– Susceptible to inclusion of spuriously related variables
– Obliterates the meaning of classical statistical tests
• Hypothesis tests depend on prior hypothesis identification
• Example: The infamous Swedish “EMF study”
• Hundreds of diseases regressed on proximity to EMF sources
– about 5% found to be “statistically significant”
• Enormous policy implications -- but method was buried in report
Spring 2007 Lecture 9 Slide #11
• Dummy variables allow for tests of the differences in overall value of the Y for different nominal groups in the data (akin to a difference of means)
• Coding: 0 and 1 values (e.g., men versus women)
Y
X
2,0
X
2,1
Spring 2007
X
1
Lecture 9 Slide #12
Source | SS df MS Number of obs = 1173
-------------+-----------------------------F( 2, 1170) = 983.14
Model | 21362735.3 2 10681367.7 Prob > F = 0.0000
Residual | 12711578.7 1170 10864.5972 R-squared = 0.6269
-------------+-----------------------------Adj R-squared = 0.6263
Total | 34074314 1172 29073.6468 Root MSE = 104.23
-----------------------------------------------------------------------------rob | Coef. Std. Err. t P>|t| [95% Conf. Interval]
-------------+---------------------------------------------------------------density | 96.56459 2.2606 42.72 0.000 92.12931 100.9999
shall | -50.08407 7.141642 -7.01 0.000 -64.09593 -36.07221
_cons | 139.9945 3.635576 38.51 0.000 132.8616 147.1275
------------------------------------------------------------------------------
Robbery rates are systematically lower in “Shall Issue” states
(is this necessarily causal??)
Spring 2007 Lecture 9 Slide #13
• Implies a comparison (the omitted group)
– Be clear about the “comparison category”
• Multinomial Dummies
– When categories exceed 2
• Importance of specifying the base category
• Examples of Category Variables
– Experimental treatment groups
– Race and ethnicity
– Region of residence
– Type of education
– Religious affiliation
– “Seasonality”
• Adds to modeling flexibility
Spring 2007 Lecture 9 Slide #14
• Interactions occur when the effect of one X is dependent on the value of another
• Modeling interactions:
– Use Dummy variables (requires categories)
– Use multiplicative interaction effect
• Multiply an interval scale times a dummy (also known as a
“slope dummy”)
• Example: the effect of Population Density
(density) on Robbery Rate (rob) may be affected by whether the respondent in in a Shall Issue state or not
– Re-code the interaction; run it.
Spring 2007 Lecture 9 Slide #15
Source | SS df MS Number of obs = 1173
-------------+-----------------------------F( 3, 1169) = 706.02
Model | 21956237.5 3 7318745.84 Prob > F = 0.0000
Residual | 12118076.5 1169 10366.1903 R-squared = 0.6444
-------------+-----------------------------Adj R-squared = 0.6435
Total | 34074314 1172 29073.6468 Root MSE = 101.81
-----------------------------------------------------------------------------rob | Coef. Std. Err. t P>|t| [95% Conf. Interval]
-------------+---------------------------------------------------------------density | 96.13354 2.208874 43.52 0.000 91.79974 100.4673
shall | -103.8476 9.957383 -10.43 0.000 -123.3839 -84.31126
i | 648.0681 85.64837 7.57 0.000 480.0264 816.1098
_cons | 140.1835 3.551295 39.47 0.000 133.2159 147.1512
------------------------------------------------------------------------------
Spring 2007 Lecture 9 Slide #16
3000
2000
1000
0
0
8000
7000
6000
5000
4000
9000
Non-Shall Issue Shall Issue
2 4 6
Population Density
8 10 12
Spring 2007 Lecture 9 Slide #17
• Assumptions of OLS Revisited
• Visual Diagnostic Techniques
• Non-linearity
• Non-normality and Heteroscedasticity
• Outliers and Case Statistics
• Aren’t we having fun now?
Spring 2007 Lecture 9 Slide #18