Uploaded by sijiam2

123A Mid1 Cheatsheet

advertisement
Regression residuals:
Minimize sum of squared regression residuals:
Ordinary Least Squares (OLS) estimates:
Fitted values and residuals:
Algebraic properties of OLS regression:
Measures of Variation:
1) Total sum pf squares (total car in Y):
2) Explained sum of squares (var explained by regression):
3) Residual sum of squares (car not exp by regression):
Decomposition of total variation:
Goodness-of-Fit measures:
R-squared interpretation: fraction of the total variation in Y explained by the regression
Lin-Lin: change by $𝛽1
Log-Lin: change by 𝛽1(%)
Log-Log: change by 𝛽1 %
Change scale of Y by constant k: multiply all coefficient estimate by k
Change X by a multiplicative constant k, leave Y alone: if multiply X by k, then the slope coefficient change by
multiplying 1/k, no impact on intercept
Homoskedasticity:
The value of Y must contain no info about variability of error term
Variance of the OLS estimators:
Estimating error variance:
Standard errors (how precisely the regression coefficients are estimated) for regression coefficients:
OLS est of multiple regression model:
1) Regression residuals:
2) Min sum of squared residuals:
MLR – Algebraic properties of OLS regression:
MLR – Decomposition of total variation:
1) R-squared:
2) Alternative expression for R-squared:
Download