Supplemental Material of Lecture 7 and 8: 1. Sampling Distributions of OLS Estimators πππ(ππ₯)= π2 πππ(π₯) a is a constant. πππ(π₯ + π¦) = var(x) + var(y) + 2cov(x, y) πππ(π₯π )= π 2 π₯ π2 πππ( ππ )= π2 1 1 π πππ(π₯Μ ) = πππ (π ∑π π=1 π₯π ) = π 2 πππ(∑π=1 π₯π ) 2 If π₯π is independent with other observations, πππ(∑π π=1 π₯π ) = ππ 1 1 2 2 Therefore πππ(π₯Μ ) = π2 πππ(∑π π=1 π₯π ) = π 2 ∗ ππ = π /π χ2-distribution: In probability theory and statistics, the chi-squared distribution (also chisquare or χ2-distribution) with k degrees of freedom is the distribution of a sum of the squares of k independent standard normal random variables. The chi-squared distribution is a special case of the gamma distribution and is one of the most widely used probability distributions in inferential statistics, notably in hypothesis testing and in construction of confidence intervals. T-distribution--In probability and statistics, Student's t-distribution (or simply the tdistribution) is any member of a family of continuous probability distributions that arise when estimating the mean of a normally-distributed population in situations where the sample size is small and the population's standard deviation is unknown. The variance of π½Μπ οΌ πππ(π½Μπ |x) = π½Μπ − π½π π2 ππππ (1 − π π2 ) ~π(0,1) √πππ(π½Μπ |x) The estimated variance of π½Μπ Μ Μπ |x) = πππ(π½ πΜ 2 ππππ (1 − π π2 ) π½Μπ − π½π Μ Μπ |x) √πππ(π½ ~π‘π−π−1 Since we are testing H0οΌπ½π , it is only natural to look at our unbiased estimator of π½π , π½Μπ . The point estimate , π½Μπ will never exactly be zero, whether or not the H0 is true. The question is Μ π very far from zero provide evidence against how far , π½Μπ is from zero. A sample value of π· H0. However, we must recognize that there is sampling error in our estimate π½Μπ , so the size of π½Μπ must be weighed against its sample error. Since the standard error of π½Μπ is an estimate of Μ π is away standard deviation of π½Μπ , T-π·Μπ measures how many estimated standard deviations π· from zero. 2. Testing hypothesis: t test Define a rejection rule so that, if it is true, H0 is rejected only with a small probability (= significance level, e.g. 5%) 2.1 Testing against one-sided alternatives (greater than zero) For a right-tailed test, c is chosen to make the area in right tail of t distribution equal 5%. In other words, c is the 95 th percentile in the t distribution with n-k-1 degrees of freedom. A pattern in the critical value: as significance level falls, the critical value increases, so that we require a larger and larger value of T-π½Μπ in order to reject the H0.So, if H0 is rejected at, say 5% level, then is automatically rejected at 10% level. 2.2 Testing against one-sided alternatives (less than zero) For a left-tailed test, c is chosen to make the area in left tail of t distribution equal 5%. In other words, c is the 5 the percentile in the t distribution with n-k-1 degrees of freedom. 2.3 Testing against two-sided alternatives Using the regression estimates to help us formulate the null or alternative hypotheses is not allowed, because classical statistical inference presumes that we state the null and alternative about the population before looking at the data. For a two-tailed test, c is chosen to make the area in each tail of t distribution equal 2.5%. In other words, c is the 97.5 th percentile in the t distribution with n-k-1 degrees of freedom. 2.4 Normal approximation If a regression coefficient is different from zero in a two-sided test, the corresponding variable is said to be “statistically significant” If the number of degrees of freedom is large enough so that the normal approximation applies, the following rules of thumb apply: 2.5 Testing more general hypotheses dlogY/dlogX=(dY/Y)/ (dX/X) 2.6 P-value The smallest significance level at which the null hypothesis is still rejected, is called the pvalue of the hypothesis test. The p-value is the probability of observing a t statistic as extreme as we did if the null hypothesis is true. A small p-value is evidence against the null hypothesis because one would reject the null hypothesis even at small significance levels A large p-value is evidence in favor of the null hypothesis. P-values are more informative than tests at fixed significance levels 2.7 Confidence interval π {−π‘πΌ,π−π−1 < 2 π½Μπ − π½π < π‘πΌ,π−π−1 } = 1 − πΌ 2 π π(π½Μπ ) π {π½Μπ − π‘πΌ,π−π−1 π π(π½Μπ ) < π½π < π½Μπ + π‘πΌ,π−π−1 π π(π½Μπ )} = 1 − πΌ 2 2 π {π½π ∉ (π½Μπ − π‘πΌ,π−π−1 π π(π½Μπ ), π½Μπ + π‘πΌ,π−π−1 π π(π½Μπ ))} = πΌ 2 2 The probability that π½π does not fall into the confidence interval is πΌ. If random sample were obtained over and over again, with π½π and π½Μ π computed each time, then the (unknown) population value π½π would be in the interval (π½π , π½Μ π ) for 95% of the samples. Unfortunately, for the single sample that we use to contruct CI, we do not know whether π½π is actually contained in the interval. We hope we have obtained sample that is one of the 95% of all samples where the interval estimate contains π½π , but we have no guarantee. If the null hypothesis is H0: π½π = πΌπ , then H0 is rejected against at the 5% level if, and only if, πΌπ is not in the 95% confidence interval. 2.8 Examples in academic paper Gaurav, Liang, Mobarak and Song (2020) 2.9 Testing linear combination 3 F test 3.1 The definition of F distribution 3.2 The R-Squared Form of the F Statistic 2 ) SSR π = SST(1 − π π2 ) SSR π’π = SST(1 − π π’π 2 2 (π π’π (π π’π − π π2 )/π − π π2 )/π πΉ= = 2 )/(π − π − 1) 2 )/ππ (1 − π π’π (1 − π π’π π’π 3.3 P values for F test p − value = P(ΟΜ > F) ΟΜ denotes the F random variable with (q, n-k-1) degrees of freedom. F is the actual value of the test statistic. Reference: Khanna, G., Liang, W., Mobarak, A. M., & Song, R. (2019). The productivity consequences of pollution-induced migration in China. Working Paper, Yale University. Wooldridge, J. M. (2016). Introductory econometrics: A modern approach. Nelson Education. Wikipedia