Given name:____________________ Family name:___________________

advertisement
Given name:____________________
Student #:______________________
Family name:___________________
Section #:______________________
BUEC 333 MIDTERM
Multiple Choice (2 points each)
1) If the covariance between two random variables X and Y is zero then
a) X and Y are not necessarily independent
b) knowing the value of X provides no information about the value of Y
c) E(X) = E(Y) = 0
d) a and b are true
e) none of the above
2) A negative covariance between X and Y means that whenever we obtain a value of X that is greater than
the mean of X:
a) we will have a greater-than-50% chance of obtaining a corresponding value of Y which is
greater than the mean of Y
b) we will have a less-than-50% chance of obtaining a corresponding value of Y which is
smaller than the mean of Y
c) we will obtain a corresponding value of Y which is greater than the mean of Y
d) we will obtain a corresponding value of Y which is smaller than the mean of Y
e) none of the above
3) In the regression specification, Yi = β0 + β1 X i + ε i , which of the following is a justification for
including epsilon?
a) it accounts for potential non-linearity in the functional form
b) it captures the influence of some of the omitted explanatory variables
c) it incorporates measurement error in the betas
d) it reflects deterministic variation in outcomes
e) all of the above
4) In the Capital Asset Pricing Model (CAPM):
a) β measures the sensitivity of the expected return of a portfolio to systematic risk
b) β measures the sensitivity of the expected return of a portfolio to specific risk
c) β is greater than one
d) α is less than zero
e) R2 is meaningless
5) Which of the following is an assumption of the CLRM?
a) the model is correctly specified
b) the independent variables are exogenous
c) the errors have mean zero
d) the errors have constant variance
e) all of the above
1 6) Suppose that in the simple linear regression model Yi = β0 + β1Xi + β2Zi + εi on 100 observations, you
calculate that R2= 0.5, the sample covariance of X and Y is 10, and the sample variance of X is 15. Then
the least squares estimator of β1 is:
a) not calculable using the information given
b) 1/3
c) 1 / 3
d) 2/3
e) none of the above
7) The law of large numbers says that:
a) the sample mean is a biased estimator of the population mean in small samples
b) the sampling distribution of the sample mean approaches a normal distribution as the
sample size approaches infinity
c) the behaviour of large populations is well approximated by the average
d) the sample mean is a consistent estimator of the population mean in large samples
e) none of the above
8) Suppose you have the following information about the pdf of a random variable X, which takes one of
4 possible values:
Value of X pdf 1 0.1 2 0.2 3 0.3 4 Which of the following is/are true?
a) Pr(X = 4) = 0.4
b) E(X) = 2.5
c) Pr(X = 4) = 0.2
d) all of the above
e) none of the above
9) For a model to be correctly specified under the CLRM, it must:
a) be non-linear in the coefficients
c) include all irrelevant independent variables and their associated transformations
c) have a multiplicative error term
d) all of the above
e) none of the above
10) Which of the following is not a linear regression model:
a) Yi = α + β X i + γX i2 + ε i
b) Yi = β 0 + β1 X i
β2
+ εi
c) log(Yi ) = β 0 + β1 log( X i ) + ε i
d) Yi = β 0 + β1 log( X i ) + ε i
e) none of the above
2 11) Suppose you have a random sample of 100 observations from a normal distribution with mean 10 and
variance 5. The sample mean ( x ) is 9 and the sample variance is 6. The sampling distribution of x has:
a) mean 9 and variance 0.06
b) mean 9 and variance 0.05
c) mean 10 and variance 0.06
d) mean 10 and variance 0.05
e) none of the above
12) To be useful for hypothesis testing, a test statistic must:
a) be computable using population data
b) have a known sampling distribution when the alternative hypothesis is true
c) have a known sampling distribution when the null hypothesis is false
d) a and b only
e) none of the above
13) If two random variables X and Y are independent:
a) their joint distribution equals the product of their marginal distributions
b) the conditional distribution of X given Y equals the conditional distribution of X
c) their covariance is zero
d) a and c
e) a, b, and c
14) The OLS estimator of the sampling variance of the slope coefficient in the regression model
with one independent variable:
a) will be smaller when there is less variation in ei
b) will be smaller when there are fewer observations
c) will be smaller when there is less variation in X
d) will be smaller when there are more independent variables
e) none of the above
15) If q is an unbiased estimator of Q, then:
a) Q is the mean of the sampling distribution of q
b) q is the mean of the sampling distribution of Q
c) Var[q] = Var[Q] / n where n = the sample size
d) q = Q
e) a and c
3 Short Answer #1 (10 points – show your work!)
Consider the simple univariate regression model,
Yi = β0 + β1 X i + ε i .
Demonstrate that the sample regression line passes through the sample mean of both X and Y.
We estimate the linear regression model Yi = β 0 + β1 X i + ε i as
Y = βˆ + βˆ X + e
i
0
1
i
i
We also know that
Yˆ = βˆ + βˆ X
i
0
1
i
and that
βˆ0 = Y − βˆ1 X
We can evaluate this second expression when X i = X to prove the statement above,
Yˆ = βˆ + βˆ X
i
0
1
Yˆi = Y − βˆ1 X + βˆ1 X = Y
So, by construction, the estimated regression line always passes through
the sample means when using OLS.
(Alternatively, you could start with the expression for Yi, considers it sum, and proceed to evaluate its
mean value).
4 Page intentionally left blank. Use this space for rough work or the continuation of an answer.
5 Short Answer #2 (20 points – show your work!)
For a homework assignment on sampling you are asked to program a computer to do the following:
i) Randomly draw 25 values from a standard normal distribution.
ii) Multiply each these values by 5 and add 10.
iii) Take the average of these 25 values and call it A1.
iv) Repeat this procedure to obtain 500 such averages and call them A1 through A500.
a) What is your best guess as to the value of A1? Explain your answer.
b) What is the value of the variance associated with A1? Explain your answer.
c) If you were to compute the average of the 500 A values, what should it be approximately equal to?
d) If you were to compute the sampling variance of these 500 A values, what should it be approximately
equal to?
a) A standard normal random variable is one for which the expected value is zero and the variance is
equal to one, so we have Z ~ N(0,1) where Z is one of he draws in i). But in ii), we are applying the
following transformation: X = 5 + 10*Z. We know from our formula sheet that
E ( a + bX + cY ) = a + bE ( X ) + cE (Y ).
It stands to reason then that our best guess (or expected value) of A1 = 10 + 5*E(Z) = 10.
b) Likewise, we know from the formula sheet that
Var(a + bY ) = b 2Var(Y )
It stands to reason then that the value of the variance associated with A1 = 52*Var(Z) = 25.
c) We know that
⎛ σ 2 ⎞
X ~ N ⎜ µ , ⎟
n ⎠
⎝
That is, the sample mean should have an expected value equal to the mean in the population of our
random variable. In this case, this is nothing more than the value reported in a) above. So, in this case,
the average of the averages should be approximately equal to 10.
d) From part c), we know that the sample mean has a sampling variance equal to the variance of the
underlying random variable divided by our sample size. We have calculated the numerator above as 25
and in applying n = 25, we find that the sampling variance should be equal to 25/25 or 1.
6 Page intentionally left blank. Use this space for rough work or the continuation of an answer.
7 Short Answer #3 (20 points – show your work!)
Suppose we have a random sample with 10 observations. These 10 observations (X1, X2, …, X10) are
drawn from a population distribution with mean, µ, and variance, σ2. We know that
x − bar = X =
X 1 + X 2 + ... + X 10 10 X i
=∑
10
i =1 10
E(X ) = µ
Var ( X ) =
σ2
10
X + X3 + X5 + X7 + X9
x − tilde = X%= 1
4
a) Is x-bar a random variable? Explain your answer.
b) Is x-bar an unbiased estimator of the population mean? Explain your answer.
c) Is x-tilde an unbiased estimator of the population mean? Explain your answer.
d) Which one of the two estimators is more efficient? Explain your answer.
a) Yes. It is a function of random variables. Different values of different observations generate different
values of sample statistics…
b) Yes. The unbiasedness of an estimator is determined by the condition that its expected value must equal
the thing which we are trying to estimate. In this case, the expected value of x-bar is given to us as being
equal to mu.
c) No. X-tilde suffers from the fact that we have five observations but are dividing not by n, but rather the
constant 4. In this case, we are likely to be consistently over-estimating the value of mu in repeated
sampling.
d) The criteria of efficiency is only relevant for unbiased estimators. Thus, we really cannot say much
about which estimator is more efficient or not (allow for partial marks if x-bar is the answer in light of
what is answered in part c)).
8 Page intentionally left blank. Use this space for rough work or the continuation of an answer.
9 Short Answer #4 (20 points – show your work!)
Consider the following estimated regression equation that describes the relationship between a student’s
weight and their height:
ˆ
Weight
= 100 + 6.0* Height
a) The sum of squared residuals is equal to 15,000. There are 1,002 observations in the data set. The
covariance between height and weight is equal to 36. The variance of weight is equal to 60.
What is your best guess about the value of the variance of the error term in the population
regression model?
b) Using the same information as in a), compute the value of R2 and briefly explain what the calculated
value of R2 means.
c) Suppose an additional variable is included, the last digit of a student’s ID number. We think this
second variable should not determine variation in a student’s weight. The new estimated
regression equation is
ˆ
Weight
= 100 + 6.04* Height + 0.02* ID
How is it possible that the R2 of the second regression is higher than that of the first?
d) If a student’s ID number is indeed not related to a student’s weight, should the estimated coefficient be
equal to zero? That is, do our results suggest we have done something wrong with this
regression?
a) Here, we make use of the equation for
s2 =
∑e
2
i
( n − k − 1)
=
15,000 15,000
=
= 15.0
1,002 − 2 1,000
b) Here, we make use of the equation for
ESS TSS − RSS
RSS
∑ ei2 where
=
= 1−
= 1−
2
TSS
TSS
TSS
∑ (Yi − Y )
R2 =
∑ (Y − Y )
i
R2 = 1 −
2
= (n − 1) *Var (Yi ) = 1, 001*60 = 60, 060 ⇒
15, 000
≈ 1 − 0.25 = 0.75
60, 060
The value of approximately 0.75 indicates that 75% of the observed variation in weights is “explained”
by variation in heights.
c)We know that R 2 =
ESS
. We know that by adding independent variables we can only increase the
TSS
value of ESS as observed variation in heights is still doing its job in explaining variation in weights. We
also know that by adding independent variables—even irrelevant ones—we can increase the value of ESS
and, therefore, the value of the R2.
10 Page intentionally left blank. Use this space for rough work or the continuation of an answer.
d)No, our results do not suggest that we have done something wrong. Even if we expect that the “true”
beta in between weight and student ID numbers is equal to zero, sampling variation in our estimated
coefficient virtually ensures that we will generate non-zero values in finite samples.
11 Useful Formulas:
k
E( X ) =
[
2
k
] ∑ (x − µ
Var ( X ) = E ( X − µ X ) =
∑p x
i i
Pr(Y = y | X = x) =
Pr( X = x) = ∑ Pr ( X = x, Y = yi )
X
)2 pi
i =1
i =1
k
i
i =1
Pr( X = x, Y = y )
Pr( X = x)
m
k
(
)
E
Y
=
E (Y | X = xi ) Pr ( X = xi )
E (Y | X = x ) =
yi Pr (Y = yi | X = x )
i =1
i
=
1
k
2
Var (Y | X = x) = [yi − E (Y | X = x )] Pr (Y = yi | X = x ) E(a + bX + cY ) = a + bE( X ) + cE(Y )
i =1
k
m
Var(a + bY ) = b 2Var(Y )
x j − µ X ( yi − µY )Pr X = x j , Y = yi
Cov( X , Y ) =
i =1 j =1
Cov( X , Y )
Var(aX + bY ) = a 2Var( X ) + b 2Var(Y ) + 2abCov( X ,Y )
Corr ( X , Y ) = ρ XY =
Var( X )Var(Y )
2
2
Cov(a + bX + cV ,Y ) = bCov( X ,Y ) + cCov(V ,Y )
E Y = Var(Y ) + E (Y )
X −µ
X −µ
⎛ σ 2 ⎞
t=
(
)
E
XY
=
Cov
(
X
,
Y
)
+
E
(
X
)
E
(
Y
)
Z
=
X
~
N
⎜ µ , ⎟
s/ n
σ
n ⎠
n
⎝
n
1
1 n
2
2
1
(xi − x )
X =
xi
s =
(xi − x )( yi − y )
rXY = sXY / sX sY
s XY =
n i =1
n − 1 i =1
n − 1 i =1
n
X i − X Yi − Y
i =1
ˆ
For the linear regression model Yi = β 0 + β1 X i + ε i , β1 =
& βˆ0 = Y − βˆ1 X
n
2
Xi − X
i
=
1
Yˆ = βˆ + βˆ X + βˆ X + ! + βˆ X
i
0
1 1i
2 2i
k ki
e2
e2 / (n − k − 1)
RSS
2 ESS TSS − RSS
2
i i
i i
R =
=
= 1−
= 1−
R = 1−
2
2
TSS
TSS
TSS
Y
−
Y
Yi − Y / (n − 1)
i
i
i
e2
e 2 / ( n − k − 1)
i i
i i
s 2 =
ˆ
ˆ
where E ⎡⎣ s 2 ⎤⎦ = σ 2
Var ⎡⎣ β1 ⎤⎦ =
2
( n − k − 1)
Xi − X
i
βˆ j − β H
βˆ − β H
Z =
t= 1
~ tn −k −1
~ N ( 0,1)
∑
∑
∑
∑∑ (
)
(
)
( )
∑
∑
∑
∑(
)(
∑(
∑
∑(
)
s.e.( βˆ1 )
Var[ βˆ j ]
Pr[ βˆ j − tα* /2 × s.e.( βˆ j ) ≤ β j ≤ βˆ j + tα* /2 × s.e.( βˆ j )] = 1 − α
T
∑ (e − e )
d=
∑ e
t =2
2
t =1 t
)
∑
)
∑(
(∑ )
)
∑(
∑
t
T
)
t −1
F=
ESS / k
ESS (n − k − 1)
=
RSS / (n − k − 1)
RSS
k
2
≈ 2(1 − ρ )
12 
Download