LN6-Applications in

advertisement
Lecture Notes #6
AN APPLICATION OF OPTIMIZATION AND MATRIX ALGEBRA:
REGRESSION
Regression analysis, applied in econometrics, is a basic part of economics. Here we will study how the concepts we
learned in the discussion of matrix algebra and derivatives are applied in regression analysis. We will start with
simple regression and then extend the discussion to multiple regression.
1. Simple Regression
In simple regression we want to find out about the relationship between two variable: If the variations in the values
of one variable are associated with the variations with the values of another variable. Here we designate one variable
as the dependent variable and the other as the independent variable. We will use the following simple example, to see
if, and to what extent, the variations in the quantity of a good sold by a firm (the dependent variable) is related to the
annual advertising expenditure (the independent variable).
Year
1
2
3
4
5
6
7
8
9
10
11
12
Sales
y
38
48
52
35
30
56
63
46
61
68
72
65
Ad
Expenditure
x
56.7
63.9
60.7
59.7
55.9
68.7
69.2
65.5
72.5
73.4
74.1
76.2
A scatter diagram of the data is shown in Figure 1.
LN6—Applications in Econometrics
Page 1 of 11
Figure 1
Scatter Diagram of Sales and Advertising Data
As the first step, the scatter diagram indicates that there is a direct relationship between sales volume and advertising
expenditure. We want, however, obtain a mathematical relationship between x and y. The mathematical relationship
is represented by the regression equation, which is the equation of a line that is fitted to the scatter diagram. This
fitted line is shown in Figure 2.
Figure 2
The Fitted Regression Line
The general format of the regression line is as follows:
๐‘ฆฬ‚ = ๐‘โ‚ + ๐‘โ‚‚๐‘ฅ
Note the symbol “^” above y. The symbol ลท (y-hat) represents the y value that lies on the regression line, which is
obtained by plugging in values for x after we determine the values for the slope of the line (b2) and the vertical
intercept (b1). It is called the predicted value. The “plain”, hatless, y represents the value of the dependent variable
observed in data set, called the observed value. For now, the equation of the fitted regression line is
LN6—Applications in Econometrics
Page 2 of 11
๐‘ฆฬ‚ = −67.4299 + 1.8119๐‘ฅ
To show the difference between observed and predicted value of y, consider, for example, the observed sales volume
in year 7 is: y = 63 million units, which corresponds to the advertising expenditure of x = $69.2 thousand. The
predicted sales is:
๐‘ฆฬ‚ = −67.4299 + 1.8119(69.2) = 57.95
The difference between the observed and predicted values of y is called the prediction error or the residual and is
denoted by e.
๐‘’ = ๐‘ฆฬ‚ − ๐‘ฆฬ‚
The prediction error provides the theoretical basis to determine the mathematical process to obtain the values for
slope and vertical intercept (the coefficients) of the regression equation. The mathematical process is called the least
squares method.
Note that for each value of x in the data set there is an observed y value and a predicted y value. Thus, given the
number of observations in the data set, n, there are n residuals. All predicted values lie along the fitted regression line.
There is only one line that fits the scatter diagram the best. The criterion for the best fitting line is that the sum of the
residuals equal zero, that is, the combined positive prediction errors exactly balance the combined negative errors.
๐‘›
๐‘›
∑ ๐‘’๐‘– = ∑(๐‘ฆฬ‚๐‘– − ๐‘ฆฬ‚๐‘– ) = 0
๐‘–=1
๐‘–=1
Since the best-fitting line provides for the sum of residuals being zero, then the sum of squared residuals must be the
minimum, or the least. Thus, we need the method to obtain the coefficients of the regression line by minimizing the
sum of square errors (SSE),
๐‘›
๐‘†๐‘†๐ธ =
๐‘›
∑ ๐‘’๐‘–2
๐‘–=1
2
= ∑(๐‘ฆฬ‚๐‘– − ๐‘ฆฬ‚๐‘– ) = 0
๐‘–=1
2. Least Square Method
The objective here is to find the formulas to compute the values for b1 and b2. We start with the SSE.
๐‘›
∑ ๐‘’๐‘–2
๐‘–=1
๐‘›
2
= ∑(๐‘ฆฬ‚๐‘– − ๐‘ฆฬ‚๐‘– )
๐‘–=1
Substituting for ๐‘ฆฬ‚๐‘– = ๐‘1 + ๐‘2 ๐‘ฅ๐‘– on the right-hand-side, we have,
๐‘›
∑ ๐‘’๐‘–2
๐‘–=1
๐‘›
= ∑(๐‘ฆฬ‚๐‘– − ๐‘1 − ๐‘2 ๐‘ฅ๐‘– )2
๐‘–=1
Since we are interested in determining the values of the two coefficients b1 and b2 such that ∑ ๐’†๐Ÿ is minimized, we take
the partial derivate of the right hand side with respect to b1 and b2 and set them equal to zero.
LN6—Applications in Econometrics
Page 3 of 11
๐œ•
(∑๐‘’ 2 ) = −2∑(๐‘ฆฬ‚ − ๐‘1 − ๐‘2 ๐‘ฅ)
๐œ•๐‘1
๐œ•
(∑๐‘’ 2 ) = −2∑๐‘ฅ(๐‘ฆฬ‚ − ๐‘1 − ๐‘2 ๐‘ฅ)
๐œ•๐‘2
Rewriting the right-hand-side and setting them equal to zero, we have:
∑๐‘ฆฬ‚ − ๐‘›๐‘1 − ๐‘2 ∑๐‘ฅ = 0
∑๐‘ฅ๐‘ฆฬ‚ − ๐‘1 ∑๐‘ฅ − ๐‘2 ∑๐‘ฅ2 = 0
which lead to:
๐‘›๐‘1 + (∑๐‘ฅ)๐‘2 = ∑๐‘ฆฬ‚
(∑๐‘ฅ)๐‘1 + (∑๐‘ฅ2 )๐‘2 = ∑๐‘ฅ๐‘ฆฬ‚
These are called the normal equations of the regression. Here we have two equations with two unknowns ๐‘1 and ๐‘2 .
We can solve for ๐‘1 and ๐‘2 two ways.
Using the matrix notation, the equation system is written a ๐‘‹๐‘ = ๐‘ s, where
๐‘‹=[
๐‘›
∑๐‘ฅ
]
∑๐‘ฅ ∑๐‘ฅ 2
๐‘
๐‘ = [ 1]
๐‘2
๐‘=[
“Coefficient” matrix
“Variable” matrix
∑๐‘ฆฬ‚
]
∑๐‘ฅ๐‘ฆฬ‚
“Constant” matrix
Thus,
[
๐‘›
∑๐‘ฅ
∑๐‘ฅ ๐‘1
∑๐‘ฆฬ‚
]
2 ] [๐‘ ] = [
∑๐‘ฅ
∑๐‘ฅ๐‘ฆฬ‚
2
The solutions for ๐‘1 and ๐‘2 is then found by finding the inverse of the coefficient matrix and the post-multiplying the
inverse by the constant matrix.
๐‘ = ๐‘‹ −1 ๐‘
To find ๐‘‹ −1, first find the determinant of ๐‘‹.
|๐‘‹| = ๐‘›∑๐‘ฅ 2 − (∑๐‘ฅ)2 = ๐‘›∑๐‘ฅ 2 − (๐‘›๐‘ฅฬ… )2
(Substituting for ∑๐‘ฅ = ๐‘›๐‘ฅ
ฬ…)
|๐‘‹| = ๐‘›(∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 )
Next find the Cofactor matrix,
[๐ถ] = [
∑๐‘ฅ 2 −∑๐‘ฅ
]
− ∑๐‘ฅ
๐‘›
LN6—Applications in Econometrics
Page 4 of 11
Since the square matrix is symmetric about the principal diagonal, the Adjoint matrix, which is the transpose of the
cofactor matrix, is the same as [๐ถ]. The inverse matrix ๐‘‹ −1 is then,
๐‘‹ −1 =
1 ∑๐‘ฅ 2
[
|๐‘‹| −∑๐‘ฅ
−∑๐‘ฅ
]
๐‘›
๐‘‹ −1 =
1
∑๐‘ฅ 2
[
๐‘›(∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 ) −∑๐‘ฅ
−∑๐‘ฅ
]
๐‘›
∑๐‘ฅ 2
๐‘‹ −1
−∑๐‘ฅ
๐‘›(∑ −
๐‘›(∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 )
=
−∑๐‘ฅ
๐‘›
2
2
2
[๐‘›(∑๐‘ฅ − ๐‘›๐‘ฅฬ… ) ๐‘›(∑๐‘ฅ − ๐‘›๐‘ฅฬ… 2 )]
๐‘ฅ2
๐‘›๐‘ฅฬ… 2 )
∑๐‘ฅ 2
๐‘‹ −1 =
๐‘›(∑
๐‘ฅ2
๐‘›๐‘ฅฬ… 2 )
−
−๐‘ฅฬ…
[ ∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2
−๐‘ฅฬ…
∑ − ๐‘›๐‘ฅฬ… 2
1
∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 ]
๐‘ฅ2
Thus,
∑๐‘ฅ 2
๐‘
๐‘›(∑ −
[ 1] =
๐‘2
−๐‘ฅฬ…
2
[ ∑๐‘ฅ − ๐‘›๐‘ฅฬ… 2
๐‘ฅ2
๐‘›๐‘ฅฬ… 2 )
−๐‘ฅฬ…
∑ − ๐‘›๐‘ฅฬ… 2 ∑๐‘ฆฬ‚
[
]
1
∑๐‘ฅ๐‘ฆฬ‚
∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 ]
๐‘ฅ2
Using the matrix multiplication rule, first solve for ๐‘2 .
๐‘2 =
−๐‘›๐‘ฅฬ… ๐‘ฆฬ‚ฬ…
∑๐‘ฅ๐‘ฆฬ‚
+
2
2
∑๐‘ฅ − ๐‘›๐‘ฅฬ… ∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2
๐‘2 =
∑๐‘ฅ๐‘ฆฬ‚ − ๐‘›๐‘ฅฬ… ๐‘ฆฬ‚ฬ…
∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2
(∑๐‘ฆฬ‚ = ๐‘›๐‘ฆฬ‚ฬ…)
Next, solve for ๐‘1 .
๐‘1 =
๐‘ฆฬ‚ฬ… ∑๐‘ฅ 2
๐‘ฅฬ… ∑๐‘ฅ๐‘ฆฬ‚
−
∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 ∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2
๐‘1 =
๐‘ฆฬ‚ฬ…∑๐‘ฅ 2 − ๐‘ฅฬ… ∑๐‘ฅ๐‘ฆฬ‚
∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2
๐‘1 =
๐‘ฆฬ‚ฬ…∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 ๐‘ฆฬ‚ฬ… − ๐‘ฅฬ… ∑๐‘ฅ๐‘ฆฬ‚ + ๐‘›๐‘ฅฬ… 2 ๐‘ฆฬ‚ฬ…
∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2
๐‘1 =
๐‘ฆฬ‚ฬ…(∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 ) − (∑๐‘ฅ๐‘ฆฬ‚ − ๐‘›๐‘ฅฬ… ๐‘ฆฬ‚ฬ…)๐‘ฅฬ…
∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2
๐‘1 = ๐‘ฆฬ‚ฬ… − ๐‘2 ๐‘ฅฬ…
LN6—Applications in Econometrics
Page 5 of 11
Alternatively, dividing both sides of the first normal equation by n
๐‘1 + ๐‘2
∑๐‘ฅ
=
๐‘›
∑๐‘ฆฬ‚
๐‘›
๐‘1 + ๐‘2 ๐‘ฅฬ… = ๐‘ฆฬ‚ฬ…
๐‘1 = ๐‘ฆฬ‚ฬ… − ๐‘2 ๐‘ฅฬ…
Now divide the second normal equation by n and substitute for b1.
(๐‘ฆฬ‚ฬ… − ๐‘2 ๐‘ฅฬ…)
∑๐‘ฅ
๐‘›
+ ๐‘2
∑๐‘ฅ 2
๐‘›
=
∑๐‘ฅ๐‘ฆฬ‚
๐‘›
๐‘›(๐‘ฆฬ‚ฬ… − ๐‘2 ๐‘ฅฬ…)๐‘ฅฬ… + ๐‘2 ∑๐‘ฅ 2 = ∑๐‘ฅ๐‘ฆฬ‚
๐‘›๐‘ฅฬ… ๐‘ฆฬ‚ฬ… − ๐‘›๐‘2 ๐‘ฅฬ…2 + ๐‘2 ∑๐‘ฅ 2 = ∑๐‘ฅ๐‘ฆฬ‚
๐‘2 (∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2 ) = ∑๐‘ฅ๐‘ฆฬ‚ − ๐‘›๐‘ฅฬ… ๐‘ฆฬ‚ฬ…
๐‘2 =
∑๐‘ฅ๐‘ฆฬ‚ − ๐‘›๐‘ฅฬ… ๐‘ฆฬ‚ฬ…
∑๐‘ฅ 2 − ๐‘›๐‘ฅฬ… 2
From the data compute the following:
๐‘ฅฬ… = 66.375
๐‘2 =
∑ ๐‘ฅ๐‘ฆฬ‚ = 43066.4
๐‘ฆฬ‚ฬ… = 52.833
∑ ๐‘ฅ² = 53411.13
43066.4 − 12(66.375)(52.833)
= 1.8119
53411.13 − 10(66.375)²
๐‘1 = 52.833 − 1.8119(66.375) = −67.4299
Alternatively, compute the quantities in the matrix notation:
[
๐‘›
∑๐‘ฅ ๐‘1
∑๐‘ฆฬ‚
][ ] = [
]
∑๐‘ฅ ∑๐‘ฅ 2 ๐‘2
∑๐‘ฅ๐‘ฆฬ‚
The solutions for ๐‘1 and ๐‘2 is then found by finding the inverse of the coefficient matrix and the post-multiplying the
inverse by the constant matrix.
X=[
๐‘›
∑๐‘ฅ
∑๐‘ฅ
]
∑๐‘ฅ 2
๐‘
b = [ 1]
๐‘2
c=[
∑๐‘ฆฬ‚
]
∑๐‘ฅ๐‘ฆฬ‚
๐‘ = ๐‘‹ −1 ๐‘
From the data compute the following:
∑๐‘ฅ = 796.5
12
[
796.5
∑๐‘ฆฬ‚ = 634
∑๐‘ฅ² = 53411.13
∑๐‘ฅ๐‘ฆฬ‚ = 43066.4
634.0
796.50 ๐‘1
][ ] = [
]
43066.4
53411.13 ๐‘2
LN6—Applications in Econometrics
Page 6 of 11
Using Excel, find
X −1 = [
8.1902
−0.1221
−0.1221
]
0.0018
๐‘
8.1902
[ 1] = [
๐‘2
−0.1221
−0.1221
634
][
]
0.0018 43066.4
๐‘
−67.4299
[ 1] = [
]
๐‘2
1.8119
3. Multiple Regression
In multiple regression there are two or more independent variables. With two independent variables, the regression
equation is written as
๐‘ฆฬ‚ฬ‚ = ๐‘1 + ๐‘2 ๐‘ฅ2 + ๐‘ฅ3
To determine the coefficients of the multiple regression the only feasible approach is using matrix algebra. Here,
again, we start with finding the normal equations.
∑๐‘’ 2 = ∑(๐‘ฆฬ‚ − ๐‘ฆฬ‚)2
∑๐‘’ 2 = ∑(๐‘ฆฬ‚ − ๐‘1 − ๐‘2 ๐‘ฅ2 − ๐‘3 ๐‘ฅ3 )2
๐œ•
(∑๐‘’ 2 ) = −2∑(๐‘ฆฬ‚ − ๐‘1 − ๐‘2 ๐‘ฅ2 − ๐‘3 ๐‘ฅ3 ) = 0
๐œ•๐‘1
๐œ•
(∑๐‘’ 2 ) = −2∑๐‘ฅ2 (๐‘ฆฬ‚ − ๐‘1 − ๐‘2 ๐‘ฅ2 − ๐‘3 ๐‘ฅ3 ) = 0
๐œ•๐‘2
๐œ•
(∑๐‘’ 2 ) = −2∑๐‘ฅ3 (๐‘ฆฬ‚ − ๐‘1 − ๐‘2 ๐‘ฅ2 − ๐‘3 ๐‘ฅ3 ) = 0
๐œ•๐‘3
The normal equations are:
๐‘›๐‘1 +
(∑๐‘ฅ2 )๐‘1 +
(∑๐‘ฅ2 )๐‘2 +
(∑๐‘ฅ3 )๐‘3 = ∑๐‘ฆฬ‚
(∑๐‘ฅ22 )๐‘2 + (∑๐‘ฅ2 ๐‘ฅ3 )๐‘3 = ∑๐‘ฅ2 ๐‘ฆฬ‚
(∑๐‘ฅ3 )๐‘1 + (∑๐‘ฅ2 ๐‘ฅ3 )๐‘3 +
(∑๐‘ฅ32 )๐‘3 = ∑๐‘ฅ3 ๐‘ฆฬ‚
In matrix format, we have,
๐‘›
[∑๐‘ฅ2
∑๐‘ฅ3
∑๐‘ฅ2
∑๐‘ฅ22
∑๐‘ฅ2 ๐‘ฅ3
∑๐‘ฅ3
∑๐‘ฆฬ‚
๐‘1
∑๐‘ฅ2 ๐‘ฅ3 ] [๐‘2 ] = [∑๐‘ฅ2 ๐‘ฆฬ‚]
๐‘3
∑๐‘ฅ3 ๐‘ฆฬ‚
∑๐‘ฅ32
Labeling the three matrices as before, to find the values of the regression coefficients,
b = X −1 k
LN6—Applications in Econometrics
Page 7 of 11
Example
The following data shows the price of houses as the dependent variable, and the size (in square feet) and the age
(years) as the independent variables. We want to determine to what extent the price varies with the size and the age
of the house.
PRICE
y
89950
138950
87000
165000
210000
108000
89000
79000
124500
135000
105500
133650
83500
101000
151500
88500
198000
135000
79500
135050
175000
71000
76000
99250
98500
117500
97000
125000
115000
145000
SQFT
x2
917
1684
1800
1900
2000
1050
1057
954
1350
2134
1313
1671
1200
1314
1877
1132
2198
1525
1208
1450
2000
1267
1088
1159
1255
1386
1400
1442
1477
1566
AGE
x3
39
10
20
13
20
40
45
6
47
13
12
37
18
45
10
38
2
10
29
35
9
18
12
35
4
30
63
12
23
6
The following are the quantities needed to generate the X and k matrices.
๐‘› = 30
∑๐‘ฅ22 = 67540602
∑๐‘ฆฬ‚ = 3556850
30
43774
X = [43774 67540602
701
955209
∑๐‘ฅ2 = 43774
∑๐‘ฅ32 = 23573
∑๐‘ฅ2 ๐‘ฆฬ‚ = 5494952850
701
955209]
23573
∑๐‘ฅ3 = 701
∑๐‘ฅ2 ๐‘ฅ3 = 955209
∑๐‘ฅ3 ๐‘ฆฬ‚ = 77587100
3556850
k = [5494952850]
77587100
b = X −1 k
LN6—Applications in Econometrics
Page 8 of 11
๐‘1
1.0387523
[๐‘2 ] = [−0.0005537
๐‘3
−0.0084550
−0.0005537
0.0000003
0.0000031
−0.0084550
3556850
0.0000031] [5494952850]
0.0001682
77587100
๐‘1
−3609.450
[๐‘2 ] = [
83.459]
๐‘3
16.802
The estimated regression equation is then,
๐‘ฆฬ‚ฬ‚ = −3609.450 + 83.459๐‘ฅ2 + 16.802๐‘ฅ3
Thus, for example, the price of a 2,000 square feet house of 10 years of age is predicted to be
๐‘ฆฬ‚ฬ‚ = −3609.450 + 83.459(2000) + 16.802(10) = $163,476.57
4. Functional Forms in Econometrics
In the simple linear regression model the population parameters βโ‚ (the intercept) and βโ‚‚ (the slope) are linear—that
is, they are not expressed as, say, β22 , 1⁄β2 , or any form other than βโ‚‚—and also the impact of the changes in the
independent variable on ๐‘ฆฬ‚ works directly through ๐‘ฅ rather than through expressions such as, say, ๐‘ฅ², 1⁄๐‘ฅ , or ln ๐‘ฅ.
In many cases, even though the parameters are held as linear, the variables may take on forms other than linear. In
many economic models the relationship between the dependent and independent variables is not a straight line
relationship. That is the change in y does not follow the same pattern for all values of x. Consider for example an
economic model explaining the relationship between expenditure on food (or housing) and income. As income rises,
we do expect expenditure on food to rise, but not at a constant rate. In fact, we should expect the rate of increase in
expenditure on food to decrease as income rises. Therefore the relationship between income and food expenditure is
not a straight-line relationship.
The following is an outline of various functional forms encountered in regression analysis. First we start with the
straight-line (linear) relationship between ๐‘ฅ and ๐‘ฆฬ‚ and then point out various non-linear functional relationships.
4.1. Linear Functional Form
The linear functional form is the familiar,
๐‘ฆฬ‚ = β1 + β2 ๐‘ฅ
The slope of the function is:
๐‘‘๐‘ฆฬ‚
= β2
๐‘‘๐‘ฅ
The slope represents the change in y per unit change in x.
The elasticity of the function shows the proportional or percentage change in y relative to a percent change in x:
ε=
๐‘‘(ln ๐‘ฆฬ‚)
๐‘ฅ
= β2
๐‘‘(ln ๐‘ฅ)
๐‘ฆฬ‚
To show this, let ๐‘ข = ln ๐‘ฆฬ‚, and ๐‘ฃ = ln ๐‘ฅ. But, since ๐‘ฃ = ln ๐‘ฅ, then ๐‘ฅ = ๐‘’ ๐‘ฃ . Therefore,
๐‘‘(ln ๐‘ฆฬ‚) ๐‘‘๐‘ข ๐‘‘๐‘ข ๐‘‘๐‘ฆฬ‚ ๐‘‘๐‘ฅ
=
=
๐‘‘(ln ๐‘ฅ) ๐‘‘๐‘ฃ ๐‘‘๐‘ฆฬ‚ ๐‘‘๐‘ฅ ๐‘‘๐‘ฃ
LN6—Applications in Econometrics
Page 9 of 11
๐‘‘(ln ๐‘ฆฬ‚)
1
๐‘ฅ
= ( ) (β2 )๐‘’ ๐‘ฃ = β2
๐‘‘(ln ๐‘ฅ)
๐‘ฆฬ‚
๐‘ฆฬ‚
๐‘ฅ
Note that to obtain elasticity, simply multiply the slope by .
๐‘ฆ
ε=
๐‘‘๐‘ฆฬ‚ ๐‘ฅ
๐‘ฅ
= β2
๐‘‘๐‘ฅ ๐‘ฆฬ‚
๐‘ฆฬ‚
4.2. Reciprocal Functional Form
The reciprocal functional form is,
๐‘ฆฬ‚ = β1 + β2
1
๐‘ฅ
The slope is,
๐‘‘๐‘ฆฬ‚
β2
=− 2
๐‘‘๐‘ฅ
๐‘ฅ
and the elasticity is,
ε=
๐‘‘(ln ๐‘ฆฬ‚)
1
= −β2
๐‘‘(ln ๐‘ฅ)
๐‘ฅ๐‘ฆฬ‚
Note that, as usual, using the general definition of elasticity,
ε=
๐‘‘๐‘ฆฬ‚ ๐‘ฅ
β2 ๐‘ฅ
1
= (− 2 ) ( ) = −β2
๐‘‘๐‘ฅ ๐‘ฆฬ‚
๐‘ฅ
๐‘ฆฬ‚
๐‘ฅ๐‘ฆฬ‚
4.3. Log-Log Functional Form
Many relationships between variables are naturally expressed in percentages. Logarithms convert changes in
variables into percentage changes. The log-log functional form is,
ln ๐‘ฆฬ‚ = β1 + β2 ln ๐‘ฅ
The slope and elasticity of the function are:
1 ๐‘‘๐‘ฆฬ‚
1
= β2
๐‘ฆฬ‚ ๐‘‘๐‘ฅ
๐‘ฅ
ε=
๐‘‘๐‘ฆฬ‚
๐‘ฆฬ‚
= β2
๐‘‘๐‘ฅ
๐‘ฅ
๐‘‘๐‘ฆฬ‚ ๐‘ฅ
๐‘ฆฬ‚ ๐‘ฅ
= (β2 ) ( ) = β2
๐‘‘๐‘ฅ ๐‘ฆฬ‚
๐‘ฅ ๐‘ฆฬ‚
Consider, for example, the function,
ln ๐‘ฆฬ‚ = 0.5 + 2 ln ๐‘ฅ
For ๐‘ฅ0 = 1.5,
ln ๐‘ฆฬ‚ = 0.5 + 2 ln 1.5 = 0.5 + 2(0.4055) = 1.3109
๐‘ฆฬ‚0 = exp(1.3109) = 3.7096
The slope of the function at the point ๐‘ฅ0 = 1.5 is then,
LN6—Applications in Econometrics
Page 10 of 11
๐‘‘๐‘ฆฬ‚
๐‘ฆฬ‚
3.7096
= β2 = 2
= 4.9462
๐‘‘๐‘ฅ
๐‘ฅ
1.5
This means that for each small unit increase in x, y increases by 4.9462.
Now let’s compute the elasticity. First, assume x increases by 1 percent point from ๐‘ฅ0 = 1.5 to ๐‘ฅ1 = 1.5 × 1.01 =
1.515. Then,
ln ๐‘ฆฬ‚1 = 0.5 + 2 ln 1.515 = 1.3308
๐‘ฆฬ‚ = exp(1.3308) = 3.7842
Thus, the percentage change in y when x increases by 1 percentage point is,
3.7842 − 3.7096
= 0.0201 ๐‘œ๐‘Ÿ 2.01%
3.7096
Elasticity is then, ε = 0.0201⁄0.01 = 2.01. For a very small percent change in x, elasticity will approach β2 = 2, as
shown in the following table.
“Percent”
change in x
(1)
0.01
0.005
0.0025
0.001
0.0001
0.00001
๐’™๐Ÿ
(2)
1.515
1.5075
1.50375
1.5015
1.50015
1.500015
๐’š๐Ÿ
(3)
3.7842
3.7468
3.7282
3.7170
3.7104
3.7097
โˆ†๐’š⁄๐’š๐ŸŽ
(4)
0.0201
0.010025
0.005006
0.002001
0.0002
0.00002
Elasticity
(4)/(1)
2.01
2.005
2.0025
2.001
2.0001
2.00001
The following will be added:
Linear-log (semi-log) model
Log-inverse model
LN6—Applications in Econometrics
Page 11 of 11
Download