Time-Varying Parameters and Endogenous Learning Algorithms Eric Gaus July 2015

advertisement
Time-Varying Parameters and Endogenous
Learning Algorithms
Eric Gaus∗
July 2015
Abstract
The adaptive learning literature has primarily focused on decreasing gain
learning and constant gain learning. As pointed out theoretically by Marcet
and Nicolini (2003) and empirically by Milani (2007) an endogenous learning
mechanism may explain key economic behaviors, such as recurrent hyperinflation
or time varying volatility. This paper evaluates the mechanism used in those
papers and the adaptive step size algorithm proposed by Kostyshyna (2012)
in addition to proposing an alternative endogenous learning algorithm. The
proposed algorithm outperforms both alternatives in simulations and may result
in exotic dynamics in macroeconomic models and models of hyperinflation.
Keywords: Adaptive Learning, Rational Expectations, Endogenous Learning.
∗
Ursinus College, 601 East Main St., Collegville, PA 19426-1000 (e-mail: egaus@ursinus.edu)
1
Introduction
Macroeconomic empiricists have long been concerned with the potential for time variation in the parameters of their econometric models. This time variation has been
modeled as structural breaks in the Markov switching literature and using a basic
Kalman filter. Macroeconomists are also concerned with accurately capturing how
agents form expectations. Bernanke (2007) suggests that monetary policy research
should use an adaptive learning framework. Adaptive learning is a bounded rationality
exercise that assumes that agents use basic econometric techniques to form forecasts of
the economic variables of interest. Given the concern over time-varying parameters in
the real world, one might be interested in a plausible learning process that addresses
the potential for time-variation.
In learning, agent’s econometric forecasts are updated using a recursive least squares
(RLS) algorithm, which has a critical component called a gain parameter. When a
researcher assumes agents believe the structure of the economy is stable it is common
to use a decreasing gain, since this reduces forecasting errors, but if agents believe the
structure varies over time a constant gain is used, since this type of gain tracks varying
coefficients more efficiently. An alternative notion of the gain is that it relates to how
much data the agent uses. Under a decreasing gain agents increase the amount of data
and a constant gain keeps the size of the data set constant. From a behavioral point
of view, one might expect agents to adjust the size of the data set endogenously with
respect to some statistical or econometric rule.
There have been a few recent papers that consider endogenous gains of this sort.
Marcet and Nicolini (2003) suggest that in the presence occasional structural breaks,
agents might switch between a constant gain, to track the change at a break, and a
decreasing gain, to reduce forecasting error, in stable periods.The switch to a constant
gain occurs when the forecasts errors rise above a certain threshold. According to
Marcet and Nicolini (2003) this type of behavior may help account for recurrent hyperinflations.1 Kostyshyna (2012) uses a similar model, but an alternative endogenous
gain. Drawing on the vast engineering literature on recursive algorithms Kostyshyna
(2012) proposes using the adaptive step size algorithm. Instead of switching between
constant and decreasing gain, the adaptive step size algorithm adjusts the value of the
gain depending upon the series of forecast errors. High forecast errors lead to higher
gains and low forecast errors lead to lower gains. This endogenous algorithm yields
much the same results as Marcet and Nicolini (2003) though with lower mean squared
errors and matches the stylistic facts of the data.
This paper adds to this growing literature on endogenous gains by proposing an
alternative motivated by agent level econometrics. Recall that these endogenous gains
are primarily motivated by situations where agents believe there are occasional structural breaks. When econometricians are concerned about structural breaks they pay
particular attention to the coefficient estimates, not just the forecast errors. This point
is directly evident in Stock and Watson (2007). While forecast errors may be indicative
of structural breaks, in our adaptive learning framework agents update the coefficients
1
This type of algorithm has been used by Milani (2007) as a potential explanation of the Great
Moderation.
1
each period, which suggests that they might be more attentive to large departures from
recent coefficient estimates. This is the basic notion that underlies the proposed gain.
Specifically, if the most recent coefficient estimate is many standard deviations away
from a recent mean of the coefficients then the gain is relatively high, and otherwise it
is relatively low.
We seek to answer two questions. First, would agents choose to use an endogenous
gain over a constant gain? The results of several simulations suggest that in the presence of occasional structural breaks both gain sequences in Marcet and Nicolini (2003)
and Kostyshyna (2012) rarely outperform a constant gain in terms of mean squared
forecast error (MSFE). The proposed algorithm frequently outperforms a constant gain
over a wide range of parameter values with a potential improvement of MSFE of 7 percent. In addition, coefficient estimates of the proposed gain are relatively closer to
the rational expectations solution than the alternatives. Further, the proposed algorithm performs better in the presence of structural breaks as opposed to an AR(1)
time-varying parameter process suggesting that it is in fact responding to structural
breaks and not time variation in general. We also consider a game theoretic scenario
where we test whether a single agent considering the use an endogenous gain might
have better forecasting performance using the endogenous gain in an economy where
the dynamics are driven by agents using a constant gain. We find that such an agent
will always have an incentive (based on MSFE) to switch to the proposed endogenous
gain for the model described below. Second, what economic importance might an
endogenous gain algorithm have? To answer this question we simulate the hyperinflation model presented in Marcet and Nicolini (2003) and a small New Keynesian (NK)
model. We find that the endogenous gain results in qualitatively the same dynamics as
both alternatives in the hyperinflation model. In the NK model we find, under certain
conditions, recurrent episodes of deviations from rational expectations. Whereas the
hyperinflation model studied by Marcet and Nicolini (2003) and Kostyshyna (2012)
impose structural change by way of an exchange rate regime, this NK model has no
inherent structural change yet still results in temporary deviations from rational expectations. This suggests that endogenous gains on their own may result in bubbles
or hyperinflations, without any other structural changes forcing the model back to
the rational expectations equilibrium. This result relies on the E-instability of “large”
constant gain values.
The rest of the paper is organized as follows. The next section sets up the model
used to test forecasting performance, describes the time varying parameter processes
and presents the endogenous gain algorithms. Section three provides the results of
several head to head comparisons of MSFE of the endogenous gains. The fourth section
describes the stability properties of the endogenous gain algorithms in a model of
hyperinflation and a simple NK model. Section five concludes.
2
Framework
This section presents a simple univariate model that provides an opportunity to examine the performance of each of the endogenous gain algorithms. Two time-varying
processes are used to demonstrate the type of time variation that yields the best results.
2
In addition, this section describes all three endogenous gain algorithms.
2.1
The Time-Varying Parameters Model
Consider the following univariate system,
yt = αt xt−1 + δyte + ηt ,
(1)
where xt is a mean zero random variable, η is normally distributed with mean zero
and variance ση , and the superscript e denotes expectations conditional on information
available at t − 1: yte = E(yt |It−1 ).2 We will consider two potential time varying
parameters scenarios. The first is a model of occasional structural breaks,
(
αt−1 with Prob. (1-ε),
αt =
(2)
υt
with Prob. ε,
where υt is and iid Gaussian error with variance στ2 . Evans and Ramey (2006) use
this particular process because it is straightforward yet econometrically challenging to
estimate. The second scenario we will consider makes the assumption, common in the
time-varying parameters literature, that αt follows an AR(1) process.
αt = ψαt−1 + ωt ,
(3)
where ωt is an iid, mean zero, variance σω white noise process and −1 < ψ < 1. We consider the AR(1) scenario to demonstrate that the endogenous algorithm is particularly
suited to occasional structural breaks not just time variation in general.
We model the expectation formation process using adaptive learning. The adaptive
learning framework considers the conditions under which agents are able to “learn” the
rational expectations solution of a model. These so called E-stability conditions serve
as parameter constraints of the model.3 The E-stability conditions are found by setting
up differential equations based on the relationship between the agents perceived law of
motion (PLM) to the resulting actual law of motion (ALM) when agents expectations
are realized in the model. For the model at hand, the PLM is,
yt = axt−1 + et ,
(4)
yt = (αt + δa)xt−1 + ηt .
(5)
and the resulting ALM is,
If αt = α then we can apply the standard E-stability principle result which, in this
case, implies the equilibrium is E-stable when δ < 1. One might be concerned whether
this model will converge to rational expectations as Bullard (1992) finds that if agents
believe there are time varying parameters they may never learn the rational expectations equilibrium. Instead agents converge to a restricted perception equilibrium since
2
The results found here are robust to setting xt = 1 ∀t. However, the magnitude of the improvements are less.
3
For more on E-stability see Evans and Honkapohja (2001).
3
they are unaware of the underlying time varying process. As pointed out by McGough
(2003) under the right conditions on the stochastic process and the stability parameter
(δ in our case) convergence to the rational expectations equilibrium still obtains.4
In the adaptive learning literature agents recursively updated their estimates of a
using the following algorithm,
at = at−1 + γt Rt−1 xt−1 (xt−1 − xt−1 at−1 ),
Rt = Rt−1 + γt (xt−1 xt−1 − Rt−1 ).
(6)
(7)
where Rt is the variance of xt−1 , and γt is the so called gain parameter. If γt = 1/t
these equations form the standard RLS algorithm and is referred to as decreasing gain
learning. Another common assumption is to set γt equal to some constant. This is
referred to as constant least squares or constant gain learning. In this paper, we assume
that γt will vary endogenously as described below.
2.2
Endogenous Algorithms
Marcet and Nicolini (2003) posit that in a world with occasional structural breaks
agents would prefer to switch between a constant gain and a decreasing gain. They
suggest using an average of past forecast errors to determine when switches occur. In
this context, they propose an algorithm of the following form,
Pt
(
1
i=t−J−1 |yi −ai |
if
< v,
γ̄ −1 +k
J
Pt
(8)
γt =
|y
−a
|
i
i
≥
v
γ̄
if i=t−J−1
J
where k denotes the number of periods since the switch to a decreasing gain, J is
the number of forecast errors used in the average, and v is an arbitrary cutoff point.
Hereafter we refer to this algorithm as MN. Milani (2007) further endogenizes the
algorithm by suggesting that the arbitrary cutoff v be a historical average of forecast
errors. That is, some window of forecast errors larger than J is used to calculate
average forecast errors for the right hand side of the inequalities. In this simple case
such an algorithm, makes considerable sense, should α or δ change value agents would
react if the their forecast errors increase. However, when there are multiple coefficients,
this approach seems rather limited.
By contrast Kostyshyna (2012) draws on the large engineering literature to offer the
adaptive step size algorithm. This algorithm allows the gain parameter to vary based
on probability distributions of the data, and the observation noise. A small gain is
preferred when observation noise is high. The algorithm takes the following form,
Y
e
γt =
(γt−1 + µ(yt−1 − yt−1
)Vt−1 ),
(9)
[γ̄− ,γ̄+ ]
e
Vt = Vt−1 − γt−1 Vt−1 + (yt−1 − yt−1
), V0 = 0.
4
(10)
The specific conditions on the stochastic process are that agents believe that the coefficient of
interest follows a random walk and the associated error decreases over time
4
Q
where µ is the step size parameter, [γ̄− ,γ̄+ ] bounds the value gain, and Vt is the “derivative” of the estimated parameter. The lower bound is not important in applications,
but the upper bound can be influential depending on the stability properties of the
model. The simplicity of our model makes the choice of the upper bound less critical.
Hereafter we refer to this algorithm as K.
Note that both of these methods rely on the forecast errors to adjust the gain
parameter. When an econometrician examines data for potential structural breaks
(s)he looks at the coefficients not the forecast errors. Motivated by this methodology
and the fact that the RLS algorithm and adaptive learning assume that agents have
access to past coefficient estimates, the proposed endogenous gain algorithm uses a
recent coefficient estimate to adjust the gain by the normalized standard deviation.
This results in the following endogenous gain algorithm,
ât −āt σ̄a ,
(11)
γt = γ̄lb + γ̄sf
t
1 + âtσ̄−ā
a
where γ̄lb is the lower bound of the of the endogenous gain, γ̄sf is a scaling factor, ât is
a recent coefficient estimate, āt and σ̄a are the mean and standard deviation of the w
most recent coefficient estimates, respectively. If the recent coefficient is very close to
the mean γt = γ̄lb and as the deviation from the mean increases the value of the gain
approaches γ̄lb + γ̄sf . Therefore, as long as 0 < γ̄lb , γ̄lb < 1 and γ̄lb + γ̄sf < 1 then γt will
be bounded between zero and one. As time progresses agents will increase the value of
the gain in times when their coefficient estimates are different from the recent past and
decrease the value of the gain when their coefficient estimates are similar. Hereafter
we refer to this algorithm as PEG.
3
Forecasting Results
Our results are found via simulation. The simulation procedure starts with an optimization routine to determine the values of the parameters for each of the learning
algorithms. The parameters are optimized to minimize the MSFE of a 50,000 period
simulation. The resulting optimal parameters are then used in 100 independent simulations of 50,000 periods. For MN we optimize over v, j, and γ̄, for the adaptive step
size gain we optimize over µ and the for the proposed algorithm we optimize over w,
γ̄lb , and γ̄sf . We then report the mean and standard deviation of the MSFE for the
last 30,000 periods of the 100 simulations. One could increase the number of periods
or the number of simulations for improved standard deviations, but this would add
a significant amount of computing time. In addition, we calculate the deviations of
the coefficient estimates from the rational expectations solution relative to a constant
gain. The optimal values of the gains are not reported mainly to save space. The
optimal PEG values straddle the optimal constant gain and the gain values increase
as volatility increases. We consider three different expectational environments, no expectational feed back, expectational feedback, and expectational feedback with game
theoretic behavior.
5
Consider (1) with a AR(1) process (3) with ψ = 0.99 and δ = 0. In this case there
will be no expectational feedback and we can focus on the quality of the forecasting.
Table 1 presents the baseline results of several simulations where we have varied the
variance of the AR process. We observe little improvement relative to a constant gain
by all the endogenous gain algorithms. However, the endogenous gains were motivated
by structural breaks and therefore an occasional structural break process (2) may yield
greater improvements. We first demonstrate that these all these algorithms do not
perform well when the breaks are frequent as documented in Table 2. One can find
little difference over a wide range of the variance of the structural break , στ relative
to the variance of the underlying process, ση and over the frequency of the breaks
Table 3 displays the results of simulations of less frequent breaks. The results show
that as the probability of a structural break decreases both MN and PEG improve
performance relative to a constant gain and PEG appears to perform as well or better
than the alternatives. Strikingly, K seems to be equivalent to a constant gain over all
simulations. Performance of PEG increases as the relative variance increases, whereas
MN performance decreases. Table 4 shows the results of the deviations from the RE
solution for the same simulations. Not surprisingly the results mimic the same pattern
found in Table 3. However, one might not expect the same pattern when we add in
expectational feedback.
To that end, we simulate the same model when δ = 0.5 and find a larger improvement
in MSFE for the endogenous gain. Table 5 reports those results for the MSFE. While
we observe a similar pattern for the proposed gain as found with no expectational
feedback, we find the opposite with the alternatives. The MN algorithm decreases
performance as the likelihood of a structural break decreases. Also performance gets
drastically worse as the relative variance increases. The result for the K algorithm is
more subtle. While it does perform worse relative to a constant gain, the performance
is more variable when the relative variances are small. Table 6 gives some insight to
why we observe these striking differences by comparing agents coefficient estimates
to what the RE solution would be at each point in time (that is, for each αt ). The
proposed gain results in coefficient estimates that are much closer to the RE solution
than a constant gain, which is where the improved performance comes from. In contrast
the MN algorithm gets progressively worse. This occurs to a lesser extent with the K
algorithm. These results suggest that with expectational feedback the PEG algorithm
hovers around the closer to the RE solution, while the alternatives wander quite far
away.
These conditions impose that all agents use the same gain process. In order to
determine whether agents would truly choose to use this endogenous gain, consider a
model where all the agents are using some optimal constant gain. Would an agent
be able to improve their forecast if they used and endogenous gain? Figure 1 plots
the values of the optimal response, in terms of MSFE, when the data are generated
with a particular value of constant gain with expectational feedback. It turns out,
that the Nash equilibrium is exactly equal to the value of the optimal constant gain
with expectational feedback. One might also ask whether the proposed gain would
always be a best response over the same range. Figure 2 presents the MSFE within the
model (red line), of the optimal constant gain response (dashed line), and the optimal
proposed gain response (solid line). Just as was found in the previous results the
6
optimal parameters for PEG resulted in a range the encompassed the optimal constant
gain response. Specifically, at the Nash Equilibrium gain value of 0.436 the optimal
endogenous response is γ̄lb = 0.314 and γ̄sf = 0.249. Over all the constant gain values
that generated the data agents will always have an incentive, in terms of MSFE, to
switch to the proposed gain instead of the optimal constant gain. As an example, at
the Nash Equilibrium the MSFE of the constant gain is 5.671 compared with 5.548
with PEG, or about a three percent improvement. This conforms with what one might
perceive about the real world, people who are adaptable, i.e. adjust the information
set relative to past information, tend to predict more accurately than those with more
rigid information sets.
Finally we note that this improvement occurs in a univariate model with only one
coefficient estimate. With a constant gain, MN and K, the gain value will be identical
across all estimated coefficients, however, with PEG there will be a different value for
each coefficient estimate. The agents assess the potential for a structural break in
each of the coefficients separately. Therefore, as the model increases in complexity the
proposed algorithm should have even larger improvements in forecast error.
4
Economic Significance
This section provides two models to demonstrate the economic significance of endogenous gains. The first utilizes the framework of Marcet and Nicolini (2003) to demonstrate that PEG results in similar dynamics as K and MN. This result should not be
surprising since the key feature of Marcet and Nicolini (2003) that results in recurring
hyperinflation episodes is structural change imposed by the exchange rate regime. The
second example demonstrates that changes in the gain parameter over time can, on
their own, result in temporary deviations from the rational expectations equilibrium
in a simple New Keynsian (NK) model.
4.1
Hyperinflation Example
Marcet and Nicolini (2003) develop a model of recurrent hyperinflation that relies
a Cagan-style money demand specification, and a money supply process where the
government either implements an exchange rate regime or seignorage. The reduced
form of the model in terms of inflation and similar notation follows:
πt =
e
1 − γπt+1
1 − γπte − dt /ϕ
(12)
where γ and ϕ are parameters of the money demand equations and dt is as stochastic
process governing seignorage. Under rational expectations there are two deterministic
steady states. Marcet and Nicolini (2003) assume that agents form expectations over
the mean of past inflation rates. The recursive formula for updating the mean βt is as
follows:
βt = βt−1 + gt (πt−1 − βt−1 )
(13)
where gt is the gain parameter.
7
Using the same parameter values as Marcet and Nicolini (2003), specifically γ = 0.4,
φ = 0.37, E(d) = 0.049, and σd = 0.01, one can generate similar results as discussed in
both Marcet and Nicolini (2003) and Kostyshyna (2012) as demonstrated by Figure (3).
The values for the endogenous algorithm were set as γ̄lb = 0.01 and γ̄sf = 0.39, which is
the same range as Kostyshyna (2012). The top panel of the figure displays the time path
of the inflation process and expectations. The bottom panel displays the value of the
gain for each time period. Typically the hyperinflation episodes coincide with a period
where the gain is relatively high. However, a high gain need not precipitate an episode.
Recall that the gain increases when the most recent data causes large changes in the
coefficient estimate. During the hyperinflation episodes the gain increases because
agents make poor predictions of inflation. When the exchange rate regime is enforced
the gain remains high until expectations converge towards the current inflation rate.
Finally, we would like to note that the frequency and severity of the hyperinflation
episodes will be related to the size of the scaling factor (γ̄sf ).
4.2
NK Example
Gaus (2013) shows that MN creates some exotic dynamics when applied to the framework of Duffy and Xiao (2007) and Evans and Honkapohja (2009). In a simple New
Keynesian model, certain parameter sets may be E-stable under a decreasing gain,
but not E-stable under large constant gains. Gaus (2013) demonstrates through simulations that if the constant part of MN is a value that would not be E-stable then
temporary deviations from the rational expectations equilibrium may occur. These
deviations are economically significant, as they exhibit far greater volatility (up to 6
times more) than what would have occurred under rational expectations. A similar
type of dynamics may occur with the proposed gain studied above.
In order to assess the dynamics consider the following NK model, presented in
section 3 of Evans and Honkapohja (2009),5
e
xt = xet+1 − ϕ(it − πt+1
) + gt ,
e
πt = βπt+1 + λxt + ut ,
ϕλ e ϕαx e
π +
x,
it =
αi t
αi t
(14)
(15)
(16)
where ut and gt are AR(1) processes. The following equations govern these processes:
ut = ρut−1 + ũt , and gt = µgt−1 + g˜t ,
where g˜t ∼ iid(0, σg2 ), ũt ∼ iid(0, σu2 ), and 0 < |µ|, |ρ| < 1. Here ϕ is the inter temporal
elasticity of substitution, β is the discount factor, λ is the slope of the Phillips curve, and
αi and αx are relative weights in the monetary policy makers loss function. Substituting
(16) and rewriting the model in reduced form,
e
ξt = M0 ξte + M1 ξt+1
+ P υt
5
See Woodford (2003) for derivation.
8
(17)
where ξt = (xt , yt )0 , υt = (gt , ut )0 , υt = F υt−1 + υ̃t ,
!
ϕ2 αx
ϕ2 λ
1
ϕ
1 0
αi
αi
M0 = ϕ2 λαx ϕ2 λ2
M1 =
P =
λ β + ϕλ
λ 1
α
α
i
F =
µ 0
.
0 ρ
i
Consequently, the MSV solution that serves as the agents PLM is,
yt = a + cυt ,
(18)
with following rational expectations solution,
are = 0
vec(cre ) = (I2 ⊗ M0 + F 0 ⊗ M1 )vec(P )
Note that with the proposed gain there are six coefficients with six different gain value
sequences.
Since PEG requires previous information, all simulations begin with a 40 period
burn-in. During the burn-in each equation receives an additional exogenous error each
period for each equation. This allows for enough variability in the data to generate
variance covariance matrices necessary for the construction of the endogenous-gain.
After the burn-in, the additional exogenous variation shuts down and the simulation
continues without any extraneous noise. Similar to Evans and Honkapohja (2009)
the calibrated parameter values are drawn from Table 6.1 of Woodford (2003), with
αx = 0.048, αi = 0.077, ϕ = 1/0.157, λ = 0.024 and β = 0.99. In addition, µ =
ρ = 0.8 and σg2 = σu2 = 0.2. Evans and Honkapohja (2009) find that this particular
parameterization results in instability if agents use a constant-gain greater than or
equal to 0.024. Therefore we consider γ̄lb = 0.005 and γ̄sf = 0.035 so that the proposed
algorithm may take on a value in the unstable region.
Figure 4 presents the deviations from the rational expectations equilibrium for one
such simulation. Notice the recurrent deviations from the rational expectations, these
continue indefinitely. Whereas Gaus (2013) reports that the deviations occur very
infrequently with MN, these fluctuations occur once every 300 periods on average.
This suggests that PEG implies even greater economic impact of expectations. Also
note there are no structural time varying parameters in this particular model. These
fluctuations occur due to the assumptions on expectation formation. In addition, the
values of PEG remain within the E-stable range, then the deviations do not exist. For
a large enough value of the scaling factor (γ̄sf ) the model will not be stable. For the
intermediate cases as the range between the lower bound, γ̄lb , and the scaling factor,
γ̄sf , increases the departures from RE become more frequent and more severe.
5
Conclusion
This paper systematically explored a proposed endogenous gain that outperforms MN
and K. Specifically we examine a model with occasional structural breaks and find
that the proposed gain is capable of improving forecasting accuracy because it remains
9
closer to the RE equilibrium. We have also demonstrated that agents will have an
incentive, through lower forecasting errors, to switch from a pure constant gain to
an endogenous gain. In addition, simulation of a hyperinflation model and a simple
NK model suggest that the endogenous gain may create additional recurrent dynamics
with significant economic impact. The intuition behind the dynamics found in the NK
model is particularly compelling, agents, wary of structural breaks, may use too little
information to form expectations leading to a period of non-rational economic outcomes. Eventually agents realize their mistakes and reacquire the rational expectation
equilibrium by using more data to form expectations. Endogenous gains seem to be
a natural outgrowth of the adaptive learning literature and have potential to explain
macroeconomic phenomena.
References
Bernanke, Benjamin, “Inflation expectations and inflation forecasting,” Monetary
Economics Workshop of the National Bureau of Economic Research Summer Institute July 2007.
Bullard, James, “Time-varying parameters and non-convergence to rational expectations under least square learning,” Economic Letters, 1992, 40 (2), 159–166.
Duffy, John and Wei Xiao, “The value of interest rate stabilization policies when
agents are learning,” Journal of Money, Credit and Banking, 2007, 39 (8), 2041–
2056.
Evans, George W and G Ramey, “Adaptive expectations, underparameterization
and the Lucas critique,” Journal of Monetary Economics, 2006, 53 (2), 249–264.
and Seppo Honkapohja, Learning and expectations in macroeconomics, Princeton: Princeton University Press, January 2001.
and
, “Robust learning stability with operational monetary policy rules,” in
Karl Schmidt-Hebbel and Carl Walsh, eds., Monetary Policy under Uncertainty
and Learning, Santiago: Central Bank of Chile, 2009, pp. 145–170.
Gaus, Eric, “Robust Stability of Monetary Policy Rules under Adaptive Learning,”
Southern Economic Journal, January 2013, pp. 1–29.
Kostyshyna, Olena, “Application of an Adaptive Step-Size Algorithm in Models of
Hyperinflation,” Macroeconomic Dynamics, November 2012, 16, 355–375.
Marcet, Albert and Juan P Nicolini, “Recurrent hyperinflations and learning,”
American Economic Review, 2003, 93 (5), 1476–1498.
McGough, Bruce, “Statistical Learning with Time-Varying Parameters,” Macroeconomic Dynamics, 2003, 7 (01), 119–139.
Milani, Fabio, “Learning and time-varying macroeconomic volatility,” Manuscript,
UC-Irvine, 2007.
10
Stock, James H and Mark W Watson, “Why Has U.S. Inflation Become Harder to
Forecast?,” Journal of Money, Credit and Banking, February 2007, 39 (1), 3–33.
Woodford, Michael, Interest and prices: foundations of a theory of monetary policy,
Princeton: Princeton University Press, 2003.
11
Table 1: Relative Performance: AR(1), No Feedback
MSFE (Std.)
Deviation from RE
σω
PEG
MN
K
PEG MN
K
0.1 1.000(0.000) 1.000(0.000) 1.000(0.000)
1.000 1.000
1.000
0.2 1.000(0.000) 1.000(0.000) 1.000(0.000)
0.999 1.000
1.000
0.998 1.000
1.000
0.3 0.999(0.001) 1.000(0.000) 1.000(0.000)
0.993 1.001
1.000
0.4 0.998(0.001) 1.000(0.000) 1.000(0.000)
0.5 0.996(0.002) 1.001(0.001) 1.000(0.000)
0.990 1.003
1.000
0.987 1.005
1.000
0.6 0.995(0.002) 1.002(0.001) 1.000(0.000)
0.7 0.993(0.002) 1.005(0.002) 1.000(0.000)
0.979 1.016
1.000
0.8 0.993(0.003) 1.012(0.003) 1.170(0.021)
0.986 1.032
1.469
0.9 0.992(0.003) 1.021(0.004) 1.000(0.000)
0.981 1.073
1.000
0.969 1.115
1.003
1.0 0.991(0.004) 1.039(0.006) 1.000(0.001)
Values reported in the first panel are the MSFE relative to the optimal constant
gain benchmark with the standard deviations in parenthesis. The second panel
displays the squared deviations from the RE solution relative to a constant gain.
PEG stands for proposed the endogenous gain, MN stands for the Marcet and
Nicolini algorithm, and K represents the Adaptive Step-Size algorithm. Note that
for the MN algorithm is optimized over v, j and the γ̄, the Adaptive Step-Size
algorithm optimized over the µ parameter and PEG is optimized over w, γ̄lb and
γ̄sf .
Figure 1: The Nash equilibrium constant-gain value. στ = 5, ση = 2, and ε = 0.01.
Note: The red line is the 45-degree line and the black line is the optimal constant
gain response (based on MSFE) to the data generating gain value.
12
Table 2: Relative MSFE Performance: Structural Break, Without Feedback
ε
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
1.000
1.000
1.000
1.000
1.000
1.000
1.000
1.000
PEG
(0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000)
1.000
0.999
0.999
0.999
1.000
0.999
1.033
1.154
1.25 MN
(0.000) (0.000) (0.001) (0.000) (0.000) (0.002) (0.009) (0.016)
1.268
1.005
1.153
1.108
1.209
1.157
1.000
1.006
K
(0.013) (0.001) (0.010) (0.009) (0.013) (0.012) (0.000) (0.002)
1.008
1.000
1.000
1.000
1.000
1.000
1.000
1.000
PEG
(0.001) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000)
1.009
1.000
1.000
0.999
0.999
1.015
1.181
1.381
2.5 MN
(0.001) (0.000) (0.000) (0.001) (0.001) (0.014) (0.025) (0.023)
1.010
1.198
1.197
1.195
1.115
1.060
1.000
1.000
K
(0.004) (0.013) (0.012) (0.014) (0.010) (0.008) (0.001) (0.000)
1.000
1.000
1.000
1.000
1.000
1.000
1.000
0.999
PEG
(0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.001)
0.999
1.000
0.999
0.999
1.000
1.084
1.291
1.402
5 MN
(0.000) (0.000) (0.001) (0.001) (0.001) (0.021) (0.024) (0.025)
1.179
1.171
1.154
1.162
1.087
1.000
1.002
1.000
K
(0.013) (0.011) (0.012) (0.012) (0.010) (0.000) (0.002) (0.001)
1.000
1.000
1.000
1.000
1.000
1.000
0.999
0.996
PEG
(0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.001) (0.003)
0.998
0.996
1.000
1.000
1.047
1.153
1.238
1.337
10 MN
(0.001) (0.001) (0.000) (0.001) (0.020) (0.017) (0.021) (0.026)
1.026
1.171
1.182
1.035
1.001
1.000
1.001
1.000
K
(0.004) (0.012) (0.013) (0.005) (0.001) (0.001) (0.001) (0.001)
1.000
1.001
1.000
1.000
1.000
1.000
0.998
0.995
PEG
(0.000) (0.000) (0.000) (0.000) (0.000) (0.001) (0.005) (0.004)
1.000
0.996
1.000
0.999
1.039
1.115
1.150
1.252
20 MN
(0.000) (0.001) (0.000) (0.001) (0.013) (0.015) (0.015) (0.021)
1.124
1.183
1.116
1.101
1.000
1.000
1.015
1.001
K
(0.010) (0.012) (0.010) (0.010) (0.000) (0.000) (0.005) (0.002)
Values reported are the MSFE relative to the optimal constant gain benchmark with the
standard deviations in parenthesis. PEG stands for proposed the endogenous gain, MN
stands for the Marcet and Nicolini algorithm, and K represents the Adaptive Step-Size
algorithm. Note that for the MN algorithm is optimized over v, j and the γ̄, the Adaptive
Step-Size algorithm optimized over the µ parameter and PEG is optimized over w, γ̄lb and
γ̄sf .
στ /ση
13
Table 3: Relative MSFE Performance: Structural Break, Without Feedback
ε
0.1
0.09
0.08
0.07
0.06
0.05
0.04
0.03
0.02
0.01
στ /ση
1.000
1.000
1.000
1.000
1.000
1.000
1.000
0.999
0.998
0.996
PEG
(0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.001) (0.002) (0.002)
1.226
1.216
1.169
1.165
1.117
1.083
1.044
1.016
0.998
0.997
1.25 MN
(0.013) (0.012) (0.010) (0.011) (0.009) (0.008) (0.006) (0.005) (0.003) (0.001)
1.000
1.000
1.000
1.025
1.006
1.000
1.000
1.000
1.000
1.014
K
(0.000) (0.000) (0.000) (0.005) (0.002) (0.000) (0.000) (0.000) (0.000) (0.004)
0.997
0.996
0.996
0.995
0.995
0.994
0.993
0.990
0.987
0.983
PEG
(0.003) (0.002) (0.003) (0.003) (0.002) (0.003) (0.002) (0.005) (0.004) (0.005)
1.360
1.331
1.327
1.282
1.253
1.177
1.121
1.060
1.005
0.993
2.5 MN
(0.018) (0.021) (0.020) (0.019) (0.017) (0.017) (0.013) (0.013) (0.008) (0.004)
1.000
1.000
1.000
1.002
1.000
1.000
1.000
1.002
1.000
1.000
K
(0.000) (0.000) (0.001) (0.002) (0.000) (0.000) (0.000) (0.002) (0.001) (0.001)
0.993
0.992
0.987
0.986
0.983
0.981
0.979
0.976
0.971
0.966
PEG
(0.003) (0.005) (0.005) (0.005) (0.006) (0.006) (0.006) (0.009) (0.009) (0.012)
1.432
1.387
1.380
1.363
1.329
1.285
1.233
1.143
1.082
1.001
5 MN
(0.028) (0.030) (0.025) (0.027) (0.029) (0.025) (0.024) (0.022) (0.023) (0.011)
1.000
1.000
1.001
1.039
1.000
1.000
1.000
1.001
1.001
1.001
K
(0.000) (0.000) (0.002) (0.011) (0.000) (0.000) (0.001) (0.001) (0.001) (0.003)
0.984
0.983
0.982
0.981
0.977
0.976
0.971
0.972
0.960
0.952
PEG
(0.006) (0.007) (0.007) (0.007) (0.008) (0.010) (0.010) (0.015) (0.015) (0.015)
1.436
1.423
1.424
1.420
1.411
1.410
1.355
1.315
1.214
1.045
10 MN
(0.032) (0.034) (0.034) (0.037) (0.036) (0.039) (0.040) (0.041) (0.040) (0.024)
1.000
1.000
1.000
1.000
1.001
1.000
1.000
1.000
1.000
1.002
K
(0.000) (0.001) (0.001) (0.000) (0.002) (0.001) (0.000) (0.001) (0.000) (0.004)
0.987
0.986
0.983
0.983
0.979
0.976
0.972
0.972
0.959
0.950
PEG
(0.007) (0.008) (0.008) (0.009) (0.010) (0.010) (0.017) (0.013) (0.017) (0.018)
1.416
1.446
1.426
1.436
1.436
1.472
1.478
1.437
1.353
1.107
20 MN
(0.035) (0.040) (0.041) (0.040) (0.047) (0.048) (0.061) (0.057) (0.060) (0.043)
1.000
1.007
1.000
1.000
1.000
1.000
1.000
1.000
1.000
1.000
K
(0.000) (0.005) (0.000) (0.000) (0.001) (0.001) (0.000) (0.001) (0.002) (0.001)
Values reported are the MSFE relative to the optimal constant gain benchmark with the standard deviations in
parenthesis. PEG stands for proposed the endogenous gain, MN stands for the Marcet and Nicolini algorithm,
and K represents the Adaptive Step-Size algorithm. Note that for the MN algorithm is optimized over v, j and
the γ̄, the Adaptive Step-Size algorithm optimized over the µ parameter and PEG is optimized over w, γ̄lb and
γ̄sf .
14
Table 4: Relative deviations from the RE solution: Structural Break, Without Feedback
ε
0.1
0.09
0.08
0.07
0.06
0.05
0.04
0.03
0.02
0.01
PEG 1.000 1.000 1.000 0.999 0.998 0.998 1.000 0.992 0.988 0.972
1.25 MN 1.874 1.851 1.648 1.677 1.469 1.337 1.195 1.060 0.970 0.983
K
1.000 1.000 1.000 1.113 1.018 1.000 1.000 1.000 1.000 1.072
PEG 0.983 0.978 0.982 0.977 0.975 0.971 0.983 0.977 0.952 0.904
2.5 MN 2.086 2.006 1.953 1.940 1.777 1.584 1.362 1.158 0.974 0.963
K
1.000 1.000 1.002 1.001 1.000 1.001 1.000 1.000 0.999 1.007
PEG 0.951 0.952 0.955 0.937 0.943 0.935 0.959 0.942 0.965 0.923
5 MN 2.438 2.324 2.209 2.016 1.973 1.793 1.701 1.462 1.262 1.018
K
1.000 1.000 1.003 1.124 1.000 1.000 1.000 0.998 1.001 1.009
PEG 0.902 0.882 0.903 0.863 0.897 0.869 0.854 0.921 0.906 0.945
10 MN 2.944 2.958 2.917 2.787 2.605 2.572 2.141 2.054 1.553 1.097
K
1.000 0.998 1.000 1.000 1.002 1.000 1.000 0.999 1.000 1.005
PEG 0.872 0.878 0.915 0.873 0.873 0.845 0.837 0.769 0.766 0.937
20 MN 4.010 4.033 3.861 3.610 3.323 3.467 3.487 3.119 2.493 1.368
K
1.000 1.050 0.999 1.000 1.001 0.999 1.000 1.003 0.998 1.000
Values reported are the squared deviations from the RE solution relative to a constant gain.
PEG stands for proposed the endogenous gain, MN stands for the Marcet and Nicolini algorithm, and K represents the Adaptive Step-Size algorithm. Note that for the MN algorithm
is optimized over v, j and the γ̄, the Adaptive Step-Size algorithm optimized over the µ
parameter and PEG is optimized over w, γ̄lb and γ̄sf .
στ /ση
Figure 2: The MSFE of best responses of constant and endogenous gains to particular
constant gain values. στ = 5, ση = 2, and ε = 0.01.
Note: The red line indicates the errors within the data generated by the gain
value on the x-axis. The dashed line represents the MSFE of the optimal constant
gain response and the solid black line represents the MSFE of the optimal PEG
response.
15
Table 5: Relative MSFE Performance: Structural Break, With Feedback
ε
0.1
0.09
0.08
0.07
0.06
0.05
0.04
0.03
0.02
0.01
στ /ση
0.999
0.999
0.998
0.997
0.996
0.996
0.995
0.994
0.992
0.992
PEG
(0.001) (0.001) (0.001) (0.001) (0.002) (0.001) (0.002) (0.002) (0.002) (0.002)
1.398
1.434
1.476
1.528
1.587
1.646
1.727
1.816
1.945
2.122
1.25 MN
(0.030) (0.029) (0.033) (0.039) (0.045) (0.048) (0.054) (0.064) (0.089) (0.138)
1.529
1.588
1.695
1.722
1.851
1.781
1.672
1.889
1.000
1.873
K
(0.148) (0.213) (0.173) (0.231) (1.461) (0.198) (0.186) (0.313) (0.000) (0.303)
0.988
0.987
0.987
0.983
0.982
0.980
0.978
0.975
0.974
0.973
PEG
(0.003) (0.003) (0.003) (0.004) (0.004) (0.004) (0.004) (0.005) (0.006) (0.006)
2.045
2.149
2.270
2.417
2.602
2.804
3.093
3.458
3.993
4.835
2.5 MN
(0.056) (0.068) (0.072) (0.085) (0.093) (0.117) (0.157) (0.172) (0.259) (0.488)
1.000
1.485
1.000
1.483
1.501
1.608
1.610
1.589
1.725
1.826
K
(0.000) (0.105) (0.000) (0.086) (0.094) (0.549) (0.194) (0.144) (0.386) (0.206)
0.978
0.976
0.973
0.971
0.967
0.965
0.958
0.953
0.947
0.945
PEG
(0.006) (0.006) (0.005) (0.006) (0.007) (0.008) (0.007) (0.009) (0.009) (0.012)
2.779
2.977
3.222
3.518
3.874
4.389
5.064
6.076
7.606
10.750
5 MN
(0.106) (0.118) (0.128) (0.166) (0.208) (0.215) (0.295) (0.353) (0.564) (1.109)
1.457
1.430
1.438
1.467
1.462
1.533
1.518
1.558
1.595
1.611
K
(0.048) (0.067) (0.067) (0.072) (0.051) (0.054) (0.074) (0.072) (0.140) (0.184)
0.979
0.975
0.973
0.970
0.967
0.961
0.958
0.952
0.940
0.925
PEG
(0.005) (0.006) (0.008) (0.010) (0.008) (0.010) (0.009) (0.011) (0.012) (0.013)
3.322
3.590
3.961
4.396
4.959
5.732
6.945
8.568
11.780 19.304
10 MN
(0.141) (0.158) (0.174) (0.213) (0.248) (0.338) (0.468) (0.625) (1.106) (2.195)
1.562
1.464
1.606
1.499
1.491
1.499
1.687
1.667
1.656
1.652
K
(0.040) (0.043) (0.044) (0.041) (0.040) (0.049) (0.061) (0.089) (0.088) (0.572)
0.981
0.979
0.980
0.976
0.973
0.971
0.966
0.967
0.949
0.932
PEG
(0.018) (0.013) (0.007) (0.010) (0.030) (0.009) (0.014) (0.012) (0.012) (0.017)
3.661
3.996
4.406
4.928
5.677
6.661
8.034
10.376 14.719 26.505
20 MN
(0.171) (0.198) (0.211) (0.257) (0.351) (0.418) (0.586) (0.885) (1.275) (3.320)
1.612
1.563
1.553
1.592
1.697
1.798
1.631
1.627
1.639
1.764
K
(0.045) (0.043) (0.046) (0.044) (0.074) (0.068) (0.064) (0.066) (0.079) (0.127)
Values reported are the MSFE relative to the optimal constant gain benchmark with the standard deviations in
parenthesis. PEG stands for proposed the endogenous gain, MN stands for the Marcet and Nicolini algorithm,
and K represents the Adaptive Step-Size algorithm. Note that for the MN algorithm is optimized over v, j and
the γ̄, the Adaptive Step-Size algorithm optimized over the µ parameter and PEG is optimized over w, γ̄lb and
γ̄sf .
16
Table 6: Relative deviations from the RE solution: Structural Break, With Feedback
ε
0.1
0.09
0.08
0.07
0.06
0.05
0.04
0.03
0.02
0.01
PEG 1.000 1.000 1.000 0.999 0.998 0.998 1.000 0.992 0.988 0.972
1.25 MN 1.874 1.851 1.648 1.677 1.469 1.337 1.195 1.060 0.970 0.983
K
1.000 1.000 1.000 1.113 1.018 1.000 1.000 1.000 1.000 1.072
PEG 0.951 0.957 0.955 0.941 0.951 0.939 0.944 0.930 0.936 0.884
2.5 MN 4.125 4.344 4.597 4.618 5.644 6.228 7.014 7.786 9.973 13.72
K
1.000 2.046 1.000 2.539 2.404 2.453 3.489 2.403 3.203 3.392
PEG 0.921 0.925 0.926 0.909 0.926 0.915 0.907 0.903 0.911 0.874
5 MN 6.093 6.403 7.163 7.627 7.917 9.347 10.96 12.77 17.62 25.19
K
2.234 2.104 2.110 2.011 2.209 2.261 2.239 2.141 2.257 2.581
PEG 0.917 0.903 0.915 0.909 0.908 0.895 0.894 0.881 0.876 0.853
10 MN 7.868 8.946 10.32 10.34 11.83 13.97 14.92 18.03 26.09 33.61
K
2.783 2.527 2.820 2.442 2.326 2.318 2.522 2.454 2.494 2.489
PEG 0.927 0.948 0.936 0.926 0.899 0.898 0.911 0.883 0.846 0.899
20 MN 10.91 11.66 12.46 14.26 14.11 18.14 22.50 28.37 38.51 65.68
K
3.196 2.981 3.040 2.798 2.825 3.264 3.059 3.012 2.384 2.841
Values reported are the squared deviations from the RE solution relative to a constant gain.
PEG stands for proposed the endogenous gain, MN stands for the Marcet and Nicolini algorithm, and K represents the Adaptive Step-Size algorithm. Note that for the MN algorithm
is optimized over v, j and the γ̄, the Adaptive Step-Size algorithm optimized over the µ
parameter and PEG is optimized over w, γ̄lb and γ̄sf .
στ /ση
Figure 3: Recurring hyperinflation episodes
17
Figure 4: Economic fluctuations of an endogenous gain.
18
Download