In the wake of St. Louis Fed President James Bullard’s statement on the output gap [0] (which I frankly did not understand), there was renewed debate over output gap measurement [Duy] [Thoma] [Krugman]. I thought this was a good time to recap and update some of the material I’d written on this subject of output gaps.
DSGEs, Detrending, and Forecasting the Output Level
As somebody who has served on many dissertation committees where the dissertation involves cutting edge DSGEs (dynamic stochastic general equilibrium models), I can attest to the fact that such models can be very useful in providing insights into the workings of the macroeconomy, as well as the welfare implications associated differing policy regimes.
However, I think the observation that DSGE’s are at the cutting edge of academic research highlights several misconceptions, and one important drawback of DSGEs. (An excellent review of the use of DSGEs in policy is provided by C. Tovar)
Misconceptions
Regarding the treatment of expectations, DSGEs usually incorporate model consistent expectations. However, ever since John Taylor’s pathbreaking work in the early 1990s [0], we have had model-consistent expectations imbedded in certain structural macroeconometric models. Hence, a DSGE is not necessary for operationalizing rational expectations.
Regarding microfoundations, if one examines the guts of the standard New Keynesian versions (including the one recently used by John Taylor [1]), one usually finds lots of ad hoc additions. Consumption is definitely not described by a simple Euler equation as implied by the pure rational expectations-life cycle hypothesis; usually there are some hand-to-mouth consumers floating around. Prices are not freely flexible; rather Calvo pricing is often assumed for tractability. Capital adjustment costs, and other frictions are often included as well. Why not leave these frictions out? Because, without them, it is well nigh impossible to replicate the impulse response functions of real world data. In other words, the bright line of microfoundations versus ad hoc functions is in fact pretty fuzzy.
(And from an international finance perspective, it’s troubling that the real exchange rate is usually linked one-for-one with the ratio of the marginal utilities of consumption, something that is as counterfactual as one can get. And don’t get me started on the risk premium gets introduced into these models, if indeed there is one.)
The Big Drawback
DSGEs (and their predecessors, RBCs) are models of the business cycle. As such, they focus on the deviations from trend. However, in order to predict where the economy will be in one year, given current conditions and policies, one needs to know what the trend is. In other words, extracting the cycle from the trend is critically important. This is a point that James Morley made in his “Emperor has no clothes” paper.
This issue has long been recognized in the policy community. From Camilo Tovar:
Econometricians often fail to be able to observe the theoretical concepts modeled (eg the output
gap). So a first question is: how to match the theoretical concepts in DSGE models with those of
the observed data? This is not trivial (and certainly the problem is not exclusive to these models).
In the DSGE literature the theoretical concepts have been captured not against specific data
figures (say GDP levels or inflation) but against filtered data (eg by using the Hodrick-Prescott
filter). Filtering decomposes the data into a cyclical component and a trend component. The
cyclical component is what is frequently fed into the model. By doing so the analysis focuses
on the business cycle frequencies, mainly because it is considered that DSGE models are
better suited to explain short-run rather than long-run cycles. However, filtering has important
implications (see discussion in Del Negro and Schorfheide (2003)). One is that, forecasting
stays out of the reach of DSGE models since the goal is to forecast actual rather than filtered
data. The second is that the dynamics obtained do not match those required by policy makers,
weakening the usefulness of DSGE models as policy tools. Alternatives often used are the
linear detrending and demeaning of variables, as well, as transforming variables so that they
are stationary around a balanced growth path. In this last case, the problem is that the trend is
often assumed to follow an exogenous stochastic process, which is imposed by the researcher.
In order to highlight the real-world complications involved in this issue, consider two popular cycle-trend extraction methods used in the “business”: the Hodrick-Prescott (HP) filter, and band pass (BP) filter. I apply the HP and BP filters over the 1967Q1-11Q4 sample, and HP filter to the 1967Q1-09Q1 sample, and plot the resulting cycle components in Figure 1.
Figure 1: Log deviation from CBO measure of potential GDP (blue), log deviation from trend GDP, obtained using HP filter over entire sample, lambda=1600, estimated over 1967Q1-2011Q4 (dark red), using HP filter over sample extended 8 quarters (green), over sample ending 2009Q1 (purple), band pass filter with Baxter-King symmetric fixed length = 4 qtrs (orange). NBER defined recession dates shaded gray. Source: BEA 2011Q4 advance release, CBO, Budget and Economic Outlook (Jan 2012) supplemental data, NBER, and author’s calculations.
In order to mitigate the end-point problem associated with using a two-sided filter like the Hodrick-Prescott filter, the log GDP series are extended 8 quarters using an ARIMA(1,1,1), and then the filter applied (green line).
Figure 1 demonstrates how the inferences regarding the output gap differ wildly depending on the filter used; and the sensitivity of the results — even for a given filter — to the endpoints (notice how the HP filter gives different output gaps for 2009Q1 depending on the sample).
Appealing to simple theories does not necessarily yield uncontroversial results. In Figure 2, I plot the gap implied if one thinks GDP and consumption (here services only) are cointegrated. That approach, outlined by Cogley and Schaan (1995) following Cochrane (1994), implies that the economy was 2.7% above trend in 2011Q4.
For comparison, I include the CBO’s estimate from January 2012. Large scale macroeconometric models usually rely upon some sort of production function approach to calculating the trend. The implied output gaps from the consumption-cointegration approach and the CBO’s version of the production function approach (described here) are displayed in Figure 2.
Figure 2: Log deviation from trend GDP, obtained using Engle-Granger cointegrating regression between log services consumption and log GDP (red), and log deviation from CBO’s potential GDP, January 2011 version (blue). NBER defined recession dates shaded gray. Source: BEA 2011Q4 advance release, CBO, Budget and Economic Outlook (Jan 2012) supplemental data, NBER, and author’s calculations.
For more on the scary things the HP filter can do, see T. Cogley, and J. Nason, 1995, “Effects of the Hodrick-Prescott filter on trend and difference stationary time series Implications for business cycle research,” Journal of Economic Dynamics and Control 19(1-2): 253-278. See also Simon van Norden on the use of these types of filters in current analysis. And for an excellent, albeit technical, exposition of the issues Gorodnichenko and Ng.
What Is It We Are Trying to Measure?
In his response to Duy, Bullard appeals to New Keynesian and two sector models of natural output and gaps, in a St. Louis Fed symposium on potential output (see this post for a discussion of the papers).
I found insights into this debate in this 2010 paper by Michael Kiley. From the paper:
What is the output gap? There are many definitions in the economics literature, all of which have a long history. I discuss three alternatives: the deviation of output from its long-run stochastic trend (i.e., the “Beveridge-Nelson cycle”);
the deviation of output from the level consistent with current technologies and
normal utilization of capital and labor input (i.e., the “production-function
approach”); and the deviation of output from “flexible-price” output (i.e., its
“natural rate”). Estimates of each concept are presented from a dynamic-
stochastic-general-equilibrium (DSGE) model of the U.S. economy used at the
Federal Reserve Board. Four points are emphasized: The DSGE model’s es
timate of the Beveridge-Nelson gap is very similar to gaps from policy institutions, but the DSGE model’s estimate of potential growth has a higher
variance and substantially different covariance with GDP growth; the natural
rate concept depends strongly on model assumptions and is not designed to
guide nominal interest rate movements in “Taylor” rules in the same way as
the other measures; the natural rate and production function trends converge to
the Beveridge-Nelson trend; and the DSGE model’s estimate of the Beveridge-
Nelson gap is as closely related to unemployment fluctuations as those from
policy institutions and has more predictive ability for inflation.
The output gaps are shown in three graphs from the paper.
Figure 3 from Kiley (2010).
Figure 4 from Kiley (2010).
Figure 6 from Kiley (2010).
Based upon his analysis, Kiley makes the following four conclusions:
- The EDO model’s estimate of the output gap (according to either a Beveridge-Nelson or production-function approach) is very similar to gaps from the Congressional
Budget Office or the Federal Reserve’s large-scale macro-econometric model (FRB/US) model, but the DSGE model’s estimate of potential growth is considerably more variable; the latter result stems from the significant degree of fluctuation in aggregate technology estimated by the DSGE model, a result consistent with the significant role such fluctuations play in model’s descended from those of the real-business-cycle tradition (from Kydland and Prescott (1982)).
- The flexible-price/natural-rate gaps are highly dependent on modeling assumptions,
and their use in policy applications or forecasting requires a deep understanding
of a specific model’s structure. (This result is closely related to the
critique of DSGE models of Chari, Kehoe, and McGrattan (2009), who highlight
the sensitivity of policy applications of such models to controversial modeling
assumptions). In particular, a natural-rate gap does not provide the same type
of guidance to a “Taylor” rule for nominal interest rates as other concepts of
gaps; indeed, the signals from the Beveridge-Nelson gap provide a better sense
of movements in the “natural rate of interest” than do the signals from the
natural rate of output gap.
- “Equilibrium” or trend expected growth is highly variable in the flex-price/naturalrate
case, implying that a focus on the current level of such gaps can be misleading
in a policy discussion. In contrast, expected trend growth for the Beveridge-
Nelson concept is exogenous and constant; moreover, all other notions of “trend”
converge to the Beveridge-Nelson trend.
- The DSGE model’s estimate of the Beveridge-Nelson gap is as closely related to
unemployment fluctuations as those from policy institutions (e.g., obeys Okun’s
law) and has more predictive ability for inflation (e.g., has a tighter reduced form
Phillips curve relationship).
For me, one important observation is that the output gap for both the production function based approach from EDO and the CBO estimate are of similar magnitude; 6 ppts and 7.5 ppts of GDP at 09Q2. The natural rate output gap is quite different, and this highlights the fact that this gap embodies a very different concept (discussed at length in the paper). I stress this, exactly because I know somebody is going to take some output gap measure and mis-interpret the implications of that gap — especially as concerns of inflation rise.
Concluding Thoughts
How do these observations relate to Bullard’s speculation? Here are my key take-aways:
- The observation that inflation has been higher than what is anticipated from traditional output gap measures is consistent with the natural rate output gap, rather than the deviation from potential GDP usually used. I wonder how a depeciating dollar and higher input prices fit into this story.
- Output gaps from DSGEs sometimes look different from production function based gaps (e.g., CBO), and sometimes don’t. It’s hard to generalize because there is considerable heterogeneity.
- When the DSGE output gaps are derived on series that have been detrended, one has to wonder abou the biases induced through the detrending process; it also complicates the forecasting of output levels, which is sometimes of primary interest.
A good review of issues is provided by the OECD.
The Bullard paper is just … odd. He says that we should view the recession as due to a large negative wealth shock, which I agree with. He then treats reduced output as an inevitable consequence, without any discussion of the mechanism (increased savings desired by the household sector), which prevents him from seeing the actual solution (increased borrowing by the govt sector). And goes on to thoroughly entangle stock variables and flow variables — comparing a bubble in the NASDAQ level with a “bubble” in GDP.
I can’t even figure out if he believes in aggregate demand. He translates the wealth effect into reduced output without invoking AD, and suggests that we should simply accept reduced output and high unemployment for a long time to come.
“In the wealth shock view, there is no “large output gap” rationale for keeping interest rates near zero.”
Um, why not? Does the origin of the gap mean that millions aren’t unemployed and underemployed? That there aren’t 4 job seekers for every opening?
He then closes with the converse of “expansionary austerity” — namely the hand-waving (I won’t call it an argument, as there’s no data and no model and no references to either) that low interest rates may actually be reducing aggregate demand, because savers are losing income while would-be borrowers are unemployed. So I guess he does believe in aggregate demand, when it’s a hand-waving for tighter monetary policy….
You left out the Kalman Filter approaches that connect inflation to the output gap.
John Hall: Yes, good point regarding the multivariate Kalman Filter. See paragraph 41 in the OECD document referenced at the end of the post.
As standing now the output gap, is the happiness of pursuit as opposed to the pursuit of happiness.
The President of the Federal Reserve bank of Saint Louis is staying within the confines of realities, the founding principles of economics and the declaration of independence of the USA.
The ZLB and depleting savings may be on the agenda,reviving economies meeting the non tested ability of Central Banks to mop up orderly the excess liquidities will be on the agenda.Economies are leveraged and so are the private and the corporations,income ,revenues are needed.
Evidences are shown that liquidities and solvencies are independent in adversities.The CBO report is making provision for the random knowledge and uncertainties surrounding capital productivity and consequently,the output of forecasters.
“Another problem is that in order to estimate the coefficients statistically, the
capital input must be adjusted (over time) to reflect how intensively the capital stock
is used. That adjustment introduces a source of measurement error because of the
difficulty in accurately measuring the rate at which the capital stock is used. Since
most analysts who employ the econometric approach check to see whether their
estimates are reasonable by comparing them with income shares, the payoff to using
the econometric approach is small.”
Throughout the last decade, potential capital gains were supplying a better benchmark of capital utility than a mere return on owned funds,incomes,revenues or capital stocks.
The macro economics assessment will be much easier when banks, financial corporation,central banks, political powers will perform their functions independently.The statistical filters could be all useful.
If anyone else is confused, “Brian” appears in this post:
https://econbrowser.com/archives/2011/07/dsges_detrendin.html
This is obviously the (or at least a) post Menzie is recapping and updating. I suspect he meant to paste in a link, but if so it didn’t paste correctly.
Andrew Burday: Yes, thanks — I saw that and edited the text, so no need to reference that blogpost, unless you are interested in Brian’s original comment.
In my comment on Econospeak on this matter, I pointed out one issue that might play toward making Bullard’s general argument more correct, if not in its details or its justification (which Menzie has laid out quite well), is the matter that a period of low GDP usually accompanies low real capital investment, which in turn implies a lower future capital stock than would otherwise be there, which means a lower potential output, however one wants to measure it.
Nice post, Menzie. I’ll add one observation that might be of interest.
Much of the modeling of potential output relies on understanding both the level and the expected growth rate of productivity. However, many modelers fail to realize how poorly we measure aggregate productivity and its growth trends. US productivity data are notorious for undergoing substantial revisions, often years after initial publication, that have profound effects on the estimated levels and trend growth rates. (Paper at http://www.cirano.qc.ca/pdf/publication/2010s-46.pdf)
To understand the extent of the problem, consider Kahn and Rich’s work at the FRB NY. They give policy makers the probability that the economy is in a High or Low productivity growth state; the Fed keeps their model updated at http://www.newyorkfed.org/research/national_economy/richkahn_prodmod.pdf. Chart 3 keeps track of how their results change over time. I’m always struck by how volatile their assessments are. For example, early estimates for 2008 put the probability of low growth at well over 90%, but within a year that had fallen to about 10%, only to later be revised to over 90%. Their latest estimates put the probability of low productivity growth at nearly 100%. But based on past performance, that doesn’t mean that their model won’t change its mind.
Given such high and persistent uncertainty about the most essential building block of potential output, its not surprising that estimates of the output gap are imprecise.
I bought Taylor’s 1993 book awhile back based on Menzie’s earlier post on this topic because I was interested in how to incorporate rational expectations in structural models. As more of a computer programmer on the job front, my main interest is in how one actually goes about solving these models to get real numerical results. In that vein, Judd’s numerical methods book also has a good chapter on solving rational expectations models.
What really surprised me, once I started thinking about rational expectations from a computational standpoint, is just how strong an assumption it is to think that a representative consumer can somehow behave in such a manner. My interpretation is that RE requires that the agents actually know what the equations of the model are and at any point in time can actually work through the model enough periods into the future so that they can figure out the “optimal” decision at that point in time. Relying solely on anecdotes, I know a lot of people who have trouble planning more than a couple of months into the future.
I’m most familiar with Mark Zandi’s structural macroeconometric model, and in models like his, consumers tend to spend a large chunk of additional income from, say, a temporary payroll tax cut. Thinking of concepts like consumption smoothing, the permanent income hypothesis, etc., it seems to me that consumers would in fact just save that extra income. That’s what got me thinking about structural models with rational expectations, but the more I think about what RE really implies, the more I realize it’s not a panacea. My prediction intervals are rather wide (bad joke).
“As somebody who has served on many dissertation committees where the dissertation involves cutting edge DSGEs (dynamic stochastic general equilibrium models), I can attest to the fact that such models can be very useful in providing insights into the workings of the macroeconomy, as well as the welfare implications associated differing policy regimes.”
Really? Did these “very useful insights” come before or after the fact of the “workings of the macroeconomy” they were so useful in explaining? If so, we’d all appreciate hearing of the predictive power of these cutting-edge models — in the absence of which I would assume you mean that they were impressively “logical” or “novel” or any other adjective that can be applied to something not subject to factual verification.
P.S. You might also share them with the Commerce Department, which seems to be revising GDP estimates with large margins of error, despite their own cutting edge models.
jcb: You are under the misapprehension that most of what we economists do has to do with forecasting. Macroeconomic forecasting is a field within economics, and atheoretical models can be very useful for forecasting, while not useful at all at teasing out welfare implications.
Brian I’m no particular fan of strong version RE models, but I believe the usual assumption is that only the marginal players need to behave rationally because only decisions at the margin have economic impact.