15 Multivariate Time Series
15.1 Introduction
A multivariate time series
Some excellent textbooks and review articles on multivariate time series include Hamilton (1994), Watson (1994), Canova (1995), Lütkepohl (2005), Ramey (2016), Stock and Watson (2016), and Kilian and Lütkepohl (2017).
15.2 Multiple Equation Time Series Models
To motivate vector autoregressions let us start by reviewing the autoregressive distributed lag model of Section
Similarly, an AR-DL model for
These two equations specify that each variable is a linear function of its own lag and the lag of the other variable. In so doing we find that the variables on the right hand side of each equation are
We can simplify the equations by combining the regressors stacking the two equations together and writing the vector error as
where
The above derivation assumed a single lag. If the equations include
Furthermore, there is nothing special about the two variable case. The notation in (15.1) allows
The error
The VAR model falls in the class of multivariate regression models studied in Chapter 11.
In the following several sections we take a step back and provide a rigorous foundation for vector autoregressions for stationary time series.
15.3 Linear Projection
In Section
We will call the projection errors
The innovations
Theorem 15.1 If
The innovations
The uncorrelatedness of the projection errors is a property of a multivariate white noise process.
Definition 15.1 The vector process
15.4 Multivariate Wold Decomposition
By projecting
Theorem 15.2 If
where
We can write the moving average representation using the lag operator notation as
where
A multivariate version of Theorem
Theorem 15.3 If
where
and the coefficients satisfy
For a proof see Section 2 of Meyer and Kreiss (2015).
We can also provide an analog of Theorem 14.6.
Theorem 15.4 If
The moving average and autoregressive lag polynomials satisfy the relationship
For some purposes (such as impulse response calculations) we need to calculate the moving average coefficient matrices
Theorem 15.5 For
To see this, suppose for simplicity
where
where
where
where
15.5 Impulse Response
One of the most important concepts in applied multivariate time series is the impulse response function (IRF) which is defined as the change in
There are
Recall the multivariate Wold representation
We can calculate that the projection onto the history at time
We deduce that the impulse response matrix is
Here we have defined the impulse response in terms of the linear projection operator. An alternative is to define the impulse response in terms of the conditional expectation operator. The two coincide when the innovations
Typically we view impulse responses as a function of the horizon
In a linear vector autoregression the impulse response function is symmetric in negative and positive innovations. That is, the impact on
The impulse response functions can be scaled as desired. One standard choice is to scale so that the innovations correspond to one unit of the impulse variable. Thus if the impulse variable is measured in dollars the impulse response can be scaled to correspond to a change in
Closely related to the IRF is the cumulative impulse response function (CIRF) defined as
The cumulative impulse response is the accumulated (summed) responses on
This is the full (summed) effect of the innovation over all time.
It is useful to observe that when a VAR is estimated on differenced observations
which is the impulse response for the variable
15.6 VAR(1) Model
The first-order vector autoregressive process, denoted VAR(1), is
where
We are interested in conditions under which
Theorem 15.6 If
The proof is given in Section 15.31.
15.7 Model
The
where
We can write the model using the lag operator notation as
where
The condition for stationarity of the system can be expressed as a restriction on the roots of the determinantal equation of the autoregressive polynomial. Recall, a
Theorem
The proof is structurally identical to that of Theorem
15.8 Regression Notation
Define the
and the
This is a multivariate regression model. The error has covariance matrix
We can also write the coefficient matrix as
In general, if
This holds whether or not
The projection coefficient matrix
Theorem
The proof is given in Section 15.31.
15.9 Estimation
From Chapter 11 the systems estimator of a multivariate regression is least squares. The estimator can be written as
Alternatively, the coefficient estimator for the
The least squares residual vector is
(This may be adjusted for degrees-of-freedom if desired, but there is no established finite-sample justification for a specific adjustment.)
If
and
Since the latter is positive definite by Theorem
Theorem 15.9 If
VAR models can be estimated in Stata using the var command.
15.10 Asymptotic Distribution
Set
By the same analysis as in Theorem
Theorem 15.10 Suppose that
Notice that the theorem uses the strong assumption that the innovation is a martingale difference sequence
If we further strengthen the MDS assumption to conditional homoskedasticity
then the asymptotic variance simplifies as
In contrast, if the VAR(p) is an approximation then the MDS assumption is not appropriate. In this case the asymptotic distribution can be derived under mixing conditions.
Theorem 15.11 Assume that
This theorem does not require that the true process is a VAR. Instead, the coefficients are defined as those which produce the best (mean square) approximation, and the only requirements on the true process are general dependence conditions. The theorem shows that the coefficient estimators are asymptotically normal with a covariance matrix which takes a “long-run” sandwich form.
15.11 Covariance Matrix Estimation
The classic homoskedastic estimator of the covariance matrix for
Estimators adjusted for degree-of-freedom can also be used though there is no established finite-sample justification. This variance estimator is appropriate under the assumption that the conditional expectation is correctly specified as a
The heteroskedasticity-robust estimator equals
This variance estimator is appropriate under the assumption that the conditional expectation is correctly specified as a
The number
Traditional textbooks have only used the homoskedastic variance estimation formula (15.9) and consequently existing software follows the same convention. For example, the var command in Stata displays only homoskedastic standard errors. Some researchers use the heteroskedasticity-robust estimator (15.10). The Newey-West estimator (15.11) is not commonly used for VAR models.
Asymptotic approximations tend to be much less accurate under time series dependence than for independent observations. Therefore bootstrap methods are popular. In Section
15.12 Selection of Lag Length in an VAR
For a data-dependent rule to pick the lag length
where
In Stata the AIC for a set of estimated VAR models can be compared using the varsoc command. It should be noted, however, that the Stata routine actually displays
15.13 Illustration
We estimate a three-variable system which is a simplified version of a model often used to study the impact of monetary policy. The three variables are quarterly from FRED-QD: real GDP growth rate
Examining the coefficients in the table we can see that GDP displays a moderate degree of serial correlation and shows a large response to the federal funds rate, especially at lags 2 and 3. Inflation also displays serial correlation, shows minimal response to GDP, and also has meaningful response to the federal funds rate. The federal funds rate has the strongest serial correlation. Overall, it is difficult to read too much meaning into the coefficient estimates due to the complexity of the interactions. Because of this difficulty it is typical to focus on other representations of the coefficient estimates such as impulse responses which we discuss in the upcoming sections.
15.14 Predictive Regressions
In some contexts (including prediction) it is useful to consider models where the dependent variable is dated multiple periods ahead of the right-hand-side variables. These equations can be single equation or multivariate; we can consider both as special cases of a VAR (as a single equation model can be written as one equation taken from a VAR system). An
The integer
There is an interesting relationship between a VAR model and the corresponding
Theorem 15.12 If
The proof of Theorem
There are several implications of this theorem. First, if
The predictive regression (15.12) can be estimated by least squares. We can write the estimates as
For a distribution theory we need to apply Theorem
For a distributional theory we can apply Theorem 15.11. Let
Theorem 15.13 If
15.15 Impulse Response Estimation
Reporting of impulse response estimates is one of the most common applications of vector autoregressive modeling. There are several methods to estimate the impulse response function. In this section we review the most common estimator based on the estimated VAR parameters.
Within a VAR(p) model the impulse responses are determined by the VAR coefficients. We can write this mapping as
We then set
This is the the most commonly used method for impulse response estimation and it is the method implemented in standard packages.
Since
The asymptotic approximations, however, can be poor. As we discussed earlier the asymptotic approximations for the distribution of the coefficients
Consequently, asymptotic approximations are less popular than bootstrap approximations. The most popular bootstrap approximation uses the recursive bootstrap (see Section 14.46) using the fitted VAR model and calculates confidence intervals for the impulse responses with the percentile method. An unfortunate feature of this choice is that the percentile bootstrap confidence interval is biased since the nonlinear impulse response estimates are biased and the percentile bootstrap accentuates bias. Some advantages of the estimation method as described is that it produces impulse response estimates which are directly related to the estimated
A disadvantage of this estimator is that it is a highly nonlinear function of the VAR coefficient estimators. Therefore the distribution of the impulse response estimator is unlikely to be well approximated by the normal distribution. When the
Impulse response functions can be calculated and displayed in Stata using the irf command. The command irf create is used to calculate impulse response functions and confidence intervals. The default confidence intervals are asymptotic (delta method). Bootstrap (recursive method) standard errors can be substituted using the bs option. The command irf graph irf produces graphs of the impulse response function along with
15.16 Local Projection Estimator
Jordà (2005) observed that the impulse response can be estimated by a least squares predictive regression. The key is Theorem
The method is as follows. For each horizon
Theorem
Jordà (2005) speculates that the local projection estimator will be less sensitive to misspecification since it is a straightforward linear estimator. This is intuitive but unclear. Theorem
One implementation challenge is the choice of
An advantage of the local projection method is that it is a linear estimator of the impulse response and thus likely to have a better-behaved sampling distribution.
A disadvantage is that the method relies on a regression (15.12) that has serially correlated errors. The latter are highly correlated at long horizons and this renders the estimator imprecise. Local projection estimators tend to be less smooth and more erratic than those produced by the conventional estimator reflecting a possible lack of precision.
15.17 Regression on Residuals
If the innovations
where
The variables
In practice the innovations
This idea originated with Durbin (1960).
This is a two-step estimator with generated regressors. (See Section 12.26.) The impulse response estimators are consistent and asymptotically normal but with a non-standard covariance matrix due to the two-step estimation. Conventional, robust, and Newey-West standard errors do not account for this without modification.
Chang and Sakata (2007) proposed a simplified version of the Durbin regression. Notice that for any horizon
where
The regressor
Similar to the Durbin regression the Chang-Sakata estimator is a two-step estimator with a generated regressor. However, as it takes the form studied in Section
Chang and Sakata (2007) also point out the following implication of the FWL theorem. The least squares slope estimator in (15.14) is algebraically identical
15.18 Orthogonalized Shocks
We can use the impulse response function to examine how the innnovations impact the time-paths of the variables. A difficulty in interpretation, however, is that the elements of the innovation vector
The natural solution is to orthogonalize the innovations so that they are uncorrelated and then view the orthogonalized errors as the fundamental “shocks”. Recall that
To distinguish
When
with non-negative diagonal elements. We can write the Cholesky decomposition of a matrix
Equivalently, the innovations are related to the orthogonalized shocks by the equations
This structure is recursive. The innovation
When using the Cholesky decomposition the recursive structure is determined by the ordering of the variables in the system. The order matters and is the key identifying assumption. We will return to this issue later.
Finally, we mention that the system (15.15) is equivalent to the system
where
15.19 Orthogonalized Impulse Response Function
We have defined the impulse response function as the change in the time
We can write the multivariate Wold representation as
where
This is the non-orthogonalized impulse response matrix multiplied by the matrix square root
Write the rows of the matrix
and the columns of the matrix
There are
The cumulative orthogonalized impulse response function (COIRF) is
15.20 Orthogonalized Impulse Response Estimation
We have discussed estimation of the moving average matrices
We first estimate the VAR(p) model by least squares. This gives us the coefficient matrices
The estimator
As discussed earlier, the asymptotic approximations can be quite poor. Consequently bootstrap approximations are more widely used than asymptotic methods.
Orthogonalized impulse response functions can be displayed in Stata using the irf command. The command irf graph oirf produces graphs of the orthogonalized impulse response function along with
15.21 Illustration
To illustrate we use the three-variable system from Section 15.13. We use the ordering (1) real GDP growth rate, (2) inflation rate, (3) Federal funds interest rate. We discuss the choice later when we discuss identification. We use the estimated VAR(6) and calculate the orthogonalized impulse response functions using the standard VAR estimator.
In Figure
15.22 Forecast Error Decomposition
An alternative tool to investigate an estimated VAR is the forecast error decomposition which decomposes multi-step forecast error variances by the component shocks. The forecast error decomposition indicates which shocks contribute towards the fluctuations of each variable in the system.
- Impulse Response Function
- Cumulative IRF
Figure 15.1: Response of GDP Growth to Orthogonalized Fed Funds Shock
It is defined as follows. Take the moving average representation of the
The best linear forecast of
The
The variance of this forecast error is
To isolate the contribution of the
Thus the contribution of the
Examining (15.18) and using
The forecast error decomposition is defined as the ratio of the
The
A forecast error decomposition requires orthogonalized innovations. There is no non-orthogonalized
The forecast error decomposition can be calculated and displayed in Stata using the irf command. The command irf graph fevd produces graphs of the forecast error decomposition along with
15.23 Identification of Recursive VARs
As we have discussed, a common method to orthogonalize the VAR errors is the lower triangular Cholesky decomposition which implies a recursive structure. The ordering of the variables is critical this recursive structure. Unless the errors are uncorrelated different orderings will lead to different impulse response functions and forecast error decompositions. The ordering must be selected by the user; there is no data-dependent choice.
In order for impulse responses and forecast error decompositions to be interpreted causally the orthogonalization must be identified by the user based on a structural economic argument. The choice is similar to the exclusion restrictions necessary for specification of an instrumental variables regression. By ordering the variables recursively we are effectively imposing exclusion restrictions. Recall that in our empirical example we used the ordering: (1) real GDP growth rate, (2) inflation rate, (3) Federal funds interest rate. This means that in the equation for GDP we excluded the contemporeneous inflation rate and interest rate, and in the equation for inflation we excluded the contemporenous interest rate. These are exclusion restrictions. Are they justified?
One approach is to order first the variables which are believed to be contemporaneously affected by the fewest number of shocks. One way of thinking about it is that they are the variables which are “most sticky” within a period. The variables listed last are those which are believed to be contemporanously affected by the greatest number of shocks. These are the ones which are able to respond within a single period to the shocks or are most flexible. In our example we listed output first, prices second and interest rates last. This is consistent with the view that output is effectively pre-determined (within a period) and does not (within a period) respond to price and interest rate movements. Prices are allowed to respond within a period in response to output changes but not in response to interest rate changes. The latter could be justified if interest rate changes affect investment decisions but the latter take at least one period to implement. By listing the federal funds rate last the model allows monetary policy to respond within a period to contemporeneous information about output and prices.
In general, this line of reasoning suggests that production measures should be listed first, goods prices second, and financial prices last. This reasoning is more credible when the time periods are short, and less credible for longer time periods. Further justifications for possible recursive orderings can include: (1) information delays; (2) implementation delays; (3) institutions; (4) market structure; (5) homogeneity; (6) imposing estimates from other sources. In most cases such arguments can be made but will be viewed as debatable and restrictive. In any situation it is best to be explicit about your choice and reasoning.
Returning to the empirical illustration it is fairly conventional to order the fed funds rate last. This allows the fed funds rate to respond to contemporeneous information about output and price growth and identifies the fed funds policy shock by the assumption that it does not have a contemporenous impact on the other variables. It is not clear, however, how to order the other two variables. For simplicity consider a traditional aggregate supply/aggregate demand model of the determination of output and the price level. If the aggregate supply curve is perfectly inelastic in the short run (one quarter) then output is effectively fixed (sticky) so changes in aggregate demand affect prices but not output. Changes in aggregate supply affect both output and prices. Thus we would want to order GDP first and inflation second. This choice would identify the GDP error as the aggregate supply shock. This is the ordering used in our example.
In contrast, suppose that the aggregate supply curve is perfectly elastic in the short run. Then prices are fixed and output is flexible. Changes in aggregate supply affect both price and output but changes in aggregate demand only affect output. In this case we would want to order inflation first and GDP second. This choice identifies the inflation error as the aggregate supply shock, the opposite case from the previous assumption!
If the choice between perfectly elastic and perfectly inelastic aggregate supply is not credible then the supply and demand shocks cannot be separately identified based on ordering alone. In this case the full set of impulse responses and error decompositions are not identified. However, a subset may be identified. In general, if the shocks can be ordered in groups then we can identify any shock for which a group has a single variable. In our example, consider the ordering (1) GDP and inflation; (2) federal funds rate. This means that the model assumes that GDP and inflation do not contemporeneously respond to interest rate movements but no other restrictions are imposed. In this case the fed funds policy shock is identified. This means that impulse responses of all three variables with respect to the policy shock are identified and similarly the forecast error composition of the effect of the fed funds shock on each variable is identified. These can be estimated by a VAR using the ordering (GDP, inflation, federal funds rate) as done in our example or using the ordering (inflation, GDP, federal funds rate). Both choices will lead to the same estimated impulse responses as described. The remaining impulse responses (responses to GDP and inflation shocks), however, will differ across these two orderings.
15.24 Oil Price Shocks
To further illustrate the identification of impulse response functions by recursive structural assumptions we repeat here some of the analysis from Kilian (2009). His paper concerns the identification of the factors affecting crude oil prices, in particular separating supply and demand shocks. The goal is to determine how oil prices respond to economic shocks and how the responses differ by the type of shock.
To answer this question Kilian uses a three-variable VAR with monthly measures of global oil production, global economic activity, and the global price of crude oil for
Kilian argues that these three variables are determined by three economic shocks: oil supply, aggregate demand, and oil demand. He suggests that oil supply shocks should be thought of as disruptions in production, processing, or shipping. Aggregate demand is global economic activity. Kilian also argues that oil demand shocks are primarily due to the precautionary demand for oil driven by uncertainty about future oil supply shortfalls.
To identify the shocks Kilian makes the following exclusion restrictions. First, he assumes that the short-run (one month) supply of crude oil is inelastic with respect to price. Equivalently, oil production takes at least one month to respond to price changes. This restriction is believed to be plausible because of technological factors in crude oil production. It is costly to open new oil fields; and it is nearly impossible to cap an oil well once tapped. Second, Kilian assumes that in the short-run (one month) global real economic activity does not respond to changes in oil prices (due to shocks specific to the oil market), while economic activity is allowed to respond to oil production shocks. This assumption is viewed by Kilian as plausible due to the sluggishness in the response of economic activity to price changes. Crude oil prices, however, are allowed to respond simultaneously to all three shocks.
Kilian’s identification strategy is similar to that described in the previous section for the simple aggregate demand/aggregate supply model. The separation of supply and demand shocks is achieved by exclusion restrictions which imply short-run inelasticities. The plausibility of these assumptions rests in part on the monthly frequency of the data. While it is plausible that oil production and economic activity may not respond within one month to price shocks, it is much less plausible that there is no response for a full quarter. Kilian’s least convincing identifying assumption (in my opinion) is the assumption that economic activity does not respond simultaneously to oil price changes. While much economic activity is pre-planned and hence sluggish to respond, some economic activity (recreational driving, for example) may immediately respond to price changes.
Kilian estimates the three-variable VAR using 24 lags and calculates the orthogonalized impulse response functions using the ordering implied by these assumptions. He does not discuss the choice of 24 lags but presumably this is intended to allow for flexible dynamic responses. If the AIC is used for model selection, three lags would be selected. For the analysis reported here I used 4 lags. The results are qualitatively similar to those obtained using 24 lags. For ease of interpretation oil supply is entered negatively (multiplied by -1) so that all three shocks are scaled to increase oil prices. Two impulse response functions for the price of crude oil are displayed in Figure
What is noticeable about the figures is how differently crude oil prices respond to the two shocks. Panel (a) shows that oil prices are only minimally affected by oil production shocks. There is an estimated small short term increase in oil prices, but it is not statistically significant and it reverses within one year. In contrast, panel (b) shows that oil prices are significantly affected by aggregate demand shocks and the effect cumulatively increases over two years. Presumaby, this is because economic activity relies on crude oil and output growth is positively serially correlated.
The Kilian (2009) paper is an excellent example of how recursive orderings can be used to identify an orthogonalized VAR through a careful discussion of the causal system and the use of monthly observations.
15.25 Structural VARs
Recursive models do not allow for simultaneity between the elements of
- Supply Shock
- Aggregate Demand Schock
Figure 15.2: Response of Oil Prices to Orthogonalized Shocks
exclusively on recursiveness. Two popular categories of structural VAR models are those based on shortrun (contemporeneous) restrictions and those based on long-run (cumulative) restrictions. In this section we review SVARs based on short-run restrictions.
When we introduced methods to orthogonalize the VAR errors we pointed out that we can represent the relationship between the errors and shocks using either the equation
where (in the
(Note: This matrix
Written out,
The diagonal elements of the matrix
The system as written is under-identified. In this three-equation example, the matrix
We will illustrate by using a simplified version of the model employed by Blanchard and Perotti (2002) who were interested in decomposing the effects of government spending and taxes on GDP. They proposed a three-variable system consisting of real government spending (net of transfers), real tax revenues (including transfer payments as negative taxes), and real GDP. All variables are measured in logs. They start with the restrictions
This is done so that that the relationship between the shocks
We just described six restrictions while nine are required for identification. Blanchard and Perotti (2002) made a strong case for two additional restrictions. First, the within-quarter elasticity of government spending with respect to GDP is zero,
Blanchard and Perotti (2002) make use of both matrices
Taking the variance of the variables on each side of (15.21) we find
This is a system of quadratic equations in the free parameters. If the model is just identified it can be solved numerically to find the coefficients of
While most applications use just-identified models, if the model is over-identified (if there are fewer free parameters than estimated components of
Given the parameter estimates the structural impulse response function is
The structural forecast error decompositions are calculated as before with
The structural impulse responses are nonlinear functions of the VAR coefficient and covariance matrix estimators so by the delta method are asymptotically normal. Thus asymptotic standard errors can be calculated (using numerical derivatives if convenient). As for orthogonalized impulse responses the asymptotic normal approximation is unlikely to be a good approximation so bootstrap methods are an attractive alternative.
Structural VARs should be interpreted similarly to instrumental variable estimators. Their interpretation relies on valid exclusion restrictions which can only be justified by external information.
We replicate a simplified version of Blanchard-Perotti (2002). We use
- Spending
- Taxes
Figure 15.3: Response of GDP to Government Spending and Tax Shocks
In panel (a) we see that the effect of a
Structural vector autoregressions can be estimated in Stata using the svar command. Short-run restrictions of the form (15.21) can be imposed using the aeq and beq options. Structural impulse responses can be displayed using irf graph sirf and structural forecast error decompositions using irf graph sfevd. Unfortunately, Stata does not provide a convenient way to display cumulative structural impulse response functions. The same limitations for standard error and confidence interval construction in Stata hold for structural impulse responses as for non-structural impulse responses.
15.26 Identification of Structural VARs
The coefficient matrices
It is difficult to see if the system is identified simply by looking at the restrictions (except in the recursive case, which is relatively straightforward to identify). An intuitive way of verifying identification is to use our knowledge of instrumental variables. We can identify the equations sequentially, one at a time, or in blocks, using the metaphor of instrumental variables.
The general technique is as follows. Start by writing out the system imposing all restrictions and absorbing the diagonal elements of
Take the equations one at a time and ask if they can be estimated by instrumental variables using the excluded variables as instruments. Once an equation has been verified as identified then its shock is identified and can be used as an instrument since it is uncorrelated with the shocks in the other equations.
In this example take the equations as ordered. The first equation is identified as there are no coefficients to estimate. Thus
Consider another example based on Keating (1992). He estimated a four-variable system with prices, the fed funds rate, M2, and GDP. His model for the errors takes the form
where the four shocks are “aggregate supply”, “money supply”, “money demand”, and “I-S”. This structure can be based on the following assumptions: An elastic short-run aggregate supply curve (prices do not respond within a quarter); a simple monetary supply policy (the fed funds rate only responds within quarter to the money supply); money demand only responds to nominal output (log price plus log real output) and fed funds rate within a quarter; and unrestricted I-S curve.
To analyze conditions for identification we start by checking the order condition. There are 10 coefficients in the system (including the four variances), which equals
We check the equations for identification. We start with the first equation. It has no coefficients so is identified and thus so is
We find that the system is identified if
15.27 Long-Run Restrictions
To review, the algebraic identification problem for impulse response estimation is that we require a square root matrix
One important class of such structural VARs are those based on long-run restrictions. Some economic hypotheses imply restrictions on long-run impulse responses. These can provide a compelling case for identification.
An influential example of a structural VAR based on a long-run restriction is Blanchard and Quah (1989). They were interested in decomposing the effects of demand and supply shocks on output. Their hypothesis is that demand shocks are long-run neutral, meaning that the long-run impact of a demand shock on output is zero. This implies that the long-run impulse response of output with respect to demand is zero. This can be used as an identifying restriction.
The long-run structural impulse response is the cumulative sum of all impulse responses
A long-run restriction is a restriction placed on the matrix
Another way of thinking about this is that Blanchard-Quah label “aggregate supply” as the long-run component of GDP and label “aggregate demand” as the transitory component of GDP.
The relations
This is a set of
In many applications, including Blanchard-Quah, the matrix
The plug-in estimator for
More generally if the restrictions on
In either case the estimator is
Notice that by construction the long-run impulse response is
so indeed
Long-run structural vector autoregressions can be estimated in Stata using the svar command using the lreq option. Structural impulse responses can be displayed using irf graph sirf and structural forecast error decompositions using irf graph sfevd. This Stata option does not produce asymptotic standard errors when imposing long-run restrictions so for confidence intervals bootstrapping is recommended. The same limitations for such intervals constructed in Stata hold for structural impulse response functions as the other cases discussed.
Unfortunately, a limitation of the Stata svar command is that it does not display cumulative structural impulse response functions. In order to display these one needs to cumulate the impulse response estimates. This can be done but then standard errors and confidence intervals are not available. This means that for serious applied work programming needs to be done outside of Stata.
15.28 Blanchard and Quah (1989) Illustration
As we described in the previous section, Blanchard and Quah (1989) estimated a bivariate VAR in GDP growth and the unemployment rate assuming that the the structural shocks are aggregate supply and aggregate demand imposing that that the long-run response of GDP with respect to aggregate demand is zero. Their original application used U.S. data for 1950-1987. We revisit using FRED-QD (1959-2017). While Blanchard and Quah used a VAR(8) model the AIC selects a VAR(3). We use a VAR(4). To ease the interpretation of the impulse responses the unemployment rate is entered negatively (multiplied by -1) so that both series are pro-cyclical and positive shocks increase output. Blanchard and Quah used a careful detrending method; instead we including a linear time trend in the estimated VAR.
The fitted reduced form model coefficients satisfy
and the residual covariance matrix is
We calculate
Examining
Using this square root of
Let’s examine and contrast panels (a) and (b) of Figure 15.4. The response to a supply shock (panel (a)) takes several quarters to take effect, peaks around 5 quarters, and then decays. The response to a demand shock (panel (b)) is more immediate, peaks around 4 quarters, and then decays. Both are near zero by 6 years. The confidence intervals for the supply shock impulse responses are wider than those for the demand shocks indicating that the estimates of the impulse responses due to supply shocks are not precisely estimated.
Figure
- Supply Shock
- Demand Shock
Figure 15.4: Response of Unemployment Rate
It is fascinating that the structural impulse response estimates shown here are nearly identical to those found by Blanchard and Quah (1989) despite the fact that we have used a considerably different sample period.
15.29 External Instruments
Structural VARs can also be identified and estimated using external instrumental variables. This method is also called Proxy SVARs. Consider the three-variable system for the innovations
In this system we have used the normalization
Suppose we have an external instrumental variable
Equation (15.29) is the relevance condition - that the instrument and the shock
Suppose
- GDP
- Unemployment Rate
Figure 15.5: Forecast Error Decomposition, Percent due to Supply Shock
is uncorrelated with
This estimation method is not special for a three-variable system; it can be applied for any
While
The structure (15.26)-(15.28) is convenient as four coefficients can be identified. Other structures can also be used. Consider the structure
If the same procedure is applied we can identify the coefficients
15.30 Dynamic Factor Models
Dynamic factor models are increasingly popular in applied time series, in particular for forecasting. For a recent detailed review of the methods see Stock and Watson (2016) and the references therein. For some of the foundational theory see Bai (2003) and Bai and
In Sections 11.13-11.16 we introduced the standard multi-factor model (11.23):
where
In the time-series case it is natural to augment the model to allow for dynamic relationships. In particular we would like to allow
where
Furthermore we may wish to generalize (15.32) to allow
where
Define the inverse lag operators
where
for some
To estimate the explicit dynamic model (15.32)-(15.33)-(15.34) state-space methods are convenient. For details and references see Stock and Watson (2016).
The dynamic factor model (15.32)-(15.33)-(15.34) can be estimated in Stata using df actor.
15.31 Technical Proofs*
Proof of Theorem 15.6 Without loss of generality assume
By the Jordan matrix decomposition (see Section A.13),
For eigenvalues with multiplicity
Define
For single dimension blocks
For blocks with dimension two, by back-substitution we find
Partitioning
The series
Blocks with multiplicity
Proof of Theorem
Proof of Theorem 15.12 The first part of the theorem is established by back-substitution. Since
We then substitute out the first lag. We find
We continue making substitutions. With each substitution the error increases its MA order. After
To recognize that
with
Notice that
15.32 Exercises
Exercise 15.1 Take the VAR(1) model
Exercise 15.2 Take theVAR(2) model
Exercise 15.3 Suppose
Exercise 15.4 Suppose
Exercise 15.5 In the VAR(1) model
Exercise 15.7 Derive a VAR(1) representation of a VAR(p) process analogously to equation (14.33) for autoregressions. Use this to derive an explicit formula for the
Exercise 15.8 Let
Exercise 15.9 Continuting the previous exercise, suppose that both
Exercise 15.10 Suppose that you have 20 years of monthly observations on
Exercise 15.11 Let
Exercise 15.12 Cholesky factorization
Derive the Cholesky decomposition of the covariance matrix
.Write the answer for the correlation matrix (the special case
and ).Find an upper triangular decomposition for the correlation matrix. That is, an upper-triangular matrix
which satisfies .Suppose
, and , and . Find the orthogonalized impulse response OIRF(h) using the Cholesky decomposition.Suppose that the ordering of the variables is reversed. This is equivalent to using the upper triangular decomposition from part (c). Calculate the orthogonalized impulse response OIRF(h).
Compare the two orthogonalized impulse responses.
Exercise 15.13 You read an empirical paper which estimates a VAR in a listed set of variables and displays estimated orthogonalized impulse response functions. No comment is made in the paper about the ordering or the identification of the system, and you have no reason to believe that the order used is “standard” in the literature. How should you interpret the estimated impulse response functions?
Exercise 15.14 Take the quarterly series gdpcl (real GDP), gdpctpi (GDP price deflator), and fedfunds (Fed funds interest rate) from FRED-QD. Transform the first two into growth rates as in Section 15.13. Estimate the same three-variable VAR(6) using the same ordering. The identification strategy discussed in Section
Exercise 15.16 Take the monthly series permit (building permits), houst (housing starts), and realln (real estate loans) from FRED-MD. The listed ordering is motivated by transaction timing. A developer is required to obtain a building permit before they start building a house (the latter is known as a “housing start”). A real estate loan is obtained when the house is purchased.
Transform realln into growth rates (first difference of logs).
Select an appropriate lag order for the three-variable system by comparing the AIC of VARs of order 1 through 8.
Estimate the VAR model and plot the impulse response functions of housing starts with respect to the three shocks.
Interpret your findings.
Exercise 15.17 Take the quarterly series gpdicl (Real Gross Private Domestic Investment), gdpctpi (GDP price deflator), gdpcl (real GDP), and fedfunds (Fed funds interest rate) from FRED-QD. Transform the first three into logs, e.g.
Write down the matrix
similar to (15.22), imposing the identifying constraints as defined above.Is the model identified? Is there a condition for identification? Explain.
In this model are output and price simultaneous, or recursive as in the example described in Section
?Estimate the structural VAR using 6 lags or a different number of your choosing (justify your choice) and include an exogenous time trend. Report your estimates of the
matrix. Can you interpret the coefficients?Estimate and report the following three impulse response functions:
The effect of the fed funds rate on GDP.
The effect of the GDP shock on GDP.
The effect of the GDP shock on prices.
Exercise 15.18 Take the Kilian2009 dataset which has the variables oil (oil production), output (global economic activity), and price (price of crude oil). Consider a structural VAR based on short-run restrictions. Use a structure of the form
Is the model identified? Is there a condition for identification? Explain.
Estimate the structural VAR using 4 lags or a different number of your choosing (justify your choice). (As described in that section, multiply “oil” by
so that all shocks increase prices.) Report your estimates of the matrix. Can you interpret the coefficients?Estimate the impulse response of oil price with respect to the three shocks. Comment on the estimated functions.
Exercise 15.19 Take the quarterly series gdpc1 (real GDP), m1realx (real M1 money stock), and cpiaucsl (CPI) from FRED-QD. Create nominal M1 (multiply m1realx times cpiaucsl), and transform real GDP and nominal M1 to growth rates. The hypothesis of monetary neutrality is that the nominal money supply has no effect on real outcomes such as GDP. Strict monetary neutrality states that there is no short or long-term effect. Long-run neutrality states that there is no long-term effect.
To test strict neutrality use a Granger-causality test. Regress GDP growth on four lags of GDP growth and four lags of money growth. Test the hypothesis that the four money lags jointly have zero coeffficients. Use robust standard errors. Interpret the results.
To test long-run neutrality test if the sum of the four coefficients on money growth equals zero. Interpret the results.
Estimate a structural VAR in real GDP growth and nominal money growth imposing the long-run neutrality of money. Explain your method.
Report estimates of the impulse responses of the levels of GDP and nominal money to the two shocks. Interpret the results.
Exercise 15.20 Shapiro and Watson (1988) estimated a structural VAR imposing long-run constraints. Replicate a simplified version of their model. Take the quarterly series hoanbs (hours worked, nonfarm business sector), gdpcl (real GDP), and gdpctpi (GDP deflator) from FRED-QD. Transform the first two to growth rates and for the third (GDP deflator) take the second difference of the logarithm (differenced inflation). Shapiro and Watson estimated a structural model imposing the constraints that labor supply hours are long-run unaffected by output and inflation and GDP is long-run unaffected by demand shocks. This implies a recursive ordering in the variables for a long-run restriction.
Write down the matrix
as in (15.24) imposing the identifying constraints as defined above.Is the model identified?
Use the AIC to select the number of lags for a VAR.
Estimate the structural VAR. Report the estimated
matrix. Can you interpret the coefficients?Estimate the structural impulse responses of the level of GDP with respect to the three shocks. Interpret the results.