Forecasting Economics and Business PDF
Document Details
Uploaded by RaptTechnetium
University of Sydney
David Ubilava
Tags
Summary
This document covers forecasting methods for economics and business, specifically focusing on inter-related economic variables. It explores models like Vector Autoregression (VAR) and Autoregressive Distributed Lag (ARDL), and explains their application for forecasting. The lecture also details Granger causality.
Full Transcript
Forecasting for Economics and Business Lecture 10: Forecasting Inter-related Series David Ubilava University of Sydney Economic variables are inter-related Economic variables are often (and usually) inter- related; for example, w...
Forecasting for Economics and Business Lecture 10: Forecasting Inter-related Series David Ubilava University of Sydney Economic variables are inter-related Economic variables are often (and usually) inter- related; for example, when incomes increase, people consume more; when interest rates increase, people invest less. We can model these relationships by regressing one variable on the other(s). 2 / 31 Autoregressive distributed lag (ARDL) Because of possible delays in adjustment to a new equilibrium, we typically need to include lagged values in the regression. This leads to an autoregressive distributed lag (ARDL) model, as follows: yt = α + β1 yt−1 + … + βp yt−p + δ0 xt + … + δq xt−q + εt Note that an autoregressive model and a distributed lag model are special cases of the ARDL. As usual, we can use an information criterion to select p and q. 3 / 31 A relationship between oil prices and inflation 4 / 31 ARDL lag length and estimated parameters Obtain AIC and SIC for lags 1 AIC suggests four lags while SIC indicates two lags. through 4 Let's go with SIC. AIC SIC ` α ` ` β1 ` ` β2 ` ` δ0 ` ` δ1 ` ` δ2 ` 1.997 2.058 estimate 0.251 0.986 -0.098 0.016 -0.010 -0.007 1.932 2.023 s.e. 0.048 0.065 0.063 0.002 0.003 0.002 1.946 2.068 1.913 2.065 5 / 31 Forecasting with ARDL Point forecasts: ^t+1 + δ1 xt + δ2 xt−1 yt+1|t = α + β1 yt + β2 yt−1 + δ0 x ^t+2 + δ1 x yt+2|t = α + β1 yt+1|t + β2 yt + δ0 x ^t+1 + δ2 xt yt+h|t = α + β1 yt+h−1|t + β2 yt+h−2|t ^t+h + δ1 x + δ0 x ^t+h−1 + δ2 x ^t+h−2 Note, and this is one of the caveats of using ARDL in forecasting, we need to have a knowledge of future realized values of x. Often, that involves also (separately) forecasting those values, or setting them to some values (e.g., scenario forecasting). 6 / 31 Forecasting with ARDL If we assume (somewhat unrealistically) that we know the future realized values of x, then the forecast errors and forecast error variances from the ARDL model will be the same as those from an AR model. This is, in a way, akin to working with an AR model that includes a trend or a seasonal component. Of course, obtaining direct multistep forecasts is always an option. 7 / 31 If we knew future realized values of oil prices 8 / 31 If we set future realized values to the current oil price 9 / 31 Vector autoregression More generally, dynamic linkages between two (or more) economic variables can be modeled as a system of equations, better known as a vector autoregression (VAR). 10 / 31 Vector autoregression To begin, consider a bivariate first-order VAR. Let {X1t } and {X2t } be the stationary stochastic processes. A bivariate VAR(1), is then given by: x1t = α1 + π111 x1t−1 + π121 x2t−1 + ε1t x2t = α2 + π211 x1t−1 + π221 x2t−1 + ε2t where ε1t ∼ iid(0, σ12 ) and ε2t ∼ 2 iid(0, σ ) 2 , and the two can be correlated, i.e., Cov(ε1t , ε2t ) ≠ 0. 11 / 31 Vector autoregression An n-dimensional VAR of order p, VAR(p), presented in matrix notation: xt = α + Π1 xt−1 + … + Πp xt−p + εt , where xt = (x1t , … , xnt ) ′ is a vector of n (potentially) related variables; εt = (ε1t , … , εnt ) ′ is a vector of error terms, such that E (εt ) = 0, ′ E (εt ε ) = Σε t , and E (εt ε′s≠t ) = 0. 12 / 31 A parameter matrix of a vector autoregression Π1 , … , Πp are n-dimensional parameter matrices such that: π11j π12j ⋯ π1nj ⎡ ⎤ ⎢ π21j π22j ⋯ π2nj ⎥ ⎢ ⎥ Πj = ⎢ ⎥, j = 1, … , p ⎢ ⎥ ⎢ ⋮ ⋮ ⋱ ⋮ ⎥ ⎣ ⎦ πn1j πn2j ⋯ πnnj 13 / 31 Features of vector autoregression General features of a (reduced-form) vector autoregression are that: only the lagged (i.e., no contemporaneous) values of the dependent variables are on the right-hand-side of the equations. Although, trends and seasonal variables can also be included. each equation has the same set of right-hand-side variables. However, it is possible to impose different lag structure across the equations, especially when p is relatively large. This is because the number of parameters increases very quickly with the number of lags or the number of variables in the system. the autregressive order, p, is the largest lag across all equations. 14 / 31 Modeling vector autoregression The autoregressive order, p, can be determined using system-wide information criteria: 2 2 AI C = ln|Σε | + (pn + n) T ln T 2 SI C = ln|Σε | + (pn + n) T where |Σε | is the determinant of the residual covariance matrix; n is the number of equations, and T is the total number of observations. 15 / 31 Estimating vector autoregression When each equation of VAR has the same regressors, the OLS can be applied to each equation individually to estimate the regression parameters - i.e., the estimation can be carried out on the equation-by- equation basis. Indeed, taken separately, each equation is just an ARDL (albeit without any contemporaneous regressors in the right-hand side of the equation). When processes are covariance-stationarity, conventional t-tests and F- tests are applicable for hypotheses testing. 16 / 31 Testing in-sample Granger causality Consider a bivariate VAR(p): x1t = α1 + π111 x1t−1 + ⋯ + π11p x1t−p + π121 x2t−1 + ⋯ + π12p x2t−p + ε1t x2t = α1 + π211 x1t−1 + ⋯ + π21p x1t−p + π221 x2t−1 + ⋯ + π22p x2t−p + ε2t {X2 } does not Granger cause {X1 } if π121 = ⋯ = π12p = 0 {X1 } does not Granger cause {X2 } if π211 = ⋯ = π21p = 0 17 / 31 Testing in-sample Granger causality In our example: gc Crude Oil − → Inflation : F statistic is 12.12 gc Inflation − → Crude Oil : F statistic is 1.98 So, we have a unidirectional Granger causality (from crude oil prices to 5- year expected inflation). 18 / 31 Forecasting with VAR models: one-step ahead To keep things simple, we illustrate using bivariate VAR(1). Point forecasts: x1t+1|t = α1 + π111 x1t + π121 x2t x2t+1|t = α2 + π211 x1t + π221 x2t Forecast errors: e1t+1 = x1t+1 − x1t+1|t = ε1t+1 e2t+1 = x2t+1 − x2t+1|t = ε2t+1 19 / 31 Forecasting with VAR models: one-step ahead Forecast variances: 2 2 2 2 σ = E(e1t+1 |Ωt ) = E(ε ) = σ 1t+1 1t+1 1 2 2 2 2 σ = E(e2t+1 |Ωt ) = E(ε ) = σ 2t+1 2t+1 2 Assuming normality of forecast errors, interval forecasts are obtained the usual way. 20 / 31 Forecasting with VAR models: multi-step ahead Point forecasts: x1t+h|t = α1 + π111 x1t+h−1|t + π121 x2t+h−1|t x2t+h|t = α2 + π211 x1t+h−1|t + π221 x2t+h−1|t Forecast errors: e1t+h = π111 e1t+h−1 + π121 e2t+h−1 + ε1t+h e2t+h = π211 e1t+h−1 + π221 e2t+h−1 + ε2t+h Forecast variances are the functions of error variances and covariances, and the model parameters. 21 / 31 Forecasting oil prices and inflation 22 / 31 Direct multi-step ahead forecasts For a given horizon, h, the estimated model: x1t = ϕ1 + ψ111 x1t−h + ψ121 x2t−h + υ1ht x2t = ϕ2 + ψ211 x1t−h + ψ221 x2t−h + υ2ht where υ1ht ∼ iid(0, σ1h 2 ) and υ2ht ∼ iid(0, σ 2 2h ) , and the two can be correlated, i.e., Cov(υ1ht , υ2ht ) ≠ 0. Point forecasts: x1t+h|t = ϕ1 + ψ111 x1t + ψ121 x2t x2t+h|t = ϕ2 + ψ211 x1t + ψ221 x2t 23 / 31 Direct multi-step ahead forecasts Forecast errors: e1t+h = υ1t+h e2t+h = υ2t+h Forecast variances: 2 2 2 2 σ = E(e1t+h |Ωt ) = E(υ ) = σ 1t+h 1t+h 1h 2 2 2 2 σ = E(e2t+h |Ωt ) = E(υ ) = σ 2t+h 2t+h 2h Assuming normality, interval forecasts are obtained directly from these variances, the usual way. 24 / 31 Out-of-Sample Granger Causality The previously discussed (in sample) tests of causality in Granger sense are frequently performed in practice, but the 'true spirit' of such test is to assess the ability of a variable to help predict another variable in an out- of-sample setting. 25 / 31 Out-of-Sample Granger Causality Consider restricted and unrestricted information sets: (r) Ω ≡ Ωt (X1 ) = {x1,t , x1,t−1 , …} t (u) Ω ≡ Ωt (X1 , X2 ) = {x1,t , x1,t−1 , … , x2,t , x2,t−1 , …} t Following Granger's definition of causality: {X2 } is said to cause {X1 } if (u) (r) 2 σx 1 (Ω t ) < σx 2 1 (Ω t ) , meaning that we can better predict X1 using all available information on X1 and X2 , rather than that on X1 only. 26 / 31 Out-of-Sample Granger Causality Let the forecasts based on each of the information sets be: (r) (r) x = E (x1t+h |Ω ) 1t+h|t t (u) (u) x = E (x1t+h |Ω ) 1t+h|t t 27 / 31 Out-of-Sample Granger Causality For these forecasts, the corresponding forecast errors are: (r) (r) e = x1t+h − x 1t+h 1t+h|t (u) (u) e = x1t+h − x 1t+h 1t+h|t The out-of-sample forecast errors are then evaluated by comparing the loss functions based on these forecasts errors. The out of sample Granger causality test is, in effect, a test of relative forecast accuracy between the two models. 28 / 31 Accuracy Tests for Nested Models In testing Out-of-sample Granger causality, we compare nested models. For example, consider a bivariate VAR(1). A test of out-of-sample Granger causality involves comparing forecasts from these two models: (A) : x1t = β0 + β1 x1t−1 + β2 x2t−1 + υt (B) : x1t = α0 + α1 x1t−1 + εt Obviously, here model (B) is nested in model (A). Under the null of Granger non-causality, the disturbances of the two models are identical. This leads to 'issues' in the usual tests. 29 / 31 Accuracy Tests for Nested Models There is a way to circumvent these issues, which involves an adjustment of the loss differential. In particular, the loss differential now becomes: 2 2 2 d(et+h,ij ) = e − e + (yt+h|t,i − yt+h|t,j ) , t+h,i t+h,j where model i is nested in model j. The testing procedure is otherwise similar to the Diebold-Mariano test. 30 / 31 Key Takeaways Many economic time series are inter-related. We can model such series as a system of equations known as the Vector Autoregressive (VAR) model. Autoregressive Distributed Lag (ARDL) model is a special case (which is, in effect, a single equation) of a VAR model. We can use information on related variables to help better forecast a variable of interest. If improved accuracy is achieved, we have a case of Granger causality. 31 / 31