Home The Rescaled VAR Model with an Application to Mixed-Frequency Macroeconomic Forecasting
Article
Licensed
Unlicensed Requires Authentication

The Rescaled VAR Model with an Application to Mixed-Frequency Macroeconomic Forecasting

  • Andrea Giusto EMAIL logo and Talan B. İşcan
Published/Copyright: April 30, 2018

Abstract

This paper introduces the rescaled representation of VAR models (R-VARs) and demonstrates its application in forecasting mixed-frequency macroeconomic data. We develop the model, illustrate how to implement it, and derive the asymptotic properties of the estimates. We show that R-VARs provide reliable estimates of the prediction error bands while maintaining the precision of the point forecasts. We illustrate these features by comparing it to a mixed-frequency Bayesian VAR model, the leading alternative in the existing literature.

JEL Classification: C15; C32

Acknowledgement

We thank two anonymous reviewers for their comments and suggestions. This research was not externally funded. The authors declare no competing financial interests.

A Appendix

A.1 Proof of Theorem 1

The solutions to non-stochastic homogeneous systems are given by the matrix exponential function defined, for complex t and square stable matrix A, as exp(tA)=n=0tnn!An. Hence, to find a particular solution to (3) – given initial conditions X0 at t0 = 0 – we define y(t) as

(6)X(t)=etAy(t).

which implies that dX(t) = AetAy(t)dt + etAdy, and, using (3), it follows that

Bdt+AXdt+ΓdW=AeAty(t)dt+eAtdy,
dy=eAtBdt+eAtΓdW,

integrating between 0 and t, we have

y(t)=0teAsBds+0teAsΓdW(s).

Using now the initial condition X(0) = X0 the particular solution to (3) is

X(t)=etAX0+etA0tesABds+eAt0tesAΓdW(s),

where the last integral is a normally distributed random variable. Therefore, our DGP maps into that of Sims (1971) and Geweke (1978) by specifying the functions x and b in those papers as

x(s)b(ts){etAX0+e(ts)ABs[0,t]0otherwise.

Now let δ > 0 denote the interval of time between observations of this process. The solution is

X(t+δ)=e(t+δ)AX0+e(t+δ)A0t+δesABds+e(t+δ)A0t+δesAΓdW(s)=eδA[etAX0+etA0tesABds+etA0tesAΓdW(s)]+e(t+δ)Att+δesABds+e(t+δ)Att+δesAΓdW(s)=eδAX(t)+[eδAI]A1B+tt+δe(t+δs)AΓdW(s).

This implies that X(t) and X(t + δ) satisfy a VAR(1) recursion; now define s = t + δz to obtain

X(t+δ)=eAδX(t)+[eAδI]A1B+δ0eAz(1)ΓdW(z)=eAδX(t)+[eAδI]A1B+0δeAzΓdW(z),

where the integral is a Gaussian random variable that is uncorrelated with X(t), with zero mean and covariance matrix

(7)0δeAzΓΓeAzdz.

The proof is completed by letting δ equal 1 first and then 1τ. Note that the inverse of eδA exists for all δ and A, thus showing uniqueness of Φ1 and Φτ.

A.2 Proof of Theorem 2

Define Q = ΓΓ, and

(8)S=0ezAQezAdz.

Noticing that Q is positive definite, we then have

AS+SA=0AezAQezA+ezAQezAAdz=0ddz[ezAQezA]dz
AS+SA=[ezAQezA]z=0z=,

and taking the limit yields the following Lyapunov equation

AS+SA+Q=0.

Lyapunov equations have a unique solution for S if A is stable. Moreover, the solution S is symmetric and positive definite. To solve for S, we use a vectorization formula [see Hamilton (1994) equations 10.2.13–18]

(IA+AI)vec(S)=vec(Q).
vec(S)=(AA)1vec(Q).

where ⊕ denotes the Kronecker sum.[9] Consider now the integral on the left side of (4)

0δezAΓΓezAdz=0ezAΓΓezAdzδezAΓΓezAdz.

and using the definition of S in equation (8) we have

0δezAΓΓezAdz=S0e(s+δ)AΓΓe(s+δ)Adz,

where we applied the change of variable z = s + δ; using (8) once more proves the theorem.

A.3 Proof of Theorem 3

A.3.1 Preliminary Lemmas

The next two lemmas are well-known but are included here to make the paper self contained.

Lemma 1

Let Δ be a complex number. Then

  1. limN(I+ΔNA)N=eΔA.

  2. limNi=0N1(I+ΔNA)iΔN=[eΔAI]A1.

Proof.

Since the matrices A and I trivially commute we can apply the binomial theorem

(I+ΔNA)N=n=01n!k=0n1(Nk)(ΔNA)n=I+ΔA+N(N1)2!N2Δ2A2+N(N1)(N2)3!N3Δ3A3.

and letting N go to infinity proves (a). Using again the binomial theorem for each of the terms of the summation above yields

limNi=0N1(I+ΔNA)iΔN=limNi=0N1n=01n!k=0n1(ik)(ΔN)nAnΔN=limNi=0N1n=01n!k=0n1(ik)(ΔN)n+1An.

Notice that the terms with matching indexes i and n are equal to zero:

i,n0123N1N0:ΔNI+11!(00)(ΔN)2A+0+0+0+01:+ΔNI+11!(10)(ΔN)2A+12!(10)(11)(ΔN)3A2+0+0+02:+ΔNI+11!(20)(ΔN)2A+12!(20)(21)(ΔN)3A2+0+0+0N1:+ΔNI+11!(N1)(ΔN)2A+12!(N1)(N2)(ΔN)3A2+(N1)!(N1)!(ΔN)NAN1+0.

Thus we have

i=0N1(I+ΔNA)iΔN=n=0N1ΔNI+n=1N1n(ΔN)2A+n=2N1(n1)(n2)2!(ΔN)3A2++(ΔN)NAN1.

Letting N → ∞, this expression becomes

ΔI+12!Δ2A2+13!Δ3A2+.

Post-multiplying it by AA−1 proves (b).    □

Lemma 2

Let Δ be a complex number, and let S be a matrix that solves the discrete-time Lyapunov equation

(9)(I+ΔNA)S(I+ΔNA)S+ΓΓ=0.

Then vec(ΓΓ)=(AA)limNΔNvec(S).

Proof.

The solution to the Lyapunov equation satisfies

(10)vec(ΓΓ)=[I(I+ΔNA)(I+ΔNA)]vec(S)=NΔ[IΔNAΔNAIΔNAΔNA]ΔNvec(S).

Taking the limit gives

limNvec(ΓΓ)=limN[IAAIAΔNA]ΔNvec(S)vec(ΓΓ)=[IAAI]limNΔNvec(S).

This proves the Lemma.    □

A.3.2 Theorem 3

Let θ = [A′, B′, Γ′]′. Process (3) is linear and Gaussian and therefore there exists a unique family of probability measures {Pθ,s,x: s ≥ 0, x ∈ ℝK} induced by the solutions of (3) for ts, given θ and Xs = x – see Pedersen (1995). Define the transition probability function implied by this family of probabilities as

P(s,x,t,D;θ)=Pθ,s,x(XtD)

for a Borel set D ∈ ℝK. Define Δ = ts and let p(s, x, t, y; θ) denote the density of P with respect to the Lebesgue measure in ℝK. The Euler-Maruyama approximation of order N = 1, 2, … under Pθ,s,x to the continuous time stochastic process, then, is

(11)nj=jNΔ+s,Xs(N)=x,Xnj(N)=Xnj1+ΔN[AXnj1(N)+B]+Γ(Wnjθ,sWnj1θ,s),

where Wtθ,s=st(ΓΓ)1/2d[XuxsuAXv+Bdv], for t ≥ 0, is a K-dimensional Wiener process starting at time s, under Pθ,s,x. Then it can be shown that XnNXt in L1 norm (with measure Pθ,s,x) as N diverges (Pedersen, 1995). The transition probabilities of the Euler-Maruyama approximation are such that

(12)p(N)(s,x,t,y;θ)=E[p(1)(nN1,XnN1(N),t,y;θ)],

where the expectation is taken with respect to Pθ,s,x. Intuitively, p(N) is equal to the probability that the Euler-Maruyama process will transition to y at time t, from the position XnN1(N) that it is expected to take at t − Δ/N. Then p(N)p as N → ∞ (Pedersen, 1995, Theorem 1). Backward substitution in the Euler-Maruyama N-approximation yields

XnN1(N)=(I+ΔNA)N1x+i=0N2(I+ΔNA)iΔNB+i=0N2(I+ΔNA)iΓεnN1+i,

where εnj=Wnjθ,sWnj1θ,s. It follows that XnN1(N) is normal with

E[XnN1(N)]=(I+ΔNA)N1x+i=0N2(I+ΔNA)iΔNB,var[XnN1(N)]=ΔNi=0N2(I+ΔNA)iΓΓ(I+ΔNA)i.

Together with (12), these imply that the probability that XnN(N)=y at nN = t can be found by taking the expectation of the following expression

(I+ΔNA)iXnN1(N)+ΔNB+ΓεnN=y.

Moreover, because XnN1(N) is normal, we have that p(N)(s, x, t, y; θ) also is normal with mean (I+ΔNA)Nx+i=0N1(I+ΔNA)iΔNB and covariance matrix ΔNi=0N1(I+ΔNA)iΓΓ(I+ΔNA)i.

Define now S=i=0(I+ΔNA)iΓΓ(I+ΔNA)i. Under Assumption 2 and Assumption 3, none of the eigenvalues of I+ΔNA can be the reciprocal of another one and therefore S is the unique solution to the Lyapunov equation (9), and it satisfies equation (10). Now, from the definition of S it follows that

i=0N1(I+ΔNA)iΓΓ(I+ΔNA)i=S(I+ΔNA)NS(I+ΔNA)N,

which shows that p(N)(s, x, s + Δ, y; θ) is normal with mean

μ(N)=(I+ΔNA)Nx+i=0N1(I+ΔNA)iΔNB,

and covariance

Σ(N)=ΔN(S(I+ΔNA)NS(I+ΔNA)N).

From Lemma 1 the limiting distribution of p(N) as N → ∞ is normal with mean

μ=limN(I+ΔNA)Nx+i=0N1(I+ΔNA)iΔNB=eΔAx+[eΔAI]A1B,

which is the desired result for the mean. To derive the covariance matrix, use Lemma 1 and Lemma 2 and since Assumption 2 ensures that AA is invertible, we have

Σ=limNΔN(S(I+ΔNA)NS(I+ΔNA)N).

Finally,

vec(Σ)=limNΔN[vec(S)(I+ΔNA)N(I+ΔNA)Nvec(S)]=limN[I(I+ΔNA)2N(I+ΔNA)N]ΔNvec(S)=(IeΔ(AA))(AA)1vec(ΓΓ).

This completes the proof.

A.4 Proof of Theorem 4

Let ({Xi}i=1T;C1,Φ1,Σ1) denote the likelihood function for the observed data set, given X0; we have

(13)log({Xi}i=1T;C1,Φ1,Σ1)=KT2log2πT2log|Σ1|12t=1T[(Xtm)Φ1(Xt1m)]Σ11[(Xtm)Φ1(Xt1m)],

where m = [I − Φ1]−1C1. The maximum likelihood estimates for the LF-VAR are (C~1,Φ~1,Σ~1)=argmaxlog({Xi}i=1T;C1,Φ1,Σ1). Under the assumption that (3) is the DGP we can also express the likelihood of these data as follows

log({Xi}i=1T;A,B,Γ)=t=1Tlogp(t,X(t),t+1,X(t+1))

where p is the transition probability function defined in the proof of Theorem 3: given X(t) at t the distribution of X(t + 1) at t + 1 is normal with mean eAX(t) + [eAI]A−1B and (vectorized) variance vec(Σ) = (IeA⊕A)(−AA)−1vecΓ). Define now X0(t) ≡ X(t) + A−1B, so that p(t, X0(t), t + 1, X0(t + 1)) is normal with mean eAX0(t) and variance Σ. It follows that this likelihood satifies

(14)log({Xi}i=1T;A,B,Γ)=KT2log2πT2log|Σ|12t=1T[X0(t)eAX0(t1)]Σ1[X0(t)eAX0(t1)].

Under the mapping defined by the rescaling algorithm, we have that (i) eA = Φ1; (ii) C1 = (eAI)−1A−1B so that m = −A−1B; and (iii) vec(Σ) = (eA⊕AI)(AA)−1vecΓ) = vec1); and therefore the rescaling algorithm maps one-to-one the maximizers of (13) into the maximizers of (14). It is now immediate from Theorem 3 that the rescaling algorithm also maps one-to-one the maximizers of (14) into the maximizers of the likelihood of the (unobserved) data set in which all the variables are measured at the higher frequency

log({Xi/τ}i=1Tτ;A,B,Γ)

and therefore the matrices C̃τ, Φ~τ, and Σ~τ are, asymptotically, ML estimates of the R-VAR’s parameters. The distributions of the estimators follow from standard results; see, for example, (Lütkepohl, 2007, Proposition 3.1).

References

Clark, Todd E., and Michael W. McCracken. 2017. “Tests of Predictive Ability for Vector Autoregressions Used for Conditional Forecasting.” Journal of Applied Econometrics 32 (3): 533–553.10.1002/jae.2529Search in Google Scholar

Geweke, John. 1978. “Temporal Aggregation in the Multiple Regression Model.” Econometrica 46 (3): 643–661.10.2307/1914238Search in Google Scholar

Ghysels, Eric. 2016. “Macroeconomics and the Reality of Mixed Frequency Data.” Journal of Econometrics 193 (2): 294–314.10.1016/j.jeconom.2016.04.008Search in Google Scholar

Ghysels, Eric, Andrew C. Harvey, and Eric Renault. 1996. “Stochastic Volatility.” Handbook of Statistics 14: 119–191.10.1016/S0169-7161(96)14007-4Search in Google Scholar

Giacomini, Raffaella, and Halbert White. 2006. “Tests of Conditional Predictive Ability.” Econometrica 74 (6): 1545–1578.10.1111/j.1468-0262.2006.00718.xSearch in Google Scholar

Götz, Thomas B., Alain Hecq, and Stephan Smeekes. 2016. “Testing for Granger Causality in Large Mixed-Frequency VARs.” Journal of Econometrics 193 (2): 418–432.10.1016/j.jeconom.2016.04.015Search in Google Scholar

Gourieroux, Christian, and Joann Jasiak. 2016. “Filtering, Prediction and Simulation Methods for Noncausal Processes.” Journal of Time Series Analysis 37 (3): 405–430.10.1111/jtsa.12165Search in Google Scholar

Hamilton, James Douglas. 1994. Time Series Analysis, vol. 2. Princeton: Princeton University Press.10.1515/9780691218632Search in Google Scholar

Higham, Nicholas J. 2002. “Computing the Nearest Correlation Matrix – A Problem from Finance.” IMA Journal of Numerical Analysis 22 (3): 329–343.10.1093/imanum/22.3.329Search in Google Scholar

Hyndman, Rob J., Anne B. Koehler, Ralph D. Snyder, and Simone Grose. 2002. “A State Space Framework for Automatic Forecasting Using Exponential Smoothing Methods.” International Journal of Forecasting 18 (3): 439–454.10.1016/S0169-2070(01)00110-8Search in Google Scholar

Knuth, D. E. 1981. The Art of Computer Programming Vol. 2: Seminumerical Methods. Reading, MA: Addison-Wesley.Search in Google Scholar

Lütkepohl, Helmut. 2007. New Introduction to Multiple Time Series Analysis. Springer.Search in Google Scholar

Marcellino, Massimiliano. 1999. “Some Consequences of Temporal Aggregation in Empirical Analysis.” Journal of Business and Economic Statistics 17 (1): 129–136.10.1080/07350015.1999.10524802Search in Google Scholar

Nyberg, Henri, and Pentti Saikkonen. 2014. “Forecasting with a Noncausal Var Model.” Computational Statistics & Data Analysis 76: 536–555.10.1016/j.csda.2013.10.014Search in Google Scholar

Pedersen, Asger Roer. 1995. “A New Approach to Maximum Likelihood Estimation for Stochastic Differential Equations Based on Discrete Observations.” Scandinavian Journal of Statistics 22 (1): 55–71.Search in Google Scholar

Press, William H., Saul A. Teukolsky, William T. Vetterling, and Brian P. Flannery. 2007. Numerical Recipes 3rd Edition: The Art of Scientific Computing. Cambridge: Cambridge University Press.Search in Google Scholar

Qian, Hang. 2016. “A Computationally Efficient Method for Vector Autoregression with Mixed Frequency Data.” Journal of Econometrics 193 (2): 433–437.10.1016/j.jeconom.2016.04.016Search in Google Scholar

Ripley, B. D. 1990. “Thoughts on Pseudorandom Number Generators.” Journal of Computational and Applied Mathematics 31 (1): 153–163.10.1016/0377-0427(90)90346-2Search in Google Scholar

Schorfheide, Frank, and Dongho Song. 2015. “Real-Time Forecasting with a Mixed Frequency VAR.” Journal of Business and Economic Statistics 33 (3): 366–380.10.3386/w19712Search in Google Scholar

Sims, Christopher A. 1971. “Discrete Approximations to Continuous Time Distributed Lags in Econometrics.” Econometrica 39 (3): 545–563.10.2307/1913265Search in Google Scholar

Published Online: 2018-04-30

©2018 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 8.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/snde-2017-0047/pdf?srsltid=AfmBOorNV7Ri-Q5udi2EFRYBcDaSDPUI5GKD7t6GsxKyA7Izl_rvshIc
Scroll to top button