MIT14 384F13 Lec11 PDF
MIT14 384F13 Lec11 PDF
Lecture 11
VARs
Notation and some Linear Algebra
Let
yt =
p
aj ytj + et
(1)
j=1
where yt and et are k 1, and aj is k k. et is white noise with Eet et = and Eet es = 0
p
p
Lemma 1. yt is stationary if det Ik j =1 aj z j = 0 for all |z | 1, i.e. all roots of det Ik j=1 aj z j
are outside the unit circle.
Denition 2. Companion form of (1) is:
Yt = AYt1 + Et ,
where
Yt =
yt
yt1
...
ytp+1
,A=
a1
I
0
..
.
a2
0
I
..
.
...
...
0
ap
0
...
..
.
et
, Et = ..
vec(Y ) = (I (A A))
vec(E )
Cite as: Anna Mikusheva, course materials for 14.384 Time Series Analysis, Fall 2007. MIT OpenCourseWare (http://ocw.mit.edu),
Massachusetts Institute of Technology. Downloaded on [DD Month YYYY].
Estimation
Estimation
=
Lemma 3. MLE (with normal error assumption)= OLS equation-by-equation with
1
T
et et
Intuition: all variables are included in all equations, so there is nothing gained by doing SUR. This also
implies that OLS equation by equation is asymptotically ecient. The usual statements of consistency and
asymptotic normality hold, as well as OLS formulas for standard errors.
Granger Causality
Granger Causality is a misleading name. It would be better called Granger predictability.
Denition 4. y fails to Granger cause x if its not helpful in linear predicting x (in MSE sense). More
formally,
(xt+s |xt , ..., yt , yt1 , ...) , s > 0
(xt+s |xt , xt1 , ...) = MSE E
MSE E
(xt |z) denotes the best linear prediction of xt given z
where E
A test of Granger causality is to run OLS:
xt = 1 xt1 + ... + p xtp + 1 yt1 + ... + p ytp + et
and test H0 : 1 = 2 = ... = 0.
Note that:
Granger causality is not related to economic causality, its more about predictability.
There could be simultaneous casualty or omitted variable problems. For example, there may be a
variable z that causes both y and x but with the dierent lag(sluggish response). If one does not
include z (omitted variable), it may look like x causes y.
Forward looking (rational) expectations may even reverse the causality. For example, suppose analysts
rationally predict that a stock is going to pay high dividends tomorrow. That will provoke people
to buy the stock today, and the price will rise. In the data you would observe that the price rise is
followed by high dividends. So, we would nd that prices Granger cause dividends, even though it was
really that anticipated high dividends caused high prices. Or increase in orange juice price Granger
causes bad weather in Florida.
How to do Granger causality in multivariate case?
Assume y1t is k1 1 vector and y2t is k2 1. Assume that we have VAR system
y1t
A1 (L)y1t1 + A2 y2t1
e1t
=
+
y2t
B1 (L)y1t1 + B2 y2t1
e2t
Group of variables y2 fails to Granger cause y1 if A2 = 0. To perform this test we have to run unrestricted
regression y1t = A1 (L)y1t1 +A2 y2t1 +eut and restricted regression y1t = A1 (L)y1t1 +ert . Then we estimate
T
T
the corresponding variance-covariance matrix u = T1 t=1 etu etu and r = T1 t=1 ert ert . The test statistic
compares these matrices:
LR = T (log |r | log |r |)
Under the null (absence of Granger Causality) LR statistic is asymptotically 2 with the degrees of freedom
equal to the number of restrictions imposed.
Cite as: Anna Mikusheva, course materials for 14.384 Time Series Analysis, Fall 2007. MIT OpenCourseWare (http://ocw.mit.edu),
Massachusetts Institute of Technology. Downloaded on [DD Month YYYY].
Reporting Results
Impulse Response
2.5
2 orting Results
Rep
Amplitude
Reporting the matrix of coecients is not very informative. There are too many of them, and the coecients 1.5
are dicult to interprete anyway. Instead, people present impulse-response functions and variance
decompositions.
1
Impulse-resp
onse
Suppose
yt = a(L)yt1 + et
0.5
with MA representation
0
0
10
yt = c(L)et
Samples
Eu
u = I.
t t
15
20
25
yt = c(L)ut
Denition 5. The impulse-response function is
yti
.
k
utj
in uktj holding all other shocks constant. We can plot the impulse-response function as in gure 1.
To estimate an impulse-response, we would
1. Estimate VAR by OLS a
2. Invert to MA
3. Find and apply rotation D to get orthonormal shocks the impulse response is given by c
Cite as: Anna Mikusheva, course materials for 14.384 Time Series Analysis, Fall 2007. MIT OpenCourseWare (http://ocw.mit.edu),
Massachusetts Institute of Technology. Downloaded on [DD Month YYYY].
Impulse-response
Standard Errors
Delta-method To calculate standard errors, we can apply the delta-method to a
the c are just some
complicated function of a
. In practice, we can do this recursively:
yt =a1 yt1 + ... + ap ytp + et
=a1 (a1 yt2 + ... + ap ytp1 + et1 ) + a2 yt2 + ... + ytp + et
..
.
so, c1 = a1 , c2 = a2 + a21 , etc. We can apply the delta-method to each of these coecients. Wed also need to
apply the delta-method to our estimate of D. Sometimes, this is done in practice. However, it is not really
the best way, for two reasons:
We estimate many aj from not all that big of a sample, so our asymptotics may not be very good.
This is made even worse by the fact that ck are highly non-linear transformations of aj
Instead of the delta-method, we can use the bootstrap.
Bootstrap
=
c+
a1 yt1,b + ... +
ap ytp,b + et,b , where et,b is sampled randomly with replacement
(a) Form yt,b
from {et }
cj cj
s.e.(
cj )
cj c j
tj,([B(1/2)])
s.e.(cj )
or [
cj tj,([B(1/2)]) s.e.(
cj ),
cj tj,([B/2]) s.e.(
cj )]
Remark 6. The bootstrap is still an asymptotic procedure. One advantage of the bootstrap is its simplicity.
There is no need to apply the delta-method.
Remark 7. There are variations of the bootstrap that also work. For example, you could sample the errors
This would be called a parametric bootstrap because wed be
from a normal distribution with variance .
relying on a parametric assumption to create our simulated samples.
Cite as: Anna Mikusheva, course materials for 14.384 Time Series Analysis, Fall 2007. MIT OpenCourseWare (http://ocw.mit.edu),
Massachusetts Institute of Technology. Downloaded on [DD Month YYYY].
Impulse-response
Bootstrap-after-bootstrap Simulations show that bootstrap works for impulse responses somewhat better than asymptotic (delta-method). This is due to nal sample correction - remember that the dependence
between {aj } and {cj } is non-linear. However, the coverage of these intervals is still very far from ideal,
especially for very persistent processes. The main reason for that is aj are very biased estimates of aj . To
correct this a bootstrap-after-bootstrap was suggested.
a1 , ...,
ap and residuals et
1. run regression yt = c + a1 yt1 + ... + ap ytp + et to get c,
2. Invert the estimated AR process to get the estimates of impulse response cj from
a1 , ...,
ap
3. For b = 1..B
1
B
B
aj,b
b=1
5. For b = 1..B
(c) Invert the estimated AR process to get the estimates of impulse response cj,b from
a1,b , ...,
ap,b
6. Form condence interval.
Cite as: Anna Mikusheva, course materials for 14.384 Time Series Analysis, Fall 2007. MIT OpenCourseWare (http://ocw.mit.edu),
Massachusetts Institute of Technology. Downloaded on [DD Month YYYY].
MIT OpenCourseWare
http://ocw.mit.edu
For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.