Chapter5 Solutions

Download as doc, pdf, or txt
Download as doc, pdf, or txt
You are on page 1of 16

Solutions to the Review Questions at the End of Chapter 5

PACF- Another important measure is called partial autocorrelation, which is the correlation
between Xs and Xt with the linear effect of “everything in the middle” removed.

ACF- ACF measures the linear predictability of Xt using only Xs .

White Noise
– A simple time series could be a collection of uncorrelated random variables, {wt }, with zero mean
μ = 0 and finite variance σw2 , denoted as wt ∼ wn(0, σw2 ).

• Gaussian White Noise


– A particular useful white noise is Gaussian white noise, wherein the wt are independent normal
random variables (with mean 0 and variance σw2 ), denoted as wt ∼ iid N (0, σw2 ).

• White noise time series is of great interest because if the stochastic behavior of all time
series could be explained in terms of the white noise model, then classical statistical methods
would suffice.

The simplest AR process is AR(0), which has no dependence between the terms. In fact,
AR(0) is essentially white noise.

1. Autoregressive models specify the current value of a series yt as a function of


its previous p values and the current value an error term, ut, while moving
average models specify the current value of a series yt as a function of the
current and previous q values of an error term, ut. AR and MA models have
different characteristics in terms of the length of their “memories”, which has
implications for the time it takes shocks to yt to die away, and for the shapes of
their autocorrelation and partial autocorrelation functions.

2. ARMA models are of particular use for financial series due to their
flexibility. They are fairly simple to estimate, can often produce reasonable
forecasts, and most importantly, they require no knowledge of any structural
variables that might be required for more “traditional” econometric analysis.
When the data are available at high frequencies, we can still use ARMA
models while exogenous “explanatory” variables (e.g. macroeconomic
variables, accounting ratios) may be unobservable at any more than monthly
intervals at best.

1. Autoregressive загвар нь цуврал Yt-ийн одоогийн утгыг өмнөх p


утгын функц, одоогийн утгын алдааны нэр томъео, ut гэж зааж
өгдөг бол дундаж загваруудыг хөдөлгөхдөө цуврал YT-ийн
одоогийн утгыг алдааны нэр томъеоны одоогийн болон өмнөх q
утгын функц гэж тодорхойлдог.
AR ба MA загварууд нь "дурсамж" - ын уртын хувьд өөр өөр шинж
чанартай байдаг бөгөөд энэ нь YT-ийн цочролыг арилгах хугацаа,
тэдгээрийн автокорреляци ба хэсэгчилсэн Автокорреляцийн
функцүүдийн хувьд үр дагавартай байдаг.

1/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


AR- Autoregressive models are based on the idea that current value of the series, Xt, can be
explained as a linear combination of p past values, Xt−1, Xt−2, ..., Xt−p, together with a random error in
the same series.

An autoregressive model of order p, abbreviated AR(p), is of the form

MA- • The name might be misleading, but moving average models should not be confused
with the moving average smoothing.

• Motivation

 – Recall that in AR models, current observation Xt is regressed using the previous


observations Xt−1,Xt−2,...,Xt−p, plus an error term wt at current time point.

 – One problem of AR model is the ignorance of correlated noise structures (which is


unobservable) in the time series.

 – In other words, the imperfectly predictable terms in current time, wt, and previous
steps,wt−1,wt−2,...,wt−q, are also informative for predicting observations.

 An important property of MA(q) models in general is that there are nonzero


autocorrelations for the first q lags, and ρh =0foralllagsh>q.

 In other words, ACF provides a considerable amount of information about the order
of the dependence q for MA(q) process.

 Identification of an MA model is often best done with the ACF rather than the
PACF.

• Recall that we have seen for AR(1) process, if |φ1| < 1 and supt Var(Xt ) < ∞,


X
j
X t= φ 1 w t − j j=0

• In fact, all causal AR(p) processes can be represented as MA(∞); In other words, infinite
moving average processes are finite autoregressive processes.

• All invertible MA(q) processes can be represented as AR(∞). i.e. finite moving average
processes are infinite autoregressive processes.

2. ARMA загварууд нь уян хатан байдлаас шалтгаалан санхүүгийн


цувралд онцгой ашиглагддаг. Тэдгээр нь тооцоолоход нэлээд энгийн,
ихэвчлэн боломжийн урьдчилсан мэдээг гаргаж чаддаг б өг өөд хамгийн
чухал нь илүү "уламжлалт" эконометрикийн шинжилгээнд шаардагдах
бүтцийн хувьсагчдын талаар ямар ч мэдлэг шаарддаггүй. Өгөгд өл өнд өр
давтамжтай байх үед бид ARMA загварыг ашиглаж болох боловч экзоген
"тайлбар" хувьсагчид (жишээлбэл, макро эдийн засгийн хувьсагчид,
нягтлан бодох бүртгэлийн харьцаа) хамгийн сайн үед сарын интервалаас
илүү ямар ч үед ашиглах боломжгүй байж болно.

2/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


Autoregressive and moving average models can be combined together to form ARMA
models.

3. yt = yt-1 + ut (1)
yt = 0.5 yt-1 + ut (2)
yt = 0.8 ut-1 + ut (3)

(a) The first two models are roughly speaking AR(1) models, while the last is
an MA(1). Strictly, since the first model is a random walk, it should be called
an ARIMA(0,1,0) model, but it could still be viewed as a special case of an
autoregressive model.

(b) We know that the theoretical acf of an MA(q) process will be zero after q
lags, so the acf of the MA(1) will be zero at all lags after one. For an
autoregressive process, the acf dies away gradually. It will die away fairly
quickly for case (2), with each successive autocorrelation coefficient taking on
a value equal to half that of the previous lag. For the first case, however, the acf
will never die away, and in theory will always take on a value of one, whatever
the lag.

Turning now to the pacf, the pacf for the first two models would have a large
positive spike at lag 1, and no statistically significant pacf’s at other lags.
Again, the unit root process of (1) would have a pacf the same as that of a
stationary AR process. The pacf for (3), the MA(1), will decline geometrically.

(c) Clearly the first equation (the random walk) is more likely to represent
stock prices in practice. The discounted dividend model of share prices states
that the current value of a share will be simply the discounted sum of all
expected future dividends. If we assume that investors form their expectations
about dividend payments rationally, then the current share price should
embody all information that is known about the future of dividend payments,
and hence today’s price should only differ from yesterdays by the amount of
unexpected news which influences dividend payments.

Thus stock prices should follow a random walk. Note that we could apply a
similar rational expectations and random walk model to many other kinds of
financial series.

If the stock market really followed the process described by equations (2) or
(3), then we could potentially make useful forecasts of the series using our
model. In the latter case of the MA(1), we could only make one-step ahead
forecasts since the “memory” of the model is only that length. In the case of
equation (2), we could potentially make a lot of money by forming multiple
step ahead forecasts and trading on the basis of these.

Hence after a period, it is likely that other investors would spot this potential
opportunity and hence the model would no longer be a useful description of
the data.

3/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


(d) See the book for the algebra. This part of the question is really an extension
of the others. Analysing the simplest case first, the MA(1), the “memory” of
the process will only be one period, and therefore a given shock or
“innovation”, ut, will only persist in the series (i.e. be reflected in yt) for one
period. After that, the effect of a given shock would have completely worked
through.

For the case of the AR(1) given in equation (2), a given shock, ut, will persist
indefinitely and will therefore influence the properties of yt for ever, but its
effect upon yt will diminish exponentially as time goes on.

In the first case, the series yt could be written as an infinite sum of past
shocks, and therefore the effect of a given shock will persist indefinitely, and
its effect will not diminish over time.
( a) эхний хоер загвар нь ойролцоогоор AR(1) загварыг ярьж байгаа бол
сүүлчийнх нь MA(1) байна. Дархан, анхны загвар нь санамсаргүй алхалт
тул үүнийг ARIMA(0,1,0) загвар гэж нэрлэх естой боловч үүнийг
autoregressive загварын онцгой тохиолдол гэж үзэж болно.

(b) ма(q) процессын онолын acf нь q хоцролтын дараа тэг байх тул ма(1) -
ийн acf нь нэг бүрийн дараа бүх хоцролт дээр тэг болно гэдгийг бид
мэднэ. Ауторегрессив процессын хувьд acf нь аажмаар үхдэг. Энэ нь
Тохиолдол (2) - ийн хувьд нэлээд хурдан үхэх бөгөөд дараалсан
Автокорреляцийн коэффициент тус бүр нь өмнөх хоцролтын хагастай
тэнцүү утгыг авна. Гэхдээ эхний тохиолдолд АЦС хэзээ ч с өнөхг үй б өг өөд
онолын хувьд үргэлж нэг л үнэ цэнэтэй, ямар ч хоцрогдолтой байх болно.

Pacf одоо эргэж, эхний хоер загвар pacf лаг үед том эерэг баяжуулалтын
байх болно 1, болон бусад хожимдол нь ямар ч статистик ач
холбогдолтой pacf-ийн. Дахин хэлэхэд (1) - ийн нэгж эх процесс нь
суурин AR процесстой адил pacf-тэй байх болно. Pacf нь(3), MA (1) нь
геометрийн хувьд буурах болно.

( c) эхний тэгшитгэл (санамсаргүй алхалт) нь практик дээр хувьцааны


үнийг илэрхийлэх магадлал өндөр байдаг. Хувьцааны үнийн
хөнгөлөлттэй ногдол ашгийн загвар нь хувьцааны өнөөгийн үнэ цэнэ нь
ирээдүйн хүлээгдэж буй бүх ногдол ашгийн хөнгөлөлттэй нийлбэр байх
болно гэж заасан байдаг. Хэрэв хөрөнгө оруулагчид ногдол ашгийн
төлбөрийн талаархи хүлээлтийг оновчтой гэж үзвэл одоогийн хувьцааны
үнэ нь ногдол ашгийн төлбөрийн ирээдүйн талаар мэдэгдэж буй б үх
мэдээллийг агуулсан байх естой бөгөөд иймээс өнөөдрийн үнэ нь з өвхөн
өнгөрсөн өдрүүдээс ногдол ашгийн төлбөрт нөлөөлдөг гэнэтийн
мэдээний хэмжээгээр ялгаатай байх естой.

Тиймээс хувьцааны үнэ санамсаргүй алхаж дагах естой. Бид ижил т өстэй
оновчтой хүлээлт, санамсаргүй алхалтын загварыг бусад олон т өрлийн
санхүүгийн цувралд ашиглаж болохыг анхаарна уу.

Хэрэв хөрөнгийн зах зээл нь (2) эсвэл (3) тэгшитгэлээр тодорхойлсон үйл
явцыг үнэхээр дагаж мөрдвөл бид загвараа ашиглан цувралын ашигтай

4/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


урьдчилсан мэдээг гаргаж болно. MA(1)-ийн сүүлчийн тохиолдолд
загварын "санах ой" нь зөвхөн ийм урттай тул бид з өвх өн нэг алхам
урагшилж чадна. Тэгшитгэлийн хувьд (2), бид болзошг үй олон алхам
өмнө урьдчилсан бүрдүүлэх, эдгээрийн үндсэн дээр арилжаа хийх замаар
маш их мөнгө хийж болох юм.

Тиймээс хэсэг хугацааны дараа бусад хөр өнгө оруулагчид энэ боломжит
боломжийг олж харах магадлалтай бөгөөд ингэснээр загвар нь өг өгдлийн
ашигтай тайлбар байхаа болино.

(d) алгебрт зориулсан номыг үзнэ үү. Асуултын энэ хэсэг нь үнэхээр
бусдын өргөтгөл юм. Хамгийн энгийн тохиолдлыг шинжлэхэд
нэгдүгээрт, MA (1), үйл явцын" санах ой "нь зөвхөн нэг үе байх тул
өгөгдсөн цочрол буюу" инноваци " болох ut нь з өвхөн цувралд үлдэх
болно (өөрөөр хэлбэл yt-д тусгагдсан байх) нэг хугацаанд. Үүний дараа
тухайн цочролын нөлөө бүрэн ажиллах байсан.

Тэгшитгэл(2) - д өгөгдсөн AR (1) - ийн хувьд өгөгдс өн цочрол болох ut нь


тодорхойгүй хугацаагаар үргэлжлэх тул YT-ийн шинж чанарт м өнхөд
нөлөөлөх боловч цаг хугацаа өнгөрөх тусам yt-ийн нөлөө эрс буурах
болно.

Эхний тохиолдолд YT цувралыг өнгөрсөн цочролын хязгаарг үй нийлбэр


гэж бичиж болох тул тухайн цочролын нөлөө тодорхойг үй хугацаагаар
үргэлжлэх бөгөөд цаг хугацааны явцад Түүний нөлөө буурахгүй.

4. (a) Box and Jenkins were the first to consider ARMA modelling in this
logical and coherent fashion. Their methodology consists of 3 steps:
Identification - determining the appropriate order of the model using
graphical procedures (e.g. plots of autocorrelation functions).

Estimation - of the parameters of the model of size given in the first stage.
This can be done using least squares or maximum likelihood, depending on
the model.

Diagnostic checking - this step is to ensure that the model actually estimated
is “adequate”. B & J suggest two methods for achieving this:

- Overfitting, which involves deliberately fitting a model larger than


that suggested in step 1 and testing the hypothesis that all the additional
coefficients can jointly be set to zero.

- Residual diagnostics. If the model estimated is a good description of


the data, there should be no further linear dependence in the residuals of the
estimated model. Therefore, we could calculate the residuals from the
estimated model, and use the Ljung-Box test on them, or calculate their acf. If
either of these reveal evidence of additional structure, then we assume that the
estimated model is not an adequate description of the data.

5/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


If the model appears to be adequate, then it can be used for policy analysis
and for constructing forecasts. If it is not adequate, then we must go back to
stage 1 and start again!

(b) The main problem with the B & J methodology is the inexactness of the
identification stage. Autocorrelation functions and partial autocorrelations for
actual data are very difficult to interpret accurately, rendering the whole
procedure often little more than educated guesswork. A further problem
concerns the diagnostic checking stage, which will only indicate when the
proposed model is “too small” and would not inform on when the model
proposed is “too large”.

(c) We could use Akaike’s or Schwarz’s Bayesian information criteria. Our


objective would then be to fit the model order that minimises these.

We can calculate the value of Akaike’s (AIC) and Schwarz’s (SBIC) Bayesian
information criteria using the following respective formulae

AIC = ln ( ) + 2k/T
SBIC = ln ( ) + k ln(T)/T

The information criteria trade off an increase in the number of parameters


and therefore an increase in the penalty term against a fall in the RSS,
implying a closer fit of the model to the data.

(a) Хайрцаг ба Женкинс энэ логик, уялдаа холбоотой загвараар ARMA


загварчлалыг хамгийн түрүүнд авч үзсэн. Тэдний аргачлал нь 3 алхамаас
бүрдэнэ:
Тодорхойлох-график процедурыг ашиглан загварын зохих дарааллыг
тодорхойлох (жишээлбэл, Автокорреляцийн функцүүдийн талбай).

Тооцоо-эхний шатанд өгөгдсөн хэмжээ загварын Үзүүлэлтийн. Үүнийг


загвараас хамааран хамгийн бага квадрат эсвэл хамгийн их магадлалыг
ашиглан хийж болно.

Оношлогоо шалгах-энэ алхам нь үнэндээ тооцоолсон загвар нь


"хангалттай"гэдгийг баталгаажуулах явдал юм. B & J нь үүнд хүрэх хоер
аргыг санал болгож байна:

- Overfitting, санаатайгаар алхам 1-д санал болгосон гэж ил үү том


загварыг холбох, бүх нэмэлт коэффициент хамтран тэг тохируулж болно
гэсэн таамаглалыг шалгах явдал юм.

- Үлдэгдэл оношлогоо. Хэрэв тооцоолсон загвар нь өгөгдлийн сайн


тодорхойлолт юм бол тооцоолсон загварын резидент үүдэд цаашид
шугаман хамаарал байх есгүй. Тиймээс бид тооцоолсон загвараас
residuals-ийг тооцоолж, дээр нь Ljung-Box тестийг ашиглах эсвэл acf-ийг
тооцоолох боломжтой. Хэрэв эдгээрийн аль нэг нь нэмэлт б үтцийн
нотолгоог илчлэх юм бол тооцоолсон загвар нь өгөгдлийн хангалттай
тодорхойлолт биш гэж бид үзэж байна.

6/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


Хэрэв загвар нь хангалттай харагдаж байвал бодлогын дүн шинжилгээ
хийх, урьдчилсан таамаглал гаргахад ашиглаж болно. Хэрэв энэ нь
хангалтгүй бол бид 1-р шат руу буцаж очоод дахин эхлэх естой!

(b) B & J аргачлалын гол асуудал нь тодорхойлох шатны х өд өлб өрг үй


байдал юм. Autocorrelation чиг үүрэг, бодит өгөгдлийн хэсэгчилсэн
autocorrelations үнэн зөв тайлбарлах нь маш хэцүү байдаг, боловсролтой
таамаглал илүү ихэвчлэн бага зэрэг илүү бүх журмыг амилуулж.
Цаашдын асуудал нь оношлогооны шалгалтын үе шаттай холбоотой
бөгөөд энэ нь санал болгож буй загвар нь "хэтэрхий жижиг" байх үед л
зааж өгөх бөгөөд санал болгож буй загвар нь "хэтэрхий том"байх үед
мэдээлэхгүй болно.

(c) бид Akaike-ийн эсвэл Schwarz-ийн Bayesian мэдээллийн шалгуурыг


ашиглаж болно. Дараа нь бидний зорилго бол эдгээрийг багасгах
загварын захиалгад багтах явдал юм.

Бид Akaike-ийн (AIC) болон Schwarz-ийн (SBIC) Bayesian-ийн


мэдээллийн шалгуурын утгыг дараах холбогдох формулуудыг ашиглан
тооцоолж болно

AIC = ln () + 2k / T
SBIC = ln () + k ln(T) / T

Мэдээллийн шалгуур үзүүлэлтүүд нь параметрийн тоо нэмэгдэж, улмаар


RSS-ийн уналтын эсрэг торгуулийн хугацаа нэмэгдэж байгаа нь загварт
өгөгдөлд илүү ойр нийцэж байгааг илтгэнэ.

5. The best way to check for stationarity is to express the model as a lag
polynomial in yt.

Rewrite this as

We want to find the roots of the lag polynomial and


determine whether they are greater than one in absolute value. It is easier (in
my opinion) to rewrite this formula (by multiplying through by -1/0.682,
using z for the characteristic equation and rearranging) as

z2 + 1.177 z - 1.466 = 0

Using the standard formula for obtaining the roots of a quadratic equation,

= 0.758 or 1.934

7/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


Since ALL the roots must be greater than one for the model to be stationary,
we conclude that the estimated model is not stationary in this case.

6. Using the formulae above, we end up with the following values for each
criterion and for each model order (with an asterisk denoting the smallest
value of the information criterion in each case).

ARMA (p,q) model order log ( ) AIC SBIC


(0,0) 0.932 0.942 0.944
(1,0) 0.864 0.884 0.887
(0,1) 0.902 0.922 0.925
(1,1) 0.836 0.866 0.870
(2,1) 0.801 0.841 0.847
(1,2) 0.821 0.861 0.867
(2,2) 0.789 0.839 0.846
(3,2) 0.773 0.833*
0.842*
(2,3) 0.782 0.842 0.851
(3,3) 0.764 0.834 0.844

The result is pretty clear: both SBIC and AIC say that the appropriate model is
an ARMA(3,2).

7. We could still perform the Ljung-Box test on the residuals of the estimated
models to see if there was any linear dependence left unaccounted for by our
postulated models.

Another test of the models’ adequacy that we could use is to leave out some of
the observations at the identification and estimation stage, and attempt to
construct out of sample forecasts for these. For example, if we have 2000
observations, we may use only 1800 of them to identify and estimate the
models, and leave the remaining 200 for construction of forecasts. We would
then prefer the model that gave the most accurate forecasts.

Бид одоо ч гэсэн манай postulated загвар тооцоог үй үлдсэн ямар нэгэн
шугаман хамааралтай байсан эсэхийг тооцоолсон загвар residuals дээр
Ljung хайрцаг туршилтыг хийж болох юм.

Бидний ашиглаж болох загваруудын зохистой байдлын өөр нэг туршилт


бол тодорхойлох, тооцоолох үе шатанд зарим ажиглалтыг орхиж,
эдгээрийн дээжийн урьдчилсан таамаглалаас гаргахыг оролдох явдал
юм. Жишээлбэл, хэрэв бид 2000 ажиглалттай бол з өвхөн 1800 загварыг
тодорхойлж, тооцоолоход ашиглаж болох бөгөөд үлдсэн 200-г
урьдчилсан таамаглал гаргахад үлдээж болно. Дараа нь бид хамгийн з өв
урьдчилсан мэдээг өгсөн загварыг илүүд үзэх болно.

8. This is not true in general. Yes, we do want to form a model which “fits” the
data as well as possible. But in most financial series, there is a substantial

8/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


amount of “noise”. This can be interpreted as a number of random events that
are unlikely to be repeated in any forecastable way. We want to fit a model to
the data which will be able to “generalise”. In other words, we want a model
which fits to features of the data which will be replicated in future; we do not
want to fit to sample-specific noise.

This is why we need the concept of “parsimony” - fitting the smallest possible
model to the data. Otherwise we may get a great fit to the data in sample, but
any use of the model for forecasts could yield terrible results.

Another important point is that the larger the number of estimated


parameters (i.e. the more variables we have), then the smaller will be the
number of degrees of freedom, and this will imply that coefficient standard
errors will be larger than they would otherwise have been. This could lead to a
loss of power in hypothesis tests, and variables that would otherwise have
been significant are now insignificant.

Энэ нь ерөнхийдөө үнэн биш юм. Тийм ээ, бид аль болох өг өгдлийг
"тохирох" загвар бүрдүүлэхийг хүсч байна. Гэхдээ ихэнх санхүүгийн
цувралд ихээхэн хэмжээний "дуу чимээ"байдаг. Үүнийг урьдчилан
таамаглах боломжгүй байдлаар давтагдах магадлал багатай хэд хэдэн
санамсаргүй үйл явдал гэж ойлгож болно. Бид "ер өнхийлөн" боломжтой
байх болно өгөгдөлд загварыг тохирох хүсэж байна. Өөр өөр хэлбэл, бид
ирээдүйд хуулбарлагдах өгөгдлийн онцлог шинж чанарт тохирсон
загварыг хүсч байна; Бид түүврийн тодорхой дуу чимээнд тохирохыг
хүсэхгүй байна.

Ийм учраас бидэнд "parsimony" гэсэн ойлголт хэрэгтэй - өг өгд өлд


хамгийн бага боломжит загварыг тохируулах. Үгүй бол бид дээж дэх
өгөгдөлд маш сайн тохирох болно, гэхдээ урьдчилсан мэдээний загварыг
ашиглах нь аймшигтай үр дүнд хүргэж болзошгүй юм.

Өөр нэг чухал зүйл бол тооцоолсон параметр үүдийн тоо их байх тусам
(өөрөөр хэлбэл бидэнд илүү олон хувьсагч байх тусам) бага байх нь
чөлөөний градусын тоо байх бөгөөд энэ нь коэффициентийн стандарт
алдаа нь өөрөөр байснаас их байх болно гэсэн үг юм. Энэ нь таамаглалын
туршилтанд хүч чадал алдагдахад хүргэж болзошгүй бөгөөд өөр өөр
хэлбэл ач холбогдолтой байсан хувьсагчид одоо ач холбогдолг үй болжээ.

9. (a) We class an autocorrelation coefficient or partial autocorrelation


coefficient as significant if it exceeds =  0.196. Under this rule, the
sample autocorrelation functions (sacfs) at lag 1 and 4 are significant, and the
spacfs at lag 1, 2, 3, 4 and 5 are all significant.

This clearly looks like the data are consistent with a first order moving average
process since all but the first acfs are not significant (the significant lag 4 acf is
a typical wrinkle that one might expect with real data and should probably be
ignored), and the pacf has a slowly declining structure.

9/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


(b) The formula for the Ljung-Box Q* test is given by

using the standard notation.

In this case, T=100, and m=3. The null hypothesis is H0: 1 = 0 and 2 = 0 and
3 = 0. The test statistic is calculated as

The 5% and 1% critical values for a 2 distribution with 3 degrees of freedom


are 7.81 and 11.3 respectively. Clearly, then, we would reject the null
hypothesis that the first three autocorrelation coefficients are jointly not
significantly different from zero.

10. (a) To solve this, we need the concept of a conditional expectation,

i.e.

For example, in the context of an AR(1) model such as ,


If we are now at time t-1, and dropping the t-1 subscript on the expectations
operator

=
=

=
=

=
etc.

10/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


To forecast an MA model, consider, e.g.

= E

So ft-1,1 =

But
= E
= 0

Going back to the example above,

Suppose that we know t-1, t-2,... and we are trying to forecast yt.
Our forecast for t is given by

=
= 0.036 +0.693.4+0.42(-1.3)
= 1.836
ft-1,2 =

But we do not know yt or ut at time t-1.

Replace yt with our forecast of yt which is ft-1,1.

ft-1,2 = 0.036 +0.69 ft-1,1


= 0.036 + 0.69*1.836
= 1.302

ft-1,3 = 0.036 +0.69 ft-1,2


= 0.036 + 0.69*1.302
= 0.935
etc.

(b) Given the forecasts and the actual value, it is very easy to calculate the
MSE by plugging the numbers in to the relevant formula, which in this case is

if we are making N forecasts which are numbered 1,2,3.


Then the MSE is given by

11/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


Notice also that 84% of the total MSE is coming from the error in the first
forecast. Thus error measures can be driven by one or two times when the
model fits very badly. For example, if the forecast period includes a stock
market crash, this can lead the mean squared error to be 100 times bigger than
it would have been if the crash observations were not included. This point
needs to be considered whenever forecasting models are evaluated. An idea of
whether this is a problem in a given situation can be gained by plotting the
forecast errors over time.

(c) This question is much simpler to answer than it looks! In fact, the
inclusion of the smoothing coefficient is a “red herring” - i.e. a piece of
misleading and useless information. The correct approach is to say that if we
believe that the exponential smoothing model is appropriate, then all useful
information will have already been used in the calculation of the current
smoothed value (which will of course have used the smoothing coefficient in
its calculation). Thus the three forecasts are all 0.0305.

(d) The solution is to work out the mean squared error for the exponential
smoothing model. The calculation is

Therefore, we conclude that since the mean squared error is smaller for the
exponential smoothing model than the Box Jenkins model, the former
produces the more accurate forecasts. We should, however, bear in mind that
the question of accuracy was determined using only 3 forecasts, which would
be insufficient in a real application.

11. (a) The shapes of the acf and pacf are perhaps best summarised in a table:

Process acf pacf


White No significant coefficients No significant coefficients
noise
AR(2) Geometrically declining or First 2 pacf coefficients
damped sinusoid acf significant, all others
insignificant
MA(1) First acf coefficient significant, Geometrically declining or
all others insignificant damped sinusoid pacf
ARMA(2 Geometrically declining or Geometrically declining or
,1) damped sinusoid acf damped sinusoid pacf

12/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


A couple of further points are worth noting. First, it is not possible to tell what
the signs of the coefficients for the acf or pacf would be for the last three
processes, since that would depend on the signs of the coefficients of the
processes. Second, for mixed processes, the AR part dominates from the point
of view of acf calculation, while the MA part dominates for pacf calculation.

(b) The important point here is to focus on the MA part of the model and to
ignore the AR dynamics. The characteristic equation would be

(1+0.42z) = 0

The root of this equation is -1/0.42 = -2.38, which lies outside the unit circle,
and therefore the MA part of the model is invertible.

(c) Since no values for the series y or the lagged residuals are given, the
answers should be stated in terms of y and of u. Assuming that information is
available up to and including time t, the 1-step ahead forecast would be for
time t+1, the 2-step ahead for time t+2 and so on. A useful first step would be
to write the model out for y at times t+1, t+2, t+3, t+4:

The 1-step ahead forecast would simply be the conditional expectation of y for
time t+1 made at time t. Denoting the 1-step ahead forecast made at time t as
ft,1, the 2-step ahead forecast made at time t as ft,2 and so on:

since Et[ut+1]=0. The 2-step ahead forecast would be given by

since Et[ut+1]=0 and Et[ut+2]=0. Thus, beyond 1-step ahead, the MA(1) part of
the model disappears from the forecast and only the autoregressive part
remains. Although we do not know yt+1, its expected value is the 1-step ahead
forecast that was made at the first stage, ft,1.

The 3-step ahead forecast would be given by

and the 4-step ahead by

13/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


(d) A number of methods for aggregating the forecast errors to produce a
single forecast evaluation measure were suggested in the paper by Makridakis
and Hibon (1995) and some discussion is presented in the book. Any of the
methods suggested there could be discussed. A good answer would present an
expression for the evaluation measures, with any notation introduced being
carefully defined, together with a discussion of why the measure takes the
form that it does and what the advantages and disadvantages of its use are
compared with other methods.

(e) Moving average and ARMA models cannot be estimated using OLS – they
are usually estimated by maximum likelihood. Autoregressive models can be
estimated using OLS or maximum likelihood. Pure autoregressive models
contain only lagged values of observed quantities on the RHS, and therefore,
the lags of the dependent variable can be used just like any other regressors.
However, in the context of MA and mixed models, the lagged values of the
error term that occur on the RHS are not known a priori. Hence, these
quantities are replaced by the residuals, which are not available until after the
model has been estimated. But equally, these residuals are required in order to
be able to estimate the model parameters. Maximum likelihood essentially
works around this by calculating the values of the coefficients and the
residuals at the same time. Maximum likelihood involves selecting the most
likely values of the parameters given the actual data sample, and given an
assumed statistical distribution for the errors. This technique will be
discussed in greater detail in the section on volatility modelling in Chapter 8.

12. (a) Some of the stylised differences between the typical characteristics of
macroeconomic and financial data were presented in Chapter 1. In particular,
one important difference is the frequency with which financial asset return
time series and other quantities in finance can be recorded. This is of
particular relevance for the models discussed in Chapter 5, since it is usually a
requirement that all of the time-series data series used in estimating a given
model must be of the same frequency. Thus, if, for example, we wanted to
build a model for forecasting hourly changes in exchange rates, it would be
difficult to set up a structural model containing macroeconomic explanatory
variables since the macroeconomic variables are likely to be measured on a
quarterly or at best monthly basis. This gives a motivation for using pure time-
series approaches (e.g. ARMA models), rather than structural formulations
with separate explanatory variables.

It is also often of particular interest to produce forecasts of financial variables


in real time. Producing forecasts from pure time-series models is usually
simply an exercise in iterating with conditional expectations. But producing
forecasts from structural models is considerably more difficult, and would
usually require the production of forecasts for the structural variables as well.

(b) A simple “rule of thumb” for determining whether autocorrelation


coefficients and partial autocorrelation coefficients are statistically significant
is to classify them as significant at the 5% level if they lie outside of

14/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


, where T is the sample size. In this case, T = 500, so a particular
coefficient would be deemed significant if it is larger than 0.088 or smaller
than –0.088. On this basis, the autocorrelation coefficients at lags 1 and 5 and
the partial autocorrelation coefficients at lags 1, 2, and 3 would be classed as
significant. The formulae for the Box-Pierce and the Ljung-Box test statistics
are respectively

and .

In this instance, the statistics would be calculated respectively as

and

The test statistics will both follow a 2 distribution with 5 degrees of freedom
(the number of autocorrelation coefficients being used in the test). The critical
values are 11.07 and 15.09 at 5% and 1% respectively. Clearly, the null
hypothesis that the first 5 autocorrelation coefficients are jointly zero is
resoundingly rejected.

(c) Setting aside the lag 5 autocorrelation coefficient, the pattern in the table is
for the autocorrelation coefficient to only be significant at lag 1 and then to fall
rapidly to values close to zero, while the partial autocorrelation coefficients
appear to fall much more slowly as the lag length increases. These
characteristics would lead us to think that an appropriate model for this series
is an MA(1). Of course, the autocorrelation coefficient at lag 5 is an anomaly
that does not fit in with the pattern of the rest of the coefficients. But such a
result would be typical of a real data series (as opposed to a simulated data
series that would have a much cleaner structure). This serves to illustrate that
when econometrics is used for the analysis of real data, the data generating
process was almost certainly not any of the models in the ARMA family. So all
we are trying to do is to find a model that best describes the features of the
data to hand. As one econometrician put it, all models are wrong, but some are
useful!

(d) Forecasts from this ARMA model would be produced in the usual way.
Using the same notation as above, and letting fz,1 denote the forecast for time
z+1 made for x at time z, etc:

Model A: MA(1)

15/16 “Introductory Econometrics for Finance” © Chris Brooks 2008


Note that the MA(1) model only has a memory of one period, so all forecasts
further than one step ahead will be equal to the intercept.

Model B: AR(2)

(e) The methods are overfitting and residual diagnostics. Overfitting involves
selecting a deliberately larger model than the proposed one, and examining
the statistical significances of the additional parameters. If the additional
parameters are statistically insignificant, then the originally postulated model
is deemed acceptable. The larger model would usually involve the addition of
one extra MA term and one extra AR term. Thus it would be sensible to try an
ARMA(1,2) in the context of Model A, and an ARMA(3,1) in the context of
Model B. Residual diagnostics would involve examining the acf and pacf of the
residuals from the estimated model. If the residuals showed any “action”, that
is, if any of the acf or pacf coefficients showed statistical significance, this
would suggest that the original model was inadequate. “Residual diagnostics”
in the Box-Jenkins sense of the term involved only examining the acf and
pacf, rather than the array of diagnostics considered in Chapter 4.

It is worth noting that these two model evaluation procedures would only
indicate a model that was too small. If the model were too large, i.e. it had
superfluous terms, these procedures would deem the model adequate.

(f) There are obviously several forecast accuracy measures that could be
employed, including MSE, MAE, and the percentage of correct sign
predictions. Assuming that MSE is used, the MSE for each model is

Therefore, since the mean squared error for Model A is smaller, it would be
concluded that the moving average model is the more accurate of the two in
this case.

16/16 “Introductory Econometrics for Finance” © Chris Brooks 2008

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy