Ch6 Slides Ed3 Feb2021
Ch6 Slides Ed3 Feb2021
Ch6 Slides Ed3 Feb2021
1
„Introductory Econometrics for Finance‟ © Chris Brooks 2008
„Introductory Econometrics for Finance‟ © Chris Brooks 2008
I. Time Series: Introduction
“Don’t never prophesy: if you prophesy right, nobody is going to remember and if
you prophesy wrong, nobody is going to let you forget”
Mark Twain
• So if the process is covariance stationary, all the variances are the same and all
the covariances depend on the difference between t1 and t2. The moments
E ( yt E ( yt ))( yt s E ( yt s )) s , s = 0,1,2, ...
are known as autocovariances.
• s approximately N(0,1/T) where T = sample size and s denotes the
autocorrelation at lag s
• We can use this to do significance tests for the autocorrelation coefficients
by constructing a confidence interval.
1
• For example, a 95% confidence interval would be given by .196
T.
• If the sample autocorrelation coefficient, s , falls outside this region for
any value of s, then we reject the null hypothesis that the true value of the
autocorrelation coefficient at lag s is zero.
Joint Hypothesis Tests: 1 =0, 2=0,…, m =0
(Data are independently distributed)
• We can also test the joint hypothesis that all m of the k autocorrelation
coefficients are simultaneously equal to zero using the Q-statistic developed
by Box and Pierce: m
Q T k2
k 1
where T = sample size, m = maximum lag length
• The Q-statistic is asymptotically distributed as a m.
2
• However, the Box Pierce test has poor small sample properties, so a variant
has been developed, called the Ljung-Box statistic:
m
k2
Q T T 2
~ m2
k 1 T k
• This statistic is very useful as a portmanteau (general) test of linear
dependence in time series.
An ACF Example
• Question:
Suppose that a researcher had estimated the first 5 autocorrelation coefficients
using a series of length 100 observations, and found them to be (from 1 to 5):
0.207, -0.013, 0.086, 0.005, -0.022.
a) Test each of the individual coefficient for significance
b) Use both the Box-Pierce and Ljung-Box tests to establish whether
they are jointly significant.
• Solution:
a) A coefficient would be significant if it lies outside (-0.196,+0.196) at the
5% level, so only the first autocorrelation coefficient is significant.
b) Q=5.09 and Q*=5.26
Compared with a tabulated 2(5)=11.07 at the 5% level, so the 5 coefficients
are jointly insignificant (low power of the joint test when 1 single
coefficient is significant).
III. Moving Average Processes
Var(Xt) = E[Xt-E(Xt)][Xt-E(Xt)]
but E(Xt) = 0, so
Var(Xt) = E[(Xt)(Xt)]
= E[(ut + 1ut-1+ 2ut-2)(ut + 1ut-1+ 2ut-2)]
= E[ u t2 12 u t21 22 u t2 2 +cross-products]
So Var(Xt) = 0= E [ ]
=
= (1 1 2 )
2 2 2
= 1 2 1 2 2
= ( 1 1 2 ) 2
Solution (cont’d)
2 = E[Xt-E(Xt)][Xt-2-E(Xt-2)]
= E[Xt][Xt-2]
= E[(ut +1ut-1+2ut-2)(ut-2 +1ut-3+2ut-4)]
= E[( 2 u t2 2 )]
= 2
2
3 = E[Xt-E(Xt)][Xt-3-E(Xt-3)]
= E[Xt][Xt-3]
= E[(ut +1ut-1+2ut-2)(ut-3 +1ut-4+2ut-5)]
=0
So s = 0 for s > 2.
Solution (cont’d)
(iii) For 1 = -0.5 and 2 = 0.25, substituting these into the formulae above
gives 1 = -0.476, 2 = 0.190.
ACF Plot
1.2
0.8
0.6
0.4
acf
0.2
0
0 1 2 3 4 5 6
-0.2
-0.4
-0.6
s
IV. Autoregressive Processes
y t 1 y t 1 2 y t 2 ... p y t p u t
where ut is a white noise process with zero mean.
• Or using the lag operator notation:
Lyt = yt-1 Liyt = yt-i
p
y t i y t i u t
i 1
p
• or y t i Li y t u t
i 1
y t ( L)u t
where,
( L) (1 1 L 2 L2 ... p Lp ) 1
Characteristics of an Autoregressive Process
yt 1 yt 1 ut
(i) Given that the series is stationary, calculate the (unconditional)
mean of yt
• So kk measures the correlation between yt and yt-k after removing the effects
of yt-k+1 , yt-k+2 , …, yt-1 .
• The PACF is useful for telling the difference between an AR process and
an ARMA process.
• In the case of an AR(p), there are direct connections between yt and yt-s only
for s p.
where ( L) 1 1 L 2 L2 ... p Lp
or y t 1 y t 1 2 y t 2 ... p y t p 1u t 1 2 u t 2 ... q u t q u t
0
1 2 3 4 5 6 7 8 9 10
-0.05
-0.1
-0.15
acf and pacf
-0.2
-0.25
-0.3
acf
-0.35
pacf
-0.4
-0.45
Lag
ACF and PACF for an MA(2) Model:
yt = 0.5ut-1 - 0.25ut-2 + ut
0.4
0.3 acf
pacf
0.2
0.1
acf and pacf
0
1 2 3 4 5 6 7 8 9 10
-0.1
-0.2
-0.3
-0.4
Lags
ACF and PACF for a slowly decaying AR(1) Model:
yt = 0.9yt-1 + ut
0.9
acf
0.8 pacf
0.7
0.6
acf and pacf
0.5
0.4
0.3
0.2
0.1
0
1 2 3 4 5 6 7 8 9 10
-0.1
Lags
ACF and PACF for a more rapidly decaying AR(1)
Model: yt = 0.5yt-1 + ut
0.6
0.5
acf
pacf
0.4
acf and pacf
0.3
0.2
0.1
0
1 2 3 4 5 6 7 8 9 10
-0.1
Lags
ACF and PACF for a more rapidly decaying AR(1)
Model with Negative Coefficient: yt = -0.5yt-1 + ut
0.3
0.2
0.1
0
1 2 3 4 5 6 7 8 9 10
acf and pacf
-0.1
-0.2
-0.3
-0.4
acf
-0.5 pacf
-0.6
Lags
ACF and PACF for a Non-stationary Model
(i.e. a unit coefficient): yt = yt-1 + ut
0.9
acf
pacf
0.8
0.7
acf and pacf
0.6
0.5
0.4
0.3
0.2
0.1
0
1 2 3 4 5 6 7 8 9 10
Lags
ACF and PACF for an ARMA(1,1):
yt = 0.5yt-1 + 0.5ut-1 + ut
0.8
0.6
acf
pacf
0.4
acf and pacf
0.2
0
1 2 3 4 5 6 7 8 9 10
-0.2
-0.4
Lags
VII. Building ARMA Models
- The Box Jenkins Approach
• Box and Jenkins (1970) were the first to approach the task of estimating
an ARMA model in a systematic manner. There are 3 steps to their
approach:
1. Identification
2. Estimation
3. Model diagnostic checking
Step 1: Identification
Step 2: Estimation
- Estimation of the parameters
- Can be done using least squares or maximum likelihood depending
on the model.
• The information criteria vary according to how stiff the penalty term is.
• The three most popular criteria are Akaike’s (1974) information criterion (AIC),
Schwarz’s (1978) Bayesian information criterion (SBIC), and the Hannan-
Quinn criterion (HQIC).
AIC ln( 2 ) 2 k / T
k
SBIC ln(ˆ 2 ) ln T
T
2k
HQIC ln(ˆ 2 ) ln(ln(T ))
T
where k = p + q + 1, T = sample size. So we min. IC s.t. p p, q q
SBIC embodies a stiffer penalty term than AIC.
• Which IC should be preferred if they suggest different model orders?
– SBIC is strongly consistent but inefficient.
– AIC is not consistent, but more efficient, and will typically pick “bigger”
models.
ARIMA Models
• Examples:
- Forecasting tomorrow‟s return on a particular share
- Forecasting the price of a house given its characteristics
- Forecasting the riskiness of a portfolio over the next year
- Forecasting the volatility of bond returns
• The distinction between the two types is somewhat blurred (e.g, VARs).
In-Sample Versus Out-of-Sample
• Say we have some data - e.g. monthly FTSE returns for 120 months:
1990M1 – 1999M12. We could use all of it to build the model, or keep
some observations back:
• A good test of the model since we have not used the information from
1999M1 onwards when we estimated the model parameters.
How to produce forecasts
• Structural models
e.g. y = X + u
yt 1 2 x2t k xkt ut
To forecast y, we require the conditional expectation of its future
E yt t 1 E 1 2 x2t k xkt ut
value:
= 1 2 E x2t k E xkt
But what are ( x 2t ) etc.? We could use x 2 , so
E yt 1 2 x 2 k x k
= y !!!
problem!!!
Models for Forecasting (cont’d)
The current value of a series, yt, is modelled as a function only of its previous
values and the current value of an error term (and possibly previous values of
the error term).
• Models include:
• simple unweighted averages
• exponentially weighted averages
• ARIMA models
• Non-linear models – e.g. threshold models, GARCH, bilinear models, etc.
Forecasting with ARMA Models
ft, 4 = E(yt+4 t ) =
• For example, say we predict that tomorrow‟s return on the FTSE will be 0.2,
but the outcome is actually -0.4. Is this accurate? Define ft,s as the forecast made
at time t for s steps ahead (i.e. the forecast made for time t+s), and yt+s as the
realised value of y at time t+s.
• Some of the most popular criteria for assessing the accuracy of time series
forecasting techniques are:
N
1
Mean Squared Error:
MSE
N
t 1
( yt s f t , s ) 2
N
Mean Absolute Error: MAE 1 yt s f t , s
N t 1
1 N yt s ft ,s
Mean Absolute Percentage Error: MAPE 100
N t 1 yt s
Example
How can we test whether a forecast is accurate or not?
(cont’d)
• It has, however, also recently been shown (Gerlow et al., 1993) that the
accuracy of forecasts according to traditional statistical criteria are not
related to trading profitability.
1 N
% correct sign predictions = zt s
N t 1
• Given the following forecast and actual values, calculate the MSE, MAE and
percentage of correct sign predictions:
• Optimal Approach
To use a statistical forecasting model built on solid theoretical
foundations supplemented by expert judgements and interpretation.