0% found this document useful (0 votes)
13 views

vertopal.com_econometrics_module_4_lesson_3

This lesson covers the autoregressive (AR) model in financial econometrics, focusing on its definition, properties, and comparison with moving average models. It introduces the AR(p) model, estimation methods including Ordinary Least Squares and the Method of Moments, and discusses the significance of ACF and PACF plots in model selection. The lesson also highlights the conditions for stationarity and the representation of moving average processes as autoregressive processes.

Uploaded by

Cliff Boogië
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
13 views

vertopal.com_econometrics_module_4_lesson_3

This lesson covers the autoregressive (AR) model in financial econometrics, focusing on its definition, properties, and comparison with moving average models. It introduces the AR(p) model, estimation methods including Ordinary Least Squares and the Method of Moments, and discusses the significance of ACF and PACF plots in model selection. The lesson also highlights the conditions for stationarity and the representation of moving average processes as autoregressive processes.

Uploaded by

Cliff Boogië
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

econometrics_module_4_lesson_3

FINANCIAL ECONOMETRICS
MODULE 4 LESSON 3

AUTOREGRESSIVE MODEL
@>
p(- 2) * >
p(- 2) * @
Reading Time & 90 minutes
Prior Knowledge & Basic Time Series knowledge, White noise, Random walk
Keywords & Autoregressive AR(p) process or model, Autoregressive operator, Mean,
Variance, Autocovariance and Autocorrelations for AR(1), Invertible MA, Method of
Moments, First order moment, second order moment, Yule-Walker estimation, Maximum
Likelihood estimation (MLE)

In the last lesson, we discussed white noise, random walk, and moving average (MA) time
series models. In this lesson, we will continue to explore more time series models. Specifically,
we will talk about the autoregressive model (AR model). We will go over the definition and
properties of the AR model. We will also compare the autoregressive model with the moving
average model. We will end the lesson with one application.

1. Ordinary Least Square Regression for Time Series Analysis


From previous lessons, we know that often observations in a time series are correlated
with each other. If we want to use the past observations to predict today’s observations, the
first thing that comes to mind is to use OLS (ordinary least square regression). In regular
OLS, we have a dependent variable Y and an independent variable X . We can write a linear
regression as follows:
Y =C +b 1 X 1 + ε

where C is the intercept and ε is the error term.


To apply OLS to a time series X t as a dependent variable, and its lag 1, X t −1 as an
independent variable, we can write the OLS as follows:
X t =C +b1 X t − 1+ ε t

where C is the intercept and ε t is the normal white noise and absolute value of b 1<1 .
We then can apply OLS to estimate the coefficients and the residuals and to conduct model
diagnostics. The above regression model for time series X t is called an autoregressive
(AR) model.

2. Autoregressive (AR) Model


An autoregressive model is a regression model to predict a time series using a linear
combination of its pass values as independent variables, like what we demonstrated in the
last section for time series X t . We set up a regression model to predict X t using its lag 1
variable X t −1 as the independent variable. Because there is only one lag 1 variable on the
right side of the equation, we also denote this autoregressive model as AR(1).
We just described the concept of an autoregressive model. Now let’s formally define an
autoregressive model.

2.1 Definition of Autoregressive Model AR( p)


The formula for an autoregressive model with the order of p, denoted as AR( p), is as
follows:
X t =α 1 X t − 1+ α 2 X t − 2+ ⋯+ α p X t − p +W t

Where X t is stationary, α 1 , α 2 ,⋯ , α p are constants and W t is normal white noise.


If we use backshift operator B, we can rewrite AR( p) as follows:

α ( B ) X t =( 1 − α 1 B − α 2 B 2 −⋯ − α p B p ) X t =W t

α ( B ) =( 1 −α 1 B −α 2 B2 −⋯ −α p B p ) is defined as an autoregressive operator.

There are restrictions for the value of coefficients α to make sure X t can stay stationary.

For AR(1), −1< α 1 <1.


For AR(2), $
-2 < _1 < 2,
-1 < _2 < 1,
_1 + _2 < 1,
_2 - _1 < 1
$.
For a higher degree of AR, the requirement is as follows:
1. Set autoregressive operator = 0 as follows: $ (1 - {1} B - {2} B^{2} - - _{p} B^{p})
= 0 $. This equation is called a characteristic equation.
2. Solve the characteristic equation for B and get all the roots of B.
3. The absolute values of all the roots of B need to be ¿ 1 to ensure X t is stationary.
We will talk more about these requirements later in this course.
We can see from the AR definition above that random walk is very similar to AR(1) model
except that |α 1 )=1. Hence, random walk is not stationary. All rights reserved WQU
WorldQuant University QQQQ

2.2 Mean, Variance, ACF, and PACF for AR(1)


AR(1) model can be written as follows:
X t =C +α t X t − 1+ W t

Where C is a constant and W t is normal white noise with mean = 0 and variance = σ 2
Now let’s look at AR(1)’s mean, variance, autocovariance, ACF, and PACF.
The mean function is:
C
μ=
1 −α 1

When there is no intercept C in the process, μ is 0 .


The variance is:
2
σ
2
1− α
Autocovariance for X t and $ X_{s}$ is:

( )
2
σ |h)
γ ( h )= 2
α , where|h )=|s – t )
1− α

Autocorrelation for X t and $ X_{s}$ is:

ρ ( h ) =( α|h) ) , where|h )=|s – t )


Partial Autocorrelation for X t and $ X_{s}$ is:

ϕ h h=
{
α for|h ) ≤ 1
0 for|h ) >1 )
2.3 AR(1) and AR(2) Examples
Figure 1: AR(1) with α =0.4 and AR(1) with α =− 0.4
Figure 1 gives us two AR(1) graph examples. They both move around the mean = 0 and
within a horizontal band. The AR(1) with negative α is choppier than AR(1) with positive α .
We know the correlation of two observations next to each other is autocorrelation ρ ( 1 ). For
AR(1) model, ρ ( 1 ) is α . Hence, when α is positive, the two observations next to each other
move in the same direction while when α is negative, the two observations next to each
other move in the opposite direction. That is why we see the plot on the left is smoother
than the plot on the right.
Figure 2: ACF and PACF of AR(1)
Figure 2 shows the ACF and PACF plots for AR(1) with α =0.4 . From both plots, we see ACF
and PACF for lag 1 are all significant and their values are between 0.5 and 0.6. Since the
example is simulated data from AR(1) with α =0.4 , they are random sample data, and ACF
and PACF usually are not exactly the same as 0.4.
Now, let’s try to simulate AR(2) and see its ACF and PACF plots. The following figure 3
shows the graphs for AR(2).
Figure 3: ACF and PACF of AR(2)
Let’s focus on the ACF and PACF plots for AR(2) in figure 3. In the ACF plot, we can see the
ACF decreases gradually. On the other hand, the PACF in the plot drops to almost 0 after lag
2. This interesting PACF behavior for AR(2) process will help us decide what time series
model we will use. We will discuss more about model selection later in this module.

2.4 Use AR to represent MA


One interesting feature about AR process or AR model is that we can represent an MA
process using an infinite AR process. Let’s use MA(1) as an example.
We have the following MA(1) process:
X t =W t +θ W t −1

where W is the normal independent white noise with mean = 0 and variance = 1. If and only
if |θ )< 1, we can write X t in an infinite AR form as follows:
2 3
X t =θ X t −1 −θ X t −2 +θ X t − 3 − ⋯− W t

Here is how to derive the above result:


X t =W t +θ W t −1

X t =( 1+θ B ) W t by using the backshift operator

1
Wt= X by rearranging the terms
1+θ B t
We also know the formula of infinite geometric series:
2 3 ∞ a
S∞ =a+ a r +a r +a r +⋯+ a r =
1− r
when |r )< 1.
1
If |θ )< 1, we can rewrite X as follows:
1+ θ B t
Xt 2 3 ∞
= X t + X t ( −θ B )+ X t ( −θ B ) + X t (− θ B ) +⋯+ X t ( −θ B )
1+ θ B
2 2 3 3 ∞ ∞
¿ X t − θ B X t +θ B X t − θ B X t +⋯+θ B X t
1
We then can plug this back into W t = X and rearrange the terms and we will get:
1+θ B t
2 2 3 3
X t =θ B X t −θ B X t +θ B X t +⋯+W t

Then, we can place the backshift operator back as the subscript for X t and we get:
2 3
X t =θ X t −1 −θ X t −2 +θ X t − 3+ ⋯+W t

Now we know that X t can be written as an infinite AR model from an MA(1) model.
The key point for MA(1) to be able to be represented as an infinite AR process is to make
sure its coefficient |θ )< 1. We call this MA(1) invertible if the MA(1) process has the
absolute value of coefficients less than 1 and can be represented by an infinite AR process.
The other benefit when an MA process is invertible is that the MA process is also unique.
Let’s see an example below to demonstrate this concept.
MA A: X t =W t +5 W t − 1 with W ∼ normal white noise ( 0 , 1 )
1
MA B : X t =W t + W t −1 with W ∼ normal white noise ( 0 , 25 )
5
Both MA A and MA B time series have the same autocovariance values shown as follows:

{ )
26 for h=0
γ ( h )= 5 for h=1
0 for h>1

In the real world, we only observe the autocovariance values of the time series, not the
actual model. We cannot tell if MA A or MA B is the correct one. Both MA A and MA B are the
model candidates for the underlying time series data. So which model should we choose? If
we need the MA model to be invertible, then we will limit our model choice to MA B , which
has the absolute value of coefficient less than 1. Then, we call the MA B model unique.
2.5 ACF and PACF Plots for AR and MA Processes
We briefly talked about the special features ACF and PACF plots exhibit for the AR process
earlier. In this section, we are going to compare ACF and PACF plots for both the AR
process and MA process together. We are going to use an AR(2) process and an MA(2)
process as examples here.
Figure 4: ACF and PACF Plots for AR(2) and MA(2)
In figure 4 above, the top two plots are ACF and PACF plots for an AR(2) process. The
bottom two plots are ACF and PACF plots for an MA(2) process. In an ACF plot, the farthest
left bar shows autocorrelation is 1 when h=0, which means it’s the correlation for the
observation itself. In the PACF plot, it starts with h=1, which means the farthest left bar is
for PACF lag 1.
For the AR(2) process, PACF plot shows a drop to almost 0 after lag 2. For the MA(2)
process, ACF plot shows a drop to almost 0 after lag 2. This is a particularly important
phenomenon. We can actually use this phenomenon to decide the model specification for a
time series. If the PACF plot of a time series shows a sudden drop in value to close to 0 after
lag p, we can assume the time series follows an AR( p) process. If the ACF plot of a time
series shows a sudden drop in value to close to 0 after lag q , we can assume this time series
follows an MA(q ) process. With the above information, we can use ACF and PACF plots to
decide which process and what lag to use to model a time series. Figure 5 summarizes what
we just discussed about ACF and PACF plots.
Figure 5: ACF and PACF Plot Features for AR(𝑝 ) and MA(𝑞)

ACF Plot Feature PACF Plot Feature


AR($ p $) Tail off Cut off after $ p $ lags
MA($ q $) Cut off after $ q $ lags Tail off

2.6 Estimation of AR Model


We have introduced what an AR process model is. We discussed its property and how to
select the number of lags p for an AR( p) model. Now we need to estimate the coefficients of
the model. There are a few methods to estimate the model. We saw in section 1 that we can
use ordinary least square (OLS) method to estimate the model. In this section, we will go
over two other methods here: method of moments and maximum likelihood method.

2.6.1. Method of Moments (MOM)


Before talking about method of moments, let’s briefly review what moments are. Moments
are statistical parameters used to describe the distribution of a random variable.
The most popular ones are mean and variance. Mean is the first order moment. Variance
and covariance are the second order moments. Why do we refer to these moment
parameters with order? It is because the functions of these parameters include the
expected values of the order of random variable( E ( X n )). For example, the mean of X is
E ( X ). The variance of X can be written as E ( X 2) − E ( X )2. Hence, mean is the first order
moment and variance is the second order moment for a random variable. Here is the list of
popular moments we have used so far:
• First Order Moment: Mean
• Second Order Moments: Variance, covariance, correlation, autocovariance,
autocorrelation
• Third Order Moment: Skewness
• Fourth Order Moment: Kurtosis
Now we know what a moment is. Next, we are going to explain what method of moments
is. Method of moments is an estimation method that equates a sample moment to its
theoretical counterpart and allows us to estimate the parameters for the model. Let’s use a
very simple example in AR(1) model to explain how it works.
Assume we have a time series X t and we have removed the trend and the seasonality. The
time series is stationary. We have also run ACF and PACF and have decided AR(1) is a good
model specification. We assume the mean of X t is 0 so there is no constant in the AR(1)
model. We can write our AR(1) model as follows:
X t =α t X t −1 +W t

Where W t is normal white noise with mean = 0 and variance = σ 2

From the time series X t , we can calculate the following information:


• Sample mean: X
• Sample variance: γ^
(0)
• Sample autocovariance of lag 1: γ^
(1)
From last section, we know variance and autocovariance formulas are as follows:
The variance is:
2
σ
2
1− α
Autocovariance for X t and $ X_{s}$ is:

( )
2
σ |h)
γ ( h )= 2
α
1− α

where $h = s – t $
We can use the above sample information and formulas to solve for AR(1) parameter α .
γ^( 1)
α^ =
^
γ ( 0)

2.6.2 Yule-Walker Estimators for AR Model


We introduced MOM in the last section to estimate an AR(1) model. We can generalize the
method we used in the last section to solve for a higher order AR( p) model. The method we
used above is a very simple case of Yule-Walker estimation. The Yule-Walker estimation
method connects the coefficients of an AR( p) model to the autocovariances of the time
series by using a system of equations called Yule-Walker equations.
When we have an AR( p) process as follows:
X t =α 1 X t − 1+ α 2 X t − 2+ ⋯+ α p X t − p +W t

Yule-Walker equations can be defined as follows:


γ ( 1 ) ¿ α 1 γ ( 0 )+ α 2 γ (− 1 ) +α 3 γ ( −2 ) +⋯+α p γ ( 1 − p )
γ ( 2 ) ¿ α 1 γ ( 1 ) +α 2 γ ( 0 ) +α 3 γ ( −1 )+⋯+ α p γ (1 − p+1 )=α γ ( p − 1 )+ α γ ( p − 2 )+ α γ ( p −3 )+ ⋯+α γ ( 0 ) ¿ σ 2 ¿=γ ( 0 ) −
1 2 3 p
γ ( 3 ) ¿ α 1 γ ( 2 ) +α 2 γ ( 1 ) +α 3 γ ( 0 ) +⋯+ α p γ ( 1 − p+2 )
¿ ¿
2
With the above p+1 Yule-Walker equation system, we can solve for {α 1 , α 2 , α 3 , ⋯ , α p , σ }
p+1 parameters. We can see the last equation in the Yule-Walker equation system is an
equation to link the white noise variance to model coefficients and autocovariance.
All statistical applications can calculate the estimated coefficients using Yule-Walker
equations, so you do not need to calculate them manually.

2.6.3 Maximum Likelihood Method (Box-Jenkins Method)


The next estimation method we will introduce is the maximum likelihood method (MLE).
The maximum likelihood method is used to maximize the joint probability of the observed
data points with respect to the model parameters. The joint probability is called a
likelihood function. The likelihood function will usually take the log form so that the
function will be easier to handle. Please read the required reading to learn the general
definition and properties of the maximum likelihood estimation method.
The maximum likelihood estimation method is the selected estimation method in the Box-
Jenkins method. We will talk more about the Box-Jenkins method in the next lesson.
To apply the maximum likelihood method to an AR(1) model, we assume we have the
following AR(1) model:
X t =α t X t −1 +W t

Where W t is normal white noise with mean = 0 and variance = σ 2 and |α ) <1
With the above model, we can write our likelihood function as follows:
L ( α , σ 2) =f ( x1 , x2 , x 3 , ⋯ , x n ∨α , σ 2 )

Where x 1 , x 2 , x 3 ,⋯ , x n are data points from the time series.

We then maximize L ( α , σ 2) with respect to α and σ 2 to solve for α . All the statistical
applications can solve the optimization for us.

2.6.4 Comparison with Yule-Walker Estimators and Maximum Likelihood Estimators


Now we know how Yule-Walker estimation and maximum likelihood estimation work for
an AR model. However, what are the differences between the two methods?
First, Yule-Walker estimation only uses information from first moments, second moments,
etc., while the maximum likelihood method uses all data in the likelihood function.
Secondly, when the sample size is larger, the coefficients from the Yule-Walker estimation
will have larger variances than coefficients from the maximum likelihood method.
Third, we can only apply the Yule-Walker method to AR models. However, you can apply
MLE to most time series models.
However, MLE is computationally complicated while Yule-Walker estimation is relatively
straight forward. The choice of estimation method will depend on the computer power and
sample size of the data available.

3. Application of Autoregressive Models


In this section, we will try to apply an AR model to simulated time series data. We’ve
simulated an AR(2) time series for demonstration because we can do comparison of the
actual parameters and estimated parameters. The analytical process is the same for real-
world time series data as the simulated time series data. Figure 6 shows the simulated
AR(2) time series plot and its actual parameters.
Figure 6: Simulated AR(2) Time Series
We next draw ACF and PACF plots as figure 7.
Figure 7: ACF and PACF Plots for Simulated AR(2) Time Series
From figure 7, ACF decreases gradually while PACF drops suddenly. This indicates we
should use the AR(2) model, which fits the actual time series process.
Now, let’s run an AR(2) time series model. Figure 8 shows the AR(2) model result.
Figure 8: AR(2) Model Result
From figure 8, we can see both coefficient estimations are close to the actual values and
their p-values are all significant. Now, let’s check the residuals to see if the residuals follow
normal white noise.
Figure 9: Standardized Residuals from AR(2) Model
From figure 9, we can see standardized residuals oscillate around 0 and move within a flat
channel. It indicates the residuals are stationary.
Now let’s check the QQ plot for the normality of the residuals.
Figure 10: QQ Plot for AR(2) Residuals
From figure 10, we can see that residuals are normally distributed. Hence, we can confirm
that the residuals are normal white noise. Lastly, let’s draw a plot to show simulated time
series vs. fitted time series.
Figure 11: Simulated Time Series vs. Fitted Time Series
In figure 11, the black line is the simulated time series, and the orange line is the fitted time
series. From the plot, we can see that the fitted line captures the general trend of the
simulated time series but not the variation in the simulated time series. The differences are
covered by the residuals.

4. Conclusion
In this lesson, we went through the autoregressive model. We first give a brief description
of what an AR model is. We then provided the formal definition of an AR model, provided
some examples of AR processes, and gave the moments of an AR model. We then moved to
discuss how to use ACF and PACF plots to identify an MA model or an AR model. We also
explained why and when we can use an infinite AR process to represent an MA model.
Then, we talked about two estimation methods for an AR model: the Yule-Walker method
and maximum likelihood method. We ended the lesson with a simulated application to
show how to model an AR model and conduct post model diagnostics. In the next lesson,
we will introduce one comprehensive model and steps to run the model.

Copyright 2024 WorldQuant University. This content is licensed solely for personal use.
Redistribution or publication of this material is strictly prohibited.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy