ECONOMETRICS An Introduction PDF
ECONOMETRICS An Introduction PDF
ECONOMETRICS An Introduction PDF
Edward Omey
HUB - Stormstraat 2 - 1000 Brussel
e-mail: edward.omey@hubrussel.be
web: www.edwardomey.com
Academic year 2010 - 2011
1. What is econometrics?
2. The method of least squares
3. Selection of variables
4. The basic assumtions
5. Checking the basic assumptions
6. Making Predictions
7. Some references
Together with this text is an example that is called
EconometricsExampleText.xls.
There is also a le called
BBA3EconometricsAndExcel
and there is a table with critical values of the Kolmogorov-Smirnov test.
Both are available in "Teaching material BBA3" on www.edwardomey.com.
1
1 What is econometrics?
1.1 Introduction
Econometrics can be seen as scientic research in which we complete the
results of economic assumptions and theories with quantitative information
that is based on real data.
In economic theory, usually there is a qualitative analysis based on some
reasonable assumptions. As an example, in economic theory one argues that
an increase in the price of a product will result in a decrease of the demand
of that product. In econometrics we want to quantify this theoretical result
to know what will be the resulting demand if we increase the price by, for
example, 1 euro.
In mathematical economics one tries to present economic theory in a
mathematical, formal way. In econometrics we want to verify these mathe-
matical models by using real data. As a result we can eventually decide that
the theory is a good or a bad representation of the real world.
Economic statisticiens usually collect data and present data in various
ways. Usually these data form the basis of further statistical analysis.
1.2 Econometric approach
In general, the econometric approach can be represented in the following
table. The table consists of 5 levels and 3 main types of input.
I (economics) II (data) III (methodology)
1 economic theory facts mathematics & probability
2 economic model data statistics
3 econometric model worked data econometric methods
4 operational econometric model
5 verication prediction evaluation
We briey discuss the entries of this table.
1.3 Input from economics
In economics, people develop "rules" about how people, companies and states
behave. Usually this analysis is qualitative and presents the inuence of -
2
resp. causal relations between - one group of variables on other groups of
variables.
Sometimes one uses economic models which give a "simple" representa-
tion of the real life.
Example. Keynes examined the relations between consumption and sav-
ings and he found that "people tend to spend more when income goes up".
A simple model is the following model: C = c + /1 , 0 < / < 1, where
C = consumption and 1 = income.
This formula is an example of a mathematical model. It consists of 2
variables and 1 relation. On a graph the model corresponds to a straight line
with slope /. The graph intersects the vertical axis in c.
In econometric models it is important to answer the following questions:
- how many variables are we going to include in a model?
- which variables are we going to include in a model?
- how many relations and which relations are we going to include in the
model?
Example
In studying the income Y of individual workers, economic theory leads to
a formal relation of the form
1 = ,(age, sexe, level of education, years of experience, knowledge
of languages,...)
To formulate an econometric model it is necessary to know how many
variables and which variables to include nd to specify the relation 1 =
,(....).
1.4 Input from data
In all econometric analysis, we need data. In some cases we have to collect
raw data and in other cases we can use data that have been collected by
others - statistical institutions.
1.4.1 There are dierent types of data.
Usually one consider the following 4 types of data
a) Qualitative - the data can not be represented by meaningful numbers
a1) Nominal data: we use groups or categories depending on a qualitative
property
Examples: the colour of hair; sexe of people; the type of car,...
3
a2) Ordinal data: these are nominal data which can be sorted in a mean-
ingful way
Examples: the stars of hotels; the size of shoes; the highest diploma;...
b) Quantitative - the data can be represented by meaningful numbers
b1) Interval data
b2) Ratio data
Ratio data are based on real variables that have a natural zero. The
number zero means that the characteristic is absent.
Example: income (my income is zero); points (I obtained zero points);
revenue (I sold zero products);...
When we deal with ratio data, then the words "double" or "halve" make
sense.
In dealing with interval data we have a variable without a natural zero
and it doesnt make sense to use the words "double" or "halve".
Typical examples are:degrees in
C or in
1; all IQ-scales.
1.4.2 Dummy
In many econometric models we have to deal with qualitative variables. In
order to include such variables in a mathematical model it will be useful to
quantify these qualitative variables. To this end we are going to use 1 or
more dummy variables. A dummy variable is a variable that can take on
only 2 values: 1 or 0.
Example. When we have a qualitative variable with 2 categories, we use
1 dummy variable. Suppose that we have to deal with the sexe of people, we
can dene 1 = 1 if the person is female, and 1 = 0 if the person is
male.
Example. When we have a qualitative variable with 3 categories, we use
2 dummies. Suppose that we want to model the colours at a trac light. We
can dene
1(1) = 1 if light is red 1(1) = 0 otherwize
1(2) = 1 if light is green 1(2) = 0 otherwize
In this case we nd the following possibilities
1(1) 1(2) result
1 0 :cd
0 1 q:cc:
0 0 o:c:qc
1 1 i:jo::i/|c
4
1.4.3 Quality of data
Many companies and government organisations collect data and construct
(huge) databases. Some of these are freely available on internet.
Using these data one has to be careful about the quality of the data. In
many cases the published data suggest a higher precision than in reality. One
often publishes rounded numbers or (weekly, monthly,...) averages. Some
always round down and others round in another way.
In collecting data one often has problems with validity: do the data
measure what we really want to measure? For one variable it is often the
case that people use several proxies. To measure - for example - poverty in
a country, one can use several variables:
- Gini - coecient
- the proportion of people living below the poverty line
- the ratio Q(3),Q(1)
etc.
1.4.4 Reworking data
Sometimes we have to modify the data before using them in a model. Ex-
amples of modications are
- replace nominal values by real values
- to calculate weekly mean values from daily values
- calculate relative changes and/or proportions
- transform variables to increase correlation coecients
etc.
1.4.5 Time-series versus cross-sections
In general we can distinguish two types of analysis in econometrics
In a time series we measure variables on dierent moments in time (daily,
weekly, yearly,...).
In a cross section we measure variables in the same moment.
Example
- We study the monthly unemployment rate in Belgium and use (for ex-
ample) the monthly ination, the monthly interest rate, the monthly relative
change in the working population,...
5
- We study the 2010 unemployment rate in the dierent countries of
the EU and we use (for example) the mean income, the interest rate, the
population density, the ratio new companies/failed companies, etc.
1.5 Input from methodology
In our analysis we are going to need some techniques from mathematics,
probability theory and statistics.
1.5.1 Stochastic models
In econometrics we often deal with relations of the form
1 = ,(A. 2. l. \. ...)
where 1 is the dependent variable and A. 2. l. \ ,... are independent or
explanatory variables. Because it is very plausable that we "forgot" 1 or
more variables, usually one adds an error term in the formula:
1 = ,(A. 2. l. \. ).
This error term summarizes all errors that can occur.
- measurement errors
- observation errors
- rounding of data
- variables that we "forgot" in our analysis (because we didnt take them
into account or because we didnt want to take them into account)
- irrelevant variables that we included in our analysis
- stochastic behaviour: in all social, biological, ... environments there is
some indeterminism present. Under the same conditions, the sales can be
dierent from day to day because of random decisions of people
- etc.
1.5.2 Linear models
In this course we consider only linear models. These are mathematical
relations that are linear in the parameters.
Examples
* 1 = c +/A
6
* 1 = c +/A +cA
2
* 1 = c +/A +c2 +d exp(A)
* 1 = c +/ log(A) +c log(2)
etc.
1.5.3 "Good" models
We will us the KISS principle in econometrics! We consider a model a "good"
model if
* it is simple: a model is stronger if it explains a lot by only a few variables
* it is theoretical consistent: if from the theoretical point of view we
expect that a variable has a positive inuence on 1 , we expect that the
model reects also this positive eect!
* it has predictive power: we wish that the predictions made by the model
are "close" to reality!
* explanatory power: we wish that the results of the model are close to
the data that we used to obtain the model!
1.5.4 Econometric methods
Model: In general we are going to use models of the form
1 = c +/A +c2 +dl +c\ +... +
= ,(A. 2. l. \. .... c. /. c. .... ).
Here
* 1 is the variable that we want to explain;
* A. 2. l. \. ... are the explanatory variables, the variables that we use to
explain 1 ;
* c. /. c. d. ... are the parameters in the model;
* is the (stochastic) error term.
Data. To estimate the parameters, we need data about 1. A. 2. l. \. ...We
suppose that we have : datapoints:
* 1
1
. A
1
. 2
1
. l
1
. \
2
* 1
2
. A
2
. 2
2
. l
2
. \
2
...
* 1
n
. A
n
. 2
n
. l
n
. \
n
.
7
By denition, we assume that each of these data points satises the "for-
mula":
1
i
= ,(A
i
. 2
i
. l
i
. \
i
. .... c. /. c. ....
i
), for i = 1. 2. .... :.
We assume that the parameters are independent of the index i but that the
error term may be dierent for dierent values of the index.
Method
In the formula
1
i
= ,(A
i
. 2
i
. l
i
. \
i
. .... c. /. c. ....
i
), for i = 1. 2. .... :,
we know the values of 1
i
. A
i
. 2
i
. ... but not the values of c. /. c. ... and
i
.
Hoping that
i
is "small", we delete
i
and we approximate 1
i
by
1
i
:
1
i
= ,(A
i
. 2
i
. l
i
. \
i
. .... c. /. c. ...), for i = 1. 2. .... :.
The approximation error that we make will be denoted by c
i
:
c
i
= 1
i
1
i
, for i = 1. 2. .... :.
We want the errors to be "as small as possible". In order to do this, we can
dene the "global" error in dierent ways. We can take for example
mean error = c =
1
:
n
i=1
c
i
or
1 = absolute deviation =
n
i=1
[c
i
[ ,
or
oo1 = sum or squared errors =
n
i=1
c
2
i
.
8
1.5.5 Least squares method
A popular and attractive method in econometrics is the least squares
method (LSE). In this method we estimate the papameters by solving the
following mathematical problem:
min
a;b;c;:::
oo1.
We consider a model to be the best model, if the sum of the squared
errors is smaller than any other model.
Example. Let us consider the following data:
X Y
3 10
6 15
9 30
12 35
15 25
18 30
21 45
24 45
We consider the following 2 models:
Model 1:
1 = 8 + 1. 6 + A
Model 2:
1 = 8. 5 + 1. 5 + A
Using these formulas, we nd the following results:
X Y Model 1 errors model 1 Model 2 errors model 2
3 10 12,8 -2,8 13 -3
6 15 17,6 -2,6 17,5 -2,5
9 30 22,4 7,6 22 8
12 35 27,2 7,8 26,5 8,5
15 25 32 -7 31 -6
18 30 36,8 -6,8 35,5 -5,5
21 45 41,6 3,4 40 5
24 45 46,4 -1,4 44,5 0,5
Simple calculations show the following results:
9
Model 1 Model 2
c -0,225 0,625
1 39,4 39
oo1 241,96 243
If we use AD, then model 2 performs better than model 1. If we use SSE,
then model 1 performs better than model 2.
10
2 The method of least squares
2.1 Introduction
Recall from the previous chapter that we are going to use the method of least
squares. To this end we need a model and data!
Model
In general we are going to use models of the form
1 = c +/A +c2 +dl +c\ +... +.
Data
To estimate the parameters, we need data about 1. A. 2. l. \. ...We sup-
pose that we have : datapoints:
* 1
1
. A
1
. 2
1
. l
1
. \
2
* 1
2
. A
2
. 2
2
. l
2
. \
2
...
* 1
n
. A
n
. 2
n
. l
n
. \
n
.
By denition, we have
1
i
= c +/A
i
+c2
i
+dl
i
+... +
i
, for i = 1. 2. .... :.
We assume that the parameters are independent of the index i but that the
error term may be dierent for dierent values of the index.
Method
We delete
i
and nd the following approximation:
1
i
= c +/A
i
+c2
i
+dl
i
+... + , for i = 1. 2. .... :.
The approximation error that we make is c
i
:
c
i
= 1
i
1
i
, for i = 1. 2. .... :.
The "global" error is given by oo1:
oo1 = sum or squared errors =
n
i=1
c
2
i
.
11
The least squares method (LSE): we estimate the papameters by solv-
ing the following mathematical problem:
min
a;b;c;:::
oo1.
We consider a model to be the best model, if the sum of the squared
errors is smaller than any other model.
2.2 Example 1: the simple model 1 = c +/A +
2.2.1 The least squares method
In the simplest model we have only 1 explanatory variable (A) and 2 para-
meters (c. /). Now we have
Model: 1 = c +/A +;
Data: 1
1
. A
1
. 1
2
. A
2
. ...... 1
n
. A
n
and 1
i
= c +/A
i
+
i
;
Approximation:
1
i
= c +/A
i
;
Approximation error: c
i
= 1
i
1
i
= 1
i
c /A
i
;
Global error: oo1 =
c
2
i
=
(1
i
c /A
i
)
2
;
LSE: we have to solve the problem
min
a;b
oo1 = min
a;b
(1
i
c /A
i
)
2
.
Solution (details are here for completeness)
To solve the problem, we rewrite oo1 as follows: we insert A and 1 in
the expression for oo1 and we nd
oo1 =
(
_
1
i
1 ) /(A
i
A) + (1 c /A)
_
2
=
(1
i
1 )
2
+/
2
(A
i
A)
2
+
(1 c /A)
2
2/
(1
i
1 )(A
i
A)
+2
(1
i
1 )(1 c /A)
2/
(A
i
A)(1 c /A).
For the last two terms, we nd
(1
i
1 )(1 c /A) = (1 c /A)
(1
i
1 ) = 0,
(A
i
A)(1 c /A) = (1 c /A)
(A
i
A) = 0.
12
Simplifying, we nd that
oo1 =
(1
i
1 )
2
+/
2
(A
i
A)
2
+:(1 c /A)
2
2/
(1
i
1 )(A
i
A)
For the other terms, we use the following notations:
\ (1 ) =
(1
i
1 )
2
and :
2
(1 ) =
\ (1 )
: 1
\ (A) =
(A
i
A)
2
and :
2
(A) =
\ (A)
: 1
\ (A. 1 ) =
(1
i
1 )(A
i
A) and Co(A. 1 ) =
\ (A. 1 )
: 1
.
We call \ (1 ) the variation of 1 and \ (A. 1 ) the covariation of A and
1 . It is clear that this is highly related to covariance and correlation:
Co(A. 1 ) =
1
: 1
(1
i
1 )(A
i
A),
:(A. 1 ) =
Co(A. 1 )
:(A):(1 )
=
\ (A. 1 )
_
\ (A)\ (1 )
.
Using these new notations, we have calculated that
oo1 = \ (1 ) +/
2
\ (A) 2/\ (A. 1 ) +:(1 c /A)
2
.
Note that the last term in this expression is always zero or positive for
all values of c. /.
To omit any contribution of this term, we impose the condition that this
term is zero:
1 c /A = 0
or equivalently
1 = c +/A.
We can simplify oo1 and nd
oo1 = \ (1 ) +/
2
\ (A) 2/\ (A. 1 ).
13
This is a quadratic relation in / and from mathematics we know that
oo1 is a convex parabola.
(See e.g. at the website: http://en.wikipedia.org/wiki/Parabola)
This parabola reaches a minimum in the value
/ =
\ (A. 1 )
\ (A)
,
and the value at this mimimum is given by
oo1 = \ (1 )
\
2
(A. 1 )
\ (A)
.
As a conclusion, we have solved the mathematical problem. The optimal
values of c and / will be denoted by c and
/. We nd
/ =
\ (A. 1 )
\ (A)
=
Co(A. 1 )
:
2
(A)
,
c = 1
/A,
1 = c +
/A.
The expressions for c and
/ are called the least squares estimates for c
and /.
The nal result
1 = c +
/A. Since 1 = c +
/A, it
follows that (A. 1 ) is a point on the regression line.
For the errors we nd
c
i
= 1
i
1
i
= 1
i
c
/A
i
.
It follows that the mean error is zero:
c = 1 c
/A = 1 c
/A = 0.
14
For the variation of the errors \ (c), we nd
\ (c) =
(c
i
c)
2
=
c
2
i
= oo1.
Recall that
\ (c) = oo1 = \ (1 )
\
2
(A. 1 )
\ (A)
.
For the variation of
1 , we nd that
\ (
1 ) = \ (c +
/A)
=
/
2
\ (A)
=
\
2
(A. 1 )
\ (A)
.
Combining these two formulas, we nd that
\ (c) = \ (1 ) \ (
1 ),
or
\ (1 ) = \ (
1 ) +\ (c).
For this simple model the variation of 1 can divided into two parts:
the total variation is the sum of the variation explained by the model
(\ (
1 ) = the variation of
1
oo1 = \ (c) = the variation of c.
The previous remark shows that oo1 = oo1 +oo1.
2.2.3 Numerical example
We take the example of the previous section. We have the following data:
15
X Y
3 10
6 15
9 30
12 35
15 25
18 30
21 45
24 45
and we consider the following model: 1 = c +/A +.
In this example we nd:
A = 13. 5 \ (A) = 378 :
2
(A) = 378,8
1 = 29. 375 \ (1 ) t 1121. 9 :
2
(1 ) t 1121. 9,8
\ (A. 1 ) = 577. 5 Co(. 1 ) = 577; 5,8
The least squares estimators are given by
/ =
Co(A. 1 )
:
2
(A)
= 1. 5278
c = 1
/A = 8. 75
and the regression line of 1 on A is given by
1 = 8. 75 + 1. 5278A.
For the model we nd the following errors and estimates (we rounded the
numbers)
X Y
3 10
6 15
9 30
12 35
15 25
18 30
21 45
24 45
1 c c
2
13,3334 -3,3334 11,11
17,9168 -2,9168 8,51
22,5002 7,4998 56,25
27,0836 7,9164 62,67
31,667 -6,667 44,45
36,2504 -6,2504 39,07
40,8338 4,1662 17,36
45,4172 -0,4172 0,17
One can verify that c = 0 and that \ (
1 ) = 1. In
econometrics it is common to use :
2
(1.
1 ). We call :
2
(1.
1 ) the 1
2
value of
the model. Alternatively 1
2
is also called the determination coecient:
to what extent is 1 determined by
1 . We always have 0 _ 1
2
_ 1 and the
ideal model has 1
2
= 1 or 1
2
= 100%.
In our example we nd 1
2
= 78. 6%.
Scatter plot As a second indicator, we can make a scatter plot of 1
1 .
In the ideal situation, all points are on the rst diagonal.
ANOVA As a third indicator we can analyse the variations or variances
in the model. Recall that for our model we have
\ (1 ) = \ (
1 ) +\ (c).
The third indicator is the ratio of what we can explain divived by what
we have to explain:
1
2
=
\ (
1 )
\ (1 )
.
It is not a mistake to use the same notation 1
2
: one can provethe following
property
Property
For linear models with a constant term, we have
1
2
=
\ (
1 )
\ (1 )
= :
2
(1.
1 ).
17
2.3 Example 2: The model 1 = c +/A +c2 +
2.3.1 The least squares method
In this model we have only 2 explanatory variables (A and 2) and 3 para-
meters (c. /. c). Now we have
Model: 1 = c +/A +c2 +;
Data: 1
1
. A
1
. 2
1
. 1
2
. A
2
. 2
2
. ..... 1
n
. A
n
. 2
n
and 1
i
= c +/A
i
+c2
i
+
i
;
Approximation:
1
i
= c +/A
i
+c2
i
;
Approximation error: c
i
= 1
i
1
i
= 1
i
c /A
i
c2
i
;
Global error: oo1 =
c
2
i
=
(1
i
c /A
i
c2
i
)
2
;
LSE: we have to solve the problem
min
a;b;c
oo1 = min
a;b;c
(1
i
c /A
i
c2
i
)
2
.
Solution
From Example 1 recall that the optimal values of c and / were given by
Co(A. 1 ) = /:
2
(A),
1 = c +/A
In the new 3parameter model one can show that the optimal values of
c. /. c are the solutions (if they exist) of the following system of equations:
Co(A. 1 ) = /:
2
(A) +cCo(A. 2),
Co(2. 1 ) = /Co(2. A) +c:
2
(2),
1 = c +/A +c2.
We are going to use EXCEL to solve this mathematical problem.
2.3.2 Multicolinearity and quasi-multicolinearity
Examples From the mathematical point of view, the system of equations
doesnt always has a unique solution!
Example 1
Let us consider the system
5 = /6 +c10
7 = /12 + c20
18
It is clear that the second equation can be simplied and we nd
5 = /6 +c10
3. 5 = /6 +c10
and there is no solution!
Note that if we look at the coecients of / and c in the system, we have
6 10
12 20
and we see that the second row is a multiple of the rst row!
Example 2
Let us consider the system
5 = /6 +c10
10 = /12 + c20
In this case, we can simplify again and we nd
5 = /6 +c10
5 = /6 +c10
or just only one equation: /6 + 10c = 5. It is clear that we nd many
solutions!
Note that if we look at the coecients of / and c in the system, we have
6 10
12 20
and we see that the second row is a multiple of the rst row!
Example 3
Let us consider the system
5 = /6 +c10
10 = / +c
In this case the second equation gives / = 10c and as a result, for equation
1 we nd that
5 = (10 c)6 +c10
19
or 55 = 4c, or c = 55,4 and then / = 10 c.
Note that if we look at the coecients of / and c in the system, we have
6 10
1 1
and we see that the second row is a multiple of the rst row!
Denition of MC and QMC Looking at the system of equations that
we have to solve, we have the following coecients of / and c:
:
2
(A) Co(A. 2)
Co(A. 2) :
2
(2)
We have a problemof multicolinearity (MC) if the second rowis a multiple
of the rst row! This is
n(:
2
(A). Co(A. 2)) = (Co(A. 2). :
2
(2)),
where n ,= 0 is a constant. In this case we nd that
n + :
2
(A) = Co(A. 2)
n + Co(A. 2) = :
2
(2)
From these equations, we can nd n. We have
n =
Co(A. 2)
:
2
(A)
n =
:
2
(2)
Co(A. 2)
and we nd that
Co(A. 2)
:
2
(A)
=
:
2
(2)
Co(A. 2)
or
Co
2
(A. 2)
:
2
(A):
2
(2)
= 1
or
:
2
(A. 2) = 1.
Recall that when :
2
(A. 2) = 1, then :(A. 2) = 1 and then A and 2
are perfectly linearly related: there is a linear relationship between A and
2.
20
Denition
In the model 1 = c +/A +c2 +, we have MC (multicolinearity) if
A and 2 are linearly related, i.e. if :
2
(A. 2) = 1.
When the variables A and 2 are not linearly related, but almost linearly
related, we say that we have a problem of quasi-multicolinariry (QMC).
This happens when :
2
(A. 2) is "close" to 1.
In econometrics, we want to avoid QMC!
We make the following agreement:
If :
2
(A. 2) _ 0. 36 or if 0. 60 < :(A. 2) < 0. 60, we dont have
problems of QMC;
If :
2
(A. 2) 0. 36 or if :(A. 2) 0. 60 or :(A. 2) < 0. 60, we have
a problem of QMC.
For models with 3 or more explanatory variables, we are also going to
discuss QMC. We have a problem of QMC if one of the variables is highly
linearly related with one or more other explanatory variables.
Example We consider the model 1 = c +/A +c2 +, where
1 = the half year salesvolume
A = the time
2 = a dummy (2 = 1 in the rst half year and 2 = 0 in the second
half year)
Y X Z
4 1 1
1 2 0
6 3 1
2 4 0
11 5 1
5 6 0
11 7 1
7 8 0
15 9 1
9 10 0
We nd (cf.Computer seminar): :
2
(A) = 8. 25, :
2
(2) = 0. 25 and :(A. 2) =
0. 17. For this model we dont have problems with QMC and we can solve
21
the system of equations. We nd
c = 2. 4;
/ = 1. 2 and c = 5. 8;
1 = 2. 4 + 1. 2A + 5. 81.
We call
1 the regression of 1 on A and 1. In the next table we calculate
1 c
4. 6 0. 6
0 1
7 1
2. 4 0. 4
9. 4 1. 6
4. 8 0. 2
11. 8 0. 8
7. 2 0. 2
14. 2 0. 8
9. 6 0. 6
The 1
2
of the model is given by 1
2
= :
2
(1.
1 = c +
/ =
\ (A. 1 )
\ (A)
,
c = 1
/A.
The parameters of the model are c. / and o
2
. The estimates for c and /
are c and
/ s `(/. o
2
b
b
),
where
o
2
ba
= \ c:(c) = o
2
A
2
::
2
(A)
,
o
2
b
b
= \ c:(
/) = o
2
1
::
2
(A)
,
Co(c.
/) = o
2
A
::
2
(A)
.
33
I would like to stress that we are making all calculations in EXCEL. These
formulas are important to obtain some information about the quality of the
estimators.
We see that the variance of
/ depends on 3 elements:
dependence on o
2
: if o
2
= \ c:() is larger, then \ c:(
/) is larger. This
means that if we allow larger uctuations in the error term, there will
be larger uctuations in the parameter estimates.
dependence on :: if : increases, then \ c:(
/) decreases. This
means that if we are far away from MC and QMC, then the estimates
are better.
The previous result can be used to obtain condence intervals for c and
for /. We nd the following 95% c.i.:
c = c .
2;5%
o
ba
,
/ =
/ .
2;5%
o
b
b
.
Although these formulas are correct, the are without use! In order to use
the formulas, we need o
2
ba
and to nd this,we need o
2
!
Theorem 2 Suppose that all basic assumptions hold. Then we have that o
2
can be estimated by :
2
e
, where
:
2
e
=
oo1
: j
.
Moreover, o
2
ba
and o
2
b
b
can be estimated by :
2
ba
and :
2
b
b
, where
:
2
ba
= :
2
e
A
2
::
2
(A)
, :
2
b
b
= :
2
e
1
::
2
(A)
.
As a compensation, in the condence statements we have to replace the
normal distribution by a tdistribution. We nd the following 95% c.i.:
c = c t
np;2;5%
:
ba
,
/ =
/ t
np;2;5%
:
b
b
.
In practise, EXCEL calculates all things that we need.
34
4.4 Example
We have a look at Model 3B of the previous section. The model was the
following:
model : 1 = c +/A(3) + cA(8) + dA(7) + .
The full excel-output is given by the following 3 parts:
SUMMARY OUTPUT
Regression Statistics
Multiple R 0,879
R square 0,774
Adjusted R square 0,767
Standard Error 4356,7
Observations 106
ANOVA
df SS MS F Signicance F
Regression 3 6,6E09 2,21E09 116,32 8,55E-33
Residual 102 1,9E09 1,9E07
Total 105 8,6E09
Coe. Stand.Err t-stat P-value Lower 95% Upper 95%
Intercept -4611,4 2558 -1,78 0,077 -9745 522
X(3) 22,68 1,723 13,11 1,2E-23 19,24 26,11
X(8) 4401,99 1317 3,34 0,001 1789 7014
X(6) -1982,2 937,7 -2,11 0,037 -3842 -122
Part 1
In part 1 we get the regression statistics. Important for us is the 1
2
value.
In our case it is 1
2
= 77. 4%.
The standard error of the model is given by :
e
= 4356. 7. It means that
with the model we make errors that have uctuations around 0 and the size
of the uctuations is around :
e
= 4356. 7.
It is wize to compare the size of the errors with the size of 1 . The mean
price of the cars in our example is given by 1 = 32473. 6. The relative error
in our model is given by
:
e
1
=
4356. 7
32473. 6
t 0. 13.
35
With the model we make relative errors around 13%. As a rule of the thumb,
we are glad if the relative error is less than 10%.
Part 2
Part 2 is devoted to the analysis of the variance
In our notations, we have
(regression) SSR = 6623984149
(residuals) SSE =1936113261
(total) SST = 1936113261
The abbreviation "df" means degrees of freedom. When we calculate the
mean square, we nd for example that
oo1
: j
=
1936113261
106 4
= 18981502. 56
This is :
2
e
= 18981502. 56. Calculating the square root gives :
e
= 4356. 7
which we got in part 1.
The 1 c|nc is the 1statistic that we use to see if 1
2
is suciently
large. The prob-value of the 1value is given by "F-signicance".
Part 3
Part 3 gives the detailed statistical analysis of the parameter estimates.
We consider /, the coecient of A(3).
From the excel output, we see that
/ = 22. 68
:
b
b
= 1. 723
Using a 95% c.i. we obtain that
/ =
/ t
np;2;5%
:
b
b
= 22. 68 1. 98 + 1. 723
= [19. 24 26. 11]
which is given in the last columns of the table in Part 3.
The results of Part 3 also allow to test the following type of hypotheses.
Let us consider
H
0
: / = 0 versus H
a
: / ,= 0
36
and let us use c = 5%.
The rst method to choose between H
0
and H
a
is based on condence
intervals. Since the 95% c.i. is given by [19. 24; 26. 11],we conclude that we
reject H
0
.
A second method is based on calculating prob-values. The tvalue of the
sample result is given by
t value =
/ /
0
:
b
b
=
/
:
b
b
=
22. 68
1. 723
t 13. 11.
Using the t
102
distribution, we nd that
1([t
n2
[ 13. 11) = 1. 261 23
This small value allows us to reject H
0
.
Remark. Excel always calculates a two-sided prob-value.
37
5 Checking the basic assumptions
5.1 Introduction
In view of the previous section, the basic assumptions are crucial for all
statistical properties of the estimators and the models in econometrics. In
the section we discuss how the check whether or not the basic assumptions
hold. If the basic assumptions do not hold, we have a problem with the
statistical properties and this may contaminate any conclusions we make.
5.2 Basic assumption 1
If basic assumption 1 doesnt hold, then our estimates are biased and most
of the regression output is not reliable any more.
We check this basic assumption by making a scatter plot of 1 and the
calculated residuals or errors c.
We hope to see a graph without outliers and without clusters.
If we see outliers, we have to check the data again: did we make an
input-error? is the data correct? do we have a special data-point?
It is possible that we decide to delete the data line that generates the
outlier.
If we see clusters, we forgot a variable in the model. Careful checking the
data in the cluster(s) may lead to the introduction of a new variable or a
new dummy.
In our example we nd the following graph:
5.3 Basic assumption 2
If basic assumption 2 doesnt hold, then all estimates related to variances
(:
2
e
. :
2
ba
. 1
2
,...) are not longer valid and mostof the statistical analysis (con-
dence intervals, 1-value) is not reliable any more.
We check this basic assumption in several ways
5.3.1 Scatter plots
We maka a scatter plot of c
2
and each of the variables 1. A. 2. .... In a time-
seriesanalysis, we also make a scatter plot (i. c
2
i
). In each plot we hope to see
a horizontal box without any systematic pattern or trend.
38
BA 1
-12000
-8000
-4000
0
4000
8000
12000
10000 15000 20000 25000 30000 35000 40000 45000 50000 55000
Y
e
39
BA 2
0
20000000
40000000
60000000
80000000
100000000
120000000
140000000
10000 20000 30000 40000 50000 60000
Y
e
[1
i
1
i
[ ,
`11 =
1
1
1
i
1
i
1
i
.
In the example, we nd the following numbers:
We have
1 = 29487. 9
1 = 31213. 05
`1 = 3606. 5
`11 = 0. 134.
Our data uctuate around the central value 1 = 29487. 9 and we have
`1 = 3606. 5. Our predictions are predictions with an absolute error of
around 3606. Related to 1 this is an error of around 12%.
Looking at the relative deviations, we nd a relative error of around
13. 4%.
6.4 Mean squared errors
In the place of taking absolute deviations, one can also look at the squared
errors:
`o1 =
1
1
(1
i
1
i
)
2
1`o1 =
_
`o1
In our example, we nd
`o1 = 18807115
1`o1 = 4336. 72
We can also use squared relative errors and we nd:
`o11 =
1
1
_
1
i
1
i
1
i
_
2
= 0. 0278
1`o11 =
_
`o11 = 0. 167
50
These measure are used when we want to compare the quality of predic-
tions of several models.
6.5 More
More measures can be found on the following website:
http://en.wikipedia.org/wiki/Forecasting#Forecasting_accuracy
This website refers also to some interesting examples.
51
7 Some references
7.1 Internet
1. http://en.wikipedia.org/wiki/Econometrics
2. Econometric Theory on Wikibooks:
http://en.wikibooks.org/wiki/Econometric_Theory
3. B.E. Hansen, Econometrics:
http://www.ssc.wisc.edu/~bhansen/econometrics/
4. Econometrics Resources on internet:
http://www.oswego.edu/~kane/econometrics/
5. Empirics and Econometrics:
http://homepage.newschool.edu/het//schools/metric.htm
6. Links to Online Texts and Notes in Econometrics:
http://www.economicsnetwork.ac.uk/teaching/text/econometrics.htm/
7. Startpagina econometrie:
http://econometrie.startpagina.nl/
7.2 Books
1. A.P.Barten,Econometrische lessen Schoonhoven: Academic Service, Economie
en Bedrifjskunde, 1989
2. W.S. Brown, Introducing econometrics. West Publishing Company,
1991.
3. P. Kennedy, A quide to econometrics, 3rd edition, MIT Press, Cam-
bridge, Mass. 1992.
4. D.N.Gujarati, Essentials of Econometrics. Mc Graw Hill International
Edition, New York, 2006
5. D.N.Gujarati, Basic Econometrics, 4th editions, Mc Graw Hill Inter-
national Edition, New York, 2003.
52
6. G.S. MADDALA, Econometrics, McGraw-Hill Ltd., New York, 1977.
7. E. Omey, Inleiding tot de econometrie. Den Arend, Bonheiden, 2003.
8. J. Schmidt, Econometrics, Mc Graw Hill, New York, 2005.
9. M.F. Triola and L.A. Franklin, Business Statistics, Addison-Wesley,
New York, 1994.
10. R.J. Wonnacott and T.H. Wonnacott, Econometrics, 2nd edition, Wi-
ley, New York, 1979.
53