Econometrics Notes PDF
Econometrics Notes PDF
Econometrics Notes PDF
1
where “Ui” is known as the disturbance, or error, is a random (stochastic) variable that has
well-defined properties. The disturbance term “U” may well represent all those factors that
affect consumption but are not taken into account explicitly. The econometric consumption
function hypothesizes that the dependent variable Y (consumption) is linearly related to the
explanatory variable X (income) but the relationship between the two is not exact; it is
subject to individual variation.
4. Obtaining Data:
We need data for the variables above. This can be obtained from government statistics
agencies and other sources. A lot of data can also be collected on the Internet in these days.
But we need to learn the art of finding appropriate data from the ever increasing huge loads
of data. To estimate the econometric model that is, to obtain the numerical values of β1 and
β2, we need data. Data are classified into time series data, cross section data and pooled data.
5. Estimation of the Econometric Model
Once the data are collected, next task is to estimate the parameters of the consumption
function. The numerical estimates of the parameters give empirical content to the
consumption function. For now, note that the statistical technique of regression analysis is
the main tool used to obtain the estimates. Here, we quantify and , i.e. we obtain
numerical estimates. This is done by statistical technique called regression analysis.
Estimation of the econometric model is usually done using Ordinary Least Square methods.
6. Hypothesis Testing:
According to “positive” economists like Milton Friedman, a theory or hypothesis that is not
verifiable by appeal to empirical evidence may not be admissible as part of scientific enquiry.
We use hypothesis testing techniques for verifying whether the estimated regression equation
and the parameter estimates are reliable. .
7. Forecasting or Prediction
If the chosen model does not refute the hypothesis or theory under consideration, we may use
it to predict the future value(s) of the dependent, or forecast, variable Y on the basis of
known or expected future value(s) of the explanatory, or predictor, variable X.
8. Use of the Model for Control or Policy Purposes:
As these calculations suggest, an estimated model may be used for control, or policy,
purposes. By appropriate fiscal and monetary policy mix, the government can manipulate the
control variable X to produce the desired level of the target variable Y.
2
error term (with well-defined probabilistic properties) is required in regression analysis for the
following important reasons.
1. Vagueness of theory: The theory, if any, determining the behaviour of Y may be, and often is,
incomplete. We might know for certain that weekly income X influence weekly consumption
expenditure Y, but we might be ignorant or unsure about the other variables affecting Y.
Therefore, ui may be used as a substitute for all the excluded or omitted variables from the
model.
2. Unavailability of data: Even if we know what some of the excluded variables are and
therefore consider a multiple regression rather than a simple regression, we may not have
quantitative information about these variables. So we may be forced to omit these variables from
our model despite its great theoretical relevance in explaining consumption expenditure.
3. Core variables versus peripheral variables: Assume in our consumption income example that
besides incomeX1, the number of children per familyX2, sex X3, religion X4, education X5, and
geographical region X6, also affect consumption expenditure. But it is quite possible that the joint
influence of all or some of these variables may be so small that as a practical matter and for cost
considerations it does not pay to introduce them into the model explicitly. Their combined effect
can be treated as a random variable ui.
4. Intrinsic randomness in human behaviour: Even if we succeed in introducing all the relevant
variables in to the model, there is bound to be some "intrinsic" randomness in individual Y's that
cannot be explained no matter how hard we try. The disturbances, the u's may very well reflect
this intrinsic randomness.
5. Poor proxy variables: Although the classical regression model assumes that the
variables Y and X are measured accurately, in practice the data may be plagued by errors of
measurement. Consider, for example Milton Friedman's well-known theory of the consumption
function.
He regards permanent consumptions (Yp) as a function of permanent income (Xp). In practice we
use proxy variables, such as current consumption(Y) and current income(X) which are
observable. Since the observed variables may not be equal to the actual permanent consumption
and income, we use the disturbance term. The disturbance term, u in this case may also represent
errors of measurement.
6. Principle of Parsimony : Usually the behaviour of Y is explained substantially with two or
three explanatory variables and if the theory is not strong other variables can be included. But ui
can represent all these variables. Just to keep the regression model simple, one cannot avoid
relevant and important variables.
7. Wrong functional form : Very often we do not know the form of functional relationship
between the regressand and regressors. For example, Is consumption expenditure a linear
function of income or a nonlinear function? If it's the former then, Yi= β1+β2 Xi +ui is the proper
functional relationship between Y and X, but if it's the latter Yi=β1+β2Xi2+ui may be the correct
functional form. In two variable models functional form can be judged using a scatter diagram.
But in case of multiple regression models, as the appropriate functional forms cannot be judged
this way.
For these reasons, it is important to use the stochastic disturbance term.
SPECIFICATION BIAS
3
other independent variables in the equation. If a variable is omitted, then it is not included as an
independent variable, and it is not held constant for the calculation and interpretation of the
coefficients. This omission can cause bias: It can force the expected value of the estimated
coefficient away from the true value of the population coefficient.
2. Including an irrelevant variable: If the model includes an irrelevant variable, then it has
more independent variables in the estimated equation than in the true one. The addition of a
variable to an equation where it doesn’t belong does not cause bias, but it does increase the
variances of the estimated coefficients of the included variables.
3. Choosing the wrong functional form: If we choose the wrong functional form or make the
wrong assumptions about the variables, the estimated regression may not be correct. To get an
idea of this, we look at the following example: Suppose we choose the following two models to
depict the underlying relationship between the rate of change of money wages and the
unemployment rate: where Yi = the rate of change of money wages, and Xi = the unemployment
rate. Then consider two models: The first regression model (1) is linear both in the parameters
and the variables, Yi= β0+β1Xi+Ui whereas the second model (2) is linear in the parameters (hence
a linear regression model by our definition) but nonlinear in the variable X Yi= β0+β1(1/Xi) +Ui
If (2) is the “correct” or the “true” model, fitting (1) to the scatter-points shown in the
diagram below will give us wrong predictions: Between points A and B, for any given Xi, (1) is
going to overestimate the true mean value of Y, whereas to the left of A (or to the right of B) it is
going to underestimate (or overestimate, in absolute terms) the true ean value of Y. The
preceding example is an instance of what is called a specification bias or a specification error;
here the bias consists in choosing the wrong functional form.
4
3. Zero mean value of disturbance ui. Given the value of X, the mean, or expected, value of
random disturbance term ui, is zero. Technically, the conditional mean value of ui is zero.
Symbolically we have E(ui/Xi) = 0.
The figure shows few values of X and Y, where each Y corresponds to a given X and is
distributed around its mean value. There are some Y values above and below the mean
value and the distances between them is the error terms ui. The mean value of the
deviations is equal to zero.
4. Homoscedasticity or equal variance of ui. Given the value of X, the variance of ui is the
same for all observations. That is, the conditional variances of ui are identical.
var(ui/Xi) = σ2
var(ui/Xi) = E[ui - E(ui/Xi)]2
= E{ui2 + [E(ui/Xi)] 2 - E(ui/Xi)}
= E(ui2/Xi)] [since E(ui/Xi) = 0, by assumption 3 ]
= σ2
The equation states that variance of ui for each Xi is some positive constant number equal
to σ2. This represents the assumption of homoscedasticity, or equal (homo) spread
(scedasticity) or equal variance.
F(y/x)
Demonstration of the Homskedasticity Assumption
Predicted Line Drawn Under Homoskedasticity
x1
x2
x3
x4
x
In contrast consider the below figure, where the conditional variance of the Y varies with
X. This is known as heteroscedasticity, or with unequal spread, or variance.
F(y/x)
Demonstration of the Homskedasticity Assumption
Predicted Line Drawn Under Heteroskedasticity
x1
x2
x3
x4
x
5
5. No autocorrelation between the disturbances. Given any two X values, Xi and Xj, the
correlation between any two ui and uj ,(i≠j) is zero. Symbolically,
Cov(ui uj/XiXj) =0
Cov(ui uj/XiXj) =E{[ui-E(ui)]/Xi}{[uj-E(uj)]/Xj} [cov-covariance]
=E{(ui/Xi)( uj/Xj) }
= E(ui/Xi) E(uj/Xj) [since ui and uj are generated independently]
=0
This equation postulates that the disturbances ui and uj are uncorrelated. Technically, this
is the assumption of no serial correlation or no auto correlation.
6
PROPERTIES OF THE OLS ESTIMATORS
Gauss Markov theorem states that given the assumptions of the classical linear regression
model, the least square estimators, in the class of unbiased linear estimators, have minimum variance.
In other words they are BLUE (Best Linear Unbiased Estimator).
According to the theorem that the OLS estimator is the best unbiased estimator with minimum
variance assuming the model is linear, the expected value of the error term is zero, errors are
homoskedastic and not autocorrelated, and there is no perfect multicollinearity.
The OLS estimator is widely used because it is BLUE. That is, among all unbiased linear estimators, it has
the lowest variance. The BLUE properties of the OLS estimator are often referred to as the Gauss
Markov Theorem. However, non-linear estimators may be superior to the OLS estimator, but since it is
often difficult or impossible to find the variance of unbiased nonlinear estimator, OLS estimator is most
widely used. The OLS estimator being linear is also easier to use than non linear estimators.
1. Linearity: The estimator is linear, that is, a linear function of a random variable such as the
dependant variable Y in the regression model. The regression model is linear in the parameters Yi =
ß1 + ß2 + Ui, conditional expectations of Y, E(Y/Xi) is a linear function of the parameters, the ß’s; it
may or may not be linear in the linear in the variable X. In this interpretation E(Y/Xi)= ß1 + ß2Xi2 is a
linear regression model. The term linear regression will always mean a regression that is linear in its
parameters, that is, the powers of the parameters are raised to the first power only.
2. Unbiasedness: The OLS estimator is unbiased, that is, the average or expected value of the
estimates is equal to the true value of the parameter. An estimator is unbiased if the mean of its
sampling distribution equals the true parameter. The mean of the sampling distribution is the
expected value of the estimator. Thus, lack of bias means that E(ßcap)= ß where ß is the estimator of
the true parameter, ß. Bias is then defined as the difference between expected value of the
estimator and the true parameter; that is, bias = E(ß)- ß. Note that lack of bias does not mean that
ßcap = ß, but that in repeated random sampling, we, get on average, the correct estimate. The hope is
that the sample actually obtained is close to the mean of the sampling distribution of the estimator.
3. Efficiency: The best unbiased or efficient estimator refers to the one with the smallest variance
among unbiased estimators. It is the unbiased estimator with the most compact or least spread out
distribution. An efficient estimator has the smallest confidence interval and is more likely to be
satisfied than any other estimator. The OLS estimator has minimum variance in the class of all such
linear unbiased estimators. An estimator (a function that we use to get estimates) that has a lower
variance is one whose individual data points are those that are closer to the mean. This estimator is
statistically more likely than others to provide accurate answers.
F(βx)
Unbiased and
efficient estimator
of β Biased estimator
High Sampling
of β
Variance means
inefficient
estimator of β
0
True β β + bias
7
4. Consistency: In addition to the BLUE properties, OLS estimators possess an additional property for
large samples (size more than 30). Conditions for consistency are
As the sample size increases, the estimator must approach more and more the true parameter
(This is referred to as the asymptotic unbiasedness)
As the sample size approaches infinity in the limit, the sampling distribution of the estimator
must collapse or become a straight vertical line with height (probability) of 1 above the value of
the true parameter.
In the figure, β is a consistent estimator of β because as n increases, β approaches β, and as n
approaches infinity in the limit, the sampling distribution of β collapses on β.
Demonstration of
F(βx)
Consistency
n= large
n= medium
n = small
0
True β