Econometrics - Classical Regression Assumptions
Econometrics - Classical Regression Assumptions
a) Identify the least squares estimates of the slopes and intercept coefficients. (2 marks)
b) Test the hypothesis 𝐻0 : 𝛽1 = 0 with 𝛼 = 0.1. Does there appear to be a relationship
between profits per employee and number of employees? Explain. (5 marks)
c) Interpret 𝑟 2 . (3 marks)
d) Is this fitted regression function likely to be a good tool for forecasting profits per
employee for a given number of publishing firm employees? Explain your answer.
(5 marks)
2
Assumptions of the Regression Model
•
1. The relationship between and is linear
2. The error term has a mean of zero for a given value of
3. The error terms are independent (No autocorrelation)
4. For an given the distribution of is normal
5. The spread of for each is constant. That is the standard
deviation for each error term is equal (homoscedasticity)
3
4
Homoscedastic
5
Heteroscedastic
6
•6. No Multicollinearity – The x variables are not related to each
other.
Motor accidents = (number of cars) + (number of residents)
7. Exogeneity – All the explanatory variables are unrelated with
the error terms.
Cov (x,u) = 0 (X is exogenous)
Cov (x,u) ≠ 0 (X is endogenous)
7
•
X1 Y
u1
8
Main Assumptions of the Error Terms
•1. The error terms are random variables with a mean of 0 i.e.
2. The variances of the error terms are equal i.e.
(Homoscedasticity)
3. The error terms are not correlated i.e.
4. The error terms are normally distributed
9
Desirable properties of an Estimator
1. Unbiased
2. Efficient
3. Consistent
10
Guass Markov Theorem
This theorem states that given the classical assumptions the least
square estimator is BLUE. It requires that the first six of the
seven assumptions be met.
BLUE – Best linear unbiased estimator.
11
The Gauss-Markov Theorem
•The
theorem shows that the least squares estimators have the
smallest variance of any other linear estimators. If represents
any linear estimator of and represents the OLS estimator of , it
can be shown that,
12
Assumptions of the GM theorem
•1. Linear in parameters -
2. Random sampling
3. Zero conditional mean - The error u has an expected value of
zero given any value of the explanatory variable. In other words
E(u/x) = 0
4. The error u has the same variance given any value of the
explanatory variable. In other words Var (u/x) =
13
No perfect collinearity
14