Econometrics I: TA Session 5: Giovanna Ubida
Econometrics I: TA Session 5: Giovanna Ubida
Giovanna Úbida
INSPER
1 / 20
Greene 3.3
2 / 20
Greene 3.3
ẽ = Mz y = (I − Z (Z 0 Z )−1 Z 0 )y
= (I − XP(P 0 X 0 XP)−1 P 0 X 0 )y
P ns
= (I − XPP −1 (X 0 X )−1 P 0−1 P 0 X 0 )y
= (I − X (X 0 X )−1 X 0 )y
= Mx y
=e
Thus, if Z is a linear transformation of X s.t. Z = XP and P is
non-singular, the residual vector are identical. Hence, the linear
transformation don’t add any information.
3 / 20
Greene 3.3
SSR e 0e ẽ 0 ẽ
R2 = 1 − = 1 − 0 0 = 0 0 = R̃ 2
SST yM y yM y
Where, following the Greene notation1 , M 0 is the n × n
idempotent matrix that transforms observations into deviations
from sample means. Note that since the dependent variable y does
not change (only the dependent one), y 0 M 0 y and ẽ = e the R 2 is
same for both cases. That is, the fit of the regression is not
affected by a linear transformation of the independent variable
1
See Section A.2.8.
4 / 20
PS02-Q3-Past exams
True or false: In the classical regression model with stochastic X, E[UT · U|X]
< E[ÛT · Û|X], since U ≡ Y − Xβ is based on the true beta parameter β, while
Û ≡ Y − Xβ̂ uses the estimated value β̂.
5 / 20
PS02-Q3-Past exams
n−k
E [Û 0 Û|X ] = E [Û 0 Û|X ]
n−k
0
Û Û
=E |X (n − k)
n−k
= E [s 2 |X ](n − k)
PS01−Q9
= (n − k)σ 2
6 / 20
PS02-Q3-Past exams
Moreover,
7 / 20
Q4 - past exams
True or False
Determine if each of the statements below is correct. If so, show why; if not,
propose a modification to correct it:
I. (Short vs. Long Regression) Let X be the matrix of regressors and
assume that it is partitioned in the usual way, X = [X1 , X2 ], where X1 and
X2 are respectively N × k1 and N × k2 with the corresponding least
squares estimator partition in β̂1 and β̂2 .
a. Suppose that instead of regressing Y on X, you run Y against
X∗1 and X∗2 together, where X∗1 = (I − X2 (XT −1 T
2 X2 ) X2 )X1 and
X∗2 = (I − X1 (XT X
1 1 ) −1 T
X1 )X 2 . Although the least squares
∗
estimator resulting from this last regression, β̂ , is such that
β̂ ∗ 6= β̂, its conditional average will be the same, E[β̂ ∗ |X] =
E[β̂|X].
b. Suppose now that instead of regressing Y on X, you regress Y
on (X1 − X∗1 ) and X∗2 together. In this case, the OLS estimator
of β2 in this regression will coincide with the estimator of the
short regression of Y on X2 . However, this will not be true for
the estimator of β1 , which will be different from the estimator
of the short regression of Y on X1 .
8 / 20
Q4 - past exams
9 / 20
Q4 - past exams
10 / 20
Q4 - past exams
Thus,
0 0
β̂1∗ = (X1∗ M2∗ X1∗ )−1 X1∗ M2∗ y
11 / 20
Q4 - past exams
Let’s work a little bit more on it
∗ 0 0
βˆ1 = (X1∗ M2∗ X1∗ )−1 X1∗ M2∗ y
= (X10 M20 M2∗ M2 X1 )−1 X10 M20 M2∗ y
@
= (X10 M20 M2 X1 )−1 X10 M20 y
= (X10 M2 X1 )−1 X10 M2 y
= β̂1
∗
By symmetry βˆ2 = β̂2 . Thus β̂ ∗ = β̂ and E [β̂ ∗ |X ] = E [β̂|X ]
False
(@) M2 is orthogonal to X2
X2∗ = M1 X2 is the part of X2 that is orthogonal to X1 (cleaning X1 of X2 )
Thus, X2∗ is smaller than X2
and M2∗ is bigger than M2
Finally, M2∗ M2 = M2
12 / 20
Q4 - past exams
Ib) Regressing Y on (X1 − X∗1 ) and X∗2 together
False
13 / 20
Q4 - past exams
IIa) and IIb) Regressing Y − y · 1 on Û
Y − y · 1 = Y − 1y
= Y − 1(10 1)−1 10 y
= Y − P 0y
= (I − P 0 )y
M 0y
β̂ = (y 0 Mx Mx y )−1 y 0 Mx M 0 y
= (y 0 Mx y )−1 y 0 Mx y
= I 6= R 2
False
14 / 20
Q8 - past exams
RLS
Define β̂R , the restricted least squares estimator, as it follows:
subject to Rc = θ0 ,
where R is a non-stochastic p × k matrix (with full rank equal to p). Show that
β̂R has a smaller conditional variance if we compare it with the OLS estimator,
β̂, under the Gauss-Markov hypothesis and under Rβ̂ = θ0 . Argue if this result
contradicts the Gauss-Markov theorem.
15 / 20
Q8 - past exams
By FOC
∂L
= −2X 0 y + 2X 0 X β̂R − R 0 λ = 0
∂c 0
2X 0 X β̂R = 2X 0 y + R 0 λ
(X 0 X )−1 R 0 λ
β̂R = (X 0 X )−1 X 0 y +
2
(X 0 X )−1 R 0 λ
β̂R = β̂ +
2
16 / 20
Q8 - past exams
∂L
= R β̂R = θ0
∂λ0
R(X 0 X )−1 R 0 λ
R β̂ + = θ0
2
λ = [R(X 0 X )−1 R 0 ]−1 2(θ0 − R β̂)
17 / 20
Q8 - past exams
18 / 20
Q8 - past exams
19 / 20
Q8 - past exams
(I)
0 −1 0 0 0 −1 0 0 −1 0 −1 0 −10 0 ( −1
(0((
−1 0 −1
ZR(X X ) R Z = (X X ) R [R(X X ) R ] R(X
X) R ( [R(X
( X ) R ] R(X X )
| {z } | {z }
Z Z0
0 −1 0 0 −1 0 −1 0 −1
= (X X ) R [R(X X ) R ] R(X X )
(II)
0 −1 0 −1 0 0 −1 0 −1 0 −1
ZR(X X ) = (X X ) R [R(X X ) R ] R(X X )
Thus,
Finally, Var (β̂R |X ) ≤ Var (β̂|X ) and Gauss-Markov still holds since β̂R is biased.
20 / 20