Stats-Basic Proves
Stats-Basic Proves
Stats-Basic Proves
Here you can find proofs for several properties about random variables
frequently used in the course. This is important material of the course.
Most of them have been extracted from the reference book by Stock &
Watson. For a summary of concepts and notation, see table in last page
(D) 𝑃(𝑋 = 𝑥𝑖 ) = ∑𝑚
𝑗=1 𝑃(𝑋 = 𝑥𝑖 ∩ 𝑌 = 𝑦𝑗 )
= 𝑎 ∑𝑘𝑖=1 𝑥𝑖 ∑𝑚 𝑚 𝑘
𝑗=1 𝑃(𝑥𝑖 , 𝑦𝑗 ) + 𝑏 ∑𝑗=1 𝑦𝑗 ∑𝑖=1 𝑃(𝑥𝑖 , 𝑦𝑗 ) + 𝑐
=
𝐷 𝑎 ∑𝑘𝑖=1 𝑥𝑖 𝑃(𝑋 = 𝑥𝑖 ) + 𝑏 ∑𝑚
𝑗=1 𝑦𝑗 𝑃(𝑌 = 𝑦𝑗 ) + 𝑐
=
𝐴 𝑎𝐸[𝑋] + 𝑏𝐸[𝑌] + 𝑐 ≡ 𝑎𝜇𝑋 + 𝑏𝜇𝑌 + 𝑐
Let´s define
(E) 𝑐𝑜𝑣(𝑋, 𝑌) = 𝐸[(𝑋 − 𝜇𝑋 )(𝑌 − 𝜇𝑌 )] ≡ 𝜎𝑋𝑌
then
(5) 𝑐𝑜𝑣(𝑋, 𝑌) =
𝐸 𝐸[(𝑋 − 𝜇𝑋 )(𝑌 − 𝜇𝑌 )] =
= 𝐸(𝑋𝑌 − 𝑋𝜇𝑌 − 𝑌𝜇𝑋 + 𝜇𝑋 𝜇𝑌 )
=
4 𝐸(𝑋𝑌) – 𝜇𝑌 𝜇𝑋 − 𝜇𝑋 𝜇𝑌 + 𝜇𝑋 𝜇𝑌
= 𝐸(𝑋𝑌)−𝜇𝑋 𝜇𝑌
(6) 𝑐𝑜𝑣(𝑋, 𝑋) =
𝐸 𝐸[(𝑋 − 𝜇𝑋 )(𝑋 − 𝜇𝑋 )]
=
𝐵 𝑣𝑎𝑟(𝑋) ≡ 𝜎𝑋2
(7) 𝑐𝑜𝑣(𝑎𝑋, 𝑐 + 𝑏𝑌) 𝐸==4 𝐸[(𝑎𝑋 − 𝑎𝜇𝑋 )(𝑐 + 𝑏𝑌 − 𝑐 − 𝑏𝜇𝑌 )] or use (5)!
= 𝐸[𝑎𝑏(𝑋 − 𝜇𝑋 )(𝑌 − 𝜇𝑌 )]
=
4 𝑎𝑏 𝐸[(𝑋 − 𝜇𝑋 )(𝑌 − 𝜇𝑌 )]
=
𝐸 𝑎𝑏 𝑐𝑜𝑣(𝑋, 𝑌) ≡ 𝑎𝑏 𝜎𝑋𝑌
(8) 𝑐𝑜𝑣(𝑎𝑋 + 𝑏𝑌 + 𝑐, 𝑍) =7 𝑐𝑜𝑣(𝑎𝑋 + 𝑏𝑌, 𝑍)
==
𝐸 4 𝐸[{(𝑎𝑋 + 𝑏𝑌) − (𝑎𝜇𝑋 + 𝑏𝜇𝑌 )}{𝑍 − 𝜇𝑍 }] or use (5)!
and, in general, if we have random variables {X1, X2, …, Xs, …, XS } & {Y1,
Y2, …, Yt, …, YT} and constants { a1, a2, …, as, …, aS ;b1, b2, …, bt, …, bT ; c}
=
𝐺 ∑𝑚 𝑘
𝑗=1[∑𝑖=1 𝑥𝑖 𝑃(𝑋 = 𝑥𝑖 |𝑌 = 𝑦𝑗 )]𝑃(𝑌 = 𝑦𝑗 )
=
𝐹 ∑𝑚 𝑘
𝑗=1 ∑𝑖=1 𝑥𝑖 𝑃(𝑋 = 𝑥𝑖 ∩ 𝑌 = 𝑦𝑗 )
= ∑𝑘𝑖=1 ∑𝑚
𝑗=1 𝑥𝑖 𝑃(𝑋 = 𝑥𝑖 ∩ 𝑌 = 𝑦𝑗 )
= ∑𝑘𝑖=1 𝑥𝑖 [∑𝑚
𝑗=1 𝑃(𝑋 = 𝑥𝑖 ∩ 𝑌 = 𝑦𝑗 )]
=
𝐷 ∑𝑘𝑖=1 𝑥𝑖 𝑃(𝑋 = 𝑥𝑖 )
=
𝐴 𝐸(𝑋) ≡ 𝜇𝑋 Law of Iterated Expectations (LIE)
Then
(13) 𝑃(𝑋 = 𝑥𝑖 ∩ 𝑌 = 𝑦𝑗 ) ==
𝐹 𝐼 𝑃(𝑋 = 𝑥𝑖 ) ∙ 𝑃(𝑌 = 𝑦𝑗 ) ∀ 𝑥𝑖 , 𝑦𝑗
Finally, remember
𝜎𝑋𝑌
(J) 𝑐𝑜𝑟𝑟(𝑋, 𝑌) = ⁄𝜎𝑋 𝜎𝑌 ≡ 𝜌𝑋𝑌
Thus,
(14) 𝐼𝑓 𝑋 & 𝑌 𝑖𝑛𝑑𝑒𝑝𝑒𝑛𝑑𝑒𝑛𝑡 , 𝑡ℎ𝑒𝑛 𝑐𝑜𝑟𝑟(𝑋, 𝑌) = 0
X & Y are random variables (not constants), therefore 𝜎𝑋 > 0 & 𝜎𝑌 > 0.
Thus, 𝑐𝑜𝑟𝑟(𝑋, 𝑌) = 0 iff 𝑐𝑜𝑣(𝑋, 𝑌) = 0 or, by (5), 𝐸(𝑋𝑌) = 𝐸(𝑋)𝐸(𝑌).
𝐸(𝑋𝑌) =𝐴 ∑𝑘𝑖=1 ∑𝑚
𝑗=1 𝑥𝑖 𝑦𝑗 𝑃(𝑋 = 𝑥𝑖 ∩ 𝑌 = 𝑦𝑗 )
=
𝐼 ∑𝑘𝑖=1 ∑𝑚
𝑗=1 𝑥𝑖 𝑦𝑗 𝑃(𝑋 = 𝑥𝑖 ) 𝑃 (𝑌 = 𝑦𝑗 )
= ∑𝑘𝑖=1 𝑥𝑖 𝑃(𝑋 = 𝑥𝑖 ) ∑𝑚
𝑗=1 𝑦𝑗 𝑃(𝑌 = 𝑦𝑗 )
=
𝐴 𝐸(𝑋)𝐸(𝑌)
Econometrics I Academic Year 2020-2021
1
𝐸(𝑋) = ∑𝑘𝑖=1 𝑥𝑖 𝑃(𝑋 = 𝑥𝑖 ) ≡ 𝜇𝑋 𝑥 = ∑𝑛𝑖=1 𝑥𝑖 ≡ 𝜇̂ 𝑋
𝑛
1
𝑣𝑎𝑟(𝑋) = 𝐸[(𝑋 − 𝜇𝑋 )2 ] ≡ 𝜎𝑋2 𝑠𝑋2 = ∑𝑛𝑖=1(𝑥𝑖 − 𝑥)2 ≡ 𝜎̂𝑋2
𝑛−1
𝑐𝑜𝑣(𝑋, 𝑌) = 𝑠𝑋𝑌 =
1
= 𝐸[(𝑋 − 𝜇𝑋 )(𝑌 − 𝜇𝑌 )] ≡ 𝜎𝑋𝑌 = ∑𝑛𝑖=1(𝑥𝑖 − 𝑥)(𝑦𝑖 − 𝑦) ≡ 𝜎̂𝑋𝑌
𝑛−1
𝜎𝑋𝑌 𝑠𝑋𝑌
𝑐𝑜𝑟𝑟(𝑋, 𝑌) = ⁄𝜎𝑋 𝜎𝑌 ≡ 𝜌𝑋𝑌 𝑟𝑋𝑌 = ⁄𝑠𝑋 𝑠𝑌 ≡ 𝜌̂𝑋𝑌
Moreover, the statistics on the right are also functions of the data, thus
random variables whose realization depends on the sample.
For example, X is the height of Europeans and we want to estimate 𝜇𝑋 ,
the unknown mean or expected value of the height of Europeans, E(X).
Juan selects randomly a representative sample of 500 Spaniards and
obtains 𝑥 = 1.67: this is his estimate of 𝜇𝑋 (𝜇̂ 𝑋 ). But Giampiero takes
another sample of 500 Italians and gets 𝑥 = 1.65. And Catiana draws a
representative sample of 500 Germans that gives 𝑥 = 1.70. Thus, {1.65,
1.67, 1.70} are realizations of the random variable 𝑋, the average height
of a representative sample of 500 Europeans. This random variable 𝑋
also has moments: mean, variance, etc; that we can estimate with data.
𝐸(𝑋) = 𝜇𝑋 ≡ 𝜇𝑋̅ 𝜇̂ 𝑋̅ = 𝑥
1 2 1
𝑣𝑎𝑟(𝑋) = 𝜎𝑋 ≡ 𝜎𝑋2̅ 𝜎̂𝑋2̅ = 𝜎̂ 2 ≡ 𝑠𝑋2̅
𝑛 𝑛 𝑋
1 1
𝑠𝑑(𝑋) = √ 𝜎𝑋2 ≡ 𝜎𝑋̅ 𝜎̂𝑋̅ = √ 𝜎̂𝑋2 ≡ 𝑆𝐸(𝑋)
𝑛 𝑛