Two Stage Sampling 1
Two Stage Sampling 1
22 August 2020
h i h i
V (θ̂) = V1 E2 θ̂ + E1 V2 θ̂
Proof:
2
V (θ̂) = E1 E2 θ̂ − θ , θ̂ is unbiased for θ
2
E2 θ̂ − θ = E2 (θ̂2 ) − 2θE2 (θ̂) + θ2
h i2
= E2 (θ̂2 ) − θ2 = E2 (θ̂) + V2 (θ̂) − θ2
Theorem
Suppose that Yi is the total of ith primary sampling unit (PSU) and has
an unbiased estimate Ŷi considering only the second stage sampling.
Moreover, let wi is the weight for the ith PSU depending on the first
stage sampling. Then an unbiased estimate of population total Y can be
obtained based on a two stage sampling as
n1
X
Ŷ2ss = wi Ŷi ,
i=1
where the subscripts 1 and 2 indicate the expectation and variability are
to be computed for 1st and 2nd stage sampling respectively.
Now,
n1
!
X
E1 V2 Ŷ2ss = E1 V2 wi Ŷi
i=1
N1
!
X 0
= E1 V 2 wi Ŷi
i=1
(N )
1
X 0
2
= E1 wi V2 Ŷi ; (subsampling is independent)
i=1
N1
X 0
= E1 wi2 V2 Ŷi
i=1
Again,
n1
!
X
V1 E2 Ŷ2ss = V1 E2 wi Ŷi
i=1
(n )
X 1
= V1 wi E2 Ŷi
i=1
n1
!
X
= V1 wi Yi
i=1
Finally we have,
n1
! N1
X X 0
V Ŷ2ss = V1 wi Yi + E1 wi2 V2 Ŷi
i=1 i=1
Theorem
An unbiased estimate of V (Ŷ2ss ) can be obtained by,
n1
! n1
X X
v Ŷ2ss = v1 wi Ŷi + wi v2 Ŷi ,
i=1 i=1
where
E2 [v2 (Yi )] = V2 Ŷi , and
" n1
!# n1
!
X X
E1 v1 wi Ŷi = V1 wi Ŷi .
i=1 i=1
n1
! N1
!
X X 0
V1 wi Yi = V1 wi Yi
i=1 i=1
N1 N1 X
N1
X 0 X 0 0
= Yi2 V1 (wi ) + 2 Yi Yj cov(wi wj ) (1)
i=1 i=1 j>i
and,
Using the idea in equation 1, the 1st stage estimated variance can be
written as,
n1
! N1 N1 X
N1
X X 0 X 0
v1 wi Yi = ai Yi2 + 2 bij Yi Yj . (2)
i=1 i=1 i=1 j>i
Comparing the equations (1) and (2) and using the given unbiasedness
condition (second) we can write,
0 0
E1 ai = V1 wi
Now, using the second stage estimators the equation (2) takes the
following form,
n1
! N1 N1 X
N1
X X 0 X 0
v1 wi Ŷi = ai Ŷi2 + 2 bij Ŷi Ŷj .
i=1 i=1 i=1 j>i
Now,
h i h i
E v Ŷ2ss = E1 E2 v Ŷ2ss
" n1
! n1
#
X X
= E1 E2 v1 wi Ŷi + wi v2 Ŷi
i=1 i=1
N1 N1 X
N1 n1 h i
X 0 X 0 X
2
= E1 E2 ai Ŷi + 2 bij Ŷi Ŷj + E1
wi E2 v2 Ŷi
i=1 i=1 j>i i=1
N1 N1
N1 X
X 0 X 0
= E1 ai E2 Ŷi2 + 2 bij E2 Ŷi E2 Ŷj
i=1 i=1 j>i
n1
X
+ E1 wi V2 Ŷi
i=1
N1 N1 XN1 n1
X 0
n o2 X 0
X
= E1 ai V2 Ŷi + E2 Ŷi +2 bij Yi Yj + E1 wi V2 Ŷi
i=1 i=1 j>i i=1
N1 h 0 0 i X N1 N1 X
N1
X 0 X 0
= E1 ai + E1 wi V2 Ŷi + E1 ai Yi2 + 2 bij Yi Yj
i=1 i=1 i=1 j>i
N1 h
( n1
!)
X 0 0 i X
= E1 ai + E1 wi V2 Ŷi + E1 v1 wi Yi
i=1 i=1
N1 n1
!
X 0 n 0 o2 X
= V1 wi + E1 wi V2 Ŷi + V1 wi Yi
i=1 i=1
N1 n1
!
X 0
X
2
= E1 wi V2 Ŷi + V1 wi Yi
i=1 i=1
= V Ŷ2ss
E1 (wi0 )
if and only if = 1 for ∀i. Note that n is the number of PSU in
the sample and N is the number of PSU in the population.
Anower Hossain Sampling
Two stage sampling
An unbiased estimator of V Ŷ2ss is given by
n
! n
X X
v Ŷ2ss = v1 wi Ŷi + wi v2 Ŷi
i=1 i=1
n
X N
Ŷ2SRS = Ŷi
i=1
n
N (N − n) XN N
N X Mi (Mi − mi ) 2
Var Ŷ2SRS = S12 + S2i
n i=1
n i=1 mi
where !2
N N
1 X 1 X
S12 = Yi − Yi
N − 1 i=1 N i=1
and !2
Mi M
2 1 X 1 Xi
S2i = Yij − Yij
Mi − 1 j=1 Mi i=1
n n
c Ŷ2SRS = N (N − n) N X Mi (Mi − mi ) 2
X
Var s21 + s2i
n i=1
n i=1 mi
where !2
n n
1 X 1X
s21 = Ŷi − Ŷi
n − 1 i=1 n i=1
and !2
mi mi
1 X 1 X
s22i = Yij − Yij
mi − 1 j=1 mi i=1
n
!
X N
E Ŷ2SRS = E1 E2 Ŷi
i=1
n
n
!
X N
= E1 E2 (Ŷi )
i=1
n
n
!
X N
= E1 Yi
i=1
n
n
!
N X
= E1 Yi
n i=1
N
!
N X n
= Yi
n i=1
N
N
X
= Yi = Y
i=1
Horvitz-Thompson estimator
n
X yi
ŶHT =
π
i=1 i
N N XN
X 1 − πi X Yi Yj
V1 ŶHT = Yi2 + (πij − πi πj )
i=1
πi i=1
πi πj
j6=i
n n X
n
X 1 − πi X πij − πi πj Yi Yj
V
b 1 ŶHT = Yi2 +
i=1
πi2 i=1 j6=i
πij πi πj
Horvitz-Thompson estimator
n
X yi
ŶHT =
π
i=1 i
N X
N 2
X Yi Yj
V2 ŶHT = (πi πj − πij ) −
i=1 j<i
πi πj
n X n 2
X πi πj − πij Yi Yj
V
b 2 ŶHT = −
i=1 j<i
πij πi πj
Let
πi : prob. that ith PSU is in the sample
πij : prob. that both ith and jth PSUs are in the sample
πi,k : prob. that kth subumit of the ith PSU is in the sample
πi,kl : prob. that both kth and lth subunits of ith PSU are in the sample
Then
E1 (wi0 ) = wi × πi + 0 × (1 − πi ) = 1
which imples
1
wi =
πi
Similarly, we get
1
E1 (wi02 ) = wi2 × πi + 02 × (1 − πi ) =
πi
Then
N X
N 2
X Yi Yj
V Ŷ2upwor = (πi πj − πij ) −
i=1 j<i
πi πj
N M M 2
1 Xi Xi
X Yil Yik
+ (πi,l πi,k − πi,kl ) −
π
i=1 i
πi,l πi,k
l=1 k<l
and
(
uij , if the jth SSU in the ith PSU is in the sample
u0ij =
0, otherwise.
if and only if E1 (wi0 ) = 1 and E2 u0ij = 1 for ∀i, j. Note that n is the
Unbiasedness:
Xn mi
X
E(Ŷ3ss ) = E1 E2 E3 wi uij Ŷij
i=1 j=1
XN Mi
X
= E1 E2 E3 wi0 u0ij Ŷij
i=1 j=1
N
X Mi
X
= E1 (wi0 ) E2 (u0ij )E3 (Ŷij )
i=1 j=1
Mi
N X
X
= Yij
i=1 j=1
=Y
Variance:
Var(Ŷ3ss ) = E1 E2 V3 (Ŷ3ss ) + E1 V2 E3 (Ŷ3ss ) + V1 E2 E3 (Ŷ3ss )
1st Term:
Xn mi
X
E1 E2 V3 (Ŷ3ss ) = E1 E2 V3 wi uij Ŷij
i=1 j=1
XN Mi
X
= E1 E2 V3 wi0 u0ij Ŷij
i=1 j=1
XN Mi
X
= E1 E2 wi02 u02
ij V3 (Ŷij )
i=1 j=1
N
X Mi
X
= E1 (wi02 ) E2 (u02
ij )V3 (Ŷij )
i=1 j=1
2nd Term:
Xn mi
X
E1 V2 E3 (Ŷ3ss ) = E1 V2 E3 wi uij Ŷij
i=1 j=1
n
X mi
X
= E1 V2 wi uij E3 (Ŷij )
i=1 j=1
n
X mi
X
= E1 V2 wi uij Yij
i=1 j=1
XN mi
X
= E1 wi02 V2 uij Yij
i=1 j=1
N
X Xmi
= E1 (wi02 )V2 uij Yij
i=1 j=1
3rd Term:
Xn mi
X
V1 E2 E3 (Ŷ3ss ) = E1 V2 E3 wi uij Ŷij
i=1 j=1
n
X mi
X
= V1 E2 wi uij E3 (Ŷij )
i=1 j=1
n
X mi
X
= V1 E2 wi uij Yij
i=1 j=1
Xn Mi
X Xn Mi
X
= V1 wi E2 (u0ij )Yij = V1 wi Yij
i=1 j=1 i=1 j=1
n
!
X
= V1 wi Yi
i=1