Linear Estimation For Random Delay Systems
Linear Estimation For Random Delay Systems
Linear Estimation For Random Delay Systems
School of Control Science and Engineering, Shandong University, Jinan 250061, Shandong, PR China
Department of Manufacturing Engineering and Engineering Management, City University of Hong Kong, Hong Kong
School of Control Science and Engineering, University of Jinan, Jinan 250022, Shandong, PR China
article
info
Article history:
Received 2 June 2010
Received in revised form
27 November 2010
Accepted 25 March 2011
Available online 10 May 2011
Keywords:
Linear estimation
Random delay
Reorganized innovation analysis
Riccati equation
Convergence
Stability
abstract
This paper is concerned with the linear estimation problems for discrete-time systems with random
delayed observations. When the random delay is known online, i.e., time-stamped, the random delayed
system is reconstructed as an equivalent delay-free one by using measurement reorganization technique,
and then an optimal linear filter is presented based on the Kalman filtering technique. However, the
optimal filter is time-varying, stochastic, and does not converge to a steady state in general. Then an
alternative suboptimal filter with deterministic gains is developed under a new criteria. The estimator
performance in terms of their error covariances is provided, and its mean square stability is established.
Finally, a numerical example is presented to illustrate the efficiency of proposed estimators.
2011 Elsevier B.V. All rights reserved.
1. Introduction
As the result of the increasing development in communication
networks, control and state estimation over networks have
attracted great attention during the past few years (see e.g. [1]).
The feedback control systems wherein the control loops are closed
through a real-time network are called networked control systems
(NCSs) (see e.g. [26]). In NCSs, data typically travel through
communication networks from sensors to controllers and from
controllers to actuators. As a direct consequence of the finite
bandwidth for data transmission over networks, time delay, either
from sensors to controllers, called sensor delay, or from controllers
to actuators, called controller delay, or both, is inevitable in
networked systems where a common medium is used for data
transfer. This time delay, being either constant, time-varying, or
random, can degrade the performance of a control system if it is not
given due consideration. In many instances it can even destabilize
the system. Hence time delay is one of the challenging problems
faced by control practitioners in NCSs.
The filtering problem for networked control systems with
sensor delays especially for random sensor delays has received
0167-6911/$ see front matter 2011 Elsevier B.V. All rights reserved.
doi:10.1016/j.sysconle.2011.03.009
much attention during the past few years (See e.g. [7,8]). On
the filtering problems with intermittent observations, the initial
work can be traced back to Nahi [9] and Hadidi [10]. Recently,
this problem has been studied in [11,12], respectively, where the
statistical convergence property of estimation error covariances
was studied, and the existence of a critical value for the arrival
rate of observations was shown. For the situation that the
one-step sensor delay was described as a binary white noise
sequence, a reduced-order linear unbiased estimator was designed
via state augmentation in [13]. When the random delay was
characterized by a set of Bernoulli variables, the unscented filtering
algorithms [14], the linear and quadratic least-square estimation
method [15], the optimal H2 and Kalman filtering [16,17], and
the H filtering [18,19] have been developed. The rational of
modeling the random delay as Bernoulli variable sequences has
been justified in those papers. On the other hand, modeling the
random delay as a finite state Markov chain is also a reasonable
way. The relevant estimation results for this type of modeling can
be found in [20,21], and the reference therein. It can be noted
that the results mentioned above mainly focus on the systems
just with one- or two-step delay. To the best of our knowledge,
there exist few estimation results on multiple random delayed
systems [22]. Furthermore, most existing results employ state
augmentation method to deal with random delays. In fact, the
reorganization observation method which was developed in our
previous works [23] is also an effective tool to deal with the
random delay without state augmentation.
In this paper we will investigate the optimal and suboptimal linear estimators for discrete-time systems with random observation
451
delays, where the delay steps are larger than one sampling period
and known online via time-stamped data. The key technique to be
used for dealing with the random delay is the reorganization observation method, and by this method the random delayed system
is transformed into a delay-free one. An optimal linear filter is presented with time-varying stochastic gains, and the solution to this
estimator is also discussed. Furthermore, an alternative suboptimal estimator with deterministic gains is developed under a new
performance index. Convergence and mean square stability of this
estimator are established. It is shown that the stability of this estimator does not depend on observation packet delay but only on the
overall observation packet loss probability. Note that both filters
(optimal and suboptimal) have the same dimension as the original
systems.
The remainder of this paper is organized as follows. In Section 2,
we state the problem formulation. In Section 3, the reorganized
observations are defined, and an optimal filter is designed by
the projection formula. In Section 4, an alternative suboptimal
linear filter is developed under a new criteria. Its convergence and
stability are discussed. Finally, a numerical example is given in
Section 5, which is followed by some conclusions in Section 6.
Notations: Throughout this paper, Rn denotes the n-dimensional Euclidean space, Rmn denotes the set of all m n real matrices. A real symmetric matrix Q > 0(0) denotes Q being a
positive-definite (or positive semi-definite) matrix, and A > ()B
means A B > ()0. A stands for the transpose of the matrix A.
Q 1 and Q 2 represent the reverse and a square root of the positivedefinite matrix Q . diag{...} denotes a block-diagonal matrix, P
indicates the occurrence probability of an event, and E {.} represents the mathematical expectation operator. As usual, we define
H = L2 ( , F , P) as the Hilbert space of all square summable in
the probability space ( , F , P), equipped with the inner product
x, y = E {xy }, x, x = E {xx } = x2 , where . represents the
standard Euclidean norm. L{y1 , . . . , yn } denotes the linear space
spanned by y1 , . . . , yn . Proj{.} denotes the projection operator, and
ts means the Kronecker delta function. In addition, for a realvalued function with domain X , arg minxX (x) = {x X :
(x) = minyX (y)}.
y(t ) = [y0 (t )
yr (t )] ,
(2.3)
where
yi (t ) = i,t Hx(t i) + i,t v(t i),
i = 0, . . . , r ,
(2.4)
i,t =
0,
(2.5)
i,t +i j,t +j = 0,
i = j.
(2.6)
y(t ) = [y0 (t )
yt (t )
0] ,
(2.7)
2. Problem formulations
3.1. Measurements reorganization
Consider the following discrete-time systems with randomly
delayed measurements
x(t + 1) = x(t ) + Gw(t ),
x(0),
(2.1)
(2.2)
y r (s) ,
y0 (s)
..
, 0 s t r,
.
yr (s + r )
y0 (s)
.
y t s (s) , .. , t r < s t .
yt s (t )
(3.1)
(3.2)
Assumption 2.1. The initial state x(0), w(t ), and v(t ) are null
mean white noises with covariance matrices
E [x(0)x (0)] = P0 ,
E [w(t )w (s)] = Q ts ,
0 s t r,
where
0,s H
.
Hr (s) = .. ,
r ,s+r H
0,s H
.
Ht s (s) = .. ,
t s,t H
t r < s t,
0,s v(s)
..
vr (s) =
,
.
r ,s+r v(s)
0,s v(s)
..
vt s (s) =
,
.
(3.3)
(3.4)
t s,t v(s)
(3.5)
(3.6)
452
with vr (s) and vt s (s) being white noises of zero means and
respective covariance matrices
Rr (s) = diag{0,s R, . . . , r ,s R},
(3.7)
and
Rt s (s) = diag{0,s R, . . . , t s,t R}.
(3.8)
(3.22)
r
{{yr (s)}ts=
yt s (s)}ts=t r +1 },
0 ; {
(3.9)
For 0 s t r, define
r (s) , y r (s) y r (s),
(3.10)
(3.11)
(3.23)
0 s t r,
(3.24)
s > t r.
(3.25)
(3.12)
K0 (t )Q0 (t ) = P0 (t )H0 (t ),
with
y t (s1) (s 1)}.
(3.13)
(3.26)
(3.27)
Q0 (t ) = H0 (t )P1 (t )H0 (t ) + R0 (t )
and the estimation x (t , 1) is computed by the following iteration.
x (0, r ) = 0,
(3.28)
where
Definition 3.2. Consider the given time instant t.
(3.14)
(3.30)
with
Qr (s) = Hr (s)Pr (s)Hr (s) + Rr (s),
(3.15)
(3.29)
(3.16)
(3.31)
;
(3.32)
where
x (s, r ) = x(s) x (s, r ).
(3.17)
x (s + 1, t s) = t s (s)x(s, t (s 1))
+ Kt s (s)yt s (s),
Ht s (s)x(s, t s + 1)],
where
(3.18)
(3.34)
Ew,v x(s + 1) x (s + 1, t s)
with
(3.33)
(3.19)
(3.35)
(3.36)
Pt s (s + 1) = Pt (s1) (s)
Kt s (s)Qt s (s)Kts (s) + GQG .
(3.37)
(3.20)
where
x (s, t s + 1) = x(s) x (s, t s + 1).
(3.21)
(3.38)
(3.39)
where K0 (t ) is to be determined.
Case 1: When the covariance matrix of the innovation 0 (t ),
denoted by Q0 (t ), is invertible, Proj{x(t ) | 0 (t )} is given by
Proj{x(t ) | 0 (t )} = P1 (t )H0 (t )Q01 (t )0 (t ),
(3.40)
Remark 3.3. Compared to the existing work in [22], the contributions of our work can be described as follows: a new model
for the system with finite multiple random observation delays is
constructed, and another efficient methodthe reorganized innovation analysis method is employed for the design of the optimal
estimator with random observation delays.
Remark 3.4. The LMMSE estimator described above is optimal
for any observation packet delay generating process. However
from an engineering perspective it is also desirable to further
characterize the performance of the estimator. In this scenario a
natural performance metric is the expected error covariance, i.e.,
E [Pt +1|t ], where the expectation is performed with respect to the
arrival binary random variable {i,s : i = 0, . . . , r ; s = 0, . . . , t }.
Unfortunately, it is not clear whether such quantity converges to
a steady state. Thus instead of trying to obtain the upper bound
on the expected error covariance of the time-varying LMMSE
estimator, we will, in the next section focus on a suboptimal filter
with deterministic gains. That is the filter gains will not contain the
observation packet arrival binary random variable .,. , instead the
gains will be determined by the probability of the random variable.
(3.41)
4. Suboptimal linear estimator
(3.42)
(3.43)
(3.44)
(3.45)
453
(3.46)
which is (3.32).
Finally, following the similar procedure as in the determination
of Kr (s) and Pr (s), we shall obtain (3.33)(3.37). This completes the
derivation of the solution to the Problem 1.
Remark 3.2. It should be pointed out that the Riccati equation
(3.32) is not in a standard form as the innovation covariance matrix
Qr (s) may be singular. In order to solve the Riccati equation (3.32)
for the filter design, we should first solve the linear equation (3.30),
which may be unsolvable or solvable but with infinite number of
solutions. Fortunately it can be shown that Eq. (3.30) is solvable and
any solution, if there exist more than one solution, to the equation
yields the same result for the Riccati equation (3.32). More detailed
discussion can be seen in Appendix.
0 s t r,
Rt s = diag{0 R, . . . , t s R},
t r < s t.
(4.1)
(4.2)
(4.3)
t
s=0
}.
Definition 4.1. Given (2.1), (3.3) and (3.4), the linear suboptimal
estimation x e (t | t ) of x(t ) with deterministic gains is defined as
x e (t | t ) , x e (t , 1) + K0 (t )[y0 (t ) H0 (t )xe (t , 1)],
(4.4)
(4.5)
For 0 s t r, define
x e (s + 1, r ) , x e (s, r ) + Kr (s)[yr (s) Hr (s)xe (s, r )],
(4.6)
(4.7)
454
(4.21)
(4.8)
Pt s (s + 1) = Pt s+1 (s)
(4.10)
(4.11)
where
1
Pt s+1 (s)H HPt s+1 (s)H + R
HPt s+1 (s) + GQG ,
is minimized.
Remark 4.1. Note from the criteria index of (3.15) (or (3.19)) and
(4.7) (or (4.9)) that, the estimation x e (s, r ) or x e (s, t s + 1) defined
in Definition 4.1 is different from the one defined in Definition 3.2.
The expectation in (4.7) (or (4.9)) is taken over w, v and {.,. }
simultaneously.
i =0
(4.9)
t s
x e (s + 1, t s) = x(s + 1) x e (s + 1, t s).
(4.12)
(4.23)
s t r,
(4.24)
s > t r.
(4.25)
x e (s + 1, r ) = x(s + 1) x e (s + 1, r ),
(4.22)
(4.26)
where M (t ) = e0 (t ), e0 (t ) and
K0 (t ) = xe (t , 1), e0 (t )M 1 (t ).
(4.13)
where
K0 (t ) = P1 (t )H0 [0 HP1 (t )H + 0 R]1 ,
(4.14)
H0 = E [H0 (t )] = 0 H ,
(4.15)
y r (s), x e (0, r ) = 0,
(4.16)
where
Kr (s) = [ Pr (s)Hr ]
Pr (s + 1) = Pr (s)
i Pr (s)H
i=0
HPr (s)H + R
Pr (0) = P0 , (4.18)
(4.19)
(4.28)
M (t ) = H0 (t )xe (t , 1) + v0 (t ), H0 (t )xe (t , 1) + v0 (t )
= 0 HP1 (t )H + 0 R.
(4.29)
Kr (s)vr (s).
(4.27)
(4.30)
E ,w,v x(s + 1) x e (s + 1, r )2
= Pr (s) Pr (s)Hr Kr (s)
Kr (s)Hr Pr (s) + GQG + Kr (s)
[diag{0 HPr (s)H , . . . , r HPr (s)H } + Rr ]Kr (s)
= [Kr (s) Kr (s)]
[diag{0 HPr (s)H , . . . , r HPr (s)H } + Rr ]
(4.31)
where
Kr (s) = [ Pr (s)Hr ]
(4.32)
Pr (s + 1) = Pr (s)
455
i Pr (s)H
(4.36)
where
i=0
(4.33)
g0 ,...,r (P ) = P
i PH
i=0
+ R)1 H P + GQG .
(H PH
Then for any initial condition Pr (0), the Riccati difference equations
for Pr (t r ), Pr 1 (t r + 1), . . . , P0 (t ) converge to a unique set of
algebraic equations when t ,
Pr = Pr
i Pr H (HPr H + R)1
i=0
P r 1
HPr + GQG ,
r 1
= Pr
i Pr H (HPr H + R)1
(4.37)
i=0
HPr + GQG ,
(4.38)
..
.
(4.39)
x e (t r + 1, r ) = [ Kr Hr (t r )]xe (t r , r )
In the following our aim is to show that for an arbitrary but fixed
nonnegative symmetric Pr (t0 ), Pr (t r ), Pr 1 (t r + 1), . . . , P0 (t )
are convergent when t . At first, we present some useful
lemmas without proof since similar derivation lines can be found
in [11].
+ Kr y r (t r ),
..
.
x e (t , 1) = [ K1 H1 (t 1)]xe (t 1, 2) + K1 y 1 (t 1),
(4.40)
x e (t |t ) = [I K0 H0 (t )]xe (t , 1) + Ko y 0 (t ),
(4.42)
where
g (X )
= X + GQG XH (HXH + R) HX ,
(K , X ) = [ KH ]X [ KH ] + (1 )KHXH K
+GQG + KRK ,
(4.35)
where 0 1. Assume X S = {S R
nn
(4.34)
|S 0}, R > 0, Q
(X ) = [ KH ]X [ KH ] + (1 )KHX (KH ) .
Suppose there exists X > 0 such that X > L(X ).
(i) For all W 0, limt t (W ) = 0.
(ii) Let V > 0 and consider the linear system Xt +1 = (Xt ) + V
initialized at X0 . Then, the sequence Xt is bounded.
Lemma 4.3. Consider the operator g (X ) defined in (4.34). Suppose
there exists a positive-definite matrix P such that
P > g (P ).
Then for any P0 , the sequence P (t ) = gt (P0 ) is bounded, i.e., there
exists MP0 0 dependent on P0 such that P (t ) MP0 , t.
With these lemmas, one can obtain the following results on
the convergence properties of the Riccati difference equations in
Theorem 4.1.
(4.41)
..
.
K1 = P2 H1 [diag{0 HP2 H , 1 HP2 H } + R1 ]1 ,
K0 = P1 H0 [0 HP1 H + 0 R]1 .
Proof. To derive the claimed results, we just need to analyze
the convergence of the Riccati difference equations for (4.18) and
(4.22). The whole process in the derivation are divided into three
stages.
(i) To show the convergence of the Riccati equation (4.18). First,
denote
g0 ,...,r (X ) = X + GQG
i XH (HXH + R)1 HX ,
i=0
(4.43)
then
Pr (s + 1) = g0 ,...,r (Pr (s)) = gs 0 ,...,r (Pr (0)).
Recalling Assumption 2.2, we obtain that 0
i=0 i 1. Thus
(4.43) satisfies the conditions of Lemma 4.1. Consider the Riccati
equation (4.18) initialized at Qr (0) = 0 and let Qr (s) = gs 0 ,...,r (0),
then 0 = Qr (0) Qr (1). It follows from Lemma 4.1(iii) that
456
Here, we have used the Lemma 4.3 to bound the sequence {Qr (s)}.
We now have a monotone nondecreasing sequence of matrices
bounded above. It is obvious that
Pr = Pr + GQG
(4.44)
i KP r H X
i =0
i 1
i=0
i KPr H
i =0
r
i=0
i =0
(4.45)
P r , P r , . . . .
However, we have shown that for any initial condition (4.18)
converges to Pr . Thus, P r = Pr . The uniqueness of the solutions
is proved.
1
(F , Y )
Then in light of Lemma 4.2 and taking limit on both sides of the
above inequalities, one obtains that
0 lim (Rr (s + 1) Pr ) 0.
s.
proving (4.37).
(ii) To show the convergence of the expressions in (4.22), that
is,
r
(1 )FH
1
2
F ( R) 2
0
YGQ
0
(4.50)
Moreover, the suboptimal estimators become the steady-state ones.
Proof. First, we prove that the following statements are equivalent:
(i) X > 0 such that X > g (X ),
(ii) K , X > 0 such that X > (K , X ),
(iii) F and 0 < Y I, such that (F , Y ) > 0.
Let K = X H (H X H + R)1 , then we will obtain the equivalence between (i) and (ii) immediately. Further, by using the Schur
complements, we can prove that (F , Y ) > 0 if and only if
Y > 0,
(4.51)
,F = X
(4.52)
X > 0,
i Pr (t r 1)H
(4.53)
X [ KH ]X [ KH ] (1 )KHXH K
i=0
(4.49)
Y [Y FH ]Y 1 [Y FH ]
Pr (t r ) = Pr (t r 1)
Y > 0, 0 < 1,
(Y FH )
( R) 2 F
(Y FH )
= (1 )H F
Q 2 G Y
(4.48)
Observe that
(4.47)
P0 (t ) = P1 (t 1) 0 P1 (t 1)H
i=0
..
.
i Pr (t r )H
lim Qr (s) = Pr .
(X ) =
r 1
i =0
Pr 1 (t r + 1) = Pr (t r )
(4.46)
(4.54)
HP (s) + GQG ,
Pr =
s 0.
A+
Ai i (t ) x(t ),
i=0
(4.55)
i =0
Theorem 4.3. Under Assumptions 4.1 and 4.2, if the Riccati equations (4.18) and (4.22) converge, then the corresponding constant gain
estimators (4.40)(4.42) are mean square stable.
Proof. It is noted that if the filter x e (t r + 1, r ) is stable, the future
finite iterations x e (t r + 1, r 1), . . . , x e (t , 1) are stable as well.
Thus to show the stability of (4.40)(4.42), one just needs to show
the stability of (4.40). Note that (4.40) can be rewritten as
x e (t r + 1, r ) =
i FPr H x e (t r , r )
(i i,t r +i )FPr H x e (t r , r ) + Kr y r (t r ),
i=0
i FPr H
i =0
i 1
i =0
i =0
i FPr RFP r .
i =0
Pr >
i FPr H Pr
i=0
i FPr H
i =0
i 1
i =0
i FPr HPr H FP r ,
i =0
1
5. An illustrative example
In this section, we present a simple numerical example to illustrate the developed theoretical results. Consider a dynamic system
described in (2.1) and (2.2) with the following parameters
1.02
=
0
0.05
,
0. 9
1
G=
,
0.5
H = 2
1 ,
where w(t ) and v(t ) are white noises with zero means and covariance matrices Q = 1 and R = 1, respectively. The initial value x0
and its covariance matrix are set to be
[ ]
1
x0 =
,
1
1
P0 =
0
0
.
1
P{r (t ) = 1} = 0.1
i=0
i FPr H Pr
P{r (t ) = 0} = 0.8,
Ai QAi < 0.
i=0
457
r
where FPr = Pr H (HPr H + R)1 . It can be seen that E [ i=0 (i
r
r
r
i =0 i ) .
Based on the theorems hypotheses, we know that Pr (s) con-
(4.56)
458
Fig. 2. The first state component x1 (t ) and its optimal filter x 1 (t | t ) based on the
given path of r (t ) as shown in Fig. 1.
Fig. 3. The second state component x2 (t ) and its optimal filter x 2 (t | t ) based on
the given path of r (t ) as shown in Fig. 1.
Fig. 4. The first state component x1 (t ) and its suboptimal filter x o1 (t | t ) based
on the given probabilities of r (t ) : P{r (t ) = 0} = 0.8, P{r (t ) = 1} = 0.1 and
P{r (t ) = 2} = 0.1.
Fig. 5. The second state component x2 (t ) and its suboptimal filter x o2 (t | t ) based
on the given probabilities of r (t ) : P{r (t ) = 0} = 0.8, P{r (t ) = 1} = 0.1 and
P{r (t ) = 2} = 0.1.
Fig. 6. The sum of the optimal and suboptimal estimation error covariances.
0,s H
.
Hr (s) = .. ,
r ,s+r H
(A.4)
(A.5)
Thus the solution to the Riccati equation exists and is unique. Note
from the property (2.6), the remaining case is that there exist one,
say {0, . . . , r } such that ,+s = 1, and the others are zero. In
this case, it is easily known that there exists an elementary matrix
1
Ar (s) with Ar (s) = A
r (s) such that
Ar (s)Hr (s) =
[ ]
Ar Rr (s)Ar (s) =
H
,
0
R
0
(A.7)
0
.
0
(A.8)
0
.
0
(A.9)
(A.10)
Let
Kr (s)Ar (s) = [K r (s)
K r (s)].
(A.11)
[K r (s)(HPr (s)H + R)
0] = [ Pr (s)H
0].
(A.12)
(A.13)
K r (s)]Ar (s),
(A.15)
References
Pr (s + 1) = Pr (s) + GQG .
(A.2)
Pr (s + 1) = Pr (s) Pr (s)H
(A.1)
(A.3)
0,s v(s)
..
vr (s) =
,
.
r ,s+r v(s)
where
459
(A.14)