Linear Feedback Control

Download as pdf or txt
Download as pdf or txt
You are on page 1of 14

Int. J. Appl. Math. Comput. Sci., 2009, Vol. 19, No. 2, 219232 DOI: 10.

2478/v10006-009-0018-2

INPUT CONSTRAINTS HANDLING IN AN MPC/FEEDBACK LINEARIZATION SCHEME


J IAMEI DENG , V ICTOR M. BECERRA , R ICHARD STOBART

Department of Aeronautical and Automotive Engineering Loughborough University, Leicestershire LE11 3TU, UK e-mail: {j.deng,r.k.stobart}@lboro.ac.uk
School of Systems Engineering University of Reading, Reading, RG6 6AY, UK e-mail: v.m.becerra@reading.ac.uk

The combination of model predictive control based on linear models (MPC) with feedback linearization (FL) has attracted interest for a number of years, giving rise to MPC+FL control schemes. An important advantage of such schemes is that feedback linearizable plants can be controlled with a linear predictive controller with a xed model. Handling input constraints within such schemes is difcult since simple bound contraints on the input become state dependent because of the nonlinear transformation introduced by feedback linearization. This paper introduces a technique for handling input constraints within a real time MPC/FL scheme, where the plant model employed is a class of dynamic neural networks. The technique is based on a simple afne transformation of the feasible area. A simulated case study is presented to illustrate the use and benets of the technique. Keywords: predictive control, feedback linearization, neural networks, nonlinear systems, constraints.

1. Introduction
The great success of predictive control is mainly due to its handling of constraints in an optimal way. There is early work on the integration of a feedback linearizing controller in an unconstrained MPC scheme (Henson and Seborg, 1993). However, in practical control schemes constraints have to be dealt with. Del-Re et al. (1993) dealt with the integration of output constraints by mapping the problem into a linear MPC problem. The basic idea of this method is to linearize the process using feedback linearization so that linear MPC solutions can be employed. However, in most of the cases, this mapping transforms the original input constraints of the process into nonlinear and state-dependent constraints, which cannot be handled by means of quadratic programming (Nevistic, 1994; Oliveiria et al., 1995). Some methods have been presented in order to deal with such nonlinear constraints mapping while using QP routines to solve an approximate linear MPC problem (Henson and Kurtz, 1994; Oliveiria et al., 1995). However, these approaches suffer from an important limitation as convergence to a feasible solution over the optimization horizon

within the available time becomes a problem (Nevistic and Morari, 1995). Ayala-Botto et al. (1999) proposed a new optimization procedure that guarantees a feasible control solution without input constraints violation over the complete optimization horizon, in a nite number of steps, while allowing only a small overall closed-loop performance degradation. Ayala-Botto et al. (1996) integrated in two complementary iterative procedures the solution of the QP optimization converting the nonlinear state-dependent constraints into linear ones. Van den Boom (1997) derived a robust MPC algorithm using feedback linearization at the cost of possibly conservative constraint handling. Nevistic and Primbs (1996) considered the problem of model/plant mismatch of MPC with feedback linearization. Guemghar et al. (2005) proposed a cascade structure of predictive control and feedback linearization for unstable systems. Scattolini and Colaneri (2007) presented a multi-layer cascaded predictive controller to guaranteed stability and feasibility. Ayala-Botto et al. (1999) presented a new solution for the problem of incorporating an input-output linearization scheme based on static neural

220 networks in a predictive controller, considering the presence of level and rate inequality constraints applied to the plant input. However, this scheme can only handle singleinput single-output systems. Moreover, Kurtz and Henson (1997) presented an input constraint mapping technology, the transformed constraints at each sampling time were decided by solving a constrained optimization problem. Also Casavola and Mosca (1996) tried to nd box constraints by employing an optimization algorithm. In this paper, focus will be positioned on input constraint handling on an integrated MPC+FL scheme based on dynamic neural networks. The aim is to nd a constraint handling method that is suitable for real time application and avoids the shortcomings of the method by Ayala-Botto et al. (1999). Solving an additional optimization problem to nd the transformed constraints is avoided in the proposed method. The paper is organised as follows: Section 2 briey describes the model predictive control formulation employed in this work. Section 3 describes the neural network structure used and discusses the training problem. Section 4 discusses the feedback linearization technique employed. Section 5 presents the proposed method for handling input constraints within an MPC/FL scheme. Section 6 discusses how the dynamic neural network is employed as a closed-loop observer. Section 7 presents a simulated case study. Concluding remarks are given in Section 8.

J. Deng et al.
ym (k|k 1) is the estimate of the plant output at time k based on information at time k 1, L is the Kalman lter gain matrix and e(k|k) is the estimated error: e(k|k) = ym (k) ym (k|k 1). The prediction model of MPC is similiar to Eqn. (2), without e(k|k). This formulation is inspired in the con strained algorithm presented in (Maciejowski, 2002). The cost function V minimised by the predictive controller penalises deviations of the predicted controlled outputs z (k + i|k) from a reference trajectory r(k + i|k), and it also penalises changes in the future manipulated inputs u(k + i|k). Dene the cost function as follows:
p

V (k) =

i=1 m1

||(k + i|k) r(k + i|k)||2 w (i) z Q (3) ||u(k + i|k)||2 w (i) , R

+
i=0

where the prediction and control horizons are p and m, respectively, Qw (i) and Rw (i) are output weight and input weight matrices, respectively. The cost function is subject to the inequality constraints: umax (k) u(k + i 1|k) umin(k), umax (k) u(k + i 1|k) umin (k), zmax (k) z(k + j|k) zmin (k), (4)

2. Model-based predictive control


The predictive control formulation employed (Maciejowski, 2002) is based on a linear, discretetime state-space model of the plant. This makes sense within the framework of this paper as the plant is assumed to be feedback linearized prior to the application of predictive control. The model has the form xm (k + 1) = Axm (k) + Bu(k), ym (k) = Cy xm (k), z(k) = Cz x(k), where xm (k) RN is the state vector at time k, u(k) Rm is the vector of inputs, ym (k) Rn is the vector of measured outputs, and z(k) R is the vector of outputs which are to be controlled to satisfy some constraints, or to particular set-points, or both. In this work, a Kalman lter was used that can be described as follows: xm (k + 1|k) = Am (k|k 1) + Bu(k) + L(k|k), x e ym (k|k 1) = Cy xm (k|k 1), z (k|k 1) = Cz xm (k|k 1), (2) where xm (k + 1|k) is the estimate of the state at future time k + 1 based on the information available at time k,

where i = 1, 2, . . . , m and j = 1, 2, . . . , p. The constrained predictive control algorithm has been implemented in the C programming language using quadratic programming (Deng and Becerra, 2004), and it has been interfaced to SIMULINK through an S-Function as it is difcult to handle constraints using the MPC toolbox available in Matlab and in a real time case.

3. Dynamic neural networks


A dynamic neural network is used as a model of the plant for control synthesis in this paper as it can easily approximate any nonlinear systems. It can be expressed as a vector differential equation: x(t) = f (x(t), u(t), ), y(t) = h(x(t), ), (5)

(1)

where x RN represents the state vector, y Rn is the output vector, u Rm is the external input vector, Rl is a parameter vector. Here f is a vector-valued function that represents the structure of the network, and h is a vector-valued function that represents the relationships between the state vector and the output vector. The structure of the dynamic neural network used in this paper is a particular case of Eqn. (5). Any nite time trajectory of a given dynamic system can be approximated

Input constraints handling in an MPC/feedback linearization scheme


1u

221

cast as a nonlinear unconstrained optimization problem:


1i

M
mu

M
Ni

() = tanh()

( xi )

Fig. 1. Illustration of a dynamic neuron.

by the internal state of the output units of a continuous time dynamic neural network with N dynamic units, m inputs and n outputs (Garces et al., 2003). These networks are neural unit arrays which have connections both ways between a pair of units, and from a unit to itself. The model is dened by a one-dimensional array of N neurons or units, in which each neuron of the dynamic neural network can be described as follows:
N m

xi = i xi +
j=1

ij (xj ) +
j=1

ij uj ,

where i , ij and ij are adjustable weights, with 1/i as a positive time constant and n N , xi the activation state of Unit i, and u1 , . . . , um the input signals (see Fig. 1). The states of the rst n neurons are taken as the output of the network, leaving N n units as hidden neurons. The network is dened by the vectorised expression of (6) as follows: x = x + (x) + u, y = Cx, (7) (8)

where x are the coordinates on RN , RN N , RN m , u Rm , (x) = [(x1 ), . . . , (xN )]T , () is a sigmoidal function, such as the hyperbolic tangent, Inn is the n n identity matrix, 0n(N n) is an n (N n) matrix of zeros. C = [Inn 0n(N n) ], and = diag(1 , . . . , N ) is a diagonal matrix. This paper used a dynamic neural network described by Eqns. (7) (8). The state vector x of the dynamic neural network of Eqn. (7) can be partitioned into the output state xo and the hidden states xh : x= xo xh . (9)

A dynamic neural network training problem can be

ix

iv

mi

1i

N x(

1 x(

s'norueN

stuptuo

rotcev

tupnI

min

FM (, ZM ) =

1 2M

||y(tk ) y (tk |)||2 ,


k=1

1 s +

(6)

(10) where ZM = [y(tk ), u(tk )]k=1,M is a training data set, y(tk ) represents the measured output, y(tk |) is the dy namic neural network output, and is a parameter vector. The optimization problem associated with training usually exhibits local minima. Hence, training dynamic neural networks is typically performed using unconstrained local optimization with multiple random starts, global optimization methods, or hybrid methods. Global optimization or hybrid methods are usually better choices for training DNNs when dealing with multivariable plants (Garces et al., 2003). In this work dynamic networks are trained using a hybrid method involving the DIRECT algorithm (Perttunen et al., 1993), which is a deterministic global optimization method, and a gradient based local optimization method. The solutions obtained by DIRECT are improved using gradient based local optimization. The training algorithm will nd the parameters of the network in Eqn. (7) for which the error function FM is minimized. The weight vector used by the algorithm is the aggregate of the neurons feedback weight matrix N N , the input weight matrix N m , the state feedback weight vector N 1 = [1 , . . . , N ]T and the initial values of the hidden states of the dynamic neural network, xh (t0 ). That is, N 1 vec(N m ) = (11) vec(N N ) , xh (t0 )(N n)1 where vec(.) transfers a matrix into a vector.

4. Approximate input-output feedback linearization and MPC integration


Many nonlinear control methods are based on state space models where the time derivative of the states depends nonlinearly on the states and linearly on the control inputs (Isidori, 1995), which are known as control afne systems and are described as follows: x = f (x) + g(x)u, y = h(x), (12)

where x RN is a state vector, u Rm is a vector of manipulated inputs, f and g are differentiable vector elds, and y Rn is a vector of outputs variables. The purpose of input-output linearization is to introduce a new input variable v and a nonlinear transformation that uses state feedback to compute the original input u, so

222 that a system described by Eqn. (12) behaves linearly from the new input v to the output y. It is of considerable importance to assess which systems can be input-output linearized. The necessary and sufcient conditions for the existence of a feedback law are presented in (Isidori, 1995). Input-output linearization and decoupling is a particular case of input-output linearization. An appropriate selection of the design parameters leads to a feedback linearized system where the i-th output depends on the ith external input only. The basic linearizing-decoupling technique is described in (Isidori, 1995). In this paper, we apply the linearizing-decoupling technique to the dynamic neural network model, which is a control afne system, as formulated by Garces et al. (2002). The control afne system mappings related to Eqn. (12) can be related to the dynamic neural network parameters as follows: f (x) = x + (x), g(x) = , h(x) = Cx. (13)

J. Deng et al.
where yi , not yi , is used as it is based on a neural network. If the relative vector is well dened, S(x) is invertible and det diag 1r1 , 2r2 , . . . , nrn = 0. (18)

Consider the dynamic neural network described by Eqns. (7) and (8), and suppose that the dynamic neural network has a vector relative degree given by r = [r1 , r2 , . . . , rn ] at a point x0 . Assume that the dynamic neural network has the same number of inputs and outputs (n = m). For arbitrary values ik (i = 1, . . . , n and k = 0, . . . , ri ), a state feedback law with u = S(x)1 E(x) + S(x)1 v, where S(x) 1r1 Lg1 Lr1 1 x1 f . . = . nr Lg Lrn 1 xn
n 1

It should be noted that the approximate feedback linearization and decoupling laws described above require state information. Since the model employed in this work to generate the linearizing-decoupling laws is a dynamic neural network, the same model can also be used as a closed loop observer to provide state information to the feedback linearizing laws. Once the feedback linearization is tested and validated on the actual plant, the constrained predictive controller can be used to control the linearized system. This controller can deal with disturbances arising from modelling errors and other sources, and it can handle constraints on plant variables. While handling output constraints using the method presented in this paper is straightforward, input constraints require special treatment due to the presence of the nonlinear transformations associated with feedback linearization. This aspect will be discussed in detail in Section 5. The scheme is illustrated in Fig. 2.

5. Algorithm for handling input constraints


Output constraints are straightforward for feedback linearization. Therefore, only input constraints are discussed in this paper. The feedback linearization-decoupling law (14) can be written as follows in discrete time based on the sampling time Ts , which is chosen using the NyquistShannon theorem: u(k + 1|k) = S(x(k + 1|k)) (v(k + 1|k) E(x(k + 1|k))),
1

(14)

.. .

1r1 Lgn Lr1 1 x1 f . . . nrn Lgn Lrn 1 xn nn f (15) , (16)

(19)

and

E(x) =

r1 k=0 rn k=0

1k Lk x1 f . . . nk Lk xn f

where i = 1, 2, . . . , m, S(x(k + 1|k)) and E(x(k + 1|k)) are given by Eqns. (15) and (16), and x(k + 1|k) is a sampled state estimate obtained from an observer as as described in Section 6. Consider the following constraints on the input uk : umin (k) uk umax (k), umin (k) uk umax (k).

n1

where Lk h denotes a Lie derivative of order k of a scalar f function h(x) along vector eld f , ik s are scalar design parameters, and ri is the relative degree of the i-th output yi , produced when applied to a dynamic neural network described by Eqns. (7) and (8) a linearized-decoupled system that obeys
ri k=0

(20)

d yi ik k = vi , dt

i = 1 . . . n,

(17)

From Eqns. (19) and (20), it is possible to see that the resulting constraints on v(k + i 1|k) are state dependent and hence are not suitable to be enforced by means of an MPC algorithm such as the one presented in Section 2, which assumes constant bounds. Although MPC based on quadratic programming could solve linear-variant, even nonlinear-variant constraints, it is not suitable for a realtime control purpose. This paper proposes the use of an

Input constraints handling in an MPC/feedback linearization scheme

223

DNN x

reference

Predictive Controller

Feedback linearizationdecoupling

Nonlinear process

Feedback Linearized Process

Fig. 2. Hybrid MPC/FL control scheme based on a dynamic neural network.

afne transformation to solve this problem for a real-time application. In order to describe the technique in a simple way, this paper considers a two-input two-output system. However, this technique can be used in cases with more inputs and outputs, simply by increasing the dimensions. Assume that u = [u1 (k + i 1|k) u2 (k + i 1|k)]T . Then, at sampling time k, u1 min u1 (k + i 1|k) u1 max , u2min u2 (k + i 1|k) u2max , u1 min u1 (k + i 1|k) u1 max , u2min u2 (k + i 1|k) u2max . (21)

d1 d2

= S(x(k + 1|k))

u1 max u2 min

+ E(x(k + 1|k)). The feasible area in Fig. 4 of the transformed inputs is not suitable for use with the predictive control algorithm described in Section 2, as the box is oblique with respect to the co-ordinate system xc yc . From Eqn. (22), it is not difcult to see that the side OT is parallel with the side RS; the side OR is parallel with the side T S. Therefore, the feasible area of the transformed inputs can be expressed in a suitable form for the predictive control algorithm described in Section 2, after an afne change of coordinates, which involves in the 2D case a rotation and a translation. Figure 5 shows the new co-ordinates. Axis xc is parallel with sides OT and RS. Axis yc is parallel with the sides OR and T S. In the new coordinates, the points O, R, S, T map into (a1 , a2 ), (b1 , b2 ), (c1 , c2 ), (d1 , d2 ). It is not difcult to prove that a1 = b1 , a2 = d2 , b2 = c2 and c1 = d1 . Suppose the afne change of variables is done by means of a 2 2 afne matrix H and a translation. Matrix H can be obtained using the following equations: a1 a2 b1 b2 c1 c2 d1 d2 =H =H =H =H a1 a2 b1 b2 c1 c2 d1 d2 E(x(k + 1|k)) , E(x(k + 1|k)) , E(x(k + 1|k)) , E(x(k + 1|k)) . (23)

Figure 3 illustrates the feasible area of u1 (k + i 1|k) and u2 (k + i 1|k) at sampling time k. It has four corners labelled by O, R, S, T . The coordinates of the corners are (u1 min , u2 min ), (u1 min , u2 max ), (u1 max , u2 max ), (u1 max , u2 min ), respectively. Note that since these are bound constraints, the sides of the box are parallel to the co-ordinates axes xc and yc . Figure 4 illustrates the feasible area of v1 and v2 after each point in the feasible area of u1 (k+i1|k) and u2 (k+i1|k) is multiplied by the matrix S(x(k + 1|k)) and translated by E(x(k + 1|k)). The coordinates of these four corners O, R, S, T have changed into (a1 , a2 ), (b1 , b2 ), (c1 , c2 ), (d1 , d2 ). The new coordinates have the following relationship with the old ones: a1 a2 b1 b2 c1 c2 = S(x(k + 1|k)) u1 min u2 min u1 min u2 max u1 max u2 max

+ E(x(k + 1|k)), = S(x(k + 1|k)) (22)

+ E(x(k + 1|k)), = S(x(k + 1|k))

+ E(x(k + 1|k)),

Notice that by choosing to have the origin of the xc yc co-ordinate system centered in the box, the effect of the shift E(x(k + 1|k)) from Eqn. (22) is cancelled by Eqn. (23). Hence it is safe to ignore the effect of the shift E(x(k + 1|k)) when computing the new bounds for

224

J. Deng et al.

Fig. 3. Illustration of the feasible area of the manipulated inputs: Feedback linearization incorporated in a predictive control scheme after the afne transformation.

Fig. 4. Illustration of the feasible area of the transformed inputs.

Fig. 5. New coordinates after the afne transformation.

v(k + i 1|k) in the xc yc co-ordinates. The following relationships are also true: b 2 a2 = c1 b 1 = c2 d2 = d1 a1 = (b1 a1 )2 + (b2 a2 )2 , (b1 c1 )2 + (b2 c2 )2 , (d1 c1 )2 + (d2 c2 )2 , (d1 a1 )2 + (d2 a2 )2 . (24)

Eqn. (24), H can be easily obtained from Eqn. (23). Thus any vector z in the xc yc co-ordinates can be transformed into a vector zc in the xc yc co-ordinates as follows: (25) z = H(z E(x(k + 1|k))), while any vector z in the xc yc co-ordinates can be transformed into a vector in the xc yc co-ordinates as z = H 1 z + E(x(k + 1|k)). (26)

After a1 , a2 , b1 , b2 , c1 , c2 , d1 , and d2 are obtained from

Input constraints handling in an MPC/feedback linearization scheme


After the afne transformation, for a given time k, v(k + i 1|k) will be bounded as follows: a1 a2 v(k + i 1|k) c1 c2 . (27) v max = Hvmax (k + i 1|k) + Hu(k 1|k 2)(S(x(k + 1|k)) S(x(k 1|k 2)))

225

(34)

Therefore, we could change the rst line of Eqn. (4) into Eqn. (27) as this is the real input constraint for MPC. For a given k, v(k 1|k 2) = S(x(k 1|k 2))u(k 1|k 2) + E(x(k 1|k 2)), v(k|k 1) = S(x(k|k 1))uk + E(x(k|k 1)). (28) Hence, v(k + i 1|k) = v(k + i 1|k) v(k 1|k 2) = S(x(k + 1|k))uk S(u(k 1|k 2))u(k 1|k 2) + E(x(k + 1|k)) E(x(k 1|k 2)). v(k + i 1|k) is obtained by transforming v(k + i 1|k) into the new coordinates xc and yc , i.e., v(k + i 1|k) = Hv(k + i 1|k). Combining Eqns. (29) and (30), we get v(k + i 1|k) = HS(x(k + 1|k))u(k + 1|k) + Hu(k 1|k 2)(S(x(k + 1|k)) S(x(k 1|k 2))) + H(E(x(k + 1|k)) E(x(k 1|k 2)). Equation (31) also can be writen as v(k + i 1|k) = Hv(k + i 1|k) + Hu(k 1|k 2)(S(x(k + 1|k)) S(x(k 1|k 2))) + H(E(x(k + 1|k)) E(x(k 1|k 2))). The following two equations will be obtained by using constraints on Eqn. (31): v min = Hvmin (k + i 1|k) + Hu(k 1|k 2)(S(x(k + 1|k)) S(x(k 1|k 2))) + H(E(x(k + 1|k)) E(x(k 1|k 2))). (33) (32) (30) (29)

+ H(E(x(k + 1|k)) E(x(k 1|k 2))). Therefore, we could change the second line of Eqn. (4) into Eqns. (33) and (34) as this is the real rate constraint for MPC. Now that the constraints have been transformed into a new coordinate frame. The next step is to transform the inputs to the predictive controller. Figure 6 illustrates the control scheme, where the forward transformation and its inverse are denoted by H blocks and H 1 blocks, respectively. The use of the afne transformation to dene the bounds for v(k + i 1|k) does not affect the transfer function of the plant seen by the predictive controller, provided the output and the reference signals undergo the same transformation. This can be proved as follows. For the scheme in Fig. 6, the matrix transfer function can be written as follows: y(s) v(s) (35) 1 0 n = c1n s ++c11 s+c10 , 1 0 c2n sn ++c21 s+c20 Here, c1n , . . . , c10 and c2n , . . . , c20 are the arbitrary values chosen when applying feedback linearizationdecoupling. From Fig. 6, y1 (s) can be obtained by y1 (s) = H G(s) H 1 v1 (s), Let G1 (s) = H G(s) H 1 , (36) (37)

(31)

where G1 is the new transfer function for the predictive controller, which can easily be transformed into a state space form. Table 1 gives a step-by-step summary of the method for handling input constraints in an MPC/FL. Consider now the multi-input multi-output case and suppose the number of inputs and outputs is m. The number of inequalities in Eqn. (21) will be 2m. The dimension of Fig. 3 should be m and the corner number of the feasible area will be 2m . The dimension of Fig. 4 is also m and the corner number of new coordinates is 2m . The number of equations in Eqn. (22) will become 2m , so will the number of equations in Eqn. (23). H can still be obtained using equations which are similar to (22) and (23). Therefore, the multi-input and multi-output case can easily be solved.

6. Using the dynamic neural network model as a closed-loop observer


Poznyak et al. (2001) proposed the use of dynamic neural networks as closed loop observers. In this paper, a

226

J. Deng et al.

v reference H Predictive v1 Controller y1 H Feedback Linearized Process H-1 Feedback linearizationdecoupling u Nonlinear process y

Fig. 6. Feedback linearization incorporated in a predictive control scheme after the afne transformation.

Table 1. Method for handling input constraints in an MPC/FL scheme.

Step

Description

Step 1 Set k = 0. Obtain a1 , a2 , b1 , b2 , c1 , c2 , d1 , and d2 according to Eqn. (24) given that a1 = b1 , a2 = d2 , b2 = c2 and c1 = d1 . Step 2 Obtain H is from Eqn. (23). Step 3 Dene the bounds for v(k + i 1|k) as given in Eqn. (27). Step 5 Calculate vmin and vmax according to Eqns. (33) and (34). Step 6 Measure the current plant output yk and transform it using Eqn. (25) to obtain yk . Step 7 Transform the current reference rk using Eqn. (25) to obtain rk . Step 8 Execute one step of the predictive control algorithm to obtain v(k + i 1|k) . Step 8 Transform v(k + i 1|k) to obtain v(k + i 1|k) using Eqn. (26). Step 9 Provide v(k + i 1|k) to the feedback linearization/decoupling algorithm. Step 10 Set k = k + 1 and return to Step 1.

this work as a state estimator, where a dynamic neural network is employed to represent the plant dynamics. We used a simplied extended Kalman lter which assumes the following stochastic model of the plant dynamics and discrete observations with a sampling interval Ts : x(t) = F (x, u) + Gw(t) = x(t) + (x(t)) + u(t) + Gw(t), x(0) = x0 , y(tk ) = Cx(tk ) + v(tk ),

(38)

where x is the state vector, y is the measurement vector, u is the external input vector, w is assumed to be zero mean continuous white noise with covariance matrix Q, v is assumed to be zero mean discrete white noise with covariance matrix R, and x0 is a random vector with zero mean and covariance matrix P0 . The algorithm assumes that the Jacobian matrix remains constant between samples, so that (k) = exp F (x, u)) x Ts .
tk

(39)

A summary of the algorithm is as follows: dynamic neural network is employed as a closed-loop observer to provide state information to be used by the feedback linearization laws. A dynamic neural network can be used as either an open-loop observer or a closedloop observer. A dynamic neural network was used as an open-loop observer in (Garces et al., 2003; Garces, 2000), where the state vector of the dynamic neural network was used to provide state information to the feedback linearization law, with no plant output information used to correct the state estimates. A closed-loop observer with output feedback can reduce the state estimation error and improve the robustness of the estimator. An extended Kalman lter algorithm (Maybeck, 1982) is employed in Algorithm initialisation: Give values to Q, R, P0 and G. Set t = 0, k = 0, P (0) = P0 , x(0) = x0 . Step 1: Time update. Integrate from tk1 to tk the following differential equation to obtain x(k) : x = x(t) + (x(t)) + u(t). Also compute the a priori state covariance from P (k) = (k)P (k 1)(k) + Q.
T

(40)

(41)

Step 2: Measurement update. Measure the current output y(k) and then compute the Kalman lter gain

Input constraints handling in an MPC/feedback linearization scheme

227

Fig. 7. Schematic diagram of the two tank mixing process.

K(k), the corrected state estimate x(k), and the cor rected covariance matrix P (k): K(k) = P (k)C (CP (k)C + R)
T T 1

algebraic and differential equations: Qc = Cc vc /100, Qh = Ch vh /100, (45) Qo1 = K1 h1 h2 , dh1 = (Qc + Qh Qo1 )/A1 , dt dT1 = [Qc (Tc T1 ) + Qh (Th T1 )]/(A1 h1 ), dt where vc is the opening of the cold water control valve (%), vh is the opening of the hot water control valve (%), Cc is the constant of the cold water control valve (cm3 /s%), Ch is the constant of the cold water control valve (cm3 /s)%), Qc is the cold water ow rate (cm3 /s), Qh is the hot water ow rate (cm3 /s), Qo1 is the outlet ow rate from Tank 1 (cm3 /s), h1 is the liquid level in Tank 1 (cm), T1 is the liquid temperature in Tank 1 (o C), Tc is the temperature of the cold water stream (o C), Th is the temperature of the hot water stream (o C), K1 is the restriction of the interconnection valve and pipe (cm5/2 /s), A1 is the cross-sectional area of Tank 1 (cm2 ). The model for the second tank is given by the following algebraic and differential equations: Qo2 = K2 h2 , dh2 = (Q1 Qo2 )/A2 , dt dT2 = [Qo1 (T1 T2 )]/(A2 h2 ), dt

(42) (43) (44)

x(k) = x (k) + K(k)(y(k) y (k)), P (k) = (I K(k)C)P (k). Set k = k + 1 and return to Step 1.

7. Case study
This section describes a case study based on a simulated mixing process. A section of the process is shown in Fig. 7. Two streams of cold and hot water enter the rst tank, where a motorized mixer operates. These streams are controlled by means of two pneumatic valves. The two tanks are interconnected via a lower pipe with a manual valve that is normally open. The outlet valve from Tank 2 is also normally open. The two measured variables are the level of Tank 1 and the temperature of Tank 2. A simple modular model of the mixing process is described below (Becerra et al., 2001). The model is obtained from mass and energy balances. The main simplifying assumptions that were made to derive this model were as follows: The mixing is perfect. Heat losses are negligible. The valves have linear characteristics. The temperatures of the cold and hot water streams are constant. The model for the rst tank is given by the following

where Qo2 is the outlet ow rate from Tank 2 (cm3 /s), h2 is the liquid level in Tank 2 (cm), T2 is the liquid temperature in Tank 2 (o C), K2 is the restriction of the Tank 2 outlet valve and pipe (cm5/2 /s), A2 is the cross-sectional area of Tank 2 (cm2 ). The following values were used for the model parameters: A1 = 289 cm2 , A2 = 144 cm2 , Cc = 280.0 cm3 /s%, C h = 100.0 cm3 /s%, T = 20o C, Th = 72o C, K1 = 30 cm5/2 /s, K2 = 30 cm5/2 /s. Qh

2oQ

2K

2knaT

1oQ

2T

2h

1K

hT,hQ 1knaT

cT,cQ

IT Ih

(46)

228
1c retnec htiw ebuc-repyh tinu eht eb ot niamod eht ezilamroN

J. Deng et al.
ters. At the same time, values obtained from the DIRECT algorithm are close to different local minima. A combined algorithm starts with the DIRECT algorithm and then netunes the result using a faster local search procedure. This is illustrated in Fig. 8. Figure 9 compares the network output obtained for the training input with the training output using the DIRECT algorithm combined with the gradient descent algorithm. Figure 10 compares the network output obtained for the validation input with the validation output using the DIRECT algorithm combined with the gradient descent algorithm. After identifying the dynamic neural network model, a feedback linearization based on this DNN has being calculated. In order to improve the feedback linearization, the dynamic neural network is also used as a closed loop observer. The two-tank system yields a 2unit network: h(x) = x = x1 x2 , 12 22 , (51)

Fig. 8. Descriptive diagram of a combined DIRECT gradient based training algorithm.

and Qc are the manipulated inputs. An identication experiment was carried out on this two-tank system model by using random steps in the inputs u1 (k + i 1|k) and u2 (k + i 1|k) over 4000 sampling points. The rst 2000 sampling points data were used for training and the second 2000 sampling points data were used for validation. The sampling time is 20 s. Training was carried out by means of the DIRECT algorithm combined with the gradient descent algorithm. Fifty training runs were carried for each network size, starting from the random initial weights, and the best network in terms of mean square error for the training data was selected for each network size. The values of the parameters of the best model found were = = = 0.0059 0 0 0.0106 0.0084 0.0030 0.0034 0.0038 0.0891 0.0783 0.1488 0.1850 , , , (47) (48) (49) (50)

C = [1 0].

In order to explore a reduced region of the error surface, some methods, such as Quasi-Newton ones, result in nding a local minimum from the given initial parame-

tluser gninut-enif eht rof mhtirogla tnecsed tneidarG

seY

dnE

elgnatcer lamitpo yllaitnetop fo E tes eht yfitnedI


nimf etadpU iE elgnatcer eht fo retnec eht elpmaS iE elgnatcer

)1c(f= nimf tes ,)1c(f dniF seY 0=iE-E tegrat hcaeR a esoohC oN oN

g(x) = =

11 21

f (x) = x + (x). It has the relative degree of r = [1 1]. The control law is given by u(t) = S(x(k + 1|k))
1 1

(52)

E(x(k + 1|k)) v(t),

+ S(x(k + 1|k)) where S(x(k + 1|k)) = =

(53)

11 Lg1 L0 x1,k f 21 Lg1 L0 x2,k f 11 11 21 21 11 12 21 22

11 Lg2 L0 x1,k f 21 Lg2 L0 x2,k f , (54)

E(x(k + 1|k)) =
=

10 L0 x1,k + 11 Lf x1,k f 20 L0 x2,k + 21 Lf x2,k f


10 x(k + 1|k) + 11 (1 x(k + 1|k) 20 x(k + 1|k) + 21 (1 x(k + 1|k)(x(k + 1|k))) +w11 (x(k + 1|k)) + w12 (x(k + 1|k))) +w21 (x(k + 1|k)) + w22 (x(k + 1|k))) , (55)

where 10 = 0.000062861, 11 = 0.0165, 20 = 0.000062861, 21 = 0.0165. The design parameters ij are chosen so that 10 = 20 and 11 = 21 , and the static gain and dominant time constant of the linearized system v y is as near to that of the Jacobian linearization of the

Input constraints handling in an MPC/feedback linearization scheme

229

Temperature

40 20 0 0 40 500 1000
Sampling Points

DNN Tank

1500

2000 DNN Tank

Level

30 20 0

500

1000
Sampling Points

1500

2000

Fig. 9. Comparison of training trajectories using the DIRECT algorithm combined with the gradient descent algorithm.

50
Temperature

40 30 20 10 0 40 35 500 1000
Sampling Points

DNN Tank

1500

2000

DNN Tank

Level

30 25 0

500

1000
Sampling Points

1500

2000

Fig. 10. Comparison of validation trajectories using the DIRECT algorithm combined with the gradient descent algorithm.

dynamic neural network model as possible. For this case, S(x(k + 1|k)) is a constant matrix: S(x(k + 1|k)) = 0.0001394 0.0000485 0.0000565 0.0000628 . (56)

S(x(k + 1|k)), the coordinates of the four corners are (a1 , a2 ) = (0, 0), (b1 , b2 ) = (0.004983, 0.0062799), (c1 , c2 ) = (0.01889, 0.00063129), (d1 , d2 ) = (0.0139387, 0.0056489). (58)

The input constraints are given as 0 0 u 100 100 . (57)

Therefore, (a1 , a2 ) = (0, 0),

After the constraint area is transformed using

230 (b1 , b2 ) = (0, 0.0079984), (c1 , c2 ) = (0.01504, 0.0079984), (d1 , d2 ) = (0.01504, 0). According to Eqn. (23), H= 0.81764 0.64492 0.391138 0.96513 . (60) (59)

J. Deng et al.
feasiblity of MPC is important (Rossiter, 2003) in the context of the work presented in this paper. This is currently being investigated by the authors and will be the subject of a future paper.

Acknowledgment
This project is co-funded by the UK Technology Strategy Boards Collaborative Research and Development programme, following an open competition. The authors would like to thank the Technology Strategy Board Program under Grant Reference No. TP/3/DSM/6/I/152. Thanks also are given to the UK Engineering and Physical Science Research Council under Grant Reference No. GR/R64193 and the University of Reading Chinese Studentship Programme for their support.

In this example, S(x(k + 1|k)) and H are constant matrices. However, if S(x(k+1|k)) and H are state-dependent, the proposed method for handling input constraints within an MPC/FL scheme remains applicable, only that the constraint values change at every sampling time. The Kalman lter gain was computed using the linearized model and the covarience matrices Qf = diag(1, 1), Rf = diag(0.001, 0.001). The references used in the simulation were two square waves with different frequencies. The sampling time is Ts = 20 s, the control horizon is m = 2, the prediction horizon is p = 320, the input and output weights are uwt = [1, 1]T , ywt = [0.0001, 0.0001]T , the manipulated variables constraints are umin = [0, 0]T , umax = [100, 100]T , the incremental constraints on the manipulated variable are umax = [0.01, 0.01]T . The tuning parameters of the extended Kalman lter are P0 = diag(0.1, 0.1), Q = diag(0.1, 0.1), R = diag(100, 100). The diagonal elements of the output noise covariance matrix R were chosen to account for the covariance of the measurement noise. The diagonal elements of the process noise covariance matrix Q were chosen small and positive in order to prevent the covariance of the state error from becoming zero. The initial covariance of the state error P0 is chosen to reect some uncertainty in the initial state estimate. Figure 11 shows the outputs responses under the MPC/FL scheme. Figure 12 shows the actual inputs to the process under the MPC/FL control method. In this case, the input magnitude constraints do not become active. Figures 13 and 14 show the output of the controller and input to the process under the proposed scheme in a scenario where one of the input magnitude constraints becomes active as can be seen in Fig. 13.

References
Ayala-Botto, M., Boom, T. V. D., Krijgsman, A. and da Costa, J. S. (1999). Predictive control based on neural network models with I/O feedback linearization, International Journal of Control 72(17): 15381554. Ayala-Botto, M., Braake, H. T., da Costa, J. S. and Verbruggen, H. (1996). Constrained nonlinear predictive control based on input-output linearization using a neural network, Proceedings of the 13-th IFAC World Congress, San Francisco, CA, USA, pp. 175180. Becerra, V. M., Roberts, P. D. and Grifths, G. W. (2001). Applying the extended Kalman lter to systems desribed by nonlinear differential-algebraic equations, Control Engineering Practice 9(3): 267281. Casavola, A. and Mosca, E. (1996). Reference governor for constrained uncertain linear systems subject to bounded input disturbances, Preceedings of the 35-th Conference on Decision and Control, Kobe, Japan, pp. 35313536. Del-Re, L., Chapuis, J. and Nevistic, V. (1993). Stability of neural net based model predivtive control, Proceedings of the 32-nd Conference on Decision and Control, San Antonio, TX, USA, pp. 29842989. Deng, J. and Becerra, V. M. (2004). Real-time constrained predictive control of a 3d crane system, Proceedings of the 2004 IEEE Conference on Robotics, Automation and Mechatronics, Singapore, pp. 583587. Garces, F. (2000). Dynamic Neural Networks for Approximate Input-Output Linearisation-Decoupling of Dynamic Systems, Ph.D. thesis, University of Reading. Garces, F., Becerra, V., Kambhampati, C. and Warwick, K. (2003). Strategies for Feedback Linearisation: A Dynamic Neural Network Approach, Springer, London. Guemghar, K., Srinivasan, B., Mullhaupt, P. and Bonvin, D. (2005). Analysis of cascade structure with predictive control and feedback linearisation, IEE Proceedings: Control Theory and Applications 152(3): 317324. Henson, M. A. and Kurtz, M. J. (1994). Input-output linearisation of constrained nonlinear processes, Nonlinear Control, AICHE Annual Meeting, San Franciso, CA, USA, pp. 120.

8. Conclusion
This paper has introduced a technique for handling input constraints within a real time MPC/FL scheme, where the plant model employed is a class of dynamic neural network. Handling input constraints within such schemes is difcult since simple bound contraints on the input become state dependent because of the nonlinear transformation introduced by feedback linearization. The technique is based on a simple afne transformation of the feasible area. A simulated case study was presented to illustrate the use of the technique. The issue of recursive

Input constraints handling in an MPC/feedback linearization scheme


Temperature(Celsius) Level(cm)

231

40 30 20 0 50 40 30 0 200 400 600 200 400 600

Level Reference

800 1000 1200 1400 1600 1800 2000 Sampling Points Temperature Reference 800 1000 1200 1400 1600 1800 2000 Sampling Points

Fig. 11. Mixing process outputs under the MPC/FL method. The reference signals are square waves.
50
Qc(cm /s)
3

Q (cm /h)

0 0 100 50 0 0

500

1000 Sampling Points

1500

2000

500

1000 Sampling Points

1500

2000

Fig. 12. Inputs applied to the mixing process the MPC/FL method.
35

Level (cm)

30 25 20 15 0 500 Level Reference

1000

1500

2000

Temperature (Celsius)

60 50 40 30 0 Temperature Reference 500 1000 Sampling Points 1500 2000

Fig. 13. Mixing process outputs under the MPC/FL control method.

Henson, M. A. and Seborg, D. E. (1993). Theoretical analysis of unconstrained nonlinear model predictive control, International Journal of Control 58(5): 10531080. Isidori, A. (1995). Nonlinear Control Systems, 2nd Edition, Springer, Berlin/New York, NY. Kurtz, M. and Henson, M. (1997). Input-output linearizing control of constrained nonlinear processes, Journal of Process Control 7(1): 317. Maciejowski, J. M. (2002). Predictive Control with Constraints, Prentice Hall, London.

Maybeck, P. S. (1982). Stochastic Models, Estimation and Control, Academic Press, New York, NY. Nevistic, V. (1994). Feasible Suboptimal Model Predictive Control for Linear Plants with State Dependent Constraints, Postdiploma thesis, Swiss Federal Institute of Technology, Automatica Control Laboratory, ETH, Zurich. Nevistic, V. and Morari, M. (1995). Constrained control of feedback-linearizable systems, Proceedings of the European Control Conference, Rome, Italy, pp. 17261731. Nevistic, V. and Primbs, J. A. (1996). Model predictive con-

232
40

J. Deng et al.

Qc(cm /s)

20 0

20 100

200

400

600

800

1000

1200

1400

1600

1800

2000

Sampling Points

Qh(cm /s)

50 0

50

200

400

600

800

1000

1200

1400

1600

1800

2000

Sampling Points

Fig. 14. Inputs applied to the mixing process under the MPC/FL control method.

trol: Breaking through constraints, Proceedings of the 35-th Conference on Decision and Control, Kobe, Japan, pp. 39323937. Oliveiria, S. D., Nevistic, V. and Morari, M. (1995). Control of nonlinear systems subject to input constraints, Preprints of the IFAC Symposium on Nonlinear Control Systems, NOLCOS95, Tahoe City, CA, USA, Vol. 1, pp. 1520. Perttunen, C. D., Jones, D. R. and Stuckman, B. E. (1993). Lipschitzian optimization without the Lipschitz constant, Journal of Optimization Theory and Application 79(1): 157 181. Poznyak, A. S., Sanchez, E. N. and Yu, W. (2001). Differential Neural Networks for Robust Nonlinear Control, World Scientic, Singapore. Rossiter, J. A. (2003). Model Based Predictive Control: A Practical Approach, CRC Press, Boca Raton, FL, USA. Scattolini, R. and Colaneri, P. (2007). Hierarchical model predictive control, Proceedings of the 46-th IEEE Conference on Decision and Control, New Orleans, LA, USA, pp. 48034808. van den Boom, T. (1997). Robust nonlinear predictive control using feedback linearization and linear matrix inequalities, Proceedings of the American Control Conference, Albuquerque, NM, USA, pp. 30683072.

Victor M. Becerra obtained his Ph.D. degree in control engineering from City University, London, in 1994, for the development of novel optimal control algorithms, a B.Eng. degree in electrical engineering (Cum Laude) from Simon Bolivar University, Caracas, Venezuela, in 1990, and an M.Sc. degree in nancial management from Middesex University, London, in 2001. He worked in power systems analysis for CVG Edelca, Caracas, between 1989 and 1991. He was a research fellow at City University between 1994 and 1999, before joining the University of Reading, UK, as a lecturer in 2000, where he is currently a reader in cybernetics. His main research areas include optimal control, nonlinear control, intelligent control, and system identication. Dr. Becerra has published over 100 papers in international journals and conferences, and he has co-authored a Springer book on neural network based control. Since 2003 he has been the head of the Cybernetics Intelligence Research Group at the University of Reading. He is a member of the Technical Committees on Optimal Control as well as Adaptive Control and Learning of the International Federation of Automatic Control (IFAC). He is a member of the IASTED technical committee on robotics. He is also a Senior Member of the IEEE, a member of the IET, and a Chartered Engineer.

Jiamei Deng received the Ph.D. degree in cybernetics from the University of Reading in 2005, the B.Eng. degree in electrical engineering from the Hua Zhong University of Science and Technology, Wuhan, China, in 1988, and the M.Eng. degree in mechanical engineering from the Shanghai Institute of Mechanical Engineering, China, in 1991. She currently works as a research associate at Loughborough University. She was a research fellow at the University of Sussex between 2006 and 2007. Her research interests include predictive control, nonlinear control, sliding mode control, optimal control, system identication, optimization, process control, power control, black-box modeling, automotive control, and energy systems.

Richard Stobart is a Ford professor of automotive engineering at Loughborough University in the Department of Aeronautical and Automotive Engineering. As a professor, he is responsible for the development of research and education in the eld of vehicle engineering, with a particular focus on propulsion and power generation systems. In 2007 he completed his term as the chair of the Powertrain, Fuels and Lubricants Activity of the US based Society of Automotive Engineers (SAE) and was a programme co-chair (with Prof. Li of Tongji University) for the 2008 SAE Congress on Powertrain Fuels and Lubricants in Shanghai in 2008. He was a professor of automotive engineering at the University of Sussex (Brighton, UK) from 2001 to 2007, where he was also the head of Engineering and Design from 2003 to 2006. Professor Stobart was a member of the team who in 1997 received Arthur D Littles Ketteringham prize for their work in developing fuel cell technology. He is a graduate from the University of Cambridge with a rst class honours degree in mechanical engineering. He was elected a Fellow of the Institution of Mechanical Engineers in 2000.

Received: 12 May 2008 Revised: 11 November 2008 Re-revised: 18 November 2008

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy