Optimal Control Of Nonlinear Differential Algebraic ... - CiteSeerX

0 downloads 0 Views 94KB Size Report
such as DASSL, for their solution as described in the text books by Brenan ... Consider the following continuous optimal control problem. (OCP), defined over the ...
OPTIMAL CONTROL OF NONLINEAR DIFFERENTIAL ALGEBRAIC EQUATION SYSTEMS P D Roberts1 and V M Becerra2 1

Control Engineering Research Centre City University London EC1V 0HB, United Kingdom

Department of Cybernetics University of Reading Reading RG6 2AY, United Kingdom

Email: [email protected]

Email: [email protected]

Abstract A novel iterative procedure is described for solving nonlinear optimal control problems subject to differential algebraic equations. The procedure iterates on an integrated modified linear quadratic model based problem with parameter updating in such a manner that the correct solution of the original non-linear problem is achieved. The resulting algorithm has a particular advantage in that the solution is achieved without the need to solve the differential algebraic equations . Convergence aspects are discussed and a simulation example is described which illustrates the performance of the technique.

1. Introduction When modelling industrial processes often the resulting equations consist of coupled differential and algebraic equations (DAEs). In many situations these equations are nonlinear and cannot readily be directly reduced to ordinary differential equations. In chemical process modelling, for example, a common situation where such DAEs arise is when relatively fast transient differential equations are approximated by quasi-steady-state algebraic equations in order to avoid numerical integration problems in simulation (see Kumar and Daoutidis [1]). High differential index DAEs, where the index is defined as the minimum number of differentiations required to obtain an equivalent ordinary differential equation [1], are particularly difficult to solve. Even index-one DAEs require particular numerical methods, such as DASSL, for their solution as described in the text books by Brenan, Campbell and Petzold [2] and Asher and Petzold [3]. A foundation for the optimal control of nonlinear DAEs has been provided by Jonckheere [4] who provided first-order necessary conditions based on a Hamiltonian characterisation. This paper also uses a variational approach to determine necessary optimality conditions which are solved in an iterative fashion. The novelty of the technique is that the iterations are performed on the basis of an

2

appropriately modified linear quadratic model based problem which contains time dependant modifiers and parameters calculated so that the final converged solution is that of the nonlinear DAE system optimal control. The formulation is based on dynamic integrated system optimisation and parameter estimation (DISOPE) which is a technique for solving complex nonlinear optimal control problems by iterating on simplified model based representations (see Roberts [5], Becerra [6]). An advantage of the resulting algorithm is that it avoids any requirement to numerically solve the DAEs during the iterative procedure.

2. Problem Formulation Consider the following continuous optimal control problem (OCP), defined over the time horizon t ∈[ t o , t f ] ,

J 1

6

min Φ x ( t f ), z ( t f ), t f + *

u(t )

I

tf

to

L

L ( x ( t ), z ( t ), u( t ), t )dt (1) *

subject to the semi-explicit type one differential-algebraic equations: * x& ( t ) = f ( x ( t ), z ( t ), u ( t ), t ) ; x ( t o ) = xo (2) * g ( x ( t ), z ( t ), u( t ), t ) = 0

%& '

() *

where u( t ) ∈ ℜ , z ( t ) ∈ ℜ and x ( t ) ∈ ℜ are the control, algebraic variable and state vectors respectively, m

nz

nx

Φ : ℜ × ℜ × ℜ → ℜ is the system terminal measure, *

nx

nz

L : ℜ × ℜ × ℜ × ℜ → ℜ is the system performance *

nx

nz

m

function, f :ℜ × ℜ × ℜ × ℜ → ℜ , the system state equations, and *

measure represents

g :ℜ × ℜ × ℜ × ℜ → ℜ algebraic equations. *

nx

nz

m

nz

nx

nz

represents

m

nx

the

system

Because, in complex situations, OCP is difficult to solve, the following simplified model based problem (MOP) is considered min u(t )

{Φ ( x (t

f

), z (t f ), t f , γ 1 )

+ ∫t L ( x (t ), z (t ), u (t ), t , γ 2 (t ) ) dt tf o

}

(3)

5 %& 0 ' g0 x(t ), z(t ), u(t ), t , α (t )5 = 0

subject to x&( t ) = f x ( t ), z ( t ), u ( t ), t , α 1 ( t ) ; x ( t o ) = xo 2

where

γ 1 ∈ ℜ, γ 2 ( t ) ∈ ℜ, γ 3 ∈ ℜ , α 1 ( t ) ∈ ℜ q

α 2 (t ) ∈ ℜ

qn

are

z

identifiable

qn

x

() *

*

*

(4)

*

and

parameters,

L: ℜ × ℜ × ℜ × ℜ × ℜ → ℜ is the model performance nz

m

f :ℜ × ℜ × ℜ × ℜ × ℜ → ℜ , model state equations, and nx

measure function, represents the

nz

m

nx

nx

g: ℜ × ℜ × ℜ × ℜ × ℜ → ℜ represents the model algebraic equations, where it is now assumed g z (.) = ∂g (.) ∂z is non-singular. The mathematical structure of (3) and (4) is deliberately chosen so that MOP is readily solvable using standard techniques [7]. For instance, f(.).and g(.) may be linear and L(.) and Φ(.) may be quadratic. In particular it is assumed that the algebraic variables z(t) can be eliminated from (3) and (4) using g(.) = 0. These assumptions and simplifications are not usually possible in the generally intractable original problem OCP defined by (1), (2). The objective, however, is to obtain the solution of OCP by appropriately iterating on MOP, matching the model to the original system through the parameters γ 1 , γ 2 ( t ), α 1 ( t ), and α 2 ( t ) . However, simply iterating between successive solutions of parameter estimation and solving MOP will not, in general, achieve the correct optimal solution of ROP. It is necessary to properly integrate the two problems taking account of the their mutual interaction. Dynamic Integrated System Optimisation and Parameter Estimation, DISOPE, provides the means for performing this integration and producing an appropriate iterative procedure [5]. nx

nz

m

nz

2

f

nz

nx

*

1

2

Φ: ℜ × ℜ × ℜ × ℜ → ℜ is the model terminal measure, nx

%K f 0w(t ), y(t ), v(t ), t , α (t )5 = f (w(t ), y(t ), v(t ), t )(K K& g0w(t ), y(t ), v(t ), t , α (t )5 = g (w(t ), y(t ), v(t ), t ) K) (6) KK L0w(t ), y(t ), v(t ), t , γ (t )5 = L (w(t ), y(t ), v(t ), t ) KK 'Φ1w(t ), y(t ), t , γ 6 = Φ 1w(t ), y(t ), t 6 * %Kv(t ) = u(t ) (K (7) &Kw(t ) = x(t ))K ' y (t ) = z(t ) *

nz

f

f

m

+ ∫t

tf o

{ L ( x (t ), z (t ), u (t ), t , γ

2

(t ) )

H (.) = L x ( t ), z ( t ), u( t ), t , γ 2 ( t )

*

( w(t ), y (t ), v(t ), t )

+ 12 r1 u (t ) − v (t )

2

0 5 + q ( t ) g0 x ( t ), z ( t ), u( t ), t , α ( t ) 5 + p ( t ) f x ( t ), z ( t ), u ( t ), t , α 1 ( t )

+ 12 r3 z (t ) − y (t ) subject to (3) and (4) together with:

2

} } dt

(8)

T

2

− λ ( t ) u( t ) − β ( t ) x ( t ) − θ ( t ) z ( t ) T

T

T

where p ( t ) ∈ ℜ is the co-state vector, q ( t ) ∈ ℜ is a time dependent Lagrange multiplier attached to the algebraic nx

nz

equations; and λ ( t ) ∈ ℜ , β ( t ) ∈ ℜ and θ ( t ) ∈ℜ are modifiers. Application of first-order variational calculus and algebraic manipulation then produces the following subsets of the necessary optimality conditions: m

nz

nx

%K∇ H (.) + r (u(t ) − v(t )) + r g (.) 1 g(.) − g (.)6 = 0(K K& p& (t ) = −∇ H (.) − r ( x(t ) − w(t )) K) (9) KK − r g (.) 1 g(.) − g (.)6 KK '∇ H (.) + r (z(t ) − y(t )) + r g (.) 1 g(.) − g (.)6 = 0* u

1

u

0

z

T

*

T

*

2

*

x

3

z

0

[∇ Φ (.) + Γ x

1

− p (t f )

]

T

δ x (t f )

+ [ ∇ z Φ (.) + Γ 3 ] δ z (t f ) = 0 T

(K %Kλ (t ) = 1 f (. ) − f (.)6 p(t ) KK + 1 g (. ) − g (.)6 q(t ) + 1∇ L(. ) − ∇ L (.)6 KK K Kβ (t ) = 1 f (.) − f (. )6 p(t ) &K + 1 g (.) − g (.)6 q(t ) + 1∇ L(.) − ∇ L (.)6)K KK KKθ (t ) = 1 f (. ) − f (.)6 p(t ) KK + 1 g (. ) − g (.)6 q(t ) + 1∇ L(.) − ∇ L (.)6 KK * '

(10)

*

v

T

*

v

2

+ 12 r2 x (t ) − w(t )

5

T

v

+ 12 r0 g ( x (t ), z (t ), u (t ), t , α 2 (t ) ) −g

nz

0

Define:

0

u (t )

f

nx

T

The key to the DISOPE approach is initially to define the following integrated problem, equivalent to OCP, which is known as the expanded optimal control problem (EOCP): min {Φ ( x (t f ), z (t f ), t f , γ 1 )

f

where v ( t ) ∈ ℜ w( t ) ∈ ℜ and y ( t ) ∈ ℜ are introduced to separate the controls, algebraic variables and states in the model based optimal control problem from the respective signals in the parameter estimation problem, defined by (6), and r0 ∈ ℜ , r1 ∈ ℜ , r2 ∈ ℜ and r3 ∈ ℜ , denoted as convexification parameters, are positive weighting parameters introduced to improve convexity and aid convergence of the resulting iterative algorithm.

x

3. DISOPE Approach

f

1

v

w

T

*

w

(5)

*

w

w

y

w

T

*

y

*

y

v

T

*

w

2

*

v

y

T

*

y

y

(11)

%KΓ = − ∇ Φ(.) − ∇ Φ (.) (K &KΓ = − ∇ Φ(.) − ∇ Φ (.) )K ' * *

1

w

w

2

y

y

−g

(12)

*

*

( w(t ), y (t ), v(t ), t )

+ 12 r2 x (t ) − w(t )

together with a repeat of (4), (6) and (7).

T

;1 6 + I : L0 x ( t ), z ( t ), u( t ), t , γ ( t ) 5 + r g 0 x ( t ), z ( t ), u( t ), t , α ( t ) 5

min Φ x ( t f ), z ( t f ), t f , γ 1 + Γ1 x ( t f ) + Γ2 z ( t f )

s. t.:

+ 21 r1 u( t ) − v ( t )

2

+ 21 r3 z ( t ) − y ( t )

2

under the co-state terminal condition

T

+ S 2 z ( t f ) + Γ2 δz ( t f ) = 0

2

2

T

@@

T

%Kx&(t ) = Ax(t ) − BR B p(t ) + α (t )(K &K p& (t ) = −Qx~ (t ) − A p(t ) + β~(t ) )K ' x ( t ) = x , p( t ) = S x ( t ) + Γ *

specified specified

parameters modifiers

satisfy (10) under specified modifiers Γ1 and Γ2 . Note that the function g(.) in (4) can be deliberately chosen to facilitate the solution of (13), for instance by expressing the algebraic variables z(t) in terms of x(t) and u(t) and then applying elimination. The parameters γ 1 , γ 2 ( t ), α 1 ( t ) and α 2 ( t ) are determined from (6) while (11) and (12) define the calculations for computing the modifiers λ ( t ), β ( t ) , θ ( t ),

Γ1 and Γ2 .

0

where

o

f

1

%K f (.) = Ax(t ) + Bu(t ) + Cz(t ) + α (t ) (K KK g(.) = Dx(t ) + Ez(t ) + α (t ) KK &K L(.) = 1 x(t ) Qx(t ) + u(t ) Ru(t ) + z(t ) S z(t )6)K (14) KK + γ (t ) KK 'Φ(.) = 1 x(t ) S x(t ) + z(t ) S z(t )6 + γ * 1

2

T

T

0

2

2

T

f

T

1

f

f

2

f

1

noting that the dependence of g(.) on u has been deliberately ignored in order to simplify the problem further, then the modified model based optimal control problem becomes

{ ( x (t

f

)

) S1 x (t f ) + z (t f ) S 2 z (t f ) + γ 1 T

1

T

+Γ1 x (t f ) + Γ 3 z (t f )

∫ { ( x (t ) 1

2

T

Qx (t ) + u (t ) Ru (t ) + z (t ) S 0 z (t ) T

+γ 2 (t ) + 12 r0 Dx (t ) + Ez (t ) + α 2 (t )

1

α ( t ) = α 1 ( t ) − CE α 2 ( t ) + BR λ ( t ) + r1v ( t ) ~ −T −1 T Q = Q + r2 I n + D E S0 + r3 I n E D ~ β ( t ) = β ( t ) + r2 w( t ) −1

−1

3

x

−D E T

−T

z

8

3θ (t ) + r y(t ) + 3S + r I 8E S = 1 S + D E S E D6 3

0

3

−T

1

nz

−1

α 2 (t )

−1

2

−T

T

If we deliberately choose a linear quadratic model for MOCP in (4) such that

1 2

1

Γ1 = Γ1 − D E Γ3 + D E S2 E α 2 ( t f )

3.1 Linear Quadratic MMOCP Situation

T

)

(18)

−1

T

T

f

A = A − CE D R = R + r1 I m

1

1

T

T

(13)

λ ( t ), β ( t ) and θ ( t ) and where the co-state p(tf) has to

tf

(17)

Application of the Maximum Principle to (15), (16) and (17) produces the two-point boundary value problem (see Lewis and Syrmos [7]) −1

subject to (4) under γ 1 , γ 2 ( t ), α 1 ( t ), and α 2 ( t ) ;

to

(16)

2

T

+ 21 r2 x ( t ) − w( t )

T

+

(15)

S1 x ( t f ) + Γ1 − p ( t f ) δx ( t f )

− λ ( t ) u( t ) − β ( t ) x ( t ) − θ ( t ) z ( t ) dt

u (t )

@@

T

2

*

2

2

1

2

− g ( w( t ), y ( t ), v ( t ), t )

min

2

%&x&(t ) = Ax(t ) + Bu(t ) + Cz(t ) + α (t )() 'Dx(t ) + Ez(t ) + α (t ) = 0 *

tf

0

+ 12 r3 z (t ) − y (t )

2

T

u(t )

to

+ 12 r1 u (t ) − v (t )

− λ ( t ) u( t ) − β ( t ) x ( t ) − θ ( t ) z ( t ) dt

From (8), it is seen that conditions (9) and (10) can be satisfied by solving the modified model based optimal control problem (MMOCP):

1 2

2

T

−T

−1

6

8

(19) (20) (21) (22)

(23) (24) (25)

with the optimal control law

u( t ) = R

−1

1 − B p( t ) + λ ( t ) + r v ( t ) 6 T

1

(26)

Applying Riccati techniques, we assume p( t ) = K ( t ) x ( t ) + k ( t ), K ( t f ) = S1 , k ( t f ) = Γ1 (27) which on substituting into (18) and (26) produces the Riccati equations −1 T K& ( t ) = K ( t ) BR B K ( t ) − K ( t ) A ~ T − A K ( t ) − Q , K ( t f ) = S1 (28)

1

6

−1 T T k& ( t ) = K ( t ) BR B − A k ( t ) ~ − K ( t )α ( t ) + β ( t ), k ( t f ) = Γ1

and a combined feedback - feedforward control law u( t ) = G ( t ) x ( t ) + h( t ) where −1

G (t ) = − R B K (t ) h( t ) = − R

−1

T

B k ( t ) + λ ( t ) + r1v ( t ) T

(29) (30) (31) (32)

From (16), (18) and (27) we can write

2

−1

7

x&( t ) = A − BR B K ( t ) x ( t ) T

+ α ( t ) − BR B k ( t ), x ( t 0 ) = x0 −1

z(t ) = − E

−1

T

1 Dx(t ) + +α (t )6 2

(33) (34)

non-linear differential algebraic equation optimal control problem defined by (1) and (2) by repeated solution of the modified linear quadratic optimal control problem defined by (15), (16) and (17), and is stated as follows: (note: superscript (i) refers to iteration i.) Data

Solution of (27) to (34) produces an estimate of the optimal solution of u(t), x(t), z(t) and p(t) in terms of parameters, α1(t) and α2(t), and modifiers, λ(t), β(t), θ(t), Γ1 and Γ2. Expressions for the parameters and modifiers are derived using (14) with (6), (11) and (12) to give

%Kα (t ) = f (w(t ), y(t ), v(t ), t ) (K K& − Aw(t ) − Bv(t ) − Cy(t )K) (35) KKα (t ) = g (w(t ), y(t ), v(t ), t ) KK ' − Dw(t ) − Ey(t ) * noting that it is not necessary to compute γ or γ (t), and %Kλ (t ) = 1 B − f ( w(t ), y(t ), v(t ), t )6 p(t ) (K KK − g ( w(t ), y(t ), v(t ), t ) q(t ) KK KK +1 Rv(t ) − ∇ L ( w(t ), y(t ), v(t ), t )6 KK KKβ (t ) = 1 A − f ( w(t ), y(t ), v(t ), t )6 p(t ) KK &K +1 D − g (w(t ), y(t ), v(t ), t )6 q(t ) )K (36) KK +1Qw(t ) − ∇ L (w(t ), y(t ), v(t ), t )6KK KKθ (t ) = 1C − f ( w(t ), y(t ), v(t ), t )6 p(t ) KK KK +1 E − g ( w(t ), y(t ), v(t ), t )6 q(t ) KK ' +1S y(t ) − ∇ L ( w(t ), y(t ), v(t ), t )6*

Step 0

*

1

*

2

1

2

Step 1

T

*

v

Step 2

v

*

v

T

*

Step 3

w

T

*

for instance, by solving the model based optimal control problem defined by (15) and (16) with all parameters and modifiers set to zero. Set i = 0 . Parameter Estimation: Use (35) to compute the parameters α1(i)(t), and α2(i)(t). Modifier Computation: Use (36) and (37) to compute the modifiers λ(i)(t), β(i)(t), θ(i)(t), Γ1 and ~ (i ) (i ) (i ) (i ) Γ2 . Then compute α ( t ) , β ( t ) and Γ1 from (21), (23) and (25), respectively. Modified Optimisation (a) Use (29) to compute k(i)(t) and (32) to compute h(i)(t). (b) Use (33) to compute the predicted optimal (i )

T

*

A, B, C, D, E, Q, R, S0, S1, S2, ku, kx, kz, kp, kq, r0, r1, r2, r3, and means for computing f*(.), g*(.), L*(.), Φ∗(.) together with their required derivatives. Initialisation: Compute the matrix expressions A , ~ R , Q , and S1 from (19), (20), (22) and (24) respectively. Then compute K(t) from (28) and G(t) from (31). Then calculate or choose a nominal (0) (0) solution v(0)(t), w(0)(t), y(0)(t) p$ ( t ) and q$ ( t ) ;

w

*

state response x

w

T

*

costate p

T

y

z

*

Γ1 = −  S1 w(t f ) − ∇ w Φ ( w(t f ), y (t f ), v(t f ), t f )     (37) * Γ = − S y ( t ) − ∇ Φ ( w ( t ), y ( t ), v ( t ), t )    y f f f f   2 f  2 *

An expression for q(t) is obtained from the final equation in group (9) which, using (8), (14) and (34), produces

q (t ) = E

−T

1S E 0 Dx(t ) + α (t )5 − C p(t ) + θ (t )6 −1

0

T

+ r0 g ( w( t ), y ( t ), v ( t ), t ) *

Step 4

(38)

4. DAE-DISOPE Algorithm Equations (27) to (38), together with (7), provide enough information to compute the solution of (1) subject to (2). The equations are tightly coupled and require an iterative procedure to achieve the solution. There are several possible alternative algorithms which may be employed. The advantage of the particular algorithm presented here, denoted by the acronym DAE-DISOPE, is that the final solution is obtained without any requirement to integrate the non-linear DAE equations (2). The algorithm is designed to solve the

( t ) from (30),

( t ) from (27), algebraic variable

( t ) from (34) and Lagrange multiplier

q ( t ) from (37). Update Use the following relaxation scheme to (i ) (i ) update v(i)(t), w(i)(t), y(i)(t) p$ ( t ) and q$ ( t ) .

%Kv KKw &K y KK p$ K'q$

( i + 1)

( i + 1)

( i + 1)

2

(t ) = v (t ) + k u u (i )

( i + 1) 1

( i +1)

2

( i +1)

( i +1 )

( i +1)

( i +1)

y

0

( t ) . Then compute the

predicted optimal control u

y

*

( i +1)

( i + 1)

1

(t ) = w (t ) + k x x (i )

1 ( t ) = p$ ( t ) + k 1 p ( t ) = q$ ( t ) + k 2q (t ) = y ( t ) + k z z (i)

(i )

( i + 1)

( i + 1)

( i +1 )

q

(i )

( i + 1)

p

(i )

7 (K ( t ) − w ( t ) 6K K (t ) − y ( t )6 ) K ( t ) − p$ ( t ) 6 K K* ( t ) − q$ ( t ) 7 K

(t ) − v (t ) (i )

(i)

(39)

(i)

(i )

Then repeat from Step 1 until convergence is achieved.

5. Convergence Considerations A full analytical analysis of the convergence properties of the DAE-DISOPE algorithm has not yet been performed. However, by comparison with previous work, it is expected that convergence performance will be enhanced if the overall

problem is at least locally convex [8] [9] [10]. Furthermore, the existence of the optimal solution and convergence to it will probably require that the non-linear system descriptions (1) and (2) are Lipschitz continuous. In the practical implementation of the algorithm relaxation gains, ku, kx, kz, kp, kq and kp and convexification term coefficients r0, r1, r2 ,and r3 are provided in order to regulate stability and convergence. In the initial application of the algorithm to a given problem, the relaxation gains would, in general, be chosen as unity, and the convexification term coefficients set to zero; and adjusted only if convergence difficulties arise.

6. Illustrative Example The DAE-DISOPE algorithm as described in the previous section has been implemented in MATLAB. In order to illustrate the behaviour of the algorithm a seventh - order non-linear system with a quadratic performance index and two control inputs has been simulated. There are four algebraic variables together with four algebraic equations. The non-linear optimal control problem (OCP) is defined as: 1

min u 1 ( t ), u 2 ( t )

%&∑ x (1) I J∑ x (t ) + 0.1∑ u (t )L() ' * 7

+

2

i

2

7

1

0

i =1

1

2

2

i

i =1

2

i

(40)

%K (K KK x& (t ) =+0−.25xx ((tt)) ++ z0.5(tx) (t ) − 0.5x (t ) + 0.2 x (t )KK KK −0.1x (t ) + z (t ) KK KK x& (t ) = 0.1x (t ) − 1.5x (t ) + x (t ) + 0.5x (t ) KK KK +0.1x (t ) + z (t ) KK & &K x (t ) = 0.2 x (t ) − 0.5x (t ) − x (t ) + 0.2 x (t ) )K (41) KK x& (t ) =+00..22ux ((tt)) + 0.17 x (t ) − x (t ) + x (t ) KK KK + x (t ) x (t ) KK KK x& (t ) = 0.1x (t ) − 0.2 x (t ) − x (t ) − 0.5x (t ) KK KK x& (t ) = 0.4 x (t ) + 0.1x (t ) − x (t ) − 0.5x (t ) KK ' − x (t ) + 0.1u (t ) * %Kz (t ) − 2z (t ) x (t ) x (t ) + 15. z (t ) x (t ) x (t ) (K KK − 0.5x (t ) x (t ) = 0 KK &Kz (t ) − 2z (t ) x (t ) + 15. z (t ) x (t ) − 0.5x (t ) = 0)K (42) KKz (t ) − 0.2 x (t ) − 0.2z (t ) = 0 KK z (t ) − u ( t ) = 0 ' *

subject to: x& 1 ( t ) = −5 x 1 ( t ) + 0.2 x 2 ( t ) + 0.5x 3 ( t ) + 0.1x 5 ( t ) 6

2

4

7

5

3

6

4

5

3

2

3

4

5

1

1

5

3

1

7

1

7

3

2

1

1

5

7

6

6

2

4

6

2

3

5

2

1

2

3

3

1

2

3

2

3

2

2

1

1

2

2

2

1

2

6

9

1

1

2

3

4

3

(43)

2

3

4

3

1

1

However, this fact is not employed in the implementation of the DAE-DISOPE algorithm. It is employed to test the accuracy of the final results by comparing them with the results obtained from a standard DISOPE implementation [5] applied to (41) and (42) where the algebraic variables are substituted using (43). In the implementation of the algorithm the matrices A, B, C, D and E are set as Jacobian matrices derived by linearisation of (41) and (42) about x(t) = z(t) = u(t) = 0. The matrices Q, R and S1 are chosen to match (41) with S0 and S2 both null. Figures 1 and 2 compare the resulting optimal control and state signals while Figure 3 shows the corresponding algebraic signals. For this example it is observed that the DAE-DISOPE algorithm produces results that are in good agreement with the DISOPE results. The DAE-DISOPE results were obtained using relaxation gain and convexification parameters ku=0.5, kx=1, kz=1, kp=1, kq=1, r0=0.4, r1=0.7, r2=0 and r3=0. The corresponding satisfactory convergence behaviour, plotted using absolute and semilogarithmic scales, is demonstrated in Figures 4 and 5 which show how the norms

u (t ) − u (i )

( i −1)

(t )

and

(i )

g (t )

6

2

5

1

converge as the iterations proceed.

2

2

4

3

2

1

2

3

%K (K K&z (t ) = x (t ) K) KKz (t ) = 0.2 x (t ) + 0.2u (t )KK 'z (t ) = u ( t ) *

Note the algebraic equations (42) have the real solution z1 ( t ) = x1 ( t ) x 2 ( t )

4

1

with initial conditions x1(0) = 1.0, x2(0) = 0.8, x3(0) = 0.5, x4(0) = 0.6, x5(0) = 1.5, x6(0) = 1.1 and x7(0) = 1.2.

7. Conclusions An algorithm has been described which achieves the solution of nonlinear optimal control problems subject to differential algebraic equations. A significant advantage of the technique is that there is no requirement to solve the differential algebraic equations during the iterations. The utility of the algorithm has been demonstrated successfully. The technique as described is limited to type-one differential algebraic systems and research is continuing to remove this restriction and to perform local and global convergence analysis.

References [1] Kumar, A. and P Daoutidis, Control of Nonlinear Differential Algebraic Equation Systems with Applications to Chemical Process, 1999, (Chapman & Hall/CRC, London, UK) [2] Brenan K.E., Campbell S.E. and L.R. Petzold, Numerical Solution of Initial-Value Problems in Differential-Algebraic Equations, 1996, (SIAM, USA) [3] Ascher U.M. and L.R. Petzold, Computer Methods for Ordinary Differential Equations and DifferentialAlgebraic Equations, 1998, (SIAM, USA)

[4] Jonckheere E., Variational calculus for descriptor problems, IEEE Transactions on Automatic Control, 1988, Vol. 33, No. 5, pp. 491-495. [5] Roberts P.D., An algorithm for optimal control of nonlinear systems with model-reality differences, Proceedings of 12th IFAC World Congress on Automatic Control, Sydney, Australia, 18-23 July, 1993, Vol. 8, pp 407-412. [6] Becerra V.M., Development and Applications of Novel Optimal Control Algorithms, Ph.D. Thesis, 1994, City University, London. [7] Lewis F.L. and V.L. Syrmos, Optimal Control,. 1995, (Wiley, New York USA) [8] Brdys M. and P.D. Roberts, Convergence and optimality of modified two-step algorithm for integrated system optimisation and parameter estimation, Int. J. Systems Sci. 1987, Vol. 18, No. 7, pp 1305-1322. [9] Brdys M., Ellis J.E. and P.D. Roberts, Augmented integrated system optimisation and parameter estimation techniques: derivation, optimality and convergence, IEE Proceedings Pt. D, 1987, Vol. 134, No. 3, pp 201-209. [10] Becerra V.M. and P.D. Roberts, Dynamic integrated system optimization and parameter estimation for discrete time optimal control of nonlinear systems, Int.J.Control, 1996, Vol. 63, No. 2, pp257-281.

25 20

z4(t) 15 10

z3(t)

5

z2(t) z1(t)

0 -5 -10 0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

time t (secs) Fig.3 Optimal algebraic signals

u(i ) (t ) − u(i −1(t ) 20 15 10 5 0 0

10

20

30

40

50

60

10

20

30

40

50

60

2

10

0

10 40 30

-2

10

20

0

iteration i Fig.4 Control signal convergence

u 1 (t)

10 0 -10

u 2 (t) -20

(i )

D A E -D I S O P E o o o D IS O P E

-30 -40 0

0 .1

0 .2

0 .3

0 .4

0 .5

0 .6

0 .7

0 .8

0 .9

g (t ) 80 1

tim e t (se c s)

60 40

Fig.1 Optimal control signals

20

2

0 0

1.5

10

20

30

40

50

60

10

20

30

40

50

60

2

x(t)

10 1

0

0.5

10

0 -2

10 -0.5

-1 0

0.1

0.2

0.3

0

iteration i Fig.5 Algebraic constraints convergence

DAE-DISOPE o o o DISOPE 0.4

0.5

0.6

0.7

time t (secs)

Fig.2 Optimal state signals

0.8

0.9

1