• Ei tuloksia

2. Well-posed linear systems and state feedback

N/A
N/A
Info
Lataa
Protected

Academic year: 2022

Jaa "2. Well-posed linear systems and state feedback"

Copied!
23
0
0

Kokoteksti

(1)

c

°2005 Birkhäuser Verlag Basel/Switzerland and Operator Theory

State-feedback stabilization of well-posed linear systems

Kalle M. Mikkola

Abstract. A nite-dimensional linear time-invariant system is output-stabiliz- able if and only if it satises the nite cost condition, i.e., if for each initial state there exists at least oneL2 input that produces anL2 output. It is ex- ponentially stabilizable if and only if for each initial state there exists at least oneL2 input that produces anL2 state trajectory. We extend these results to well-posed linear systems with innite-dimensional input, state and output spaces. Our main contribution is the fact that the stabilizing state feedback is well posed, i.e., the map from an exogenous input (or disturbance) to the feedback, state and output signals is continuous inL2locin both open-loop and closed-loop settings. The state feedback can be chosen in such a way that it also stabilizes the I/O map and induces a (quasi) right coprime factorization of the original transfer function. The solution of the LQR problem has these properties.

Mathematics Subject Classication (2000). Primary 93D15, 49N10; Secondary 93C25.

Keywords. Exponential stabilization, output stabilization, nite cost condi- tion, LQR problem, quasi-coprime factorization.

1. Introduction

To illustrate the philosophy behind our results while avoiding undue technicalities, in this introductory section we start with the (more or less well known) nite- dimensional case.

The standard model of a nite-dimensional linear time-invariant system is

˙

x(t) =Ax(t) +Bu(t),

y(t) =Cx(t) +Du(t), t≥0, x(0) =x0.

(1.1)

This work was written with the support of the Academy of Finland under grant #203946.

(2)

A B

C D

F 0

˙

x=Ax+Bu y=Cx+Du

? x

¾x˙ 6 R

¾y

¾z=F xr

?c

+-+

uª r6u=F x+uª -

Figure 1. State-feedback connection

Here the generators[A BC D]∈ B(X×U,X×Y)are matrices, andU=Cp,X=Cnand Y=Cqare called the input space, the state space and the output space, respectively.

We calluthe input (or control),xthe state andy the output of the system.

The Laplace transform ofuis dened by bu(s) := R

0 e−stu(t)dt. One easily observes that with zero initial statex0= 0equation (1.1) leads toyb= ˆDbu, where D(s) :=ˆ D+C(sI−A)−1B (1.2) is called the transfer function of the system. Conversely, every rational matrix- valued function has a nite-dimensional realization, i.e., it is the transfer function of a nite-dimensional system.

State feedback means that we add a second output, sayz(t) =F x(t), where F ∈ B(X,U), and feed this signal to the input, as in Figure 1. Under an exogenous input (perturbation) uª, we get u=F x+uª. Solving forx,˙ y andz in terms of xanduª, we get the following closed-loop system (fort≥0)

˙

x(t) = (A+BF)x(t) +Buª(t), y(t) = (C+DF)x(t) +Duª(t), z(t) =F x(t),

x(0) =x0.

(1.3)

(By the open-loop system we mean the original system (1.1) with the additional outputz=F x, i.e., as in Figure 1 without the dashed connection.)

The (original) system is called exponentially stable i there existM, ² >0such thatkx(t)kX≤Me−²tkx0kX(t0)for each initial statex0X, or equivalently, i the spectrumσ(A)is contained in the open left half-plane C :={s∈

¯Res <

0}. By Datko's Theorem, an equivalent condition is that x∈L2(R+;X) ifu= 0 (for all x0 X). The system is called exponentially stabilizable i there exists a state feedback operator F ∈ B(X,U) such that the closed-loop system (1.3) is exponentially stable.

The following condition is called the state-FCC (state Finite Cost Condition):

For each x0X, there existsu∈L2(R+;U)such thatx∈L2(R+;X). (1.4) Recall thatxstands for the solution of (1.1).

(3)

Next we state three theorems for the system (1.1). We shall generalize them in Section 5.

Theorem 1.1. The state-FCC (1.4) holds i the system is exponentially stabilizable.

The above result is a special case of the following result that involves alsoC andD (the terminology will be explained below):

Theorem 1.2. The output-FCC (1.5) holds i the system is output-stabilizable.

The output-FCC means the following:

For each x0X, there existsu∈L2(R+;U)such thaty∈L2(R+;Y), (1.5) that is, some stable (i.e., L2) input makes the output stable. This condition is strictly weaker than the state-FCC (1.4).

The system is called output-stabilizable if there existsF ∈ B(X,U)such that the state-feedbacku(t) =F x(t)makesuandy stable for each initial statex0X (with no exogenous input:uª = 0). In fact, thenF can actually be chosen so that uandy become stable for eachx0Xand eachuª L2(R+;U), and, in addition, the mapsuª 7→[uy]become (right) coprime.

Indeed, from (1.3) we obtain, forx0= 0, thatsbx(s) = (A+BF)bx(s)+Bucª(s), hence

b

x(s) = (s−A−BF)−1Bucª(s), yb= ˆN ucª, bu=Mˆucª, (1.6) whereNˆucª = (C+DF)bx+Ducª andMˆucª=bz+ucª =Fxb+ucª, hence

Nˆ(s) :=D+(C+DF)(s−A−BF)−1B, Mˆ(s) :=I+F(s−A−BF)−1B. (1.7) By [NMˆˆ] :ucª7→[yubb] being right coprime we mean that there existf, g∈H that satisfy the Bézout equationfMˆ+gNˆ≡I. Recall thatHdenotes the space of bounded holomorphic functions on the right half-plane C+:={s∈

¯Res >

0}.

We express the above as follows:

Theorem 1.3. If the output-FCC (1.5) holds, then there exists an output- and I/O-stabilizing state feedback with Nˆ andMˆright coprime.

By output- and I/O-stabilizing we mean that (see Figure 1)

k[yz]k2≤M(kx0kX+kuªk2) (x0X, uªL2). (1.8) The map Dˆ : bu7→ by can be written as Dˆ := NˆMˆ−1. We express this as follows:

Corollary 1.4. Any function having an output-stabilizable realization has a right- coprime factorization.

The state-feedback operatorF used above is usually obtained by solving the so-called (algebraic) LQR Riccati equation

PB(I+DD)−1BP =AP+PA+CC, (1.9)

(4)

sinceF =−(I+DD)−1BP for the minimal nonnegative solutionP. This F is the unique state-feedback operator that minimizes the LQR cost function

J(x0, u) = Z

0

¡ky(t)k2Y+ku(t)k2U¢

dt (1.10)

for each initial statex0X.

In Section 5 we shall extend the above theorems to arbitrary well-posed linear systems (WPLSs). These systems are a generalization of the systems of type (1.1) and allow unbounded generators and innite-dimensional input, state and output spaces; see Section 2 for details.

The state-FCC has been studied under the name optimizability in, e.g., [WR01]. Thus, our generalization of Theorem 1.1 shows that optimizability is equivalent to exponential stabilizability.

It was already known that under the output-FCC an output-stabilizing con- trol is produced by another WPLS, as shown in [Zwa96] (the case withC∈ B(X,Y) was shown in [FLT88]), or by ill-posed state feedback. It was not known that this other system can be obtained from the original one by a well-posed state feedback.

It means that the state-feedback loop is well posed with respect to external distur- bance, i.e., that the mapsMˆ:ucª7→buand Mˆ−1are well dened (see Figure 1).

In fact, the functionsMˆandMˆ−1 become proper (bounded on some right half- plane).

Any bounded state-feedback operator (F ∈ B(X,U)) generates a well-posed state feedback, but so do some unbounded ones. IfB andCare bounded, then the stabilizing state feedback is given by a boundedF; this special case of Theorems 1.1 and 1.2 was already known. The two theorems were known also for fairly un- boundedB's in the case thatAgenerates an analytic semigroup andCis bounded [LT00].

To extend Theorem 1.3 and Corollary 1.4 to arbitrary WPLSs, we must replace coprimeness by quasi-coprimeness, which we dene below.

For any ω Rwe set C+ω :={s∈

¯Res > ω}. By H2ω(U) we denote the Hilbert space of holomorphic functionsC+ω Ufor which

khk2H2

ω := sup

r>ω

Z

−∞

kh(r+it)k2Udt <∞. (1.11) Moreover, C+ := C+0, H2 := H20. Bounded holomorphic functions Nˆ : C+ B(U,Y)andMˆ:C+→ B(U)are called quasiright coprime i

·Nˆ Mˆ

¸

h∈H2 h∈H2 for everyh∈H2ω(U) and ω∈R. (1.12) In other words, quasiright coprimeness means that if h is in some H2ω and its image is in H2, thenh must actually have been in H2. We identify any function with its holomorphic extension (if any) to a right half-plane, so H2 means that is the restriction of an element ofH2.

(5)

For a quasiright coprime factorization Dˆ = NˆMˆ−1, the image [MNˆˆ] [H2] equals the graph [DIˆ] [{f H2¯

¯Dfˆ H2}]. In fact, also the converse holds. See also Lemma 4.4.

Whenever Nˆ and Mˆ are (Bézout) right coprime, they are quasiright co- prime. Indeed,[NMˆˆ]h∈H2andfMˆ+gNˆ≡I imply thath

f g¤

[NMˆˆ]h∈H2. The two forms of coprimeness are equivalent if Nˆ and Mˆ are rational [Mik02, Lemma 6.5.3]. However, quasi coprimeness is in certain sense a more natural ex- tension of coprimeness to the non-rational case. We shall treat dierent forms of coprimeness in detail in future articles.

Our proof is based on showing that the control minimizing (1.10) is given by well-posed state feedback, but due to the unboundedness ofB,C and F, we must use the integral Riccati equations of Lemma 3.8 instead of the algebraic one above. Those equations allow us to reduce the minimization problem to a stable one by replacingAbyA−αforαbig enough (we must add some cost on the state to keep the minimal cost the same). The stable LQR problem can then be solved by using a spectral factorization.

See Section 6 for generalizations and further historical comments.

2. Well-posed linear systems and state feedback

In this section we present our notation and denitions except for those concerning optimization and coprimeness. The denitions and claims in this section and fur- ther details can be found in, e.g., [Sta05], [Sta98a], [Wei94] or [Mik02, Chapter 6].

ByB(U,Y)we denote the set of bounded linear operatorsUY, and we write B(U) :=B(U,U)(similarly for something else in place ofB).

LetU,X,Ybe arbitrary complex Hilbert spaces. If the generators of the system (1.1) are bounded, i.e.,[AC BD]∈ B(X×U,X×Y), then the unique solution of (1.1) is obviously given by the system

½x(t) = Atx0+Btu

y = Cx0+Du, (2.1)

where

At= eAt, Btu= Z t

0

At−sBu(s)ds, (Cx0)(t) =CAtx0, (Du)(t) =CBtu+Du(t).

(2.2) The above formulas are actually valid for fairly unbounded operators, but in the most general case the right-hand-sides (at least CBt) become meaningless.

Therefore, the WPLSs (also known as SalamonWeiss systems or abstract linear systems) were dened by requiring the system to be linear and time-invariant and A to be strongly continuous; in addition, one requires that£

AtBt C D

¤is causal and continuousX×L2loc(R+;U)X×L2loc(R+;Y)for eacht≥0, or equivalently, that

kx(t)k2X+ Z t

0

ky(s)k2Yds≤Kt

¡kx0k2X+ Z t

0

ku(s)k2Uds¢

(2.3)

(6)

A C D

? x0

? u

Initial state Input (control) State

Output

¾ x x=Ax0+Bτ u

¾ y y=Cx0+Du Figure 2. Input/state/output diagram of a WPLSh

ACB D

i

for all (equivalently, some) t > 0, where Kt depends on t only. An equivalent formulation, from [Sta98a], is given in Denition 2.1. There we use the natural extensionsB(ofBtτ−t) andD that allow the inputs to be dened on the whole real line, thus simplifying some formulae.

We use the notationL2ω= eω·L2={f¯

¯e−ω·f L2},(τtu)(s) :=u(t+s)and π±u:=χR±u, whereχE(t) :=

n1, tE;

0, t6∈E . WhenE R, we set πEu:=χEu. We identifyL2ω(E;U)with functions inL2ω(R;U)that vanish outsideE.

We study the following generalization of systems of type (2.2):

Denition 2.1 (WPLS and stability). Let ω R. An ω-stable well-posed linear system on (U,X,Y) is a quadruple Σ = £A B

C D

¤, where At, B, C, and D are bounded linear operators of the following type:

1. A· is a strongly continuous semigroup of bounded linear operators on X satisfyingsupt≥0ke−ωtAtkX<∞;

2. B: L2ω(R;U)XsatisesAtBu=tπufor allu∈L2ω(R;U)andt∈R+; 3. C:XL2ω(R;Y)satisesC Atx=π+τtCxfor allx∈Xandt∈R+; 4. D: L2ω(R;U) L2ω(R;Y) satises τtDu = tu, π+u = C Bu, and

π+u= 0for allu∈L2ω(R;U)andt∈R.

The dierent components of Σ = £A B

C D

¤ are named as follows: A is the semigroup,Bthe input map,C the output map, andDthe I/O map (input/output map) ofΣ.

We say thatA (resp.B,C,D) is stable if 1. (resp. 2., 3., 4.) holds forω= 0.

Exponentially stable meansω-stable for someω <0. The system is output stable (resp. I/O-stable) ifC (resp.D) is stable.

We setΣτ:= [AC D ], Bt:=tπ+=tπ[0,t),Dt:=π[0,t)[0,t).

For any x0 Xandu∈L2loc(R+;U)we associate the state (trajectory)x:=

Ax0+Bτ u and output y := Cx0+Du on R+ (i.e., [xy] = Στ[xu0]), as in (2.1) and Figure 2. (By causality, also D is dened for any u L2loc(R+;U) through π[0,t)Du=π[0,t)[0,t)u(t0).)

From Denition 2.1 we easily obtain thatB=,C =π+C and

π+τty=Cx(t) +Dπ+τtu (t0). (2.4) Indeed,π+τty=π+τtCx0+π+D(π+π+tu=C Atx0+C Bτtu++τtu= Cx(t) +Dπ+τtu. This says that the output is time-invariant, i.e., the remaining output (at timet) depends only on the current state and the remaining input.

(7)

The same holds for the state:π+τtx=Ax(t)+Bτ π+τtu(t0). Conversely, any linear system satisfying these two equations and (2.3) is a (restriction of a) WPLS.

By A we denote the innitesimal generator ofA. One can show that there existB∈ B(U,Dom(A))andC∈ B(Dom(A),Y)such that the middle formulas in (2.2) hold (foru∈L2ω andx0Dom(A)). Moreover,x˙ =Ax+BuinDom(A).1 An ω-stable WPLS is an ω0-stable WPLS for any ω0 > ω (we identify the unique extensions/restrictions ofA,B,C andD for dierentω).

Exponential stability of a system is equivalent to that of its semigroup, hence Datko's Theorem [Dat70] leads to the following:

Lemma 2.2. The WPLS Σ = [AC BD] is exponentially stable i Ax0 L2(R+;X)

for allx0X. ¤

Exponential stability implies that σ(A) C := {z

¯Rez < 0}. The converse holds if, e.g.,Ais bounded or analytic [Sta05].

Now it is the time to present our assumptions:

Standing Assumption 2.3. Throughout this article we assume thatU,XandYare complex Hilbert spaces,Σ = [AC BD]is a WPLS on(U,X,Y), andJ =J∈ B(Y).

(The operatorJ will not be needed before Section 3.)

Let ω R. We dene TICω(U,Y) to be the (closed) subspace of operators D ∈ B(L2ω(R;U); L2ω(R;Y))that are causal (i.e., π+ = 0) and time-invariant (i.e. τtD = t for all t R). The I/O maps of WPLSs are exactly all such operators (TIC(U,Y) := ω∈RTICω(U,Y)). The Laplace transform u 7→ub is an isometric (modulo

2π) isomorphism ofL2ω ontoH2ω. ByH(Ω;X)we denote the Banach space of bounded holomorphic functionsΩXwith supremum norm. We setC+ω :={s∈

¯Res > ω}.

For each D TICω(U,Y), there exists a unique function Db Hω(U,Y) :=

H(C+ω;B(U,Y)), called the transfer function of D, such that Duc = Dbuˆ on C+ω for every u L2ω(R+;U). The mapping D 7→ Dˆ is an isometric isomorphism of TICω(U,Y)ontoHω(U,Y). IfB is bounded, thenD(s) =ˆ D+C(s−A)−1B.

A function is called proper if it is bounded and holomorphic on some right half-plane. Thus,H(U,Y) :=ω∈RHω(U,Y)is the set of all properB(U,Y)-valued functions. (We identify functions that coincide on some right half-plane.)

By G we denote the group of invertible elements. Thus, e.g., GTICω(U,Y) stands for{D TICω(U,Y)¯

¯D−1 TICω(Y,U)}, i.e., it corresponds to GHω, the set of bounded holomorphic functionsDˆ :C+ω → B(U,Y)for whichDˆ−1 exists and is (uniformly) bounded.

1This is based on the fact thatA: Dom(A)Xextends to a continuous mapXDom(A)and generates a semigroup onDom(A). Similarly,A|Dom(A2) generates a semigroup onDom(A), and all three semigroups are isomorphic. It also follows that(sA)−1Bbecomes well dened forsρ(A|X) =ρ(A|Dom(A)). However, we only need theseA's for examples.

(8)

A C D F G

Σ

ext

Σ ª

? x0

¾x

¾y

¾z=Fx0+rGu

?c

++-

uª r6u=(I−G)−1uª+(I−G)−1Fx0 - Figure 3. State-feedback connection for a general WPLS

Now we shall return to state feedback. When the input operatorBis bounded, the generators³A B

C DF 0

´and³A+BF B

C+DF D

F 0

´, determine the open-loop and closed-loop systems, respectively, corresponding to a state feedback operatorF, as in Figure 1 and equations (1.3).

Thus, state feedback means adding an extra output to the system and feeding that output back to the input (likez(t) =F x(t)in Figure 1).

For a general WPLS, the denition is the same: a state feedback means a pair[F G ]such that the extended systemΣext in Figure 3 is a WPLS.

The state feedback is called admissible if (I−Gˆ)−1 exists and is proper, or equivalently, if the map (I−G) :u7→uª has a bounded and causal inverse on L2ωfor someω∈R. This means that the (closed) state-feedback loop is well posed under external disturbance.

Thus, the denition of an admissible (or well-posed or proper) state feed- back is as follows:

Denition 2.4 (Σª,[F G ]). A pair[F G ]is called an admissible state-feedback pair forΣif the extended system

Σext:=



A B C D F G

 (2.5)

is aWPLS andI−G ∈ GTIC(U).

We setM := (I−G)−1, N :=DM, and denote the corresponding closed- loop system (see Figure 3) by

Στª=



Aª Bªτ Cª Dª

Fª Gª

=



A +BτM F BMτ C +DM F DM

M F M −I

 (2.6)

= Στext

· I 0

−F I−G

¸−1

= Στext

· I 0

M F M

¸ :

·x0

uª

¸ 7→

x y u−uª

. (2.7)

(9)

We call [F G ] exponentially stabilizing if Σª is exponentially stable. If there exists an exponentially stabilizing state-feedback pair forΣ, thenΣis called exponentially stabilizable (similarly for output-stabilizing or I/O-stabilizing).

(The systemΣªis necessarily a WPLS. Note thatΣª is output stable iCª

andFª map XintoL2.)

Any F ∈ B(X,U) determines an admissible state feedback (with F =FA, G =FBτ), but so do also some unbounded operators. The above denition also allows for feedthrough terms (i.e.,z(t) =F x(t) +Gu(t)or G =FBτ+G, where G∈ B(U)), but ifF ∈ B(X,U), then admissibility is equivalent to I−G∈ GB(U), and essentially the same feedback is obtained by using the state-feedback operator (I−G)−1F with zero feedthrough (see Lemma A.5). However, in the case thatF is unbounded, the feedthrough term G= lims→+∞Gˆ(s)need not exist [WW97, Example 11.5].

3. Optimal control and Riccati equations

In this section we shall present certain necessary and sucient conditions in terms of Riccati equations for a control to be optimal. In Section 4, these conditions will be applied to establish the existence of a (1.10)-minimizing state feedback.

The cost (1.10) is nite iu∈ U(x0), where U(x0) :={u∈L2(R+;U)¯

¯y∈L2(R+;Y)}. (3.1) We call this the set of admissible controls. Note that the output-FCC (1.5) holds iU(x0)6=∅ for eachx0X, hence the name nite cost condition.

The following is obvious (and given in [Zwa96]):

Lemma 3.1. The setU(0)is a subspace of L2. Ifu∈ U(x0), thenU(x0) =u+U(0).

¤

A control uopt is called J-minimizing for x0 if J(x0, uopt) ≤ J(x0, u) for every u ∈ U(x0). It is well known (see [Zwa96, the proof of Theorem 6]) that under the output-FCC a (1.10)-minimizing control exists:

Lemma 3.2. Assume the output-FCC (1.5). DeneJ by (1.10).

Then there exists a unique J-minimizing control uxopt0 for every x0 X and a nonnegative operator P ∈ B(X) such that J(x0, uxopt0 ) = hx0,Px0iX for every

x0X. ¤

In Section 4 we shall show that the above minimizing control is given by admissible state feedback. For that purpose we need certain integral Riccati equation conditions for minimality, also over other cost functions than (1.10) (see Lemma 3.9). Therefore, we shall introduce a cost operatorJ =J∈ B(Y)and the (generalized) cost function

J(x0, u) :=hy, JyiL2 = Z

0

hy(t), Jy(t)iYdt (x0X, u∈ U(x0)). (3.2)

(10)

As a by-product, our proofs and formulas actually apply in a much more general optimization setting (possibly indenite, such as the minimax H control of [Sta98c] and [Mik02]). The explicit inclusion ofuin (3.2) would not only reduce generality (see Lemma 3.4) but also lengthen numerous formulae below by half.

A controlu∈ U(x0)is calledJ-optimal forx0 (andΣ) ifhy, JDηiL2 = 0for eachη∈ U(0). (By [Mik02, Lemma 8.3.6], this corresponds to a zero of the Fréchet derivative ofhy, JyiL2.)

WhenJ =I, this orthogonality condition implies thatyis of minimal norm.

More generally:

Lemma 3.3. A control uminimizes J(x0,·)over U(x0) iu isJ-optimal for x0

andJ(0,·)≥0.

Proof. The suciency follows from Lemma 3.1 and the fact that for anyJ-optimal uforx0 we have

J(x0, u+η) =J(x0, u) +J(0, η) (η∈ U(0)), (3.3) which follows from the identityhy+Dη, J(y+Dη)i=hy, Jyi+ 0 + 0 +hDη, JDηi.

Ifhy, JDηi 6= 0for someη∈ U(0), then dJ(x0, u+αη)is nonzero atα= 0,

henceJ-optimality is also necessary. ¤

The cost (1.10) is a special case of (3.2):

Lemma 3.4. SetC˜:= [C0], D˜ :=£D

I

¤. Then Σ := [˜ AC˜ BD˜] is a WPLS on(U,X,Y× U). Moreover, a control isI-optimal forΣ˜ i it is (1.10)-minimizing. ¤ Indeed,Σ˜ has the outputy˜:= ˜Cx0+ ˜Du= [yu], henceJΣ˜(x0, u) :=h˜y, Iyi˜ = kyk22+kuk22, so the optimality claim follows from Lemma 3.3 (applied toΣ˜ andI;

note thatU(x0)is the same for both ΣandΣ). The WPLS claim is obvious.˜ Naturally, a minimal cost is unique (for any xed x0 X). In fact, the J- optimal cost is unique also for indeniteJ:

Lemma 3.5. IfuandvareJ-optimal controls forx0X, thenJ(x0, u) =J(x0, v). Proof. By Lemma 3.1, u˜ := v−u ∈ U(0). Buthy+Du, J˜ Dηi = 0 (η ∈ U(0)), hencehDu, J˜ Dηi= 0, also forη = ˜u. This and (3.3) imply that J(x0, u+ ˜u) =

J(x0, u). ¤

When using the dynamic programming principle, we need the following:

Lemma 3.6. Letx0Xandu∈L2loc(R+;U). Thenu∈ U(x0)iπ+τtu∈ U(Atx0+ Btu)for some (equivalently, all) t≥0.

This says that u is admissible for some initial state x(0) = x0 i at some (hence any) momentt the remaining part ofuis admissible for the current state x(t).

Proof. Obviously,u∈L2iπ+τtu∈L2. By (2.4),y∈L2iCx(t)+Dπ+τtu∈L2,

henceu∈ U(x0)iπ+τtu∈ U(x(t)). ¤

(11)

The (Riccati) operatorP in Lemma 3.2 is called theJ-optimal cost opera- tor:

Denition 3.7. We call P ∈ B(X) the J-optimal cost operator for Σ if, for each x0X, there exists at least one J-optimal controluwithJ(x0, u) =hx0,Px0iX.

Obviously, then the output-FCC holds and P = P. By Lemma 3.5, then J(x0, u) =hx0,Px0iXfor every x0 and everyJ-optimalu, henceP is unique.

We callωA:= inft>0¡

t−1logkAtk¢

the growth rate ofA. By [Sal89, Lemma 2.1], the whole systemΣisω-stable for anyω > ωA.

Now we can derive certain necessary and/or sucient conditions forPand for J-optimal controls. The conditions (3.5) and (3.8) below are integral versions of the standard algebraic Riccati equation (if, e.g.,BorCis bounded, we can dierentiate the integral equations to obtain the algebraic ones; see [Mik02, Sections 9.11&9.7]).

The other, non-standard Riccati equations with parameter r∈ R will be used later below to reduce the optimization of Σ to optimization of another, stable system. The convergence conditions (3.7) and (3.9) can be used to distinguish the stabilizing solution of the Riccati equation from other solutions [Mik02].

Lemma 3.8 (Riccati equations). Assume that theJ-optimal cost operatorP exists.

Let x0, x1Xand r∈R. Letu∈ U(x0) be arbitrary and recall that x:=Ax0+ Bτ u, y:=Cx0+Du.

(a): If uk is aJ-optimal control forxk (k= 0,1), then

hCx1+Du1, J(Cx0+Du0)iL2 =hx1,Px0iX. (3.4) (b): If u is a J-optimal control for x0, then π+τtu is J-optimal for x(t) and

(3.5)(3.11) hold.

hx0,Px0iX = hy, π[0,t)JyiL2+hx(t),Px(t)iX ∀t≥0. (3.5) hy, JyiL2 = h[yx],[J0 2rP0 ] [yx]iL2

r if 0≤r > ωA. (3.6)

hx(t),Px(t)iXt→+∞ 0. (3.7)

(c): The controlu∈ U(x0)isJ-optimal forx0 i (3.8) and (3.9) hold.

hx(t),PBtηiX=−hy, JDtηiL2 ∀t≥0, ηL2loc(R+;U), (3.8) hx(t),PBtηiXt→+∞−→ 0 ∀η∈ U(0). (3.9) (d): We have (3.8)(3.10).

−e−2rthx(t),PBtηiX=hy, π[0,t)JDηiL2

r+ 2rhx, π[0,t)PBτ ηiL2

r

∀t≥0, η L2loc(R+;U). (3.10) (e): We have (3.5)⇔(3.11).

hy, π[0,t)JyiL2+hx(t),Px(t)iX

=h[xy], π[0,t)[J0 2rP0 ] [yx]iL2

r+ e−2rthx(t),Px(t)iX ∀t≥0. (3.11)

(12)

Before the proof, we give here some kind of intuitive explanations for the above Riccati equations. Part (a) says thathy1, Jy0i=hx1,Px0i. IfJ =I, then equation (3.5) says that the minimal costhx0,Px0iX equals the cost until now Rt

0ky(t)k2Ydt plus the minimal cost over the remaining time interval [t,∞). The latter cost equals the minimal costhx(t),Px(t)iXwith initial statex(t). This is often called the principle of dynamic programming (or the principle of optimization).

By (3.5), it also applies to the indenite case (to generalJ =J∈ B(Y)).

Similarly, (3.7) says that while Rt

0ky(t)k2Ydt converges to the minimal cost R

0 ky(t)k2Ydt, the remaining costhx(t),Px(t)iX converges to zero.

One could derive from (3.5) and (3.8) that hx0,Px˜0iX = hx(t),Px(t)i˜ X+ hy, π[0,t)Jyi˜L2 when u is J-optimal for x0 and u˜ ∈ Ux0) (and x˜ := Ax˜0 + Bτu,˜ y˜:=Cx˜0+Du); this is an indenite form of (3.5). Equation (3.8) is the˜ special case of this withx˜0= 0,u˜=η. Similarly, (3.9) says thathx(t),Px(t)i˜ X0;

it is equivalent to the orthogonality conditionhy, JDηiL2 = 0(under (3.8)).

Equation (3.11) follows from (3.5) by partial integration (through (3.19));

the proof of (d) is analogous. Equations (3.10) and (3.11) are actually exactly the equations (3.8) and (3.5) for the system Σ+ (and JP) introduced in Lemma 3.9 below; see its proof.

Proof of Lemma 3.8: (a) To obtain (a), expand the equalityhy1+y0, J(y1+y0)i= hx1+x0,P(x1+x0)i, whereyk :=Cxk+Duk (k= 0,1), and then replacex0 by ix0.

(b)1 Letη∈ U(0). Thenτ−tη∈ U(0), by Lemma 3.6, hence

0 =hJy,Dτ−tηi=hJy, π+τ−tDηi. (3.12) Sinceπ[t,∞)=π+−π[0,t), it follows that

hJπ+τty,Dηi=hJy, τ−tπ+Dηi=hJy, π[t,∞)τ−tDηi=−hJy, π[0,t)τ−tDηi= 0, (3.13) because π[0,t)τ−t =τ−tπ[−t,0) = 0 (since π+ = 0). By (2.4), equation (3.13) says thatπ+τtuisJ-optimal forx(t), hence

hx(t),Px(t)iX=J(x(t), π+τtu) =hπ+τty, Jπ+τtyi=hy, π[t,∞)Jyi (3.14) (use Denition 3.7). This proves (3.7) and (3.5)

2 Claims (3.8)(3.11) follow from (c), (d) and (e) (whose proofs only use 1). Let t +∞ in (3.11) to obtain (3.6) (case r = 0 is trivial; for r > 0 we can use the fact that e−r·x→0, because Σis r−² stable ande−rttukL2

r−² = e−²tkukL2

r−² 0).

(c) By lettingt→+∞in (3.8) for anyη∈ U(0), we get0 =−hy, JDηi, hence if holds. Assume then thatuis J-optimal. Givenη L2((0, t);U), extend it by setting η˜ := π[0,t)η+τ−tu˜opt for some J-optimal ˜uopt for Btη. By Lemma 3.6,

˜

η∈ U(0). By (2.4),π+τt˜=C Btη+Dtu˜opt=: ˜yopt; by (b),π+τtuisJ-optimal

(13)

forx(t). Therefore,

0 =hJy,Dηi˜ =h(π[0,t)+τ−tτtπ[t,∞))Jy,Dηi˜ (3.15)

=[0,t)Jy,Dηi˜ +hJπ+τty,Dτtηi˜ (3.16)

=hy, JDtηi˜ +hCx(t) +Dπ+τtu, Jy˜opti (3.17)

=hy, JDtηi+hx(t),PBtηi, (3.18) where the last equality is from (a). Thus, (3.8) holds (for any η L2([0, t);U), hence for anyη∈L2loc(R+;U), because (3.8) depends onη|[0,t)only).

Lett→+∞to obtain (3.9) (becausehy, JDηi= 0).

(d) The proof is analogous to that of (e) (but simpler) and hence omitted.

(e) If (3.5) holds, thenhx,PxiX=hx0,Px0iX+Rt

0hy, JyiY(t)dt, hence then hx,PxiXACloc &hx,Pxi0X(t) =−hy(t), Jy(t)iY for a.e. t≥0. (3.19) (Here ACloc stands for locally absolutely continuous functions.) Conversely, if (3.19) holds, so does (3.5) too, because its both sides are equal fort= 0.

Similarly, if (3.11) holds, then the facts that (1e−2rt)hx,PxiX(t) =h[yx], π[0,t)[J0 2rP0 ] [yx]iL2

r− hy, π[0,t)JyiL2 =:f(t) (3.20) andf ACloc imply that

hx,PxiXACloc and hx,Pxi0X(t) + 2re−2rthx,PxiX(t) =f0(t) a.e. (3.21) (on (0,∞), hence on [0,∞), because x is continuous (hence hx,PxiX too) and f02re−2r·hx,PxiXL1([0,∞))). Conversely, if (3.19) holds, then the derivatives

of both sides of (3.11) are equal a.e. ¤

The optimal control problem has already been solved for stable systems.

Therefore, we want to replaceAbyA−αto make the system exponentially stable.

To retain the sameJ-optimal cost operatorP, we must add the cost2rhx,PxiL2: Lemma 3.9. Let α∈C be such that0≤r:= Reα > ωA. Then the system

Σ+:=

"

A+ B+

C+ D+

# :=



e−α·A Beα·

e−α·C e−α·Deα·

e−α·A e−α·eα·

, (3.22)

is an exponentially stable WPLS on (U,X,Y×X).

Assume thatP is theJ-optimal cost operator forΣandJ. ThenP is theJP- optimal cost operator forΣ+, where JP := [J0 2rP0 ], and if x0X andu∈ U(x0) isJ-optimal, thene−α·uisJP-optimal forΣ+.

Proof. The rst claim is from [Sta05, Example 2.3.5] or [Mik02, Remark 6.1.9].

Assume thatx0Xand thatuis aJ-optimal control. Setu+:= e−α·u. We have eα·τTu= eα·u(·+T) = e−αTeα(·+T)u(·+T) = e−αTτT(eα·u), (3.23)

(14)

henceB+τTu= e−αTTeα·u, i.e., B+τ = e−α·Bτeα·. Consequently,

x+:=A+x0+B+τ u+= e−α·x and y+:=C+x0+D+u+= e−α·[yx]. (3.24) Therefore, (3.10) equals (3.8) with Σ+ in place of Σ and JP in place of J (and e−α·η in place ofη). By (c1) and (c2), it follows thatu+ isJP-optimal forx0and Σ+.

By (3.24), we havehy+, JPy+iL2=hy, JyiL2

r+hx,2rPxiL2

r. But, by (3.6), this equalshy, JyiL2 =hx0,Px0iX. Sincex0 was arbitrary andu+ wasJP-optimal, the operatorP is theJP-optimal cost operator forΣ+. ¤

4. Minimizing state feedback

In this section we deduce certain properties of minimizing state feedback.

An admissible state-feedback pair[F G ]forΣis calledJ-optimal (resp.J- minimizing) if for anyx0Xthe controlFªx0 isJ-optimal (resp.J-minimizing) forx0.

Now we can establish our main result:

Lemma 4.1. If the output-FCC (1.5) holds, then there exists a (1.10)-minimizing state-feedback pair[F G ] forΣ. The pair is unique modulo (A.6).

Proof. By Lemmata 3.2 and 3.4, there exists anI-optimal cost operatorP forΣ.˜ Fix some αas in Lemma 3.9; then P isJP-optimal for Σ˜+ (which is dened by (3.22) withΣ˜ in place ofΣ).

The output ofΣ˜+contains a copy of the input (because£ 0 I¤

e−α·De˜ α·=I).

Therefore, the system is JP-coercive in terms of [Sta98d]. Consequently, [Sta98d, Lemma 2.5 & Theorem 2.6(i)] imply that there exists aJP-optimal state-feedback pair[F+ G+]forΣ˜+and that theJP-optimal control forΣ˜+is unique for every x0X.

It easily follows that[ F G ]is an admissible state-feedback pair forΣ˜(hence for Σ too), where F := e·αF+, G := e·αG+e−·α, Fª = e·αF and F :=

(I−G+)−1F+ [Mik02, Remark 6.1.9].

By uniqueness and Lemma 3.9, the controlFªx0= e−α·Fx0 must equal uxopt0 for any x0X, hence[F G ] isI-optimal forΣ, i.e., (1.10)-minimizing for˜ Σ, andFª is unique.

SinceFªis unique, the pair[F G ]is unique modulo (A.6), by Lemma A.5.

¤ We could deduce Theorems 1.2 and 1.1 from Lemma 4.1 (and the fact that Fªx0∈ U(x0)), but to avoid unnecessary details, we rst establish one more useful Riccati equation, which we anyway need for Theorem 1.3.

To anyJ-optimal state-feedback pair corresponds a unique signature operator S:

(15)

Lemma 4.2. Let [F G ] be a J-optimal state-feedback pair for Σ. Then Cª and Fª are stable, P =CªJCª is the J-optimal cost operator and there existsS = S∈ B(U)such that for eacht≥0we have

π[0,t)S=N tJNt+BªtPBtª. (4.1) If J =I, thenP, S≥0,N is stable,N Cª= 0 andNN =S.

Recall from (2.6) that M := (I−G)−1 =Gª+I and N := DM = Dª. Note that we identifyS∈ B(U)with the multiplication operatoru7→Su.

Proof. 1 Foru=Fªx0 we havey=Cx0+Du=Cªx0and for anyx0X. But Fªx0∈ U(x0), henceu, y∈L2. Sincex0Xwas arbitrary, the mapsCªandFª

are stable, by Lemma A.1. ButJ(x0, u) =hy, JyiL2 =hx0,CªJCªx0iX, for every x0X, henceP =CªJCª.

By Lemma A.2 (with Σª in place of Σ), we have £

DI

¤M[L2c] = £

NM

¤[L2c] L2, henceM[L2c(R+;U)]⊂ U(0). Consequently,J-optimality implies that

hN ηª, JCªx0iL2=hDMηª, JCªx0iL2= 0 for every ηªL2c(R+;U). (4.2) By 4. of Denition 2.1, hN π+v, JNπui=hNπ+v, JCªBªui= 0 for all u, v L2c. By Lemma A.4, it follows that there exists a uniqueS =S ∈ B(U) such thathN v, JN ui=hv, Sui(u, vL2c). This implies that

π[0,t)N +)JN π[0,t)=π[0,t)S (t0) (4.3) (whereπ+is redundant). From the identities P=CªJCª and

π[t,∞)N π[0,t)=π[t,∞)τ−ttπ[0,t)=τ−tπ+Ntπ[0,t)=τ−tCªBªτtπ[0,t)

(4.4) it follows that

BtªPBtª= (π[t,∞)N π[0,t))[t,∞)N π[0,t)=π[0,t)N +−π[0,t))JN π[0,t). (4.5) Combine this with (4.3) to observe that (4.1) holds for anyt≥0.

2 Assume that J = I. Then P, S 0, by (4.3) and the fact that P = CªCª. From (4.3) we observe thatkN π[0,t)uk22=kS1/2π[0,t)uk22(uL2). Letting t +∞, we observe that kN uk22 =kS1/2uk22 <∞(u L2), by the Monotone Convergence Theorem, i.e.,NN =S. From (4.2) we get thatN JCª = 0. ¤ In the case J =I (or J ≥²I), we obtain from y =Cªx0+Nuª and the above that

J =hy, JyiL2 =hx0,Px0iX+huª, SuªiL2. (4.6) Thus, theJ-optimal cost is then particularly robust with respect to any external disturbanceuªL2(R+;U)in the state-feedback loop.

Given a nonnegative cost function (e.g.,J 0), the minimal costhx(t),Px(t)i vanishes to zero, for any admissibleu:

Viittaukset

LIITTYVÄT TIEDOSTOT

Kirjallisuudesta löytyneiden tulosten mukaan kaksikerroksisilla taloilla rungon vaakasuuntaiset ja lattian pystysuuntaiset värähtelyt ovat keskimäärin noin 1,5-kertaiset ja

Lannan käsittelystä aiheutuvat metaanipäästöt ovat merkitykseltään vähäisempiä kuin kotieläinten ruoansulatuksen päästöt: arvion mukaan noin 4 prosenttia ihmi- sen

• Suoritustasoilmoitus ja CE-merkintä, mahdollinen NorGeoSpec- tai muun kolmannen osapuolen laadunvalvontasertifikaatti sekä NorGeoSpec-tuotemäärittelysertifikaatti tai muu

In chapter eight, The conversational dimension in code- switching between ltalian and dialect in Sicily, Giovanna Alfonzetti tries to find the answer what firnction

By studying the modified DLS instead of the original, we can apply the Riccati equation theory of [6] for systems whose critical feedback operator could be unbounded in the

Jarmo Malinen : Discrete time Riccati equations and invariant subspaces of linear operators; Helsinki University of Technology Institute of Mathematics Research Reports A407

For any breed taking part in the program, it is required that all dogs used for breeding must have a certificate that they have been tested for certain genetic defects and /

Varmistakaa, että kaikki tavoit- teet, laaja-alaiset tavoitteet ja kohderyhmä on huomioitu menetelmiä valittaessa. Ennen kuin jatkatte, varmistakaa, että osaatte vastata