Control Theory - Robust Systems, Theory and Applications
Control Theory - Robust Systems, Theory and Applications
| 0 >0 (B84) eo c Hence Ux0 such, that oO ct 0° (=e) DT} >0 (685) cop ot Pre- and postmultiplying by 1 OQ 0 0 “4 var o 0 "0 vat yields the equivalent condition a [@ 9 cr M(ao)2} 0 fy oD" | >0, where f(a) £ a(n ~ 0) CQ, “Dol Given any ay oy € Sy, and A € (0,1) define & 8 Aa + (1 ~ Aap. Suppose that there exist oy and oy such that May, ey) >Oand Mao.) = 0. Then, forany 4c (.1),AM(ano;) + (-A)M(ay on) > or equivalently, @ 0 ocr 8 flor) +0—AVfalos))! EDT) >0 (887) co =D al where Q = AQ(a) + (1 = A)O(ar) and & = Ag + (1 — A)oy, Note that the function f,(o) = (nm ~ 0) is concave in o, that is, fa(0) > Afylor) + (I — AVfalan). Therefore 0 fel DT (888) cg oD al Qo o "| APPROXIMATESOLUTONS 273, Multiplying both sides by oo 10 (8.89) or yields oo ct 0 flat aDT| >0 (8.90) Cc aD at Using the explicit expression for the solution of the Lyapunov equation (8.82) A'BBT AT it can easily be established that Q, is a convex function of «. Thus O(a@)-! > O and | Hence a € 5, which implies that S, is convex on a. Thus V(a) is quasicon vex on the interval (0 dng). o 0 cr @' oO c s(oyt ad") >| 0 flat avr} >u 91) oD at c oD “a This lemma suggests that an upper bound of Gli can be computed by minimizing V(a). Following the approach in {3} we will define this upper bound as the * norm of G, that is, inf V(a)!? (892) It can easily be shown that the ||. satisfies all the properties of @ norm. Moreover, computing its value only entails solving a LMI optimization prob- lem and the scalar minimization of the function V(a) in the interval (0,4). 8.4.3 Full State Feedback In this section we consider the problem of synthesizing full state feedback controllers that minimize the * norm of the closed-loop system. The main result shows that these controllers are static and can be found by combining @ LMI optimization with a scalar optimization in (0,1)274 £! conROL ‘Theorem 86 Assume shat G has the following state-space realization: A|By Bs GO Da 4893) where the pair (A, By) is stabilizable. Then the following statements are equiv- alent 1, There exists a finite-dimensional, ful state-feedback internally stabilizing LTI controller such that |\Toulle <7. 2. There exists a static contro! law u = Kx such that |Tewle < ¥. 3. There exists a scalar a, 0 3) Suppose that there exists a finite-dimensional LTI controller K that internally stabilizes the system and renders |'Tzn!|. < 7. Assume @ realization of K is given by ‘The corresponding realization of Tey is given by ABD, BC.|B1) [ale Tw=| Be Ac [ol = [a4] (895) C+ D2D_e Di2G 0 clo From Lemma 88, there exist a scalar « and a symmetric matrix Q > 0 such that AQAT ~ 0+ 188" <0 and Cole vt APPROXIMATESOLUTIONS 275, By Schur complements, this is equivalent to [ere 4g |: o co) ear (a1) ” oe ° 897 i ol > 97) Partition Q as (8 ee Oh On Since Q > 0, it follows that Q,, > 0, Multiplying on the left-hand side of (8.96) by [le o} 0 | 0 Fo and on the right by its transpose yields <0 (898) On BBE AQn +BY OuaT eVTBE (a~1)0n where V = DyQy; + C.Of, Pre- and postmultiplying (897) by (4) one obtains : [4 %]>° 99) Finally, multiplying on the left by [!" 6] ‘| oor and on the right by its transpose yields (eocte aa mame les (100) (3+ 2 and 3 1) Suppose there exist Q > 0 and V that satisfy the LMI (894), With the controller given by u = VQ-'x, the closed-loop system be- comes Ay} a (8.101) cto _ [As Bvo" [Bi] _ T= Les paveto |276 C' conTRot Table 8.3, Optimal versus staticeedback “” norm for Example 8.4 « Na 1eh, Taw Gap y z 2 421 465 10% 52 ist 1 304 348, 14% 391 1301 16 301 346 13% 390 1500 300 346 15%, 390 Since Q and V satisy the LMI (8.4), it can easily be verified that 4 Ajat -0+1B,BF <0 and COC} -¥71<0 ‘The proof follows now from Lemma 88 o Nest, we illustrate these results by synthesizing «norm controllers for the system used in Example 84 and comparing their performance against that of the optimal é controller. Table 83 shows a comparison of the optimal # norm corresponding to different values of x versus the ¢” norm achieved by the static optimal s-norm controtier. Its worth noting that the gap 1e- ‘mains constant at about 15%, even when the order ofthe optimal controller approaches co. 8.4.4 All Output Feedback Controllers for ‘Optimal »-Norm Problems In this section we establish necessary and sufficient conditions for the exis- tence of y-suboptimal »-norm output-feedback controllers. ‘Theorem 8.7 Consider a discrete time FDLTI plant G of McMillan degree with a minimal realization: A| BB G=|G)0 Dr (8.102) C:| Dy Da Assume that the pairs (A, Bz) and (A,C3) are stabilizable and detectable, re- spectively. The suboptimal s-norm control problem with parameter y, that i { Toulle < i8 solvable if and only if there exist pairs of symmetric matrices (R,S) in °°" and a real number > 0 such that the following inequalities APPROXIMATESOLUTIONS 277 are feasible: wi (5 1 spat = Re LinBT) wi <0 yfATSA~Q-a)S ATSB, . nl BISA wel s6, [No o RI), [F s}2° ion WY (ol ~ GRCT)W, >0 Sct le, Si]>o where Wi, No, and Ws are any matrices whose columns span the null spaces of BE, (Cz Dz), and Df, respectively. Moreover, the set of y-suboptimal controllers of order k is nonempty if and only if the LMIs (8.103) hold for some R, S, which further satisfy he rank constraint rank (I ~ RS) < k Proof. Assume, without toss of generality, that D2: = 0.7 Proceeding. as in Section 64, given a controller with state-space realization: ul x= [Ale] (e109 combine its parameters into the single variable o-[% 3] As before, in the sequel we use the following shorthand notation: Ay ¢ a): a= [%]: CHIC, Ol: oo ° ok (6 4] e-(2 4] [] ho " Recal hat this assumption can always be removed thoagh a loop shiing transformation Dis= [0 Dys|:278 Cc‘ CONTROL In terms of these variables the closed-loop matrices A,j, By. Ca and Dy corresponding to the controller K can be written as Ay = Au + BOC, By =B, + BOD, (8.105) Cut D2@C, Dy = Dx®Px From Lemma 838 we have that the + norm of the closed-loop transfer function is less than y if there exists X, > 0 in RU" such that exer Och [‘ vrcor Bi) >0 (106 er XA By ai (@-)x, 0 | co (107) Bho al Rewriting these inequalities in terms of the matrices (8.105) yields + PTOQ+OT@TP <0 (8.108) ¥+PTOQ+ OTOP 0 where a Ag By oF Al (a -1)X 0 BE Oe ai. pal Me Srximty Oram “lero 0 0 hoo 0 Dn (8209) os (7 -o)lm, 0 In nm 0 Dh 0 Oem, 0 Dy 0 APPROXIMATESOLUTIONS 279 From Lemma 6.9 it follows that (8.108) is feasible in @ if and only if WEOWe <0: WEbWy <0 Cees (8.110) WIVWe > 0: WEWWo>0 where Wp, Wo, Wp, and Wo denote bases of Ker P, Ker Q, Ker P, and Ker , respectively I can easily be seen that Wp has the form wo 0 0 0 Ona) 0 7a a (an) oe where Ws denotes any basis ofthe null space of BY. Proceeding as in Theo: rem 6.3, partition X, and X>! as y{| AT 0 Be oo alm, Using Schur's complement formula we have that this is equivalent to aoe our (Lara -eeLmay)m, 6m By ct in sia sion, can be eae he neil Wao cObs cout Nb [* SA ~(1-a)S ATSB, BISA atvatse,] lo<0 @B.116)280! contRot where Ny denotes any basis of the null space of [Cy Dy). The condition X, © ROK 5 (is equivalent to RS satisfying Ba Jo and santas) 0 reduces i eo we) Gren, Eo! |meze eum Ic, 0} 0 | where af tu 0 AO as) wm 3 8 ‘After rearranging rows and columns this can be rewritten as, a 0. io me & Y we 0 | igo (8.119) a where : _ [Ww 0 0 M210 Im 0 (120) 00 Ing Using Schur complements once more, this is equivalent to 1 wr oo crcl wo] Hl wore J>o 2) PE adPsee oy ][F Adee Carrying out the block multiplication explicitly yields WE (of -GRCL)Ws> 0 Fo (1m) APPROXIMATESOLUIONS 281 Similarly, from W5¥We > 0, one can obtain that [é S]>o To recap, if X, > 0 of dimension m+ k solves (8.106) and (8.107), then the inequalities (8.103) hold for some R, S, and 0. Conversely, ifthe set (8.103) admits a solution (R, S, c), then a matrix X, >0 satisfying the inequalities (8.106) and (8.107) can be feconstructed from R, S. o Remark Theorem 8.7 implies that if ||7;.||. < y is achieved by some con- troller of order k > n, there exists a controller of order m also rendering Tew. < ¥. I follows that in the output-feedback case the optimal + norm can always be achieved with controllers having the same order as the gener- alized plant The LMl-based approach introduced in Theorem 8.7 is also useful for synthesizing reduced-order controllers. These y-suboptimal controllers of order k 0 (ie. A= AM x £4(R,)), equip ped with the norm l[élia = tll, + lll. A denotes the space of Laplace transforms of elements in A with the norm defined as Ala = Wella where His) is the Laplace transform of A). B(R.) denotes the space of bounded functions on Ry equipped withthe nor fia = sucn COL Tecan be shown (82) that A = BUR.) x C=(R,). Each element (fs) € A" defines a bounded linea functional on A, with its vluc given by (Chan Rio) = Soha) + [COC a Proceeding as in Section 83, the first step toward applying duality isto iden- tify the sets M and M+. To that effect, we need the following continuous time equivalent of Lemma 84. TE CONTINUOUSTIME CASE 283 Lemma810 Assume tht T3(s) © A havm single cers 2i,20:+-y2m Ree) > 0, and no zeros om the jo axis Then given RCA, there exists Oe A such that R=; Q ifand only if R@)=O, 1212.40 (8.124) Let $= {REA:RE)=0. i= 1.n} en with this definition the continuous time C' problem can be reduced (via the Youla parametrization) to Ho = ink IT + Ria (6.126) ‘We will solve this model matching problem by exploiting the dual of Theorem 83. In this case it is lear that X = A and S = M. Moreover, lel File) =Re {e™"} (= Im fe} ¢ sequences Mis given by M! Span {fy fs(68))} Combining these results with the dual of Theorem 83 yields [82] the fol- lowing. (8.127) It can easily be shown that in terms of th ‘Theorem 88 Let T3(s) have m zeros z; in the open right-half plane and no zeros on the jw axis. Then uo = max Seameiriso) + So matmtrit)] (8.128) subject 10 been Lareley+Paadmley] ct veer. — (6129) Furthermore, the following facts hold: (i) the oprimal extremal functional r°(0) equals at only finite points: t,....tm; (i) an optimal solution (s) = Ty(s) — T,(3)Q(s) (0 problem (8.126) always exists; and (iii) the optimal @ has the following form: = S450), HER. finite (8130) and satisfies the following conditions:284 C' CONTROL @ dr) 20. 0) El = Ho (©) Vide = Ta), k=l. un. Remark It can be shown [82] that the constraints (8.129) need to be satisfied only for all < tan, Where fmas is finite and can be determined a priori. Even 50, there are still infinite constraints, and therefore the dual problem is a semi-infinite linear programming problem, Example 8.6 Consider the following unstable, nonminimum phase plant The control objective is to minimize {bly = ||PC(L + PC)". By using the Youla parametrization, this problem can be recast into the form (8.27) with sth (8131) The dual problem ts guven by Ho = Maxey (6.132) subject IN{t,a)| = |e + ape") <1 for allt e Ry (8.133) By differentiating (8.133) it can easily be shown that r(t,<2) achieves local extrema at the points t = 7, where Te me eO (8.134) with the corresponding value r(r,a) = ~a/4ap. Thus problem (81132) is equivalent to do = maxes (6.135) subject 10 lay +091 <1 fax] <* (8.136) a) <0 THE CONTINUOUS TIME CASE 285, The solution to this last problem is given by era -2+2V%), ma 3e2V (137 The constraint (8133) saturates only at two points, 1=0 and t= 7 = ~In(t + V2)/t3-+2v2). Thus the optimal closed-loop system has the form b= bo S(t) + H(t 7) Finally, the values of ¢, and ¢ can be obtained from the interpolation con- straints at § = Land § ~2 yielding i+ v2 342V2 b, a vi sith bull, = a2 = 5.8284. The corresponding controller is given by (5 = 290.7071 — 4.1213 (e=D(-O70T1 + 4.1213 HY 55) Ke 8.5.2. Rational Approximations to the Optimal £, Controller From equation (8.130) It follows thar, unlike in the discrete-time case, the y-optimal controller is irrational even if the plant is rational. Prompted by the difficulty in physically implementing a controller with a nonrational transfer function, a rational approximation method was developed in [41]. In the sequel we will briefly review this result, based on the observation that the C; norm of a continuous time system is (uniformly) upper bounded by the £! norm of its discretization using the forward Euler approximation, Definition 86 Consider the continuous time system G(s). Its Euler opproxi- ‘mating system (EAS) is defined as the following discrete time system: otean= [A I From this definition it is easily seen that we can obtain the EAS of G(s} by the simple variable transformation s = (z ~ 1}/7. that is, ofan =6(5*) ‘On the other hand, for any given 7 we can relate a discrete time system to a continuous system by the inverse transformation z = 1 +75. Itis obvious that the discrete time system is, in fact, the EAS of the continuous time system ‘obtained in this form. (8.138)286 C' contRoL Next, we recall the main result of [41] showing that the £! norm of a stable transfer function is bounded above by the £' norm of its Euler approximating system (EAS). Moreover, this bound can be made arbitrarily tight by taking the parameter + in (8.138) small enough. This result is the basis for the approximation procedure proposed in (41) Theorem 8.9 Consider a continuous time system with rational Laplace trans: form ®(s) and its EAS, (2,1). If © (2,7) is asympuotically stable, then ®(s) is also asymptotically stable and such that HOG) SHOE Cz, 7h Conversely if (s) is asymptotically stable and such that |§P(s)I]v © po, then for all u > we there exists 1° >0 such that for all 0< 7 <7, OF(2,7) is asymptotically stable and such that \}"(z, rl < w- ‘Theorem 8.10 [41] Consider a strictly decreasing sequence 7, +0, and define HS iat, HOGG 7h where (2,7) denotes the closed-loop transfer function. Then the sequence Jus 18 noninereasing and such that a, the optima Cy cost Corollary 83 A suboptimal rational solution to the £; optimal control prob= lem for continuous time systems, with cost arbitrarily close to the optimal cost, can be obtained by solving a discrete time £' optimal control problem for the corresponding EAS. Moreover, if K(z) denotes the optimal &, compensator for the EAS, the suboptimal £, compensator is given by K(rs +1). 1 Example 8.7 °Consider again Example 8.6. For 011 the EAS is given by 120 01 T (8.139) 1A 2.9091 and L = 0.3667 yields GF (2,01) [ ‘and the Youla parametrization with F = 0.9091 0.2909 | 0 0.0909 0 0.8333 |0.3667 0 TOOT 2509170 TT 0 -0.8333/0.8333 0 (8.140) z 00900 | 1.6091 | ~0.3667_~0,2824 yale 1.5909] 0.8333 0.7576 THE CONTINUOUS TIMECASE 287 Hence we have that 16(2 — 1.1) ‘s(t - N2e—) (cn Liye = 12) = Le "= ig ze) Solving for the optimal € compensator yields optimal cost jy = 6.184 and ‘optimal error: o(e) = 1.8414 — 4.342327 ‘The corresponding optimal Q and compensator Keas are given by Q(z) = 2.4309~ 0.05252"! + 0.06072? +0.20892-* + 0.40042-* + 0.654225 + 0.95542°%4+ 1.34582"? + 1.83432-* — 3.28052? (3142) Keas = FiO) Finally, the transformation z = 7s +1 yields the corresponding compensator for the continuous time system. Figure 86 shows |\6\|4 versus 7. Note that as + ~» 0, \blla —* Hor Moreover, the error —+0-as (7) [316]. ‘Acnorm 5 ° 107 107 10) Figure 8.6. Optimcl cost ebtoined using he EAS method versus >288 C' conTROL 8.6 PROBLEMS 1. Consider the uncertain linear system 2(k +1) = [Ay + w(K)DE) x(k) A,=05+[ 1 HE o- [i] E={1 oO) G43) Jw(k)| < Wem where w(K) represents memoryless (arbitrarily fast) time varying uncer tainty (a) Use Theorem 8.2 to show that if wma. = 0.6 then the system (8.143) is BIBO stable (b) Define G(z) = E(zl ~ A,)-"D and recall from Problem 6.10 that if Yan S (Glen)! then there exist P > 0 such that V(x) = x" Px is a Lyapunov function for the system (8.143), Use this result to reduce the conservatism in the estimate of the maximum value of Way guar anteeing stability (© Consider now the following polyhedral Lyapunov function: Vp(xe) = max | P82 (8.144) me: where oo4 1 1 0 1 ee ies (8.145) 1 0.3390. oon. ‘ Show that if wus < 0.975 then V,(x) is a Lyapunov function for the system (8.143). In fact, using similar techniques it can be shown that (8.143) is stable if and only if wax < 1. (See [42] for a thorough treat- ment of polyhedral Lyapunov functions). 2 The purpose of this problem is to investigate ¢" controt problems with zeros on the unit crcl (317). Consider the SISO é" model-matching prob- lem: 0 = inf, T(z) ~ Ta(2)QIh Ho = jaf ITi(2) ~ Ta(2)h ‘Assume that T,(2) has two complex conjugate zeros z; = e!® and 22 = e-!# 6 the unit circle and no other unstable zeros. (2) Show that bo = inf Ti — Tr lh on = max [arRe{ Fi(21)} + atm{y(z2)}] mons 289 subject |r = fv onthe) ~ sink] <1 £=0.1,2 (8.147) (by Show that for the case @ = yx where y is a rational number the infimum in the primal problem is always achieved and the optimization problem can be solved via finite-dimensional linear programming. (©) Show that if @ = yx where 7 isan irrational number, there exist values ‘of 7; such thatthe infimum in the primal problem can not be achieved (a) Show that when 7 is irrational, the set of «satisfying the constraints (6.147) 5, does not depend on the exact value of y (i. Sy is the same for all irrational y). Denote this as 8}. Similarly, let Si (y) denote the constrained area of the dual problem for the case where y = m/n Show that $#(y) does not depend on m and that S89) 28h, for any of rational y (6.148) (e) Use these results to conclude that the optimal approximation error j., does not change continuously when the zeros of T(z) move continu- ously on the unit circle. 3. Consider the pitch axis control of a forward-swept wing X29 aircraft [95]. ‘A simplified model of the plant is given by: (s+3)__ 20 (5-26) PO) = By 106 =6) G+ 20) (5 +26) (8.149) We are interested in synthesizing a controller to minimize (|W(s)S(s)|) where 5 denotes the closed-loop sensitivity function and where (+1) WO) = Bor (8.150) ‘Assume that the plant is discretized via a zero-order hold at the inputs ‘and sampling at T, = 4 (a) Design a controller to minimize ||W,(z)S(z)Jh where the weighting function W,(z) is obtained from W/(s) using a bilinear transformation with 7, = 30-1 (b) Use a CAD package to set up a simulation diagram where the con- tinuous time plant (8.149) is connected via sample- and hold-devices to the discrete-time controiler found in part (b). Find the impulse re- sponse of the resulting sampled-data system and compare it against the discrete impulse response of the system obtained by connecting the discretized version of (8.149) and the controller found in (b),290! contRoL 4. Consider the following discrete time plant P: A| BB P=|G]Dn Dn | (151) | Du 0 where (A, B:) and (A2, C2) are stabilizable and detectable. (a) Proceeding as in Section 8.4.4 show that the suboptimal «norm control problem with parameter y is solvable, if and only if, there exist pairs of symmetric matrices (R, S) in R"*" and a real number o > Oso that the following inequalities are feasible (66} RAT — Re TBBT) W) <0 7 [ATSA —(1-a@)S ATSB, Ng BISA at +B7sB,|No< RI > [i see (9152) WYO) [1-ZGRCT Du | |W, Ol Lg or Di, (-ollo 7 7 [os ce, Cyn MDE oa acorn) N02 where W, No, Wo ate any matrices whose columns form bases of the null spaces of BY, (Cz D2], and DF, respectively (©) Show that if Dy, =0 then the inequalities (8.152) reduce to the LMIs (8.103) ‘5. Consider the optimal closed-loop system found in Problem 3. (2) Find the worst ease peak of the control action when the input to the system isa persistent signal wth magnitude bounded by one. What is the worstease input signal? (©) The peak control action can be reduced by synthesizing a controller K(z) that minimizes WaG)Ste) | 01K G)St2) This leads to a two-block MIMO ¢' problem. Find a suboptimal solution to this problem using the inequalities (8.152) and compare its performance against that of the controller found in Problem 3 PROBLEMS 291 6. In order to obtain a rational controller, the X29 plant in Problem 3 was discretized using sample-and-hold elements, and the resulting discrete- time problem solved using €! theory. However, this approach could result in significant intersampling ripple. To avoid this, use the EAS technique to synthesize a suboptimal, continuous-time rational £' controller for the plant (8.149) and the weight (8.150),MODEL ORDER REDUCTION 9.1 INTRODUCTION ‘Computers play an important role in the modeling, analysis, and design of ‘modern control systems. Thus, it is necessary to understand not only theit advantages but also the new problems that appear with their use. These problems will be considered in this chapter. In particular we will concentrate ‘on problems that concern state-space realizations of FDLTI systems, since this is how they are represented in a computer. I is easy to understand in other fields the problems to which we are referring. Take, for example, matrix inversion. A nonsingular matrix can be theoretically well defined by means of its detéemmimant, However, front a numerical point of view the situation is not as clear, That is, a badly condi tioned matrix can be cheoretically nonsingular but numerically singular as will be seen in the next example. This is a simple consequence of using truncated rational numbers in a computer instead of real numbers, as in the supporting theory, Example 9.1 Consider the following matrix: Mt 1 A= eR” (1) 1 00.05 Its eigenvalues, hy =---= Ay =—1, and its determinant, det(A) = (-1)", indicate that in theory its a nonsingular matrix. Nevertheless, if we multiply 293294 MODEL ORDER REDUCTION it by vector u = [11/2 1/2")? we obtain Au = —2'-"|L yy Ay the size of A increases, ||Aullz tends 10 zero. This implies thar oumeri cally this is a “near singular” mairix. The latter is confirmed by the fact that aA) x2" Therefore a theorctca zero may be represented computationally by avery small (nonzero) number that depends onthe existing computing technology ‘The problem canbe solved by quantifying the meaning of “small” and “large” in a computational numerical sense. In this example, its well Known that this can be accomplished using the singular valves ofthe matrix. Te smallest singular from singularity depending on this value, a quantitative property. Instead, theoretically, it can be either singular or nonsingular, a qualitative distinction. The same problem appears when we work with FDLTI dynamical sys tems, represented in a computer by a state-space realization. The basic the- ‘ory of minimal realizations (165, 166)) suffers from the problem of structural instability. This means that “small” numerical perturbations can change the properties of a system. The properties we are interested in are controllabil- ity, observabilty, stabilizabilty, and detectability. As in the case of matrix singularity, measures of each of these properties can be computed, using the singular values of specific matrices (see (67, 108, 171, 202, 223, 269, 311) {As in the matrix inversion case, each of these values measures the minimal numerical perturbation that cancels a particular propery. For example, py provides the smallest complex perturbation (8, 8B) on the stabilizable sy- tem (A, B) such that it sno longer stabilizable. It ean be computed as follows (269) os inf @{[(A~s1) B)} (92) ‘This and other similar measures quantify a property of a system but do not distinguish separately each of its states. In this chapter, we present a way to distinguish the states of a given system according to a property of interest. For example, we might want to order the states according to their controllability, with the first state being the “most” controllable and the last one the “least” controllable. This provides a real- ization useful when the less controllable states need to be eliminated, From all possible realizations of a certain system whose order is fixed, itis often convenient to select the one having its states ordered by their relative com trollability and observability. This is the case when we need to verify how “minimal” a realization is from the numerical point of view, or when pet forming model reduction, The latter problem is of fundamental practical importance. The following is a list, by no means extensive, of the need for model order reduction. GEOMETRY OF STATESPACE REALIZATIONS 295 ‘# The series. parallel, or feedback interconnection of systems (see Appendix B). even when these are represented by state-space minimal realizations, may produce a nonminimal one, ‘© Many design methodologies (e.g. u-synthesis, € optimal control), result, in high-order controllers, + The state-space realizations of time-delayed systems or systems described by partial differential equations are infinite-dimensional, There ‘are many methods that generate (high-order) finite-dimensional approximations. Thus sometimes a further model reduction step needs to be performed, ‘¢ In many cases. when working with a high-dimensional model, the numerical errors may distort the analysis or simulation results ‘* Because of computation time, limited memory, or simplification of the controller hardware, itis important to have a small number of states in its realization, In all the above cases, itis also important to have a measure of the error between the nominal and reduced-order model, which ean be considered as ‘model uncertainty in a robust control context 9.2 GEOMETRY OF STATE-SPACE REALIZATIONS 9.2.1. Controllable/Unobservable Spaces Consider the following FDLTI realization: 8) = Ax() 4 Bul), x) Py 3) ye) = Cx (04) Define in R" the controllable and unobservable subspaces as follows ([61). Definition 91 The controllable subspace Xe contains all final states corre sponding to all piecewise continuous inputs w €U and zero taital condiiions, Definition 92 The unobservable subspace Xz contains all initial condivions 4 € RY, which produce null ouiputs, y@) 7 Itcan be proved that Xc = R(e*’B) = R(W,)-and Xz = N(Ce**) = N(W,) (see [61]. Lemma 1, page 75). Here W, and W, ate the controllability and observability Gramians, respectively: This has been fepresenied graphically in Figure9.1,296 MODEL ORDER REDUCTION Re Xo = Re) “BE u ett {0} Xy=N(CeM) Figure 9.1. Contiollable and unobservable spaces ‘The state space R" can be divided as follows: R" = Xco@Xéo 05) Xco = Xen (Xe Xp) 06) ‘with Xco the minimalsubspace needed to describe the input-output behavior of the system. Therefore each state can be represented as the sum of two orthogonal vectors—one in Xco, the other in Xzo, Note that we have not define (Xe NXp)!. This is due 10 the fat pservable sats is no a vector space. Take, for example, two observable initial states x,(0) and ‘xp(0) with Zero input (u(r) = 0) and outputs y(r) and —y(1), respectively. The sum of both initial conditions xy = x4(0) + x9(0) gives a null output and, as @ consequence, is not observable. In general, the sum of observable states of the form x4(0) = x,+m and xp(0) = ~x, + m, where y(t) = Cex, 40 and mim € Xqu may not be observable. ‘The subspace X 4 (XcMXq)* contains the zero element as well as observable states_Italso contains unobservable and wwe prove next. Divide the space R" in X and X' and consider nonzero states %y EX* and xp CX. For any state x — xq +xp, we have Cet!x = CeM'xp. If Ce*"xy 40, then it is observable. On the other hand, if Ce*'xp = 0, then xp € Xz and xp ¢ Xc because it cannot be in X and X simultaneously un- less xp ~ 0. Therefore all unobservable states in X are also uncontrollable Asa consequence Xco ineludes only controllable and observable sates ‘The subspace Xco contains all states of a minimal realization of the system and therefore completely characterizes the input-output behavior from the state-space standpoint. Related with this input-output behavior, we can also describe both subspaces X'c and Xin terms of input signals, as follows ontralable sates 25” [GEOMETRY OF STATESPACE REALIZATIONS 297 # Define 21(1), # = 1,...,p, as the state response to input vectors u(t) = e7B(0), = 1,...,p, where ey is the Fth column of the identity matrix in RP? and 5(:) is Dirac’s delta distribution, Then Xe can be defined as the smallest dimension subspace that contains the range of all outputs, that is, RUX(®); > 0), with xo= [kOe] a) Define»), €= I... as the output response to iil condition vectors x(t) = er, € = 1,...,m, with ey the fth column of the identity matrix in R"*". Then X,; can be defined as the largest dimension sub- space contained in the null space of all outputs, that is, N[Y (0); « > O}, with yo = po yo 8) 9.2.2 Principal Components In this subsection we describe the geometrical characteristics of the state space in terms of the input-to-state transformation and state-to-output trans- formation. For simplicity we define an operator and its corresponding, Gramian, Which can be interpreted as any of the latter transformations. In particular, the input-to-state transformation is related to the controllability Gramian and_ uutput transformation to the observability Gramian, ‘continuous Tunction F : RR" and its associated we [Faron 9) ‘The above matrix is symmetric and positive semidefinite (prove); therefore it has nonnegative real eigenvalues. As a consequence we can select the following set of orthonormal eigenvectors (Problem 1) W = v3tvT (9.10) oe Ba (uy 0 where the eigenvalues have been ordered for convenience (o > @ > += > o) and V is a unitary matrix whose columns are the corresponding eigen: vectors. To clarify, we present next specific examples of the above and the State-space interpretation of the corresponding298 MODEL ORDER REDUCTION Example 92 Consider the problem of reaching a final controllable) state 2) 20 al Lime 1, =0, starting from zero initial conditions at ty» —>e with a ‘minimum energy input signal u © £x(-~D0,0) iin, Lllssth that (0) ~ ay 912) This problem is similar to the following leas squares one ina finite dimensional space: min ula subject to 2 = Mue BY (ony sin ll sub ) with M/€ R™" and m > n. Denote by M* the adjoint operator of M (in this case, the transpose). Using the geometrical fact that R can be separated into the direct sum of R(M") and N(M), it is clear that the minimum norm typ, should be in the range of M". The fact that B" can he separated into the direct sum of R(M) and N(M*) implies tap = MC, 89) = M27, x, CRM), sy CNM) (9.18) y= Muy = MM, (915) Its easy to prove that R(M) = R(MM") as follows If y C RIMM") + yc R(M) Because y = MM "xy = Maz for x2 = Mn. On the other hand, ify R(MM*) + y © N(MM") therefore MM"y =0. For 2 = M"y we obtain 2 € N(M) and z ¢ ROM"), thavis, 2 =O. This in urn implies that y belongs tothe Kernel of M* and as a consequence is notin the range of M. When M has full rank, this fact combined with equation (9.15) yields: Hayy = MMM bay (9.16) leo = x8 (MM xa (9.17) where M>(MM")"" is known asthe pseudoinverse 0 “taney Sean eat be oben oho ihe input space is L(-,0) (infinite-dimensional, although the operator is different. In this case we define the infinite-dimensional operator M as soa) = MW © fe Buln)de (9.18) = [er ou-nar (919) Therefore in (9.9) F(t) = eB and all eigenvalues of A should be in the open lef-half plane for W to exist. In this case W is exactly the controllability GEOMETRY OF STATESPACE REALIZATIONS 299 Gramian W.. The minimum energy input is Hop) = BIE WEN ay (920) ney Ieopil = xg We xo (921) In this case BT e""W-" is the pseudoinverse of the operator M. It is clear from the above equation that if W-" is “large” (a [W.| “small”), certain states will be reached only by high-energy inputs, being usp, the optimal one. This gives a quantitative measure of controllability, since the most controllable states ‘an be reached with the least input energy, and vice versa. In the limit, as W, tends to singularity, [Wd ~ 0 (922) = [luapll> + 00 (923) = 49 > noncontrollable (924) Example 9.3 The dual problem is as follows. Using the output equation (9.4) with u(t) = 0, € > 0, we want to find the inital condition xo that mavimizes the energy of the output: xe vo Ax(0), x(0) = x0 apy be a ar [20020 (925) ‘In this case the transformation can be defined as y(t) © M(x) = CeA'x. Using duality and the solution of the last example, we obtain (026) As before, a quantitative measure of observability can be obtained from the size” of the Gramian W,, If itis near singularity (|We] 0), rio matter how large certain nil states may Be, He oul energy wil edt zero, that fs, this wal é aaa a Next we represent the function F(t) in terms of the orthonormal basis formed by the eigenvectors of the Gramian W. Feo = Soro 20 fy © oT FO (9.28)300 sooeL onoER DUCTION Vector functions v;f! (0) are called the principal components and fF(t) are defined Ue urcional somone 909) The Totowing property's Uefa aS an interpretation of the optimal projection and “energy” distribution of this operator (for simplicity we use fp = 0 from now on). Lemma 9.1 The set {f\(0),---»falt)} 18 an orthogonal basis of functions. The norm of each element of the set is elated tothe singular values of the Gramian W, as follows a hh (9.29) If + 00 and the following integral converges, then it coincides with the norm PIE: - trace [FTO] de = (930) For each 1, the argument of the integral is defined as the Frobenius matrix norm |IF (OI Proof. ronoa otf ror wade of VE2VT wy, of of, oe {a wie) oan which proves (9.29). Since the trace of a matrix is the sum of its eigenvalues, we obtain trace(W) = FIR (932) [fs [Porro] ar ‘The above result can be interpreted a follows. The total the operator Fl), that is, [Fis distributed spacially accord éalions given by THE constant 3 uj, The relative import ‘each direction is quantified by the eigenvalues a. Therefore the Gramian W. (GEOMETRY OF STATESPACE REALIZATIONS 301 ‘through its eigenvectors and values plays an important role in the ‘geometrical interpretation of F(¢): in particular, as will be seen next, in the description of the controllable and unobservable subspaces and also in the input-output priority of each state (see Section 9.4) ‘To this end define the following subspace of R Se = ROW) (9.33) which can be spanned by vectors v; corresponding to all g; > 0. Assume Without loss of generality, that the whole space can be spanned, that is, S- = R*. Then the optimal projection satisfies the following property. Lemma 92 Take 1 <1 0) > --- > oy > 0 in (9.11). Then there exists q = VE~!p such that 2 = Wa = V3p € §, with [pl = 1. Therefore the input signal a) = FT ng a vey (a1) =a [fl Ferner od =4"Wa [pl = 1 (9.42) takes the state 2(0) = 0 to 2(s) = Sos. In Example 92 we have proved that the above input u(t) is the minimum ‘energy one! satisfying 2 = fy F( ~ u(t) dr. Due to linearity and jus = 1. this means that there is no juz <1 that generates a final state outside the ‘boundary of the hyperellipsoid S: therefore this boundary is S. 2 € S (Problem 2). This proves that From the above results we can note the following: Notes ri.then q = W-!2 = 9;"'v, which in turn implies Tr=du _ filr-1) un =F (043) ‘Therefore the set {a)v;} generates the reachable states in R" and the functions /,} generate the orthonormal set of corresponding input signals. 2. When F(1) = eB. the boundary represents the reachable'states start ing from x(0) = Owith normalized inp li» ~ 1. The boundary of the hyperellipsoid in the dual case—F(t) = e*"'CT—represents the initial "Take the final state #45 4p and W = W. Then Ill? = 4 Wig = af Wy tay = TWH (GEOMETRY OF STATESPACEREAUZATIONS 303 conditions from which normalized outputs [yj =1 canbe obtained, with zero input. In each of these cases, the values (04 /0,,) reflect the condition numb of control and observation, respectively. By this we mean the relati ontrollability and’ observability with Tespect to all directions 3. The specific coordinates and magnitudes for controllability and obsery- ability depend on the Gramians. which in turn depend on the specific set of selected states; that is they are not invariant under state trans- formations. Therefore it is important to select a particular state trans. formation that can be useful to reflect the relative controllability and observability “measures” simultaneously. To this end, note the following fact. A state transformation changes the system matrices as follows: A|B Tr TAT"! | TB LEE] ow ‘The Gramians Mo) ate changed to (TW, TT, T-TW,T-), respec- tively: Except for special cases (unitary T), the Gramian eigenvalues will ange_under sate tansfomatons, Neverteles, ie /, femain unchanged, since the transformed product is TW.WaT~! Example94 Consider the following realization of the system 1/(s + 2)(s + 1): = [? S]ao [Se] (045) YQ) = [10% 10-4] 20) (046) For this realization, the Gramians are [Osx g mem OO paste] om 025x108 = [PAT os do] os with similar condition numbers in the order of fe = ky ~ 10". It is clear that for this realization x is “almost” uncontrollable and x “almost” unobservable. With the following state transformation, ‘ wo 6.49 10 =['% Jao304 MODEL ORDER REDUCTION ve obtain the following realization iW = [e ‘Jim [ 4] (950) yo = (1 10) (ast) with a better “balance” between both condition numbers, which have decreased (0 R. = Ry = 38. Clearly, this new realization has much better numerical prop: erties. From the geometrical point of view, the highly eccentric ellipsoids of controllability and observabilty in R? have now a more “spherical” descrip: tion 9.3 HANKEL SINGULAR VALUES 9.3.1 Continuous Systems As indicated in the last section, itis useful to represent the dynamic system in terms of a particular set of states, which “balances” the quantitative measures of controllability and observability. In this way it is easy to order the states from the most controllable observable to the least one. This is useful from the point of view of model reduction because this order provides the relative priority of each state in terms of its input-output effect ‘The quantitative measure of relative controllability observability should depend now on both Gramians W, and W,. In particular, these measures should be a property of the system, that is, they should be independent of the specific realization selected, According to the notes mentioned in the last section, the eigenvalues of the product of Gramians are invariant under state transformations. This is coherent with the definition of the Hankel singular values of Chapter 6 (Definition 6.2), as the square roots of the eigenvalues of the product WW, (or W,W.), indicated as (wWees = (a) en =the (932) For convenience, they are usually ordered as follows of! > off > ---> of! In (209) these values are called second-order modes. ‘They Fepresent the fundamental measures of gain and complexity of a LTI system ((132)). The preferred state realization, which we seek for, should make both Gramians ‘equal and diagonal (see Section 9.4) such that of 0 w= We (053) oo HANKELSINGULAR VALUES 305 Hankel Operator As mentioned in Section 62.1, the Hankel singular vat interpreted in terms of the singular values of the Mankel operator fin Detintian-B1-This operator provides an intresting Tlation between the iaput-outpat pains of a system and the controllability and ob- servability of its states. I is easly seen that Definition 6.1 is equivalent to: y(Q) =Tev {Rrcemnnude, 130 cE 1<0, where v € £(0, 0) ‘The interpretation of this operator in terms of inputs and outputs is as fol- lows, Take the input u(t) = o(-1), ¢ < 0, therefore the input-output transfor. ‘mation can be performed in terms of the Hankel operator: y(¢) = (F@v)(t) (g + 4)(0). This is the convolution of a system whose input is defined for neg- ative times (in £3(—00,0)) and its output for positive times. The final state at r= 0 for input u(t) is x) = f. eM Bus) dz (955) 1 corresponds (0 the initial state for the unforced system whose output is (0), > 0. This operator transforms past inputs into future outputs through the state (0). ‘The Hankel singular values can be defined either way, as in Chapter 6 (Definition 62) or in terms of the Hankel operator. Furthermore, it can be shown that the Smith-MeMillan order of a system is equal to the number of positive Hankel singular values. Recall that the Hankel norm of a system G(s) has been defined (Defi tion 6.3) as the largest singular value of its corresponding Hankel operator 1G(S)lln = @ ['c}. The Hankel norm can be interpreted as an induced norm between past inputs and future outputs [see equation (6.19)}: max WPale* wdla welieol Tle NGC) = (056) which also equals /p(W.W.)- Note that the Hankel operator, its singular values, and the Hankel norm are all independent of the constant D term of the realization, This is the reason why, without loss of generality, we have considered only strictly proper systems Next, we present the results for discrete time systems. In turns out that if the latter is obtained as a bilinear transformation of a continuous system, the Hankel singular values and Hankel norm remain unchanged.306 MODEL ORDER REDUCTION 9.3.2 Discrete Systems Consider the discrete time, linear, shift invariant, asymptotically stable system: AUK T] = AgakT) + Bul kT) (951) WAT) = Cyx(kT) + DgulkT) (9.58) Since [A,(Ay)| <1, Vi = 1,...41, the Gramians can be defined as we © atauat (a4)! (959) 2 (As)! cfcuat (960) Both are obtained also as solutions ofthe discrete Lyapunov equations we — A,WiAl — B,BT = 0 (961) Wy -AlWiAy—C}C, = 0 (9.62) respectively. Similar to the continuous case, the Hankel singular values are defined as follows. Definition 93 Consider the asymptotically stable system Gaz) = [ eit | 063) with controllability and observability Gramians Wand Wi, respectively. The set of Hankel singular values of G(s) are the square roots of the eigenvalues of the product WW, indicated as (wewan, = (fo, v.20 (964) For convenience they are usually ordered as follows aff > af >. > of! Next, we show that the equivalent continuous system, obtaines bilinear transform of the above, preserves the controllability and observability Gramians. Therefore the Hankel operator, singular values, and norm ave preserved as well. Also, due to the fact that the McMillan degree can be defined in terms of the Hankel singular values, the order of both discrete and continuous systems is the same WANKEL SINGULAR VALUES 307 ‘The bilinear function is (965) (9.66) which maps the interior of the unit disk |z| < 1 into the open left-half plane oi) = Gx(e= 44) a and for the state-space realization matrices (Problem 3) A= (4 Ag Ag—D) (9.68) B= VUI4 Ag 'By (9.69) C = VIC {I+ AG (9:70) D = Dy Gilde Agy'Bg (on) Next, pre-and postmltiply the continuous controllability Lyapunov equa- tion by (J+ Ay) and (1+ Af) as follows (14+ AAW + WAT + BBTVI+ AT on) By substituting (9.68) and (9,69) we obtain equation (9.61) (Problem 3). Its unique solution therefore coincides with the continuous case, that is, We = W4, Similarly, we can prove that W, = W. Furthermore, if we evaluate both systems in the boundary of their re- spective stability regions (|z| = 1 and s = jw), we obtain from the bilinear ofrm(4)] on This indicates that both frequency responses coincide in @ € (0, 7), and there fore their respective infinity norms. Due to the fact that the McMillan degree ccan be defined in terms of the Hankel singular values, the order of both dis crete and continuous systems is the same, Gale" 2 For simplicity consider the sa308 MODEL ORDER REDUCTION From the above, we conclude that any model approximation of systems G(s) and Gq(2) of arder n by others G,(s) and G¥(z) of order r 0) can be obtained by combining the above equations: (0) (1) HT)) = KoKe | M27) (979) ae As before, the Hankel singular values of = 0,(H) provide a measure of the relative controllability observability of each state x(0). Geomet- rically, we can represent the above equation as a hyperellipsoid with each axis representing an eigenvector of Hf with a relative magnitude given by off. Therefore the relative controllability /observability of x(0) depends on its direction relative to the principal axes of the hyper ellipsoid. Finally, we define the Hankel norm of a diserete time system. Definition 95 The Hankel norm of a stable system Ges) is the largest singu lar value of its corresponding Hankel operator: \Gals)\n = GH) (or equiv alently the largest Hankel singular value of")310 MODEL ORDER REDUCTION 9.4 MODEL REDUCTION 9. 1 Introduction The purpose ofthis chaper is to provide @ model reduction procedure sich that, given a model Gl) of MatMilan depee n, we btan a reduced ——r——e——e T_T Pen. Woate ako interested in aving 9 hound on the appx G(s) — G.(s)la in terms of a meaningful norm a. By meaningful we mean that the specific norm chosen should have relevance from the modeling and contel pint of view "ake for example, the Hankel norm defined inthe latter section It po r-rel mmentl or model reduction, is main Objective bel relevant sates from the input-OUIput perspecive. that i. iy” oF & fealization from the numerical point of view. a r—t——r—r— be explained quite imply, we ise a reducedorder mode! o design acon —— se _ it fits naturally as (global) dynamic uncertainty. As we have seen Fnchapters2 and the obust stay coniton for his ype of uncertain brebocedon te M-norm in lace of the rests weave so far (Gee Lemma 92) the wuncation of states that correspond to small Hankel singular sales proves an optimal projection in the norm Tn th section me describe the balanced-truncation model reduction a )— Clr ———§__ Cro in terms ofthe infinity and Hankel norms Although the procedure oes not sve the eptimal problem, min||G(s) — Gy(sNlas @ = Hyco (980) it guarantees an approximation error that is very close to the optimal, with a smaller computational effort. The optimal Hankel norm approximation prob- lem has been solved in (132). An extension of the balanced-truncation method ‘and the optimal Hankel norm approximation applied to the approximation Of infinite dimensional systems can be found in [79] 9.4.2, Hankel Operator Reduction For simplicity, consider the discrete time system described by matrices (Ag, By, Ca,Dq) with Hankel operator H of rank n. Due to the fact that the rank of the latter is defined as the order of the model. we can state the MODEL REDUCTION 317 ‘model order reduction problem in terms of this operator as follows |H — Helle (ai) where H, is a Hankel operator of rank r. This problem applied to finite dimensional matrices was one of the early uses of the singular value decom- position. The solution is due to Mirsky ((208)) and is easy to prove. Theorem 9.1 Take MeO" with a singular value decomposition M URV*, 3 = diaglay, =. -50] such that 0) > --- > om >0. Also, con: sider a unitarily invariant matrix norm, that is, |(M lly = ||UM Ilo = IM Ul with unitary U. Then fon 0 int {\\a# — atl, , ROM) aXe”) (993) and rearranging, we obtain K(WeW,) = 2 (994) Be) gclT)Ko(T) (0.95) (We)314 MODEL ORDER REDUCTION where we have used in (994) the symmetry of W, and W,, to obtain 03 (W.W,) = 4 (W,W2W,) = AZ (WW, ). Furthermore, we have: max{xe(T),o(T)] > Yre(Fwo(T) = (3) (996) ‘Therefore the minimum is achieved by T = T.y for which x¢(Tin) = wo(Tis) = (of! jaf!) a ‘This property of balanced realizations indicates that they offer the best comproinise in terms of conditioning of controllability and abservability mea sures Graphically, this means that they Tend to make both controllability and 0 and ~2BB" z, = WB" |} <0, we obtain Re(A.) <0. (9.99) MODEL REDUCTION 315 If (A, B) is not controllable, using the Hautus test we have (9.100) and using (998), we obtain Re(A,) = 05 that i, Ais unstable with imaginary unstable eigenvalues. Now assume A is nt stable and agai, using (295) all unstable eigenvalues are located inthe jo axis, say, at A= yoy-From (9.98) we obtain |B" sy], = 0 which implies (4, 8) is not completly controllable. a By duality we can prove a similar result for observability using the corre- sponding Lyapunov equation. Next, we prove the main result Theorem 92 Consider a minimal, asymptotically stable balanced realization with states x(1) € RY, ¢ > 0: (9.101) ow = L244 with Gramians W. = Wy =% = diag (of! into 1wo subvectors as follows oft). Divide the sate veetor 10)= [20] wie nt emt i= 12, 20 such that ny +nq =n, and partion the system and Gramians accordingly An Av By a-[4 4]: 5=(3] c=[a a] (0.102) xy 0 eo (9.103) Assume the Gramians of both subsystems do not share common diagonal ‘elements. Under these conditions both realizations, [He ‘are minimal, internally balanced, and asymptotically stable Gis) 2 (9.108) Proof. Separating both Lyapunov equations into two subsystems, we obtain for the first block-diagonal entries Bay +AhE+ CFO = 0 (2.05) An Si +3,Ah + BBP = 0 (0.106) Br. to, a8"316 MODEL ORDER REDUCTION and for the lower-eft block entries AB, +EAh+ By = 0 (@.107) ALE, +¥adn+ CFC = 0 (0.108) Both (9.105) and (9.106) imply that subsystem (Ay), B1,C;) is balanced, ‘Assume it is not stable; therefore by the last lemma Ay; has at least one imaginary egenave sy ay yo, Then (An poh¥ =0 ox) V" (Ale sot) <0 ox) where the columns of V span the right mullspace of (Ay) ~ jo). In (9.105) and (9,106) add and subtract jw. Then pre- and postmultiply the former by V* and V, respectively, and the latter by V*E) and 5\V, respectively, Then: VIS (Ay —s01)V + VAT, + J01)EV =-Viclov =0 ot) 2CV 205 (Ah + 0D3,V (0.112) VeE(An ~0DEV+VSHAT + 10)EV (0.13) S1B\BTEAV =0 (9.14) 3 BIZV = 0 (Ay ~ jo)2V =0 (9.115) Therefore the columns of 32V are included in the nullspace spanned by V and we can always find a matrix 3, such that 32V = VS}. In particular, V can be selected so that Eis diagonal with a subset of the diagonal entries of. Now postmultiply (9.107) by 3,V and (9.108) by V to get An3V +E2AR3V = 0 (9.116) AREV +2,Anv = 0 on) ‘which in turn implies BAnv (8) Because £, and ¥; do not share common entries, it follows that A2)V ‘Therefore rearranging equation (9.118) and the eigenvalue equation for Ai1. MODELREDUCTON 317 Ay An [a a2] [5]-[0] one which contradicts the stability of the original system, Therefore (Aj), B,C) is asymptotically stable. Similarly, we can prove the stability and balance of system (Azo, Bp, Cy) as well. Using the previous lemma, both subsystems are controllable and observable, therefore minimal o The above result is useful as a model reduction method. The objective is to separate the full model G(s) into two subsystems G,(s) and Gy-r(3). Cate must be taken to order the states according fo Their Hankel singular Values, such that of > --a!" > off, ---> a In this way, the truncated model with the first r states G(s) contams the str Hankel singular values, Therefore the mosf controllable and observable states are contained in this reduced system, which are more relevant in an input-output sense. The assumption that both subsystems should not share common Han- kel singular values is extical. The following example (240]) indicates that, otherwise, one of the subsystems could be uncontrollable, unobservable, or unstable. Consider the balanced system xo= [22 Eas [ha w= [1 0] xt) Both Hankel singular values are 0.5 and, clearly, the subsystem (0,0,0) is neither controllable, nor observable, nor stable. From the practical point of view it is not a problem, due to the fact that, generically, any matrix has distinct eigenvalues, Also, for model reduction, it is important to eliminate ‘the subsystem that has very small Hankel singular values (a/ > of,), which is not this case. This last point is important when considering tie model reduction error. It can be proved ({132)) that the approximation error for the balanced and truncated method, defined as the norm of the residual system Gy-,(s), is given by IG(s) - G()Ila <2 of, «= H,c0 (0.120) for both the Hankel and infinity norms, when all Hankel singular values are distinct (otherwise we may have an unstable error). Although this error is Rot minimal, it is close to the optimal solution. In (132) the family of optimal reduced. order models, measured in terms of the Hankel norm approximation, is obtained. The corresponding minimum approximation error is min |IG(s) — Gils)Ilw = off, (9.121)318 MODEL ORDER REDUCTION It is possible to find a D, matrix for the Hankel optimal reduced-order model G,{s) such that G(s) ~ (Gels) + Dill = So of (9.122) This is so because the Hankel norm does not depend on the constant D term of the realization Therefore the balanced truncation algorithm produces a solution subopti- ‘mal up to a factor of 2 in the infinity norm, with less computational work 9.5 ALGORITHMS ‘Moore's algorithm ((209)) for balancing an asymptotically stable system [42 aw = [244] (9123) is a follows. 1, Compute the controllability Gramian W, from the corresponding Lya- punov equation, 2. Compute its Jordan form with ordered eigenvalues (largest first) and orthogonal eigenvectors: Ve (9.124) This is used to form the transformation matrix T) 2 3c'Vz. which produces the new system matrices A = TAT; ', B= T,B, C= CT," ‘The new controllability Gramian is W. = 1 3. Compute the eigenvalue decomposition of the new observability Gramian as before, Wy = Vo3Vg. This is used to form the transfor- mation matrix Vg, which produces the balanced realization: Ay = TATs! = TAT, "Ts! (9.125) By = nb =r. iB (9.126) Gye ETS C7 S'T5! a2) liza under It is not difficult to verify that this is_an internally hatance tion and that the transformation matrix T = T>T is nonsingular certain assumptions (Prableni'8)~~—~ ALGORTHMS 319 Another version of this algorithm was developed by Laub {180] and Glover (133) 1. Solve the corresponding Lyapunov equations to compute the Gramians W. and W,. 2. Compute the Cholesky form W, = R'R (also known as the matrix square root). 3. Compute the Jordan form of RW.R* = US*U", selecting U to be unitary and ordering the eigenvalues as before. The transformation ‘matrix that takes the system to its balanced form is T= 3-"?U"R, with Ay = TAT!, B, = TB, Cy = CT Its also easy to verify that this is an internally balanced realization and that the transformation matrix T = 737; is nonsingular under certain assumptions (Problem 9), In both cases, after a balanced realization is obtained, a truncation step needs to be performed to reduce the system's order, A, Ap B, we[h a me [E] C= [a CG) Gls) = i Be | (0:28) 5.1, Approximation Error Selecting the order r of the reduced model entails trading-off model size versus approximation error. It is desirable to have a small number of states due to computer speed and memory size constraints, but this may increase the approximation error. To this end it is important to consider the following, ‘bounds, taking into account that the r selected states corsespond to the larger Hankel singular values {0}, i = 1,7) NGC) ~ Gels)Ily <2 of! (9.129) G(s) ~ G6} <2 of! (9.130)320 MODEL ORDER REDUCTION If the original system is not strictly proper, that is, D # 0, we select for the reduced-order system D, © D. This choice does not modify the Hankel sin- {gular values or the approximation error bound and causes the approximation error at higher frequencies tend to zero, that is, lim,» |G(s) ~ Grls))la — 0. If the system has a stable and an unstable part, we can reduce the order of the stable part and retain the unstable one Gls) = Gls) + Guls) (9.131) Gls) = Gals) + Gals) (9.132) G(s) ~ Gio = G8) ~ Garb)» Hoo (9.133) Again, the approximation error remains unchanged as well asthe number of unstable poles. This is an important assumption for robust stability analysis (Chapters 2 and 4). I we consider G,(s) as the nominal model from which a robust controller will be designed, we may interpret the error (Gu) ~ G, (jo) as frequency-weighted additive dynamic uncertainty. This family of models should “cover” the plant G(s) by means of a nominal reduced-order model G,(s) and a frequency-weighted bound. Additionally, the numberof unstable poles of each member ofthis family should remain the same, in order forthe necessary and sufficient condition that guarantees robustness ofthis class of (additive) dynamic uncertainty sets to be vali ‘Equivalently. if G, (3) isinverible, we may consider the trequency-weighted ‘multiplicative uncertainty [G(Jw) ~ G,()|G, "Gu. In most cases, the weight for multiplicative uncertainty is lower at low ftequencies and increases at higher ones, up to a eitcal « for which it increases above 100%.-On the other hand, the behavior of the balanced-truncated approximation error is just the opposite: its higher at lower frequencies and goes to zer0 as w ~ 00. ‘Therefore, for robust control applications, its important to invert this fre- quency characteristic of the approximation error. This can be achieved by balancing the realization of G(s), but truncating @ new system G(s) = G(i/s). The reduced model is piven by G,(s) = G,(1/s). ‘This transforms lower into higher frequencies and preserves stability, Gra- mians, and, asa consequence, also balancing (see [24). Furthermore, the error bounds remain the same, although the error asa function of frequency changes, Specifically, the algorithm proceeds as follows: 1, Transform the balanced system G(s) to G(s): Het] [eatoteee] ono ave [A 2)'-[ ee] ons PROBLEMS 321 ‘based on the partition of the original balanced system in (9.102). Then the truncated system G(s) is 6.0) [Reon Enea GRi-GRa | D-Cate | 136) 3. Finally, invert the frequencies on the reduced-order model G,(s). Sim- Plifying we obtain (prove) Au ~AvAg An | Bi-AnAgB (9.137) Gis) = | Ae] 08 C= GAAn | D- GAB: a Note that the matrix Az, is always invertible, because it is stable, Thus Gy(s) is well defined, There are other algorithms that solve some of the numerical problems inherent in obtaining balanced realizations ((263)).It is also possible to per form balanced truncation over badly conditioned realizations or nonminimal cones. These algorithms are implemented in commercial CAD tools: Mu-Tools and Robust Toolbox from MathWorks ([20, 262). 9.6 PROBLEMS Prove that the Gramian Win (99) is symmetric and postive semi- ‘definite, and therefore its eigenvalues are all real and nonnegative, 2,/In the proof of Lemma 9.3, verify that signal u(t) takes the initial state ~' 2(0) to 2(7) with minimum energy. 3. Obtain the expressions in (9.68) and (9,69) as functions of the matrices of the equivalent discrete time system. Prove that substituting in (9.72) ‘we obtain the controllability Gramian Lyapunov equation (9.61). 4. Show, by means of an example, that the Hankel singular values of the difference of a system and its balanced-truncated version are not equal to the Hankel singular values of the difference of their Hankel operators, In which case could they be equal? ‘&/ Prove the relation between matrices Tip, Tiqy and Ton, a8 indicated in (9:88) and (989). \@ Consider a stable, strictly proper minimal realization represented by ma- trices A, B, and C. Prove that any balanced realization of this system can be transformed to another balanced realization by means of a ut satisfying TE = 37, anced Testization tary ith W, = W, = 3. Thus there is not a unique bal-322 MODEL ORDER REDUCTION 7. Prove tha Moore’s algorithm transforms a realization into its balanced form, Which is the assumption that guarantees Tis a transformation matrix? 8, Prove that the Laub/Glover algorithm transforms a realization into its balanced form. Which is the assumption that guarantees T is a transfor- mation matrix? From the point of view of computational cost, what is the difference between both algorithms? 10, Consider the Jordan form of the product of Gramians ofa system, W.Wo = TET" with unitary 7. Can T be used asthe transformation matrix which takes the realization to its internally balanced form? IL. Verify the equations for the algorithm that provides zero error at « = 0 presented in Section 95.1, ROBUST IDENTIFICATION 10.1 INTRODUCTION In the first part of this book, the analysis and synthesis of robust controllers were considered. A basic point that has not yet been addressed is the method by which a set of models that represents a specinc physical process can be obtained. Before the appearance of systematic methods, families of models were obtained by ad hoc procedures (see Example 11.2). At the end of the 1980s, the first algorithmic strategies were introduced, based on approxima: tion techniques that provide a uniform error bound ((134, 139, 321)). Classical identification procedures (184, 293]) are not adequate for ro- bust control, because usually they identify, using stochastic methods, a set of parameters of a fixed mathematical structure. These procedures are used to identify parametric models in adaptive control. Within this framework, a fixed model order is assumed, which is not the case when using robust design and analysis tools. Even if families of models with parametric uncertainty could be obtained in this way, there is a limited design machinery for robust analysis and synthesis of this class of uncertain plants. In general, the analysis of parametric sets of models is a NP-hard problem ([60, 248). Robust control analysis and design tools are based on a deterministic worst-case approach, with no previous assumption on the order of the system. For this reason intense research was initiated in the 1990s on deterministic identification procedures, which can be used as a first step in a robust control design methodology. This area is called robust identification, ‘The robust identification problem has been posed in (143, 144, 145, 146] and has attracted considerable attention since then. It considers model uncer- tainty due to two different sources: measurement noise and lack of knowledge 323324 ROBUST IDENTIFICATION of the system itself due to the limited information supplied by the experimen- tal data. Therefore these new identification procedures are based not only ‘on the experimental data (a posteriori information) but also on the a priori assumptions on the class of systems to be identified. The algorithms produce ‘2 nominal model based on the experimental information and a worst-case bound over the set of models defined by the a priori information. If time response experiments are performed, an é; error bound can be de- fined ((75, 193, 194, 197), which is adequate for & optimal control. Further- ‘more, these types of experiments are useful in verifying the (approximate) stability, ime invariance, and linearity of the systems ((272), for example, before a frequency response experiment! is performed. ‘The uncertainty bounds can also be measured in terms of the Ma, norm (9, 137, 138, 146, 147, 191, 196)), which is more adequate for 7. optimal control or w-synthesis procedures. In this case, the experimental data are obtained from the frequency response of the physical system, whichis also assumed to be linear and time invariant ‘A general abstract formulation of the problem will be presented next. We postpone the specific algorithms and convergence results using Mx. and metrics to later sections of this chapter. 10.2 GENERAL SETUP In this chapter, we present algorithms for solving the problem of (robust) control-oriented identification of physical plants. The outcome of these pro- cedures is a family of models, represented by a nominal one and an identifi- cation error. We start with an abstract and general formulation ((304)), which does not take into consideration specific metrics (used to evaluate identifi cation errors and convergence) or model and noise sets. The objective of presenting a general formulation is to point out the differences with classical parametric identification methods ((184, 293]). It is also useful to understand the fundamental limitations that appear when working with corrupted and incomplete data and model ser inclusion, from an information theoretic point of view. As a consequence, the structure of the information on the plant will be stressed. Furthermore, this general setup is useful to indicate the similarities with approximation procedures of infinite-dimensional models. These will be in- cluded as particular cases ofthe identification ones, Recall that approximation is loosely defined as a procedure that represents a known large-order mode! (even infinite-dimensional) by a simplified smaller-order one. On the other hhand, identification is meant as model computation from (possibly) a priori information and experimental data, "Athough all experiments are performed in the time domain, «sinusoidal swcep atthe input ofthe plant obtain an output magnitode and phase response i deiaed here as» equeney ‘esponte experiment (sce Seton 1032), GENERALSEIUP 325 10.2.1 Input Data ‘The outcome of a robust identification procedure is a family of models that should include the real physical plant behavior. This family is specified by 2 nominal model and an uncertainty error measured in a certain norm. For the problem to be nontrivial, classes of candidate models and measurement noise should be assumed? as will be explained later in greater detail. The uncertainty bound should be valid for any plant included in the class of mod: els or noise inside the class of measurement noise. This naturally produces a worst-case approach. Therefore the input data to the problem arc the class of candidate models and measurement noise assumed, called a priori infor- ‘mation, and the experimental data, called a posterior’ information, ‘The a posteriori information is a vector y € C™ of experimental data cor: rupted by noise 7. The data can be frequency and/or time noisy samples of the system to be identified. For a model g and a given noise vector 7, the experiment can be defined in terms of the operator E(-,-) as follows: y= E(g.n) (10a) For simplicity we assume E(.,-) linear with respect to both variables, the noise being additive, Note that this is not an injective operator because the same outcome y may de produced by different combinations of mode! and noise. This 18 4 restatement of the fact that the information provided by y is incomplete (M samples) and corrupt (noise »). Therefore the operator is not invertible and 1no direct operation over y will provide the mode! g. Instead, a type of set inversion will be attempted. By this we mean the computation of a set of models, described by a (central) nominal one gy and an uncertainty bound ¢y- This set “covers” all possible sets of plants in the a priori class, which could have produced the a posteriori information y. In order 10 make this a nontrivial problem, itis necessary to add extra information on the class of models and measurement noise. Otherwise the combinations of models g and measurement noises 7, which could have produced y, would not only be infinite but would also form unbounded sets. The set of candidate models would have an uncertainty bound ¢y —+ e. This last point can be exemplified as follows. Assume we have an ex- periment that produces the vector of real numbers y = [yo dual" ‘These represent the (truncated) impulse response of a certain (unknown) system g, measured with (unknown) noise m, at times ty = KAY, that is, Yi = Bulls) +76), k=0,...,M — I. If the only available information isthe vector y € RY, there is no way to identify plant g.. Note that we can always Fepresent every measurement point by yx = glt4) + m(t). =0,--.)M— 1, where g(t) = yu-+z and ns.) = —2, for any arbitrary real value z. There: Otherwise the uncetsnty bound could be infrite and both the maxing and contr problems would not make seme326 ROBUST IDENTIFICATION fore any (possibly unbounded) model or noise could generate this same set of data points Even without measurement noise, that is, 7, =0, the candidate models _g() could be almost anything, as long as yy ~ gli). k = 0,..-5M ~ 1. We can interpret this result in terms of the more familiar aliasing concept. Specifically it is related to Shannon's theorem, which poses an upper bound on the class of plants that could be identified, using this experimental information y. Only if g, belongs to the FDLTI class of models with bandwidth less than 1/24 ccan it be computed from the available data. If the bandwidth is greater than 1/241 but bounded, a smaller (and bounded) set of candidate models (which include g,) could be determined, The above argument justifies the need to define a class S of models and N of measurement noise. Specific classes will be described in further sections of this chapter, according to the norms used to bound the uncertainty error. Hence the input data to a robust identification algorithm are composed of both @ priori (S,N) and a posteriori (y) information. Duc to the fact that the assumed @ priori information is a quantification of the engineering common sense or simply a “leap of faith,” there is no guarantee that it will be coherent with the experimental a posteriori data. Therefore consistency of both types of information should be tested before an algorithmic procedure can be applied to compute the family of models. 10.2.2 Consistency Consistency is a toncept that can easily be understood if we first define the set of all possible models, which could have produced the @ posteriori information y. in accordance with the class of measurement noise: SY) = {gE S | y= Elen), EN} (102) ‘Therefore S(y) CS isthe smallest set of models, according to all the available input data (a priori and a posteriori), which are indistinguishable from the point of view of the input information. This means that with the knowledge of (y,5,.N) there is no way to select a smaller set of candidate models. Note that the above set is independent of the metric selected and depends on all the input information. ‘As we mentioned in the last section, for a robust identification problem to make sense, the identification error should be bounded. Since the “size” of the family of models is defined by this error bound and should include the set S(y), this set should be bounded as well. For this itis sufficient to have a bounded class $ of models (see Figure 10.1). Furthermore, the “size” of the set S(y) places a lower bound on the identification error, which cannot be decreased unless we add some extra information to the problem. The latter can take the form of new experimental data or more detailed information on. the sets S and/or NV. This lower bound on the uncertainty error holds for GENERALSEIUP 327 noise (corruptness) Son . $0) gorithm 2'@ System partins) —— a Figure 10.1. Gonotal statement of he problem any identification algorithm and represents a type of uncertainty principle of identification theory (see Lemma 10.4), It is clear from the above arguments that consistency of @ priori and a posteriori information can be defined as follows. Definition 10.1 The a priori information (S,N) is consistent with the exper- imental a posteriori information y if and only if the set S{y) is nonempty. It should be noted that once consistency between « priori and a posteriori information is established, both are equally reliable. Nevertheless, this docs not imply that the “real” plant g. belongs to S. This statement deserves a more detailed explanation, which will be presented next. 1k could happen that, although (S,V) and y are consistent, the set S does not include the real plant g., that is, the @ prior’ information is false. This could be the case of very noisy measurements (A’“large"), which allows the set S(y) not to be empty, although g, ¢ S(y), that is, there exists gy € S(y) such that E(@1,m) = Elgasm) =y with m,n EN. From @ practical point of view we can detect this problem by performing more experiments, say, {y',92,...9"}, ¥ ERY .i= 1,...,N. If we define this set of experiments a6 A, the new consistency set is Sa)={eeS | y= Elen!) a eNandye a} (103) Its clear that if one of the experiments in A is y in equation (10.2), then S(A) € S(y). In fact, we can equivalently define S(A) =n yS{y'). If the sets $ and AV meet certain conditions (detailed in further sections) and the experimental operator E(-,-) makes sense, as we increase the number of experiment, the consistency set might get “smaller.” Also, the a posteriori local error bound may decrease (see equation (10.5) Nevertheless, from @ worst-case point of view, although N’ new exper iments ate performed, we could always obtain the same set of measure- ments y. In this sense, the new consistency set would not “decrease,” that is, S(A) = Sty). This can be interpreted in terms of the a priori lowest ertor bound (see Lemma 104). Because this a priari errar is based on the worst- case experiment, it will not change with new experimental data, Therefore to328 ROBUST IDENTIFICATION decrease the consistency set (and the a priori lower error bound) new types of experiments need to be performed, for example, increasing the number M of measurements, decreasing the measurement noise bound, and incorporating ‘2 new type of experimental data (time and/or frequency). However, and this concerns any empirical theory, no matter how many experiments we perform there is always the possibility of having the real plant excluded from the set S, even for consistent input information, This is ‘an epistemological problem that relates to the following facts: '* Any empirical theory (or model) can only be falsified (see (245) ‘+ The number of experiments is always finite At this point, the following practical remark can be made: To reduce the possibilities of having the real plant g. excluded from S, as many different types of experiments (time, frequency, changing sensors, and/or data points) as possible should be performed. The final verification is always through a new consistency test. This isthe best that can be done in the framework of the scientific methodology. In the case of approximation, the set S has only a single element (usually infinite-dimensional). Therefore the “experiment” that consists of @ partial sample of this system (§M data points), “measured” without noise, is always consistent. The uncertainty in this case arises only from the partialness (M data points) of the information. The only element of S (the “real” system) cannot be used as the desired model g,, because of computational reasons. In later sections, the consistency problem will be stated as determining the existence of a model g € S, such that the difference between its output and y is included in set NV, that is, it is only due to measurement noise, This may be solved as an interpolation problem. The latter holds for additive measurement noise and linearity ofthe operator E(-,-) with respect to model and noise, which we have assumed. 10.2.3. Identification Error The a priori knowledge of the real system and measurement noise present in the experiment y is stated in terms of sets S and AV. The statement of the problem does not assign probabilities to particular models or noises; therefore it is deterministic in nature. In addition, the modeling error should be valid no matter which model g S is the real plant (or € NV the noise vector) leading to a worst-case approach. In this deterministic worst-case framework, the identification error should “cover” all models g © S that, combined with all possible noise vectors 7 € JV, are consistent with the experiments. Ideally then, the error should only include the models g € S(y). In practice, however, the family of models conservatively covers this “tight” uncertainty set. Hence it provides an upper bound for the distance (in a certain metric) from 2 model (consistent or not) to the seal plant. In this framework, the worst-case ertor GENERAL SETUP 329 is defined as follows: A) E sup. 8 ATE(S, MSN} (104) where £(-,:) is the experiment operator and m(.,:) a specific metric. The identification algorithm A maps both a priori and a posteriori infor: mation to a candidate nominal model. In this case the algorithm is said to be tuned to the a priori information. On the other hand, i it only depends ‘on the experimental a posteriori data, itis called untuned. Almost all class cal parameter identification algorithms belong to the latter class ({184)). For simplicity, in the sequel, the identification algorithm will be denoted as A, without specifying its arguments unless needed, ‘The identification error (10.4) can be considered as @ priori, in the sense that it takes into account all possible experimental outcomes consistent with the classes A’ and S, before the actual experiment is performed, Since it considers all possible combinations of models g and noises 7, that is, all experimental data y, it is a global identification error in terms of the ex: Periments. A focal error that applies only to a specific experiment y can be defined as follows: Ay) 2 sp mig. ASN) (105) From the above, it can be shown that the global worst-case error can be defined equivalently as follows, Lemma 10.1 (A) = sup e(A.y) (106) ye¥ where Y is the set of all possible experimental data, consistent with sets S and Nv ee By definition ¥ © {E(g, n)|g € S, 1 € N)therefore the set {g € S(y) Y€ Y} is equivalent to S, From the definition of Y and replacing y by E(g, n) in (105), we obtain sup sup m (a, A[E(G,n),5,N1) = sup sup mg. Aly,S,]} wee yeveesiy Obviously the algorithm always depends onthe posterior’ information forthe computation fan empirical model330 ROBUST IDENTIFICATION Clearly we always have e(A,y) < ¢(A). To decrease the local error more experiments need to be performed, whereas to decrease the global error new types of experiments, compatible with new a priori classes, should be performed, for example, reducing the experimental noise and changing ” accordingly. ‘As mentioned before, there is an uncertainty principle in identification, which produces a lower limit to the possible global error, closely related to the consistency set S(y). To this end we first define the radius and diameter Of a set and prove several properties. Definition 10.2. The radius and diameter of a subset A of a metric space (Xm), are 1A) = ink sup m(x,a) (107) (103) ‘The radius can be interpreted as the maximum error, measured in a certain ‘metric, when considering the set A as represented by a single “central” point (which may not belong to A). The diameter is the maximum distance between any two points in the set. From these definitions, the following properties can be proved. Lemma 10.2 1A) < (A) < 2r(A) (109) Proof. From Definition 102 we have (A) 0, VEX | YEA m(x,y) 4100) (1016) ye ‘As a consequence, there exists an experiment y., a pair of models gy, 2 € S(y,), and noises m,m € N, such that: E(u.) = Elen.) = ys (10.17) m(gi.g2) > a1S(yo)] (10.8) By symmetry, the element (2es ~ g1) € S, and by convexity the element ty © J Igo + (cs — gi)] € S as well. Hence the element hy 2 cs + $(gi ~ 82) belongs to the set S, due to its symmetry. Replacing $ by AV we can prove in a similar way that the elements jx © ew + }(m1— m2) and #2 2 ew + Hm — 1m;) belong to NV. Using the linearity of the experiment operator E(-,-), we have Yo = 5 (Ete. m) ~ Ear, m)] + Eles.cw) (10.19) ~ E(u in) (1020) and similarly E(/p, ja) = yo This proves that hy, fz € (yo), with m(h, a) = ‘m(gi, 2), which contradicts (10.18), Tis easy to verify that S(yp) is a symmetric set, and therefore supm(sy) < maxes) + mses) (1021) i 2max [m(x,cs),m(y,c5)] (10.22) which achieves equality when y = —x. The diameter is computed among the symmetric elements, which proves (10.15). o Ifthe sets belong to normed linear spaces, we may use the metric mab lla- BI, ‘The above proves that the experimental output achieved by the centers of both « prior sets provides the least amount of information (largest diameter of indistinguishable set). In the case where these centers are the nul elements, it means that the experimental instance with null output provides minimal information, which in some sense is intuitively clear. From the previous discussions, it seems important to have identification procedures that can compute models inside the consistency set. These are called interpolation algorithms and are defined as those that produce models GENERAL SETUP 333 included in the set S(y) for all possible experiments y € ¥. The name comes. from the fact that algorithms that interpolate the experimental data produce a model g that satishes y = E(g,0). and therefore g € S{y) if g € S. ‘A lower bound for the worst-case error of any identification algorithm and fan upper bound for any interpolation algorithm are presented next Lemma 10.4 The worst-case identification error defined in (10:6) satisfies the following inequality e(A) 2 R(T) > 5010) (1023) for any algorithm A The following upper bound holds for any interpolation algorithm Ay Diz) > eA) (1024) Proof. Its clear that sup e(Ay) > e(Ay) > inf _e(A.y) (10.25) agree) ates Next, take the supremum over all experiments y € ¥ on each member of the above inequality. Based on definitions for radius and diameter of information and equation (10.6), we obtain for the first inequality DU) > eA) (1026) ‘The above holds only for interpolation algorithms because the supremum has been taken over all Ay(y) € S(y). Using Lemma 10.2, we obtain for the last inequality (A) 2 RG) > Spa) (1027) Which holds for any algorithm A o ‘As mentioned before, e(A), D(Z), and R(Z) are global in terms of the experimental outcome y. Therefore the above a priori bounds hold for any experiment (consistent with S and NV) and may be used before any mea: surement is performed. As a consequence of the above lemma, interpolation algorithms can obtain the minimal error bound R(Z) up to a factor of 2, that is, D(Z) > e( As) > $D(Z), “almost” optimal In terms of local errors, an important subclass of interpolation procedures, called central algorithms, generate a model g. that satisfies (Ay) = sup m(g.g.) =r 509)) (1028) ws)334 ROBUST IDENTIFICATION ‘This “central” clement may not belong to the set itself (see Definition 10.2 may be nonunique, or may not exist. If it does exist, then itis called Cheby- shev center and represents the best approximation by a single element of the smallest indistinguishable set S(y). From Lemma 1044 it follows that these al gorithms provide the smallest possible worst-case error. Therefore they may be considered optimal in that sense. The importance of symmetric sets is that, the symmetry point is a Chebyshev center of the set Optimal Algorithms 1: is important to consider algorithms that make an clfcient use of the availabe information and produce the smallest possible identification errors. These optimal algorithms ({75) are defined next. Definition 105 The local optimal identification error of a particular experi- ment y cis ©) 3 infe(ALy) (1029) An algorithm A* is called locally optimal if and onty if A, )= ey) Definition 10.6 The global optimal identification error is defined as nf e(A) (1030) An algorithm AY is called globally optimal if and only if e(*) = e*. Further ‘more, an algorithm is called strongly optimal if itis locally optimal for all experiments y € Y. Note that strong optimality is much stronger than just global optimality. I implies that, regardless of the set of available data, the identification pro- ‘cedure provides an estimation of the model, which minimizes the worst-case error. The central algorithms defined before are strongly optimal 10.2.4 Convergence ‘A question that might be posed is the following: What happens with the family of models when the amount of information increases? It would be desirable to produce a “smaller” set of models as input data increase, that is, model uncertainty should decrease. This could be interpreted as having ‘monotonically decreasing uncertainty bound with inereasing input informa- tion, Certainly, we would expect the set of models to tend to the real system (monotonically or not) when the uncertainty of the input information goes to zero. This latter property can be stated in terms of convergence as follows’ ‘An identification algorithm A is said to be convergent when the worst case lobal identification error e( 4) in (10.4) goes to zer0 as the input information GENERALSETUP 335, tends to be “completed.” The latter means that the “partialness” and “cor ruption” of the available information, both a priori and a posteriori tend to zero simultaneously ((146) Input information is corrupted solely by measurement noise inthis scheme, ‘Therefore “corruption” tends to zero when the set NV’ only contains a single known element, say, the zero element. On the other hand, partialness of information can disappear in two different ways: by a priori assumptions or 4 posteriori measurements. In the first case, partialness disappears when the set S tends to have only one element (the real system); in the second case, when the amount of experimental information is completed by the remaining (usually infinite) data points It is convenient to unify the above concepts as follows. The available information (a priori and a posteriori) is completed when the consistency set Sty) tends to only one element: the real system. When this happens there is ‘no uncertainty that is, the input information is complete.* Hence converger can be defined as follows: Definition 10.7. An identification algorithm A converges if and only if ii nea) =0 (1031) The diameter of information tends to zero when the noise set A’ reduces toa single element (usually zero noise) and either the set S tends to the real plant or the experiment operator provides all necessary data. Note that as the consistency set S(y) reduces to a single element, the experiment operator tends to be invertible. Since the identification error is defined in a worst-case sense, its convergence is uniform with respect to the a priori sets N’ and S. The definition of convergence that has been presented reflects the in: formation structure of the problem and generalizes the ones that will be stated in the following sections, When considering convergence in terms of completing the a posteriori experimental information, care must be taken because infinite data points (time or frequency response experiments) do rot in general guarantee complete information ((236)). Take, for example, a frequency response experiment with data only at integer frequencies {o, = k, k=0,...,M~—1). By making M —+ 00 the @ posteriori information will not determine the unique real system since D(Z) does not vanish (see Deft- nition 10-7). that is, information is still incomplete. For this reason, all neces: sary (infinite) data points to complete the information should be considered. For example. it takes an infinitely countable number of points in the time response or an infinitely uncountable number of points for the frequency response of a discrete IIR time system, According to Lemma 1044, all interpolation algorithms are convergent. “Here we aesome that the tea plant ways Relongs tthe consistency set. Aocording tothe comments the end of Section 10.22. th wil evenly he the cise a= kong. 38 the et SY)336 ROBUST IDENTIFICATION Morcover, since these procedures generate a nominal model in the consis. tency set, they are always tuned to the a priori information. In later sections, convergent noninterpolating algorithms, which may be untuned, will also be sought. Convergent untuned algorithms are said to be robusily convergent, This is an important property because algorithmic convergence is dissociated from a priori assumptions. Therefore convergence is guaranteed, even if the a priori information is false, although the resulting bounds are no longer valid, This has a practical implication because most of the a priori information is bbased on a “leap of faith” or a combination of engineering “intuition” and ‘As mentioned before, model approximation can be considered as a special case of identification. In that case, the sets $ and N already have single known, clements: the real model and zero noise, respectively. Uncertainty is only due to the partialness of the a posteriori data. In model approximation the objective is not to exactly invert the experiment but to produce a lower-order ‘model (due to computational tractability). Convergence is therefore defined in the limit as the number of data points tends to be completed (M —+ 0c), rather than with the diameter of information going to zero. Convergence and a priori class S The worst-case error defined by equation (10.4) depends on the classes of models and measurement noise. In turn, the convergence of this error to zero also depends on these classes, ‘Thus there are necessary conditions that the sets S and A’ should meet in order to guarantee the etistence of convergent algorithms, Next, we present a condition on S that precludes the existence of these types of algorithms (1194). Lemma 10.5 _[f the a priori class 5, which belongs to metric space (Vym), con- ‘ains a ball of size y, say, By(go) = {g © V, m(g,go) < y} CS, then R(T) > y {for any number of measurements and any noise level. Proof. The experimental evaluation of the plant g produces y = E(go,n) = [9b + nf BRT Wl") for m €.N’. We can always find a model g. such that gf = gh, k =0,...,M—1 and m(go,g.) = 7, that is, go € By(@0) ‘Therefore independently of the value of M, we have two indistinguishable elements gp and g, inside the ball. Thus these elements should be included in the consistency set. Since m(go,g.) ~ y then r[S(y)] > a Several comments concerning the above result are presented next. «Ifa set is defined only in terms of the distance among its elements, there is no relation between the number of measurements M and the “size” Of the set. Therefore the @ priori class should not contain this type of set if convergence is desired. ‘* The element g, described above belongs to the kernel of the operator [E(,-m) ~ y). Therefore the “size” of this kernel should decrease to 2er0 GENERALSETUP 337 as M — oo and 1 ~ 0. This is related to the inversion of the experiment operator £(-,-) when we have complete information, as mentioned in Section 102.1 « In [194] and [229] specific plants «, are provided for the metrics defined by the f; and %4,, norms. + From Lemma 105 it follows that S should not have interior points if we expect to have convergent algorithms. There are yet two other elements that should be limited by the a priori information: the intersample behavior and the “tail” of the experiment, that is. (94, & > M). Both are a consequence ofthe finite numberof experimental data points obtained from systems with infinite time oF frequency responses. To this end define the following: Definition 10.8 Consider a metric space (Vm) and a family of functions Az) —+V. The family A is equicontinuous if and only if for any ¢ >0, there exits 6 > 0 such that if x,y € Y, m(x,y) < 8 implies miUf(x), f0)] < ¢ for all fea The above is a generalization of the uniform continuity of a function, extended to a whole family A. Such a family has a uniform bound over the intersample behavior of each of its functions, An equivalent characterization of bounded equicontinuous sets is given by the Arzela-Ascoli theorem ((201)), ‘Theorem 10.1 The following statements are equivalent (a) The family A is equicontinuous and uniformly bounded. (b) The closure A is compact. ‘Therefore selecting compact sets for the a priori class of models will limit the intersample behavior. To limit the “tail” of the experimental output, which is unknown due to the finite number of samples, additional restrictions should be added to the a Priori set of models. When considering the last portion of the time response, this condition is related to the stability of the class of plants. On the other hand, when considering the frequency response the “tail” of the experiment is restricted by the “properness” of the models. To recap, in order to have convergent algorithms, we need the a priori class of models 5 to have the following properties: 1. It should not contain interior points so that it converges as M —+ oo. 2. It should be an equicontinuous set to limit the intersample behavior. It should also be uniformly bounded in order to have a bounded consis: tency set, therefore it should be compact. 3. It should inelude only stable and proper models so that the (unknown) tail portions of the (time and/or frequency) experiments are bounded,338 ROBUSTIDENTIICATION For the sets considered above it can be proved that pointwise convergence is sufficient for uniform (worst-case) convergence ((19). Lemma 106 For a set uf M measurements and noise level €, any untuned denufication algorithm A satisfying im tim Els, m))~ gh =0 (10.32) for all g € V, also sauisfies (1033) lim limsup AEC.) ~ for any relaively compact set? 5 CY. Proof. See {196 c 10.2.5 Validation ‘The main objective of robust identification is 10 compute a set of models that can be used in a robust control setup. A general framework to deseribe robust control problems uses LFTs, as described in the initial chapters of this hook fn this setup, there is a block that represents the nominal model (including robustness and performance weights), a controller block, a model uncertainty one, uncertain disturbance inputs, and error outputs, Both the ‘model uncertainty block and the uncertain disturbance inputs are usually defined as bounded sets, ‘Similarly to the consistency test that verifies that a priori and a posteriori information are coherent, another procedure should establish whether of not the outcome of the identification procedure is coherent with the setup in which it will be used by being able to “explain” all available experimental information. Therefore, if a set of models is to be used in the context of a Robust Control framework, it seems reasonable to verily the existence of an element of the model uncertainty block and an uncertain input signal (both included in their respective sets) that reproduces the experimental data. This is a reasonable test irrespective of how the set of models has been obtained. This same procedure is used in general to test an empirical theory. In a scientific framework, an empirical theory must be capable of being falsified ((245)). Similarly, an empirical model® should be tested against all available experimental data. If positive, this provides a verification that the model is rnot false. Obviously this test does not prove that the model is completely valid, because a new set of experimental data may invalidate it. For this Teason, a model or set of models cannot be validated, only invalidated. The 5 sets sid to Be relatively compact itt closure compact "11 not necessary hat the mode! computation sned the sorementioned experimental date ow the mrael as heen obaned st relevant st this pnt FREQUENCY DOMAIN IDENTIFICATION 339 procedure called invulidation tests the model or set of models against all available experimental information. It provides a sufficient condition for the model to be invalid, if particular experiment does not fit in the framework. Otherwise it only provides a necessary condition for the model to be valid when all experiments can be explained in this setup. For different reasons, the problem of invalidation has usually been called the validation problem. It is not the purpose of this section to justify the use of this terminology (which can be explained probably better by psychological reasons) but only to introduce this concept, which will not be developed further. Nevertheless, care must be taken when using the word validation in this context. Ifthe analysis framework is identical to the one used for the experiment, a set of models computed using a robust identification algorithm will be always validated by the experimental data used to compute it. Therefore the only two reasons by which a robustly identified set of models may be invalidated © A different structural setup + A new set of experimental data. ‘The usual framework in which the robust idemtification experiments are performed is coherent with the use of the LFTs as a general control frame: ‘work. The consistency among both structures should include the norms used to bound disturbance inputs, model uncertainty, and output errors as well In practice, the invalidation problem involves computing the minimum norm bound on the disturbances and mode! uncertainty to validate the ex- periments. Its clear that with large enough disturbances or uncertainty, any experiment can be validated, Hence this minimization problem also measures the conservativeness of the set of models. ‘The problem of validation has been initiated in {292}, for models given in a LFT form, using the 1, norm to bound uncertainty and disturbances. The problem has a solution in terms of the structured singular value and includes structured dynamic model uncertainty in the model description. The statement of the problem is given entirely in the frequency domain, For a practical application see (291)) ‘The use of a time-domain description and bounds in the & norm have been proposed in [242]. The solution involves a convex matrix optimization problem based on the experimental data and the proposed set of models. 10.3. FREQUENCY DOMAIN IDENTIFICATION In this section we present Robust Identification procedures for a priori classes oof models in 71. and experiments performed in the frequency domain. The scheme follows the general framework presented in the last section, although specific issues pertaining to this metric will be stressed.340. ROBUSTIDENTIFICATION 10.3.1 Preliminaries We consider causal, linear time invariant, stable either continuous or dscrote time models. For simplicity we consider only SISO systems, although all results thoughout this setion can be applied to MIMO systems aswell, either clement by element or asin 76. To unify the treatment we denote the models as H(z) = HAS), A > O in the continuous time ease or H(z) ~ Hall/2), for discrete time systems, with z € C. Note thatthe latter (known a5 the transform in the bibliography (87) is the inverse of the usual z transform. Therefore in the Sequel, causal stable models H(2) wil be analytic inside the unit disk, In this context the idemtification problem reduces 0 that of computing a model, analytic in a certain region, based on partial and conrupt data measured atthe boundary of ths region In order to define the a posterior information, we introduce the following sets Mas {io analyte ll | and finite K, the following holds: Moolp, K) C Hoop © A(D) C Hoe (10.39) Note that %a.(9,K) isa strict subset of A(D) because there are elements in the subalgebra that do not belong to (p,K) (give an example) The a priori class of models is defined as the set S = Hy.(p, K). This cor- responds to exponentially stable systems, not necessarily finite dimensional FREQUENCY DOMAIN IDENTIFICATION 341 From a practical viewpoint, the models defined by the @ prior’ information have @ stability margin of (p ~ 1) and a peak response to complex exponential inputs of K. The following bound is satisfied in the time domain (prove) [mi] < Kp (1040) The a posteriori experimental information is the set of M samples of the frequency response of the system, measured with additive bounded noise: Y= HC) +n, k=, MAI cost) which represent the components of vector y € C™, The measurement noise is bounded by in the f, norm, To recap, the a priori and a posteriori input data are, respectively, (He HK), p>1, Ko) (108 Wa (ne imbce k=OouM 1} 088 y = EH.) =Hlel™) +m, k=0, M (10.44) Since the a priori class S is relatively compact, Lemma 10.6 applies. Also note that both sets, Sand A’, are symmetric with centers cs and cy in the zero model and noise vector, respectively. Based on the general statement of the last section, the global error and convergence criteria for algorithm A are as follows: (A) = sup AE(H.n),5,M]—Hln (1045) lim eA) = 0 (10.46) According to Lemma 106 if A is untuned, that is, Aly). the convergence criteria need only be applied to the error corresponding to plant H: |.A(y) — Hljx instead of to the above uniform error. Using the results of Lemmas 10.2 and 103, a simple lower bound for the information radius can be found. Lemma 108 The radius of information cannot be smaller than min(e, Proof. Define C= min(e, K) and consider the constant models Hy(2) =Ha(2) = C and noises 7) =~» = —C, which belong to the a priori classes S and N, respectively. Models H, and Hy also belong to the consistency set ‘S(y) for the null experiment yp = 0, that is, Hy 4 1) = Hz + m = 0. Hence.342 ROBUST IDENTIFICATION from Lemmas 10.2 and 103. RID) 2 |DIL) = Lals(oy) min(e.K) (10.47) o Again, this bound is global and holds for all possible experiments. A feature of Hx. identification is that it assumes that the experimental data is in the frequency domain, Real world experiments are always per- formed in the time domain. For this reason, the time-domain data needs to be pre-processed in order to obtain the required complex frequency-domain 4 posteriori information. In the next subsection, two convergent algorithms, which compute the data vector y and the equivalent frequency measurement error bound e, are presented. 10.3.2 Sampling Procedure To obtain the a posterior’ information, an experiment should be performed to compute the frequency response of the system and the data vector y € C¥ For practical reasons. this experiment involves bounded sequences of inputs ‘and outputs measured in the time domain, The usual procedure is to input, sinusoidal signals and measure the phase and magnitude of the output in relation to the input. The a priori information indicates the time at which the transient response is irrelevant and only the steady-state values are significant. Additional practical considerations involve verifying the a prior’ assumptions. (272), that is, stability, linearity, and time invariance, The computation of the a posteriori data points can be performed using the same framework as, the model identification itself, as shown next Consider the system initially at rest, a sequence of sinusoidal input signals bounded by a and measurement noise bounded in magnitude by y. Inputs. The complex sinusoidal sequence ono c . 4 w= {6 Ben (1048) for the set of frequencies (4: k= 0,...M=1} COupus. The measured vector x € CY. with components ee ee ye foals, N coe) ‘Two possible algorithms ((146}) that identify the values yy in (1041) are the following FREQUENCY DOMAIN IDENTIRCAION 343, Algorithms. The data points at each frequency are computed by algo: rithms P, or P, as follows: Pek =m = 1050) LS oe Prk =e = oy Yael (ost) Errors, The above values have the fllowing error bounds Method 1: Im| < MBp 42 2e (1052) Kelp") ya Method 2: Im) < M226 oss ints KeGme), 2 (1083) forall k= 0,...,M-1 a8 shown in (146), Note that v isthe time domain measurement noise bounded by from which a bound € over the complex frequency-domain noise m,k =0,...,M ~ 1, is calculated, For the real system the inputs are replaced by sinusoids, although the bounds and outputs are similar. For simplicity the frequency points are equally spaced , = 2/M_ Note that the first term of both error bounds is related to the transient response, which vanishes as time increases (N — 00). The second term, on the other hand, is constant and represents a signal-to-noise ratio that can only be decreased with a better measurement procedure, that is, decreasing, y ‘The above frequency point identification procedures can be recast in the same framework as model identification. The a priori information are the noise bound y and the values (p, K) of the plant set Hax(p, K). The a pos: teriori data are the time measurements x. The error is defined in a worst-case sense over the set of measurement noises v, defined above and the class of models H.u(p, K). AS in the general case, there is also a lower bound to the identification error, which can be specified in terms of the radius of infor- ‘mation (see Section 10.2.3). Specifically for all frequency point identification procedures, the following result holds Lemma 109 The worst-case error for any frequency point identfcaion al gorithm P is always greater than min(K, ¥/a. Proof. The proof is similar to the one in Lemma 10.8 by defining the con ‘models as + min(K, y/a); therefore it will not be repeated here, 3344 ROBUST IDENTIFICATION I is clear from the bounds of Methods | and 2 that both are convergent In fact, since they are untuned with respect to the a prior’ information, they are also robustly convergent im tim ec) = 0 (05a) Furthermore, according to the above lemma, both algorithms converge when K > y/a, to the (optimal) lower bound y/a as N —+ 00. Therefore these pro- cedures are robustly convergent and asymptotically optimal. We can always assume that K > y/a, since otherwise the input signal magnitude a can be scaled such that this condition is guaranteed. For practical reasons, itis clear that the frequencies cannot be gen- crated exactly. Nevertheless, the frequency error 5, can be included in the bounds as well ((229)) 10.3.3. Consistency Recall from Section 10.2 that consistency of a priori and a posteriori in- formation establishes the validity of the identification error bounds. As a consequence, the covergence of tuned algorithms depends on the solution to this problem. Although untuned identification procedures converge even for inconsistent input data, the use of the set of models in @ robust control framework is based on the validity of these bounds and hence also depends ‘on the consistency test. Given the experimental data {yx = H(e) + ng k=0,...4M —1} and the a priori sets S and N defined before, the consistency set S(y) can be specified using (10.2). The consistency problem can be rephrased as the ex- istence of a model H € S whose difference with y atthe sampling points can be attributed solely to measurement noise in A, that is, lite ye] Se for all k=0,...,M—1 (1055) We refer to this problem as an approximate interpolation one because /7(e!) matches y, up to an error of size €. For 7 identification, the consistency problem can be cast as an approximate interpolation problem and solved via the Nevanlinna-Pick procedure ({74]). A brief statement of this latter result follows ((21)) - Consider M values w, measured in the interior of the unit disk Dat points zy. There exists a function f € BM, that satisfies f(z) = my for all k=0,...,M-~1 if and only ifthe Pick matrix P, 1 wy Py = a” rat i,j =0,...M— (1056) FREQUENCY DOMAIN IDENTIFICATION 345 is powitive semidefinite. If P > 0 then there exist infinite interpolation fune: tions f, which can be described as a linear fractional function of a free parameter in BH... These solutions will be described in further detail in ‘the context of interpolation algorithms (see Section 10.3.4 and Problem 6 5). This result can be extended to Huu(p,) using the fact that f(z) € BR, + Kf(c/p) € Mou(p,K) (verify). Hence the rescaled set of Pick matrices can be defined as follows: 1 Dyn LR G+ 1 La) 1,7 =0,...M— (1037) for any vector of measurement noise 9 € NV. Therefore an equivalent char- acterization of the consistency problem can be stated as follows Lemma 10.10 A set of experimental data y ¢ ©™ is consistent with the a priori sets S and N’ defined in terms of constants (K, p,e) if and only if there exists a vector n €N such that the Pick maurix in (10.57) is positive definite! Proof. Follows from the fact that f(z) € BH, => Kflz/p) € Sand the classical Nevanlinna-Pick theory. 0 From the above lemma it follows that a necessary condition for consistency is that all experimental samples remain below K + €. This is based on only ppart of the a prior’ information and stems from the fact that for all diagonal clements in P to be positive, itis necessary to have Wet ml
lal —Iml Inl< Kee (10.58) for all =0,...,M ~1 and 9 €.M/, where a signal-to-noise ratio greater than 1 has been assumed, that i, y4| > /n| Clearly the above isan intuitive upper bound for the experimental data based only on the magnitude of the signal and the measurement noise Finally, to test consistency, at least one noise vector 7. € N’ should be found. A’computational practical solution to this problem can be obtained by solving a minimization problem ([74): Lemma 10.11 The following statements are equivalent: 1. The a posteriori data y € © and the a priori sets (S\N) defined by (K,p,€) are consistent. The pongeneric case of P can be found in = 0, which has a unique solution, will not be considered here hu346 RORUST IDENTIFICATION 2 ring. MR) <0 (1059) o: 4 XC) : Fe lavaxem “0 aes where Y= diap[ye --- yas fe X= dal ms mes] and 04, = [1~ th jp?) 3. The linear matric inequality (LMI) problem RIXmIl > 0 1061) tex) ne [xin “?] >0 (10482) thas a feasible solution for a vector 9€ CM 4 iminyea {QY2[¥ + X(N] OF} < K 1063) Proof. Alter some straightforward algebra, the Pick matrix can be written pag reer +X)" (1064) K Using the Schur complements, P > 0 <=» R > Oand the equivalence of state ments 1 and 2 follows. Matrix R(X) is linear in X, which in turn is linear in n, The positiveness condition in equation (10.62) is equivalent to |7)| < €, i =0,...,M 1, Therefore the LMI feasibility problem 3 is equivalent to consistency. Finally the equivalence between statements 4 and 1 is derived in (74 0 Statement 2 is a standard generalized eigenvalue minimization problem that can be solved via the methods developed in [214] or by semidefinite programming ({308)). Problem 3 is a standard LMI and can be solved using available commercial software ({126)). Finally, convex programming can solve problem 4, ‘An advantage of testing consistency through an approximate interpolation problem is that, from the parametrization of all interpolating functions, an interpolatory identification algorithm can be obtained dircctly. This will be detailed in the next subsection, 10.3.4. Identification Procedures ‘Once consistency has been established, the computation of a nominal model and a valid model error bound can be attempted. In this section two different FREQUENCY DOMAIN IDENTIFICATION 347 types of algorithms will be presented, as well as their corresponding worst: case model error bounds. The frst type of procedure, initiated in [144] and extended to H.. identifi cation in [137], consists of two steps, the second of which is an optimal Nehari approximation. In this approach most of the research effort has been concen: trated on the first step, aimed towards improving the algorithms perform- ance. These procedures can be untuned to the a priori information: therefore they can be robustly convergent when the worst-case error vanishes as the input information is completed. This means that even if the model error bounds are invalid (if a priori and a posteriori information are inconsistent), the convergence of the algorithm is guaranteed. On the other hand, the nom- inal model generated by these two-step procedures is not guaranteed to be in the consistency set, which in principle may lead to a larger model error bound. ‘The second type of algorithm presented in this section is interpolatory, hence the nominal model is always in the consistency set. From the results presented in Section 10.2, it follows that the algorithm is always tuned and ‘convergent, Furthermore, this procedure is also strongly optimal, in terms of the identification error, to within a factor of 2. This interpolatory algorithm is based on the consistency test performed in Section 10.3.3 and can be solved tusing the parametrization of all solutions of the same Nevanlinna-Pick prob- lem. This interpolatory algorithm was first presented in [74]. Another type of interpolation algorithm has been reported in [233, 236), but will not be presented here. As a comparison, from a computational point of view, the first step of the two-stage algorithms can be solved via a FFT for equidistant experimental data and may take longer in the case of nonuniformly spaced data ([8, 127)). The second step (Nehari approximation) has a computational cost similar to the Nevanlinna-Pick problem. Finally, note that the above identification algorithms produce a constant ‘uncertainty bound as a function of frequency. Therefore the description of the set of models does not admit any frequency-dependent uncertainty weight, In the case of model approximation since the “real” system is available, the difference with the nominal model can be computed explicitly as a function of frequency. The uncertainty weight can be computed much in the same way as the nominal model. In this process the order of both the nominal model and the uncertainty weight that constitute the augmented plant will be the same as that of the nominal model itself. This will not be explained here but is detailed in [127 In the sequel we illustrate the identification algorithms through the fol- owing example Example 10.1 The physical system under consideration is a Euler-Bernoulli beam with viscous damping, which can be described using the following phys: ‘cal model that relates the vertical displacement y to time t and the longitudinal348. ROBUSTIDENTIFICATION coordinate x: Oy ep LOY, MY wad 1 TY ocr Zh 0, ve (Ly (1065) with boundary conditions ay Bane Facey =0; aan=0 (10.66) Here ais the cross-sectional area, p the mass density of the beam, I the moment of inertia, E the Young modulus of elasticity, and E* the normal sirain rate The values adopted in this case are: ap = 46, E*1 = 0.46, and El = $5.2, and the dynamics are evaluated at x = 0. Since the PDE (10.65) is linear, by using the Laplace transform we obtain ‘he following (infinte-dimensional) transfer function: ao WCET+ ED Hs) = (sim sinh A ~ 008A cosh A ~ 1) cos Ax SinbAx~ sin Ax + "cos Asin A + sin A cosh & (cin Asinh A + cos A cosh A + 1) cosh Ax ‘cos A sinh A + sin A cosh A oe (10.68) Due to the rigid mode of the structure, this transfer function has an unstable part 1/2ps. Although in practice itis nox possible to physically separate the tuistable and stable responses, the purpose ofthis example is only 10 test the different idenification methods over the stable portion. Thus in the sequel we assume exact knowledge of the unstable portion of the model, and we subtract its response from the experimental data inorder 10 identify only the sable part At the end of this chapter some comments on identification of unstable systems in closed-loop experiments are presented. In this case, we have generated three “experimental” data sets of 180 points each from the frequency response of the flexible structure using the software package (173). Figure 102 shows the total magnitude response of the sytem (stablesunstable), the stable part, and three sets of simulated “measurement” noise in f(€) with € =8 x 10" The three sets of “experimental” measurements can be combined to pro- dluce an equivalent set of data points and a new measurement error bound. This can be achieved by considering each data point as the center of a ball of radius «, in C for each experiment i = 1,2,3. At each frequency, he center and radius of a new ball should be computed such tha t “covers” the inter FREQUENCY DOMAIN IDENTIFICATION 349 otal system x10) Stable part Nose scrote requency Figure 10.2. (0) Total froquoncy rosponse, (b) Stable frequency response. (c) Thee ses of “oxparimental” measurement noise. section of all the balls corresponding to the experiments to be combined. This produces an equivalent set of data points and a frequency-dependent error Dowd that includes all previous experimental data. This frequency-dependent noise bound should be smaller than the individual bounds for each exper ‘ment and could provide a less conservative description of the noise, that i, €(1@) < miner, 4 6) In this particular example, the equivalent set of measurements has been “superimposed on the thre inital datasets in the first llustration of Figure 10.3 The frequency-dependent error bound of this new measurement set appears in the second part of Figure 103. For simplicity throughout this chapter we will use the constant upper bound € = 8 x 10°*. Two-Stage Algorithms A first intuitive algorithmic approach to obtain a model from the a posteriori measurements would use the discrete Fourier transform (DFT) and its inverse (IDET), as follows: rt = 5 ee, ae 0360 ROBUST IDENTIFICATION x10 7| es IL — inversion ql Experments | 5 ee ai 2| o 05 1 7s 2 25 3 x10 a ° 05 1 15 2 25 3 scree frequency Figure 10:3. (2) Superpostion of three measurement sats ond their combination. (©) Equivalent measurement noise ot each Frequency, Hala) = > halkyz* (10.70) Here hy, corresponds to an “equivalent” FIR impulse response obtained from the experimental samples y and H, is the identified model. Model Hy,(2) interpolates the experimental data at M frequencies in the unit circle. that 4 = Hy (@), k=0,...,M — 1. If the real system has a finite impulse response (FIR) with less than or equal to M nonzero samples and in the absence of measurement noise, then H,,(2) coincides with the real system and only the M values H,y(e), k = 0,...,M ~ 1, are necessary to represent it If, instead, the real system has an infinite impulse response (IR) and/or the samples are affected by noise, the following two problems appear: ‘© The equivalent impulse response ht, suffers from sine aliasing. Thus the ‘identified model H(z) could be unstable, even if the real plant is stable This becomes clear by shifting the sum in (10.70) by (M ~ 1)/2 samples* Assume without los of penerahy thal A iso FREQUENCY DOMAIN IDENTIFICATION 381 0 compute the (periodic) function H(z) as follows Ha) = YI hatkyat (071) aya Hence if the (periodic) function hy1(k) # 0 for —(M ~ 1)/2< k <0 (the same as for (M + 1)/2 < & 0, 26 Mii Mee 2 Flog) +01) (10.76) AS a consequence, the worst-case interpolating function diverges as the number of samples increases. ‘The above problems motivate the 1wo-siage algorithms ({146, 137]) 1. Using the IDFT, an equivalent impulse response is obtained from the ‘experimental data y. This response, suitably weighted so that the con vergence is uniform as M — oo and ¢ —+0, is used to compute a model H(z). This part of the procedure is linear in the a posteriori data Yes k=0,.-.5M—1 2, The model H(z) may have an unstable part, although the original plant was stable, Therefore a projection of the model over the stable set 7 should be performed. ‘The details of each step are described next. ‘Stage J, The input is the set of @ posteriori information {yn, Ory M1), from which hy is computed 1S ost ath) = By me, 077) ‘The proidentfied model /1, is calculated using the following weighted Hult) = 2 w(k dha (De (10.78) FREQUENCY DOMAIN IDENIICATION 353. Here the weight values w(k) ate suitably selected to have the required convergence characteristics. Because of its reminiscence to signal pro- cessing procedures, itis often called “window” w. By replacing hy, in the last equation it is clear that this part of the algorithm is linear in the input data y, Hate) © tly) = ye So eKulze™) (10279) where Kwlz)~ whet (1080) whee is the kernel ofthe interpolation operator fy : C™(Fx.) —+ La (the input data are equipped with the f, norm). The design of the algorithm depends on the selection of the weight w or, equivalently, the Kernel Ky or the interpolation operator fy. Convergence is guaranteed for certain classes of weights w or operators fy. This will be discussed at the end of this section in further detail Stage 2. The optimal projection from Hy; € Ca, to a stable causal model Hig € Hoy ean be found by solving: War ~ Hille (1081) ‘This is precisely the Nehari problem solved in Section 6.6 (Theorem 655), The second step differentiates this identification procedure from the standard fitering problem, which arises in digital signal processing, where the noncausality of the filter can be solved via a delay. ‘The projection from £._ to Ha, of H.: should be performed only on its antistable (anticausal) part. To see this, consider H,y(2) in (10.78), which can be separated into its stable and antistable? parts H.a(z) = Hj(2)+ Ha(2). Is clear that the optimal projection only depends on H(z) because the stable partis already in 7,. and can be subtracted. AS in Section 6.6, a solution to the Nehari problem can be calculated in terms of the singular value decomposition of the Hankel matrix for the antistable part H(z) as follows FER dy, TE that Hidl2) = Hale) (10.82) * Recall thatthe anisole part hs its ples inside the unit circle inthis ese at =354 ROBUST IDENTIFICATION where d is the maximum singular value of the Hankel matrix of H,(2), (6,2) its corresponding right and left singular vectors, and N= (M 1)/2. This optimal approximation is rational and has degree less than ‘or equal to I. Nevertheless, care must be taken when using the above equation, due to the fact that the second term should cancel exactly the antistable part of H,i(z). More efficient numerical solutions pre sented in state-space form can be found in [21] and [332] for both the continuous and discrete time cases, The above procedures are untuned, since the only input data are the a pos terior’ information. The second stage is nonlinear in y, as opposed to the first stage, which is linear; hence the complete algorithm is nonlinear in the input data. In fact, it can be proved that there are no robustly convergent Ma. identification procedures that map the input data y linearly to the nominal ‘model ([234)) When solving a model approximation problem ({127)), the second step is not necessary if the number of samples is high enough, so that the val- tues of {i(k}, k= =I,...,-(M —1)/2} are negligible. This is due to the fact that there is no measurement noise. In that case only the causal values {hy(k), k=0,...,(M ~1)/2} need to be kept to compute the approximate ‘model, which corresponds to a rectangular unitary one-sided window. There- fore the whole procedure is linear in the input data, Moreover, it can be shown that itis robustly convergent ((139, 321)) In a similar way, in many practical identification applications when mea- surement noise is small enough only the first step of the identification algo- rithm needs to be applied ({9, 138]). In this later case, the divergence of the identification error due to measurement noise is Ole log( Convergence and Error Bounds 11 is possible to consider the identifica- tion errors at each stage in order to evaluate their convergence separately. It can be proved that the convergence of the whole procedure depends solely fon the convergence of the first stage. Due to the optimality of the Nehari solution to problem (10.81) we have: Wy ~ Hall < Waa ~ HL (10.83) where H € S is the system to be identified. Using the triangle inequality and the above equation, the total error can be bounded as follows: WH = Hull < Hos — Hil + WHat ~ Meal, 2 Hull, (1084) ‘Therefore the whole error converges if the identification error of the (linear) first step does, as M — 2c and ¢ — 0. Furthermore, the total worst-case error is bounded by twice the first-stage error. This indicates that the design of the algorithm and its convergence depend only on the selection of a convenient window function w for the first stage FREQUENCY DOMAIN IDENTIFICATION 355 Necessary and sufficient conditions for stage 1 to be convergent are pre sented next ({137). Lemma 10.12 The first step of the wo-stage identification algorithm is ro- busily convergent if and only if the following two conditions are satisfied: im sup pL = (1085) fim sup pill. = (10.86) where the preidentified model of step 1 (see (10.79)) has been separated into Hy, = Hps + ty, using the fact that the noise is additive. Proof. Both conditions are necessary by selecting 9 = 0 € N and H =0€ S, respectively. Sufficiency can be proved by applying the triangle inequality to Hy = Hpi + ni a8 follows sup Wa ~ Hil $ sup |p ~ Hl. + sup ll The convergence of these two conditions guarantees the convergence of stage 1. u Note that the above two conditions are met by any plant H € $ mea sured without noise and by the null plant H'=0 being robustly identified, respectively, Tn the case of model approximation, only condition (10.85) needs to hold for convergence, where Hy = Hy = Hl. Therefore the worst-case ident- fication noise-free error supycs [Hpi — H||,, i8 called approximation error. Similarly, sup yey [il is termed the nove esto, The uncertainty bound of the identincation algorithms and their convergence will depend on both these errs “There are many diferent windows that may guarantee convergence of the fst stage, and hence ofthe algorithm. For example, the triangular window, Known in digital signal processing asthe Bartlett window, given by fs ikl i087) { 0 KEN where N has been defined in stage 2. In this case, the sum in (10.78) cor- responds to the average of partial sums of the f,i(k), also known as their Cesaro sum, which guarantees uniform convergence ((139, 321) as N > oo. Iv can be proved that this type of window produces a noise error bound of wk356 ROBUST IDENTIFICATION «Hence it is optimal, because this is the lowest possible uncertainty bound, when K > € (see Lemma 10.8). Another rapidly convergent procedure uses the De La Vallée Poussin sums ((137, 233) In (137), necessary and sufficient conditions are given, guaranteeing that 4 particular window will generate a (robust) convergent identification al gorithm. These conditions therefore define a class of two-stage algorithms that includes, as special cases, all other two-stage algorithms cited in the bibliography. For cach different algorithm, there is a compromise between the approximation and the noise identification errors. From the many possi- ble convergent windows that may be generated, a subset that quantifies this ‘compromise can be parametrized as a function of an integer q as follows: ho naka nae asksn m4 w(k) (10.88) nak nq 0, otherwise with 0.< q 0) can be found in (21, 74) The algorithm depends on the noise sequence n, computed in Section 10.33. to establish consistency. Recall that if the a priori and a posteriori information are consistent, a noise vector 7. € A’ must be available such that the conditions in Lemma 10.11 hold. This is important not only to initiate the procedure. but also to guarantee that the idemification error bounds are valid, since the algorithm is tuned to the @ priori information, The algorithm proceeds as follows: ‘* Compute the Pick matrix P in (10.57) (which should be positive definite. replacing 7 by ‘# Define z, 2 et and wy 2 H(e!") for k= 0,....M ~ 1 and compute Ti) = Rep FREQUENCY DOMAINIDENTICATION 359 where po tw Ke= tu) ‘© The interpolating function is Ty(2)al2) + Trl) Hyle) = KB (1090) Tal 3)ale) + Taal) where g(z) € Bw isa free parameter In particular, if g(2) is constant, then Hj4(z) is of order M. ‘The degrees of freedom available suggest a further optimization step to select the optimal 7. and q(2), for instance to minimize certain criteria, The computational size of the problem corresponds to the number of experimen- tal data points As for the identification error, in (74] the following bound has been deter- ‘mined for the diameter of information: esky PO) 2minte, Ke) (2099) where Ky = (0). Proof. Consider the models h = mine, Ko), consistent withthe null ouput Equation (10.99) follows immediately from the general results in Section 102 a ‘The above bound is independent of the duration of the experiment and the nature of the input sequence; therefore it represents an absolute lower limit on the error. Next, we obtain bounds for the impulse response experiment that depend on the number of samples taken (230) Lemma 10.14 The information diameter for the &, identification procedure with an impulse response experiment is given by: Dy(Z) =2[ nine #00) + $290] (10.100) Proof. Define the following: fay = {Bh ESA nt = me) Next, consider the pair of models and measurement noise vectors: {mer= ay and {rns =to nk) = —n{k) math) = n(k) = hymeS, mmen Hence the above produces zer0 outputs to an impulse excitation, that is, Ehy.m) = Ella, m2) = 0. Note that this isthe worst-case model (maximum f, norm) compatible with the experimental output and in the consistency set S(0). Equation (10.100) follows now from Lemma 103 and the fact that $ and N’ are symmetric and convex by considering d(S(0)] = Wy ~ fal, = 2. (A) and « being positive a The following comments are in order: # There is a significant quantitative difference between the bound in (10.100) and the global one in (10,99), the latter being a conservative lower bound for the identification error. TIMEDOMAIN IDENTIFICATION 365 + The diameter of information tends to zero as M — 90 and € -» 0. This ‘guarantees the existence of convergent algorithms, for example, inter- polation procedures. ‘© For a given noise bound level, there exists 4 maximum number of ex: perimental samples M,, which decreases the diameter of information. In other words, for M >'M,, the diameter remains constant, which means ‘hat any extra experimental sample will not increase the amount of useful information. This fact will be used in the formulation of an interpolatory convergent algorithm ((230)) in the next subsection, Under certain conditions ((163}) it can be proved that the impulse input sequence is optimal in terms of the diameter of information, for a given experiment duration M. These conditions are related to the fact that the 4@ priori information does not exclude models in the consistency set of the null experimental outcome. The optimality stems from the fact that this in- put produces outputs directly related to each time coefficient. Therefore all the uncertainties are decoupled, that is, each measurement noise element is related to only one model coefficient. For example, for the step response experiment, the reader may verify that the diameter of information satisfies Dy (Z) ~ minle, (0)]+2 > minle,@]+ S24) (10.101) ‘The proof is based on the same arguments as before, but using the following model: min{(0),€), k=0 i(k) = } 2(—1) minf(K),€], 1k
2. Since AC) is based only on the experimental information, it cannot distinguish between both plants, but should converge for all plants in S(O), Therefore if sy lim, AQ) ~ hol = 0366 ROBUST IDENTICATION = Mo, such that [AQ <1, M > Moy € <6 Asa consequence for M > My and € < y, WAG) — Ath > Wall ~ AQIIh 2 1 (10.108) which contradicts the assumption that A converges to a plant in S(0) (see also (160) a In [237] the following necessary and sufficient condition for the existence of a robustly convergent ¢; identification algorithm is established: 36 >O-such that lh «ull > Sli, Yh Eb (10.104) ‘here lulas < 1 without loss of generality. The above condition excludes non persistent inputs for which u(t) — 0 as f — oo. Specific untuned algorithms, based on experiments performed with inputs known as Galois sequences, have been proposed in (193). Furthermore, in [238 it is proved that there are no linear robust convergent algorithms in f, as was the case for Ha, identification. 10.4.2 Identification Procedures Next, several tuned convergent algorithms for f, robust identification are presented. These are based on the impulse response experiment although, with minor changes, can be applied to the step response experiment as well. ‘The fact that these algorithms produce a nominal model in the consistency set makes it unnecessary to test consistency, as long as such a model exists. ‘Untuned convergent algorithms for additional input signals, can be found in [141, 193, 194, 306} Based on the Firstn ine, (K)] (10.106) Proof. This proof follows a similar one in [146]. For every h € S,n € N and for any algorithm A, the following holds: We AQ) < eA) (10.107) where e(A) Is the worst-case local error defined in (10.108). Consider the following models and measurement noises {nw = min((k), ¢} {ne —hy(k) m(k) = -fu() am{k) = ~ha{k) => nS, mmeN, Eth m) = Elta, m) Now, applying equation (10.107), Wry AO), < eC) and Way + A(O)I, < (A) => [hall <5 lr ~ ADI, + $s + ACI, < CA) To obtain (10.106), simply take the infimum over A. a Lemma 10.17 For the identified model Higa) = yz’, nM the following ienifation error bound holds: Wh Hh) < ne + S01) (20108)368 ROBUST IDENTIFICATION The value of n, that minimizes the above bound is the smallest integer n achieving ®(n) © € Proof. The first part follows from: We Ah = SAG) = vel + S100) < nes 20k) (10.109) For the second pat, define a(n) =ne + 370K) ‘Therefore a(t +1) ~ a(n) = € ~ O(n) is nondecreasing because © is nonin creasing. As a consequence, the minimum of a(n) is achieved at the smallest rn, satisfying a(n+ 1) > a(n) a ‘An intuitwe interpretation of the above result is the following. Since VA(K)| < (k) and & is nonincreasing, for values of k beyond the point for which (K) < ¢, there is no useful information on the system which can be obtained from the data It is clear from (10.109) that the value of n should depend not only on M but aiso on ¢ in order for the algorithm to converge. This is supported by the fact that there are no untuned idemtfication algorithms based on impulse response data ((160, 193, 194)) as proved before. There are many values of ‘n that may be selected for the algorithm to converge. In the sequel, three of them are presented. Lemma 10.18 The algorithms obtained by selecting n as either m(M,«) = min(M ,¢p ~ cy Ine) or na(M ,€) = min(M, co + c1€-"),1 € (0,1) and ¢a,¢; real constants, are tuned to € and convergent. A third, asymptotically globally op- imal algorithm tuned 10 (@,€) can be obtained using ns = min(M,n.), Me defined in Lemma 10.17. Proof, For n= ni, a(n) < (ca ~ eyln€ + Ie + Dig, PCR). The fist term vanishes as ¢ -+0, and ny 00 as M 0 and-e'~~0. Since € &), this implies that the second term also goes to ze¥0. Convergence for n =n) can bbe proved in a similar way. For m= ns, when M— 00 then 1+, and therefore the upper bound in (10.108) achieves the optimal value in (10.106) Also, for any given 8 > 0, there is always aq such that S22 y,, (Kk) < 8/2. TIMEDOMAINIDENTIICATION 369 and € > 0 such that (g + 1}e < 8/2. Because the lower bound in (10.106) is alobal, the last procedure is asymptotically globally optimal. a ‘As a special case of the above results, consider again the set of models ‘Yoolp, K), which includes the exponentially stable discrete time linear sys- tems considered in the Ha. identification setup ((146)). For p > 1 and K > 0, Hoo(p, K) € A(D), the disk algebra of functions analytic in the open unit disk and continuous on the unit circle. It is well known that the impulse response ‘of this class of systems is bounded above by the sequence {Xp-*, k > 0} This leads to the following corollary. Corollary 10.1 For the identified model #1", defined in (10.108), the identifi cation error can be bounded as follows: Kp Wh Hay < nes SO (10.110) This bound is minimized by selecting the smallest integer n > ma, with pst Inp (20111) Proof. Follows immediately from Lemma 10.17 by noting that %x(), K) CS when selecting (k) = Kp™* o Notes 1. Depending on the resulting value of n, it may be necessary to balance and truncate the identified model H",, to obtain a smaller-order model ‘This is a standard procedure, which may be simplified, avoiding the computation of Gramians, using the algorithm in [45, 127]. The new error bounds are given now in terms of the |: oF the Hankel norm, ‘Similarly although computationally more expensive, an optimal Hankel ‘model reduction may be performed, 2. For the case of systems that have all positive impulse response values, the &; bound coincides with the Ha. bound. Therefore the asymptoti: cally &; optimal algorithm is also asymptotically Hx, optimal. 3. When the value of n, is very high, for example in cases where M is high and ¢ is very small, care must be taken when balancing, in particular, when the first singular vector of the Hankel matrix is computed ([45, 127)), to avoid numerical errors that may increase the approximation 4. The identified system can be represented by a family of models, in terms of the nominal H, and the identification error bound, This family could be unnecessarily Conservative if the convergence of the algorithm370 ROBUST IDENTIFICATION is not monotonically decreasing with € -+0, In this case there might exist ¢ > « for which the identification error is smaller than the one considered and still “covers” the class of models S and noise N= Gel€) C fu(€,). All the algorithms presented here have monotonically decreasing convergence. Based on all Samples Next we present, a strongly optimal convergent algorithm, that is, locally optimal for any set of experimental data. This al- gorithm is based on the one in {160} Consider the vector of experimental data y of length M and di ho (k) = min yy + €, (8) (wo.12) fi (k) = max ly, — €, -(K)] (10.13) Note that Jy(k) is the lowest value of h(k) compatible with the a priori information, therefore consistent. Similarly, hy(K) is the upper bound on. ‘h(k) for each k, so that model h belongs to the consistency set. In addition, the consistency interval {hy (k), hu (k)] has length 2min [e, (k)} The algorithm is as follows: siy.0,0), = {AMM KEM oar The above algorithm is called central because it selects the model in the center of the consistency set; therefore it is also strongly optimal. This lat property can be proved as follows. Lemma 1019 The worst-case error for algorithm (10.114) is e[Ay, ®,€)] = > min fe, (4)] + JO, - > G9] (10.115) and tends 10 zero in the limit M +00 and € ~+0. The algorithms is strongly optimal Proof. The bound can be obtained as in Lemma 10.17, using the fact that AY. ®, €), — A(R] < mine, P(K)], & Mad since the model is in the center of the interval iz(k),/u(&)]. The above bound can be achieved with a suitable choice of system and measurement noise, It is clear that the consistency set is symmetric with respect to the ‘model Hiy = Aly, , €); therefore the algorithm is also central. TIMEDOMAIN IDENTIFICATION 371 Note that the first term of the error is equivalent to ner [6] (10.116) with n, defined in Lemma 10.17. This error converges to zero as (1/e,M) — oo. Using the above equation, it is easily seen that the total error can be rewritten as the global lower bound in equation (10.106). It follows that the algorithm is globally optimal. o In the special case where (k) = Kp-*, the bound (10.115) reduces to ((160)): we ie e[Aly,®,€)] = Zemin Ol SH (10.117) ‘The above algorithm is clearly tuned to all a prior’ information. Next, an- other interpolatory algorithm, which is tuned only to the model information, that is, ©, will be presented. This algorithm is given by: sign (4) min|ly«{,(0)] Os ks M— 0, kom AQ.) = { (10.118) Lemma 10.20. The worst-case error for algorithm (10.118) is: 14(9,0)) = 5 ming, 20(0] + ¥ ee] (20.119) which tends 19 zero inthe limit M oo and € ~+0. Proof. The proof is similar to the one above, considering the fact that [Ay ~ A(k)| < min fe, 20(6)] when [yx| < (If, however, [| 2 (k), the following holds: LA, ®)e = A(K)] < 26(4) Since the experiment and the noise have the same sign, in this ease LAly. ®), — AC} < WH) ~ yu] < € In both cases this leads to the desired bound. °372 ROBUST IDENTIFICATION For the particular case when ¢(k) = Kp-*, the bound (10.119) simplifies 10 ({166)} elAly,)] ~ Y minfe, 2049) y 10.5 FURTHER RESEARCH TOPICS Due to the fact that robust identification is a currently active research area, there are yet many theoretical and computational aspects that have not been developed fully. In this section, several items that have not been included in the main part of this chapter will be commented on and referenced. An excellent recent survey of the area of robust identification can be found in {199}, which includes an extensive list of references 10.5.1, Unstable Systems ‘The methods developed in this chapter apply to stable LTI systems. For slowly divergent systems, that is, with slow unstable poles, care must be taken when carrying out the experiment, especially when considering its duration ({127}). In general, neither time nor frequency response experiments can be Performed over open-loop unstable systems. Rather, they should be carried ut after 2 stabilizing controller has been applied to the plant. In that ease, the closed-loop input/output data should be used to identify the open-loop plant. The identification error should also take into account the fact that both. ‘model and plant are unstable. Thus the 7, norm should be replaced by a suitable distance function. In [196] a complete framework for the robust identification of strongly'® stabilizable plants is presented. The measure of identification error is com- puted in terms of the gap, graph, or chordal metrics. 10.5.2, Nonuniformly Spaced Experimental Points In the 2. identification two-stage procedures, it is assumed that the fre- quency experimental data points are equally spaced. Since this may not be the case in practical situations, the standard algorithm should be modified ac- cordingly. In (8] and (235] interpolation is used to perform identification with nonuniformly spaced experimental data points. In [127] a procedure that uses explicitly the unequally separated data points is presented, in the context of, model approximation, "A sronely sabiizable plant isthe one that can be nabilized hy a stable conte FURTHER RESEARCH TOPICS 373 10.5.3 Model Reduction In many of the algorithms presented in this chapter, the model order re- mains high. As a consequence, a further model reduction step needs to be performed and the model error bounds should be modified accordingly. This hhas been attempted in (142] for é identification and in (198, 253] for Hu. identification, 10.5.4 Continuous Time Plants ‘The robust identification framework presented so far considers both discrete and continuous time systems represented by H(z) in Section 10.3. Never- theless, in this last case, some extra considerations related to the frequency aliasing should be taken into account, Specifically, strictly proper stable LTT plants with an a priori upper bound on their roll-off frequencies should be considered. This problem has been treated in (9, 147) 10.5.5 Sample Complexity For application of the identification techniques to practical problems, it is important to know whether the experiments can be performed in a reasonable amount of time. Specifically, the following question may be asked. Which is the minimum time length of an experiment in order to arrive at a certain degree of accuracy, measured by the worst-case identification error? A general framework for this problem has been considered in [183] and specific results have been obtained in (163, 85, 243, 303]. These results are presented next Consider the FIR systems of length M, bounded experimental inputs [luli <1, and a (optimal over all algorithms and inputs) worst-case iden- tification error bounded by nf theca 20 as where € is the measurement noise uniform upper bound and the parame ter Q > 1 defines the level of uncertainty (Q = 1 corresponds to the optimal error bound 2e). For this case, the minimum length of the experiment dura- tion to achieve this bound is O@//2), with f() a strictly monotonically increasing function in (0,1), with f(0) = 0 and fim f(x) = 1 as x —» 1. Itean bbe shown that this isa tight bound, A consequence of this result is that the worst case clas of measurement noise (uniformly bounded by e) considered in time-domain experiments may be very conservative. Therefore, as suggested in [243], the class of measure~ ‘ment noise should be relaxed or noise should be considered in the average ‘On the other hand, the worst-case class of models does not need to be mod: ified. Tn the case of M., identification, the sample complexity is only quadratic in at ((199).374 ROBUST IDENTIFICATION 10.5.6 Mixed Time/Frequency Experiments ‘When applying robust identification techniques to either frequency- or time: domain experiments separately, it may happen that a model that efficiently fits the data in one domain may have poor performance in the other do- ‘main, For example, with only frequency experimental data points, a “good” frequency response fitting (small %,, error norm) may lead to a “poor” fit ting in the impulse response. Additionally, from an information theoretic viewpoint, more experiments produce a smaller consistency set of indistin- ‘guishable models, and as a consequence a smaller worst-case error, From a practical standpoint, in many cases robust identification algorithms are ap- plied to systems that may not be exactly LTI. Under these circumstances itis desirable to perform both time and frequency response experiments ((272)) to assess the validity of these assumptions. Thus, in these cases, an algorithm that combines both types of experimental data could take advantage of the additional data available Recent papers [77, 333] have proposed interpolatory algorithms that use data obtained from time-domain experiments and generate a nominal model together with an a, bound on the identification error. Therefore, although they combine time domain measurements with frequency domain error bounds, they do not use both types of experimental information simulta: neously, Alternatively, some recently propased algorithms explait hoth time and frequency domain experiments at the same time. In {273 necessary and suf- ficient conditions for the consistency of mixed time/ frequency experimental data for FIR systems are presented and an interpolatory identification algo- rithm is proposed, In [231] the above result was extended to IIR systems, based on a general interpolation theorem ([21}), which allows the simultan- cous interpolation of both the time and frequency experiments. The solution to this problem leads to a convex LMIls optimization, 10.5.7 Mixed Parametric/Nonparametric Models All ofthe results in this chapter address nonparametric identification of mod- els with a worst-case global bound. In many cases, part of the model has @ clear parametric structure. In these cases disregarding this information may lead to conservative results (see Example 10.4). This is usually the case in mechanical flexible structures, which have a well defined parametric model for the lower-frequency modes and an unknown higher-frequency behavior ‘that naturally leads to a nonparametric identification ((121)) In general, the parametric information does not appear explicitly in the 4 priori knowledge (K,p) usually considered. Therefore it is important to include it, if available, so that less conservative a priori estimates of (Kp) for the nonparametric portion can be derived. This is the case of systems ex- hibiting large peaks in the frequency response (for instance, a poorly damped FURTHER RESEARCH TOPICS 378 flexible structure), which must be “covered” by large values of K ({121)). This is a consequence of the fact that the usual a prion information characterizes only the smoothness and magnitude of the whole class of models but cannot distinguish among other properties, for example, low frequency model struc: ture. Instead, the a priori parametric knowledge provides more “structured” information, ‘There are several results that address this problem, for example, in [121. 175, 205, 232, 333]. In particular, the identification of affine parametric and nonparametric models using simultaneous time and frequency response ex- perimental data can be solved as a convex optimization problem ([232]). This includes cases of practical interest such as flexible mechanical structures, as well as models that can be described concisely in terms of a set of Laguerre ‘or Kautz functions of, more generally, any other basis of 2 ((148))APPLICATION EXAMPLES 11.1 SAC-C ATTITUDE CONTROL ANALYSIS. n Introduction The SAC-C satellite is a joint project between the Natioual Conunission of Space Activities (Argentina) and NASA (United States). It is essentially an Earth observation satellite that carries an Argentine CCD camera for Earth resources observation, a Global Positioning System (GPS) experiment from JPL (NASA), and a magnetic sensor from Denmark. The latter has a Star Imager and two Magnetometers mounted on an 8 meter boom, which will bbe deployed in orbit. The purpose of the boom is to magnetically isolate the ‘magnetometers from the spacecraft. The length of this boom makes it very flexible; therefore the model of the spacecraft cannot be considered as a rigid body (see Figure 11.1) The attitude objective is to point the camera toward the vertical with a precision of +1° around the roll and pitch axis; that is, the line of sight of the camera should be within a 1° square with respect to the vertical direction. Based on optical considerations, the jitter of the control should be on the order of 0.05%/s, Due to several practical limitations, it is important to have the simplest possible controller and the least amount of hardware, These limitations stem from the fact that the microprocessor onboard should not only implement the mission contro! mode but also several alternative safe-hold modes, as well as failure logics. Therefore a low-order controller is desirable. In addition, sensor and actuator hardware adds weight to the spacecraft, which in turn decreases the available payload. For this reason, only the minimum required rotations along the axes are sensed. Similarly only the minimum number 377378 APPUCATIONEXAMPLES Of torques are applied by the actuators. Thus only the roll and pitch axes are sensed by means of a scan wheel horizon sensor (SW) and the control torques are applied only in the pitch and yaw axes. These control torques act through magnetic torque coils (MTCs) and momentum wheels, the latter ‘built into the SW. In the design, the torque bounds of both actuators should bbe considered, due to limitations in the maximum magnetic moment of the MTC and the maximum size and acceleration of the momentum wheels ‘The desired attitude of the satellite in Mission Mode (MM) is the fol- lowing: roll along the velocity vector, yaw opposite to the vertical, and pitch perpendicular to the orbital plane (See Figure 11.1). Our goal is to design ‘@ minimal order controller to achieve this nominal attitude, This controller feeds back the roll and pitch angles and rates as well as the momentum wheel velocities, using static LOR feedback. The only dynamics involved in the controller arise from a yaw angle and rate observer (the only axis without sensor). Therefore the complete controller has only two states, is applied to only two of the three axes, and senses two out of three axes: as simple as it can be. The possibility of obtaining such a low-order controller and not hhaving to sense or actuate over all axes is due to the fact that the spacecraft and the orbital dynamics are coupled. ‘The attitude for the Earth Pointing Safe Hold Mode (EPSHM) is the same as for MM, except that the momentum wheels are kept at nominal constant speed. Here the control is applied using only the Earth magnetic held, through the MICs, The controller design is carried-out using a very simple procedure, that takes into consideration the limitations in the controller order, sensors, and actuators. This design is validated in Section 11.1.4 through an analysis involv: ing all the performance objectives as well as model uncertainties. Using the structured singular value, we assess the robustness of the controller against the unknown high-order dynamics of the boom, considered as global uncer- tainty. Robust performance is guaranteed using 2 for both modes: MM and EPSHM. Several time simulations of the complete system are presented in- volving random measurement noise and torque perturbations. Further details of this analysis are reported in [274] 11.1.2. Linear Model ‘The general nonlinear dynamic equation of the system is: Tut Teyt Ty + Ty = [pW 4 W x pW W xh tlle] SAC ATITUDE CONTROLANALYS'S 379. ‘ORBIT Xa FORWARD VELOCITY << A EARTH Figure 11.1, SAC-C Eorth observation satelite Here W is the angular rate of the spacecraft with respect to inertial space, represented as the rotation of a frame attached to the spacecraft with respect to the inertial reference frame, The inertia of the rigid part of the spacecraft is Fr and h represents the wheels momentum in the spacecraft frame. This frame is the usual roll (x), pitch (y), and yaw (z) axes mentioned before (see Figure 11.1), The external torques consist of the perturbation (7,), gravitational (T;), and magneticcontrol (Tx), that is, T = Ty + Ty + Tm The gravitational torque is due to the effect of the rigid body model of the boom and the magnetic control is produced by the MTCs. The perturbation torques are due to the aerodynamic drag (almost constant in pitch and yaw along the orbit) and the380 APPLICATION EXAMPLES magnetic and gravitational residual torques, both periodic and negligible as compared with the former: The wheels and magnetic torque col torques are used for control purposes the first one depends on the velocity change of the momentum exchange wheels and therefore does not change the total momentum of the spacecraft, the latter is an external torque that does change the total momentum. Note that these torques are applied only in the pitch and yaw axes, The reason is that a control torque inthe roll axis is unnecessary due to the orbital coupling between roll and yaw, as will be seen next. This reduces the number of actuators, therefore simplifying the design and minimizing costs The external perturbation torques are bounded but otherwise unknown Next, linearization is computed around the equilibrium conditions, cor responding to the nominal attitude of the satelite: the roll and yaw axes colinear with the velocity and local vertical vectors, respectively, and pitch perpendicular to the orbital plane, which completes the spacecraft frame (boom opposite to the local vertical). The roll, pitch, and yaw angles are de- noted as Op, Up, and Oy, respectively. The incremental values of the wheels momenta are 8h, in pitch and dh, in the yaw direction. For simplicity we can assume that Jy = diag [/, ly. Ie] Since the roll and yaw dynamics are coupled but independent from the pitch, we define the system state as xT = [x5 x7] ay ty = [Oe Oe Oy By dhe]", —xp=[ 6p Op Bhy|” ‘Adding the equations for the wheels torques Tey = —Siy and Tue = we obtain the state equations for both the {roll-yaw] and [pitch] dynamics: = Apt + Bry lies | +B.Tpe (12) pean «Bows [72 | + BTyy (13) { (0) = Ax(0) + Bult) + B,T, y= [22 Te] + aa xl + Dute) Oona Tent where x(1) has been defined in (11.1) and where: 1, wy 2 | SAC-C ATITUDE CONTROLANALYSS 381 Note that the system output consists only of the measured states, that is roll and pitch, Mattices Ay, Bry, Ap, Bpisn, By. and B, have been defined in [274 11.1.3: Design Constraints To simplify the controller as much as possible, a state feedback has been designed. The only controller dynamics will be the ones from the yaw angle and rate observer, due to the fact that only roll and pitch sensors are available (scan wheels) ‘A static LOR controller has been designed, taking into account the desired bounds for the different states and actuators when computing the weights to solve the minimization problem. These constraints are as follows Tue] [0012sing N-m d 2610- Nem Tr | | otzacose Nm aus) te) | foi Nm ‘This is due to the fact that the torque bound in each wheel is 12 mNm and that in pitch both add their torques projected through cosa (a < 10°). In yaw, on the other hand, they subtract both torques projected by sina. The bbound on the magnetic torque coils is computed as the average torque along the orbit. Each MTC has 15 Am? and the average magnetic field along the orbit is (V3/2) 0.4» 10-' Tesla, Therefore the actuator weight in the LOR problem is a diagonal matrix that has the inverse of these hounds as each element (Bryson's Rule, see Chapter 5). “The state weighting matrix is also diagonal with the inverses of the maxi- ‘mum allowable bound in each of the states. These are given by: bn oss Om r by o/s by r ah. |<] 0.06N-m-s co bp O1°/s op r ah] [006N-m-s The bounds in both 5h; and 8h, correspond to @ maximum deviation from the nominal wheel speed of 400 rpm, For a total momentum of 3.2 Nims. (2000 rpm, J = 0.0077) along the pitch axis, these bounds are 0.11 and 0.64 Noms, respectively. The observer involves only the [roll-yaw] dynamics, due to the fact that pitch is uncoupled, Furthermore, a reduced order observer will be used so382 APPLICATION EXAMPLES that the controller has the minimum possible order. The measured, estimated, ‘and input variables to this observer are, respectively, i a = af by 8 | Tw: melon 8 [#) 2 [Be] [* le rm Closed-Loop System The observer spstem is given by Ao | Bo Ao | Bor Bor fo] | au) Co | Do Co | Dor x2 with output fy and input [x2 7", In this case the observer gain can be computed analytically ([274). ‘The control signals, LOR gain matrix, and state feedback are oR Ts i. 4 | Tme| [Fe Fy Fie 03] | gy, a le Gir Oar Ons Fe |B] O18) h o ™ or a, where the feedback matrix has been partitioned according to the roll, yaw, {5h,, and pitch variables, From the above equation and the observer dynamics we compute the controller as follows: (Ao ~ Bo2FyCo)z + {Bor ~ Bor (Fy Dor + (Fe Fre )} %m Du O FrCo2~ (FrDoi + [Fe AD™[7 | =] or equivalently, a9) SAC-C ATITUDECONIROL ANALYSIS. 383, Note that the controller has the minumum number of states, given the avail able sensors (roll and pitch) and the minimum number of actuator outputs (yaw and pitch) given the the coupled dynamics. Using the above definitions for the controller, input, state, and output variables, the closed-loop system is given by: (A+ BD,C)x + BC yz + BpT» + BDyn 2 = Bre Az + Bin Ei Thus the state-space realization of the generalized plant G,(3) with states [27 27], disturbance inputs p 2 [ TF n? ]’, and e the output to be mini- mized is A+BD,C BC, | By BD, BC AL | Oe Be |_| Ac| & Teen Opn2 O54 Ones Ce | De DCG | Oe Dy 11.1.4 Robustness Anatysis The closed-loop dynamics computed in the last section correspond to the nominal system, which takes into account only the rigid body modes of both satellite and boom. Nevertheless, the flexible modes of the boom, should be taken into account explicitly in the robustness analysis, ‘These modes can be modeled as uncertainty in the high-order dynamics, using experimental data provided by the boom manufacturer ((295)), to com- pute the frequency distribution of the uncertainty. The boom dynamics is considered in this setup as follows: Hs = Grontsdrar v2 [ TH], ys 2[M] ‘The reason i that the diving input, which iis the Rexble dynamics Osilations the angla rotation of the spacecraft in rll and ich (he torsion around the yaw axis eligible) One other andthe pemtubation ofthe boom onthe stele sructre represented by the (intemal) trgues Applied nthe roland pth anc Gusto these esllation. The experiment in [295 determined lower requency yeamics ot! Hz. Therefore a convenient Weigh that “cover” the uncertain high fequency Ojai of the boom and assumes a 20% modeling eror at lower frequencies Seon Wa(s) = 04752 1,,5384 APPLICATION EXAMPLES The performance specifications call for minimizing the states and bounding the control torques of MWs and MTCs. Therefore a suitable performance ‘weight is a constant diagonal matrix with the inverse of the upper bounds in (115) and (11.6) as its elements. To perform the robustness analysis ofthe satelite the nominal and uncer. tain portions, and weights for robustness and performance are connected in the LFT form shown in Figure 11.2 ‘The closed-loop nominal system computed in Section 11.1.3 corresponds to the lower portion of the general system R(s) = Fe(P, K), connecting the inputs p to the outputs e, that is, Gy(s) = Ron(s). The total system, with- ‘out weights, connects the inputs (us,p) to the outputs (¥s,e) and has the following state-space realization: Ac Ris) = o10000000 000000100 G “feta vm To perform the robustness analysis the output weights must be added to the system as indicated in Figure 11.2. Furthermore, to normalize the anal- ysis, we scale the inputs in vector p so that the magnitude of the worst-case perturbations has unity norm. To this end we use a constant input weighting ‘matrix having the maximum perturbation torques and measurement noise as its elements. These perturbation torques are due mainly to the atmospheric mae va —Awnj2—1 Pe) ¢ fim} d y K(s) Figure 11.2. Unear toctional tanslormation setup for robustness analysis. SAC ATITUDECONTROL ANALYSIS 385, drag effect in yaw and pitch with a maximum of 10-° Nn, and tothe residual ‘magnetic torques in roll with a maximum of 10-® Nem. The angular measure iment error due to the scan wheels is 0.3°in roll and pitch, the only measured axes (recall yaw is estimated with an observer). The angular rate measure ment error in the same axes stems from the error in the LSB of the word that spans the total angle, this is 0.01". The rate is computed as the difference between two correlative angle samples, divided by the sampling time 7. For this reason, the rate “measurement” noise weight is similar to a notch fier centered al w = 2/T and scaled by 0.01*/s ols We (s+ 2/5)(s/207 +1) sels) au Finally, the rate measurement errors of the wheel velocity, which con- tribute to the error in their angular momentum in yaw (8h) and pitch (5h,), amount to a maximum of 2% of hyo. ‘The necessary and sufficient conditions to guarantee robust (internal) sta- bility, nominal performance, and robust performance are RS > [Ws(s)Ru (Wye <1 (a2) NP = [Wp(s)Ra2(5)Wir(5)I, SL Gusy RP e=> sup Ha [Wo(s)R(s)W)(s)] <1 ais) where Wo(s) = diag|Wa(s) Wp(s)], Wr(s) = diag[Wi(s) Wra(s)],and the ‘uncertainty structure A contains two full blocks, one for the boom high-order ‘dynamics, the second for the performance objectives. 5 Simulations ‘The following are simulated outputs of the main variables to be minimized, in Mission Mode (MM). Figures 11.3 to 11.6 show the angles, angular rates, and momentum wheel and magnetic control torques, all below the desired bounds. Figure 11.7 shows the robust stability and nominal and robust per- formance conditions, just below the value I ‘A different controller has been designed with the same methodology to control the spacecraft only with the MTCs, called Earth Pointing Safe-Hold Mode (EPSHM). In this case the wheels are kept at a constant speed and provide the necessary angular mmomentum to the system. In particular, the ‘weight on the MW control torques for the design is large enough so that the wheel torques are below 6 x 10° Nm, that is, almost constant speed, The robust analysis curves based on the infinity norm and the structured singular values are presented in Figure 11.8.386 ogress “3900828003605 aa09 000 000 ine) Figure 11.3. Rotation angus rors MM) otc) Yaw) Pa ete cons oo 0.009 t 009 oo oo A ! 0 360 20003004008 S003 tooo via) APPLICATION EXAMPLES Fo (i), Yam (Pach) ii Torque (nN) SAC ATITUDE CONTROL ANALYSIS. 387 i ores in Yaw (0) & Pa (-) sine 6) Figure 11.5. Momeniurn wheel contol foraues (MM), MIC toques in Yaw so) & Pe) Figure 11.4. Rotation ate erors (MM). sine (6) Figure 11.6. Magnetic contol lorques (MM).388 APPUCATIONEXAMPLES jj Nominal ana Robust Performance sis ah Robust Stabilty ESO RSEEE HEEL Av w (adie) 4 w (cadis) Figure 11.8, Robust stabilty and nominal and robust performance conaiions (EPSHM). CONTROLLER DESIGN FOR AD.O PLANT 389 11.2 CONTROLLER DESIGN FOR A D,O PLANT In this section we illustrate by step, the robust control methodology through an application (0 a D;O plant. The mathematical model of the plant is a linearized and simplified version of a nonlinear system with a delay. This particular application is a standard problem in the process control area. The Purpose of this example is to assess the incidence of model and disturbance luncertainty in this typical application. The model uncertainty arises from the time delay and the gains and poles of the transfer function, which depend on the particular linearization point. The usual control structure to handle systems with time delays is the Smith predictor ([289)). This classical method in process control has the advantage that, when the delay is known exactly, the controller designed to achieve (nominal) stability needs to take into account only the rational (nondelayed) part of the plant. In this example, based on the Smith predictor, we perform the robust analysis and design of a controller for this particular uncertain system, which exhibits all the relevant issues of process control problems. The following example is based on [268), and the interpretation of Smith predictors with uncertainty is described in [211] 11.2.1, Model of the Plant ‘The plant to be considered is an experimental process that obtains D;O by isotopic exchange between H;O and HyS (see (128, 206)). The plant has already been constructed but is not yet operational. ‘A general diagram of the process can be seen in Figure 11.9. This diagram is a simplification of the process flow sheet (there are more cold and hot towers in the actual plant) but retains the main issues of interest, The input water flow and liquid flows are marked with full lines and the gas flows with dashed ones. Two compressors have been included, as well as three heat exchangers that control the temperature in the process. The greatest D;O concentration in each stage is obtained between the hot and cold towers. A fraction of the partially enriched water of the first stage is sent to the second stage for further enrichment, The product is extracted from the second stage after the H3S stripper eliminates the gas. The isotopic exchange process that takes place in the cold and hot columns, respectively, is the following H,0+HDS “© HDO+H,S HDO+H,S C H,0+HDS The controlled variable is the ratio L/G, with L and G representing the liquid and gas operational flows, respectively. The value of G remains almost constant duc to the fact that it circulates on a closed circuit, as seen in the same figure; therefore the control is performed by changing the value of L390 APPLICATION EXAMPLES sho) Figure 11.9. Schematic dlagram of the D,0 plant. ‘A general nonlinear model of the whole process has been implemented ({206, 228}) for simulation purposes. This general model has been obtained from physical considerations and is too complicated for both analysis and control design. Therefore the controller analysis and synthesis are based on a linear model that takes differential H,O flow as an input and differential D;O ‘outlet concentration as the output. The linearization has been made at 0.01%, 0.3%, and 1% disturbances in the input water flow, under the restriction of 15% outlet concentration of D,O. The structure of the linear model remains the same for all three linearization points, but the parameters (gains and poles) differ from one point to another. Assuming continuity between the linearization points (water flow disturbance) and the model parameters, the linear process may be described by a set of models with the same linear structure but with parameters lying in real bounded intervals, ‘The delay is a consequence of the measurement procedure for the out- put variable. It consists of the collection of samples at each column, which takes around 15 minutes per column, with a transportation time between columns of another 15 minutes. All these samples are measured with a mass spectrometer, to obtain, after approximately a 4 hour test, the DO concen: LE CONTROLERDESIGNFOR ADO PLANT 391 tration. Depending on particular conditions, the elapsed time between the sample extraction and the final output value may differ significantly. There. fore for analysis purposes, we consider a nominal delay of 5 hours, obtained as an average value, with an uncertainty of £1 hour, In this linear model we also consider additive disturbances in the output. ‘These disturbances include several liquid and gas flows, which go through part of the process and which, for operational purposes, are bounded. If these bounds are exceeded, the required level of D;O extraction will not be ‘met (or we will have no extraction at all). Due to the fact that these flows circulate through most of the columns, we may assume that the dynamics of the disturbances is very slow (see Figure 11.9). For the above reasons, we ‘model the additive disturbances at the output as a set of bounded signals with unknown and slow dynamics, ‘The set of linear models representing the process is the following _— «fe, Gls) = 0 |e ot ae tye | Teme) Kee Miu 4 71 €lypwjh, vj >0, 4 ais) T=%+8, [8l<1, 8ER, where the nominal delay is x» = 5 hours. The uncertainty intervals for poles and gains are obtained from their values at the three linearization points mentioned before ([228)). The disturbances are represented as a weighted and bounded set, that is, {We(s)d{s)|ldl2 < 1). Here W,(s) takes care ofthe fact that the distur bances have a higher energy content in the lower frequencies (slow dynam ies). For the controller design step, a rational approximation of the nominal time delay needs to be made. Many different methods exiss, but @ standard simple procedure is the use of Padé approximants, Nevertheless, we will verify the robust performance condition with the exact values of the delay, without approximation errors, 11.2.2 Robustness Analysis In this section we obtain conditions on the controller such that stability, performance, and robustness are achieved. Due to the fact that the system hhas a delay, we work with the Smith predictor control structure, but including the uncertainty in the delay, poles, and gains. The classical Smith predictor structure ([28]) is presented in Figure 11.10, The stable plant G(s) consists of a rational term Gy(s,p) with a vector392 APPLICATION EXAMPLES of uncertain parameters p. p; € j.b,) (in this case the poles and gains in (11.15)) and an uncertain delay 7 = 7, + 8,[6] < 1. Define the nominal model 88 Gals) = € "*Gyy(s,po) which wil be used for the design. Next, we analyze step by step nominal and robust stability and performance Nominal Stability The Smith predictor operates by subtracting from the feedback signal the difference between the nominal model and the core. sponding minimal phase part (Figure 1110) This is equivalent tothe incl sion ofa prediction in the sytem since the feedback signal f() is exaely the output ata future time (1+), This procedure i elective only when theres Complete knowledge of the system (he parameters ake their nominal values +=, and p= pp) otherwise we will not be able to havea corect prediction oy From the same figure, we observe that, the nominal transfer function be- tween u and the feedback f, equals Gy, © Gy(s,p.). Therefore in terms of nominal stability, only a controller K(s) that stabilizes the nominal minimum phase part Gwo(s), regardless ofthe time delay 1», needs to be designed This isthe main advantage ofthe classical Smith predictor, but this esti to longer valid when considering performance of robustness issues, a5 will be seen nex Nominal Performance Based on the results in previous :chapters, nom- inal performance depends on the transfer function between the disturbance and the output error signal. Figure 11.11 yields the following transfer function in the nominal case, that is, G(s) = G(s) vals) = [10% Tu(s)] Walsdd(s) (1.16) AO Figure 11.10. Smith precictr. CONTROLER DESIGNFORAD.O PLANT 393 where we have defined the complementary sensitivity as Tals) © Gass) K(s) [1 + Gael )K(P The necessary and suffiient condition for nominal performance is Nominal performance => jlya(slz $1, ¥lld(s)I <1 si qt) Tu (s}] Wal) This leads to the following 21. control problem: Ip- sary hll “°° TaCO WAC oy An important point is that the nominal performance condition includes the nominal delay 7,.. As we mentioned before, the classical Smith predictor does ‘ot simplify the analysis step for performance, as it did with nominal stability. A controller that solves equation (11.18) can be obtained by a standard Ha. ‘optimal control algorithm, with a previous approximation of the time delay. The optimal solution without approximating the delay can be obtained from [116], but this approach cannot solve the robust performance prahlem Robust Stability In this section we analize the closed-loop stability for all Possible delays between 4 and 6 hours and all possible parameters p [poles and gains in (11.15)] of the minimum phase part of the plant. Wale)d A lyautye Ks) [ber ou(s,p) Figure 11.11. Smim predictor wit: uncertainty.394 APPLICATIONEXAMPLES A practical! way to analyze models with parametric uncertainty is to “cover” them with dynamic uncertainty. In this case we have used a mult plicative dynamic description for the set of plants, as follows: ME {[1+A.Wa(s)Gmols), eC y (Sul <1} (1119) In this application, there are two different sources of parameter uncer. tainty, which in principle could be covered separately. This will produce a ‘smaller” set of plants than in the case where the coverage is performed over both parametric uncertainties together. As a consequence, the analysis results for the “smaller” set of models will always be less conservative than in the case where both parametric uncertainties are globally covered, For the delay, we transform the dashed block of Figure 11.1] into a nom: inal system with global multiplicative uncertainty, with p = py. To this end, consider the family of models GF Gs) — Gals) + Gols) (11.20) = eles -re out, BER, jl mate =| at2t) Figure 11.12 shows the above condition in a magnitude Bode plot. A rational " ama hoc analysis considering directly parametric uncertain could be made, bu iis ot the purpose of thin example. The resus on analysis of parametnc uncertainty srutures ae ee? Snslyieal methods that solve particular eases (and donot i hs spication) oF algorithmic ‘methods that solve general cases but ae computationally involved, infact « NP-complete prob Jem (jf), 24). There i 9 yet a sates method that supports a peneral parametric anal ‘Am cxcllent survey ena he found [35] CONTROLERDESIGNFORAD,O PANT 395 7 10 10 10% 10 107 10% wads) ‘Figure 11.12. Dynamic weight W(s} covering delay uncertainty. approximation Ws(s) for the weight can be computed from the same figure, Note that the set of models that covers the delay uncertainty is independent of the nominal delay 7, Following a similar procedure, a set of models with dynamic multiplicative uncertainty, which covers the parametric uncertainty in gains and poles can be obtained. For briefness it will not be repeated here, but the corresponding, ‘weighting function may be seen in Figure 11.13. ‘The necessary and sufficient condition to guarantee the stability of all ‘members of a family of models with multiplicative uncertainty” is IMs) Tus) <1 (122) From the above equation, we observe that the critical point in terms of robust stability i the frequency w. for which |W,(ja.)| > 1. From (11.22) itis clear that, for w > a, the values of |Tiy(jw)| should be less than one to achieve the robust stability condition. In some sense this frequency defines an upper bound for the closed-loop bandwidth, the latter interpreted in terms of Ty (5) 2 For the family M the condition is necessary and sufficient when replacing W with Wy, but since @'C AM forthe actual set of models 8 only sufeient396 APPLICATIONEXAMPLES Weights for uncertain delay & poles 10! Total uncertainty weight +04 : a Uncertain pote-gain weight 107 10 (ee ay w(radis) Figure 11.13. Dynamic weights covering pole gain and delay uncertainties, Tota Uncertainty wight is Wo(5) ‘The above condition could be useful in determining the relative impor- tance of both types of uncertainties for the robust stability of the plant. By comparing the weights in Figure 11.13 we observe that, for the frequency range of interest, the curve due to the pole and gain uncertainty remains be- low one. Therefore the limiting factor in the robust stability condition is the delay uncertainty, Nevertheless, in the lower-frequency range, the uncertainty in poles and gains is larger than the one due to the delay. Therefore, from a practical point of view, we should consider the total uncertainty weight W.(s), which covers both types of parametric uncertainty (see Figure 11.13). Finally, it should be noted that the Smith predictor structure allows this analysis to be independent of the nominal delay 7. Only the uncertainty is taken into account, which appears indirectly in the weights Wa(s) (see equation (11.21)) and W, (3) Robust Performance For robust performance we need to achieve the performance objective for all models in the set {[1 + &,Wu(s)] Ga(S), [Mul < 1). From Figure 11.11 we obtain the following transfer function between d and the output signal yals) = [10 Ty (s)] [1+ AaWals)Tu(s))! Walsddts) (11.23) ‘CONTROLLER DESIGN FOR AD, PLANT 397 where we have replaced G(s) = Gy(s) + 4uWal)Grols)- A necessary and sufficient condition for robust performance can be obtained from the above equation in the following way: | et Tue) + WaT)" Wa] SA, VS = Je, Nal <1 = |(l- Ty) Wal < inf 10+ WaT wl V5 = Jo ) Wal < inf 10 DI J ee |(1— Ty) Wel +1WuTul <1, Ys = jo (1.24) where we have selected a worst-case || < 1 at each frequency. Equation (11.24) is a combination of both the robust stability and nom- inal performance conditions of previous sections. The main difference with a standard mixed sensitivity problem is that both terms of this equation are applied to different plants—one to Gyzo(s) and the other to a combination Of Gyo(s) and Go(s). To recast it as a standard mixed sensitivity problem ‘we apply the following transformation Le Ty(s) = (Ie) +S u(s) (1125) where Sys) © 1+ Gols) )}" isthe sensitivity funtion, The goal is ow {o design a controler that solves: ani ggg Sm COWA)| + Wels) Ta] (1126) with Wy =e" Wa(s), The factor y not only provides relative weight be- tween the nominal performance and robust stability conditions, Dut also compensates forthe extra term (1-~€""*) in equation (11.28). At the de sign stage in the next section, the value of 7 is computed so tht the robust performance condition (11.24) is achieved inthe least conservative way. 11.2.3 Controller Design ‘The mixed sensitivity problem of equation (11.26) can be solved in two ways. ‘Hoe optimal control theory can be used to design a controller that solves: Wa(s)Ta(s) ll Wals)Suls) a that is, it minimizes the norm of the transfer matrix from the input v2 10 the outputs (z1,22) in Figure 11.14, Recall from Section 48 that the optimal value of (11.27) is at most twice the value of (11.26).398 APPLICATION EXAMPLES wile as wy cd K Lew. Figure 11.14. Mixed sonstivy probier, Alternatively synthesis can be used to design a controller that minimizes directly the value in (11.26). This is equivalent to cosy xe) —Wals)Tu(s)— Wals)Tu(s) {[ yWals)Su(s) yWe(s)Sw(s) } a the argument being the transfer function matrix between inputs (11,3) and outputs (21,22) of Figure 11.14. ‘The sensitivity Sy (s) and its complement Ty(s) are both functions of the ‘nominal minimum phase part of the plant Gy(s), which has the following state-space description: she 0 0 o 0.395 x 10° Oe nn 0 08 | ome ings) ORT gO nace 0 [*T gar xcao-e | 0 0 oo 0.166 x 10° we =[1 11 axe ‘The uncertainty weight, obtained from Figure 11.13, and the performance ‘weight from design considerations are, respectively, ) = 191 86% 1088 41 Wee 3x 10541 (129) ‘When equation (11.26) is solved using M4. optimal control, the resulting. ‘optimal solution does not necessarily achieve the optimal robust performance in (11.24), where we have a combination of minimum and nonminimum phase models. Therefore we have iterated over until the design achieves the robust performance condition (11.24) in the least conservative way. In fact, CONTROLER DESIGN FOR AD.O PLANT 399 for this case, the best performance has been obtained for the first iteration of a w-synthesis procedure, that is, the 74. optimal design.’ and a value of Youn = 1.58. For the design, we used a third order Padé approximation of the nominal, delay. Nevertheless, 10 verify robust performance in (11.24). the exact val ues of the delay have been taken into account. The plots of the robust and nominal performance and robust stability can be seen in Figure 11.15. In the same figure we can see the structured singular value 4 of the closed-loop system. We remark that, since the design has been made using the minimum phase part of the system Gyo(s), the value of is below the actual robust performance condition (11.24), which includes the delay. For the value Yaa the latter is very close to unity at w, = 2-10~ rad/s. In Figure 11.16 we can see the response of the system to an 11 hour sinusoidal disturbance of unit magnitude for 4,5. and 6 hour time delays. The frequency of the disturbance has been selected as a worst-case value according to the peak on robust performance in Figure 11.15 (w, = 2-10 rad/s, worst-case disturbance) Due to the fact that the present design optimizes condition (11.26), a 09 08| 07| o6| os| o4 03 02 oat ogee Te wiradts) Figure 11.18, Robust and nominal performance, robust stabilly, and y (forthe design sructute) “yma sandard problem, where the semsitvity and ts complement correspond tthe same mode! the performance peadoce hy synthesis design should have ben Petter than or east equal teste 0 optimal one400 APPLICATION EXAMPLES (Oe i0 eco eeeon] 40 60 6070 80 time (hours) Figure 11.16. Time response for worst.case disturbance (tobus disturbance rejection), possible improvement could be obtained by the use of methods that con- sider directly parametric uncertainty and the fact that the system is infinite dimensional. This is beyond the scope of this example, only intended to illus trate a complete step-by-step robust control analysis and design procedure ‘on a typical process control application. 11.3 X-29 PARAMETRIC ANALYSIS In this section we perform a robust stability analysis of the Jateral-directional controlled dynamics of NASAYs X-29 aircraft. This is a true multivariable control problem due to the coupling of the roll and yaw axes. The rigid body ‘model of the aircraft with a controller in place has been taken into considera- tion and the analysis consists of computing the multivariable stability margin (or equivalently j.) for the uncertainty in the aerodynamic coefficients. Fur- ther details of this example can be found in (265). 11.3.1 Linear Model First we' introduce the notation used throughout this section ({189]) in Table 11.1, corresponding to Figure 11.17. An illustration of the aircraft is shown in Figure 11.18. X29 PARAMEIRIC ANALYSIS 401 Table 11.1. Aiteraft parameters = Pitch rate ® + Lateral ‘acceleration () + Roll rate (p) + Longitudinal acceleration (a,) Figure 11.17, Lateral acceleration (@) Reference span (0) Roll and yaw moment coefficients Lateral force coefficient Gravity acceleration (322 fy/¢) Roll, piteh, yaw axes moments of inertia (slug) Roll and yaw axis cross inertia (slug) Altitude (ft) Mass (slugs) Roll, pitch, yaw rates (rad/s) Dynamic pressure (Ib/t) Reference area (ft) Sampling period (0.025 5) Velocity (ft/s) ‘Angle of attack (rad) Sideslip angle (rad) Aileron and rudder deflection (rad) Pitch and roll attitude (rad) Rudder Elevator * Normal ‘accelerator (@n) 4 ED + Yaw ae (0) Parameters of e generic aircraft.402 APPLICATIONEXAMPLES X29 PARAMETRIC ANALYSIS 403. 4 = pegtandsind +rtand cose (11.33) ee dy = 4 y Se (11.34) eam with the aerodynamic coefficients given by Cy = CyB +B Cr, 9, a3 : _ pb eth nse GB + C84 C1,8 + 6,20 + 2? 1136) oe eb E Cu = CyB Cag Ba+ Cn Bs CBP 4 Cy (37) The subscripts fi, ,, and 8, indicate the variables with respect to which the derivatives Cy,C), and C, are taken; for example, Cy, represents the deriva. tive of the lateral force with respect to the sideslip angle. These derivatives are computed by parameter identification procedures applied to the experi mental data. In addition, error bounds on these parameters are obtained as well ‘The linearized state-space equations, considering the input from the dif ferential aileron and flap w= (8, J” and the output y= |p r ¢ ay]? a= EArt el Bu (11.38) y = Cr+Du 2139) where the state is x=[ pr 6]!. Matrices (A,B) explicitly contain the uncertain aerodynamic coeficiems Ci, Cys Ch Cras Oy and Ca Gn, Sina cosa (4/¥)e0s0 Dace ae ane 20 7 an40) nT Figure 1.18, NASAXZ ocr Ini nr, B ly Cry M41. a GC, Cay, : y The lateral-iectional linear mode of the atralt (189) is presented next oD 4 ‘ additional matrices are B= Bey epsina—reosa+ £singeose (11.30) oe 7 v 10 0 0 Ply — Hes = GSbC) + rqlly ~ Ie) + pals 131) _|o Qo -he 0 fl a O -he hb 0 11.42) He bs ~ §80Ce+ path ~h) rah 0132) 00 04406 APPUICATIONEXAMPLES o 1 0 0 0) Or 8 i, to c o 0 014 (143) Cry ~Za/8 Xo le 0 o 0 0 0 lol 0 14a) Cr. Cr, Here we have defined mn = GS/mV, 4 =4S/mg, gy = 45d (1145) 4 = GSP LV, n= (gC, rane) (1146) dy = aC, +4(ly— 1), da =a, #g(le—4) (11.87) dx = = (Gn, ~ Glas), dae tan 6 (a148) ‘This model has four open-loop poles: the dutch roll, a complex conjugate pair that applies to the roll and yaw dynamics; the roll mode, a real stable pole; and the spiral mode, a low frequency unstable pole. ‘The controller structure for all flight conditions is the following: (5) = Ka(s) Kyls) (i149) where Ke(s) is: [ore Mins «Mua /2+ Xan] one (1150) and the elements of Ky(s) are given by (Ky) = Pils) Sp(s) [Ke —a-Blend-L(s)-K3] (1151) (Kyra = Ks S365) -Fo(5) Ps) L(s) (11.52) (Ks)is = —Ks-Po(s)-L(s)-g- Blend/V (113) (Kyu = Ke Fels) Pus) (assy (Kyla: = Puls) Sp(5) -[Kig ~ a Blend - L(s)] (iss) (Kolar = Kir Sis) Fe(3)- Pals) L(s) (1156) (Ky)zs = —Kir- Pals) L(3) -g- Blend/V (asa) (Ky)nu = Kun Fala) Pals) (1138) where Blend is a parameter. The elements of both matrices are shown in Table 11.2 X29 PARAMETRIC ANALYSIS 405 Table 11.2. Mattices Ko(s) ane Ko(3) Aye) Fourth-order differential ap actuator model AiG) Fourth-order rudder actuator model Zs) Padé approximation due to Sample & Hold delay Pus) Roll, yaw, and lateral acceleration prefiters Pus) Roll attitude prefiter 5,18) Second-order Roll rate sensor model Sis) Second-order yaw rate sensor model FAs) Second-order yaw notch fter Fis) Second-order lateral acceleration notch filter Lis) ‘Analog equivalence of digital filter L(z) 11.3.2 Results ‘The robust stability analysis was performed over two different linearization points. The first one corresponds to the nominal design flight conditions, the second one to a critical condition at sea level and with large angle of attack. These conditions are shown in Table 11.3 ‘The mathematical structure of the controller and nominal model of the plant remain unchanged, although the parameters have different values ac- cording to the fight condition. In the controller these parameters are Kz to Kis, Xepav Xin, Xzpas and Blend; in the nominal plant model the parameters are the aerodynamic coefficient derivatives. In this last case not only do the ‘nominal values of the parameters change but also their uncertainty bounds. Before a branch-and-bound analysis algorithm is applied to this example ((266, 281]), we proceed to compute the structure of the characteristic poly nomial f (5,4) as a function of the uncertain parameters mentioned before: this is G(sa) = D+C [se 40a] 8) 1159 f(s,A) = det [I + K(s)G(s,4)] (11.60) {eble 1.3. Unwotionconatons _ No Nea Mach No. 09 04 was som seni ° oe 0 ¢ sss als : Bat a A os 2a | ssn ar406 APPLICATIONEXAMPLES A = diag|5, 54) 161) with 6 the bounded uncertainty of parameter p, around its nominal value Pov hati, p= Pa +5, [8 < 8, A'symbolic manipulation solver has been used to determine the mathematical structure of this equation, in particular the exponents of the uncertain parameters, to determine which are repeated ‘ones. The mapping theorem can be applied to the parameters that appear linearly. On the other hand the ones raised to powers larger than one appear in a polynomial fashion in the characteristic equation and therefore should be restricted by the equality contrain's mentioned in Chapter 7, reducing them to a moltiinear dependence. in this case, the characteristic polynomial /() 's multilinear in all parameters but to, in this latter case with a quadiatie dependence. The new set of parameters generated by the adition of equality restrictions (see Chapter 7) Pi=P2 = Cip~ Poi = Peat (11.62) >= Pa = Cay = Pos + 8s = Poa + 5s (11.3) Ps = C= Post 8s (11.64) Po = Eng = Poo + 8 (11.65) b= C= Path (166) Ps = Cny = Pow + 8x (11.67), ‘The bounds on the uncertainties are 8) = 5) = 333%, 5 =3,=137,% K= 128% By = 58%, 5, = 189%, By = 14.3% (11.68) {or the fist linearization point and Bb =39%, == 111%, H=DO% iggy A= 636%, = 685%, | = 226% for the second. ‘The multivariable stability margin Km(jw) [1a,(Jw)] Was computed for a certain range of frequencies. The minimum values for each flight condition were km(Ja) = 3.15 [ia,(s0) = 0.317) at w = 3.5 rad/s in the fist case, and Kyg(Jo) = 2.89 [19, (Jen) = 0.346] at « = 10 rad/s for the second case. Thus the design is robustly stable. ‘The worst-case parameter combination, from the stability viewpoint, can also be obtained from the algorithm. This is the vector p* whose image frst reaches the origin of the complex plane for the smallest scaling factor k. This is a very important element in the analysis and redesign (if necessary) of the controller CONTROLOF A DCTOE RESONANTCONVERTER 407 This worst-case combination of parameters for flight conditions | and 2 were, inthis case, Ai = diag[) By Be 086-5) - i &) a1) ds = diag [ds 8 8-8-8. (O18-]) Be ] LTH The design of K(s) was made at both linearization points for the set of ‘nominal parameters py = py + 4y. with Ay as follows: ay = diag(—5 -& 5B -3 -& ] -&] C1172) {As observed, this disagrees with the values obtained in (11.70) and (11.71) for both fight conditions. This is a consequence of the fact that itis difficult to develop a physical iuurion on how the parameter variations will influence the stability of the system, especially in MIMO plants with a large number of uncertain parameters, As stated previously (see Chapter 7). the exact computation of ky (ja) or 443,(70) in the cases of parameter uncertainty has an exponential time compttation with the number of parameters, Nevertheless. in many cases, as the one presented here, it can be computed in a practical way by branch-and: bound procedures. Further details of the algorithm used in this ease ean be found in (266, 281). and this particular example has been developed in (265) under a NASA grant. 11.4 CONTROL OF A DC-TO-DC RESONANT CONVERTER 11.4.1 Introduction DC-to-DC resonant converters have been the object of much attention late since they have the potential to provide high-performance conversion. al- lowing for smaller, lighter power supplies. However. they require using a control circuit capable of maintaining the desired output voltage under dif- ferent operating conditions. In this section (based on the recent paper [65)) ‘we illustrate the z-synthesis technique by designing a controller for a paral- lel resonant converter. In addition to guaranteeing stability for a wide range of load conditions, this controller rejects disturbances at the converter input while keeping the control input and the settling time within values compatible with a practical implementation, 11.4.2. The Conventional Parallel Resonant Converter Figure 11.19 shows a diagram of a conventional second-order parallel res: nant converter. The combinations of the diodes and transistors form bi- directional switches operating at 50% duty ratio, Thus. in cach switching period. the resonant circuit L-C is alternatively excited by +Vy and — Vy408 APPLICATIONEXAMPLES [ts rtp EY Figure 11 Conventional second-order PRC cireutchagram, ‘The large output inductor L,, and capacitor C, are used to minimize the load effect on the resonant capacitor voltage and to ensure the constant out put voltage through the output circuit (37). As for notation, the resistor Ry and the voltages V_ and V,, represent the load, the line (input), and the output, respectively. Throughout this section we use as nominal parameters the following val- vves, taken from the design example in Chapter 2 of (37) L = 4.18 aH C = 11300F Ry = 208.330 = 10mH C= 47 nF V, = 100V Vv, = 250V f, > 200kHz VEC = 60.390 CONTROL OF A DC-TO.DC RESONANICONVERTER 409 For convenience, we introduce the following normalized variables: v, where the resonant frequency f, 11.4.3 Small Signal Mode! Under steady-state conditions it can be shown that, for a PRC operating in the continuous conduction mode [37], there are four circuit modes in each ‘witching period. Thus the converter is a nonlinear, variable structure system, with its steady-state trajectory uniquely determined by the normalized switch: ing frequency F,, and the load condition Q,. For a given operating point, a discrete time, small signal model of the converter can be obtained by using a perturbation method. The sampling time for this discrete time model is equal to T,/2, where T, = 1/f, is the switching period. Therefore it follows that this model is correct under small-signal perturbations with frequencies up to the operating switching frequency 1, = 2nf, = 1.26 x 10®radjs ‘The discrete time model from the normalized switching frequency Fyy and the normalized line Vag to the normalized output Vp (10 simplify the nota- tion, we use the same variables for both the steady state and its perturbation) at the nominal operating point is given by the following state-space rcaliza- tion [37|: X(k +1) = AX(k) + BUR) Volk) = CX(K) where 0.2767 0.6108 0.6644 0.0053 0.0075 0.9387 [ 0.8219 0.5504 -2 “| 6.4684 0.4834 B= | 106774 1.9499 = 0.0002. 0.0162410 _-APPUCATION EXAMPLES C= |0 0 345] The state variables and inputs are defined as X= [iath) vad) tal)" UCR) = [Fal Vag] where iy, 3nd Jy, are the normalized resonant inductor current, capacitor voltage, and output current, respectively. 11.4.4 Control Objectives Figure 11.20 illustrates the diagram used for control design. In the small signal model of the converter there ate two inputs: line voltage Vy and switching frequency F,,. The objective is to synthesize a controller having 4 input the error signal (obtained by comparing the output voltage versus the reference input r) and as output the switching frequency, F,, such that the output voltage is kept at a prescribed level (in our case V,, ~ 250 Vie. Vo = 2.5) at all operating points, This problem can further be divided into four parts ‘© Line Regulation (Nominal Performance). The line voltage is often unreg, tlated and could have a substantial range of variation, typically around + 20%. This variation will be modeled as an external disturbance, thus leading to a disturbance rejection problem. Performance specifications for this type of problem are usually given in terms of time-domain quan- tities, such as: 1. Zero steady-state error 2, Small overshoot at output (usually less than 10% for reference input step response). 3. Appropriate settling times for both fine and reference input step responses (5 ms at most in our case). Line Volage 7 CONVERTER ouput Switching Frequency Reference. ——] CONTROLLER Figure 11.20. the dlogrom tor contol design CONTROL OF ADCTODC RESONANT CONVERTER 411 4. A closed-loop bandwidth of at least 360 Hz. in order 1o successfully suppress line ripple. © Load Regulation (Robust Stability). Ow the other hand, the load con- dition could also vary over a wide range. Since the load R,, enters the ‘dynamics of the model, load variations will appear as model uncertainty and could possibly lead to stability problems. Normally the load changes from 10% at low load to 90% at full load condition, Other model uncer tainties, such as unmodeled high-frequency dynamics and uncertainties in the resonant inductor £. and capacitor C, will also be considered, Robust Performance. Since the converter operates over a wide range of load conditions, the performance requirements must be satisfied at all operating points. This is equivalent to requiring satisfactory response under both line and load variations. 11.4.5 Analysis of the Plant Control Characteristics For a PRC converter operating under steady- state conditions, the input-output relationships can be represented by the control characteristics curves, relating the output voltage to the load and switching frequency. Given any two variables among the normalized output Vo, Switching frequency ratio Fy,, and output load Q,. the third variable can be determined from the curves, Thus these curves allow one to casily visualize the effects of the switching frequency and load on the converter ‘output. From a control point of view, the control characteristics curves allow us to make an initial estimate of the load change that can be tolerated and to see some of the difficulties inherent in the load regulation problem. Figure 11.21 shows the control characteristics curves for various output loads Qp, obtained analytically from the steady-state analysis. To maintain the output voltage constant in the presence of perturbations, the controller should adjust the switching frequency to keep the converter operating along the dashed line indicated in Figure 11.21. As the converter is perturbed away from the nominal operating point (marked with an asterisk in the figure) the plant dynamics may vary significantly, resulting ina difficult control problem, Remark Note that from Figure 11.21 it follows that at lighter loads (higher Ry, larger Qp, and lower J,), a small frequency change will result in larger output changes. Thus we should expect that the control problem will become more difficult at larger Q, values, In the next section we will show. through a frequency-domain analysis, that this is exaclly the case. Frequency Responses From the discrete time state-space model, we can easily get the z-transfer functions from the normalized switching, frequency412 APPUCATIONEXAMPLES 19-— O5 055 06 065 07 075 08 085 09 095 1 Fre Figure 11.21. The conventional PRC contol characterises curves Fax and the normalized line input Vrg to the normalized output Vn Vno(2) Vnol2) Fus(2) Vrg(2) | ~ [o02) Ge(2) Ct-ap"B (11.73) Following a common approach, we will carry out the analysis of the plant and the synthesis of a digital controller using a w-plane approach. To this ‘effect, the bilinear transformation _Lash/4 ran (7) is used to obtain the transfer functions in the frequency domain s. These transfer functions, still denoted as G(s) and G,(), are given by (5+ 798041)(5 ~ 792481)(s — 800003) 2.652 x 10-2 + 7950415 = 792431)(s — 800003) ot (5 +29167)(s + 83363 + 202487i) Coe) (s = 800003)(s + 484950 + 2902534) Gals) = 136710 poreny(s R365 20D8T|) LT) as) ‘The above transfer functions correspond to the nominal load R, = 208.330. AS stated before, since the load enters the dynamics of the converter, load variations result in different transfer functions, Figures 11.22 and 11.23 show ‘CONTROL OF A OC-TO-OC RESONANT CONVERTER 10 104 NOE ak at aa at Al yor 1020 10* tPF C10? wirads) Figure 11.22, Frequency responses G(s) a ferent load conditions. ior igs i iv wieadls) Figure 11.28. Frequency responses G(s) at diferent load conditions. as.414 APPLICATIONEXAMPLES the frequency responses of G(s) and G,(s) corresponding to several different load conditions, respectively. These figures show that as the load becomes lighter (larger R,), the over shoot inereases, leading to a more difficult control problem. This conclusion is consistent with the conclusion drawn in the last section from the study of the control characteristics. On the other hand, the control characteristics require that Q» be greater than Vqp in order to get the prescribed output voltage. Since in our design example the value of Vn» is chosen to be 2.5, it follows that R, should be greater than 151. Therefore, in the sequel, we will assume that R, varies within the range 151 to 1200 0. 11.4.6 Control Design ‘As mentioned before, our goal is to design a controller that satisfies the performance specifications listed in Section 11.4.4 for all load conditions in the range 151.0 < R, < 12000, assuming that the values of the components of the resonant tank are known within a 10% tolerance. In the sequel we solve this problem by recasting it into a robust performance synthesis form and using synthesis. To this effect we need first to describe the family of plants corresponding to different values of the load as a nominal plant subject to uncertainty. Plant Description and Uncertainty Weight Selection In this exam- ple we choose to model the uncertainty caused by load variations by using a single, norm-bounded, multiplicative uncertainty that covers all possible plants. Let G¥(s) and G(s) denote the transfer functions from the control input and line input to the output at operating points other than the nominal point (Ry 208.332), respectively: In the sequel we represent these transfer functions as GMs) = Gls) [t+ 8/(5)Wi)] au) GPe(s) = Gas) [1 + Bg(5)65)] (11.78) where G(s) and G;(s) are the nominal transfer functions given in equations (11.75) and (11.76), respectively, W;(s) and We(s) are fixed weighting fune- tions containing all the information available about the frequency distribution of the uncertainty, and 4y(s) and A,(s) are stable transfer functions represent- ing model uncertainty. Furthermore, without loss of generality (by absorbing any scaling factor into W,(s) and W(s) if necessary), it can be assumed that [Sr(s)lx <1 and liMe(s)[le <1. Thus W;(s) and W,(s) are such that their respective magnitude Bode plots cover the Bode plots of all possible plants. Some sample uncertainties corresponding to different values of the load R, are shown in Figures 11.24 and 11.25. We can see that in both figures the ‘multiplicative uncertainties have a peak around the resonant frequency. This Peak becomes larger and steeper as the load resistance R, increases. Magnitude Figure 11.24, Multiplicative uncertainties (contol output and weight Magnitude 108 ‘CONTROL OF A DC-IO-DC RESONANT CONVERTER 10 104 107! 10" Cn a a nT) wiradis) 10 10 108 10F 10 wiradis) Figure 11.25, Mutipicatve uncertainties (ne fo output) and weight as416 APPUCATION EXAMPLES Based on these plots, the following multiplicative uncertainty weights were chosen for control design: 4x 10-45 +065 Wy) = AXIO 06 (1179) 104s Wo) = ee (1180) The magnitude frequency responses of W,(s) and We(s) are also shown in Figures 11.24 and 11.25, respectively. These figures clearly show that attempt- ing to cover the sharp peak around the resonant frequency will result in large gaps between the weight and the uncertainty at high frequencies, introducing conservatism at that frequency range. On the other hand, a tighter fit at high frequencies using higher-order functions will result in high-order controller. The weights (11.79) and (11.80) used in our design provide a good trade-off between robustness and controller complexity. ‘We turn our attention now to the effects of changes in the values of L and CC, the resonant tank components. Since these changes affect the location of the resonant peak, they could conceivably destabilize any controller design based on its nominal location. Figure 11.26 shows the changes in the trans: fer functions due to +10% changes in the values of L and/or C. It is worth noticing that our choice of weighting functions W; and W, will also cover this family of plants, even at the extreme load conditions Ry = 12000 and 151.0. Thus a robust controller designed using these weighting functions will be able to accommodate both changes in the load condition and uncertainty in L and C. Figure 11.27 shows a block diagram of the converter, taking into account the uncertainty. Here A; and Ay are scalar blocks, representing the model uncertainty perturbations from the control and line inputs, respectively, and W; and We are the corresponding uncertainty weights. Performance Weight Selection As discussed in Chapter 7, in order to guarantee robust performance we need to add to the structure shown in Figure 11.27 an additional (fictitious) uncertainty block 4p, along with the corresponding performance weights W, and W,, associated with the track ing/regulation error and the control effort, respectively, resulting in the block diagram shown in Figure 11.28. The selection of the performance weights W, and W, entails a trade-off among good regulation versus peak control action. The weight on the control error W,(s) is usually selected to be very large at low frequencies in order to get good tracking and regulation, Additionally, the order of the weights should be kept low in order to keep the overall controller complexity low. A g00d compromise between performance and compleaity is given by weighting CONTROL OF ADC-TODC RESONANTCONVERTER 417 Control to Output fe eee ee 107 10? 109 10% 10° 108 107 10% wads) Line 1 Output 10? us —— 10! ww fo 5 F151, L-9THH,C=t00F 20? 10 xo! 10) = = “107 10 70 “6 79 10! wired) Figure 11.26. Uncertoinies due to 410% changes of L ond/or C of extreme loos conditions Re = 120010 and 155,418 APPUCATIONEXAMPLES Figure 11.27. the block diagram of the converter including the uncertainty due to 00d and component variations. functions of the form W.(s) (181) BSA where A is the desired steady-state error (A will be zero if zero steady state is required); T, approximately determines the bandwidth (w, = 1/T) and Figure 11.28. The block diagtam for synthe ‘CONTROL OF A DC-IO‘OC RESONANTCONVERER 419 hence the rising time and settling time; where the ratio 73/7; is associated with performance requirements against high-frequency noise ((188]). Note that there is no exact relationship between the parameters 7; and T> and time-domain performance specifications given in terms of rise-time, setling: time, and overshoot. The design of multiobjective robust controllers subject to both time- and frequency-domain specifications is, to a large extent, an ‘open problem, although some progress has been made recently [301] In this particular design example we selected the following weights: w= ae ns 104s we) = cay The weight on the control input W,(s) was chosen close to a differentiator to penalize fast changes and large overshoot in the control input, The weight W(s) guarantees zero steady-state error. The frequency responses of W.(s) and W,(s) are shown in Figure 11.29. The closed-loop bandwidth correspond: ing to this choice of weights is approximately 1/0.0006 = 1700rad/s. Log Magnitude 10° cr 104 Frequency (radiansisec) Figure 11.28. Performance weights We(s) and Wu(s) 10420 APPUICATIONEXAMPLES 11.4.7, Controller Synthesis Combining the uncertainty description with the performance weights yields aan uncertainty structure 4 consisting of two scalar blocks (corresponding to the robust stability requirements) and a 2x2 block (corresponding to the ro- bust performance requirements). Since the A structure has only three blocks, the upper bound of 1, inf pep IDM D> |, coincides with its exact value. The robust controller was synthesized using the . Analysis and Synthesis Toolbox (20), applied to the block diagram shown in Figure 11.28. After four D-K itera: tions with third-order D-scalings, we obtained a 18th-order controller yielding veep = 0.9721. Finally, Hankel norm model reduction yielded a sixih-order controller with virtually no performance degradation (size = 0.9753 <1). ‘The state-space description of this reduced-order controller is given by K=Oyst~A)"'By + De (11.84) where 025 1008 -1as¢ nats nis 1.296, 1708 1320045 998e+4 3190045 1460044 —2213045 Aga | HIS 9980044 —16%e+4 S2MMeS 1200004 1.983665 w=] chats 34mess “320805 “2erdews 172905 4053045, FUN6L 14600 +4 1200644 hm8645 -2664e83 LOIBer S 1296 2213e05 “LBBerS “408305 “101965 “RORSe rs BY ~(-2338 7.983 5.345 —6.610 0.543 6.060) 10-2 G.=(-0.935 3.193 -2.138 2.644 0.217 2.424) x 10 De=0 In order to benchmark the performance of the robust controller, we also designed a phase-lag controller using classical design tools, based on the plant frequency responses at the various operating points shown in Figure 11.22. To improve performance, this controller was further tuned by trial and error. ‘The transfer function of the final controller is given by 0.025 + 200, 3402 The frequency responses of both the 4. and the phase-lag controllers are shown in Figure 11.30, Both controllers have similar responses at low fre- quencies, while at high frequencies the gain of the u controller decays faster im order to accommodate the model uncertainties at high frequencies. Figure 11.31 shows the closed-loop frequency responses for the nominal plant and for the lightest load considered in the design. Note that in both cases the y controller provides lower gain and better roll-off at high frequencies. Moreover, while the response corresponding to the phase-lag controller is acceptable for the nominal plant, it exhibits a large peak at the resonant frequency at light loads. As we show next, this peak results in significant performance deterioration at these loads. K, = (115) CONTROL OF A DC-TODC RESONANTCONVERIER 421 Log Magnitude 10 1 oe 10 rg 108 Frequency radianssec) 9 7 = & so é 09 50nd 10 10 10 10 6 10 Froquency (rasianaee) Figure 11.20. Frequency responses o 1 (solid) and phase-og (dashed) controls. 11.4.8 Simulation Results Linear Simulations The closed-loop system corresponding to the j. con- troller was simulated at the nominal operating point R, = 2080) and at two ‘extreme cases R, = 151 and 12000, using the corresponding linear model of the plant. The time responses 10 20% step change in line voltage Vag and reference input r are shown in Figure 11.32, For the nominal case R, = 2080, the settling time is about 2.5 ms for both line voltage change and reference input change. The output responses are satisfactory since the settling time is smaller than the required 5 ms, with ro overshoot. The control action in these responses is also adequate, without overshoot or abrupt change. This is due to the choice of the weight Wi, penalizing fast changes and overshoots in the control action ‘When the operating point moves to R, = 151.0, the setting times for both step changes are about 4 ms, This increase is due mainly to the significant decrease in plant static gain (see Figure 11.22). The u controllers undertuned at this operating point in order to achieve robust performance. When the operating, point moves toward lighter loads, the responses are almost the same as for the nominal plant, except that for the case R, = 12000 (note422 APPLICATIONEXAMPLES 10° — _ : ~ ~~ Phase-iag contalier Gain 104 p-controlle oe 108 10 10° Frequency (rad) ° Phase-ag controller /\ 1ot w-conteoller 108 10° 10" 10 10° Frequency (ad) Figure 11.31. Closedoop trequency rospontes for x (sold) ond phaselag (dashed) controllers: (4) Ry = 20881 and (0) Re ~ 1200, CONTROL OF ADCO-0C RESONANTCONVERIER 423 409) aa x10 085 oe 0G Time (582) 10? Time (see) 492 @) (b) Figure 11.82. Linear simulation resus with 4. conkolier at diferent operating points Fo = 208 1 (oid), 181 12 (dashed), and 1200 1) (does): (a) relerence input stop change (20%) and (bine vollage step change (20%) that this load is the lightest Joad considered in our design). some chattering in both the output and control input starts to show up at the beginning of the responses. The occurrence of the chattering is linked to the large peak in the plant frequency response at lighter loads, barely covered by the uncertainty weights W; and Wy From the simulation results it follows that the controller achieves robust, performance, since all performance specifications are satisfied at all operat: ing points of interest. However, significant variation of performance is also observed. This isa direct result of the large variation in the plant dynamics. and any fixed linear controller can do very litle in this respect, To reduce ‘this variation will require using a nonlinear, gain scheduling controller. The same simulation was performed for the closed-loop system corre sponding to the phase-lag controller. The time responses to 20% step change im line voltage Vy and feference input r at three different operating points — R, = 1510, 2080, and 1200—are shown in Figure 11.33. They are424 APPLICATIONEXAMPLES om 07 aso ano _ 5 5 aso | asl 0 0) aay 0 1 2 3 4 rece) gs : —— om a 2 34 2 3 Time (660) y Tere 600) yo! (a) (b) Figure 11.38. Linea simulation resus wih phaselag controler at diferent operating pins Re = 208 0 (soli), 181.0 (dashed), ond 1200 £1 (coted): (a) reference Input step. ‘change (20%) ond (b) ne volage step change (20%), similar to the responses with the 1 controller except that the performance is far worse for Ry = 12000. This is due to the phase-lag controllers inability to provide enough attenuation to counteract the increment in the magnitude of the resonant peak of the plant at heavy loads, as shown in Figure 11.31. Furthermore, as shown in Figure 11.34, at R, = 240010 the phase-lag con- troller fails to stabilize the system, while the controller can still produce acceptable performance. Nonlinear Simulations and Validation While simulations using the linearized model of the plant corresponding to different load conditions can usually provide an approximate evaluation of load regulation performance, this is usually insufficient to assess the performance of a highly nonlinear system such as the converter. Thus to further validate the controller, a non- linear simulation of the PRC circuit was performed using P-Spice. Figure 11.35, shows the responses due to reference input and line voltage step changes. Note that these results are similar to those obtained using a linear simulation CONTROL OF ADCIO.DC RESONANICONVERIER 425, 1 har om; Oem oa 0 foo 3 od aa rr os Cectmenianate aes ee nes) yg? me ees) yg? Input 2 Opt Int 22 4 eos 0s é q $4 $004 4 i a Eo q ons| oi se 8% Os se Tine 08 49? Time ane ag? ® 4 4 co a 2] = a ce lteeas Te see9} 40? “we sees) ap? 10" _Iput2 vet ‘to? _ hpi 2 Ou? «poe meen) , E os| il i a} 2 2 os ay Tre sos) 9? Figure 11.34, Stop responses of he closecioop systom at = 2400.2: (6) 4 contoler (slab) and (©) phoseag controller (unsiabi),426 APPLICATIONEXAMPLES 220) 200] S280 260) 2405 SO nce ses) Time (560) 97 095; og toss 03] 88 Time ($20) 49? Time (see) 30? (a) (by Figure 11.38. Nonlinear simulation resus wih. contoler a ferent operating points p= 208 1 (sold), 181 11 (dashed) and 1200 11 (coted) s: (a) reference input sep ‘change (20%) and (bine vollage sep change (20%). as shown in Figure 11.32. In the responses to line voltage step change and reference step change, settling times are slightly larger than those in the li car simulation, The chattering observed in the output voltage is substantial, due to the periodic switching behavior of the converter (periodic charge and discharge of capacitor). i ro i ‘6 fu bi BIBLIOGRAPHY ‘Abdallah C. "On the robust control of uncertain nonlinear systems,” Proceed ings American Control Conference, San Diego, 1990. Abdallah C., Dawson D., Dorato P, Jamshidi M, “Survey of robust control for rigid rob0W," IEEE Control Sysiems Magazine, Vol 11, NO. 2, 1991 ‘Abedor J., Nagpal K., Poola K.,"A linear matrix inequality approach to peak to peak gain minimization,” International Journal of Robust and Nonlinear Con: ‘rol, Yo. 6, pp. 99-927, 1996. ‘Ackermann J, "Parameter space design of robust control systems,” IEEE Trans actions on Auiomatic Control, Vol 25, pp. 1058-1072, 1980. ‘Ackermann J., Robust Control: Systems with Uncertain Physical Parameters Springer-Verlag, New York, 1993, ‘Adamjan VM,, Arov DZ., Krein MG., "Analytic properties of Schmidt pairs for a Hankel operator and the generalized Schur-Takagi problem,” Math USSR Sbornik, Vol. 15, pp. 31-73, 1971 Adamjan VM. Arov D.Z., Krein M.G., “Infinite Hankel block matrices and related extension problems,” AMS Translations, Vol. 111, No. 133, 1978, Akcay H., Gu G., Khargonekar P, “Identification in ., with nonuniformly spaced frequency response measurements," Proceedings American Control Con- ference, Chicago, 1992 ‘Akcay H., Gu G., Khargonekar P, “A class of algorithms for identification in Ex: continuous-time case,” JEEE Transactions on Automatic Control, Vol. 38, No. 2, 1993, ‘Anderson B.D O., VongpanitlerdS., Network Analysis and Synthesis: A Modern Systems Theory Approach, Prentice-Hall, Englewood Clif, NI, 1973. ‘Anderson B., Kraus F, Mansour M., Dasgupta S., “Easily testable sufficient conditions for the robust stability of systems with muliaffine parameter de- pendence,” in Robustness of Dynamic Systems with Parameter Uncertainties, M Mansour, S. Baler, W. Trudl, Eds. Birkhauser, Berlin, 192 427428 BIBLIOGRAPHY 12] Anderson B.D.O., Moore. JB. Optimal Conrot: Linear Quadratic Methods, Prentice-Hall, Englewood Cis, NI, 1990. 113) Astrom K., Eykhoft P, “System identication: survey.” Automatica, Vl 7, pp. 123-162, 197. [14) Astrom K. Witenmark B.. Computer Controled Systems: Theory and Design, Prentice-Hall, Englewood Cli, NJ, 1984 115) Athans M., Falb P, primal Control: An Inroduction 1 the Theory and Is ‘Applications, MeGraw Hil, New York, 1966, [16] Athans M, KapasourisP, Kappos E, Spang H.A. IM, “Linear quadratic gaus- Sian with Toop transfer recovery methodology for the F100 engine.” AIAA Journal of Guidance, Control and Dynamic, Vo. 9, No. 1, 1986 17] Balakrishnan VS., Boyd S., Balem S, “Branch and bound algorithm for com- Puting the minimum stability degree of parameter dependent linea systems” ‘Technical Report. Information Systems Laboratory Stanford University, 1991 (18) Balarishaan VS.. Boyd .P, “On computing the worst case peak gain of linear systems," Systems and Controls Leters, Vo. 19,192 119) Balas G., Doyle J.C, "Identiication of flexible structures for robust contol,” IEEE Comiro Systems Magazine, Vo. 10, No.4 199. [20) Balas G., Doyle .C. Glover K, Packard A, Smith R, p-Analysis and Syuess Toolbox. The MathWorks Ine, Natick, MA, 1991 [21] Ball.A., Gotberg I, Rodman L, Interpolation of Rational Matrix Functions, Operator Theory: Advances and Applications, Vol. 4, Birkhauser, Beri, 1990 [22] Ball 1A, Helton J.W, Walker MLL, "7, control for nonlinear systems wth ‘output feeaback.” IEEE Transactions on Automatic Control, Vo. 38, No. 4 1998, [23] Bancroft A.R, “Heavy water G-S process research and development achiewe- ments,” Atomic Energy Canada Limited (AECL) Report No. 6215, 197. 24} Banks HLT, Burns).A., “Hereditary contol problems: numerical methods based ‘on averaging approximation,” STAM Journal of Conirol and Optimization, No. 16, pp. 169-208, 1978 [25] Banks FLT, Kappel F, “Spline approximations fo functional differential equa tions,” Journal of Differential Equations, Vo. 34, pp. 496-822, 1979. [26] Barabsnov AE. Granichin OLN. “Optimal controler fo afinear plant with bounded noise,” Aulomation and Remote Control, Vol. $, pp. 39-46, 1984 [27] Barmish BR., “Invariance of the strict Hurwite property for polynomials with perturbed cosficients," IEEE Transactions on Automatic Control, Vo. 29,1984 {First presented atthe 1983 Conference on Decision and Control, Tex.) Barmish B. Khargonckar PP. “Robust stability of feedback control systems sth uncertain parameters and unmodelled dynamics.” Proceedings American Control Conference, Georgia, 1988 (29) Barmish B R, “New tools fr robustness analysis," Plenary Session, Proceedings Conference on Decision and Control, Austin, Texas, 1988. Barmish B, Ackermann JE, Hu HZ. “The tee structured decomposition: a new approach to robust stability analysis,” Proceedings Confrence on Informa ton Setences and Systems, Johns Hopkins University, Baltimore, 1989 es} [p9) BRUOGRAPKY 429 [31] Barmish B.R., "A generalization of Kharitonov’ four polynomial concept for ‘robust stability problems with linearly dependent coefficient perturbations,” IEEE Transactions on Automatic Control, Vol. 34, No, 2, 1989, [22] Barmish B., Khargonekar PP, Shi Z.C., Tempo R., "Robustness margin need not be a continuous function of the problem data,” System and Control Letters, Vol. 15, 196. [53] Barmish B.R,, ShiZ., "Robust stability of class of polynomials with coefficient depending mulilinearly on perturbations,” JEEE Transactions on Automatic Conirol, Vol. 35, No.9, 1990, [34] Barmish B., Kang H., “A survey of extreme point results for robustness of control systems,” Automatica, Vol. 29, No, 1, 1993, [35] Barmish BR. New Tools for Robusiness Analysis, Memillan, London, 1994, [36] Bartiet A.C, HollotC, Lin H., “Root locations of an entire polytope of poly ‘nomials it suffices to check the edges.” Mathematics of Control, Signals and Systems, Vol. 1, No.1, 1988. [57] Batarseh L., Steady Site Analysis and Design of High Order Resonant Convert rs, PhD. Thesis, University of Illinois, Chicago, June 1990 [38] Beck C., Doyle JC. “Mixed s2 upper bound computation,” Proceedings Con: {ference on Decision and Control, Tucson, 1992 [89] Bertsekas DR, Rhodes LB., “On the minmax reachability of target set and target tubes,” Automatica, Vo. 7, pp. 233-247, 1971 (40) Bersekas DP, “Infinie-cime reachability of state-space reglons by using feed: back control," JEBE Transactions on Automatic Control, Vol. AC-17, No. 4, PP. 604-613, 1972 [a1] Blanchini F, Sanaier M., “Rational £” suboptimal compensators fr continuous- time systems.” JEEE Transactions on Automatic Control, Vol. 39, No. 7, pp. 1487-1492, 1994, [42] Blanchini, F, “Non-quadratic Lyapunov funetion for robust contra” Automat ‘ca, Vol 31, No. 3, 451-461, 1994 (43) Blanchini F, Sznaier M. “Persistent disturbance rejection via static state feed back," IEEE Transactions on Automatic Control, Vol. 4, No. 6, pp. 1127-1131, une 1995, [44] Becker N., Grimm WM. “On £; and £,, stability approaches for the robust control of robot manipulators,” IEEE Transactions on Auomatic Control, Nol 33, No.1, 1988. (45) BetczynskiB., Kale 1, Cain G.D., “Approximation of FIR by IR digital filters: an algorithm based on balanced model reduction,” IEEE Transactions on Signal Processing, Vol. 40, No. 3, 992. {46] Bernstein, D., Wie, B., Organizers, Invited Session on “A Benchmark Prob- lem for Robust Ms / 7. Control Design,” Session WP-1S, Proceedings of the American Control Conference, San Diego, pp. 961-973, 1950. [47] Bhattacharyya SP, Chapellat H., Keel LH, Robust Convol: The Parametric Approach, Prentice-Hall, Englewood Cliff, NJ, 1995 (48) BialasS.,“A necessary and sufficient condition for stability of interval matrices, International Journal of Control, Vol. 37, pp. 117-122, 1983,430 19 150) 154] [ssi 6 1371 Iss 9 (60) (on (62) (63) (4), (6s) (6s) 67) BisLIOGRAPHY DBialas S. Garloff J, “Stability of polynomials under coefficient perturbations,” IEEE Transactions on Automatic Control, Vol. 30, 1985. Biernacki RM., Hwang H., BhattacharyyaS-P, "Robust stability with structured real parameter perturbations,” JEEE Transactions on Automatic Control, Vol. 32, 1987 Blondel V. and Gevers, M.. "Simultaneous stabibzation of three linear systems is rationally undecidable,” Mathematics of Control, Signals and Systems, Vol. 6, pp. 135-145, 1993, Bode H.W, Network Analysis and Feedback Amplifier Design, Van Nostrand, Princeton, NJ, 1945, Bodson M., Chiasson J.N., Novotnak RT, Rekowski R.B., “High-performance nonlinear feedback control ofa permanent magnet stepper motor,” IEEE Trans. actions on Control Systems Technology, Vol. 1, No. 1, 1993, Bose NK, “A system theoretic approach to stability of sets of polynomials,” Contemporary Mathematice, Vol. $7, pp. 25-38, 1985, Boyd S., Desoer C.A., “Subharmonic functions and performance bounds in linear time invariant feedback systems,” IMA Jounal of Mathematics, Control and Information, Vol 2, pp. 153-170, 1985. Boyd SP, Doyle J.C, "Comparison of peak and RMS gains for discrete-time systems,” Systems and Controls Lentrs, Vol 21, 1983. Boyd S. Balakrishnan V., Kabamba P, “A bisection method for computing the ‘Yo norm of a transfer matrix and related problems.” Mathematics of Control, Signals and Systems, No, 2, 1989. Boyd S,,Barrat CH., Linear Controller Design Limits of Performance, Prentice: Hall Information and System Sciences Series, Prentice-Hall, Englewood Clif, NJ, 1991, Boyd S., EI Ghaoui L., Feron E., Balakrishnan V, Linear Matrix Inequalities in Spstems and Control Theory, SIAM Studies in Applied Mathematics, Vol. SIAM, Philadelphia, 1994 Braatz R.D., Young PM., Doyle .C., Morari M. "Computational complexity of ju calevlation,” IEEE Transactions on Automatic Control, Vol. 38, No, 5, 1998. Brockett RW, Finite Dimensional Linear Systems, Wiley, New York, 1970. Bruck J, Blaum M., “Neural networks, errr-correctng codes and polynomials over the binary r-cube,” IEEE Transactions on Information Theory,No.9, 1989. Bryson AE,, Ho ¥.C,, Applied Optimal Control, Hemisphere Publishing, New York, 1975. Bu J., Sznaier M,, Holmes M., "A linear matrix inequality approach or synthe- sizing low order ! controllers,” Proceedings 38th IEEE Conference on Decision and Cantrl, Kobe, Japan, pp- 1875-1880, 1996. Bu J., Senaier M. Batarseh 1, Wang Z.Q., “Robust controller design for @ parallel resonant converter using, w-synthesis.” [EEE Transactions on Power Electronics, Vol. 12, No. 5, pp- 837-853, September 1997 Bu J., MultObjectve Robust Control: Design of Mixed €°/P,. Controllers, PhD. Dissertation, The Pennsylvania State University, 1997. Burns J, Peichl G., “On robustness of controllability for nite dimensional approximations of distributed parameter systems,” Technical Report. Dept. of (63) (69 (70 m 7) 73) (74) i (76) m (78) 9) (80) isi (82) 83] aauocRarHy 431 Mathematics. Virginia Polytechnic Institute and State University. Blacksburg, 1985, Chapeliat H., Bhattacharyya SP. °A generalization of Khatitonov’s theorem: robust stability of interval plants." IEEE Transactions on Automatic Control. Vol 34, No. 3 1989 CChapeliat H.. Bhattacharyya S.P, “An alternative proof of Kharitonov’s theo- rem." IEEE Transactions on Automatic Conwrol, Vol. 4. No. 4.1989. CChapellat H, Dahleh M. Bhattacharyya SP, “Robust stability under structured ‘and unstructured perturbations." IEEE Transactions on Automatic Con. Vol 35, No. 10, 1990 Chapellat H., Dahleh M, Bhattacharyya SP, “Robust stability manifolds for ‘multilinear interval systems.” IEEE Transactions on Automatic Conirol. Vol. 38 No. 2, 1993, Chen CT. Linear System Theory and Design, Holt, Rinehart & Winston. New York, 1984, Chen J., Fan M. Nett C, “The structured singular value and stability of un- certain polynomials: @ missing link,” ASME Annual Wier Meeiig, Atlanta, 1991, Chen J. Nett C. Fan M, “Worst-case system identification in W.: validation of priori information, essentially optimal algorithms. and ertor bounds.” Proceed: ings American Control Conference, Chicago. 1992; also in IEEE Transactions ‘on Automatic Control. Vol. 40, No. 7. 1995 Chen J. Nett C, Fan M.. “Optimal non-parametric system identification from arbitrary corrupt finite time series: a control-oriented approach.” Proceedings “American Conirol Conference, Chicago, 1992. Chen J, Farrell, Nett Zhou K.,"% identification of multivariable systems by tangential interpolation methods.” Proceedings Conference on Decision and Control, Florida, 1994 Chen J., Nett C., “The Carathéodory-Fejér problem and the 1. /t! identi cation: 9 time domain approach,” JEEE Transactions on Automatic Control Vol. 40, No. 4, 1995 Chiang, R, and Safonov, MG., “Ha. Robust control synthesis for an undamped. hon-colocatedspring-mass system.” Praceedings American Control Conference. PP. 966-967, 1990. Curtain R., Glover K., “Balanced realisations for infinite dimensional systems.” Workshop on Operator Theory and lis Applications. Amsterdam. 1985, Dahleh MA., Pearson J../-Optimal controllers for discrete time systems, Proceedings American Control Conference. Seattle. pp. 1964-1968. 1986, Dahlen M.A.. Pearson J. "f"-Optimal controllers for MIMO discrete time sys tems.” IEEE Transactions on Automatic Control, Vol 32. No. 4.1987 Dahlen M.A,, Boyd Pearson J. °'-Optimal compensators fr continuous time systems.” [EEE Transactions on Automatic Conirol, Nol. 32, No. 10. pp. 88— 895, 1987, Dahlch MA.. Pearson J.. ‘Optimal rejection of persistent disturbances. robust stability and mixed sensitivity minimization.” FEEE Transactions om Autonnatic Control. Nol 33, No. 8, 1988.432 BIBLIOGRAPHY WN) Dableh M.A.. Ohta Y.,"A necessary and suflicient condition for robust BIBO. stability” Systems and Control Lens. Vel, 11. pp. 271-275, 198 {Ks} Dalen M.A... TheodosopoulosT. Tsitsiklis J, ~The sample complexity of worst ease [dentcation of FLR. linear systems,” Systems and Control Leter. Vo. 20. Dp. 157-166, 1993 'Ko) Dahleh M.A.. Khammash M.. “Controller design for plants with structured un certainty.” Automatica, Vol 29. No. 1. pp. 37-56. 193. {87| Dahleh M.A.. Diaz-Bobillo LJ... Control of Uncertain Svstems: A Linear Pro gramming Approach, Prentice-Hall, Englewood Cis. NI 1995 (88) Fu M.. Dasgupta S.. Blondel V.. “Robust stability under a class of nonlinear parametric perturbations.” IEEE Transactions on Auwomatic Consrol, Vol. 4. No, 2.1985, [89] D’Auetis CE. ouroduccién a los Sistemas Nolineales de Control y sus Aplica (a&b) € F. Operation & is associative, that is, (@& b)&e= a&(b&C) = akb&e for a,bceF Operation & is commutative, that i, a & b = b &a for a,b F. Set F contains the neutral element ng with respect 10 &, that is, there exists ng © F such that (a &ng) =a for all ac F 44s446 MATHEMATICAL BACKGROUND 5. Set F contains the inverse element al, with respect to &, that is, for all ae F, there exists aly € F such that (a8 al.) = ne Set F is closed with respect to Operation + is associative Set F contains the neutral element n, with respect to * Set F contains the inverse element al with respect t0 Operation + is distributive with respect to &, that is, (a& b) we = (a & (x0) for a,b,c F. Take, for example, set R equipped with operations (+, x) as (8, +) respec: tively, then mg =0.,= 1, ah = ~a, and al = a7! (a0), A.1.2. Linear Vector Space Definition A2._A set V is a linear vector space over the field (F,+,*) if and only if the following properties are satisfied: (Im the sequel the elements of F and V will be called scalars and vectors, respectively) Set V is closed with respect to + Operation + is associative in V. Operation + is commutative nV. Set V contains the neutral element with respect 10 + Set V contains the inverse element with respect 10 + V is closed with respect to operation x between scalars and vectors. Operation x among scalars and vectors is associative in the scalars, that is, (ax b) xv =ax (b xv) =ax bx for abe F and vey. 8. Disiributive I: (a+b) x v = (a xv) +(b xv) for a,b € F and ve ¥. 9. Distributive 2: (u+v} x a= (ux a) +(v xa) fora € F and uv € V. 10. Field F contains the neutral element of operation x between vectors and scalars, that in x0 =v form. € F and v€ V. In the sequel, equivalently. inear vector space, linear space, or vector space will be used Definition A.3_ The set of vectors {04,...,9} in the linear space V, defined over the field F, is linearly independent if and only if the linear combination ev bo ay = a = 0, with a, € Fi =1,...,m. Other wise they are called linearly dependent, Definition A.4 A set of linearly independent vectors {Uy,..-,tn} of @ linear space V constinue a base of V if and only if any v € V can be represented ‘as a linear combination of these vectors. The dimension of V is the minimum number of vectors that constitute a base, in this case n. ALGEBRAIC STRUCTURES 447 Linear spaces may have finite or infinite dimension, as will be shown next bby means of several examples. Example Al The marrices in ""' or C°!, constirue finite-dimensional lin ear spaces. The representations of position in physical space with respect 10 a certain reference frame also constitute a finite-dimensional linear space. The sequences of real numibers Sz. oF the set of continuous functions defined in a real interval Cig, form infinite-dimensional linear spaces. These last two cases may represent the pulse sequences in digital systems or certain classes of signals in analog circuits, respectively A.1.3. Metric, Norm, and Inner Products Definition A.S A metric space (V,d\-,.)) is defined in terms of a linear vec~ tor space V and a real function d{-,»):V x V = R,, satisfying the following conditions: 1 diy) 20, Vay ev. 2dyy)=0 =e xray: 3. dry) = dla), Yayev. 4 dena) x=0 3. lex||=lal-ix], Vx Via F. 4, xe yll ) is defined in terms of 4 linear vector space V and a function <_,- >: x V+ F that satisfies the following conditions ¥ x,y,z €V and a ¢ F: kxtye>acne>+ FS. 20, Vr 60 i 2, 3. a . 4 S =0 oox=0. Here @ represents the complex conjugate of element a, Example A.3- The three normed spaces from the last example can be made inner product spaces with the following definitions: 1, Take vectors in C* and define t [Foire (ayy ‘An inner product space generates a normed vector space when using its inner product (o define a norm, as follows: |x\|= =X. >. In addition, a normed vector space generates a metric space when using its norm to define the distance function, as follows: d(x, y) = lx ~ yl Definition A.B Two vectors x,y in (V,< +>) are said 10 be orthogonal if and only if =0. If, in addition, =1 and < y,y >= 1, they are called orthonormal ‘A.2 FUNCTION SPACES 4.2.1 Introduction Next, some important properties of infinite-dimensional vector spaces are briefly reviewed. Definition A.9 A sequence {x,} in a metric space (V, 4) is said to be a Cauchy sequence if and only if Ve > 0 AN such that nm > N > dn, %m) < € Definition AO A sequence {9} in a metric space (V,d) converges to x, €V if and only if ¥€ > 0 AN such that n > N => d(tq%.) < € IL is easy to prove that any convergent sequence is a Cauchy sequence. (On the other hand, the converse, is not necessarily true. Example Ad Consider among the rational numbers Q the sequence {3 3.14 34415. This isa Cauchy sequence, although it does not con: verge because = ¢ Definition A.11 A metric space (V,d) is said 10 be complete if and only if all Cauchy sequences are convergent in V. A.2.2 Banach and Hilbert spaces jon A.12 A complete normed vector space (V, | -|\) is ealled a Banach ition A.I3 A complete inner product space (V.<-,->) is called a Hilbert space450. MATHEMATICAL BACKGROUND x(t) Figure A.1. Nonconvergent sequence in Cra Example AS The infinite-dimensional vector space of continuous functions defined in the unit interval C3, #8 not complete when using the metries (x,y) = |x = ylh oF daxyy ilo, where these norms have been de: {fined in (AS). ‘Note in Figure A.J thatthe sequence X(t) of continuous functions converges ww a discontinuous one. Therefore the Space of continuous functions is not complete, under these metrics 2.3. Operator and Signal Spaces Many ofthe useful function spaces, which willbe described inthis subsection, include other than continuous functions. In these cases integration should be understood inthe Lebesgue integral sense (Riemann integral is valid ony for continuous functions). In addition, the concepts of minimum and maximum of a function (min, max) are replaced by their generalizations: infimum and supremum (int, sup). Finally, essential supremum, denoted as ess sup, refers to the supremum of a function for al but sets of measure zero in its domain These functional analysis concepts can be found in greater detail in 174) Time Domain ‘* C\(R,R) is the space of absolutely integrable scalar functions, equipped with the norm (a0) FUNCHIONSPACES 451 © £\(R,R"™*) isthe space of absolutely integrable matrix functions, equip. ped with the norm fil mol wn (Any 1* £:(R,R™") is the space of square integrable matrix functions of 1 € R, with inner product - + 74(B, RY") isthe subspace of £(R, R°*") formed by causal functions, that is, zero for: <0, The notation £,((0, 20), R”™) is sed equivalently + 715 (8, RP") isthe subspace of £3(R,R”") consisting of anricausalfonc- tions, that is, zero for 1 > 0. The notation £((—o0,0],R"*™) is used equivalently. Next, we show that for a causal stable system G, ||Gilz can be intexpreted as the RMS value ofits output z = G+ u when the input u isa white noise signal with unit covariance. To this effect, given a wide sense stationary random vector signal u(r) define its autocorrelation as ate) =€ ote] cary where £ denotes expectation. The Fourier transform of Ru(1) is called the spectral density Suu(jw), that is, Saja) = J Rane ar (a4) follows that €(u w) = trace [E(uu)] = trace [Ry(0)] Ibis a standard result that if z= Gu then Sj) = G(jo)Su(ja)G"Gw). Thus aeto-meettool= 2 ["wowtutwide Ly eS 52 = ay [Trace 610)" a) = 16 where we used the facts that Sya(jiw) = I and that trace(GG") = trace(G").452 MATHEMATICAL BACKGROUND Frequency Domain # C)R,C"™") is the space of square integrable matrix functions of jo, @ € R, with inner product, = a | se (FUG y0y"}de (Ais) and norm ||F\|2° V . Here we have defined G'(s) £ GT-s), + 7(UR,C™") is the subspace of £(jR, C""") composed of analytic func- tions in Re(s) > 0 such that spf trace|Fjw+o)Fyw+o)'|de 0 = 1,UR,C™™) is the subspace of £..()R,C™") of analytic and (essen- tially) bounded functions in Re(s) <0. £{R,-) and £,GR,-) are Hilbert spaces; £,.(JR,-) is a Banach space. In many cases, when these spaces are restricted to the set of rational transfer functions or matrices, an R is added to the notation, that is, RL3, RLao. RH, RM. In the main part of the book, the arguments of these spaces will be dropped for simplicity, these being clear from the context A.2.4 lsomorphism Definition A.14 Two vector spaces V and W are said to be isomorphic if and only if there exists a linear transformation L.: V —+ W, satisfying the following conditions: 1. L(x) = LO) > x= y EV, that is, N(L) = {Q}. This property is defined as monomorphism or one to one. 2. w= L(x) for all w € W with x € V, that is, R(L) = W. This property is called epimorphism or onto. FUNCIIONSPACES 453 Here N() and R(.) represent the kernel and range spaces of an operator, respectively. Definition AS Two normed vector spaces (V, |v) and (W,||- lw) are said to be isomette if and only if they are isomorphic and, in addition, the @ linear transformation L satisfies lly = WL(0)|w for any vector v € V. Example A The normed linear vector spaces £3(R,R) and £3(R,©) are isomorphic and isomeric under the linear transformation F : g(t) Ga) defined as the Fourier transform: Vf oper J 16003? ao This is also known as Parseval’s theorem. Furthermore, the inner product ‘among vectors in both spaces is also preserved, isle Giz (A.19) y — (A21) As a.consequence, the “algebraic structure” of both spaces is similar. In addi- tion, orthonormal vectors in one space will remain orthonormal in the trans- formed space, that is, “geometry” and “proportion” are preserved from one vector space to the other. Due to the isometry that the Fourier transform establishes between £(R) and £,()R), their subspaces can be related in the same way. Therefore %9(R) and Ha()R) are isometric, as well as 43 (R) and 13 ()R). Hence causal signals (or systems) ate isometric with stable transfer matrices and anticausal signals (or systems) with antistable ones. A.2.5 Induced Norms Definition A16 Consider two normed linear vector spaces (V,\\- Iv) and (WV, Iv) and a linear transformation L : V — W. The induced norm of the transformation is defined as follows: A IL()lw Wy 2 sup WOOl (a2) vive Telly It can be proved that equivalently the induced norm can be computed as SUP yy, 1 HEC)lw.454 MATHEMATICAL BACKGROUND Example A.7 1. Consider the operator A € R"”" and use the Euclidean vector norm js in both R" and R” A. ieTATAX Malia © 9B Tid sel V aT = YEE) | aaa) (ax) The latter will be defined in the next section as the maximum singular value of A, 2. If the norm |) is used instead in RY and R, the following induced norm is obtained: Ath x = max Yai (A2a 3. When || - oo is used in R" and R" the following induced norm is obtained: All ow = max lanl (a2s) i ‘The induced norm quantifies the maximum “amplification” of a linear operator, 8 generalization of the concept of gain in SISO transfer functions. ‘Take, for example, the following matrix: f [om 10° 0 0.001 (a26) which can be interpreted as a linear operator A :R? ~ R?. Its eigenvalues Ay 0.001 do not measure the maximum amplification of vectors in RE when multiplied by the matrix. On the other hand, its induced norm (Alla 2 = (A) ~ 10? does. For example the input w= 0. 1)", produces an ‘output having Aullp = a(A) = 108 ‘An important property of induced norms is the fact that they are sub- smulipicative, that is, [> Lill w < [!Lalv—wlLi yw Proof. From (A.22) we have that jLyllw < hLllw-—wlyllw- Ht follows that MaLayllw < Waliw av hLaylw < Law ovliLallw-awllyllw— (A27) Equivalently, Mataylh Se < Uallw_wllLslww A28) ly = Weal wile (A28) ‘The proof follows now directly from the definition of induced norms, © FUNCTION SPACES 458 A.2.6 Some Important Induced System Norms In this section we consider the £* and £? induced system norms and we show that they coincide with the €, and H.. norms introduced in Section ALB The £*-+£ Case For simplicity consider the SISO case. For any in put u(t) € £*, lull = 1, we have that the corresponding output y = g + w satisfies boots fee sandr s [enue s [nian =i ‘This shows that lglle.c. < ligll. To show that the equality holds, assume that lg = Ho. Then, given € > 0, there exist T such that (A29) bom es fella sto (a30) ider che following signal aoe {omar osrsT 0 otherwi (ey) I is easily seen that the corresponding output y(¢) satisfies jy > |y(T)] > Bo ~ €. Thus sup gulls 2 Ho ~€ (ax) Halles ex ‘The proof follows now from the fact that ¢ is arbitrary, The CC Case Consider any input w € £2,;Iuj2 = 1 and let z = Gu Using the frequency-domain definition of |- >. we have that Hells = WGu(s)3 (A33) =¢ [ L2G Yay yaya) ds ve Eff stove genie de 1p (Ge 3 fw e)W10) deo GON. (As)456 MATHEMATICAL BACKGROUND ‘This shows that the Mq norm is an upper bound of the £> induced norm, To show equality we need to find some w. € £2, |Iu.ll2 = 1, such that [22 = UiGllx. Assume that sup, @ [Gla)] is achieved at the frequency w.. Consider the input 1a(t) = k(e}e-* £05 (wal)man, WHETE Vas iS the Unitary right sin gular vector that corresponds to 4 (GUw,)] and k(e) is such that |4.(3)h for € > 0. As ¢ +0 we have that leallS = Tim G(s)u.(5)IH 5 i [GUoomek(e)[6(w +0.) + 5(w— a)? do + HG, )omex|” = 2 (Gye)] =I), (35) The 4 Norm ‘The Hz norm does not admit an interpretation as a norm induced by a single space (and thus it does not have the submuliplicaive property). Nevertheless, it i instructive to show that it can be considered fs an induced system norm, if the input and ouput spaces are allowed to have different norms. Given a signal u(t) we can define its autocorrelation and spectral density proceeding exactly as in Section A23. If Ryi(2) and Sy(jw) exist and Ry, is bounded, then w is said to be a power signal, with power (semi)norm given by |lull = trace [R,,(0)]. A power signal v is said to have bounded spectral density if ||Si«(J0}l|~ <0. In this case we can define its spectral density norm |jlls = |Sve(Jw)|}ac- Proceeding as in Section ‘A.23 it can easly be shown (by considering a signal u(?) having Sy, = 1) that Gl = IIIs Finally, in the SISO case, |j.\|2 can be interpreted as the £; — £o. norm. ‘To establish this fact consider € Cx, Illy 1. The corresponding output satisies viol = [f° ete ~ nara « ([Lat-otae)” (fiwerae)" = Isl (A36) where the second line follows from the Cauchy-Schwartz inequality, To com: plete the proof we need to show that there exists at least one signal u(t) € Lays lhully = 1 such that lly > gla To this effect take a-d Osh (a37) DUALITY AND DUAL SPACES 487 Obviously falls = 1 and Ae lla > Iy(OW TED Wolke 2 DOl= Tei Fa A.3 DUALITY AND DUAL SPACES (Aan) 4.3.1 The Dual Space Consider a linear vector space X. A functional on X is a function f:X'— R, Definition AAT Consider a normed linear space X. Its dual space X" is defined as the space of all bounded linear functionals on X, equipped with the norm | fl = stPjx4<1 If) It can be shown (187 that X" isa Banach space. In the sequel, given x € X and r © X", (x,r) denotes the value of the linear functional r at the point x. Example A.7: The Dual of f°, 1< p< oo Consider the space #” of all real sequences h = {,} for which >", /hi)? < 90, equipped with the norm Welle = COR Mul?" (i p =e then [Ase = supy !hu|)- Given 1 < p < 00, define the conjugate index g by the equation! (A39) Next, we show that every linear functional fon €? has a unique representation of the form A) = Show (A40) where = {n} € €*. Assume first that 1 0, there exists my € M such that ie ml)
QandU ecm, are unitary matrices. ‘crow pan "Thun npone ocelot ast of rbonomategemesr satay DE o viacaw = [8 8] ass Let = YEA. Without tos of generaiy we ssume thatthe eigen vale of WPA are ode. that 2 Dene Vand te itr anv -n columnsof Vesecvey Repiang m (A00) min {\jrl>, Ax =b, Ae C™™, n < mh (as) The solution to these problems can be obtained by means of the matrix pseudoinverse or generalized inverse A*. The latter can efficiently be com puted with the use of the singular value decomposition as follows: =z 0 a-u[% 8 : : EO] y. Je eer fs Qo ay ere (Ae) ‘The input vector x = (1 1/2 L2"" yields Ax = —2""1 ay that tends to the zero vector as n —+ 90. Hence for a large enough m this ‘matrix is computationally singular, although itis theoretically not. ‘The eigenvalues are Ay = --- = An = —1 and the determinant is det(A) = (-1)", none of which provide a clear measure ofthe computational singularity of this matrix. In this last case, the reason is that det(A) = []} A,, which therefore depends on all eigenvalues, not only on the smallest one. In the former case, the measure provided by the smallest magnitude eigenvalue only quantifies the effect of the matrix on the eigenvectors, but not on any arbitrary vector in the input space ‘On the other hand, by considering the minimum value [Axl can take over all possible input vectors x (normalized by {illp 4.0 10 avoid trivial solutions), the measure of singularity can be computed as Axi RTATAR int Mele ing iehzo [ie] neo (EPID way twat which is directly the minimum singular value oA) = 2!" SINGULAR VALUES 463, Condition Number Consider the linear matsix equation Ax = b. The ‘maximum relative error 6b due to an error 5x, can be found as follows 86H yp Adele ner [Ole ~ Se ART =1 with axils = lal sup IAB inf, [Axl with [Sxl = [lela = 1 A) 2 7 wa!) (Ass) where 5x and x are independent vectors. The problem is equivalent to the one of finding the maximum relative error 8x of the solution for @ given error 8b in the data vector b. The condition number x(:) provides this measure and quantifies the relation between the maximum and mini- mum “gains” of A (according to different input vector “directions”). For ‘example, orthogonal (rotation) or unitary matrices are “spherical” because U'U = 1 = oy == oy = 1s hence in this ease «(U) = 1. On the other hhand, for singular matrices « —+ oo.APPENDIX B SYSTEM COMPUTATIONS In the following sections we present the state-space representations of the ‘most usual interconnections between linear systems. These representations are used to manipulate systems in a computer code. There is standard soft ware that implements these interconnections in state-space form, as in the case of the j-Toolbox and the Robust Toolbox of Matlab. Consider the following linear systems ous) = 54] (Ba) 2 | Be | (62) Gals) = |-p with states, inputs, and outputs (x1,4,y1) and (x, u2,92), Fespe series connection can be seen in Figure B.1, where uu). yam ‘The objective is to compute a state-space realization (not necessarily min- imal) of the product system. To this end, consider the state of Ga(s}Gu(s) as ely. The and x= [xf 21)’, and proceed as follows: AU = Arante) + Bini) (B3) 465