0% found this document useful (0 votes)
297 views

Control Theory - Robust Systems, Theory and Applications

Uploaded by

surfer109
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF or read online on Scribd
0% found this document useful (0 votes)
297 views

Control Theory - Robust Systems, Theory and Applications

Uploaded by

surfer109
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF or read online on Scribd
You are on page 1/ 258
eee Panel PEIDNEAS VN3g-Z3HINYS, ria ° & RosBust SYSTEMS nme | ae sta) Dar ACA ane Waren ochre eon ee es Sea) een eer eee eat phe iene j Leetonia) ener ane A n See) Korte et ae eee ara Mario SZNAIER WY ci increasingly important subject Co ee Le ae See Pee enn een ees CO eae ee es are complemented by elaborated examples and a group of worked: De eee reac et nee Ce ee ee eon tet Ce eet eed ee) ‘expressly a5 textbook for master’s and first-year PhD ete REINA d-ZaHONYS TNE) ces basic robustness concepts inthe context of SISO systems described by Laplace nets eet ke eae ee ee ee Se eee ea a ene ee] oad es the four basic problems in robust contol and the Loop shaping design method ROBUST SYSTEMS THEORY ano APPLICATIONS ee ee ee Ce eee ee) = ee ee ee ee Te Se ee eC ee snr aes oe er and Le eee) ek ke ee eee s the basic tools of model order reduction Se eee ‘numerous endot-chapter problems and workec-out examples of robust contol See cd Activities (CONAE) and Professor of Contol Systems atthe School of Engneesng at eee ee een eee ere Kel S STORES eee a ering at Pennsylvania State University, University Patk, USA Mario SZNAIER WON wi WILSAS 1sngoy ROBUST SYSTEMS THEORY AND APPLICATIONS Adaptive and Learning Systems for Signal Processing, Communications, and Control Editor: Simon Haykin Werbos / THE ROOTS OF BACKPROPAGATION: From Ordered Derivatives to Neural Networks and Political Forecasting Kislié, Kaneliakopoulos, and Kokotovié / NONLINEAR AND ADAPTIVE CONTROL DESIGN Nikias and Shao / SIGNAL PROCESSING WITH ALPHA-STABLE DISTRIBUTIONS AND APPLICATIONS Diamantaras and Kung / PRINCIPAL COMPONENT NEURAL NETWORKS: THEORY AND APPLICATIONS Tao and Kokotovié / ADAPTIVE CONTROL OF SYSTEMS WITH ACTUATOR AND SENSOR NONUNEARITIES Tsoukalas / FUZZY AND NEURAL APPROACHES IN ENGINEERING Hryce / NEUROCONTROL: TOWARDS AN INDUSTRIAL CONTROL METHODOLOGY Beckerman / ADAPTIVE COOPERATIVE SYSTEMS Cherkassky and Mulier / LEARNING FROM DATA: CONCEPTS, THEORY, AND METHODS Passino and Burgess / STABILITY OF DISCRETE EVENT SYSTEMS SGnchez-Pefia and Sznaier / ROBUST SYSTEMS THEORY AND ‘APPLICATIONS Vapnik / STATISTICAL LEARNING THEORY ROBUST SYSTEMS THEORY AND APPLICATIONS Ricardo S. Sanchez-Pena Mario Sznaier ® ‘A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York / Chichester / Weinheim / Brisbane / Singapore / Toronto ‘This book i printed on aides paper. @) Copyright © 1998 by John Wiley & Sons Ic. Allright reserved Published simakancousy in Canada No past of this publication may be reproduced stored in revival system or tranamited in any form or by any meas, cto, mechanical. photocopying. record. eamnng cpherwise, except as permitted under Section 107 oF TOR ofthe 1976 United Sites Copyright Drive, Danvers, MA 01923, (978) 750.840 fax (978 7504744 Requents to the Publades lor Permission should be adiressed 10 the Permissions Department ke Wie) & Seon fe (05 Third Avenue, New York, NY 101580012, (212) 8504011 fax 212) 850e008 E-Mail: PERMREQ@WILEY COM, Library of Congress Caaloging-in Publication Data: Sinches Pena. Ricardo boom Includes bibliographical references and inde. ISBN 0471176073) |. Automatic control 2. Control theory. 3. System ‘entiation. L'Sensier, Mario Tile TAISTISe 1998 e9s-de2I 9.6820 c Printed inthe United States of Ameria WORTH SE TIT To our parents Lolita and Miguel, Rebeca and Marcos CONTENTS | Introduction 1 1.1 General Control Problem / 1 LL Experimental Phase / 2 11.2. Simulation Phase / 3 113. Theoretical Phase / 5 1.2. Why Feedback? / 11 13 Feedback Loop Trade-olf / 14 14 Objectives of an Applied Theory / 17 15. Objectives of This Book / 19 15.1 General Remarks / 19 152. Scope / 19 153. How to Use This Book / 21 SISO Systems 23 21 Introduction / 23 22 Well Posedness / 24 23. Nominal Internal Stability / 26 24 Robust Stability / 29 24.1 Phase and Gain Margins / 29 242 Global Dynamic Uncertainty / 32 25. Nominal Performance / 42 25.1 Known Disturbance/Noise/Reference / 42 252 Bounded Disturbances at the Output / 44 253 Other Performance Criteria / 47 26 Robust Performance / 50 2.1 Extension to MIMO Systems / 55 28 Problems / $7 3 CCONTENIS Stabilization 6 3.1 Introduction / 61 3.2 Well Posedness and Internal Stability / 63 32.1 Well Posedness / 63 32.2 Internal Stability. / 64 33. Open-Loop Stable Plants / 65 34° ‘The General Case / 68 34.1 Special Problems / 69 342 The Output Feedback Case / 75 35 Controller Structure and Separation Principle / 78 3.6 Closed-Loop Mappings / 79 37 A Coprime Factorization Approach / 80 37.1 Coprime Factorizations / 80 38 LFTs and Stability / $8 3.9 Problems / 90) Loop Shaping 93 4.1 Introduction / 93 42 Nominal Pesfosmance 97 43° Robust Stability / 98 44 Nominal Performance and Robust Stability / 99 45, Robust Performance / 100 45.1 Sensor Uncertainty / 101 45.2 Actuator Uncertainty / 102 46 Design Procedure / 105 4.7 Examples / 109 47.1 Permanent Magnet Stepper Motor / 109 47.2 Loop Shaping Q(s) / 119 48 Related Design Procedures. 123 49° Problems / 125 %_ Optimal Control 127 5.1 Introduction / 127 5.2. The Classical Linear Quadratic Regulator Problem / 128 53° The Standard 4, Problem / 133 54 Relaxing Some of the Assumptions / 141 5.5 Closed-Loop Properties / 142 55.1 The LOR Case: Kalman’s Inequality / 143 CONTENTS x 55.2 Some Consequences of Kalman’s Inequality / 145 553. Stability Margins of Optimal 4» Controllers / 146 56 Related Problems / 149 57 Problems / 152 Hoo Control 157 6. Introduction / 157 62. The Standard 4, Problem / 158 62.1 Background: Hankel and Mixed Hankel-Toeplitz Operators / 161 62.2 Proof of Theorem 6.1 / 176 63. Relaxing Some of the Assumptions / 179 64 LMI Approach to H. Control / 180 64.1 Characterization of All Output Feedback Controllers / 183 642 Connections with the DGKF Results / 188 65 Limiting Behavior / 192 66. The Youla Parametrization Approach / 193 67 Problems 203 Structured Uncertainty 207 7.1 Introduction / 207 TAA Stability Margin / 210 72. Structured Dynamic Uncertainty / 215 72.1. Computation / 215 72.2. Analysis and Design / 219 73. Parametric Uncertainty / 223 731 Introduction / 223 732. Research Directions / 225 733. Kharitonov's Theorem / 230 734 Mapping Theorem / 233 74. Mixed Type Uncertainty / 237 74.1 Introduction / 237 742 Mixed w / 239 75 Problems / 242 Control 248, 8.1 Introduction / 245 10 ‘CONTENTS 82. Robust Stability Revisited / 247 82.1 Robust Stability Under LTV Perturbations / 250 82.2. Stability Under Time Invariant Perturbations / 253 83 _A Solution to the SISO ¢' Control Problem / 254 83.1 Properties of the Solution / 262 83.2. ‘The MIMO Case / 266 84° Approximate Solutions / 267 84.1 An Upper Bound of the é Norm / 268 842 The + Norm / 269 843. Full State Feedback / 273 844 All Output Feedback Controllers for Optimal «Norm Problems / 276 85. The Continuous Time Case / 282 85.1 Solution Via Duality / 282 852 Rational Approximations to the Optimal £, Controller / 285 8.6 Problems / 288 Model Order Reduction 9.1 Introduction / 293 92 Geometry of State-Space Realizations / 295 92.1 Controllable/Unobservable Spaces / 295 922. Principal Components / 297 93 Hankel Singular Values / 304 93.1 Continuous Systems / 304 93.2 Discrete Systems / 306 94 Model Reduction / 310 94.1 Introduction / 310 9.4.2 Hankel Operator Reduction / 310 943 Balanced Realizations / 312 944 Balanced Truncation / 314 95 Algorithms / 318 95.1 Approximation Error / 319 9.6 Problems / 321 Robust Identification 10.1 Introduction / 323 102 General Setup / 324 293 323 103 104 105 contents xi 02.1 Input Data / 325 10.22 Consistency / 326 102.3 Identification Error / 328 10.24 Convergence / 334 102.5 Validation / 338 Frequency-Domain Identification / 339 103. Preliminaries / 340 1032 Sampling Procedure / 342 1033 Consistency / 344 10.34 Identification Procedures / 346 ‘Time-Domain Identification / 361 1041 Preliminaries / 362 10.42. Idemification Procedures / 366 Further Research Topics / 372 105.1 Unstable Systems / 372 1052 Nonuniformly Spaced Experimental Points / 372 1053. Model Reduction / 373 1054 Continuous Time Plants / 373 155 Sample Complexity / 373 105.6 Mixed Time/Frequency Experiments / 374 105.7 Mixed Parametric/Nonparametric Models / 374 Application Examples 377 ie n2 u3 na SAC-C Attitude Control Analysis / 377 141 Introduction / 377 11.4.2 Linear Model / 378 11.13 Design Constraints / 381 11.14 Robustness Analysis / 383 115. Simulations / 385 Controller Design for a DO Plant / 389 11.2.1 Model of the Plant / 389 11.22. Robustness Analysis / 391 11.23 Controller Design / 397 X-29 Parametric Analysis / 400 113.1 Linear Model / 400 1132 Results / 405 Control of a DC-to-DC Resonant Converter / 407 11.4.1 Introduction / 407 xl CONTENTS CONTENTS xi 11.42. The Conventional Parallel Resonant Converter / 407 BO Inverse / 469 1143 Small Signal Model / 409 B.10 Linear Fractional Transformations / 470 1144 Control Objectives / 410 BAL Norm Computations / 475 1145 Analysis of the Plant / 411 B.IL1 7% Norm Computation / 475 1146 Control Design / 414 B112 Me Norm Computation / 476 1147 Controller Synthesis / 420 B13. € Norm Computation / 476 1148. Simulation Results / 421 B.I2 Problems / 479 Bibliography 427 ¢ Riccati Equations ar ‘A Mathematical Background 445 Index 487 Al Algebraic Structures / 445 ALI Field / 445 A.L2 Linear Vector Space / 446 AL3 Metric, Norm, and Inner Products / 447 ‘A2 Function Spaces / 449 21 Introduction / 449 22. Banach and Hilbert Spaces / 449 23. Operator and Signal A24 Isomorphism / 452 A25 Induced Norms / 453 A26 Some Important Induced System Norms / 455 AB. Duality and Dual Spaces / 457 A31_ The Dual Space / 457 32 Minimum Norm Problems / 458 A4 Singular Values / 460 AAI Definition / 460 AA2 Properties and Applications / 461 paves / 450 B System Computations 465 BI Series / 465 B2 Change of Variables / 466 B3_ State Feedback / 467 Ba ‘timation / 467 B.S Transpose System / 467 B.6 Conjugate System / 467 BT Addition / 468 BS Output Feedback / 468 tate PREFACE Robustness against disturbances and model uncertainty is at the heart of control practice, Indeed, in the (completely unrealistic) case where both all external disturbances and a model of the system to be controlled are exactly knowin, there is no need for feedback: Optimal performance can be achieved with an open loop controller. The main ingredients of present day robust control theory were already present inthe classical work of Bode and in many popular frequency domain: based design techniques. With the advent of state-space methods in the mid 1960s the issue of robustness took a backseat t0 other topics, but was never completely abandoned by the contzol community. especially control practi tioners, Interest in robust control rose again in the late 1970s where it was shown that many popular control methods (including optimal LOR control and controller design based on the cascade of an observer and state feedback) Jed to closed-loop systems very sensitive to model perturbations, Moreover, widely accepted “ad-hoc recipes for “improving robustness.” such as using "nifiiallylarge noise levels in the design, had precisely the opposite effect. Robust control has undergone extensive developments in the past «wo decades, leading to powerful formalisms, such as 7, w-synthesis/analysis and, more recently "optimal contro, that, coupled with newly developed control-oriented identification techniques, have been successfully applied to challenging practical problems. A salient feature of the framework is that itis oriented towards applications and thus is based on “practical,” realistic assumptions “There are many excellent books that cover specialized topics (Hz, Ma ©, parametric uncertainty linear matrix inequalities) with others scattered in the technical journals, Our intention in writing this book is to provide a self contained overview of robust control that illustrates all the issues involved, ranging from the transformation of experimental signals from the physica plant toa set of models (robust identification), to the synthesis of a controller {or that set of models (robust control), The purpose of the book is twofold to serve as a textbook for courses at the Masters/beginning Ph.D. level and as a refesence for control practitioners. It assumes that the reader has a background inclassial and state-space control methods. In order to keep the xvi PREFACE text size at a manageable level, in some cases only basic results ate covered, and the reader is referred to more specialized literature for further coverage In all cases we have strived to link the theory with practical applications. To this end, in addition to the examples covered throughout the book, the last chapter contains several worked out application problems that stress specific practical issues: nonlinearities, unknown time delays, infinite dimen- sional plants, actuator and sensor limitations. They are all extracted from our practical experience in different engineering fields. Furthermore, due to the fact that most of the problems to be solved by the theory presented here are computer-intensive, we stress the algorithmic and computational aspect of the solution along with the mathematical theory. We (and our graduate students) have tried to eliminate obvious mistakes, However, as anyone who has tried knows, itis virtually impossible to make 1 “perfect” book. We encourage readers to send corrections, comments and general feedback to either one of the authors. Finally, as altested by the list of more than 300 references, we have tried to give credit where it is due, However, owing to the sheer volume of literature published on robust control, we may have inadvertently failed to do so on occasion, We apologize in advance to readers or authors who may fee! that this is the case, and we encourage them to send us comments, ACKNOWLEDOMENTS We owe a special debt to Professor Thanasis Sideris from whom both of us learned a great deal (not only about robust control) and who was very influential in shaping our subsequent research. We would also like (0 a: knowledge the important contribution that Professor Manfred Morari had in the development of the concepts presented in this book, and in advising Ricardo S, Sanchez Pea to make the right decisions during hard times in his career. Mario Sznaier is specially indebted to Professor Mark Damborg. for his mentoring. Among Professor Damborg’s many contributions, the two that stand out are attracting him to the field of control and making sure that he got a Ph.D. degree when he showed signs of enjoying life too much as a sraduate student The reviews by Professors Peter Dorato and Roy Smith at initial stages « of this book have been instrumental in encouraging the authors to carry out the project and in shaping its final form. Preliminary versions of the book were tested in courses offered by the au- thors at Penn State University and the University of Buenos Aires, Feedback from students enrolled in these courses was fundamental to its fine tuning We are also indebted to our graduate and post-doctoral students, Takeshi ‘Amishima, Pablo Anigstein, Juanyu Bu, Tamer Inanc, Cecilia Mazzaro, Pablo Parnlo and Zi-Qin Wang, for reviewing the book in great detail and provid- ing some of the material, In particular, Cecilia developed very efficiently the PREFACE xv ‘examples in Chapter 10 and Takeshi, Juanyu, and Zi-Qin contributed some of the material in Chapters 8 and 11 ‘The book was shaped to a large extent by interactions with our colleagues. In particular, many of the ideas and developments in Chapter 8 arose from research carried out jointly with Professor Franco Blanchini. Dr. Hector Rot- stein influenced the presentation of the, material in Chapter 6, Profes- sors S. P, Bhattacharyya, D. Bernstein, N. K. Bose, O. Crisalle, M. Dahleh, M.A. Dabileh, J. C. Doyle, M. Fan, C. V. Hollot, M. Khammash, A. Megret ski, A. Packard, J. Boyd Pearson, A. Ray, A. Saberi, A. Stoorvogel, R. Suarez, R. Tempo and J. Vagners contributed (directly or indirectly) to this book through many discussions and comments during the years. ‘The first author would like to thank his colleagues at LAE, School of Engineering of the University of Buenos Aires and at the Argentine Space ‘Agency (CONAE), especially Professors C. Godifrid and A. Perez, and Ing. Beto Alonso for important support. The financial support of CONAE, the University of Buenos Aires and Fundacién Antorchas during his stays at Penn State in 1994 and 1996 are also gratefully acknowledged, “Mario Sznaier would like to thank the National Science Foundation, and in particular Dr. Kishan Baheti, Director of the Engineering Systems Program, for supporting his research program. This book would have not been possible ‘without this support We are also indebted to Ms. Lisa Van Horn and Mr. George Telecki at John Wiley and Sons, Inc. for their assistance throughout this project. Finally we would like to thank our children, Gadiel, Lucila and Pablo, and spouses, Ménica and Octavia, for their infinite patience, support, and understanding for the time this book has stolen from them. . Ricanpo $. SANCHEZ-PESA Manio S2NAieR Buenos Aires, Argentina University Park, Pennsylvania ROBUST SYSTEMS THEORY AND APPLICATIONS INTRODUCTION 1.1 GENERAL CONTROL PROBLEM The purpose of this book isto introduce the reader to the theory of control systems, with particular emphasis on the applicability ofthe results to pract- cal problems. Therefore it considers not only the control viewpoint but also the ‘denifcation of systems as part of the problem. Also, as for any theory of systems oriented toward practical applications, robustness is essential and will be the underlying concept throughout the book, For these reasons the “tle of the book includes the following words: Theory, Applications, Robust, and Systems. It is important to start with a wide angle view of the general control picture, so that we can locate ourselves in the particular approach focused by this book. For this reason, the general problem of implementing a con- troller, which aetuates over a physical process in a prescribed way, will be divided into three well defined phases. These phases are all equally impor tant but involve very different echnical cools, which go from pure mathe- ‘matics to microprocessor technology. In today’s practical control problems, all phases should be considered and itis very difficult to have a clear knowl edge of all the technical fields involved. Although in this book we con centrale on the theoretical and computational aspects of the whole problem, we first make a general overview of the disciplines involved at teach stage and the interrelation among them. This is the basic objective ‘of robust identification and control theory: to produce theoretical results that are closely related to the computational and experimental aspects Of the control problem. Tn the next subsection we will describe each phase in detail 2 INTRODUCTION 1.1.1 Experimental Phase the control problem is generated and where the last test ‘on the controller is performed. This means that the actual control problem to be solved comes always from a physical setup: an aircraft, a distillation plant, a robotic manipulator. The performance specifications are also formulated in terms of variables involved in this physical experience: stable dynamics at high angle of attack, constant concentration of the distilled product, fast response to commanded trajectories. Also, in many cases, once the practical problem has been formulated, an experiment should be performed to:obtain ‘a mathematical model as “close” as possible to the real physical system to be controlled. Finally the last test to prove the effectiveness of the controller is performed on the physical system, ‘The experimental phase is basically technological. It involves engineer- ing disciplines (electrical, mechanical, chemical, acronautical, ete.) of even biology or economics, depending on what we define as the system to Be con ‘rolled. It also involves knowledge on the sensors, actuators, and technical aspects of microprocessors and 1/O devices, for computer control. We may as well include in this phase the low- and high-level programming, as in the case of communication programs in Assembler and control algorithm coding, respectively. ‘The interaction of the physical plant with its environment is performed by “real world” signals provided by the plant to the sensors and application ‘of physical stimulations to the plant through the actuators, which modify its behavior. These input and output signals, as well as the physical laws that describe the system behavior, provide the only information we have of the physical plant. In many cases there are no physical laws that explain the plant behavior or they are extremely complex. A trial and error contol, or direct experimentation using the input and output signals (without any analysis), is untealizable, at least for the class of problems in which we are interested. ‘A more elaborate and “scientific” procedure should (and can) be performed {0 control a given plant. Furthermore, in many cases trial and error would not even be practical because of costs (chemical plants) or simply to prevent accidents (aircraft, nuclear power plant). ‘The information on the system—input-output signals and physical laws— is never complete (to complicate things). The reason is that the physical laws, which may be applied to the (sensor, actuator, plant) behavior describe only’. 4 certain aspect of the problem, but not all. To make this point, consider the following example. To describe the output voltage on a given resistor through which a given current is circulating, we may use Ohm's law. Nevertheless, the temperature, gravitational field, and magnetic field could also influence the ‘measurement, as well as the physics of the voltage sensor and current supply. A description of a particular physical process using all the many aspects involved could, to say the least, be impractical.' Sometimes, also for practical is an epistemological question sf complete mathematical picture of physical rely GENERAL CONTROL PROBLEM 3 reasons, a simplification of the physical laws should be applied to describe the system. Actually, the laws are not modified, but instead new simplifying assumptions are added The above facts determine that all practical control problems have a cer lain amount of uncertainty in its more general meaning. It should be noted that this uncertainty is a consequence of the observation of the system and its environment, not due to the system itself. In particular, in the control jargon, this lack of knowledge is called noise when applied to the sensing, proce: dure, disturbance for the external stimulations to the plant, and uncertainty when it applies to the description of the dynamical behavior of the plant by ‘a mathematical model. These three elements—in particular, the latter—play 4 fundamental role in the orientation of control theory since the 1980s and in the motivation for Robust Identification and Control. For many different reasons it is highly convenient to have a computational ‘model of the plant, By this we mean a representation, in a computer, of the relevant inputs and outputs of the physical system and their relationship. This representation can be very elaborate as in a hardware in the loop scheme, where parts of the system are represented computationally and others are ‘connected physically to the computer I/O ports? Otherwise it can be a com- plete computational model of the plant obtained from a mathematical model that includes all relevant physical phenomena involved “To illustrate some of the many reasons justifying a simulation of the plant, we cite the following “real world” examples, which will be described in greater detail in Chapter 11 |.2 Simulation Phase 1, In many cases direct experimentation with the hardware is impossible, as in the case of satellites. In these situations, itis usual to start with a high-fidelity simulation, which includes models of the system (satellite), actuators (inertial wheels, gas jets, magnetic torque coils), sensors (s0- lar and Earth magnetic field sensors), and environment (Sun Ephemeris and Earth magnetic field model along the orbit). These models include the nonlinear dynamics, flexible phenomena, and all the necessary ele- ‘ments to represent the system. According to required tests to be per formed along the project, pieces of real hardware may be connected to this hi-fi simulator replacing their simulations, that is, sensors or actua- tors. 2. The cost of experimentation could be a strong reason to simulate a plant, ike the case of many chemical processes, where the startup of 1, could he argued that in thi case we fave » combination of experimental and simulation phases, What we mean hy simulation in ths book i im genera an eect that isnot exaetly the eal plat or envionment In any case. the important point st realize that there ae diferent pasts the onto problem, 4 IntRODUCTION the process could involve many millions of dotlars. The high costs could be a consequence of very expensive materials or long experimentation, ‘Testing over a computer model provides a very fast and inexpensive way of verifying a controller (see DzO plant in Chapter 11 and [228)), 3. A very important reason for simulation is when human lives are in volved, as in the case of an experimental aircraft, Therefore a com- puter simulation gives a safe means of testing high-risk maneuvers (see NASA's X-29 experimental aircraft in Chapter 11), The disciplines involved are related to the areas of computer science, op- timization, and numerical methods. The mathematical models to be imple- mented by a computer code could be obtained from physical laws or from direct experimentation on the plant. In this last case, the classical Parameter Identification procedures ({184, 293)) or the very recent Robust Identification ((146)) techniques (see Chapter 10) could be used. The main goal of the simulation phase is to have a copy as “close” as possible to the experimental plant. The word experimental has been empha- sized to distinguish it from the true plant, which is not the same concept, as ‘mentioned in the previous subsection and detailed next. For simulation, we are interested only in certain aspects of the physical system but not in all of them. Take, for example, the dynamics of the longitudinal axis of a rigid aircraft, which is a partienlar aspect ta he considered not at all related to the ‘thermal gradients of the wings of the same aircraft. Therefore, for simulation purposes, we only consider certain input and output signals that are relevant ‘o our problem. Furthermore, even for a particular aspect of the problem, we may consider different levels of detail. Take the same example, where in cone case we may consider the aircraft as a rigid body and in a more detailed analysis we may add the flexible modes. Therefore, for simulation purposes, we only consider the experimental plant, which is a particular aspect of the ‘rue plant, that arises in the experimental setup. ‘The trade-off in this phase of the problem is defined in terms of how “close” we want the computational model to represent the plant versus how “big” a computer will be needed to perform this task. In the latter, the main limitations are computation time and available memory. As the computa- tional model describes the plant better and better, the computation time will grow, which could pose a serious limitation in many real-time applications. For example, even for reasonable size minicomputers, a complete modef of the orbit, Earth magnetic field, Sun Ephemeris, gravitational and acro- dynamic disturbance models, and satellite dynamics could take much more time for simulation than for real time. This is the case of low-orbit satellites with short orbital periods (SAC-C in Chapter 11 has a 90 minute period), Also, in many cases. a “good” computational model could involve a great amount of memory, like in many finite element representations of physical systems, Im any case, this trade-off determines that certain simplifying assumptions should he made when representing a physical system by a computer code, GENERAL CONTROL PROBLEM 5 Typical assumptions for the nonlinear model of an aircraft are rigid body dynamics and symmetry. The simplifying assumptions mentioned above, in the simulation phase, are only related to the computational constraints. In stead, for the theoretical analysis and synthesis phase, new hypothesis should be added to simplify the problem even more. This is related to the available mathematical tools to design and analyze a feedback loop, as will be seen next 1.1.3 Theoretical Phase Introduction In history and, in particular, in the history of science, humans have tried to solve first the simpler problems and continue with the more complicated ones. This, which seems almost a natural law of human behavior, also applies to the theory of control systems. For this reason in classical con- trol theory, systems were first described as single-input single-output (SISO), ‘while later on, the more general multiple-input multiple-output (MIMO) de- scription was introduced by modern control theory through the state-space ‘description. In the same line of thought, the optimization of a quadratic func- tional (LOG optimal control) appeared earlier than the Ha: optimal control problem. In this latter case, the linear 1. control problem was stated ((329]) flmost ten years before the first solutions (|22, 157, 309, 310)) to the Ax nonlinear control problem were produced. At this point, there are not yet fully developed analysis and design tools to deal with control systems described by coupled nonlinear partial differ tential equation descriptions with (possibly) uncertain parameters. Therefore the theoretical phase of a control problem, which will be the main concern .of this book, deals with a much more simplified mathematical model of the plant. Historically, the most usual system description for design and analysis of feedback systems has been finite-dimensional (lumped), linear, time invari- ant (FDLTI) mathematical models. For this class of models there is a well established theory, which also covers many of the most usual applications? ‘A visual idea of all the simplifying assumptions the designer needs to make to obtain a FDLTI mathematical model from the physical system may bbe seen in Figure 1.1. The usual steps that should be taken to compute a FDLTI model, from a very complex mathematical representation of a partic- ular plant, are presented there, Some applications, which will be detailed in Chapter 11, illustrate a few of the steps. System Model _At this point we will distinguish two different approaches that compute a FDLTI model from a physical system. The first one obtains ‘4 simplified version of a very complex mathematical model. This could be a set of partial differential equations or a time delayed or a high order set of ‘ordinary differential equations. This complex mathematical model is usually "ot hecause physical systems ae themselses FDL ho Dect for moet practical purposes 6 ntRoDUCTION Physical System (Cavs of Physics)/(Tdentification) | ed] | Nonlinear O.D.E. } 4 Nonlinear P.D.E. Linear PDE. Ll | LC | (Gtotat (ocat linorisation) linearisation). {Ayprosimation) EO) Eo Physical System = FDLTI Model + Uncertainty Figure 1.1, “Transtaton” ofa physical system to a FDLT model ‘obtained from physical laws describing the system. This model simplification procedure is called approximation and is purely mathematical, once the model Of the system, obtained from physical laws, is known, GENERAL CONTROL PROBLEM 7 ‘There are a variety of methods that approximate systems described by partial differential equations (infinite-dimensional systems) by ordinary dif- ferential equations. Among, others, we can mention finite elements, finite dif- ferences, averaging techniques, splines, and FFT methods ((24, 25, 139, 169)). ‘More recently, the balanced realization and truncation of infnite-dimensional ‘models provides simplified finite-dimensional models ({79, 134)). In the case of systems with time delays, which are also infinite dimensional, although the above methods can be used, the most usual and practical approach is to use Padé approximants. TThe second approach is based on direct experimentation over the physical system and is called in general identificarion. This approach consists in com- puting a mathematical model, usually FDLTI, from the data measured at the system's output when a particular input signal is applied. The classical param: ‘eter identification procedures (|184, 293)) assume a particular mathematical structure for the model and, through an optimization method, obtain a set of parameters that minimize the “difference” between the model and the system's output, for the same applied input. More recent techniques called robust identification obtain ditectly a FDLT1 model, without assuming any parameter structure, from the output data of the system when a particular Sel of signals are applied to the input. These latter techniques are related di rectly to robust control and will be described in greater detail in Chapter 10, From Figure 1.1 we observe that sometimes we shoul linearize a nonlinear ‘model of the system. The classical linearization technique consists in comput- ing a linear version of the nonlinear model for a particular set of parameters (operating point) of the latter. This is called a local model ({189)), which is approximately valid only in a small neighborhood of the selected values of these parameters. In some applications, the parameters have fixed values and, “ for that reason, there is only one working point and one linear model ((268)) Otherwise, ifthe parameters can take different values, the designer should synthesize a linear controller for each linear model. Nex, a test should be made (analytical, by simulation or experimentation) to guarantee a smooth transition from one controller to another (see Example below). This is called {ain scheduling and is a well known procedure to control nonlinear plants Recent results in this area attack this problem through the use of linear parameter varying (LPV) controllers ([222)) ‘Ako, for certain classes of nonlinear models (linear analytic models), a state transformation and a nonlinear state feedback can be computed, which produce a (global) linear mode! ((89, 153, 56)). Although this is not @ robust result! it is very useful in many practical applications ([1, 2, 44, 106, 107)), for example, robotic manipulators, aitcraft/spaceeraft control, and motor control. “The state-feedhack lw, which linearis the mode, depends on exact knowledge of the non linear modck therefore Fedhack lineation wa robot against uncertainty inthe onlinest 8 introDucTON Disturbances The designer not only assumes certain properties ofthe sy ‘em (e., linearity, time invariance) to obtain a simpler model but also needs to make certain hypotheses on the disturbances applied to the system by is environment. In classical and modem control, the disturbances were usually assumed to be certain classes of signals: steps, ramps, impulses, sinusoids, In 4 stochastic approach, the covariance matrix of the stochastic proce’s repre- senting the disturbance was assumed to he known exactly. Very seldom does the designer have a clear knowledge of the disturb- ing signals acting on the system. Therefore the less one assumes about their nature, the higher the possibility of producing an elfective control, despite these disturbances. As a matter of fact, a control design procedure that op- timally rejects a step disturbing signal will no longer be optimal if the actual disturbance is a square wave. For this reason, in robust control, the only assumption on the disturbances is the fact that they are bounded (usually in energy), which is a very realistic hypothesis. Ifthe designer has some extra in- formation on the frequency distribution ofthe disturbances, that information could be incorporated into the design procedure and produce a less conser- vative result. Note that even in tis eas, there is no assumption on the form of the disturbance Uncertainty Due to the assumptions on the system and the disturbances, a fair amount of uncertainty between the mathematical model and the exper ‘mental results will appear. Inthe simulation phase, the simplifying hypotheses fre made due to computational restritions (time, memory). Iasead, in the theoretical phase, the assumptions that generate the uncertainty ate a con- sequence of the mathematical tools thatthe analyst/designer has decided to ‘We can evaluate separately the effect on the feedback loop of the uncer- tainty due to the modeling ofthe system and due to the external disturbances. From_basic contro courses we know that the latter has no effect on the sta- lity of the loop, at least in linear systems,° although it can degrade the performance. On the other hand, the unce ¢ the design of a stabilizing controller has been based on an erroneous plant ‘model, the loop can be unstable. This is clear from the basics of control systems, as in the case of models with uncertain gains or phases. Take, for example, a controller designed for a model of the plant with a gain that eX- actly doubles the gain of the actual plant. If this model has a gain margin smaller than 2 (and the designer doesnt know it), the controller could make the loop unstable, Sn nantinear systems, bath the uncertainty ia the model and the umbnown dsturances can Ustahiive the enedonp system, GENERAL CONIROL PROBLEM 9 Surprisingly, mode! uncertainty has only started to be considered quantita Lively in the assumptions of control theary since the 1980s. This recent theory, which considers uncertainty as part of the problem, is called robust control and, together with robust identification, is the main subject of this book. ‘Throughout its chapters, the following fact will be stressed: the design of a controller should be based not only on a particular model, but also on the un certainty this model has with respect to the actual physical plant, Uncertainty is related directly to feedback, as will be seen in Section 1.2. Example The three phases described above are closely related and equally important. The final solution of the control problem, from the purely specu- lative theoretical step to the final physical implementation of the controller ‘of the plant, depends on all of them. Each phase has its own characteristics Although in the experimental phase there are no approximations, it is not possible to elaborate any control strategy directly from the input and output signals. The computational model used in simulations is always an intermedi- ate step, because it already contains simplifying assumptions, which limit its applicability. Finally, no matter how strong (necessary and sulficient) the con- S(s) as) ‘This variation may represent uncertainty in the modeling Ifthe variation is not infinitesimal, we can compute the relation between open- and closed-loop ‘models considering the influence ofa signal us(s) on the output of the plant ‘model y(s). These signals are related to the relative uncertainty of the loop as follows: ts(o) = BOs 65) 16 IwiroouctiON shee the uncertain loop (1+ L(s}/L(s)] Lt). Inside the feedback toop both signals are related by AL(s) vals) = T45)-0g |, SEED ) (0) = Teo) [FE (6 This interpretation of model uncertainty will be explained in greater detait in Chaptets 2 and 4 From their definitions, it is easy to see that S(s) and T(s) satisfy the following equation [TOssO=T (17) where 1 is the identity matrix. This equation places a serious constraint when designing a controller that should guarantee stability and performance, as well as robustness to model uncertainty. According 10 equations (1.3) and (1.4), for disturbance rejection at the ‘output, to minimize the influence of reference and disturbance in the track- ing error, or the (infinitesimal) robustness problem of equation (1.5), the designer should minimize the “size” of S(s). Instead, if the objective is Te- jection of measurement noise at the output, at the tracking error signal e, or the robustness problem of equation (1.6). the “size” of T(s) should he ‘minimized, From the restriction in (1.7) this is not possible simultaneously. To solve this problem, first we should define an adequate “size” for a MIMO transfer matrix, particularly the sensitivity matrices, and by means of this measure we should try to solve the trade-off in (1.7). In later chapters we will define this measure and present different methodologies to solve the trade-off. Some of them are based on a more intuitive approach, as is the case of loop shaping (Chapter 4). The others—1a (Chapter 5) and Hx, optimal control (Chapter 6), y-synthesis (Chapter 7), and ¢! (Chapter 8) optimal contro!—solve the problem by means of an algorithmic procedure, The important point of this section is to realize that any well defined control problem should include implicitly in its requirements this trade-off, Otherwise the problem may be trivial a5 inthe following “textbook” prob Problem 11 Design a stabilizing controler for the system g(s) = (s+ As~2), which also rejets disturbing steps atthe output Clearly a controller k(s) = q/s with q > 2 solves the problem. This means that an infinite control action may be proposed being that q has no upper ‘bound. From the practical point of view, this is obviously unrealizable because of actuator bounds and high-frequency mode! uncertainty. The point is that ven in the formulation of a “theoretical” problem, the trade-off should be stated implicitly, (OBJECTIVES OF AN APPLEDTHEORY 17 1.4 OBJECTIVES OF AN APPLIED THEORY In the preceding sections we presented some ideas concerning model uncer: tainty in “real world” applications. the use of feedback, and the connections between the three hasie phases of a control problem. From them we can ex tract minimal requirements for a systems theory to be applicable in practical problems. In particular, we will concentrate on how Robust System Theory satisfies these requirements ++ Practical Assumptions. This requirement reflects the need to have sim: plifying assumptions that are related not only to the mathematical sim plicity but also to the applicability of the theory. In general, it means that the standing assumptions should be as unrestrictive as possible, in the sense of being more “realistic.” In the case of robust control it pertains to the connection between the three phases described in Section 1.1. As an example, assuming distur bbances as hounded energy signals is always more realistic than assuming they are steps. Also, from the practical point of view, instead of having 4 particular model as a description of the system, We can represent it by a family of models. This family is described by a nominal model and uses the information on the uncertainty bound, Strong Results. From the mathematical point of view, by strong we mean necessary and sufficient conditions. These conditions are therefore equiv- alent to a desired result, which can be verified (analysis) or implemented (synthesis), From a more practical point of view, we seek conditions that are neither unrealistic (too optimistic) nor conservative. In particular, the conditions we are interested in are on stability and performance of a certain control loop. As an example, the verification of ‘a necessary condition for stability is not enough to guarantee it; there fore itis optimistic. On the other hand, there could be no controller that satisfies a sufficient condition for achieving performance. That does not ‘mean there is no controller achieving that performance, therefore a con servative result. In both cases the results are inconclusive. Instead, we look for a strong mathematical condition (60 it can be tested), which is equivalent to a particular desired result (stability and/or performance) ‘Computability, The results obtained from the analysis and synthesis pro- cedures should have not only a mathematical representation but also ‘a computer representation. This means that the analysis conditions to be tested should be computable and the synthesis procedures imple mentable by means of an algorithmic code. ‘In particular, at this point we may discuss what “computable” means. exactly. We could argue that, although a code that implements an algo: rithm can be written, it may happen that either the code or the algorith are very inefficient, This could lead to a more general discussion between tractable and untractable computational problems ([226]) and efficiency 18 INTRODUCTION Of algorithms and codes. Therefore the main point could be clarified as follows. The computational implementation of a control problem (analysis or synthesis) should not be the bottleneck of the solution to the problem. In the following example we will see how robust system theory meets the above requirements. : Example Consider the closed-loop system of Figure 1.8, The system S is represented by a family of models of the form S= (Go(s)[1+W(s)4] | Al] <1} (8) ‘This description considers specifically the uncertainty between the system and the set of models by means of A and the weight W(s), The output data from an experiment performed on the actual system have been processed by 2 robust identification algorithm (Chapter 10). From this procedure we com- pute the nominal model Gy(s) and W(s). Therefore we have quantified the relation between the actual system, the nominal model, and the uncertainty. The objective is to achieve closed-loop stability for all members of the family, defined as robust stability. If from the experimental data we obtain a family of models that “covers” the actual plant, robust stability will guarantec stability of the actual closed-loop system with the controller connected to the real plant. As will be explained in Chapters 2 and 4, at equivalent condition for robust stability is the following: [ROG Us KEG WE], <1 9) By means of the 71x. optimal procedure (Chapter 6), a controller that achieves the minimum of the above analysis condition can be calculated, Both the analysis condition and the synthesis algorithm can be computed efficiently by means of commercial software ({20, 126, 262) From the point of view of the three phases described in Section 1.1, both the analysis and synthesis theoretical results provide implementable condi- tions to guarantee closed-loop stability in the final experimental verification phase, r+ w(s)a\ COBECTIVES OF THSBOOK 19 1.5 OBJECTIVES OF THIS BOOK 1.5.1 General Remarks ‘The main objective of this book is to introduce the techniques used in the theoretical phase of the control problem from the point of view of Robust System Theory. Therefore, according to the discussion in Section 1.1, the theoretical results will take into account, as well, the problems that arise in the other two phases. ‘The title of the book includes Robust System Theory because it not only considers robust control but also has a chapter devoted to robust identi cation. The words Theory and Applications appear in the title because not only theory is presented but also practical “real world” problems. This is consistent with the objective of bridging the gap between both. The book is oriented to students of engineering departments with a back ‘ground in classical and modern control. ‘The main analysis and design methods are complemented by an elabo- rated example, either in the same chapter or in Chapter 11, The latter has a ‘group of worked out applications that stress particular practical issues: nonlin- earities, unknown time delays, infinite-dimensional plants, and actuator and sensor limitations. The three appendices—Mathematical Background, System ‘Computations, and Riccati Equations—make the book as self-contained as possible 1.5.2 Scope Chapter 2 gives the first robustness concepts applied to SISO control systems described by Laplace transforms. It relates these new ideas to the classical control stability, performance, and robustness concepts. It is a simple intro- duction to the main issues of robust control, for students or engineers with a classical control background. ‘Chapter 3 presents the internal stabilization problem from two different points of view. The first i a state-space approach based on solving four special problems first introduced in {104}. This also serves as an introduction to the tools used in Chapters 5 and 6. The second is a more classical approach, based on the use of coprime factorizations. As a by-product we obtain a parametrization of all stabilizing controllers and achievable closed-loop maps. ‘These results are instrumental in solving the €" control problem addressed in Chapter 8, ‘Chapter 4 introduces the four basic problems in robust control: nominal and robust stability and nominal and robust performance for MIMO systems, in a linear fractional transformation setup. The objective is to present a prac- tical synthesis methodology that is not mathematically involved and could be helpful to the student /engineer who developed an “intuition” designing with classical control tools (Bode plots). The fact that this methodology for mult 20° INTRODUCTION variable systems is not a simple generalization of the procedure for SISO systems (Chapter 2) is also explained. This is the case of combined plant input uncertainty (uncertain actuator models) and high condition number plants, presented in its last section, Chapters 5 and 6 present the optimal 4) and 21x, control problems. These problems are solved first using a state-space approach using the tools intro dduced in Chapter 3. Alternatively, by using the Youla parametrization the problems are recast into an approximation form and solved using either pro- jections (in the 7p case) or Nehar's theorem (H.). Finally, in the Ma. ease ‘we present a third approach based on the bounded real lemma and the use of linear matrix inequalities. This approach allows us to easly address general (even singular) output feedback problems. Chapter 7 discusses more general types of uncertainties. In the case of structured dynamic, it defines the structured singular value (2) and presents its computation and use in both analysis and design. In the case of parametric uncertainty or mixed type uncertainties, it covers the analysis, the synthesis results being under development. This chapter also gives a good idea of the ‘open problems in this area, which are stil undergoing intense research Chapter 8 presents an overview of the relatively new subject of optimal € control theory. While the problem was formulated in 1986 [315], with the first solution methods for the simpler SISO case appearing around that time 180, 81. 83). this is presently a very active research area, where complete theory is just starting to emerge and where most of the results appear only in the specialized literature. The chapter starts by discussing the ¢! framework following the same induced-norm paradigm used in Chapter 2 to address stability and performance in the Has context. The € optimal problem is ‘motivated as the problem of rejecting bounded persistent disturbances (as ‘opposed to bounded energy signals in the A framework) or the problem of achieving stability against linear time varying model uncertainty. For sim- plicity most of the chapter is devoted to SISO systems, showing that in this simpler case the problem can be recast, through the use of duality as a finite- dimensional linear programming problem. Since this theory may result in very large order controllers, the chapter contains some very recent results on synthesizing suboptimal controllers using an LMI-based approach as an alter- native to the usually high-order optimal controllers. Finally, tis shown that in the continuous time case * optimal controllers are infnite-dimensional and the issue of fnite-dimensional approximations is addressed. Chapter 9 has a threefold objective. First, it gives a nice geometrical i terpretation ofthe structural properties of the state-space models in terms of principal components and directions, as related to controllability and observ ability. Second, this same interpretation is useful when evaluating numerical issues in state-space descriptions: For example, how “near” is the model from uncontrollability? This is useful for the student to distinguish between the “theoretical” model and the “numerical” model implemented in a com puter, This is also related to the “measure” given by the Hankel singular OBJECTIVES OF THSBOOK — 21 Table 1.1. Road map for self-study, by topic Robust Indentification ‘Stabilization Robust Analysis Robust Synthesis! 2 3 3 A 9 3 6 6 8 10 4 naz 7 n values. Finally, the third objective is to provide an introduction to balanced realizations as a part of model order reduction. A classical model reduction algorithm is presented. This material is closely related to the one presented in the next chapter. ‘Chapter 10 is a tutorial on a very recent (circa 1990) research subject that ‘connects the robust control methodology with the experimental modeling of systems. It presents the conceptual problem of worst-case deterministic identification but also details some of the main algorithmic procedures. Both time- and frequency-based experiments are considered, and a flexible beam example illustrates the use of some of these procedures along the chapter. It also provides a wide range of bibliographic references for the researcher. Chapter 11 gives a good idea of the issues that appear in practical prob- lems, It presents the implications of applying robust idemtiication and control procedures to several practical application problems, 1.5.3. How to Use This Book ‘This book is intended to be used either as a reference for control engineers or as a graduate level textbook. Table 1.1 gives an outline by topic of how to use the book for self-study. If used as a texbook, there are several al- tematives depending on the background of the students and the amount of twowerester courte Second Course Preteq: First Course 7 e 9 10 in * Optional ebapter 22 mtRoouction time available. At the simplest level it could be used as a one-semester, be- ginning graduate level course serving as an introduction to robust control and requiring as a prerequisite a standard course on linear systems theory. This one-semester course would cover Chapters 2-5, portions of Chapters 6 and 7, and some of the examples in Chapter 11. Since Chapter 8 is largely independent, it could replace Chapter 5 if the students have already taken an optimal control course. As a follow-up to this course, a second advanced graduate level course would cover Chapters 7-11. These options are outlined. in Table 1.2, SISO SYSTEMS 2.1 INTRODUCTION In this chapter we address the issues of nominal and robust stability and per- formance probleme in single-input single-output (SISO) systems, represented either by a transfer function or by state-space equations. The analysis is per- formed proceeding from stability of the nominal model of the plant to the final objective of robust control: robust performance. ‘As mentioned in the introductory chapter, robustness measures are related to the type of uncertainty considered. The classical uncertainty description for SISO models, gain and phase margins, are presented first, along with an analysis of their limitations. To overcome these limitations, a more general robustness analysis based on more realistic descriptions is presented. Per- formance is based on rejection or tracking of sets of disturbances, noises, or reference signals, which lead to conditions on the Toop sensitivity func- tions. The above conditions for both performance and stability will affect the selection of the loop transfer function as mentioned in Chapter 1 For the sake of simplicity and clarity, in this chapter the definitions of uncertainty and performance are restricted to SISO systems. Generalization to MIMO systems is postponed until Chapter 4, where we will also intro- duce additional issues that do not arise in the simpler SISO case. Following the standard notation, we will use lowercase to denote SISO transfer func: tions and uppercase for MIMO systems. Throughout the chapter we will use packed notation to represent state-space realizations, that is, worncer-avrerve | Ate] 23 24 sisosystEMs 2.2. WELL POSEDNESS Consider the feedback loop of Figure 2.1, where the input signals [uy(s),us(s)}. the outputs [y1(s),y2(5)}, and the errors fe1(3),ex(8)] satisfy the following equations: Jils) = als)er(s); yals) = K(ser(s)s es) =ui(s)—yas) A) ex(s) = uals) + y1(5) @2) Next, we define well posedness of a feedback loop, a standing assumption that should be made in any control problem. This avoids trivial situations in the loop, as in the case where g(s) = —-'(s) Definition 2.1 The feedback loop of Figure 2.1 is said to be well posed if and only if all possible transfer functions between the inputs ui(s) and wn(s) and a Outputs ex(s) e3)_yul6), and y2{5) exist and are proper. In other words, we can say that a feedback system is well posed when for well defined inputs, we obtain well defined outputs. In fact, it can be proved (Problem 1) that it is sufficient to verify well defined outputs [ey(s),€2(s)] or Lvi(s),y2(8))- It should be clear that well posedness of the feedback loop is a mathematical prerequisite for any control problem in which stability and performance are to be defined. It is also clear that in practical situations, feedback loops are always well posed. Before proving an equivalent condition for well posedness, we state in compact form the relationship between the inputs u; and errors e; for the feedback system of Figure 2.1 [28] = [eto “P] [88] ro [28 @3) wig als) 7 h Hs) ak Figure 2.1. Feedback loop model fo evatuate well posedness and (nominal internal ssobilty WELLPOSEONESS 25, ‘Therefore to have well defined outputs e\(s) and ¢2(s), we need the inverse of (5) to exist and to be proper. Otherwise, either the outputs will not be well defined oF could be nonproper. We will illustrate this by means of a simple example, Example 2.1 Consider the following model of a plant and a controller, con nected as in the feedback loop of Figure 2.1 as _ +7 2 = Gey KO Gay ea) The transfer function between tz and e3 is <1 _ (842s +5) Bacar! = ST es) whichis improper. Therefore the loop is not well posed. Note that, in practice, the model of the plant is stricily proper. In that case, iti easy 10 See Tha the above warsfér famction Would fave been proper Next, we state necessary and sufficient conditions for well posedness. As- sume that the plant and controller have the following state-space realizations: as) = [2 A 26) an Then we have the following result. Lemma 2.1 The following conditions are all equivalent: 1, The system in Figure 2.1 is well posed. 2. P(s) has a proper inverse. 3. 14+8(s)K(3) has a proper inverse. 4 Proof. All transfer functions between 1(s) and ¢)(s),f,)= 1,2 in (23) are well defined if and only ifthe inverse of P(s) is proper and exists for almost alls. The later and the fact that to guarantee well posediness it is enough that these four transfer functions are well defined (Problem 1) prove the equivalence between items 1 and 2. It is clear that if both the plant and the controller are proper, P(s) will have a proper inverse if and only if its determinant, (1+ g(3)k(8), has a proper inverse; which proves 2 and 3 are 26 sIsosySTEMS equivalent. For properness, (1 + g(s)k(s)]! should exist for almost all s and, in particular, for s ~+ co. Equivalently, ‘im 1 + g(s)k(s) 40 28) ‘The state-space representation of this limit is 1+ DpDg #0, which proves 3 and 4 are equivalent, The above arguments stilT hold in the case where we consider yy and y2 as the outputs of the closed loop (Problem 1), a 2.3 NOMINAL INTERNAL STABILITY The classical concept of stability of a linear time invariant system, described either as a transfer function or by state-space equations, between a given input-output pair of signals, is well known. Therefore we will not repeat here these definitions already stated by classical and modern control theories, Nevertheless, we will point out the idea of internal stability of a feedback loop, which was not stated explicitly. To this end we will clearly distinguish the stability from an input-output point of view from the stability of a feedback loop. Consider again the feedback loop of Figure 2.1. Its not difficult to see that the stability of a certain input-output pair (all the poles of the corresponding, transfer function in the open left-half complex plane C_) does not guarantee that all input-output pairs will be stable (in the same sense), This is illustrated by the following simple example. Example 22 Consider the following system and controler: (+1) (s-1) ene *= wt) oD connected as in Figure 2.1. The transfer function from the input u; to the ouput 1 is given by Ty, =1/(s+4), which is stable in the usual sense. However, the transfer function between the input uy and the output yi Ty, = (5+ 1)/(6~1)(s +4), which is obviously unstable. As we will see next, this is caused by the cancellation of the unstable plant pole at s = 1 by a zero of the controller. ‘This example shows that there is a difference between the stability of a certain system, considered as a mapping between its input and output,! which we define as input-output stability, and stability of a feedback loop, Which will be defined next. In the latter, we must guarantee that all possible input-output pairs be stable, which leads to the concept of internal stability "Even in the MIMO case with several inputs and outputs. NOMINALINTERNALSTABIITY 27 Definition 2.2. The feedback loop of Figure 2.1 is internally stable if and only if all wansfer funciions obtained from all input-oulpua pairs have ther poles nC. (input-ouput stable). Itis easy to prove (Problem 2) that to verify internal stability i is sufficient to check the input-output stability of the four transfer functions between the inputs [us(s),u2(5)] and the outputs fe:(s),e2(s)]- This is summarized in the following lemma. Lemma 22 Let ow = [245+] 9) bbe a minimal state-space realization of the closed-loop system of Figure 2.1, between the inputs |us(s),ux(s)] and outputs {ey(s),ex(s)] (or [pi(s).yal)]) Then the following conditions are equivalent: The feedback loop in Figure 2.1 is internally stable The eigenvalues of Ac are in the open left-half complex plane C The four transfer functions obtained between inputs |us(s),43(8)] and outputs {ey(s),€2(s)] (or lyi(s),y2(8))) have their poles in C_; that is, Ci(s) is input” output stable, Proof, 1 <> 3isleft as an exercise (Problem 2).2 = 3 follows immediately from Definition 2.2. Finally, ifthe four transfer functions are stable, then all the eigenvalues of A. are in C_. since minimality of (2.9) precludes pole-zero cancellations. This shows that 3 = 2. a Is important to note that the concept of internal stability, although not explicitly stated, has existed among practicing engineers for many years. A. proof of this is the basic controller design rule of no unstable pole/non- Minimum phase zero cancellaion Between plant and coniroller should be allowed. A misleading explanation for this rule is that, numerically, its im- Ossble to cancel exaetly a zero and a pole (why is it misleading?). According to the definition of internal stability, we now know that the argument behind this “practical” rule is the following: with unstable pole-zero cancellations, there will always be an unstable input-output pair in the loop (and hence ‘an internal signal could grow unbounded, damaging the system). This will be proved next. Lemma 23 The feedback loop in Figure 2.1 is internally stable if and only if (1+ e(s)k(3)|" is stable and there are no right-half plane (RHP) pole-zero cancellations between the plant and the controller. 28 sisosystems Proof, First factorize both plant and controller into coprime numerator and denominator polynomials (which do not share common roots), as follows: =) mute) als) a6)" (s) = als) (2.10) We assume that estoy = tesla) en) ing(odne(s) + de(s)dg(6) is stable. If (s) and de(s) share common roots in the RHP, these will also be roots of msimu(s) + dy(s)d,la)-The same occurs if ny(s) and ds) share contin roots in the RHP. Therefore in the first case, the transfer function y(s)ru(s) ie(s)ru(s) + da(s)dg(s) is unstable, because the RHP roots of the denominator will not be can- celed-by-the numerator [this is precluded by the coprimeness of the pairs (ro(s) (3) and ((5) dys). In the second case the transfer function [1+ g(s)k(s)}"1 (6) will be unstable for the same reason. Note that the com plementary sensitivity (1 + (s)k(s)}-"g(s)K(s) remains always stable, with or without RHP cancellations ifthe sensitivity is stable. On the other hand, if there are no RHP pole-zero cancellations, all these transfer functions are stable. As these are all possible transfer functions, the loop is internally stable, which concludes the proof. o H(s) [1+ g(R(S)™ (212) Note also that, according to the definition, any stable pole-zero cancel lation in the loop will not modify the internal stability. However, from the point of view of performance, the cancellation could be performed over a stable although undesirable mode of the system. This means that a certain input-output pair of Figure 2.1 will have “poor” performance. This can be understood better by means of an example. Example 2.3 Consider the following system and controller: (1 = s)(s? +0.2s +1) (s+ Ips +5) 89) = Cerner aes) MO = 3670241) connected as in Figure 2.1. The design objectives are to internally stabilize the Toop and to track with yy @ step input injected at uz with not more than 25% ‘overshoot. The controler selected meets the performance specification, since the closed loop is internally stable and y,(s}/us(s) = (I~ $)/(s +1), which thas an adequate time response. However the transfer function y(s) uals) ROBUSTSTABILITY 29 (5+ 3)(s-45)/(s + 1s? +025 +1) presents undesirable oscillations due to the lightly damped poles at s = —O.1¥), This situation can be avoided by simply extending the definition of inter- nal stability to encompass more general regions, rather than just the (open) leftchalf complex plane. Therefore, according to the desired location of the closed-loop poles, an adequate region in C_ can be defined, and no pole-zero cancellation between plant and controller should be performed outside this region. 2.4 ROBUST STABILITY As mentioned in Chapter 1, a given robustness margin is related to a specific type of model uncertainty. Therefore by first evaluating the stability robust- ness margins used in classical control theory we can analize their limitations and propose more realistic and general uncertainty descriptions. 2.4, Phase and Gain Margins The phase dy and gain margins gq are used as robustness measures to evalu- ‘Toosely speaking, {he abilty OF syslem To Tespond adequately, n terms of performance and stability, even when the open-loop model é(s) = (s)k(s) used in the design does not exactly match the physical system, due to the existence of uncer- tainty Both measures can be clearly interpreted in terms of the Nyquist plot and the Nyquist stability criteria, In this plot, dy and gm Fepresent the “distance” in angle and absolute value, respectively, to the critical point z= —1, as shown in Figure 22. ‘The usual practice of evaluating both margins separately assumes implic- ily that both types of uncertainty (phase and gain) act on the loop one at a time. As a consequence, these margins are effective as analysis tools only wen the model ofthe plat as ber phase or ginancereiny Care must be taken When using ¢, and gq for stability robustness analysis of SISO closed-loop systems, because they do not guarantee robust stability for the more realistic situation where both phase and gain are alfected by uncertainty. Here robust stability means the stabil sible models that are described by the combination of the nominal model 4go(s) and the uncertainty 5= ce, where cand @ are uncertain values Bontained inside the intervals Te = [émnsCmar] and Ly = [4mie: Prue» FESPEC tively (Figure 2.3). This set of plant models can be defined 38 the family of models F © {g(s):8(s) = go(s)[ee"*), C1. & € Ly}. We will clarify these concepts by means ofthe following examples. 30 siSOSYSTEMS. Figure 2.2. Phase and goin margins. Example 24 Consider the Nyquist plot of Figure 2.4, in which both $m and Bm have adequate values. Nevertheless, with small simultaneous perturbations in the phase and gain of the oop, the plot will encircle the critical point z = 1. Therefore, in this case, the minimum distance from the Nyquist plot to the critical point (the minimum return difference) gives a much better assessment of the robust stability properties of the system than dy OF Bm Example 25 Consider the Nyquist plot of Figure 2.5. In this case, the gain ‘margin gm —» 02, although the phase margin is very small. If we know that the only source of uncertainty in the loop is in the gain, a small b, poses no threat to stability. Therefore, in this case, the controller designed guarantees robust stability Figure 2.3. Phase and goin margins retated to the uncertainty in the model ROBUST STABILITY 31 Im(:) Figure 2.4, Simutianeous uncertainty n both gain and phase, Im(s) Re(s) Figure 2.5, Uncorointy inthe opentoop goin, 32. SISO SvSTEMS From both examples we conclude the following: ‘+ A given stability margin is a good indicator of the robust stability prop. Of gain and phase margins in the analysis of control systems is limited to the cases where there is @ priori knowledge that the ut s either on (that isnot ously. ‘© The uncertain system can be analyzed as a family of models that can be described mathematically by a nominal model and a bounded uncer- tainty set. In the Nyquist plot this corresponds to a nominal plot and an uncertainty region surrounding it «If there is simultaneous uncertainty in phase and gain, itis important to analyze the type of uncertainty considered, because it will determine the form of the set of Nyquist plots in the complex plane. «© The stability margin can be computed as the minimum distance between the family of models and the critical point in the Nyquist plot. Therefore it is directly related with the region describing the model set in C and as a consequence, with the type of uncertainty. This can be expressed as tollows: ‘Type of uncertainty | ~ Phase and gain margins are special measures that are useful only when either the uncertainty is “angular” or “modular,” respectively. They can be considered as special cases of a more general (and realistic) uncer- tainty description that will be presented next. 24.2. Global Dynamic Uncertainty Mote tals descriptions of mode! uncertainty have been developed, lead ing to controller designs tat perform better in practice, One seh Sesrip Mon the global dynamic unctiang. The ame arses snce thin uncertainty is related «9 the uncertainty athe system dynamics and covers globally the compete model ofthe plant. This ype of uncertainy description can be wed ‘when the order of the differential equations describing the plant is unknown ? This situation arises in practice when certain physical phenomena (last fenibliy) are not included inthe plant deseription or when the physical description of the ssem isto complicated to Tead to Wecable problems 2A ciferent branch of control theory sso dealing with uncertain systems adaptive contol IW has as one of ts base assumptions thatthe order ofthe equations describing the system is nova. Today, thee i an important effort geared toward relaxing this assumption ROBUSTSTABLTY 33 In many cases it describes situations where a constraint in the computation time (real-time control ofa fast process) imposes a limit on the order of the ‘mathematical models that can be handled. As a consequence, these models are approximated by a set of equations of lower order, and the neglected dynamics are lumped into the uncertainty Global dynamic uncertainty can also be used to describe lin errors, where a nonlinear system is linearized around a nominal point, Although the resulting model is valid only in a neighborhood of the ‘nominal operating point, the nonlinear effects outside this neighborhood can be bounded and interpreted as dynamic uncertainty. Typical examples of this situation appear in the modeling of aircrafts, which are described by a general nonlinear mathematical model. The latter is linearized at different altitudes, velocities, and angles of attack, yielding a set of linear models, each ‘one valid at a particular operating point. An uncertainty bound “covering” the linearization errors in a neighborhood of each linearization point can be interpreted as dynamic uncertainty Finally, many systems may be described by partial differential equations or time delayed equations, both called infinite dimensional systems. In these cases, an approximation ([24,25, 79, 134, 169)) of the general equation should bbe made so that a finite-dimensional model is obtained. The approximation error can be interpreted as dynamic uncertainty because the approximation process eliminates all the high-order dynamics of the general mathematical Iodel. In particular, uncertainty inthe time delay can be modeled as dynamic uncertainty ((268)) in a very simple way. Examples of the latter and of the ‘dynamic uncertainty of a flexible structure due to its approximation error are presented in Chapters 10 and 11 ‘The approach adopted by robust control theory is to describe a physical “system by means of an uncertain model. The latter is defined as a set of models and described in terms of a nominal plant together with bounded "uncertainty. In classical and modem control theories the controller is designed using the nominal model of the plant. Instead, robust control theory attempts to design a single controller guaranteeing that certain properties such as stability and a given level of performance are achieved for all members of the set of models. In this case, these properties are said to be robust. In the {following subsections, different sets or families of models will be described. Multiplicative Dynamic Uncertainty An uncertain model with multi- plicative dynamic uncertainty is shown in Figure 2.6. It describes a physical system as a set G of mathematical models, as follows: G = felsh:als) = gu(s){1+3W6)], BEC, I< 1} 13) The set G is the family of models and is characterized by a nominal plant (8), fixed weighting function W(s), and a class of bounded uncertainty 5 Te nominal model g,(s) corresponds to the case where there is no uncer tainty, that i, 8 — 0. Without loss of generality, the hound on the uncertainty 34° ssosysteMs [Family of Moses ii + 5W(s)}-+ 96(s) | SEY Figure 2.6, Dynamic mutipticative uncertain 6.can be taken to be one, because any other bound can be absorbed into the weight W(s), “The weighting function W/(s) represents the “dynamics” of the uncertainty, or in other words its “frequency distribution.” A graphical interpretation of this uncertainty weight can be scen in Figure 2.7. In this plot we can observe a \ypical weighting function, with 20% uncertainty at low frequencies and 100% at higher ones. A complete lack of knowledge of the system (100% error) ‘means in practice that there is no phase information (ie., the plant transfer function g(s) can have either a positive or negative real part). Thus, in order to guarantee stability of the closed-loop system, the controller must render the nominal loop function go(ja)k(jw) small enough at frequencies above , to guarantee that the Nyquist plot does not encifcle the critical point, regardless of the phase of g(jm)k (jon) This point illustrates an important limitation for robust stabilization that should be taken into account in any controller design, The situation where the uncertainty is small at low frequencies and in- creases at higher frequencies (although not nearly’as dranvatically-as in the example discussed above) is the usual case arising in practice. For instance, for mechanical systems a better model can be obtained to describe slow Figure 2.7. Frequency distibution of uncerionty, ROBUSTSTABLITY 38, mechanical resonances than faster ones (involving mechanical Nexibility) This situation can also be observed in models of electrical cicuits. For appli- cations in lower frequency bands, the model is simpler and usually provides a ‘good description of the physical realty. To deseribe the response of the circuit at higher frequencies, other physical phenomena should be included in the ‘model (parasitic capacitances and inductances) increasing its complexity. Al- tematively, these additional dynamics can be treated as dynamic uncertainty. The weighting function W(s) can be obtained explicitly in many applications {see Chapter 11) or by using the techniques known as robust identification, which will be described in Chapter 10. An important each mem ber of unste SraTHeat based on the Noquist graphical criteria (whichis a necessary and sufficient condition for stability of a single model) as follows. Suppose that any two members of a family have a different number of RHP poles, then their Nyquist plots will encircle the critical point a different number of times. [As a consequence, the region between both plots will include the critical point z = —1. For the family of models considered, the set of Nyquist plots of this family “covers” this region and also includes z = 1. On the other hand, a model that goes from (closed-loop) stability to instability should have its Nyquist plot go across the critical point, because it changes the number of cncirclements. Thus it will be difficult to detect if an element of the family is (closed-loop) unstable since the whole set of models already includes the point z = —1. Hence we cannot derive equivalent conditions for (closed-loop) stability of all members, based only on their frequeney response (Nyquist plot). For this reason, to guarantee that the number of RHP poles of all members of @ remains constant, we will assume, without loss of generality, that W(s) is stable (why?) and that no unstable pole-zero cancellations take place when forming gq(s)|1 + 8W(s)). Perturbations & satisfying this later condition will be termed allowable. ‘A more general uncertainty description can be obtained by replacing 8 € C by a teal rational transfer function 4(s), analytic in the RHP and such that 1A(s)Ilw = sup,, |Aw)] <1. As before, we require that no unstable pole— zero cancellations occur when forming g.(s)[1 + A(s)W(s)]. As will be seen next, robust stability of this general model set just described and the one in (2.13) yield the same condition. Thus, for simplicity, in most of this chap- ter we adopt the description in (2.13), which ass “frequency” distribution of uncertainty to W(s), and the bound to 8 € C. Finally, note that, in order for the problem to make sense, we should always assume bounded uncertainty, otherwise there is no possible design that can stabilize the whole set simultaneously (why?), A condition guarantecing stability of all elements of the family G, that is, robust stability of gu(8), is derived next. 36 sISOSvSTEMS ‘Theorem 2.1 Assume the nominal model go(s) is (internally) stabilized by a controller kis) (Figure 28). Then all members of the family G will Be (inter: nally) stabilized by the same controller if and only if the following condition is satisfied: | ITEWEIe *supl7Gw) Ge) <1 | easy Le with T(s) © go(s)k(s)[1 + go(s)k(s)]~" the complementary sensitivity function of the closed-loop system. Proof. From Lemma 23, it follows that: 1. {1 +g(s)k(s)}" stable for all members g(s) € ¢. 2. No RHP pole-zero cancellation between k(s) and any member g(s) € 6. are equivalent to robust (internal) stability. First we prove that (2.14) is equivalent to condition 1 above, when the nominal loop is internally table [1 +g(s)k(s)]* stable Vals) eG (2.15) > MH gG)KON+WE)] ZO MBl 1+ gols)k(s) # go(s)k(s)W(s)8 —¥/8J<1,seC, (2.17) gt g BoD K(S)WE) # T+ go(SK(G) o (enon <1 vec. 219) = IT(s)W(s)I] S10 (2.20) Pej bead | Ka) = Vidl 418 €,|8|<1 can vay be found sch that equality holds The “only i partis vial For equation (220) we have sed the maximum modulus eo. Fem which states that since PEpW (a) analyca Ce BORE oT eR of the nominal closed Toop-and W (3), IT @)W (3). achie the bounty of Cy thatthe yo an Neve we complete hep tiene (=) It is clear that robust (internal) stability implies nominal (internal) stability and equation (2.14) (equivalent to condition 1), (on) On te other hand t's ear that nominal (tera) ability and equation (220) imply condition 1. To complete the proof we need ts how tht eunition 2 abo bolas Flom Lemna 23 we hove that nena ser nal sbity implies no RHP pole-2ere cancelation between) and Ke) Hence for any RHP pol of the open logp pv € Gs Ti.) = Listy?) Use (218), whichis equivalent to Sondtioo TRS have Mos) < ana hence 13 Gp) £0.09] 1 Becane paring, this imbisno RAP poe zero cancelation between Ka) tnd any merther glahe 6 (eomion 9) : ols)k(s)W(s) T+ go(5)K(5) itsmaximunron f of the Using Figure 2.9, we can interpret condition (2.20) graphically, in terms of the family of Nyquist plots corresponding to the set of loops. First observe that (2.20) is equivalent to [1+ go(@)K(s)| 2 lso(S)K(G)WOS)], Y= jo 221) For a given frequency w, the locus of all points z(sw) = go(Jo)k(Jo) + BoJo)kyw)Ww)6, 6 €C, [8] <1 isa disk Dw), centered at g,(Jo)k(Jo) with radius r= |go(jo)k(Jo)W a)]. Since [1 + go(7w)k(w)| is the distance between the critical point and the point of the nominal Nyquist plot corre- sponding to the frequency jw, it follows that condition (2.20) is equivalent to requiring that, for each frequency «, the uncertainty disk D(a) excludes the critical point z = ~1. Therefore robust stability for SISO systems can be checked graphically by drawing the envelope of all Nyquist plots formed by the set of circles centered at the nominal plot, with radi [@,(/w)k(j«)W (a) and checking whether or not this envelope encloses the critical point 2 = —1 In the MIMO case, although an equivalent condition can also be obtained (Gee Chapter 4), therg is no such graphical interpretation The frequency at which |W(jw.)] = 1 corresponds to the limit for which there is 100% uncertainty in the model of the plant. Therefore another in terpretation of condition (2.14) is that frequency w, is the upper limit for 38 siso systems Figure 2.9. Sot of Nyquist plots of the fomily of models. the bandwidth of the complementary sensitivity function T's). At frequen- cies above w., the gain |7(ju)| should be less than one, to achieve robust stability Gain and phase uncertainty in the loop can be considered as special cases of the dynamic multiplicative description, W(s) = 1 and 8 € R, [8| <1 4 t0_gain uncertainty. On the other hand W(s) = }€ C6 € 0,2) = |3| <1 corresponds to phase uncertainty. Th both cases there is an exira structure on the uncertainty 6, which should be a consequence of the extra « priori information on model uncertainty that is, ‘only phase or gain is uncertain. For this reason, the condition of Theorem 2.1 could be conservative if applied to this case and if the a prior information is correct. Otherwise itis more prudent to consider the unstructured type of ‘uncertainty described by the set &. Several examples will serve o illustrate the above concepts Example 26 Consider the following nominal model go(s) and a second pos sible plant g,(s): 300 Gree H+ TO} oe HE g165) = (2.22) Bul3) = (2.23) ROBUSTSTABLTY 39 These two plants can be described using the following family, characterized by ‘multiplicative uncertainty represented by a weight W(s) and a bound on 5: fs(sh-e6s)~gols)[1+ 3WG)}, [81<1} 24) wo) = a (225) Its easily verified that gy corresponds 10 8 = 1. Note that the set G also includes many other plant models, For instance, for 6 = we obain as +50 ale) = ett) 2.26) GF IG ++ 100) Therefore if we only need to include the models g,(s) and go(s) or even a finite set, the description could be unnecessarily conservative. As a consequence, any design that applies 10 all members of G could be conservative as well (what will be sacrificed?). In the next example we illustrate the use of multiplicative uncertainty to describe uncertainty in the location of the zeros of a transfer function. Example 2.7 Consider a set of models with uncertainty in the high-frequency dynamics (w > 120 rad/s) of the numerator polynomial. o= (ewe ca} any fol) ag for 5=0 (228) CHD ‘This uncertainty can be represented as multiplicative dynamic uncertainty using the weighting function W(s) =0.2{1 + (s/20)]. Its graphical representation is similar to the one of Figure 2.7. The system is known with a 20% relative error up (0 10 rad/s. Above 100 rad/s the model has no information on the system ‘that may be useful for control design (|W (ja)| > 1). According to condition (2.20), this frequency is the upper limit for the bandwidth of the complementary sensitivity function T(s), to achieve robust stability. Other Uncertainty Descriptions Additional types of dynamic uncer- tainty descriptions, similar to (2.13), can be formulated as follows: Gu = (Gols) + 8Wa(s), 15] <1} (2.29) Gy 3 {eols) {1 + BWa(s)11, [3] <1} (230) G, = {eols) [+ 8MiCs)Ro(8)1"", [91< 1} 31) 40 siso systems ‘These descriptions are defined as additive, quotient, and inverse dynamic uncertainties, respectively, The difference between the description of uncer- tainty By ary of the above depends on the specific application under consid- eration, Take, for example, the case where a high-order model (even infinite dimensional) g(s) must be approximated by a lower-order one g,(s). The approximation error can be considered as additive dynamic uncertainty. The set of models “centered” at the nominal g,(s), which includes the high-order ‘one g(s), can be defined as follows: Go S {arls) + 8Wals), [5] <1} (232) ‘The weight W,(s) can be obtained from the frequency responses of the approximation error (or its upper bound) and g(s). A systematic way to ‘obtain this uncertainty description and its corresponding uncertainty weight will be presented in Chapter 10. In the same way that multiplicative uncertainty seems to be the “natural” structure to describe uncertainty in the zero locations, quotient dynamic un- certainty is the least conservative way to describe open-loop pole uncertainty. ‘This will be illustrated in the following example. Example 2.8 Consider the following set of mathematical models, which can be described as a family with quotient dynamic uncertainty 1 % {moar wali} 233) & feo(s)[L+ Wolo)", lol <1} (234) which leads to the following nominal model and uncertaimy weight 1 3 me MWe (235) Bo(5) If, instead, the same set would have been described by a family of multi plicative dynamic uncertain models, there would have been an unnecessarily large number of models, and as a consequence the description would have been conservative. If we take the same nominal model, the new uncertainty weight Wm(s) and bound ym should be defined such that {BmWnls), Bm EC, [Bul < Yh (236) “covers” the following set: 38 - {etry 86 ici} (237) ROBUSTSTABILITY 41 ‘obtained by considering (g(s) ~ go(5))/Bo(s). (5) € Gy- The weight Wo(s) = 3/{8 +3) and bound y= 1 achieve the above condition but include in the new set of models Gy second-order systems, which were not included in Gy We conclude that, depending on the specific application, the description of model uncertainty should be selected appropriately among the above de- scriptions. If ail the uncertainty is included in one particular description, itis called global dynamic uncertainty. Otherwise, in many cases, we may adopt ‘wo or more uncertainty descriptions to better “fit” the family of models, By this we mean that, taking advantage of the information on the “structure” of, the uncertainty in the model, we select the smallest possible set. This is im- portant because, in general, a robust controller should stabilize and provide performance for all possible plants in the set of models; therefore the larger this set, the more conservative the design. The type of uncertainty description that combines information of different portions of the plant and adopts different uncertainties for each part, that is, different 3's, is called a structured dynamic uncertainty description. For ‘example, consider the following set of models with both uncertain zero and uncertain pole locations: §+2(1+0.36,) o = {STS lal <, Fe (zn) (238) 2 fo cll +6) 7 ; 8 {eon lal<1, ra} (239) In Chapter 7, robustness analysis and design tools for this more general ype of uncertainty description willbe presented. Following the proof of Theorem?.1, necessary and sufficient conditions for the robust stabity of the model sets described in Tale 21 can be obtained (Problem 3), where S(s) © [1+ gu(9)(6)-" is the sensitivity function of the nominal closed-loop system. We can conclude from ths table that different stability conditions correspond to different uncertainty descnptions. This coherent with the fat stated before, that sabity margin ae related to 8 Table 2.1 Global dynamic uncerlainty descriptions and the conesponding robust sabilty condition (6 <1) @ Robust Stability £a(9)[1¢ Wal) WAT, <1 ss) + 3446) HWEEe)kLS) SIL < 1 sols) [1 + 8W,(5))! WW, (s)S(syIL. <1 [1+ BWSR) IMC SE), = 1 42 ssosvstems specific type of uncertainty description, This was the case with the classical phase and gain margins. The robust stability conditions in Table 2.1 can be interpreted also as stability margins. Take the case of multiplicative uncertainty and suppose we bound it as Bo] < yn, With y%, positive real number. Then its easy to prove that the equivalent robust stability condition for this new set of models is the following: IPOW Oh <5 240) The above means that [|7(5)W (s)|q, can be interpreted as a stability margin for the set of models with multipicative dynamic uncertainty in the same way gn i the tabiliy margin fora set of models with gun uncertainty. TBs isso Because the infinity norm ofthis particular function gives the measure of how much dynamic multiplicative uncertainty can be tolerated before there exists at least one model inthe set that is closed oop unstable Te follows thatthe controller that yields the largest stabity margin (with respect to multiplicative uncertainty) can be found by solving the following ‘Optimization problem: waa, yp ITO¥ Oe Gan ‘The above is an Mx. optimal control, which will be analyzed in detail in Chapter 6, 2.5 NOMINAL PERFORMANCE Performance will be defined in the frequency domain in terms of the sen: sitivity functions, according to the compromises stated in Chapter 1. First, the classical and modern control concept of performance as rejection (track ing) of known disturbances and/or noise (references) is stated. Based on the practical limitations of this approach, we state a more realistic definition of performance as rejection (tracking) of sets of disturbances and/or noise (references) 2.5.1, Known Disturbance/Noise/Reference In classical control theory, a measure of performance of a closed-loop system is based ofits ability to reject (in steady state) known disturbances, noise, ‘oF measureitient errors, which may appear at differeni parts of thé loop, that is, sensors, actuators, or outputs. In this context, we define a known signal as one having a particular form (sine, step, impulse, etc.), which is known beforehand by the designer, athough it snot known at what particular NOMINAL PERFORMANCE 43, time it will disturb the system. Similarly, a measure of performance may be posed as the ability of the loop to follow known reference signals with zet0 steady-state error. Itis a well knowin fact that rejection of exogenous disturbances and tracking references are equivalent problems. If we consider input signals that are polynomial in time, then we can clusify stable systems according to the highest degree of the input that can be tracked with finite steady-state error. An internally stable system is said to be of type n if it can track (or reject) a polynomial input of degree up t0 7: with zero steady state error I€18 easy to show that a type m FDLTI system must satisty the condition lim, .s"a(s)K(s) = K 4 0, For instance, in order to track a step input (or reject a step disturbance), the system must be at least of type I (ie, lim, 0 59(s)k(s) © K, #0). Similarly, to track a ramp reference (or reject a ramp disturbance), the system must be at least of type 2 (lim,_.9 s°g(s)k(s) K, #0), In modern control theory, the same concept of performance has been used: rejection of known disturbances in the state of output of the system or tracking a known reference signal. In addition, the concept of optimal control is introduced. In this framework the contrller seeks to minimize a certain functional, which quantifies the compromise between tracking (rejec- tion, stabilization) speed and the control signal energy. I is possible to state this optimization problem in both a deterministic and a stochastic context Inthe latter, the perturbing signat 1s considered as a stochastie process of Anown covariance. In either case, the optimal controller that minimizes the functional will remain optimal as long as the disturbance or the reference signal matches exactly the assumptions made in the design process. If this hypothesis is not met, there is no guarantee of optimal, or even “good.” _ disturbance rejection or tracking (100). We may conclude that, in both cases, the fact that the designer should have a clear knowledge of the disturbance, noise, or reference to be tracked poses a serious constraint. If he/she designs for the rejection of a particular sig- ral, but it tums out that the disturbing signal is different, performance could seriously be degraded. In the same case, if the design has been performed by an optimization method, the performance no longer will be optimal. Con- ceptually, this is the same limitation we found for closed-loop stability. If we stabilize a nominal model, there is no guarantee that the closed loop will re- main stable if the mode! is changed. For these reasons, classical and modern control techniques require a clear knowledge of the nominal plant model and the external signals for correct (or optimal) stabilization and performance, In practice, the designer does not have such detailed information. For this reason, the assumptions on the system and external signals should be relaxed. For the case of uncertainty in the system description, the approach pursued in the last section was to consider a family of models, rather than a single nom- inal model. Similarly, in the case of performance, a family of disturbances, noise, or references will be considered. An example of a practical situation 44 sso systems Be He) Fale) Figure 2.10. Feedback loop with output disturbance, where this relaxed assumption is necessary is the case of robotic manipula- tors, when different trajectories need to be tracked in situations where the ‘workspace is changing with time. Also, we could have sinusoidal disturbances ‘with frequencies contained in a certain uncertainty band, or even less infor- ‘mation, as the case of energy-bounded signals. In the next section we present the performance analysis when defined with respect to a family of external signals, acting on the nominal model of the system. 2.5.2 Bounded Disturbances at the Output Consider the block diagram of Figure 2.10, with output y(s) = S(s)d(s) = [1 +.g(s)k(s)}"" d(s). If the performance objective is to minimize the effect Of the disturbance d(s) at the output y(s), a stabilizing controller that makes Sis) as “small” as possible needs to be designed. The trivial solution to this problem is k(s) + 00, s0 that S(s) — 0, but there is no guarantee of nominal stability. Furthermore, as mentioned in the introductory chapter, @ fondamental constraint of any feedback loop is the equation S(s) + T(s) = 1 ‘Therefore in this case we would have T(s) —+ 1 at all frequencies. According to condition (2.14) ofthe last section, this implies that robust stability.can be ‘guaranteed only if the uncertainty in the plant model is less than 100% error at all frequen : = a ‘AS a consequence, we should seek a controller that makes $(s) “small” in a certain frequency range of interest for the particular application on hapd (Gee Figure 2.11). This could be the case. when output disturbance rejection needs to be achieved at low frequencies, as in the position control of Targe mechanical systems with very low natural resonance frequencies. Further- 15()| Figure 2.11. Desired frequency distibution fr 5(s), NOMINALPERFORMANCE 45 more, there could be situations where there is a certain a priori knowledge of the frequency contents of the disturbances. The frequency range where ‘S(s) needs to be “small” or the frequency content of the disturbances can be represented by the weights W,(s) and Wg(s), respectively. These weights are dynamic systems, which represent the knowledge of the frequency bands of interest for performance and the disturbance frequency content. Both can be incorporated into the block diagram of the system as in Figure 2.12. ‘Our next step is to define precisely the meaning of “small” S(s). The size of a transfer function, in particular the sensitivity function, will be measured by its induced norm. For a general LTI operator A: x A.+x the induced norm is defined as follows: Alles = sup A sala (2.42) In particular, the operation « will be equivalent to convolution, if both input ‘and output are time signals, or product, in the case where they are repre- sented by their Laplace transforms. Therefore the induced norm will depend. entirely on the norms we adopt to measure both the input and the output. For convenience, in most of this text, we will adopt the energy of the signals as a ‘measure of their size. In this case, according to Parseval’s theorem, for both representations of a signal, time or frequency, their norm will be the same, that is, x(0)l2 = l/x(s)[i. With this in mind, we define nominal performance as follows: Definition 23 Nominal performance of the feedback loop of Figure 2.12 is achieved if and only if the weighted output remains bounded by unity, that is, HW, (s)y(s)lla <4, for all disturbances in the set {4 € £2, |\d2 <1}, and for all other external inputs to the system equal t0 zero. Without loss of generality, we have considered unitary bounds in both ceases, because any other bound can be absorbed into the weights W,(s) and Wa(s), by linearity, Wale) Wale) = Figure 2.12. Augmented feedbock loop with performance weights 46 ssosvstems Checking nominal performance by using Definition 23 directly requires a search over all bounded £; disturbances, which is clearly not possible. Hence, as in the case of robust stability, it is necessary to find a computationally verifiably equivalent condition. This condition will be obtained by exploiting the relationship between the £2 to £) induced norm of a LTI system g and its frequency response, Theorem 2.2 The feedback system of Figure 2.12 achieves nominal perform ance, as defined in Definition 2.3, if and only if (1W4(s)S(s)Wa(s)|],, © sup |W,Ga)Syo)Wa(yo)] <1 (2.43) Proof. The operator that maps the input signal d(s) to the weighted output W, (s)y(3) is W,(3)S(s)W,(s). The induced norm from £; to £3 isthe infinity norm (see Appendix A). Therefore are WWyylle = Wyler cS (2.44) => [Ws (5)S(s)Walslao <1 (245) Hence nominal performance is equivalent to equation (2.43), 0 Note that this condition is similar to the one for robust stability (2.14), although applied to a different transfer function. Nevertheless, there is an important difference between both conditions. The one for robust stability can be interpreted qualitatively. In other words, either all members of Tare closed-Toop stable, or there is at least one model that is not. There is 1o~ “gradual” boundary separafion-between the set that satisfies (2.14) and its complement. Instead, nominal performance can gradually be relaxed, since it is a quantitative property of the closed loop. Specifically, this means that” the upper bound of the weighted output to be bounded (or minimized) can bbe changed to a number y > 1. This new bound will “gradually” degrade the performance, This separation between qualitative and quantitative properties of systems is well known and has nothing to do with robustness but with the definitions of stability and performance. It has been pointed out, because it will be important in the next section, when considering robust performance. ‘As before, the optimal controller, in the sense of providing optimal dis- tubance rejection, can be found by solving an H.. optimal control problem, in this case of the form stl pain [MOSH]. (2.46) The solution to this type of problem will be discussed in detail in Chapter 6. NOMINALPERFORMANCE 47 Note that the only assumption on the disturbances is that they have bounded energy, There is no hypothesis whatsoever on the shape of these signals, as in the usual setting of classical and modern control. The weight- ing functions W(s) and W,(s) only provide an a priori knowledge of the frequency content of these signals but do not determine their shape. ‘A graphical interpretation of the nominal performance condition can be ‘obtained by means of a Nyquist plot (see Figure 2.13). To this end, define W(s) 2 Wy(s)- Wy(s) and note that (2.43) is equivalent to [W(yo)| <|1+gol@)k(Yo)|, Yo (247) Consider, for each frequency J, @ disk Diya) centered at 2 = ~1, with radius 1 =|W(w)]. Then (2.47) can be interpreted graphically as nominal perform: ance being achieved if and only if, for every frequency jw, the disk D(jw) |W), vo @si) where W(s) © Wy(s) Wa(s). Hence, in this case, nominal performance is achieved if and only if, at each frequency, the disk centered at z= —1 with radius r = |W (Jw)| does not include the inverse Nyquist plot of go(je)k (Jw) Finally, we like to remark that, due to the fact that the bounds for in- put and output signals have been stated in terms of their energy, all the Conditions for nominal performance are expressed in terms of the infinity norm. An equivalent formulation can be obtained when the input signals are taken as bounded persistant signals. In this case, the natural norm to measure the “size” of the inputs and the outputs is the time-domain £,, norm (i.e. the sup,). Since the La. {0 Lay induced norm is the L; operator norm (see ‘Appendix A), in this case the conditions for nominal performance will have the form of a weighted £; norm. The problem of optimizing this norm (C, ‘optimal control theory) will be addressed in Chapter 8. 2.6 ROBUST PERFORMANCE ‘The final goal of robust control is to achieve the performance requirement on all members of the family of models (j.e., robust performance), with a single controller. Clearly, robust stability and nominal performance are pre: requisites for achieving robust performance (why?). Next, we will establish a necesary and sufficient condition for robust performance, by making use of the conditions for nominal performance and robust stability. To this end we consider the case of performance defined as disturbance rejection at the out- ut for a family of models described by dynamic multiplicative uncertainty as represented in Figure 2.16, Definition2.6 The feedback loop of Figure 2.16 achieves robust performance if and only if |W,(s)y(s)ll2 <1, for all possible disturbances in the set {d € £| ld < 1}, for all inputs to the system equal to zero, and for all models in the set G = {g: g(s) = [1+ Wa(s)8]go(s), 8€C, [6] <1), Recall that the necessary and sufficient conditions for nominal perform- ance (2.43) and (2.14) and robust stability are parominal , | <> Wals)S(ly <1 (252) REY | <> W6(S)TUIl <1 253) ROBUST PERFORMANCE 81 $ System Figure 2.16. Disturbance ejection atthe output for afar of model wih multiplicative uncertainty, Robust performance is equivalent to condition (2.43), but with the nomi- nal model go(s) replaced by the family of models g = go [1+ Ws(s)8], 5 € , [3] < 1, We use this fact to obtain a necessary and sufficient condition for robust performance, ‘Thearem 2.3 A necessary and sufficient condition for robust performance of the family of models in Figure 2.16 is |1W2ue)Sue + )7Ge)Ws(r0)| |] <1 254) Proof. Replacing the nominal model ga(s) in (2.43) by the family of models £2(8) yields the following condition for robust performance: retort +e(s) wa(s)8))"I <1, |aj |Walyw)S(y0) 1+ Tyo)We(y0)8"| <1, [8)< 1 > [Wa(y0)SU«)] < 1+ TUW)We0)8], 18] <1, Yo => |Wa(soo)S{ya)| S1—|T(yo)We(yo)], Yo (2.55) = W0)S(9)| +I TGOWsK0)1 <1, Yo 256) > | mar0)S0)) + 17Gw)WoUe) | <2 esr) where fw) = ge(w)k(w). To prove the “if” part of (2.55), we consider 5 = 6. such that T(jw)We(sw)8, is real and negative, The “only if” partis immediate, 0 52 sisosYsTEMS Figure 2.17. Nyquis pot for bust performance interpretation, As before, a graphical interpretation of the robust performance condition can be obtained by means of the Nyquist plot of Figure 2.17. Note that (2.54) is equivalent to 11+ €(40)| ~ (Ws(@)€(G0)] + Waal) 20 (258) From the figure we see the robust performance requirement combines both the graphical conditions for robust stability of Figure 29 and nominal per formance of Figure 2.13. Robust performance is equivalent to the disk cen- tered at 2) = ~1 with radius ; = [W,(jo)] and the disk centered at z> = (ju) with radius ro = |Ws(jw)E(jw)] being disjoint. Clearly, this is more restrictive than achieving the robust stability and nominal performance conditions sep- arately. "Nevertheless, in the SISO case, by scaling the nominal performance and/or robust stability conditions conveniently, we can guarantee robust perform: ance. From the practical point of view, there is an important difference when scaling one or the other. Inthe first case, as we mentioned before, the condi- tion can be relaxed by changing the upper bound of the output to be hounded. In the second case, condition (2.14) is exclusive, therefore either we guaran. tee internal closed-loop stability for all members of the set or not. Sealing can be accomplished by means of a new “control-oriented” identification of the plant, yielding a new nominal model and a “smaller” uncertainty region, This isnot a trivial task, as will be seen in Chapter 10. ROBUST PERFORMANCE 53. ‘Therefore the usual practice is to scale performance, once condition (2.14) is achieved. This is performed by means of a parameter 7, as follows [71¥0e)sy0))+1W0e)T eI] <1 ye Wt 259) [As we have seen before, nominal performance and robust stability can be verified by computing the infinity norm of an appropriately weighted closed- loop transfer function. Synthesis of an optimal controller, which maximizes cither nominal performance or robust stability, isa wel established procedure, known as Hq, optimal control (see Chapter 6). However, the condition for robust performance (2.54) cannot be expressed in terms of a closed-loop weighted infinity norm. Rather, a new measure j., the structured singular value, must be used. This measure will be defined in Chapter 7, where we will indicate how to compute it, a procedure known as jeanalysis. The synthesis of controliers that optimize this measure, synthesis, will also be explained in the same chapter. In the next paragraph we briefly describe the relation between robust performance and structured uncertainty. Consider the block diagram in Figure 2.18, where two different types of uncertainty have been considered, Ws, and W,d2. The set of models involv- ing both types of uncertainties is the following % {so(s)[1 + Wals)d,] "(1 + Wols)&], 18 <1, i= 1,2} 2.60) Next, we obtain a condition for the robust stability of this family of mod- cls. The proof follows the one in Theorem 2.1, replacing the multiplicative uncertainty set by (2.60). Lemma 2.4 Robust performance of the family of models in Figure 2.16 is Wb Ws] Figure 2.18. Robust performance interpreted as robust stabity of sets with structured, uncertainty, 54 siso systems equivalent to the robust stability of the family of models described by equation (2.60), shown in Figure 2.18. Proof, Robust stability of all members of the set G, is equivalent to 1+ €Uw) (1 + Weo)6i) [1 + Wee)8]' #0, Wor, VB) <1 || <1 = 1+ e(w) + £40)We(j0)6) + Wa0)6: #0, Yo, VIB <1 [6] <1 > [1+ G0) ~ [eGo)We0)] ~ Walo)| 20, Vo => |TQw)We(w)| + |Sw)Wew)| <1, Yo = |lroweol+iseymaef, <1 where £(s) = g(s)k(s) is the feedback loop. The above equation is identical to (254) o It follows that robust performance can alwayss be cast into robust stability of a set of models with structured uncertainty ((102]). Another interpretation is the following. The problem of rejecting a family of disturbances weighted by W,(s) can be incorporated as 4 new type of model uncertainty, as in the block diagram of Figure 2.18, As mentioned before, the measure over the closed-loop system corre- sponding to this problem is not a usual system norm. The computation of this new measure should take into account the structural information on the uncertainty of the system (see Chapter 7). The transformation from robust performance to structured robust stability can be represented by means of linear fractional transformations (LFT) as in Figure 2.19 (see Appendix B). In this case & is called the performance block. [5 3] P(e) Figure 2.19. Structured robus! stably wih extra performance Block hi EXIENSIONTO MIMO SYSTEMS 55 2.7. EXTENSION TO MIMO SYSTEMS. Concerning the extension of the above concepts to MIMO systems, we can state the following: « There is an equivalent concept for pole-zero cancellations, which is not straightforward as in SISO systems. © The infinity norm is also used to evaluate nominal performance and robust stability, by means of the singular values. «# The robust stability equivalent condition changes when the uncertainty in the model appears at the input (e.g., actuator uncertainty) or at the ‘output (¢., sensor uncertainty). The condition number of the nominal ‘model plays an important role in the former case. « In general, scaling by y the nominal performance condition, as a method to achieve robust performance, is not as effective as in the SISO case. ‘© The robust performance condition cannot be computed directly as in equation (2.54). Instead, a minimization problem should be solved. We conclude that the tools for the analysis of stability and performance robustness in MIMO systems are not a simple extension of the ones used in SISO systems, although we can find many similarities, ‘A usual procedure to operate with multivariable systems is to apply the tools for SISO models to each input-output pair of the MIMO system. We will show, by means of a simple example, that, in general, this is not correct. Example 2.9 In this example we show how the analysis by “loop at a time” ‘may lead to wrong results in the case of a MIMO control problem. Consider the following nominal plant and controller (see Figure 2.20): 5 ~ 2641) Gs) = 61) G+NG+2) 1 # 3 Theerain System Figure 2.20. Feedback loop of uncertain system 56 SISO SvsTEMS l : K(s) 542 Zs+Div+2) (2.62) 1 For the above design, the nominal loop L(s) = G(s)K(s) and the complemen: tary sensitivity functions are uy tf, ° tft a 1 HOS ml ‘| eo) The “loop at a time” analysis is based on using the tools for SISO systems ‘on the problem. To this end, each loop is analyzed by means of the gain and phase margins, while the other loop is closed. This can be seen in Figure 2.21, where the first loop has been opened, while the second one remains closed. The equations in this case are the following: mao ae a es) Similarly, losing the first loop and evaluating the second one, we obtain y2 = ~us/s. As a consequence of this analysis, thee is an infinite gain margin and 90° phase margin in both loops. This analysis is based on the fact that the uncertainties in the system do not appear simultaneously in both channels. "Next, we show that with small uncertain gains in both channels, the closed-loop system can be unstable To this end, take the multiplicative uncertainty of Figure 2.20. If we con- sider a perturbed model (I + 8)G(s) instead of the nominal, the characteristic polynomial will be modified. Take, for example, an uncertainty of the form 4 A= G a (2.66) Figure 2.21. “Loop ata ime” anaiysis, PROBLEMS 87 The new characteristic polynomial of the closed-loop system is the following: s+ 5(24 8, +) + (148 +8) =0 (261) nis easy to verify that the following uncertainty destabilizes the closed loop: [29] as The perturbed model produces a closed-loop pole ats = 0. The “size” ofthe perturbation can be measured by means of a matrix norm. In particular, we may use the maximum singular value, which results in &(S.) = v2/2. We have considered muliplicariveuncertanty, therefore we might aswell ry to use he tools described in his chapter for SISO systems. To this end we mea sure the stability margin using the equivalent condition (2.14) (the weight in this case is W(s) = 1) If we use asthe spatial norm || for the complementary Sensitivity the maximum singular value, we obtain [Tih = 1618. Therefore the “smatles” (measured with the same norm) perturbation that destabilzes the loop ist inverse hati, (Sx) ~ 0618. Ths minimum perturbation can be found i we alow & tobe a fll complex matrix. Nevertheless, by using the special structure ofthe uncertainty, i can be shown tha the smallest destabilz ing perturbation i indeed (268), which has norm a(S.) ~ 2/2. This resul i nat as opimisicas we could have expected from the “loop tate” analysis thor i tas pessimistic as the analysis neglecting the special structure of. This is reasonable because we are now allowing a much ricer (and more practical) Set of uncertainties, which affect the plant model. We will sce in Chapter 4 that the equivalent condition (2.14) for robust stability sil holds for MIMO systems, by replacing the magnitude by a matrix 2.8 PROBLEMS 1. Prove that well posedness can be verified by considering only outputs e and ¢) of Figure 2.1, 2. Prove that to verify internal stability, it is enough to check the input- ‘output stability of the four transfer functions Fis) :u;—+e) with i,j = ea 3. Prove all entries of Table 2.1 4, Represent as a linear fractional transformation diagram ((M(s), A] struc: ture, see Appendix B), specifying M(s) and A, the following set of models: (a) A model g(s) with a positive feedback of 6. 58 sisosystems (b) A model g(s) multiplied at the input and output by blocks (1 + 4) and (1 + 53), respectively. (€) A model g(s) multiplied at the input and output by blocks (1 + 6;) and (1 + 8:)-*, respectively In all cases there is a controller k(s) that stabilizes the nominal model. Is it possible to find necessary and sufficient conditions for robust stability in all the cases? 5. Consider the following family of models: : neg quo l 1 batet\t < spe yle {ake wai 0. Then a simple root-locus argu: ‘ment shows that in this case the set of stabilizing controllers of this form is not even connected, 2. Even in cases where the set of stabilizing controllers can be charac- terized, the resulting optimization problem is hard to solve since the objective function (3.1) is a nonconvex function of the optimization variable [in this case K(s)]. | 4 : a a | OSes ee eae, eal Axis Figure 3.2. Root locus iustrating « nonconnacted set of stabilizing controliors. WELL POSEDNESS AND INTERNALSTABIITY 63 Alternatively, we could attempt to use tools from optimal control theory such as LOR. However, tis will require recasting the performance specifi cations in terms of Q and R, the state and control weighting matrices used in the LOR performance index. Since there are no rules that will allow for exactly expressing specifications of the form (3.1) in terms of Q and R, in practice this leads to a “box-with-a-crank” approach, where the weighting matrices are adjusted through multiple trial-and-error type iterations In order to avoid these difficulties, we will parametize all the controllers that stabilize a given plant in terms of a free parameter Q(s) (a stable proper transfer matt). This parametrization, orginally developed by Wouls ba oe workers in the continuous time case (323, 324 (se also [94] for an alternative derivation) and, independently, by Kucera [178] for discrete time systems, is usually known as the Youla or YBJK parametrization, although sometimes itis also denoted as the Q-parametrization. A similar parametrization for ‘multidimensional systems was developed by Guiver and Bose [140] ‘The main result ofthe chapter shows that all the closed-loop mappings are affine functions ofthe parameter Q(s). Thus any convex specification given in terms of a closed-loop transfer function translates into a convex specification in Q(s), and the resulting synthesis problem reduces to a convex (albeit infinite-dimensional) optimization problem [58]. We will use these results in later chapters to solve the 743, lax, and £" optimal control problems. The fist step in achieving the parametrization isto extend the concepts of well posedness and internal stability introduced in Chapter 2 to the general case of MIMO plants 3.2. WELL POSEDNESS AND INTERNAL STABILITY ‘While the concepts of well posedness and internal stability of MIMO systems are very similar to those introduced in Chapter 2 for the SISO case, there are nevertheless some differences, mostly related to the tools used to assess whether or not a given loop exhibits these properties. In this section we briefly reexamine these concepts and introduce the appropriate changes. 3.2.1, Well Posedness For multivariable systems, the definition of well posedness is the same as in the scalar case (see Definition 21). Thus Lemma 2.1 can still be used to assess whether or not the feedback loop is well posed, provided that the scalar 1 there is replaced by the identity matrix F and in item 4, different from zer0 (#0) is interpreted as nonsingular. Nevertheless in spite the similarity with the SISO case, the luck of well posedness in MIMO systems may not be detected easily, as illustrated by the following example 64 STABILIZATION Example 32 Consider the following plant model and controller: 32) ow) pale 1 rei The elements of the corresponding ouput sensitivity function S,(3) are all improper: (s#1) {s #1) [+ G)K()7 2 63) Go Se) xo =[4, 9] 6s the sensitivity function is not well defined, because + G(s)K(s) is singular for all's 3.2.2 Internal Stability ‘The same concept of internal stability presented in Definition 2.2, as well as Lemma 22, can be applied to the multivariable case. The loop intercon- nection shown in Figure 33 is internally stable if and only if all possible input-output transfer functions are stable. On the other hand, the general- ization of Lemma 2.3 is not straightforward. The difficulty here is that the concept of RHP pole-zero cancellation (as interpreted in SISO systems) in hg et as) w » K(®) anh, ws Figure 3.3. Loop interconnection to assess well posedness ond intornal stability, COFENLOOPSIABIE PLANTS 65 transfer matrices does not always compromise internal stability. In fact, entirely possible Tor a transfer matrix location of € (but not in the same input-output “channel"), as we illustrate with the following example. Example 33. Consider the plant +1 = 0 Gs) = s42 65) sel From the definitions of multivariable poles and zeros, we have that both are “simultaneously” located at 2; = py = —1 and 23 = py = -2. However, it can easily be verified that a realization containing the modes hy = ~1 and Ay = ~2 is both controllable and observable and therefore minimal. Moreover, a loop having G(s) as the closed-loop transfer function from (u; ug] to [ey €) is certainly internally sable. A generalization of Lemma 2.3 will be deferred until Section 3.7, where we introduce the concept of coprime factorizations. In the sequel, to develop 4 {state-space based) parametrization of all stabilizing controllers we oily need the following MIMO version of Lemma 22> cate [4h] 06 be a minimal state-space realization of the closed-loop system of Figure 3.3, between the inputs [ui(s),u2(s)} and outputs [ey(s),€2(8)} (or [ya(s),y2(6)]) Then the following conditions are equivalent: 1. The feedback loop is internally stable. 2. The eigenvalues of A. are in the open lefhalf complex plane C-. 3. The four transfer functions obtained between the ints js) 43(3)] and outputs |e;(s),e2(s)] (or yi (s),¥2(s)|) have their poles in C_ 4. The waner matic [2, “Ki inertbl in RP, Proof. The proof is virtually identical to the proof of Lemma 2.2, 0 3.3 OPEN-LOOP STABLE PLANTS In the case of open-loop stable plants, the set of all linear time invariant, finite-dimensional controllers that internally stabilize the plant has a simple expression, easily interpreted 66 STABIUZATION ‘Theorem 3.1 Assume that the interconnection shown in Figure 3.3 is well posed and G(s) € RM, Then the set of all FDLTI controllers that internally stabilize the loop is given by K(s) = Q(s)1- GOUT" 6 Qls) € RA, det[f — Gloo)Q(oo)] #0 Proof. (=>) Assume that the interconnection is well posed and that K(s) internally stabilizes G(s). We need to show that there exists Q(s) € RNs. such that K(s) = O(s) |! ~ G(s)Q(s)]-*. To this effect define as) 2 By assumption, Q(s) is well defined and in RM... Solving equation (3.8) for K yields (8) = + KG) 'K = K(+ GK) Ga) K(s) = Q(s)|f — G(s) Q(s)}"' G9) (<) To complete the proof we need to show that any controller of the form (3.7) internally stabilizes the loop. Recall that checking internal stability is equivalent to checking that the four transfer matrices obtained between the inputs {ui(s),ua(s)] and the outputs [y1(5), ya(s)] in Figure 3.3 are in RH. Straightforward calculations using (3.7) yield vn] 2 [F-GQG Ge]fu [e)- ("28° [a] om ‘The proof now follows immediately from the fact that by assumption both G(s) and Q(s) are in RMLs. 0 A simple physical interpretation of the parametrization (37) can be ob- tained by elementary block manipulation techniques as follows. First note ae) Gls) Figure 3.4. Reolling O(3)1- G00)" OPENLOOPSTABLE PLANTS 67 2. a || a te) FY ate) | @ © Figure 3.5. Physica interpretation of the paromettization for stable plants that Q(s) {7 ~ G(s)Q(s)|" can be realized with the block diagram shown in Figure 3.4. Replacing the block K(s) in Figure 3.1 with this realization yields the block diagram in Figure 3.5a, Finally elementary loop manipulations lead to the block diagram in Figure 3.5b. From this last block diagram we can see that, in the case of a stable plant, the parametrization (3.7) amounts to can: celling out the plant (by subtracting in the feedback signal) and replacing it with any arbitrary stable dynamics Q(s), ‘This interpretation also highlights the reason why this approach will fail tor the case of open-loop unstable plants. While ~G(3) cancels the unstable Als) “Fi Gals) Figure 3.6. Presiabiluing an openoop unstable plant, Gay denotes the prestabilzed plant (46a) "6. 68 STABILZATION dynamics G(s) in some of the paths (such as Ty), there are some others (©, Ty,4) Where this cancellation does not take place, yielding an unstable closed-loop transfer function. Thus a different approach is needed to deal with the general case where the plant is not necessarily open-loop stable. Note in passing that one may attempt to paramettize all the controllers that stabilize an open-loop unstable plant by first prestabilizing the plant and then applying the ra Figuré_36. However, it can be shown that while this approach generates stabilizing controllers, it does not generate all the controllers that stabilize the plant, unless the prestabilizing controller K,(s) is itself open-loop stable. 3.4 THE GENERAL CASE In this section we develop a parametrization ofall stabilizing controllers for the general case where the plant is not necessarily open-loop stable. The main result shows the existence of a separation-like structure, where all the stabilizing controllers have a familiar observer-based structure. The state- space approach that we use follows after the recent paper {185}, although the observer-based structure of all stabilizing controllers was already pointed out a5 early a5 1984 [103} In order to use the results of this section in later chapters, we consider the slightly mote general structure shown in Figure 3.7. Here the inputs to the plant have been partitioned into two groups: w, containing exogenous dissrbances and reference igmals anu the contol ingus Tikes, the ‘utputs have Beer partitioned inib z, containing outpuis subject to perform ance specifications, and y, containing the outputs available to the controller. Throughout this section we assume that the corresponding state-space real- w es) 2 K() Figure 3.7. the generalized pont THEGENERALCASE 69 ization of G(s) is given by sow A| BB v) Dn Dia Gan) yl G1] Dn Da yl Gl dn 0: where the pairs (A,B) and (C:,A) are stabilizable and detectable, respec- tively. Clearly, this assumption is required for the stabilization problem to be solvable. 3.4.1, Special Problems In the sequel we consider four special cases of the general structure (3.11), ‘commonly referred to as full information (FI), disturbance feedforward (DE), full control (FC), and output estimation desired parametrization for the general ouput feedback (OF) case will be obtained by combining these cases. Full information This problem corresponds to the case where the plant thas the following form: A | BB CG | Du Dy Ger=| oh | To) Te @.12) ol | ly lo Note that the outputs available to the controller include both the states of te plant and tie exogenous perturbation, that —— he plant and tife exogenous perturbation, _ fxn w= [30 hence the name. Rather than attempting to parametrize all the controllers that stabilize a given FI plant, in this subsection we obtain a weaker result, namely, a parametrization of all the possible control actions. Specifically, we define two controllers K and K’ to be equivalent (denoted K ~ K’) ifthe corresponding closed-loop matrices are identical, that K~K! = F(G,K)=FG.K) G3) ‘This equivalence relation induces a partition of all the stabilizing controllers into equivalence classes (an equivalence class is formed by all the controllers that generate the same control action). As we will see in the sequel, in order 0 obtain the set of all output feedback stabilizing controllers it suffices 10 obtain the set of all equivalence classes. 70 STABILIZATION “ Lemma 3.2_ Let F be a constant feedback matrix such that A + BpF is stable. Then the set of all stabilizing FI controllers! can be parametrized as Kei(s)=|F Q(s)], QE Re G4) Proof, (=) Simple algebra shows that closing the loop with the controller (4) yields vy sf). Te) = TE) +806) [C+D | Div | n= espe] ME No [ A+BF | By 2 Oe Dak T Diz which is stable since A + B2F and Q are stable. (<) Assume now that a FI stabilizing controller Kp) = [Ky Kz] is given, To complete the proof we need to show that there exists Q € RM, such that Krr~[F Q)-To this effect, make the change of variable v =u ~ Fx, and let Q denote the closed-loop transfer function from w to v obtained when closing the loop with Ky. Since Key stabilizes Gr, 2 € RHay. Moreover, we have that un Peso = ke[3] uo =1F ai[5] Hence Kr, ~|F Q) o Disturbance Feedforward In this case, the plant has the following state space realization: f A| BB Gor(s)=2| Gr [Dn Dia G17) ylal| ro In addition, we assume that A ~ ByC; is stableAs we show next, with this additional hypothesis the FT and DF problems are equivalent. Lemma 33. Consider Gop and assume that A B\Cz is stable. Then the following results hold: 1. The feedback matrix Kpr(s) internally stabilizes Gpr if and only if Kop [CxT-intemally: stabilizes Gey. In addition, Fy(Gor, Kor Fu(Gry Kor(C_ 1) "nthe sense of generating all posible contol action, as dscussed before THE GENERALCASE 71 2. Given any feedback system Kyy(s) that stabilizes Gy,, then controller Koe = Fi(Pop, Krs) internally stabilizes Gor, where G.18) Moreover, Fe(Ger, Kes) = Fe (Gor, Fe(Por. Kei) Proof. Result | follows immediately from the relationship between the FI and DF problems, illustrated in Figure 38. For result 2, note that in this case the state and output equations can be solved to obtain fe] in terms of EE |. yielding @.19) Denote by é the states of the “observer” Py and by w the estimated value of the disturbance. From Figure 3.9 we have that a state-space realization of oF) cle |Grr(s)|_ Kor [oe 7) Figure 3.8. Relationship between the Fl and DF probiems. 72 stanuzation Figure 3.9 Inlorconnocting Gor(s)ond Pos). the interconnection Gar—Por is given by A(t) = Ax(0) + Byw(t) + Byult) ee) = (A~ BCE) 2(0) = Crx()+ Durw(e) + Diault) 620) H(t) = x()—e@) (0) = Cree) + w(t) where ¢ 2 x ~ 2, Finally, eliminating the (stable) noncontrollable mode ¢(t) yields A| BB Du Da r}o0 0 ol 7 6 Thus F,(Gris Ker) = Fr [Fe(Gor.Por):Kei]. The desired result follows now from the fact that Fr (Fe(Gor,Por), Kei) = Fe(Gor,Fe(Por.Ken)] (see ‘Appendix B), o FuGor,Por) = = Gr 21) ‘This lemma shows that as long as the additional assumption concerning the stability of A ~ BC; holds, the FI and DF problems are equivalent. This ‘equivalence is exploited next to parametrize all the stabilizing controllers for the DF case. Note in passing that the proof of the lemma clearly shows the necessity of the additional assumption, since the matrix A — ByC> is precisely the error dynamics in (3.20). THEGENERALCASE 73 Combining Lemmas 32 and 33, we have that a class of stabilizing con- trollers for the DF problem is given by 3.22) where F is any matrix such that A + ByF is stable. Moreover, simple algebra shows that (3.22) can be rewritten nee Kor(s) = For, Q) OE RH» \..4¥** (3.23) A+ BAF ~BiC: | By By , Jor F oer *% (3.24) -G 70 © Indeed a stronger result holds: (3.23) parametrizes all the stabilizing DF controllers. This result is summarized in the following lemma, where Lemma 34 Assume that A ~ B,C3 is stable. Then (3.23) parametrizes all the controllers that internally stabilize Gr Proof, (=) The fact that Kpr(s) internally stabilizes Gpp follows immedi- ately from Lemmas 32 and 33. (4) Consider now a DF stabilizing controller Kpr. We need to show that Kpp can be written in the form (3.23) for some Q € RM. Let - A |B B Jor =| =F 0T ¥ 25) and define Q, = FeVne, Kor). Since Jor and Gpr share the same triple (A,Bz,C2) and since Kpr stabilizes Gpp, it follows that Qy € RH. More- over, by using the formulas for composition of LFTs given in Appendix B we have that FiUpr, Qo) = Fi [Jor, For, Kor)| = FeUimp, Kor), where A+ BrP ~ BC x _ =BiCe Jing = oT e ro A-BG ~BF tsk | 0? Aver hey - 0 =F why 0G Thus F)Upe, Qn) = FrUinp, Kor) = Kar o 74 STABIUZATION Full Control in this case the plant has the following realization: A {l 0 Gre=|C | Du [OTT 8.27) G | Dy (0 0 The name stems from the fact that the control input wo= [283] affects independently the states x and the performance outputs z. Rather than obtaining directly a parametrization of all stabilizing con- rollers for this case, which would essentially follow the same steps of the previous two cases, we will exploit these results by using the concept of duality. given a plant G(s), its algebraic dual is obtained by simply transpos- ing it. Consider now the interconnection of Figure 3.7 and the duals of both the controller and the plant. It can easily be checked that, (FAG, K)}! = Fx(GT,KT) ) (328) internally stabilizes a po at_K internally sta t G if. and only if its dual KT stabilizes G™. Thus a parametrization of all controllers that stabilize a given plant G can be obtained by simply transposing the corresponding parametrization for its dual G”. In particular, comparing the realizations (3.12) and (3.27) itis immediate that the FI and FC problems are algebraic duals. It follows that @ parametrization of all FC stabilizing controllers can ‘be obtained by simply transposing (3.14). This result is summarized in the following lemma, Lemma 35. Let L be a constant feedback matrix such that A + LCz is stable. FC controllers? can be parametrized as 329) Output Estimation Finally, we consider the case where the plant has the realization 830) Note that if B; = 0, then this reduces to the problem of estimating the output z from the noisy measurements y, hence the name. nthe same sense asin the F cae a at 4 epeyrate Bi A=B:C| 6 [1 -B,) y Por = Cr 0 a (3.31) a 4 , \\ THEGENERALCASE 75 It is easily seen that the DF and OE estimation problems are algebraic duals. Thus the relationship between OE and FC is similar to that of the DF and FI problems. In particular, the following properties hold, Lemma 3.6 Assume that A — B3C, is stable. Then: 1. The feedback matrix Koe(s) internally stabilizes Goe(s) if and only if [Jw meaty sbi Got Moen hse have that F: (re [5] Kor) 2. Given any feedback matrix Kc(s) that stabilizes Grc(s), then controller Fe(Poe, Kec) internally stabilizes Gor(s), where Fi (Gor, Kor) G 1 [0 0} Blas Hdy (Goes Fe(Por.Kro)] Glee] 3. The set ofall internally stabilizing OE controllers can be parametrized a8 Kor = For: Q), Q € RH, where aa A+LG~ BC | L_-B or oT 3.32) GQ 10 3.4.2 The Output Feedback Case ‘We return now to the problem of parametrizing all the controllers that in- ternally stabilize the plant (3.11). This parametrization will be obtained by combining some of the four special cases discussed before, Jor = Theorem 32 Consider the general output feedback problem for the plant (3.11). Let F and L be constant matrices such that A+ LC; and A+ ByF are stable, Then all the controllers that internally stabilize G are given by K()=FU,Q), QE RM, det[l + DO(o0)] #0 (3.33) A+B E+LG+LDuk | -L_By+LDy Js F oT (Cr+ Dak) 1 Dn where (334) 76 STARIUZATION Proof, For simplicity, we will assume for the time being that Dz> = 0. This, assumption will be removed at the end of the proof. Consider the state-space realization of G(s), which under this assumption will be denoted G(s): HO = Ax()+ Byw() + Boule) 2(0) = Cix(0) + Durw(0) + Drault) 3.35) Y(8) = Cax(0) + Dw Using the change of variable v= u— Fx, we have that (3.35) can be patti tioned into two subsystems as illustrated in Figure 3.10, where 5 [_AtBF | Bi Be oar a Br] Ca) By By Gump =| =F T 837) G | Duo and Since Gy is stable by choice of F, it follows that a controller K(s) internally stabilizes G(s) if and only ift stabilizes Gay. Moreover, Ging ha an op estimation structure, since A + BaF is stable by construction. Hence, from Lemma 36, we have that al the controllers that stabilize Ginp (and therefore K(s) Figure 3.10. Output feedback os « cascade of an OF and a stable stern THEGENERALCASE 77 J =| (338) oc | ‘To complete the proof we need to remove the assumption D22 = 0. This can be achieved through a standard loop shifting technique, consisting in using the change of variable pay-Daur This change of variable subtracts Dzau from the output of the plant and wraps a feedback Day around the controler, resulting in the struture shown in Figure 3.11, where G has Dz, =O. Hence the controller X is given by K = FeG,0), The original controller K can be recovered by wrapping the feedback ~Dzy around X as shown in Figure 3.12. Thus, inthe general OF case, we have that K = F,(J,Q), where simple algebra shows that [oe -L_By+LDn Je F 7 T “(C+ Dak) 1 -Dn 3.39) Finally, det(! + DzxO(ce)) #0 is a necessary and sulficient condition for the interconnection of Figure 312 to be well posed 0 Gs) G(s) * . ron Da kK) LAAT Ri) Figure 3.11. Loop shiting transtormation. 78 STABILIZATION 4) Da « iy Je) as) Figure 3.12. Undoing the loop shiting lranstormation. 3.5 CONTROLLER STRUCTURE AND SEPARATION PRINCIPLE Consider again the parametrization of all OF stabilizing controllers developed in last section, K(s) = FJ, Q). A block diagram of this parametrization is shown in Figure 3.13, Note that J has a very familiar structure. Indeed, if ‘we neglect for a moment the additional output and input associated with Q (e and v), then J is precisely a classical observer-based output feedback controller, having F and L as the controller and observer gains, respectively. «| ow T() K(s) ae) Figure 3.13, Porometerzation of all sabilaing contolers. CLOSED-LOOP MAPPINGS 79 Thus the parametrization (3.33) can be interpreted as an observer-based con- troller, where a nominal classical observer is modified so that it produces an additional error signal ¢£ y — yon. Al stabilizing controllers are obtained ty Slering this eror with arbitrary sable dynamics Q and injecting the resul- tng signalo = Qe back into the observer. whee it gets aed tothe contol Sipnal u-~ Pi Recall from standard linear systems theory that the estimation es —— Sece no feedback, Sine the nominal controller stabilizes the plant and since is sable in hindsight i is aot surprising that the family of augmented Sonvlirs also intrmall stabiizes the plan The surprising est is the fact that this modified observer paradigm generates all stabiling controls “the existence ofa seperation property now apparent All outpat feed er —“ Cee edback (or F) problem to obiain a matt F such that A + BP is Saupe eatmation (or PC] poem in sian Lach at CIpREMIGL A TLC are stable. All controllers ate obtained by com: ihe tofu Thee problems to form the modified observer J and conngcting arbitrary stable dynamics Q between ¢ and v. Alternatively, this err ‘was obtained essentially using state-feedback and Ginp corresponds 10 an Ne rpoee eyes 3.6 CLOSED-LOOP MAPPINGS Recall from the example in the introduction that one of the reasons for seek- ing a parametrization of all stabilizing controllers is to recast performance specifications given in terms of closed-loop transfer functions into @ form having a simpler dependence on the free parameter (in this case Q). As we show next, an additional advantage of the parametrization is that the resulting closed-loop mapping are affine in Q. To see this, consider again the block di- gram of Figure 3.13, We have that Tew = Fe [G, FiJ,Q)] = Fe(T,Q), where renos[h th] Since, as pointed out before, ¢ is uncontrollable from v, Tap =0. Thus Te reduces to Tew = Tu + TQ ~ T2Q)"Tn = Ta + T2Ta. (3.40) ‘an affine function of Q. Simple algebra using the explicit expressions given in Appendix B yields the following state-space realization for T: 80 STABILZATION Bh ea . BiskDy 0 ~ Di Da 3.7 A COPRIME FACTORIZATION APPROACH In this section we provide an alternative derivation of the Youla paramettiza- tion based on the use of coprime factorizations [94, 324, 314). Besides its historical importance, this approach allows for a simple derivation of some well known results on the simultaneous stabilization of multiple plants and the related problem of strong stabilization, that is, stabilization with stable controllers. 3.7.1 Coprime Factorizations Definition 3.1 Two transfer matrices M, and Ny in RMxs are right coprime over Re if there exist two matrices X, € RH and ¥, € RHx such that the following Bezout identity holds: X,M, + Y,)Ny (3.42) Similary, Mj, N; € RHao are said to be left coprime over RU. if there exist matrices X; € RHae and Y; € RH, such that MiXi+ MY, =1 G43) Remark From these definitions it can easily be shown that if M, and N; have a common right factor U, € RM (ie, Mp = MU, and N,=NU,), then Uz! € Rte. A similar property holds for let-coprime matrices, Thus coprimeness among elements of the ting of stable, proper transfer matrices is.a generalization of the SISO concept of no commrion RHP zeros between numerator aid denominator (|72, 314). Every proper, real rational transfer matrix G(s) can be factored as G(s) = N(s)M(s)"! (NO) G4) where N,M\N, and M are in Ra, and where N and M ate tight coprime and NV and M are left coprime. These factorizations are known as a right- coprime_and_a lefi-coprime factorization of Gis), respectively. In the next Jemma we provide a state-space based algorithm to obtain such factorizations starting from a stabilizable and detectable realization of G(s), ACOPRIMEFACIORUANONAPPROACH 8 Lemma 37° Ler Ge = | (345) be a stabilizable and detectable realization of G(s) and F and L any matrices such that A+ BF and A+LC are stable, Define Wit stone [AeEeb ese 1) aay Then (3:46) and (3:47) are right- and left-coprime factorizaions of G(s). respectively. [Me] = (3.46) NOs) and Proof. Since A+ BF and A + LC are Hurwitz by construction, it follows that all the transfer matrices defined in (3.45) and (3.47) are in RHa.. Moreover, it is easily seen that M and Af are not singular. Using the formulas for transfer ‘matrix inversion and multiplication given in Appendix B, we have that eo a ao and ctl Y A+SBF BF| B i N(S\M“4s) = - A “C¥DF DFT D AsBE 0] 0 oO A B | =G(s) C¥DF-=cTD_| where the lat equality was obtained by using the similarity transformation Smee ef sre vy a : a9) Pete tee we{ ee ———— itary ican easly be shown that M-"(0)N() ~ G(s) To show thatthe pats M,N and AV are right and left coprime, respectively, define (X46) Y= eS G50) 82 STABILZATION and yun) Anette oe 0 es) WO" | cA) | 1 Cleary, these four matrices are in RH. Using the formulas for transfer ‘matrix multiplication yields (after some tedious algebra and eliminating un- ‘observable and uncontrollable modes) yu Xs) ¥e(s)] [M(s)_ ~¥ils) ro 8 HSI 28 1-[63] om RGIS Ra Hime, 0 mA Remark _A factorization of the form (3.52) is called a double-coprime factor- ization of the transfer matrix G(s). The state-space formulas presented here for computing double-coprime factorizations are due to Nett et al. (215}. Using these concepts we are ready now to present a MIMO extension of Lemma 23. Indes adnate foloush opine hactaatene Ope NON M~"(s)N(s), K(s) = Nu(s)M,'(s) = Mz'(s)Ng(s). Assuming well posedness, 1, The feedback loop is internally stable. MN, 2 [Me Gis] eimentiin mt My Ny 4 2 [Ae] mente cquivalent to the tanser mates [/) —*] being invertible in RPL. Expl [677-2 21 ay Using the (right) coprimeness of M,N and Mj, Nj, it can be shown that the 53) matrices Moo a [MM Oa ee ees are right coprime. i ACOPRIME FACTORIZATION APPROACH 83. Since M, My € RMx. and the matrices in the right-hand side of (3.53) only share common elements invertible in RM... it follows that: E K MN nes (3.54) | J erm “The equivalence of items 1 and 3 can be established proceeding in the same way using the let-coprime factorizations of G and K by noting that [6 K] erm — [% By] cen. cu tad (355) o Next, we show that given a stabilizable plant G(s) there exists an in- ternally stabilizing controller K,(3) = U,Vz! = Vz"U, such that the factors Up, Vo, Uo, and V, ae precisely the components of a double-coprime factor ization of G. This result constitutes the first step in the alternative derivation of the Youla paramettization. Lemma38__[f G(s) is stabilizable, and detectable then there existright-coprime {factors Up, Ve and lefi-coprime factors Ua, Vo in Rx. such that Ve Oo] [M Uv.) [1 0 -N M|[N ¥ or Moreover, the controller Kals) = Us(s\W=H(s) = Vs"s)to(s) internally stabilizes the loop. (836) Proof. The first part of the proof follows immediately from Lemma 3.7 by selecting Vp = X,, Uo = Y;, Vo = X;, and Uy = ¥;. To show that the con- troller K internally stabilizes the loop, note that with this choice of factors we have that : M -U]'_[V_ Up [x ve] =[5s ii)er Internal stability ofthe loop follows now from Theorem 33. ° G57) Lemma 38 provides a simple way of finding a stabilizing controller if a double-coprime factorization of the plant is known. As we show next, this ‘controller is exactly the familiar observer-based controller obtained by cas- cading an observer with gain L with a state-feedback gain F. A state-space realization of such an observer-based controller is given by (see Figure 3.14) a = At+ Bu+L iy -(-y)) = Ck+ Du (358) ua Fe 84 STABILZATION a) = y He Ys Bod 2 FOO A+ BP , \c-+ pr| re xe) o ow) Hrht s Bods = - i L A+ BF i en , A+ BF : c+ DF | ) 0) - . Figure 3.14, The controller of lemma 38 as an observer-based controller: (a) standard realization ond (b) factored as UaV! [ACOPRIME FACTORIZATION APPROACH 85 Define ¢ © j + and replace w by the last equation: ASBF | L w= Atle = Use 639) AsBF_[L (C: TIS = Vo(sle Ts = Us(s}e = Uals)Volsy!y (G.60) which coincides with the controller obtained in Lemma 38, Next, we use the results of Theorem 33 and Lemma 38 to obtain an alternative derivation of the Youla parametrization, Theorem 3.4 The set of all controllers that internally stabilize the plant G(s) can be paramernzed as fg _ayg)(1ANO K() 1) The fact that U-U =F implies that |Uazlho £1 (since Uss is a compression of U and matrix compressions do not increase the norm). Since ||Q].o <1 then a simple small gain argument shows that (I - U@)-! € RH. Since both U,Q € RH, well posedness and internal stability follow immediately. To show that l|Tvullm <1, note that, since UF is inner, then WEIE-16) Hel $< wi ~~ HON IMI (3.78) Since || Qlao < 1. it follows that [lz < [lw], or equivalently [Tell <1 (1 = 2) Assume now that the interconnection is internally stable and such that [[Tz. fs <1. To show that Qj <1 assume to the contrary that there exists a frequency «, and a constant vector f such that 77) bh be which implies OCian)s 2 Ill (3.79) 90 staaiuzarion (here we use the frequency-domain characterization of the Hq, norm, [lla = sup, 7 [Ou)). Define the input w = U5'U ~ UnO)s. Then the control action and outputs corresponding to this input are precisely & = Q9 and 2 = Uys + UizQ9. Since U is inner, we have that NEUE + PIE = IB + 5 (3.80) which combined with (3.79) yields W218 2 Wa G81) However, this last equation contradicts the assumption that Tz lsc < 1. Thus F[QUe)] <1, Vw <> ||Olla < 1. Finally, to show that Q € Ria, start by considering a (right) coprime factorization of Q.Q = NM~', N,M € RX. Since by assumption the interconnection is internally stable, we have that QUI ~ UnQy! = N(M = UpN)! € Rts. (3.82) follows then that the winding number of det(M — Uzp/V) as s traverses the Nyquist contour is 0, From internal stability, combined with the facts that [lUzalle: $1 and [JQ||x < 1. we have that det [7 — a /zs(j0)QUia)] #0 for all w and all « € [0,1]. Since det(f — aUrQ) = det(M — aN) det(M~"), it follows that the winding number of det [M (jw) ~ aUzo(jw)N (jw)] is also 0 for all #, a € [0,1]. In particular, for a =0 we get that the winding number of det [M (ja) is zero and thus Q © RH. 3 3.9 PROBLEMS . Consider the nonminimum phase, open loop unstable plant \} phase, open loop unstable pl s2 FL P@= (a) Parametrize all the stabilizing controllers. (b) Parametrize all the stabilizing controllers that guarantee zero steady- state tracking error to a step input. get \3/ Parametrize all the controllers that stabilize the plant 2 oo (415-2) P(s)= (s) s(s—1) Compare the parametrization against the one obtained in Problem 3.1 PROBLEMS 91 3. The purpose of this problem is to explore the effect of nonminimum phase zeros on performance. Consider the feedback system shown in Figure 3.16, where P(s) = (s ~ 10)/(s +2). After analyzing the plant it was determined that a desirable closed-loop transfer function T'(s) is PO)C(S) _ Os +1 Ts) = 1+PC sat Can this transfer function be achieved by an internally stabilizing con- troller? If yes, design the controller. If no, fully justify your answer and then design a controller that achieves a closed-loop function such that: (a) It has robustness properties as close as possible to those of the target system, (b) It has the same steady-state step-tracking characteristics as the target system, \4/A gstemis suid tobe strongly stabilizable itis stabilizable sing an open Toop stable controller Show that a SISO system s strongly stabilizable i if : of eal pols ta between every pat zeros (including po ea le between eran a ition is Known as the parity interlacing property. Hint Use the coprime factorization farm af the Youla parametrization and show that if this condition fails then ¥ ~NQ must necessarily have a RHP zero. 5. Consider now the simultaneous stabilization problem of finding a single controller K(s) that simultaneously stabilizes two SISO plants Gi(s) and Gls). Show that such a controller exists, if and only if, G(s) = Gi(s) Gy\s) is strongly stabilizable. —— 6./Consider the feedback system shown in Figure 3.16, where P(s) = \'(@~ Dis? ~4), and the following performance specifications: (a) Zero steady-state to a step input (b) Perfect rejection of a sinusoidal distusbance with frequency = — ‘Trad /sec. (©) (Open-loop) stable controller. x» 0 + cw) Ps) Figure 3.16. Feedback Loop for Problem 33. in nas been recently shown [St] that the problem of determining whether or not there exist conoller tht simulanceunly stabilizes thrce or more plant is sndeedable 2 STABILZATION Indicate which of these specifications can be achieved with an internally stabilizing controller. Design an internally stabilizing controller that simultaneously achieves as many of these specifications as possible. (Note These specifications are ordered by their desirability, with the most de sirable listed first. If you find out that all of these are not achievable simultaneously, then when designing your coatroller you should drop the least important specifications (ie, higher numbers) fist] Prove, for the following system, that a realization containing poles at s = 0 and s = ~2 is minimal (controllable and observable): + 6 ast = s+ (3.83) oO an z LOOP SHAPING 4.1 INTRODUCTION ‘This chapter extends the analysis results of Chapter 2 to MIMO systems. Both the similarities with the scalar results and the main differences will be stressed. Notable among the latter are the appearance of the condi: tion number of the plant as a new analysis parameter, and the different effects of actuator and sensor uncertainty. Based on these analysis results, the frequency-domain design methodology known as loop shaping ({101, 258]) is introduced. This is a generalization of the Bode method used in classical control ((90, 99, 120)). The problem will be stated in terms of the interconnection of linear frac- tional transformations (LFT, see Appendix B). This formulation allows for treating, in a unified framework, the models of the plant, controller, and un- certainty and performance weights. As will be shown in the sequel, any linear robust control problem can be posed in this way. Example 4.1 Consider the robust tracking problem shown in Figure 4.1. The objective is to synthesize a controller such that, for all elements of the family of plants described by the model G= {U1 + A)W5(S)] Gols), 8 Mass [AClao <1) the resulting closed-loop system is internally stable and tracks a reference signal of the form {r(s), tr(s)Ile <1} with tracking error bounded by 1, that is, |&(s)ll2 <1 93 94 LOOP SHAPING. va w We [a “9-4 xe Figure 4.1. Robust racking problem with sensor uncertainly Jn Figure 42 this problem is recast into. a LET form, with the perform- ance input w = r and ouiput z= & connected to the nominal block M(s). The laste includes the nominal model, the controller, and the uncertainty and per- formance weights, that is ll the input data of the problem. The upper block represents bounded model uncertainty affecting the nominal plant. Partitioning the nominal system M (3) into four blocks conformally to the ‘input and output vectors, (us, and (ya, 2), respectively, yields! a va » M(s) 2 Figure 4.2. Statement of the problem as a LF. "This can be obtained either proceeding diretly from the block digram of Figure 4.1 or by ‘computing the linea factional transformation M(s) = Fr [P(s) Kis} where PCS) denotes the ‘ransfer matsix fom the spats aro tthe OUI, INIRODUCTION 95. MuG) Mu(s) wor [48 ae] Wes)Tu(8) Wals)Tals) [ee Wes) | where So(s) = [I 4 Go(s)K(3)|"" and To(s) = Gols) (5) [P+ Gols) K(S))"" de mote the output sensitivity and its eompiement, respectively. Note tha according to the defintion of internal stably introduced in Chapter 3, inthis cate the Inputoutput stability ofall elements of M(s) 18 not enough to guarantee that the closed loop system i internally stable oRy?- Example 42 Consider the control system shown in Figure 4.3, where the ‘aciual plant has been modeled as a nominal plant Go(s) subject 10 inverse dynamic uncertainty, yielding the following family: = {it Gols)Ws(s)8(5))"" Gol), AE Hoe [A(S)h0 <1} The objective is 10 design a controller that internally stabilizes this family and such that, in all cases, the control action w, weighted by W,(s), remains bounded, in the presence of measurement noise n(s) described by {n(s), ln(s)ll <1} For analysis purposes, we need 1 recast this problem into the standard form of Figure 42, now with input vectors [a.m] and outputs (Yai To this end, twe open the loop and compute the matrix P(s), which takes as inputs [yt and generates the ouput [ys de) way K(s) 96 LOOP SHAPING To compute P(s), by linearity, we obtain all transfer functions between each ‘input and output, seting all other inputs equal to zero. In this case, the resulting ‘nominal system is given by t & " & Pe) i [oer 9) (3) » L[-Gols)Wals) 1] [-Gu(s)) - ee ro Pals) Pr(s) i) The above matrix has bees partitioned into four blocks, acording tothe d- mensions ofthe inputs ugth| and wand the ouspute hal and €, Once the controller K(s) is designed a Bisness analysis cat Be perforried by inter. connecting it t0 P(s) via a LF. as follows FelP(s), K(S)) = Puls) + Pials)K(s) [7 ~ Pals) (9) Pals) = SoGWs Ty = Mo [ ich, ws, “9 where Ti(s) = K(s)Go(s) [0 + K(s)Go(s)|" denotes the input complementary sensitivity matrix. The block separation in the matrix M(3} corresponds 10 the inputs us and n and the outputs ys and & ‘As in Chapter 2, in the sequel we assume that the model uncertainty is linear, time invariant, stable, and bounded. Moreover, without loss of ‘generality, we take the bound to be one, that is, [AGI 2 sup @{Aye)) <1 (44) since any other bound can be absorbed into the nominal block M(s) of Figure 4.2. Thus, for the remainder of this chapter, 4 will be assumed to be an element of BH.., the (open) unit ball in %q.. In the case of MIMO. systems, it is clear why the name of global or unstructured uncertainty is adopted. Even in the case where the uncertainty matrix is a constant, that is, 4 € C™, it contains no additional structural information we could take advantage of, as could be the case when A is diagonal, unitary, or block diagonal. In many practical problems, this structural information is available and, as a consequence, describing the family of plants by means of global uncertainty may be unnecessarily conservative. On the other hand, the tools to analyze sets of models with structured uncertainty are computationally ‘more demanding, as will be seen in further chapters. Therefore a compromise 2Sines (4) € 1, the supremum of 9) achieved atthe Boundary of ts analytic egon C. (1170), that is, 5 = jw. i 7 i NOMINALPERFORMANCE 97 between conservatism and computational complexity should be adopted for ‘each specific application. ‘As in Chapter 2, here the uncertainty and performance weights, W,(s). W,(s),and Wis) respectively, represent the frequency information available ‘on the model uncertainty and input or output signals. In the multivariable «ase, in addition to the frequency content information, they contain structural information as well since they are, in general, block-diagonal matrices. Continuing with the analysis initiated in Chapter 3, in the sequel we evalu- ate the stability, performance, and robustness of the standard interconnection shown in Figure 42. This analysis entails the following steps: Nominal Stability. Guarantee the internal stability of the nominal closed- loop system represented by M(s) (Chapter 3). Nominal Performance. Guarantee that for the nominal plant (ie. 4 = 0) the energy of the output z due to exogenous disturbances {W € £2 \jw(s)ll< 1} is bounded by 1, that is, ie(s)[l2 <1 Robust Stability. Guarantee the internal stability of the set of closed-loop systems represented by the LFT interconnection of M(s) and A\(). Robust Performance. Guarantee that, fr all possible plants, the energy of the output z due to exogenous disturbances {w € C2 , |w(s)[2 <1) is ‘bounded by 1, that is, 2(s)|>< 1 for all {A(3) ; A(6) € BAL.) 4.2. NOMINAL PERFORMANCE This second analysis step assumes that nominal stability has already been established, for instance, using the tools covered in Chapter 3. Therefore in the remainder of this section, we will assume that M (s) is stable. The problem statement of Figure 42 includes as special cases all possible performance objectives mentioned in Chapter 2. In reference to this figure, we define nominal performance as follows: Definition 41 The feedback formance if and only if for the n the output £ due 10 distur nop shown in Figure 4 minal model M (3) ( He(s)lle $1, V{w€ Cay (sla <1) (45) “This definition represents a worst-case analysis of the effect of exogenous disturbances on the closed-loop system. This means that the bound on the ‘output z should be achieved for all elements in the set of disturbances and therefore for the worst disturbance. Since the transfer matrix relating w with + for the nominal case (A(s) = 0) is Mr(s), itis natural to state an equivalent nominal performance condition in terms of this matrix. 98 LOOP SHAPING. Lemma 4.1 Assume that the feedback interconnection of Figure 4.2 is inter: rally stable. Then it achieves nominal performance if and only if Uatza(s)Ilo <1 Proof. The proof follows from the definition of nominal performance and the fact that (See Appendix A) sup {izil2 = IMz2(s)llao (46) wisi a Thus the problem of establishing whether or not a system achieves nominal performance reduces to the problem of computing the 21, norm of a transfer matrix. This can be accomplished by using the algorithms proposed in [57, 104, 247] and briefly reviewed in Appendix B. Similarly, the problem of designing a controller that optimizes the nominal performance of the closed-loop system ‘can be reduced to the problem of minimizing the Ha. norm of a suitable transfer matrix. This leads to the well known 24, control problem addressed in Chapter 6. 43 ROBUST STABILITY ‘Consider again the interconnection shown in Figure 42. The transfer matrix relating the input and output signals is given by the LFT between the nominal and the uncertainty blocks, that is, 269) = FulM(s), (6) (8) an = {Moa(s) + Mou(s)O(0) HE ~ Mix(S)S()T Mit) } w() where F,(,-} denotes the upper LFT defined in Appendix B. If we assume nominal internal stability all matrices M, (3) i,j = 1,2 in equation (4.7) are stable. Thus the only way to destabilize the interconnection is through the un- certainty A(s) appearing in the element {0 — Myx(s)A(s)|"", This observation leads to the following robust stability condition: Proof. We will prove that the condition [Mulla <1 is equivalent to = Muy(s)A(s)|"' input-output stable, for all (s) € BH. (=) From standard singular value properties we have 2 I~ Muy(s)A(s)] > 1 — 9 [Mar(s)A(s)] > 1—F[Mulsy] (48) NOMINAL PERFORMANCE AND ROBUST STABILITY 99 Since ||Myilj <1 and My, is stable, from the maximum modulus theorem wwe have that o [My;(s)] < 1, Vs €C.. Hence I~ My(o)Q6)) > 0 a9) eo det | — My (A) 40, Veer, (4:10) (<=) Assume there exists s, € C, such that o (Myi(s.)] > 1. Consider the singular value decomposition Mi,(S.) = UZV* and define A= aVU* with = 1/6(My(s,)] <1. Then A € BH. and det - My(s.)A) = det jf ~ aVEU"} (aan) det(/ ~ a3) (4.12) Tee since (1 - ae,) =0. To complete the proof we need to show that internal nominal stability of the closed-loop and input-output stability of [1 — My(s)4(s)|"', for all A(s) € BY, are equivalent to robust internal sta- bility (>) All transfer matrices that map inputs to outputs in Figure 42 contain the product ot stable matrices (because of nominal internal stability) and [7 My(s)A(9))"", which is also stable (<=) Consider us — Aya(s) + n(s) in Figure 42. The transfer function from n(s) to wa(s) is given by Tun = [0 ~ A(s)Mui(s)| . Stability of this transfer matrix follows from the assumption of robust internal stability of the interconnection. a (4.13) 4.4 NOMINAL PERFORMANCE AND ROBUST STABILITY Clearly, in any practical situation, we should require robust stability of the closed-loop system. However, since there is usually more than one controller achieving robust stability, it is natural to attempt to use these additional degrees of freedom to optimize performance. Toward this goal one could attempt to optimize performance only for the nominal model while stabilizing all members of the family of plants. From the discussion in the two previous sections, it follows that a necessary and sufficient condition for simultane: ously achieving nominal performance and robust stability can be obtained by combining Lemmas 4.1 and 42, that is, [pesenia inal + Rakasy performance |*| stability an max {Mills EM) 100 LOOP SHAPING while this condition is useful for analysis purposes (since it can easily be checked), attempting to use it for synthesis leads to a multidisk problem (a problem that entails synthesizing a controller that satisfies simultaneously two or more . conditions). At the present time there is no known analytical solution to these problems, although in some cases these are amenable to efficient numerical solutions [217]. An alternative approach is to obtain a sufficient condition exploiting the properties of singular values. Since matrix dilations do not decrease norms it follows that 1 smax {Milos |Mzalc} <1 = |e | <1 (45) The problem of synthesizing a controller such that the closed-loop system satisfies this sufficient condition is known as a mixed sensitivity minimization problem and can be solved using standard %,. tools covered in Chapter 6. Note in passing that a controller achieving nominal performance and robust stability can lead to severe performance degradation when confronted with ff-nominal conditions. This situation can be avoided by synthesizing a con- troller achieving robust performance, that is, a guaranteed performance level for all members of the family. As we will see next, it turns out that this can be accomplished by solving a scaled mixed sensitivity problem. Thus, given the difficulties involved in solving exactly a nominal performance robust stability problem and the fact that a tractable sufficient condition leads to a problem hhaving similar complexity to that of (approximately) solving a robust per- formance problem, we will devote most of the remainder of the chapter to the solution of the latter. 4.5 ROBUST PERFORMANCE Definition 42 The feedback loop shown in Figure 4.2 achieves robust per formance if and only if all members of the family of plants described by the interconnection of M(s) and A(s) achieve nominal performance. Equivalently, the performance ouiput z should have bounded energy for all bounded energy disturbances and all models in the set. As we have seen before, without loss of generality, this can be restated as follows: Webs, VOwe Lr, lw(s)le <1}, and YAG)E BH (4.16) Using Lemma 4.1 to impose that all members of the family of plants 2 The situation is diferent inthe ease of multiobjective problems, where assessing performance and stability requires using differen induced norms. Most of the resis crrenlly available for ‘ese problems are limited to nominal performance /robust stability [301 ROBUST PERFORMANCE 101 achieve nominal performance yields the following neccesary and sufficient condition for robust performance: WFeIM(s), M6} <1, VA(s) € BH ay (4.7) This condition cannot be checked easily, since it depends on the uncer. tainty A(s). Therefore it would be desirable to obtain a simpler condition depending only on the nominal system M(s). However, it turns out that condition (4.17) cannot be stated any longer in terms of the H. norm of a suitable transfer matrix involving only the nominal plant. Thus the exact assessment of robust performance will be postponed until Chapter 7, where new tools will be developed. Instead, at this point we will exploit the con- ditions for nominal performance and robust stability to obtain a sufficient robust stability condition for two different cases: (i) sensor uncertainty and. (ii) actuator uncertainty. 4.5.1 Sensor Uncertainty Consider the robust tracking problem of Figure 4.1. Applying Lemmas 4.1 and 42 to this case yields the following nominal performance and robust stability conditions: pengminal. | <> [We(s)So(S) $1 (418) [REE | WwacyTo(ota <1 (49) Imposing nominal performance on all members of the family of models leads to Robust performance @ {Welw} [1+ (+ AUe)Wae)) Goa)K Yo)!" <1 (420) = > 6 {Wels0) (C+ Gow) Kw) + Mo) Ws 4)Go40)Kw))Y} <1 9 6 {W.y0)Ss0)|l + AUa)Wao)ToGw)I} <1 (421) - [W.G0)S.4a)}& {1+ A0)W3G0)T.0)) ‘y 2 1We(y0) 8, 0))_ a+ AG@Ws0)T G09) 1 (422) 102 LOOP SHAPING. (423) > G[We(}0)5,(J0)] + 5 [Walr0)To(s0)) <1 for allo (424) where we used the fact that [&(s)|,, <1. At this point the following com- mens are in order. ‘* While condition (424) is similar to the one obtained for SISO systems, in this case itis only sufficient, ‘¢ From equation (4,23) it is clear that the nominal performance margin we had for the nominal plant Go(s) decreases as the “size” of the uncertainty A(s) increases. # As in the univariable case, a method to design for robust performance is to synthesize a controller that minimizes the robust stability condition while scaling the performance requirement (using @W, with a < 1) until (4.24) is satisfied, # If the system is recast into the standard LFT form, condition (4.24) becomes. @(Mn(ja)] +o [Miy(s0)] <1, Yo (425) This condition is no longer valid in the case of actuator uncertainty Rather, a3 we will see next, in this case a different, more rest condition should be used. 45.2. Actuator Uncertainty ‘The case of uncertainty in the input to the nominal model changes the suf ficient condition for robust performance, due to the nonconmutative nature of the systems involved.‘ Consider the control problem of Figure 44 subject to the same performance requirements as before, but where the uncertainty appears now in the actuator. For simplicity we have assumed that the plant Gg is invertible. It can be shown that in this case the nominal performance condition (4.18) remains unchanged, while the only change in the robust sta- bility condition (4.19) is that the input complementary sensitivity T; should be used instead of To(s); that is, perioratce | > UW(8)S (he $1 (426) [Rea] = IWa(S)TAS)ILe <1 (27 Imposing nominal performance over all members of the family yields the following robust performance condition: “OF course, this situation docs not arse inthe SISO case ROBUST PERFORMANCE 103, t —— wa {a 7 = 4 Fou 44, Robust rocking problem win actuator uncer. [a | = of, (rsGr+aWyKr } <1 (428) > a {W.((1+ 6K) + AWK" <1 429 W.S, 9 GaMgKS,/} <1 (430 a - a{wes.c, [reawaxas.ay'o,) 'o,'} <1 (431) = o1ws,ja (Ge [6;]a[asawery'] <1 (32) (433) > F[W.S,] [Go] + [WaT] <1 for allo (434) where x [Go(sw)} denotes the condition number of the nominal model. Note that (4.34) is more restrictive than the (sufficient) robust performance con- dition corresponding to sensor uncertainty. This is due to the fact that the condition number of the nominal model has a minimum value of 1. There- fore, for the same uncertainty level (in the sense of ||W37illx ~ [IW Tolle) the value of a [We(sa)S.(Ja)] required in order to satisfy equation (4.34) is smaller than the value requited {0 meet (424), resulting in a lower nominal performance level (since a smaller W, must be used in the first case). The condition number x (G,(jw)] represents, as a function of frequency, 104 LOOP SHAPING the relation between the maximum and minimum gain of the nominal model This characteristic is exclusive of MIMO systems and denotes the “balance” between the input-output channels of G,(s). The ideal situation from this point of view is when the nominal model has the same gain for any input vector direction. This is the case, for instance, where the system matrix is, unitary at each frequency. Note that a high condition number by itself is not a problem, except when combined with actuator uncertainty. This situation arises in practice in the case of high-purity distillation plants ((287)), leading to a challenging control problem. A conceptual explanation of the difference between this case and the sensor uncertainty one is the following. When the uncertainty appears at the output of the nominal model, the interconnection between the output of the controller and the mode! input is not subject to uncertainty. Therefore we may consider the series connection of Go(s) and K(s) as a single transfer matrix Lo(3) = Go(s)K(s). Instead, in the input uncertainty case, the uncertainty ‘A(s) “separates” the connection between K(s) and Go(s). This means that the controller output channels and model input ones are “mixed up.” This does not pose any problems for robust stability, because the controller and the ‘nominal plant can be connected through the output of the later as in L;(s K(s)Go(s). But for certain performance requirements this is not possible ‘Therefore ifthe channels are “mixed,” an output of the controller that should g0 through a small gain direction of G,(s) could conceivably enter the plant through a high-gain direction and vice versa. This effect becomes worse as the relation between the higher and lower gains of the model (condition number) becomes higher. In many cases, if « (Go) is very high in the frequency region of interest, condition (4.34) can guarantee very little performance for all the members of the family of models. In those cases, a new idemtification of the set of models should be made, attempting to reduce the size of the uncertainty A(s). Alternatively, we could seek @ less conservative robust performance condition. The latter approach is covered in Chapter 7, while the former is addressed in Chapter 10, In order to avoid these difficulties, one could be tempted to try to model 4 plant subject to actuator uncertainty as a set of models subject to output ‘multiplicative uncertainty (possibly having different bounds). In this case, we should change the input uncertainty 4,(s) to an “equivalent” output uncer- tainty matrix 4o(s) so that both descriptions cover all elements of the family of plants. This can be accomplished as follows: {Go T+ ACs], FIAC)} <1, ¥s= so} 435) jo) (4.36) 2 Ay = Gals)AGEM5) (437) Go SUI + Bols)] Gols), F[Ao(s)] <7, Vs DESIGN PROCEDURE 105, 16¢s)} Figure 4.5. Inputto output uncertainty ranstormation. where again, for the sake of simplicity, we have assumed that G,(s) is invert ible. To keep the description as “tight” as possible, the upper bound of the ‘output uncertainty size, y, should be as small as possible. While cannot be ‘computed exactly, an upper bound can be obtained as follows: 5 Bolw}] < a(Gya)]e (8,40) # [6-"G0)] < «1G0)) 2 ry Yeo (438) Note that, since y IweGa)), Vo (440) = allo(u)] 21+ Imo), Yo (ast) and say 9 a {Wsyo)Lo(0) [7+ LoGo))"} st, Yo (4.42) = a{{rts'u)]"} <1/lwsia) Yo (443) ea [F+ 15'G0)] > baled], Yon (aay = a[L;'Ge)] -12 bwaGe)), Yo (4.45) = aloo] <1/ {1+ alo)ll, Yo (446) ‘A similar procedure (see Problem 4) leads to the following necessary con- ditions: [ peNgminal | + cto(ay] > Iwelo)|=1, Yo (aan) = 9 [Lo(w)| <1/{|waw)|~1], Vo (4.48) Note that the sufficient conditions (4.41) and (4.46) conflict since @ [Lolya)] < (1 + fwaa)|]! <1 < [1+ [wee 2[Lolse)] (449) for all w. This confit can be avoided by using the fat that performance and stability requirements are usually separated in frequency. Performance re- quirements are often more stringent at low frequencies. Thus we] is large at low frequencies and ~0 at high frequencies. On the other hand, uncertainty is usually higher at high frequencies so that jwa(Jw)| > 1 for large w and ~0 DESIGN PROCEDURE 107 Figure 4.6. Loop shaping design, for low frequencies, The design procedure exploits this separation to define admissible regions for L at low and high frequencies (see Figure 4.6). Specif- ically, the minimum (maximum) singular value of L,(s) should be shaped :0 that it falls above (below) = (upper) bound at small « (high «), that we6w)| stabte, afd minimum phase, In the general case, a suitable controller can be synthesized by first sepa ene mvenable quar, sable, miniqum phase) from the Bon- invertible part of the plant as follows: Gy(8) = Gge(s)Gin(s). The controller is then given by K(s) = Ggi(s)G.(s)Lo(3), where G.(s) sae en eects alias Snes malipliation by an i rec te tareiace ihe. singular values (ina; (EoU) 9; (Gax(s)G-(6)LaUo)])s the resulting loop function satisfies the con- straints. This is illustrated by the following simple SISO example: (+N G6-5) 4 e son Sl 453 8068) = Ee ay mbox 653) From the loop shaping design itis desirable to have [Gel = 1/[p0)- The controller that satisfies the above condition for the loop transfer function (3) is Kis) = © (454) since (s ~5)(5-+3)/(s +5)(s - 3)] = 1 for all s= jo.) 2. The sufficient conditions (4.50) and (4.51) are valid only for low and high frequencies, respectively. Thus as a final step in the synthesis pro- cess, the resulting loop function should be examined over the complete frequency range. 3, For the robust stability and nominal performance bounds, in the case ‘of actuator uncertainty, there is no need to include the plant condition number. Nevertheless, it must be included with the scaling when we proceed to verify the robust performance sufficient condition. EXAMPLES 109 4, Since the method is based on the use of sufficient conditions, it may be too conservative, In cases combining. high-performance requirements with high levels of uncertainty, the bounds obtained from these condi tions may overlap, so that the admissible region for L becomes emp. ‘One solution is to carry out the design using the necessary rather than the sufficient conditions. In this case the synthesis must be followed by a verification of robust performance (which includes nominal per- formance and robust stability). Alternatively, a less conservative design procedure should be used (such as w-synthesis convered in Chapter 7), fr the uncertainty in the model should be decreased through an addi- tional identification effort (see Chapter 10). 4.7 EXAMPLES 4.7.1 Permanent Magnet Stepper Motor In this section we illustrate the use of the loop shaping technique with an application example. The system under consideration is a permanent magnet step motor driving a precision linear positioning table [53]. The objective is to track prespecified angular and velocity trajectories, with minimum track- ing error and without excessive control action, Robustness against model uncertainty is also considered, Description of the Plant Although the stepper motor is a nonlinear system, a linear model of the tracking error dynamics can be obtained through the use of feedback linearization ((156]). The state-space and transfer matrix, representation of this linear model are given by -R/L 0 0 o}10 0 (L.(Ja)], which were designed in [53] for the same performance objectives® (curves c and d in Figure 47). These will serve as "upper and lower bounds on the open-loop transfer matrix to be designed Although we have considered in this chapter only lower bounds for per- formance, in general also upper bounds can be desirable. For example, when the performance objective is to bound the control action, as in this case, we define nominal performance as ||W,(s)u(s)|2 < 1 forall bounded references lir(S)llz < 1. Here W,(s) isthe control action weighting function. Proceeding, as in previous sections we obtain the following sufficient condition over the Joop, which guarantees this performance objective. NP c= |WAOKO)IT+ GOK) | <1 (458) = [MOo@LOU +L <1 (459) & (WASIG)] Geena < Y (4.60) Since (0) = Gla) hs at mos rank 2, we consider only theft and second singular EXAMPLES 111 Loop singular values, petormance and robustness bounds w (rads) Figure 4.7. (0) {te} (b) [U0 (€) upper performance bound, (6) lower perform: {ance bound, and (e) upper robustness bound, = & [Wals)G%s)] < 1, Vs=jo (4.61) a aL) 1 = LO < sma Y= (4.62) where G!(s) is the left inverse (if it exists) of the model G(s). ‘Considering the bounds c and d in Figure 4.7, we would ideally want an open-loop transfer matrix L(s) with ag (LGw)} = 3 [Lyw)] = t/s, a € (0,1), that is, L(6) = diag( Gs]. However, itis clear from the structure of the plant G(s) that a diagonal open-loop matrix cannot be achieved (since the plant hhas no right inverse). Thus we will try instead to obtain a block diagonal transfer matrix, of the form L(s) = G(s)K(s) = (493) 112 Loo SHAPING where « represents nonzero elements. The simplest controller achieving this, ‘open-loop structure has the form ce) O = 0 0 KO 0 kaals) kaals) als) (464) Furthermore, since the the open-loop plant is also block diagonal, it follows that we can design each block separately. For the first scalar block we can achieve the loop ¢;/s and compute a value of f, such that [é,/ja fits between the upper and lower bounds in Figure 47. Before we start designing a controller that fits L(s) between the nominal performance bounds, we should consider internal stability, which places addi- tional constraints on the design. Although G(s) has no right inverse, we can design K(s) = D(s)Gi(s), where G(s) isthe left inverse of the stable part of the nominal model and D(s) = diag{d(s), da(s)] is a diagonal system. Here we consider the following stable-unstable factorization of the model: 1 see s(s-+B/J) “pe | 71 0 w= [02 ] (0s) Ky Tes) | a Kn, Ipls) Bo such that G1(s)G(s) = withthe stable left inverse being (+R/L) 0 0 0 Gy=| 9 IR vat] (4.66) Kml Kal As a consequence, T(s) = Gls)K(s) I + G(s)K(s)]"" = G(s) I + K(s)G(s))"" K(s) = GAS) [1 + Guls)D(G)}"" Guls)D(SG'(5) = o[§ g]oim=[5 ,p,]e@so aon where we used the fact that K(s}G(s) = D(s)G,(s) and we have replaced the input complementary sensitivity Ty(s) = + K(s)G(s)}"" K(s)G(s) by EXAMPLES 113 diaglt,,4], where as) 2 dy(s)lL + di(s)-" and (5) [ey Note that the input sensitivity $)(s) = [I + K(s)G(s)|"' its complement T/(s) and T(s) are all stable for an appropriate choice of K(s)G(s) diaglds(s),dz(s)/s|. The poles at the origin of G(s) and K(s) will be can- celled by the zeros at the origin of 5(s). As a consequence G(s)S;(s) and 5,()K(s) are also stable and internal stability is achieved It follows that the design specifications can be met with a controller of the form a(s\(s+R/L) 0 0 0 Ko)= JR pi) RB+K: Gey o d,(s) Kee) Keb Bhs) ‘where the design transfer functions di(s) and do(s) should be selected to meet the performance and robustness specifications. Since in this case the performance bounds have been imposed dircetly over L(3) (and not over T(s) or S(s)), there is no need to verify the sensitivity or its complement. Nevertheless, the issue of integral action will be considered next. ‘Note that the input loop L,(s) = K(s)G(s) is diagonal and depends on 4(s) and d3(s)/s. Therefore to achieve a loop of the form L,(s) = diag(¢,/s), wwe only need to have integral action in dy(s). This may seem clear from the structure of model G(s), which already has a pole at s = 0 in its element ge(s), but none in gy(s). As a consequence the block-0, Rr=R>0, [9 AE (52) Furthermore, by using the change of variable d =u ~ R-INTs, itis easily seen that (51) is equivalent to in f [e™@Ox0 +7 Ru] de 53) wcll) OF 030, RT=R>0 (54) THE CLASSICAL LINEAR QUADRATIC REGULATOR PROBLEM 129 where Q = 0 NR“INT. Here Q is a matrix that penalizes the deviation of the states x from the desired operating point, while R penalizes the control effort. Thus (5.3) represents a trade-off between regulation performance and control effort. A commonly used choice for the matrices Q and R is Bryson’s rule: the matrices are chosen to be diagonal, with elements given by = maximum acceptable value of [x,(0)}? 2 (55) = maximum acceptable value of {4(0))” >| However, we stress the fact that while this rule usually gives good results, there is no exact way to translate specifications given in terms of the peak values of x and u into specific values for Q and R. Thus several trial-and- error design iterations may be required in order to accommodate design specifications given in terms of peak values rather than £2 norms (akin to energy). While in general the system may be subject to several classes of exoge- nous disturbances, we consider first the case where the system starts from ‘an arbitrary (but deterministic) inital condition x(Q) = x,, and it is subject to no other exogenous disturbances during its evolution, leading to the clas- sical LOR formulation. The effect of additional exogenous signals will be considered later in the chapter. Tn the sequel, for simplicity we will normalize the control action by consid ‘ering a Cholesky factorization of R, R= V7V, V.> O and using the change ff variable tu» = Vu. Thus we have that u7(ORu(t) =u! (uals). If follows that without loss of generality we can assume that R— in (5.3). In terms of this new input, and using a factorization of Q = CTC, we have that, . aT Qul) + ul ual) Wee ee =o Gol [F [EB] e424). wl) vei [ (56) Yo woll “lle p [7@ wo] pr] le D) where C and D have been chosen so that D7|C_ D}=[0 1]. Using these results, we can precisely state the classical LOR problem in the general form of Figure 5.1 as follows Classical Linear Quadratic Regulator Problem Given the LTI system H() = Ax(t) + Bute) 2) = Cx(t) + Dut) x(0) = x0 © (57) 130%; OPTIMAL CONTROL with (A,B) stabilizable 68) (C.A) detectable 69) c™D=0 (510) D'D=1 (oan) find an optimal control law u € £20, 00) that minimizes? | Classically, the LOR problem has been solved using either variational cal- culus techniques or “a completion of the square” type argument (eg, see [61, 63]). Here, we will pursue a different approach (following after [104]) that exploits the orthogonality and norm preserving properties of certain linear operators. The main advantage of this approach is that itis easily ex- tended to both the general 1; and the M,. cases. Toward this goal, we begin by giving a characterization of inner systems* in terms of the solution to an algebraic Riccati equation, Lemma 5.1 Consider the following stable system: ul Als (5.12) oA] om where the pair (A,B) is controllable. Let X = XT >0 denote the solution to ‘he following Lyapunov equation: (6.13) Then U(s) is an inner system if and only if the following conditions hold: DTC +BTX =0 (6.14) D'p=1 Proof. (-+) Using the state-space formulas given in Appendix B to obtain UG) and U(s)U(s) yields rue) (315) Dicep =e |e pip) 2 From the discussion above and the definition of the given in Appendix A, it should be ‘a this point that Je is precisely given by (53) * Recall that a LTH stable Wansfer matrix U is ner if UU = J and that inner systems are norm preserving (Lemma 3.9). THE CLASSICAL LINEAR QUADRATIC REGULATOR PROBLEM 137 Applying the similarity transformation we % [ 1 ‘] . Xe XBR, to the realization (5.15) and using (5.13) and (5.14) yields A 0 Uisyu(s) = |_-XA-a?xX-cTC -al | -xB-cTD eee ee D'D aoe) aA He @ | foe ate vr [ = [6i9 oat [7 Ope Finally, eliminating the nonobservable and noncontrollable modes in (5.16) yields the desired result. To show the converse, note that if the pair (A, B) is controllable, then from the first equation in (5.16) it follows immediately that U"(s)U =I only if D'D =I and (D™C + B7X,A) is completely unob- servable, which implies (5.14). a Next, we introduce two preliminary results that will be used to solve the LOR problem. Lemma 52 Given the system (5.7), denote by X > 0 the solution to the fot- lowing algebraic ATX +XA+C"C~ XBBTX =0 (6.17) Consider now the change of variable v = u— Fx, where the feedback gain F is given by F=-B"x (6.18) ‘and let U denote the system in terms of this new input variable v, that is, uw - (412 = eqD (519) Ap = A+BF Cp = C+DF Then U is inner Proof. Simple algebra using the definitions of Ay and Cy and the orthogo- nality condition CTD = 0 shows that (5.17) is equivalent to AIX + XAp+ CFC; =0 (520) 132, OPTIMAL CONTROL Moreover, from the definition of F, using again the orthogonality condition, itis easily seen that DT Cr + BX = 0. Thus from Lemma 5.1 it follows that UGG) is inner, provided that itis stable. To show stability, recall that from the properties of the Riccati equations (see Appendix C) we have that the assumption in (5:8) is a necessary and sufficient condition for the existence of 4 unique positive semidefinite solution X' > 0 to equation (5.17). Moreover, this solution is stabilizing, that is, A(A — BBX) CC o Lemma 53 Consider again the feedback gain F = BX and denote by GAs) the following system: ous)? [4244 521 (= (521) Then U(s)Gc(s) € RHE Proof. Since Ga(s) is strictly proper, we need only to show that U"G. is strictly antistable. Proceeding as in Lemma 5.1, we have that AE -CECp | UIS)Gs) = |_0 _ a -at 0 =| 0 a |r fe) 0 elt liad (522) = eat 0 Which is antistable, since Ay is stable. a Finally, we recall (without proof) the following result from (104) Lemma 54 Consider the system (5.7). Ifthe control and output signals u,z € £,{0, 00) then its states x are also in £210,000). Moreover, x(t) —+ 0 as ¢ ~+ 00. Using these results, we are ready now to give a complete solution to the LOR problem, ‘Theorem $.1 There exists a unique stabilizing control action u(t) = F(t) that minimizes \\z lp in (5.7), where Fy = —B?X is the feedback gain defined {WE STANDARD}; PROBLEM 133 in (5.18). Moreover, we have that atin, HECOIE = UGels)x013 = xf X20 (5.23) where Ge(s), the closed-loop system, is given by (5.21). Proof. Proceeding as in the proof of Lemma 5.2, change the control input to u(t) = u(t) ~ Fox(0), In terms of this new input we have that H(t) = Apx(t)+ Bolt) — x(0) =x (524) 2(0) = Crx(0) + Dolo) Stability of Ap implies that if v € Cz then z € £2. Thus we can take the Laplace transform of (524), obtaining 2(5) = Gels)xo + U(s)uis) (5.25) where Ge(s) and U(s) are defined in (5.21) and (5.12). Using Parseval’s the- orem and the fact that U is inner yields W213 = IGels)x0 + UCs)u(s))I3 5.26) = |UGlshx0 + ¥(5)13 : Finally, using the facts that v(s) € Mf and U"Gexo ¢ 74} and that multiplica- tion by an inner matrix is 2-norm preserving, (5.26) reduces to Weil UGe(s)ol? + (s)I3 627) From this last equation it is apparent that the optimal value of 22 is obtained when v=0-e=2 u(t) = Fax() a Itis worth stressing that as a by-product of the proof we recover the well known fact that in the LOR case optimal performance among the set of all possible state-feedback stabilizing controllers is achieved by using linear static feedback u = Fx. Tis follows immediately from the fact that any other control law having,o #0 will result in higher [3 5.3. THE STANDARD 71 PROBLEM Consider again the block diagram of Figure 5.1. As we mentioned in the introduction, a very common control problem is to synthesize a controller that stabilizes the system and minimizes (in some sense) the “size” of the 134; OPMMAL CONTROL output due to a given class of input signals w. The LOR problem addressed in the last section is a special case of this problem, where the input is restricted to be of the form w(¢) =x,8(0), with x» arbitrary, and where it is assumed that the states are available for feedback, leading to the following state-space realization for G(s): 0 Da (628) All Bh o 0 (5.29) As before, the objective is to synthesize an internally stabilizing controller that minimizes the 2-norm of the closed-loop transfer function from w to z, IITzls- Recall from Appendix A that this can be given a physical interpreta- tion in terms of minimizing the root mean square (RMS) value of the output z due to a gaussian white noise input with unit covariance. This problem can be formally stated as follows, Standard’, Problem Given the plant G(s). find an proper, LTI controller that Tit Note that we fiave assumed that Dy; = 0 and Dz =0. In the sequel, for the sake of simplicity, we will make the following additional assumptions: (Al) (A,Ba) is stabilizable and (C,,A) is detectable. (A2) (A,B1) is stabilizable and (C;,A) is detectable (A3)_ C{Di2=0 and Dh, = (A4) Diz has full column rank with D/,Di2 = and Dz; has full row rank with D2,DJ, ‘The assumption Dy; = 0 guarantees that the closed-loop transfer matrix is in 4 (recall ata reat rational stable transfer matrix is in M, if and only if it is Sictly proper). A complete treatment of 2 problems with distur- bance feedforward can be found in [332]. The assumption Dz» = 0 is made to simplify the algebra. It can easily be removed by using the loop shifting tech- nique discussed in Chapter 3. Assumption (Al) is clearly necessary for the system to be stabilizable via output feedback. Assumptions (Al) and (A2) together guarantee that the control and filtering Riccati equations associated with the %2 problem admit positive semidefinie stabilizing solutions. As we THE STANDARD A> PROBLEM 135, will sce in the next section, they can be relaxed to Giz and Gay not having invariant zeros on the jw axis. The orthogonality assumption (A3) is made to simplify the development. It can be relaxed at the price of the appearance of sin the controller. Finally the rank assumptions in (A) gua is nonsingular, while the normalizing assumptions (DEDiz = 1, DziD3, = 1) d0 not entail any loss of generality since they can always Be met by redefining the inputs w and w if necessary. "Next, we introduce some notation that will be used to solve the 3 prob- Jem. Consider the following two Riccati equations: cot ATX, 4 Xo — XBDBIN + CTC, (530) (631) asi AY + YoAT ~ YoC] CaYo + BBP = Recall from Appendix C that these equations can be associated to the Hamil- tonian matrices H, and Jp, respectively, where Ab -cIG BT A Cy A ~BBE cle. -AT Under assumptions (A2) and (A3) these matrices satisfy Hs,J2 € dom(Rie) (see Appendix C), which in turn implies that (5.30) and (5.31) have unique solutions Xz = Ric(Hz) > 0 and Y2 = Ric(Y2) > 0, respectively. Moreover, these solutions are stabilizing, that is, A—ByBYX> and A~Y,CICy are Hurwitz, ‘With these assumptions we are ready now to give a complete solution to the 74 problem. Theorem $2 Under assumptions (Al)-(A4) the unique optimal H, controller is given by pr [ears * opt (8) = tet e (533) with the corresponding optimal value of \\Tzwl2 given by min Te0(3)13 = IGe(s)Bi1B + IG (SIIB (534) cmaiing 136%; OPTIMAL CONTROL where Gs) = “ “] Ap = A+ BaF; Gr = C4 Dake Ay = Atlas (535) Ary = A+ BaPy+ aC, Bu = Bit LD Ra BY 1X, = Riel) Ly = -¥,CF Ya = Ric(l2) ‘To prove this result we will decouple the output feedback problem into full information and output estimation subproblems, proceeding as in Chapter 3. ‘Thus before proving the theorem in the general case, we will consider the same four special problems (Fl, DE, FC, and OE) introduced in Chapter 3 as an intermediate step in obtaining the Youla parametrization. Full Information Problem Lemma 55 Consider the FI case, that is, the case where A | BB G | 0 Dp (5.36) (ol Ff) (} 1. The optimal 13 controller is Ker = (Fz 0} 2. The corresponding optimal value of the Ho cost is You = minue ce, WTeella = |GeBy fp = trace(BY X2 By)". 3. Given y > Yop, the set of all intemally stabilizing controllers such that Towle Sy is {K(S):K =[Fx Q(s)}, Q(8) © RM, Ol S -1GeBsIB}. THE STANDARD 11; PROBLEM 137. Proof. As in the LOR case, define a new control input v © obtaining (1) ~ Ext, fy = [ef & pw 537 «0- Leto a. | [08] cay Or equivalently, Tel) [e+e]. ara Tels) =UGe(S)By YU(s)Tow(s) (5.38) which, combined with Lentihas 5.2 and 53, yields Wow SIIB = WGes)BiI3 + [Tow( 13 (5.239) Clearly, this quantity is minimized by setting v = 0, which results in T, This is achieved by the (unique) static controller Kr, 2 [F; 0] yielding amin |Tex(3) = Ges)Bs|3 = trace(B7 X28)" (5:40) where the last equality follows from the fact that Xz is the closed-loop ob- servability Gramian. ‘To show the last property, recall from Chapter 3 that the set of all stabilizing FI control actions can be obtained using controllers of the form K(s)=|F Q], A+ByF stable, Q © RM. Selecting F =F; yields ITowl = 16.13 + 101 (541) Thus ||Tewl2 <7 <> QE RA, OI < ¥?— |GeBIB. o Remarks 1, Note that the optimal controller uses only feedback from the states. Thus, under the assumption that Dj, =0, the FI and state-feedback (LOR) problems have the same optimal solution. 2, The optimal controller depends neither on the exogenous disturbance w nor on B; (a similar situation arises in the LOR case where the optimal gain does not depend on the initial condition x,). However, the optimal value of the performance index is a function of By. 3. We have indeed proved a stronger result than intended. From the proof of the lemma it follows that the best possible control action (over the set of all stabilizing control actions u € £2) corresponds to the control action generated by linear static state feedback. By exploiting duality, we immediately get the following result 13874, OPMMAL CONTROL Full Control Problem Lemma 5.46 For the full control case A[ B10 Gres)= |G | 0 07 (5.42) G | Dn 00 the following properties hold: 1 Thea cater re~ [3] 2. The corresponding optimal value of the Wy cost is Yop: = UTewlla = WCrGylla = trace(C;¥2C7)", 3. Given y > Yop, the set of all internally stabilizing controllers such that Teall <8 {ato:«= [92] 20 «rm, tO < ¥-tevcpt} Finaly we comer the DF and OF problems, Disturbance Feedforward Problem twee Lemma 5.7 Consider the DF problem (5.43) where, as before, we make the additional assumption that A — By C2 is stable. Then the following properties hold: 1 Yop = min Teall = 1G [ A+ BaF, ~ BiG | Bi | Avert (5.44) A 0 3. The set ofall LT1 controllers such that \Tsully< y is given by K(3) = Fir, Q), QE RX, [O13 < 7 ~ |GeBills, where e Kom(s) Jor(s) = (5.45) THE STANDARD M2 PROBLEM 139 Proof. Recall from Chapter 3 (Lemma 33) that under the additional hy- pothesis of stability of A — B,C the DF and FI problems are equivalent, in the sense that if Kr) =[F Q] stabilizes Gp, then Kop = FiIpr,Q) stabilizes Gpr and yields the same closed-loop transfer function, that is, Fu(Gr1, Kr1) = F(Gpe, Kor). Moreover, Kog parametrizes all the DF sta: bilizing controllers. The proof follows now by combining these facts with Lemma 5.5 a As before, the solution to the OF case follows from duality Output Estimation Problem Lemma $8 For the OE case where G has the form A | BB Goets)=|"G | 0 7 A~B,C, stable (5.46) G | Dn 0 the following properties hold: 1. The optimal He controller is 47) At L2G - BC | Lr Koe(s) = |——_— Gq 0 2 Yon © min WTovle = WCvG le 3. The set of all internally stabilizing controllers such that |\Tewll2 < y is K(s) = Foe, Q), QE RM2, ||OI3 < ¥? — ||CiGylf, where Jor(s) = (5.48) Using the results of Lemmas 5.5-5.8, we can now give a proof of Theorem 52. Proof of Theorem 5.2. Proceeding as in Chapter 3, make the change of input variable u(r) = v(t) + Fax(1). This partitions the system into the wo 1404; OPTIMAL CONTROL K(s) Figure 8.2. Output feedback as a cascade of an OF onda stable system subsystems shown in Figure 52, where G(s) and Gimp have the following realizations: : A+ Bor | By Bs Gis) = | C+Duhk | 0 Da A |B Bm (5.49) Gels) = | | 0 oT G | Dy 0 Note that Gy is stable and that Gigy has an OE form. Moreover, K(s) internally stabilizes G(s) if and only if it internally stabilizes Gimp. From the figure we have that ) = [Pe] BB] poe Vee lo Da [ra] = Ge(s)Biw(s) + U(s)v(s) (5.50) or equivalently, Ten(8) = Gels)By + U(S)Tow(s) (551) From Lemmas 52 and 5.3 we have that amin, WTaa(9I2 = IGAS)BIB + min. Ton(s)I8 (5.52) But since Ginp has an OE form, from Lemma 5.8 we have that the controller that minimizes this last transfer function is precisely A+ Lio + BF By Kor(s) = J} os RELAXING SOME OF THE ASSUMPTIONS 141 yielding min |Tov(s)l, = [F2Gy(s)[l- It follows that ang Tem = WGCCSIBLIE + MG (SIIB (5.54) o Remark Note that the optimal controller (5.33) clearly exhibits the well- known separation structure of the 2; problem. This can be made readily apparent by rewriting its state-space realization as B= A+ Brut LalCot -y) 5.55} aA (555) ‘Thus the output of the controller is precisely the optimal estimate of the LOR control action w= F;x. Alternatively, the separation can be seen directly from the proof of the theorem, since the state feedback used in obtaining, G, corresponds to the optimal LOR state feedback, and the subsystem Ginp leads to an optimal output estimation problem. Note also that the optimal cost is precisely the optimal state-feedback (i., LQR) cost plus the optimal filtering cost. 5.4 RELAXING SOME OF THE ASSUMPTIONS. ‘As we have just seen, the controller (5:33) minimizes Tz» under assumptions (A1)-(A4). While some of these assumptions are quite practical (e.g., (A1), required for the system to be stabilizable using output feedback), others may seem rather restrictive. In this section we address the issue of relaxing some of these assumptions, in particular, (A2) and (A3). Recall from Appendix C that assumptions (A2) and (A3) are sufficient ‘conditions for the Riccati equations (5.30) and (5.31) to have positive semi- ‘definite stabilizing solutions. However, these assumptions are not necessary for this property to hold. Indeed, it can be shown (see Appendix C) that assumptions (A2)-(A4) can be relaxed to the following: eat es (az) [ a ral has full column rank forall » (556) meen (a3) [ a Bl has full row rank for all » (557) (A4)) Dz fell column rank and Day full row rank. Assumptions (Al) and (A2!) are necessary and sufficient conditions for Hy ¢ 14214; OPTIMAL CONTROL dom(Ric), where Hz now denotes the following Hamiltonian matrix af ABR DEG -ByR'BY (558) me = CU DaR'DR Cy —(A ~ B2R'D},CI), and where R © DI.D,2. Similarly, assumptions (A1) and (A3') are necessary and sufficient for Jz € dom(Ric), where af (A-BD]S CG)? ISAC: ne [ (559) ~By(I- D}S“'Dy)B]_ ~(A- DESC.) and where $ © D;,D3, ‘Our previous results hold under these relaxed hypotheses, provided that F; and L; are suitably modified. ‘Theorem 5.3 Suppose that assumptions (Al) and (A2')-(A4’) hold. Let Xy Ric(H) and Y2 © Ric(¥2), and define RBI: + DEG) cay —(%2C] + B,DI)S"! Then Lemmas 5.5-5.8 and Theorem 5.2 hold. Finally, we want to comment briefly on assumption (A4’). The full rank conditions are imposed to avoid obtaining singular control or filtering prob- ems. The treatment of such problems is beyond the scope of this book (e.g., see [298]), although we will briefly discuss a special case, the “cheap control” case, in the next section. 5.5 CLOSED-LOOP PROPERTIES Recall that we motivated the 2» problem using an optimal control argument, where the goal was to optimize a performance index related to rejecting a class of disturbances. However, it should be clear that optimizing performance in a certain sense does not necessarily guarantee that the resulting system will be “optimal” (or even have acceptable performance) under a different set of conditions. Thus itis of interest to analyze the properties of the closed-loop system. In particular, guided by practical considerations, we are interested in assessing how sensitive the system is to modeling errors, since these errors are inevitably present in real life situations CLOSEDLOOP PROPERTIES 143 5.5.1, The LAR Case: Kalman’s Inequality ‘We begin by considering the simpler LOR case shown in Figure 53. We will establish the surprising fact that although robustness considerations were not taken into account in the design, the resulting system exhibits, in a classi- cal sense, excellent properties. In order to establish this fact we need the following preliminary result. ‘Lemma 5.9 (Kalman’s Identity) Let L(jw) = F(jwl — A)"'B be the state- edback loop, where Fis the optial LOR gain. Then the following identity holds fora [+ L(je)|" RU + L(ja)) =R+|ior—ay'8]' 0 [Gor -ay"8] 61) where Q and Rae the sae and control weighting matrices respectively Proof. See (98, 168} 7 Remark Since the second term on the right-hand side of (5.61) is positive semidefinite, it follows that f+ LGw)I" RUE+ LGw)] 2 R (5.2) Moreover, recall that without loss of generality (by redefining the control input if necessary) we can always assume that R = 1. Thus (5.62) reduces to + LGu)}" [1+ LGa)) > 1 (543) + LGm))" + LGe)) 2 t which is called Kalman’s inequality. Equivalently, if we define the sensitiv ity function S(ja) mm the usual way S = [I+ L(jo))"', then we have that SG} <1 (Le, the sensitivity is smaller than or equal to 1 at all fre- quencies). B (I-A) F Figure 5.3, Assessing stability margins for the optimal L@R controll. 144 4; OPTIMAL CONTROL In the SISO case (5.63) reduces even further to [1+ LGa)f 21, vo (564) Recall from Chapter 2 that the left-hand side of (564) isthe distance from generic point on the Nyguist plot of L(jw) to the critical point (1,0). Thus (5.64) implies that the Nyguist plot of L is excluded from a unit disk centered at (~1,0). In turn, this implies (see Figure 5.4) thatthe closed-loop system has an infinite increasing gain margin and a decreasing gain margin of at least §. Additionally, exclusion from the unit disk centered atthe critical point means that the worst possible phase margin corresponds tothe situation shown in Figure 5.5, where the Nyguist plot passes through the intersection ofthis disk and the unit circle centered at the origin. Thus the LOR design has a phase margin of atleast 60°. Both margins indicate that, from a classical standpoint, the LOR design has excellent robustness properties. Let us return now to the MIMO case. Recall from Table 2.1 that ||SjJ> <1 is a necessary and sufficient condition for robust stability against quotien type uncertainty, that is, G(s) = [1+ A]"' Gy(3), [|All <1. Thus we can ex- pect good robustness properties against this type of uncertainty (eg. arising from uncertainty in the location of the poles of the plant). On the other hand, itis easy to establish that [Sl < 1 implies that |]. < 2. Moreover, itis ‘not difficult to find examples where this bound is tight (e., see Example 32 in (181)). Thus the LOR design does nor necessarily have good robusiness properties against multiplicative uncertainty, since robustness can be guar anteed only for perturbations satisfying |All < }. Additionally, it can be shown [12] that T(jw) has the relatively poor roll-off rate of w~t, which may not be sufficient to compensate for high-frequency unmodeled dynamics. Imig») Re{Liw)} Figure 5.4. Nyquist plot lusting the goin margin of 10 regulators CLOSEDLOOP PROPERTIES 145, ImtLgwo) Figure 5.5. Nyquis pot ilusrating the phase margin of 10 regulators. 5.5.2, Some Consequences of Kalman's Inequality In this section we briefly discuss an important corollary of Kalman’s inequal- ity, the minimum phase property of the optimal LOR plant, and its impl cations for the cheap control case. For simplicity we will consider only the SISO case, but the results hold for general MIMO systems. From the inequality (5.63) it follows that the sensitivity function S(s) = [1+ f(s! —A)b|" is bounded real, that is, stable and such that [[Sllex < 1 Assume now that L(s) = f(sf —A)-*b is not minimum phase. Then there exist 5,, Re(s9) > 0 such that L(s,) = 0. It follows that $(s.) = 1. Since S(s) is analytic in the closed right-half plane and bounded above by 1 in the jw axis, it follows from the maximum modulus theorem ((224}) that S(s) = 1 in Re(s) > 0, which implies that f(s/ — A)-¥b =0. Clearly this is not possible (unless the state weight Q = 0). Thus we have established that the optimal loop function cannot have zeros in the open right-half plane. Note, however, that this proof does not rule out the existence of zeros on the jw axis. ‘As an application of the minimum phase property, we will analyze the case where the control penalty is of the form R = pl and p — 0. This is usually referred to as the cheap control case, since very little penalty is imposed on the control effort. Intuitively, it seems clear that as p —> 0 the control action increases and that the corresponding value of the performance index (53) should approach 0. However, as we will see next this is true only in the case where the plant is minimum phase. Setting R = pl and factoring Q = cTc, the equality (5.61) becomes {1+ LGu)} + Ga) ij)" GW) (565) 14614; OPTIMAL CONTROL where L(s) = f(s! ~ A)-'b and G(s) = e(sf ~ A)-*b, Cleatly, as p 0 this ‘equality implies that ty) 2 2 4 + fal ~ Ay 'b| > Lletjat - ay" 5. [+ fj! ~ Ay 'd| aici 1 — AY‘) (5.66) Since L(s) is minimum phase, it follows that its zeros are the minimum phase zeros of G(s) plus the mirror image (with respect to the ja axis) ofits non- ‘minimum phase zeros. Consider first the case where G(s) is minimum phase. In this case since G and L have (asymptotically) the same zeros, it follows that as p+ 0 then /pL(s) — G(s) and thus co evar (567) Rewritting the algebraic Riecati equation (5.17) in terms of f and c we have AIX XA4Te~ pf f = (5.68) From (567) and (5.68) it follows that as p ~+ Othen AX + XA — 0 and thus X10, Hence lim Jon(to) = lima Xx 0 (5.69) (On the other hand, if G(s) is nonminimum phase then cTe — pf? f #+ 0 and thus XX 0. It follows that for nonminimum phase plants there is a nonzero LOR cost J =] Xx, even in the case where the control action is not penalized. 5.5.3 Stability Margins of Optimal 71; Controllers ‘As we saw in Section 5.5.1, the optimal state-feedback Hz controlled system has very good robustness properties, at least in the case where the hypotheses of Section 5.2 hold (it can be shown that if R is not diagonal, then the gain margins may become arbitrarily small {181]). Motivated by these results, one would expect that some of these robustness properties will carry over to the general 4, case. Unfortunately, as we illustrate with the following counter- example (100), this is not true. Ed B IESG (570) y= oO} [a] CLOSED-LOOP PROPERTIES 147 where w and v are uncorrelated white gaussian noises with covariance and 1, respectively. Assume that we want (o minimize the following performance = nte{/ [x7Qx+u" Ru] a (71) o2e[!]iu a, o>on-t 6m It is easily seen that this problem can be recast in the general Hz form of Section 5.3 by selecting the following realization for G(s) “ I t] (6.73) “Tra, Bl oo) | fo 1) 0 In this case equations (5.30) and (5.31) can be solved analytically yielding the {following optimal gains: A= st 1) 1 la=d t] sw f S 24 rq) a+r) Hence the optimal coniroer is given by : ied od xe-[ eaats 4 es rf [oe To assess the gain margin of the resulting closed loop, suppose that the control ‘action is multiplied by a scalar gain m (with nominal value 1). Closing the loop with the controller mK (s) yields the following closed-loop dynamics matrix: 11 0 0 | 0 1 mf mf oe re (76) <4 0 dof 14s 148 Hp OPTIMAL CONTROL I can be shown that the characteristic polynomial of Aq has the form POs det(sl ~ Ag) = s+ ass? +435" + a)8 + a9 (6.7) where only the last two coefficients, a and ao, are functions of m, with the following explicit expressions: ay = (d+ f +4) +2(m— If % = 14(1—m)df Recall that a necessary condition for stability is that a; > 0. But from (5.78) it can be seen that, for sufficiently large d and f, ay and a, can be rendered negative by taking m=1%+¢, with ¢ arbitrarily small. Thus the closed-loop system has an arbitrarily small gain margin. Note also that ths situation hap- pens precisely in cases where the driving noise gets larger (large @) or the control weight relative 10 the state weight (1/q) gets smaller, contrary to the conjecture that either choice (large noise or small control weight) will result in improved “robustness.” (5:78) Since LOR systems exhibit good robustness properties, motivated by the separation principle, one may attempt to recover these properties by mak- ing the dynamics of the observer very fast. It can be shown 101] that this approach will not, in general, recover the desired properties. However, they ‘may be recovered in some special situations, most notably in the case where the plant is minimum phase. The main idea of this approach consists of in troducing fictitious plant noise with a specific structure (thus giving up filter optimality”). It can be shown that as this noise intensity approaches infinity, then the closed-loop LOG loop matrix K(s)C(s! ~ A)"B approaches the LOR matrix F(s/ — A)""B. This technique, first proposed in [101], is known a8 LQG/LTR (loop transfer recovery). In its original formulation, it was lim ited to minimum phase plants since the resulting controller inverts the plant (Grom the left) substituting it with the desired LOR dynamics. Ths fact sug. gests that the technique could also be used for “mildly” nonminimum phase plants, that is, cases where the norm of the transfer matrix is small at the frequency of the nonminimum phase zeros (16, 331] Finally, itis worth noting that this technique is not restricted to the case where the transfer function recovered corresponds to the optimal LOR con- troller, Indeed, any return ratio that shares the right-half plane pole/zero structure of the plant can be recovered [294]. This opens the possibility of recovering some other state-feedback structure that could be applied to non- ‘minimum phase plants, Note, however, that at this point, in addition to giving up filter optimality (by the introduction of the fictitious noise), we will also be giving up controller optimality. Thus, in this ease, the method reduces to just aan ad hoc procedure for trading off robustness versus performance, Rather than pursuing this approach further, in the next chapters we will address the issues of robust stability and performance using an H., approach. RELATED PROBLEMS 149 5.6 RELATED PROBLEMS. Before closing this chapter we want to briefly address the related general: ized control problem [249]. Recall that in the SISO case the M2 norm of a system can be given a deterministic interpretation as the £2 10 Las (ie, energy to peak) induced norm. Thus, in this context, the Ha problem ean be thought of as the problem of synthesizing a controller that stabilizes the plant and minimizes the worst possible peak valve of the output due to exogenous, signals with unit energy: In the case of MIMO systems, the two interpreta- tions do not coincide. Specifically, assume that the output is measured using the supremum over time of the spatial 2 norm. It can be shown [320] that S0P inj (89, l2C)fo} = VBISTTeeY where & denotes the maximum singu- lar Vale. Thus, while the standard Hz problem leads to the minimization of Jnug = trace(CW.C1), where W, is the closed-loop controllability Gramian, the generalized My problem leads to the minimization of Js = a(CaW.CI). ‘Obviously, these two indexes coincide only in the case where C,W.CT is a scalar, in which case Theorems 5.1 and 5.2 provide a solution also to the gen- eralized % problem. While this approach is no longer valid in the MIMO case, some of the tools developed to solve the standard %» problem can still be used, allowing for recasting the generalized problem into a linear matrix inequality form. This LMI problem can be solved efficiently using finite-dimensional convex optimization techniques. ‘As in the conventional 7 case, the optimal state-feedback controler is static, while optimal output feedback controllers have the same order as the plant and exhibit a separation structure. As an intermediate step in estab- lishing these results we will address fits the full information case Lemma 510 The state-feedback and FI generalized Hz problems are equiv- alent in the sense that there exist a linear static state feedback controller suck that Jil < y if and only if there exists a linear full information controller such that JE, < y [243]. Remark From this lemma it follows immediately that dynamic state-feedback controllers cannot improve upon the performance of static ones. Thus in the sequel we only need to consider static state-feedback laws. Theorem 5.4 There exists a static state feedback controller such that the generalized Hz cost 12 Jen = (3 / Tonle) yj) «)| o y>o Moreover, in this case @ solution is given by the state-feedback gain Ke=WY"', with the corresponding cost given by Jey = & (C+ DK.)¥(C + DK,)"] Proof. The proof follows from combining Theorem 32 in [249] with standard Schur complement arguments to recast the equations there into the LMI form a Next, we consider the output feedback case, As in the regular Hp case the solution will he obtained hy reducing it to a disturbance feedforward problem through the use of an appropriate output injection. ‘Theorem S$ Consider a plant G(s) with he following state-space realization: A| BB TG | 0 Da (581) Gl dn 6 ‘and assume that the following conditions hold: (Al) (A, Bz) is stabilizable and (C2,A) is detectable. (A2) Day has full row rank with Dy [BT DI] =(0 1) (a3) le a Bi) sf row rank forall Then the following statements are equivalent 1. There exis an internally. stabilizing. controler Kis). such that Ieen G(s), K(S)] < 9. 2. There exists a static state feedback controller K,, that internally stabilizes RELATED PROBLEMS 151 the plant G(s) and such that 3 [C\YaCP +8(G,p,Kyp)] <7, where [Ror “S"] = Ric) ly = -¥,C] =k, B: SUG Ky) = 3 | FAG y(F0) Ky Go)]FelGylio) Ky Go)}'do ae (582) Proof. By using the output injection r = Lzy, the plant G(s) ean be decom- posed into the two subsystems shown in Figure 5.6, where AtLaC | Bit Lan Gino) 0 Dy (583) -l, Gs) = |G] 0 Da alio Note that in this decomposition Ginp(s) is internally stable while G(s) cor- responds to a disturbance feedforward problem. Using standard arguments it ccan be shown that (e.g, see [249]), regardless of the choice of controller K(s), the following orthogonality property holds for S(G, K): S(G,K) = W¥aCT +5(G1,K) (5.84) 11 follows then that there exists an internally stabilizing controller such that Jyen < 7 if and only if there exists a controller K that internally stabilizes G(s) and such that & [Ci ¥2Cf + $(Gi,K)] < 7°. By construction G(s) has DF structure and hence (Lemma 5.7) is equivalent to a FI problem. The Proof follows now from the equivalence of the FI and SF problems shown in Lemma 5.10. o Finally, we want to consider briefly the case where the norm used for 152%; OPMMAL CONTROL . |) -— 2) G(s) Ke) Figure 5.6. Output feedback as he combination of @ DF problem and a stable system. ‘measuring the output is the standard £* norm. In this case it can be shown ((320)) that ‘sup {[[2|loo/llwl2:w € L2{0, 00), w #0} = y/daax(S(Tw)} (5.85) where dmax denotes the maximum diagonal entry. It follows that the same tools used to solve the generalized Hz problem with cost (5.79) can also be used in this case, by simply replacing & BY das 5.7 PROBLEMS \ iL Tne servomotor ofa tracking antenna canbe approximately represented by the following second order model: x)= [3 a s]20# [oer] 40 (586) where u(t) is the input voltage and where the states are the angular pos tion and velocity respectively. (2) Using ascal state-space tools design a controler that places te lose loop poles a 220.20 (b) Find the peak value of the control action for an initial condition x(0) (0 1)" (©) Find the optimal LOR controller corresponding tthe choice of weights o=[f 8) maR=2%10-%7 (a) Repeat part (b). PROBLEMS 183 3h prot ts some fhe pte pias of Opin Con theo Conder te flowing ae beh (a) Find the controller that minimizes: J [F (rox utRa) a, o= aes fri) * for q = 1, 10, 100. (©) In each case, plot the Nyquist plot of the combined plant controller and find the closed-loop gain and phase margins. (©) Assume now that instead of the states the output of the plant is x; corrupted by a white Gaussian noise signal and thatthe plant is also driven by white Gaussian noise w, that i: vem 4s (Y]4Bm «fo ide [i] | vy Find the optimal 2 controller that minimizes self (Coreatmjah ory aot for q = 1, 10, 100. (@) Repeat part (b). What conclusion can you draw about the gain and phase margins of the optimal system? 3. The issues involved in controlling systems subject to model uncertainty and multiple specifications can be illustrated by the simple system shown in Figure 5.7 consisting of two unity masses coupled by a spring with nominal value ky = 1.0 but subject to uncertainty. A control force acts on body 1 and the position of body 2 is measured, resulting in a noncolocated sensor actuator problem used as a benchmark several years at the American Control Conference (46). (a) Find a state-space model of the system. 154; OPTIMALCONTROL Figure 8.7. The ACC benchmark problem, (b) Design a stabilizing controller that meets the following performance specifications: (j) the control action following a unit impulse distur- bance w acting on m is constrained by |u(t)| < 1; and (ii) for the same disturbance the displacement y of mz has a settling time of about 15 seconds, (©) Investigate the performance of your controller when k varies between 015 and 2. 4,,A common performance specification is that the closed-loop system must have a prescribed degree of stability a, that is the closed-loop poles must, lic to the left of the line Re(s) ~~. Show that this ean be accomplished using a LO regulator that minimizes e*'2(0|lp rather than llz(). ‘8. Consider the optimal LOR problem for a SISO plant with Q = cc and R=p. Let z5,1=1,...,p and p,,i=1,....m denote the zeros and poles of G=a{s! ~Ayb. (@) Show that as p ~+ 0 p of the n optimal closed-loop poles approach z) if Re(z,) < 0 or 2, if Re(z,) > 0. (b) Show that the remaining n—p poles approach infinity as the roots of 320-7) = (-1)°-"1 8. This configuration is known as a Butterworth configuration of order'n ~p. (6) Show that as p — oo the optimal closed-loop poles approach p, if Re(p,) <0 or ~p, if Re(p,) > 0 6; Consider a discrete time system with state space realization U(z) = D+ YY Cal = Ay 1B. (2) Its conjugate is defined as U"(z) = U (2). Show that if A~! exists then ve=[_2 —— 5.87 @=lirav |p? latter (687) (b)_U(z) is said to be inner if it is stable and U(2)U(z) = 1. Assume that the pairs (A,B) and (A,C) are controllable and observable respec- PROBLEMS 155 tively and let X > 0 denote the solution to the discrete-time Lyapunov equation: ATXA~X+C7TC=0 Show that U(z) is inner if and only if: DTC+BTXA=0 (5.88) D'D+BTXB=1 (©) Use the results of parts (a) and (b) to solve the discrete-time LOR problem, trie ne yz soOle Hoo CONTROL 6.1 INTRODUCTION ‘As we indicated in Chapter 5, x large number of control problems of practical importance can he described using the hlack diagram shown in Figure 6.1 Here the goal is to synthesize an internally stabilizing controller K(s) such that the worst-case output z due to a class of exogenous disturbances w is kept below a given threshold. In this chapter we consider the case where w is an £y signal and where the design objective is to keep |z|2, the energy of the output, below a given level 7. Since for LTI stable systems the £3 to z induced norm coincides with the Ha, norm of the transfer matrix (see ‘Appendix A), this problem is known as the 7 (sub)optimal control prob- lem, Recall from Chapter 2 that this problem also arises in the context of achieving robust stability and robust performance in the presence of dynamic uncertainty, ‘The approach that we follow in the ist portion of the chapter, where we give a complete solution to a class of simpler problems, follows after (104) and is similar to the approach that we pursued in Chapter 5. As there, the solution will be obtained by first solving four cases having a special structure: full information (FI), full control (FC), disturbance feedforward (DF), and output estimation (OE). The solution to the general output feedback case will then be obtained by decoupling it into a FI and an OE problem. There is, however, an important difference between the approach pursued in Chap- ter Sand the one used here. While in the 12 case we looked for the optimal tent 158 11, CONTROL a) = [| Figure 6.1. General loop structure for the %,problom. ‘optimal controllers are both theoretical, since finding Hos optimal controllers is difficult (133), and practical, since these controllers may exhibit undesirable propsstics and their computation may lead to numerical difficulties, Never. theless, at the end of the chapter we Wy analyze the limiting behavior (of the suboptimal controllers at y | yo, the optimal achievable value 6.2. THE STANDARD 71,, PROBLEM Consider a plant G with the following state-space realization: eget A| BB a | 0 Da (6.1) WGI Dn Then the Hz. control problem can formally be stated as follows: given 7, synthesize an internally stabilizing, proper, LTT controller K(s) such tharthe closed-loop transfer matrix Tz» satisfies [Tywlla < 7. Note that, as in the %p case, we have assumed Hat Dj; = 0 and Do, =0. In the sequel, we will assume that the following additional conditions hold: (A1) (A,B,) is stabilizable and (C,,A) is detectable. ¢ (A2) (A,B) is stabilizable and (C;, A) is detectable (A3) C7Di2=0 and B,Df, = 0. (A4) yz has full column rank with DY,D,. = 1 and Dp; has full row rank with DayDJ, = 1 THE STANDARD 11, PROBLEM 159 Assumption (A1) is clearly necessary for the system to be stabilizable via ‘output feedback. AS in Chapter 5, assumption (A2) is made for techni cal reasons. Together with (Al) it guarantees that the control and filtering Riccati equations associated with a related #2 problem admit positive sem definite stabilizing solutions. The rank assumptions jn (A4) guarantee that the Ha, problem is nonsingular, while the normalizing assumptions (D/-D 3 = 1, DDB, =1) do not entail any loss of generality since they can always be ‘met by redefining the inputs w and w if necessary. Similarly the orthogonality assumptions (A3) are made for simplicity. They can be relaxed at the price of introducing additional terms in the expression for the controller. Finaly, the assumptions Dj = 0 and Dz = 0 are also made for the sake of simplic- ity. As we indicated in Chapter 5 they can be relaxed through the use of loop shifting (264). However, the resulting formulas are considerable more complicated than the ones corresponding to the simpler case treated here. Paralleling the development in the 7 case, we introduce the following two Hamiltonian matrices: ee ee r na [ ae Y mals PT] eae Te ~ 2 ya] 47 rda-da . = [BB ee Note that in contrast with the matrices H and J; defined in (5.32), the (1,2) blocks in Ha, and Ja. are not sign definite. Thus we can guarantee a priori neither that Ha.,Ja. € dom(Ric) nor that Ric(H..) > 0, Ric(J~.) > 0, Indeed, as we will see in the sequel, these conditions are related to the existence of suboptimal %,. controller for the given attenuation level y. Note also in passing that as 7 — oo then Ha, —» Hz and Ju. ~* Jp ‘The next result, originally from [104], gives a complete solution to the Ha. problem, ‘Theorem 6.1. Under assumptions (Al)-(A4) there exists an internally stabi- lizing controler K(3) that renders Teall <7 if and only ifthe following conditions hol: 1. Hay € dom(Ric) and Xx. © Ric(H) > 0. 2 Ing € dom(Ric) and Ya, 2 Ric(Hy) > 0 3. [Xo Yoo) < In this case the set of all internally stabilizing controllers rendering \Twlloo < can be parametrized as K(s) = Fr(Max,Q) (63) 160 Ma. CONTROL Where Q € RH \IQlae < 7, and Mus has the following state-space realization Aw | -Zabw Zubr Ma = | Fy oO; (64) Oy 10 and where Au = A+ Y BBY Xe + BaF + ZckaCe Fy = ~BEXe _ 65) Zu = (1-7? VanXno) Remark The special controller corresponding to the choice @ = 0 in (6.3) is known as the “central controller.” This controller has the following state- space realization: Ax | ~Zooks 66 re 0 (6.6) Hence its McMillan degree does not exceed that of the (generalized) plant Gs). AAs in Chapter 5, this result will be established by decoupling the output feedback problem into ful information and ouput estimation subproblems. “Thus the first step is to consider the same four intermediate special problems (Fl, DE,FC, and OF) used there. However, before proceeding, its instructive to analyze the structure of the central controller Kea 10 gain some insight into the problem. As in the 2 case, this controller exhibits a sep structure. This can be made apparent by rewriting its state-space realization = Ais By (7 ?BT Xi) + Byu+ Zon Cot 9) = em anh We can sce now that the general frm of the controle does indeed resemble an obsererbed controller, with afew notable diferenes, namely, the we Of ZeL rather than Lay alone and the appearence of the By (y-7BT Xai) term. As we will show next_y-?B] X.ox is a worst-case disturbance for a related FI problem. Thus hiv Extra Term can Ge Though OF ws ‘stimate of this worst-case disturbance based on the present estimate of the sate Similarly. Za the optimal ler pain for estimating the optimal Fl coml input, = Fin the presence ofthe worsease dxturbance. To show ths assume thatthe following Riceti equation admits solution Xo AT Xoo 4 Xoo + CPC, + XagB BY Xo — XouBBEXqy =O (6.8) THE STANDARD 74x, PROBLEM 161 Note that Xzc = Ric(Ha.) when y = 1. This can always be assumed without loss of generality by using the transformation B, —+ y-'B, if necessary. Thus, in the remainder of this chapter, we will set y = I for the sake of notational simplicity. Consider now the quantity x” (/)X.o¥(t) and its time derivative. Using (6.8) together with the orthogonality of Cx and Dyzu, it can easily be shown that SP OXax(0] = AT OXeAO + x7 OX 37(0) (AT Kae + Xe) x(0) +207 (DBT Xsex(t) +2u" ()BEXnox(0) = lel? + fet BE Xaoxll? = BY Xoor |? + 207 (BY Xox(t) (69) A completion of the squares argument yields Ef OXeH(O] = HELO + mCP + ule) + BEXWAOIP 630) = Ihe = BF Xoo? Finally, integrating this last equation, assuming that x(0) = x(co) = O and that w € £4)0,00), yields ell — Wwll3 = Ne + BY Xoo lw ~ BY Xoox 3 (6.11) los tat wang the motte tuba inte sere ht mai mizes |\z\[3 — ||wi[. On the other hand, v — ~B7 X,..x is the optimal control ae a aoe toes quanln We se hess ‘Thus these values of u and w"Salisly @ saddle-point type condition. Note also that this suggests that when the sates are wvall@bie Tor feedback then tte Fac is indeed the optimal contol ation, As we wil seein the sequel, ths is predely the eae 62.1 Background: Hankel and Mixed Hankel-Toeplitz Operators In this section we provide some background material required to solve the FI x control problem. While this problem can also be addressed using a game- theoretic approach [136], the approach pursued here (which follows after {104)) highlights some connections between Mx. control, Hankel operators, And the problem of approximating an anticausal system with a causal one. Additionally, some of the tools developed here will be used again in Chapter 9'when we deal with the problem of model reduction. 162M. CONTROL In order to introduce the Hankel operator, given a stable, strictly proper sen a on [444] a and a point xe, consider the problem of finding the minimum energy input v(t) € £3(-00,0) such that ifx(o0) =O, then x(0) ~ Xe, that is, weftag ll (613) subject to: x(0) = x9,x(—c0) =0 Itis a standard result that if the pair (A, B) is controllable, then the solution to the problem is given by Wags = BTA Wot, (614) where W, > Os the controll ity Gramian of G. Moreover, the optimal cost [opel = x2 Wolo (615) (On the other hand, consider now the free evolution of G from a given initial condition x,. Again, itis a standard result that WzlE = WCxI = x2 Wate (6.16) where W, is the observability Gramian of G. Consider now the problem of finding the past input w € £,(—oo, 0] that maximizes the energy of the future output z(t), € (0,00), that is, Zi max 617 wey(" 0 0|a40_[fw)fh ey where P. :£,(~00,00) -+ £[0,00) isthe projection operator, that is, = {fF «20 P.flt) = { Wares 618) By combining equations (6.14) and (6.16) it follows that this problem is equiv- alent to ; Wel _ a 22 Wote wetr(ocetlwt® [wll 20 xT Wey This last equation can be given an interpretation in terms of Hankel operators! as follows, WW) (6.19) "These operators have historical importance since they were used inthe eatly 19805 to furnish a solution to 7, control problems via a reduction Loan equivalent Nehari approximation problem (see Section 66). A full treatment of Hankel operators i deferred until Chapter 9, whete they wil be discussed in the context of model rection. Here we provide an abbreviated teatment introducing some ofthe properties relevant tothe connection with H~ conta. THE STANDARD 21x, PROBLEM 163 Definition6.1 Given thesystem (6.12), ts Hankel operator Ug: £(~c0,0] ~+ £4(0,00) is defined by = 7 alt) =Pov 8 {ff Cent MBvayar 120 (620) ‘Thus ['g can be thought of as mapping the past input v(¢), 1 € (~20,0} to the fotuse output z(1), € [0,00), via the state x(0). The adjoint operator Ty is defined by the relationship (y.Pov) = (vit ay). ¥ € £2(-00,0), ¥ € £2[0,00) (621) It follows that Ty :£2[0,00) —+ £3(-00,0) is given by Be Tere heey aa 2S0 (622) Next, we show that the eigenvalues of [';I'q coincide with the eigenvalues of W.Wa. Consider 4 nonzero eigenvalue o of PeIg and let u € £3(—00,0) be the corresponding (right) eigenvector. Define 0 = 0” 'Tou. Then we have that Tou = ov (623) Tyv = ow ‘The pair (u,v) is called a Schmidt pair of Vg. Using the explicit expressions (620) and (6.22), we obtain CoM, = ov (628 BTe*li, = ou where we defined wo 2 fl et aunas = (625) “8 [rorctune Premultiplying the first equation in (6.24) by e*”"CT and integrating from 0 to ae yields [MCCA gd = Wore = 0%, (626) 164. CONTROL ‘Similarly, premultiplying the second equation in (624) by e*"B and inte- grating from ~oo to 0 yields fie where we used the fact that “BBM eA, di = Woke = 0x5 (62) 7 [emaate a = [Perper am 628) L. fo Combining (626) and (627) we have that WeWaty = 07x (629) ‘Thus an eigenvalue of [Tg is also an eigenvalue of W.W. Reversing the argument above using (614) shows that II'¢ and W.W, have precisely the same eigenvalues. Definition 62 The square roots ofthe eigenvalues of WeW, are called the Hankel singular values of the system G(s). ~~ Since the £2 -+£; induced norm of Ig is given by its largest singular value, we have that | I” — p(W-W,). Additionally, this ean also be used to define a norm as follows. Definition 63 The Hankel norm of a stable, stricly proper system G is defined as NG()Iiv = sup e'?(WeWo) (6.30) wetst-coo [lle ale where g(t) denotes the impulse response of G(s). Next, we explore several implications of |[Gil. < 1. The first one, usually known as the bounded rea! lemma (10) links this condition to a Hamiltonian matrix Hf and its associated Riccati equation. The second establishes a rela- tionship between the output norm achieved by a worst-case perturbation and the solution to this Riceati equation. Lemma 6.1_ Consider a proper, stable system G(s) = C(s1 ~ A)'B + D with ‘F(D) <1. Then |\Gilw <1 ifand only if H € dom(Ric) and X = Ric(H) > 0, where a 4+ BRD? \gt i A+BR'DTC BROB' ] CM+DR'DTVC ~(A+ BR DTC)? 31) and R21— 7D. THE STANDARD Hao PROBLEM 165 Proof. Define ®(jw) £1 - Gjw)G(jw). Then IG(s)c <1 — @Gw) >0, Yo (632) It can easily be seen that a state-space realization for ® is, (633) Since A has no eigenvalues on the jw axis, this realization has no uncontrol- lable or unobservable modes there. Then there exist a frequency w, and a vector uo such that (jor )up =O if and only if there exists ve such that [fe Sr]e[-Ep]etere ates = jan [ A+BR"DTC BR"'BT | aad T+ DR“DT)C (A+ BR'DTC)t | Ye = Javo ‘Therefore @(s) has no zer0s on the jw axis iff H has no eigenvalues there. Thus det I — GGw)GYa)] £0 for all w iff H has no eigenvalues on the jo axis, Since [f - Gja)GLje)] > 0 as w — 00, we can conclude that © (jw)G(ju) > 0 allo. I follows then that [Gill <1 <= H has no eigen: values on the jw axis, The fact that H € dom(Ric) follows now from Lemma C2 and the fact that (A + BR"'D™C, BR”) is stabilizable since A is stable. Finally, to show that X = Ric(H) > 0, start by rewritting the corresponding Riocati equation as (A+ BRAD™C)TX + X(A+BRDTC) +XBR“BTX +C7(I+DR"'DTC) 0 (635) ATX 4XA + (CTD +XB)R\BTX + D™C)+CTC = 0 Thus X is the observailty Gramian for (Atct (wascroyR”y") Since A ie stable i follows that X > 0. a Lemma62 Assume that G(s) is strictly proper with |Gllay < and x(0) Then sup (8 —twlB) = 2x6 (636) where X = Ric(H), 166 1, CONTROL Proof. Ditterentiating the quantity x7 (¢)Xx(0) along a trajectory and using the Riecati equation associated with 17 yields 4 fron] = exe soxe9 210 (ATX 4 XA) x(0) +267 (DBT XX(0) lzll} — |B Xx|]3 +207 BT Xx (637) ‘Completing the squares yields a 2 § [Xa] = “eC + bool? = he — BaP (638) Finally, integrating this last equation yields (since w € £2[0,00) and thus x € £2)0,00)) lel — Nw = 22 Xx — Ww — BTXxIE < xP Xx6 To complete the proof note that the disturbance w* & BTXel4+F8' Nx, is in £3[0,00) since A + BBTX is stable, and that for this disturbance the inequality is saturated. a We will consider now a generalization of the Hankel operator deseribed czrlier in this section. Suppose that the input to the system G is partitioned in w= [11 w2], where w € £2(~00,0] and w» € £3(-~00, 00), and that the ‘matrix B is partitioned conformally in B = (By By]. Define the space we{[n] wy € L4(-00,0), ve Lal-0,00)} and consider the following operator I: W ~+ £,{0,00) defined by 7 [] 2 where P,£2(—00,00) -+ £2(0,00) denotes the projection operator. T can be interpreted as the sum of the Hankel operator Tg and the Toeplitz operator P.G2P.. AS in the case of the pure Hankel operator, we are interested in finding the worst-case perturbation, in the sense of maximizing |2[], and ‘bounds on the corresponding value of the output 1G) Ga) (m] (6.40) Lemma 63. Assume that (A, B) is controllable. Then supzcay IF wlla <1 iff the following two conditions hold: (i) Hy © dom(Ric) and W = Rie(Hy) > 0 where A at 4 Gi) WW.) <1 THE STANDARD 21. PROBLEM 167, Proof. From the bounded real lemma (Lemma 6.1) we have that (i) is a necessary condition for having sup,esy lw f2 <1. Thus, to establish the desired result, we need to show that piven (), (i) holds iff sup, gy Fa < 1. By definition of W, for any w © W we have that WP2i88— Iwi = P21 ~ |IPowiB — IIPwil3 WP.zIK ~ IMP woll — NPwilp (41) Since the last term contributes to [[P,z|[3 only through x(0), the state at ¢ = 0, and since Wz is invertible, we can use (6.14) to solve explicitly for P_Wron where Wwons is the disturbance that maximizes the right-hand side of (6.41) ‘Thus combining (6.15) and Lemma 6.2 we have that sup {UP.2I3 ~ Ibi} = sup. {IPI ~Bwf:x(0) = xo} = sup {2x - xDWeolxe} (6.42) To prove the desired result, assume by contradiction that there exists w* € BW such that |\Pw*l|2 > 1. Then ||P,z|3 — Ilw"|3 > 0, which, together with (642), implies that p(WW.) > 1. Conversely, if p(WW.) <1 them the left- hand side of (6.42) is negative for all w, which iaiplics hat [Pw lip <1 for all we BW. a Finally, before closing this section we want to consider I*, the adjoint operator of T. We will use this operator, together with the property that UT] = |IF"], to solve the full information case. From the definition of T (6.40), we have that given any z € £[0,00) and w = ie) ew (z,Dw) = (2,P.(Giws + Gowe)) (2.Gimi) + (2, Gama) = (P_(Gire) mi) + (Gee, w2) = (2) (643) where P_ denotes the projection operator from £200, 00) to £2(-~00,0). It follows that 5 Se] [" or]. on With the tools introduced in this section we are ready now to give a solution to the FI Ha, control problem 2 hese inequalities are easily verified sing the frequency-domain expression for he inner prod ein 0 ayaa) = 2 fe lea) de 168 H.,CONTROL Full Information Problem Lemma 64 Consider the FI case, that is, the case where (645) Then the following results hold: 1. There exists an internally stabilizing controller such that Tell <1 iff Hoy € dom(Ric) and Xu = Ric{Hze) 2 0. Moreover, one such controller isKri=(Fe 0}. 2. The following family of controllers, Kri(s) : Kei(s) Fo ~ Q(s)B]Xeo O(s)], O(s) € RHao, ||Olloo <1} (6.46) inernally stabilizes the plant and renders \Teul <\ Proof. (Necessity) Assume for simplicity that the pair (C;,) is observable, rather than jst detectable (see Problem 6.1). Asia the Mz kas, define & new contol input v 2 u(t) Fox(), obtaining _ [An |B Ar, | Bo rans ete Lette = GAs)By + U(S)Tow(s) (6.47) where BT [a] Xz = Ric(H,) >0 P= BIN, C An = A+ BF Ci, = C+ DF Recall (rom Lemma $2) that U is inner. Thus there exists U, € RH. such that (UU, is unitary. For instance, it can easily be shown that one such THESTANDARD 71x. PROBLEM 169° transfer matrix is An, | -XSICTDs ue (649) Cn Di where D, is any matrix such that [Dy D,.] is orthogonal. Suppose that there exists an internally stabilizing controller that renders ||T;w||a0 <1. It follows then that sup inf lle <1 650) wea yell Ce Since the 2-norm is invariant under multiplication by unitary transfer matrix, we have that Wel = WU aT eI = | [UGB +0 ~ |i wrG.Byw b e 5 \[" Cen eae (651) ULC.Bw From this equation it is apparent that the minimizing control is v= =P.(U'G.Biw) and that sup inf |b = sup [Pee] BEI) "€C2100) wesespo lL UGB |p sup (ee wr wEBLa10,20) | ULGBy l2 = ap [Puig (652) eB) where I is the adjoint of the mixed Hankel-Toeplitz operator defined in (640). Thus we have that sup inf sles <1 <> sup wi <1 wc BC;10.00) 76221020) weB E2100) = sup Pwr <1 (653) where the last line follows from the fact that {fj = IP], and where nes caroru arocuy[]=nazery vsi[m] (30 170 Hs. CONTROL ‘Simple but tedious algebra using the state-space realizations of U, U, and G. and the formulas in Appendix A shows that a state-space for B?G."(U Us| is given by BIGo|U Us An | Br -Xy'CPD. [ : ] (655) -BIx, [0 0 Combining this expression with Lemma 63, we have that Hw € dom(Ric) and that W = Ric(Hy) > 0, where An Xp'CFCXs! tw [Lasatary, “S| ed Moreover, X;" is precisely the controllability Gramian of (6.55) since it sat- isfies the Riccati equation ApX;! +X31AT, + BBY + X5'CTC.X;! =0 (657) ‘Thus it follows from the second part of Lemma 63 that p(WX5') <1, or equivalently, Xp >W. To complete the proof note that the Hamiltonian matrices Ha, and Hy are related by the similarity transformation Hy: = THyT-!, where “1x3! [x 8] 1 follows that Ha, € dom(Ric). Moreover, using the facts that ay) =m] (658) and % (Hoo) (659) wwe have that Xz = Xo(I ~ Xp'W)-! = X9(Xp— WX, > 0. (Sufficiency) We will show first that if Hx, €dom(Ric) and X, = Ric(Hoo) 2 0, then u=—BYX..x = Fax stabilizes the system and renders |Tzwlloo <1. To establish this, note that since H., € dom(Ric), then A+ B,B] X., ~ BzB] Xx. is stable. Denote Ap, 2 A ~ ByB}X., and Cy, © Cy + DizF'n. Then the Riccati equation associated with H.. can be rewritten in temms of Ap, and Cr, as Al Xoo + XocAp, + XxoBiB] Xue + CE Cp, = (6.60) where we used the fact that Df,[Ci Diz|=[0 1]. Moreover, since A~ ByBTXa,+ BBY Xw is stable, a simple argument using the PBH test THESTANDARO Ha. PROBLEM. 171 shows that the pair (BYX..,Ay,) is detectable, which, together with (6.60), implies that Ap, is stable. This fact, combined with the bounded real lemma (Lemma 6.1), implies that the corresponding closed-loop transfer function Ar, | Bi rays [ed] wan is stable and such that ||Tsjfao <1 Finally, to show that the set of controllers Kry ={[Foo~ Q(8)B] Xoo Q(3)],O(s) € RH sey [IO <1} is admissible, use again the change of variable u = F.,x + v. Consider now a controller from the family Kry. In terms of w the control action is given by e=[-90)87% 9101[5] = 0191-87 ali] Clearly, closing the loop around G(s) with the controller Key is equivalent to closing the loop with the controller Q(s) around a system having the following state-space realization: Ar | Bi Br P0)=|—G | Da (662) -Bie | 1 0 This is illustrated in Figure 62, where r= w — BYX,ox. Note that equation (6.60), together with Lemma 5.1, implies that P is inner. Moreover, using the formulas for system inversion given in Appendix B, we have that 1 ir, + B\BTXeq | —By Pal) = | i 663) PX Thus P3)\(s) € RM It follows now from Lemma 3.10 that Toy € RHao with WTewlloo <1 iff Q € RAs2, [|Olleo <1. Thus the family Kp, has the desired Properties. o I can be shown (see [104] for details) that the family K;, parametrizes the set of all admissible control actions, albeit it does not generate the set of all admissible controllers. This can easily be seen by considering a static controtler of the form [K 0] and such that the corresponding |[Tsulln <1 Clearly this controller is not a member of the family Ke, for any finite Q(s) (unless K = F.). Recall that a similar situation arose in Chapter 3, in the 172 Ha CONTROL Figure 6.2. Loop transformation for proving he Fl case. ‘context of stabilization. As there, the fact that the family Kr, generates all possible control actions will be sufficient for our purposes. Remark Note that the central (Q =0) FI controller uses only feedback from the states. Thus, as in the 719 case, under the assumption that Dy, = 0 the FI and state feedback are, in some sense, equivalent. ‘This 1s no longer true when D3, # 0. Note also that, contrary to what happened in the M2 case, the Ma» controller depends on By through Hoo. By exploiting duality, we immediately get the following result. Full Control Problem Lemma 65 For the full control case Gres} (6.64) the following results hold: 1, There exists an internally stabilizing controler such that |Toellx <1 iff Jun €dom(Ric) and Ya, = Ric\Jq.) > 0. Moreover, one such controller ; we-[] THESTANDARD 71, PROBLEM 173, 2. The following family of controllers, {arcto)Bre= ['*~ Fel], (9) €Rikay Oe < 4 Os) (665) internally stabilizes Gre and renders |Teull <1 Next, we consider the DF problem and its dual, OF. Disturbance Feedforward Problem Lemma 66 Consider the DF problem A|B B Gorls) 0 Da (666) GQir 0 where, as before, we make the additional assumption that A ~ B,C. is stable. Then the following properties hold: 1. There exists an internally stabilizing controller such that \Tzullox <1 iff H.. € dom(Ric) and Xx, = Rie(Ina) > 0. 2. The set of all LTI controllers such that |\Tzullo. <1 can be parametrized Kor(s) = Feldpr,Q), QE Ria, [Ilo <1 (667) where Jor= (6.58) Proof. As in the H case, the proof follows from the equivalence of the FI and DF problems under the additional hypothesis of stability of A ~ By C Recall from Chapter 3 (Lemma 33) that under this additional hypothesis the FI and DF problems are equivalent in the sense that if Kop stabilizes the DF plant then Ker © Kor [C2_ 1] stabilizes the FI plant. Moreover, this case F; (Ger, Ker) = Fe(Gor, Kor). Conversely, if a controller Kr,(s) admissible for the FI plant, then Kor © Fe((Por, Ky) internally stabilizes Gop and Fy(Gri Kes) = Fe(Gor, Kop). Item 1 and the frst part of item 2 follow now from Lemma 6.4, by using the explicit expression (6.46) for Ky) tn the same sense as in Chapter 3 that isthe set of controllers generating all possible control 174 Ha. CONTROL and noting that Fy(Por, Kr) = Fang.) Finally, to show thatthe family (6.67) parametrzes all the DF controler, proceeding asin Chapter 3 given an admissible controller Kor(s) define @ = Fi(Mf,Kpr), where A+ B,BIX., | By By Fe | oF (6.69) G+BTX. | 10 Using the properties of the interconnection of LFTs given in Appendix B it can easily be shown that A+BrFo- BC: BaF Bs =BiC;—B,BT Xe A+ B,BTX B, Fi pp Mt) = (6.70) F, Fe a oT C~BYXno Co+ BEX 10 Using the simiarty transformation -[to reli] and eliminating the uncontrollable and unobservable modes from the real- ization yields “, 0 A+BzFo+ByBIX. | 0 0 eres 0 G+BT Xe | ro 1 [2 4) tn where we used the facts that A—ByC) is stable by assumption and A+ BrP, + BiBT Xe is stable since Hx. € dom(Ric). Thus FrUpr,Q) = 7 [lor Fett, Kor = FilFeJor,M), Kr] = Kor. To complete the proof we need to show that Q € RH.» ||Ollw <1. To this effect consider the in- terconnection shown in Figure 6.3. Simple algebra using the formulas in ‘Appendix B shows that F,(Gr,Jor) has the following state-space realiza- tion (where once more we have used the similarity transformation T and climinated uncontrollable modes exploiting the stability of A+ B2Fa, and A~ BC): A BaF. By B BiCy_A4+ BaF —ByCr | By Bz FuUGordor) = | Ord Dy G -G-BixX, |1 0 THESTANDARD 71, PROBLEM 175 Gog w 7 tor : Gor « y a Kor y “ Kor Ql Figure 6.3. Interconnection for solving the DF case, AtBiFo Bib 2 G+Dufe Dale “BT Xe -G-BFX_ | 1 0 A+ BF = | C+ Dike BT Xu BB 0 Dp (672) ‘Thus it follows from Lemma 3.10 and equation (6.62) that Tyw € RH, NTewlleo <1 =F QE RAce; IQlloo <1, a ‘As before, the solution to the OE case follows from duality. ‘Output Estimation Problem Lemma 6.7 For the OE case where G has the form + A= B,C; stable (673) the following properties hold: 176 a, CONTROL 1. There exists an internally stabilizing controler such that |\Teulln <1 iff Jas € dom(Rie) and Yuu = Ricllnc) > 0. 2. The set of all LT1 controlters such that Talla. <1 can be parametrized Kor(s) = Fdor.Q), QE RH, [Olle <1 (674) where AtLwy ¥.CP Joe = | (6.15) 6.2.2. Proof of Theorem 6.1 In this section we exploit the results of Lemmas 6-6,7 to prove Theorem 6.1. Motivated by the min-max problem (6.11) make the following change of variables: v= Ws BEX gx a) w= BTXx leading to the interconnection shown in Figure 6.4, where P and Ginp have the following realizations: A+BrFo | Bi Bs P= lGsDake | 0 Da “BIXe | 10 BB (677) Gun oT Dy 0 Porn pe and where we used the fact that D2, BT = 0. Since P is inner with P;;! € RH, We have that Tow € RMaoy[|Toxlle < 1 <=> Fe(Ginp K) © Roo, llFe(Gimp, K)lleo < 1. Note that Gimp has an out- put estimation form. However, in order to be able to apply the OE results, ‘we need to establish that Giny satisfies the corresponding assumptions 1. (Aimy, By 8 Sabilizable and Aygp + BaF: i stable 2. (Cy, Amp) is detectable. 3 [2 ]on-()] TNE STANDARD, PROBLEM 177 K Figure 6.4. Loop transformation for reducing OF to DF. Note that stabilzability of (Amp, B:) follows immediately from the stabiliz- ability of (A, B)) using the PBH test, while stability of Amp + BaF’ follows from the fact that H.. € dom( Ric) Finally assumption (A3) is inherited from the corresponding assumption for the general H. case. Thus we only need to Show thatthe pair (C3, Aim) i detectable. In the sequel we wll establish this fact by showing that an equivalent condition holds. To obtain thie condition start by considering the Hamiltonian matrix associated with the OE problem Gimp (6.78) Recall that as an intermediate step during the proof of the FI case we established that H. € dom(Ric) and X,. = Ric(H..) > 0 are sufficient con- ditions for the stability of A BsBYX~,. By duality we have that inp € dom(Ric) and ¥inp = Ricimp) > 0 are sufficient conditions for the stability Of Aimp ~ YinpCf Ca. The detectability of (C2, imp) follows now from the PBH test. Proof of Theorem 6.1. (Sufficiency) Using the Riccati equation for Xzo, it can easily be shown that Joy and Jip are related by the following similarity transformation: refh F] 6) Thus Jinp € dom(Ric) + Jn. € dom(Ric). Using the facts that 20m =tm[ 1 ] «6m 178 4. CONTROL dg) Ta) =Tim[ yl =m|t $M] ay we have that Ying = Yoo! ~XueYou)! = ZxY¥o. Moreover, since p(XoeYao) <1 then Yin > 0, which implies that (C>, Aimp) is detectable. Since Ginp satisfies all the OE assumptions it follows from Lemma 67 that all controllers that stabilize Gin and render ||Tylao <1 are given by K = FU,0) Fi(Ma,Q) (6.82) where we used the definitions (6.5) and the fact that Ying = Zo-Yo. (Necessity) Consider now an internally stabilizing controller K(s) such that | Tes[|. < 1. The controller K(s)[C D2i] solves the corresponding Fl problem, Thus from Lemma 644 it follows that Hq, € dom(Ric) and Xx) = Ric(H) 20. From duality it follows that J4.€dom(Ric) and ¥.. Ric(Hq) 2 0. Since Xe >0 we have that P is inner, with Px, € Rit, ‘Together with Lemma 3.10 this implies that Fu(Gemp, K) € RU, |\Fe( Gimp, K)foo <1, that is, the OF problem associated with Gimp is solvable. It follows that the Pair (C2,Anmp) i detectable, Hence Gy satisfies all the requirements for the OE case. From the output estimation results (Lemma 6.7) it follows now that IFe(Gimp, Koo <1 <=> inp © dom(Ric), Yip = RicJimp) > 0. Using the Same silanty transformation used inthe frst part of the proof, we have that Yip = (I~ YouXoo) "Yue > 0 (683) To complete the proof we need to show that p(X ¥x) <1. To establish this fact consider first the case where Yoo >0. Since ¥PXu(¥L)" 20 it follows that if pln" Xau(¥4")'| =1 there exists a right eigenvector o sueh that ¥2°X,.(¥.") v= v. Define now w = ¥z'y, We have that (J — Xa.¥a.)w = wand thus ¥,.w = 0 against the assumption that Y¥, > 0. Hence RELAXING SOME OF THE ASSUMPTIONS 179 plYs? Xa ¥42)"] <1 + plXn You) < 1. Suppose now that ¥,, > 0 and con sider its singular value decomposition: 7r[¥u 0 vcu'[Me ao an [Xn Me aoor[te BE] «6 aesases eeeeeaaeaee anes we eau Yu oe 20 (6.86) = (Xun) <1 (6.87) ‘where the last line follows by applying the argument above to the (1,1) block of UYiapU™. The proof foliows now by noting that p(Xi1¥i1) = p(X Yn). Remark As in the 212 case, the central 114, controller exhibits a separation structure. However, this structure is different from the one we observed in the M% problem. Here the controller is estimating the optimal full information, control gain in the presence of the worst-case disturbance Ween. The gain Zool corresponds to the optimal filter gain for estimating u = F..x. The corresponding formulas in the > case are simpler since in the corresponding FI problem there is no worst-case disturbance. Moreover, in this case the problem of estimating any output, including the optimal state feedback, is equivalent to estimating the states. 6.3 RELAXING SOME OF THE ASSUMPTIONS, As in the Mp case it is interesting to analyze whether or not the set of assumptions (A1)-(A4) can be replaced with a set that is less restrictive. As before, assumption (Al) is necessary for the existence of stabilizing con- trollers, while (A4) guarantees that the problem is nonsingular. On the other hhand, assumption (A2) and the orthogonality conditions can be relaxed. Re- call that assumptions (A2) and (A3) are sufficient conditions for H3,2 € dom(Ric), with X= Ric(Hs) > 0 and Y = Ric(Jz) > 0, properties that were used to establish the FI and FC results. However, assumptions (A2) and (A3) are not necessary conditions for these properties to hold. Indeed, it can be shown (see Appendix C) that these assumptions can be relaxed to the fol: lowing: (22) [A cEh BE] sat tame ak eesti 180%. CONTROL [A—jol By cay eh J bs at cow rank for aw (6.88) ‘Assumptions (A1) and (A2’ are necessary and suficient conditions for Hs € dom(Ric) and X> = Ric(H1,) > 0, where Hz denotes now the following Hamil- tonian matrix: A-BR DEC, BR 'BE CTU DaR "DEC, (A ~ BR DICT) (6.89) and where R © D¥.D,p. Similarly, assumptions (Al) and (A3’) are necessary and sutficient for Jz © dom(Ric), and Yp ~ Ric(J) > 0, where i [ (A~BSDI.CyT ~cIstG ] on Bi — DE S“"'Dy)BY_ ~(A ~ BS1DE.Cy with $2 DyDf, On the other hand, the conditions Dy, = 0 and Dz, = 0 can be relaxed by using a loop shifting technique (eg. ee [133] or [264). However, the corresponding formulas become rather involved, As an alternative, in the next section we pursue a LMI-based characterization of all suboptimal Hi controllers 6.4 LMI APPROACH TO 71,, CONTROL ‘As we indicated in the last section some of the simplifying assumptions that we made in Section 6.2 can be relatively easily removed, albeit at the price of more complicated formulas. On the other hand, assumptions (A2') and (A3), namely, the fact that Pz and Pz; do not have invariant zeros on the extended jw axis, cannot be removed easily within the framework used so far. Itis worth emphasizing that many problems of practical interest do actually Violate these assumptions. An example will be the case where it is desired to have perfect rejection of persistent disturbances with a known frequency wo, including as a special case w, = 0, that is, integral control, According to the internal mode principle, this perfect rejection necessitates including a model of the disturbance in the generalized plant. However, this will lead to zeros fon the jar axis at w = oo. In this section we use simple linear matrix inequality manipulations to ‘obtain a convex LMI characterization of all suboptimal 1. controllers, This approach does not necessitate making additional assumptions on the plant, beyond the usual stabilizability and detectability assumptions (A1) and leads 0 convex optimization problems that can be solved efficiently. Moreover, it also gives a characterization of all reduced order controllers, although in this case the convexity property is lost, resulting in harder computational IMI APPROACHTO 74s, CONTROL 181 problems. The developments in this section follow after {124] and the closely related work in [158] and [275], with the main ideas being the reformulation of the bounded teal lemma in terms of LMIs and the observation that since the controller parameters enter these LMls affinely, they can be eliminated to obtain solvability conditions that depend only on the generalized plant and on two positive definite matrix variables, R and S. ‘To establish this result we begin by reformulating the bounded real lemma in terms of linear matrix inequalities and by establishing conditions for solv- ability of a certain matrix equation. Lemma 6.8 (Bounded Real Lemma Revisited) Consider the transfer func tion G(s) = D + C(sl ~ A1B. Then the following statements are equivalent: 1, G(s) € RH and ||G(s)Ijxo < 1 2. There exists X = X™ > 0 such that BX) -1 DT c Da ATX4XA XB CT <0 (691) Proof. We will prove first that statement 1 is equivalent to (I~ D™D) > 0 and the existence of a positive definite solution to the following algebraic Riccati inequality (ARI): ATX +XA+(XB+C"D)I ~D™D)(BTX + D™C)+CTC <0 (6.92) Define the following quantities: R = (I-D"D) Ay = A+BR™DTC (693) By = BR? Gy = 4 DR DTC Consider € > 0 small enough so that GAINGEs) > 0 (694) where AB c |p (695) vat} o 182 Hx CONTROL , has the following state-space realization: ao , CTC~el ~AT (6.96) DTC BT D'D ‘Thus the Hamiltonian matrix: Hv? | _oeQe Baby (697) ChCy- el -Al, does not have eigenvalues on the jw axis. Moreover, the pair (Ay, By) is stabilizable, since Ay ~ ByR-'2DTC ~ A isstable. It follows then that Ay € dom(Ric). Let X = Ric(Hy). Then X satisfies the following Riccati equation: AUX + XAy + XBYBEX + ChCy +1 =0 (698) Since € > 0, this implies that ALX + XAy + XByBEX + ChCy <0 (699) or equivalently, eT De el Do-l (6.100) ATX+XA XB CT < Since 4 is stable and the pair (Ay, [CZ Yel") is observable it follows that X > 0. This completes the fist part of the proof. On the other hand, suppose ‘that there exists a positive definite solution X’ to the ARI (6.92). Let Q > 0 be such that ATX 4XA4(XB+C™D)R\BTX+D™C)+C7C+Q=0 (6.101) Using the similarity transformation io mld wwe have that 4a 0 -B (3) = 1 ~ G18)GIs) [ # AT _|XB+CTD (6.102) DicsBx BT] OR where 18 [exp sctD)RBTX+ DTC)+9)"" >0 (6.103) LMI APPROACHTO M, CONTROL 183, m2 {4t prey (6.108) ‘Then simple algebra shows that ®(s) = Mis)M(s) +R — (B7X + D'C)L*(C7D + XB) (6.105) Moreover, using the matrix inversion lemma it can easily be shown that R~(BTX + D'C)L*(CTD + XB) > 0. It follows then that (jw) > 0, or equivalently (Gil. < 1. Stability of A follows from the Lyapunov theory and the facts that X >0 and C7C+L'L > 0. Equivalence of the ARI (6.92) and the LMI (691) follows from Schur's complement formula, Finally, 1 —D™D > Ois a necessary condition for (6.91) to hold. o Lemma 69 Given a symmetric matrix VW € R"™*"* and two maurices P and Q of column dimension m, the following statements are equivalent 1. There exists some matrix © of compatible dimensions such that V+ PTO7Q+O7OP <0 (6.106) 2. WEWWp <0 and WEWWo <0, where Wp and Wo are any matrices whose columns span the null space of P and Q, respectively Proof. (1 = 2) follows immediately by pre- and postmultiplying (6.106) by W} and Wp or WZ and Wo. The converse can be shown through some sim ple, albeit lengthy, algebraic manipulations considering bases for the kernels of P and Q and invoking Schur’s complement (see [124] for details). 6.4.1. Characterization of All Output Feedback 7,, Controllers In the sequel we will assume that Dz = 0,4 and we will consider controllers having the following state-space realization: fee] (6.107) where A c R*, For ease of reference, we will combine all the controller parameters into the single variable Ac Bs o-[é a] K “Recall hat this assumption can always be removed through a lop shifting transformation 184 Ha, CONTROL and we will use the following shorthand notation: alt sh este a o-( 34] mee [6] In terms of these variables the closed-loop matrices Ay, Ba, Cy, and Dy corresponding to the controller K can be written as Ag = Ao + BOC Ba = By + BODy, Ca = Cot DOC Dg = Dy +Dy20Dy CC ‘Theorem 6.2 Consider a FDLTI plant G of McMillan degree n with a min- inal realization nu) [Alb Bs 2] =|G]Du Bal (*] (6.109) y G|Dn 0 where the pairs (A, B,) and (A, Cz) are stabilizable and detectable, respectively, and where AE R™", Dy RO, Dry € R™™, and Day ER". Then there exisis an internally stabilizing controller K(3) with McMillan degree k ‘hat renders the closed-loop transfer function |\Toullw> <1 if and only if there ‘exists a symmetric positive definite matrix X € R°-4*0") such that WEOyWy <0 a (6.110) WE¥xWe <0 where PE LB Odmym, Dh ba | (6.111) QE LC Dr Oesm)xm | where Wp and We are matrices whose columns span the mal spaces of P and Q, respectively, and where ” d AoX'+X1AT B, X-1CT oy S Bi 1 Dh Lt el Diy a (6.112) 4 [AIX +XA, XB, CT we oe ot) cy Dy -T LMLAPPROACHTO CONTROL 185 Proof. From Lemma 68 we have that the controller (6.107) is admissible if and only if there exist X¥ © R="), ¥ > 0 such that the following LMI holds: AIX 4 XAg XBy CT Bix -1 Dil <0 (6.113) & Dy -t Using the explicit expressions (6.108) for the closed-loop matrices in terms of the plant and controller parameters, this LMI can be rewritten as Vx + OTOP, + PTO <0 (6.114) where BTX 0 DL] QSic Dy 0} (6.115) Next, we eliminate @ to get a condition depending only on the plant. From Lemma 6.9 we have that the inequality (6.114) is solvable if and only if WEY xWp, <0 (6.116) WEWxWe <0 17) where Wp, is a matrix spanning the null space of P,.. Since x00 Py=P}O 1 0 (6.118) oor it follows that x1 0 0 We, =| 0 1 0| We (6119) o oF Hence (6.116) is equivalent to x00) [x 0 0 wei] 0 1 0l¥e| 0 1 0] }Wp=Wwiowe <0 (6.120) o oF 0 o7 a ‘While this theorem provides a very general condition for the solvability of the suboptimal 24. problem, these conditions are given in terms of inequalities involving both X and X~'. Thus checking for the existence of an appropriate matrix X leads to a hard-to-solve nonconvex optimization problem, As we show next this difficulty can be avoided by rewriting the inequalities (6.110) in terms of two positive definite matrices R and S. As an intermediate step 186M, CONTROL in this derivation we will obtain solvability conditions given in terms of two algebraic Riccati inequalities that closely resemble the two algebraic Riccati equations obtain in Section 62, in the special case where the assumptions (A1)-(A4) stated there hold, ‘Theorem 63 Given a plant G(s) with the minimal realization (6,109) where the pairs (A, B,) and (A, C2) are stabilizable and detectable, respectively, there ‘exists an internally stabilizing controller with McMillan degree k that renders the closed-loop transfer function |Tzn\ls0 <1 if and only ifthe following linear ‘matrix inequalities in the variables R and § are feasible: ‘AR+RAT RC! OB, [Eo [ ca” SF ou][s o] co wren Br Dh =r _ ‘ATS+SA SB, CT [fof ars? Sr ok] [% 9] <0 eazy nt [f Hee enn where Nx and Ns are any matrices whose columns form bases of the mull spaces of [Bf Dh] and (Cz Dai}, respectively. Moreover, the set of suboptimal controllers of order k is nonempty if and only if (6.121)-(6.123) hold for some R, S satisfying the rank constraint rank (I~ RS) 0. Partition X and X" as S oN NT Xn M MT Yn where R, $>0€ R"*" and M, N ¢ Re™*. Rewriting @y in (6.112) explicitly in terms of R and M yields lio x (6.125) ie x AR+RAT ? By RCT (6.126) ar Dh GR Dy -T UM APPROACHTO Mx, CONTROL 187 where --- indicate elements that are irrelevant to the development. Note in passing that this expression is independent of ¥z2. By rewriting P explicitly in terms of the plant data we have oh oO 0 ole ae el (6127) It follows thatthe bases of Ker P have the form wo _[ 0 Ww=19 2, (6.128) w, 0 where (6129) is any basis of the null space of [BY DJ]. Rearranging expression (6.126) and exploiting the fact that the second row of Wp is identically zero, we have (Hele) [oe” 8] (CEL 9] Oe ia Cg Die! Int “(Fa Proceeding in a similar fashion it can easily be shown that the second in- equality in (6.110) is equivalent to (6.122). Finally, to establish (6.123) and the rank condition (6.124), note that using the Schur complement formula to compute the (1,1) block of X-! yields WiO,Wp AR+RAT RCT By GR =I Dy Gp Gt By i ‘| (6130) o7 R1 = (S—NXGINT) = SR =NXQNT>0 6.131) which is equivalent to (6.123). Moreover, since Xzz > 0 and N € R™* it fol- lows that rank(S ~ R-") < k ¢=3 rank(I — RS) < k. To show the converse note that if (6.123) and (6.124) hold then there exist N € R"™* such that S-R'=NNT (6.132) The construction of X and X-% can then be completed by taking Xx = M = —RN, and Yn = (I+ N7RN), 188M. CONTROL 6.4.2. Connections with the DGKF Results In this section we explore the connections between the LMI characterization of all suboptimal 7, controllers derived in the last section with the charac- terization obtained in Section 6.2. Suppose that the assumptions (A1)-(A4) introduced in Section 6.2 hold. In particular, since Dy has full column rank and Df,Dj2 = there exists D, such that [Dyy Da)" [Diy Ds) = 1.1 follows that a basis for the null space of [BT Df, is given by 1 0 ve[ola 8] wm Using this explicit expression for Ng in (6.121) and carrying out the block ‘multiplications, exploiting the orthogonality of C, and Diz (assumption (A3)) yields f ‘AR+RAT RCP By [" ‘| ce Foal [Me] <0 Br oDh o-T AR+RAT— BB RCTD, By (6.134) DICR tr o < 0 ‘BI 0 =r 5 AR+RAT + BBY ~ByBT+RCTGR <0 where the las ine follows from using Schur’s complement and the fact that D,D! =1 ~ DypDF,, Similarly. if Dy has fll row rank and Dsy[B] D3.) = [0 /], then we have that (6.122) is equivalent to the following algebraic Riceat! inequality: SA+ATS 4 SBBTS+CTC ~ CIC <0 (6.135) Defining X ten as Ro and ¥ © $-*, we have that these inequalities can be rewri ATX +XA+X(ByBI ~ B,BI)X+CTQ < 0 (6136) AY +YAT+Y(CIC, ~CEG)¥ + BBY < 0 Rewriting (6.123) in terms of X and ¥ we have that p(XY) <1 Next, we show that the existence of positive definite solutions to these algebraic Riccati inequalities imply that the Hamiltonian matrices Hx. and Jap defined in Section 6.2 satisfy Ha, € dom(Ric), J € dom(Ric). Moreover, X > Xap = Ric(Hy.) 2 0 and Y > Yq. = RicJoe) > 0 and thus p(Xu¥ox) < 1, To establish these results we need the following lemma UMIAPPROACHTOH,, CONTROL 189 Lemma 6.10. Suppose that assumptions (Al), (A2), (A3!) and (A) hold and thatthe following algebraic Riccat inequality ATX +XA+X(BBT ~ BBI)X «CTC, <0 (6.137) admits @ solution X. Then the following are true: 1, Hac does not have eigenvalues on the ju axis: 2. IfX > 0 then Ha. € dom(Ric) and X > Xu ® Ric(H) > 0. Proof. Begin by noting that X is nonsingular, since if there exists x # 0 such that x7X = 0 then pre- and postmultiplying (6.137) by x” and x we obtain 27CFCix = iIC.x] <0, which is obviously impossible, Thus we can define R= X-". In terms of R the ARI (6.137) can be rewritten as RAT + AR + RCTCR + BBY — BBE <0 (6.138) To show that Hx. does not have eigenvalues on the jw axis, assume to the contrary that there exist w, and a vector x = le such that de MP EJ-mE] om From assumption (A2!) it follows that v # 0, Pre- and postmultiplying (6.138) by vt and v leads to the following contradiction. 0 > v'RATU4 v'ARU + "RCT C\Ro + 0*(B\BY ~ BBP 0 = U'R(-CI Cu = jogv) + (-u"CTC; + joov")Ro 4U'RCTC\Ro +" (joo ~ Au) = U'RCTC)(Ru — u) — wCTC.Ro + (jw.v* - v"A)u = WRC] C\(Re ~u) ~ WTC Ro + wCF Cue = (Ro =1)"CTC\(Ro ~ uw) > 0 (6140) ne [¥]- [Re] (6.141) 190M, CONTROL where A ¢ R** is a Hurwitz matrix with real entries. In order to show that Hos € dom(Ric) we need to show that Xp is nonsingular. To establish this, start by expanding explicitly (6.141) to obtain AX; + (ByBT ~ ByBI)X> = X\A (614) “CT OX - ATX, = XA Premultiplying the first equation by XJ and the second equation by X7 and using the fact that’ X7.X = X7.X, yields TAX, + PAT: + XT CP CX) + XI (ByBY ~ B,BE)X2=0 (6.143) 0. Pre- Hence ‘Assume that Xp is singular. Then there exists x #0 such that Xo. and postmultiplying (6.143) by x7 and x we have that CX; % = loreover, ‘iplyir second equation in [ij] +=# Mores soxmaiping th soma eqion (6142 by x we have that X2Ax = 0. Repeating the same argument as before we have that [ % ]ate=o, ea =O, (6.148) X wus, from the stains atte [tare Pam he tes here ens yO and So ht Anal xy cx, y=0 (6.145) Postmmultiplying the first equation in (6.142) by y we have that (r-ayxy=0 (6.146) ‘Assume for the time being that the pair (A,C\) does not have any stable unobservable modes. Then (AI ~ A) is nonsingular and this last equation im- psa xy ~0.Hene [i] ~ Onan hpi ht | the stable eigenspace of H, and thus has full column rank. By interchanging the first and second row of Ha. we have that [wat aon sR] - ‘Tis fact can easly be established by noting that since Hay is @ Hamiltonian matrix then Fe oT a no ] rans [2] A 187) O where S| LMI APPROACHTO Nl, CONTROL 191 Let Rx. © X,X; '. Premultiplying this last equation by [Rx,—/] and post: multiplying by X;' we have that [Re -11 av ao 1 ] _ (BiB) ~B,BT) -A | [Ru - [Re -n| 1 Jou (6.148) RAT + ARyy+ BABY ~ BBE + RCT CR Subtracting this last equation from (6.138) we have that (R— RegJAT + A(R ~ Reo) + RCETCAR~ RewCP CR <0 (6.149) or equivalently, (RRMA + ReoCP Ci)" + (A + RaxC7 CMR — Row) + (R= Rx)C] C(R = Ra) <0 (6.150) From (6.148) we have that A? + C7C)Ray = —XzAX;". Since A is stable it follows then that A? + Cf C; Ra is antistable. Thus, from the properties of the Lyapunov equations, it follows that R— Ry, < 0,or equivalently, Rx. > R > 0. Since Ray > 0, Xx = Rz! is well defined and satisfies 0 < Xx, < X. Finally, to complete the proof, we need to remove the assumption that the pair (4, Ci) does not have any stable unobservable modes. This can be accomplished by perturbing C; to | ©], with ¢ > 0 small enough so that X remains a vet Solution of the perturbed ARI. The proof follows by noting that the sta ble invariant subspace of Hs, depends continuously on eand that from the discussion above we have that 0< X%,<¢X. Thus Xu Simo XS, i well defined and stsiesO © Xa, X. Finally this inequality can be sremgthened to X.. < X by eplacing X by (1 ~ m)X in the arguments above, with y smal enough o From this lemma it follows that whenever assumptions (A1)-(A4) (or (A2), (A2'), (A3!), and (A4)) hold then if the LMIs (6.121)-(6.123) are fea- sible, the conditions in Theorem 6.1 also hold. Conversely, assume that con- ditions in Theorem 6.1 hold, Since Hq. € dom(Ric) and Xx, = Ric(Ha.) > 0, it follows that for ¢ small enough the Riccati equation ATX, + X,A+X,(ByBY — ByBY)X, + CPC) +1 =0 (6.151) has a stabilizing solution X, > 0. Proceeding as before it can easily be seen ‘that this solution cannot be singular. Thus the LMI (6.121) admits a solution 192, CONTROL X, | Proceeding in a similar fashion it can easily be established that if Ja. € ‘dom(Ric) then the LMI (6.122) admits a solution Y, '.Finaly, by continuity the condition p(Xa.¥n.) <1 implies that p(X.¥,) 1 of, equivalently, that and ¥" satisfy the LMI (6.123) ince the discussion above shows that whenever assumptions (A1)-(A4) {or their relaxed counterparts) hold the approach pursued in Section 6.2 and the LMI approach presented here are equivalent, one may wonder about their relative advantages. For regular 21,, problems the approach of Sec tion 62 is preferable since it involves solving algebraic Riccati equations, a problem with a lesser computational complexity than solving the correspond: ing conditions given in terms of LMls. On the other hand, the LMEbased approach still allows for finding numerically stable solutions even in cases where Theorem 6.1 breaks down, such as singular problems or plants having invariant zeros on the jw axis. Finally, note in passing that the LMI approach presented in this section also parametrizes all reduced order controllers. The existence of an admissible controller having order k <1 can be investigated by checking whether or not the set of LMIs (6.121)-(6.123), together with the additional rank assumption (6.124), is feasible. If so, a reduced order controller can be reconstructed from the solutions R and S. Note, however, that the additional constraint (6.124) is not convex, thus resulting in an optimization problem considerably more iffcult than in the case where k >. 6.5. LIMITING BEHAVIOR So far the treatment has been limited to suboptimal controllers. In this section ‘we briefly examine the behavior ofthese controllers as Yo. the infimal achievable Ya, norm. Note that yx. can be computed by performing a search fon y until one of the conditions in Theorem 6.1 fails. I can be shown (see {125)) thatthe conditions X. > 0 or Ya, > 0 cannot fail prior to plXx¥) < -y°. Thus, in the limit, one of the following two situations will arise: 1, Xoo OF Yoo will no longer be stabilizing, that is, either Ha, oF Ju will have eigenvalues on the jw axis 2. pXa¥n) =F In the first case Xu, You, and Zax have well defined limits. Hence the central controller Keesra is well defined. Moreover, assume that H, ¢ dom Rie) but that X,, is still well defined, although no longer stabilizing. It can easily be shown that if assumption (A3') holds then the pair (Ay. ,C) is detectable. Thus from equation (6,60) it follows that the control law u = ~BEX,. is still stabilizing and renders T:wllx. = yy. Moreover, it can also be shown that there exists no control law (even nonstabilizing) such that |[Tsal]x. < Yor THE YOULA PARAMETRIZATION APPROACH 193 (On the other hand. inthe case where p(X...) ~ yi. then the formutas in ‘Theorem 6.1 are no longer applicable since 1 — yql¥q Xx. becomes singular, leading to unbounded Aj, and ZL. This difficulty can be circumvented by using the following descriptor version of the controller [135, 264) (= yah VaoXw)é = Ack + YoCEy 6.152 w= -BTXAE 2 where Ae= (I - 743¥aXw)[A + (7—3BBF ~ B)BT)Xx)— YauCE Cy. It can be shown that this system always has index 1 and defines a reduced order controller where the unbounded modes of A¢ cancel at infinity. Note, how- ‘ever, that this formula is valid only when J ~ y~?Ya.Xx is singular. Thus for values of y close (but not equal) to Yup: We still need to use the formulas in Theorem 6.1. However, since Z,.! is nearly singular, numerical difculties and near pole-zero cancellations in the controller may appear. These problems can be avoided by modifying the central controller with the inclusion of a feedthrough term (see [123] for det On the other hand, note that the LML-based approach works well even for values of y close to the optimal. 6.6 THE YOULA PARAMETRIZATION APPROACH In this sation we consider an alternative solution to the Mx control problem based on using the Youla parametrization to recast the problem into'8 model matching form. In addition to having historical importance, since it provided the fis solutions to the Hx problem in the early 1980s, ths approach also allows for exactly solving several multiobjective control problems such as mixed Hy/Hae and mixed (1/2, problems (29, 300) Recall rom Chapter 3 that the set ofall stabilizing controllers can be parameterized as K = Fy(J,0), where Qis any arbitrary element of Hx such that + DiaQ(oe) is nonsingular and where J is given in (3.34). Moreover, the corresponding set of all achievable closed-loop map i given by the following expression (affine in Q): Tow = T+ T20Tn (6.153) Under assumptions (A1)-(A4) (oF their relaxed counterparts) there exist Diz, and Dp, such that [Dy Dy.) and [DJ, D4, ] are unitary. Define now the following Hamiltonian matrices: ne [ A~ BEG BB] ] ~ |-CFD2.Dh CG, (A= DECI (6.154) a8 i BEC)" -IG ] BD}, Dn. Bl (A - ByD},C) 194 1H. CONTROL Stabilizability of the pair (A — B,DI,Cy,Bs) follows immediately from sta bilizability of the pair (A, B2) (just consider the feedback F = DF,C, + F where F is any matrix such that A+ B3F is stable). On the other hand, if Giz does not have any invariant zeros on the jw axis, then the pair (Df, ,Ci,A ~ BaF.) cannot have unobservable modes there. To establish this fact, assume to the contrary that (Df, ,C;,A ~ B,DJ,C\) has an unob- servable mode at jw. From the PBH test this is equivalent to the existence of a vector x #0 such that (A~BDEG)x = joo 6.155 Dh,cix = 0 a But this implies that Ve" i] Loko, ifol=° —— ss0 which violates the full column rank assumption (A2), Since (A ~ B:D},C;,,B:) is stabilizable and (Df, ,C,,A ~ BzDI,C)) does not have unobservable modes on the je axis, it follows from Lemma C2 (Appendix ©) that H € dom(Ric) and X — Ric(I1) is well defined and stabilizing. By duality the same properties hold for J and ¥ = Ric(J). It follows that the controller and observer gains in the Youla parametrization can be selected P= ~(Di,C + BY X) (6137) L = -WCT +B.Dh) yielding the following explicit expression for T: Ty Te ro [RF] Ar -BF | By By 0 A, | Bu 0 © | Ge -DeF | Du Da (6.158) 0 Q Dy 0 « » Ap = ABE AL = ALG Or = + DaF By, = By+ LD sing Lemma 5.1 it can easily be seen that yz is inner and 3; is co-inner, THE YOULAPARAMETRZATION APPROACH 195: Define now the following transfer matrices: 1 Ar [-X1 1D. a Or Draw (6.159) AL Bu Tas = = Dy BEY} | Daa where ! denotes the Moore-Penrose pseudoinverse. By straightfoward but somewhat tedious manipulations, using the formulas i Appendix B, it can be seen that the transfer matrices [7i2.1(5) Tia(s)] and ny are unitary Since the },. norm is invariant under multiplication by unitary matrices, we have that, WTenlloo Wirt T2QTrllo Ta lin. rors T2072] _ ee where RE (Tia. Tr]Tu{ Tai Tai] has the following state-space repre- sentation: -AE —MBy, -MDR, = -MDI, 0 AL YIBDE -cf -DRiGX' Dh DuBl, | Dh Duh, Dh Dud} By FY +D}DyBi, | DeDuDh,, — DDudh, (6.161) where M 2 XB, +CJ,Dy). Note that R is completely antistable. It follows thatthe 1 problem can be recast asthe following approximation problem 00 int {I+ Oate:Q=[) B]. Gerab rea) that i, the problem of finding the best stable approximation to an unstable system, postbly subject to some additional stucturel constraints on O ‘Consider fis the case of a SISO aster. In this case both RCs) and Q(s) are scalar transfer functions and this problem reduces then to the clasical Nehari approximation prablem. The intially achievable approximation error is given By the folowing result 212 196M. CONTROL Theorem 64 Let RERH. and denote by Vg the corresponding Hankel ‘operator introduced in Section 6.2.1. Then inf J+ Olle: = [ell = IR 6.163) dnt UR + Olle = Il = URL (6.163) and the infimum is achieved. Next, we indicate how to construct the optimal approximant using an algorithm originally proposed by Adamjan, Arov. and Krein (6). To this effect, assume that K€ R74, and recast the problem into the problem of finding the best antistable approximation to a stable transfer function by noting that inf |R+Qllo= inf [I+ Qlla. (6.164) ORM oekm, Theorem 65 Let G= RC RH and = [Feil = p'(WeW,). Consider the Schmidt pair introduced in (6.33 Tou = ow (6.165 Tgp = ow Ce Then the (unique) optimal Q= oV(s)/U(s) G, where U(s) and V(s) denote the Laplace transforms of u and v, respectively, Proof Let 118 (G+Q)U. Then IH -ToUl3 WHIB + IPcUI ~ (H,PoW) - PoU,H) WIG + IPoUIR - (PA. To¥) —PoU, Poy WIR - PoU.ToU) = IIMB - oF MUIR (6.166) where the lat line follows from the fact that since Q € #2, and U € #4} then P.H = P.GU =[GU. On the other hand, we have that IMAI = (G+ OWUIB < G+ OVENUIB = oF 1UIE (6167) Combining equations (6.166) and (6.167) we have that WA ~ToUIG = MIB — of IIR <0 (6.168) Hence H = I'gU, or equivalently (G+ Q)U =U. From the definition of Tg we have that Ps|QU] = 0, or equivalently, QU € RX}, which implies that Q.€ RH. The proof is completed by noting that since FgU = oV then (G+Q)U =o <> Q=a,V/U -G. o THE YOULA PARAMETRIZATION APPROACH 197 A state-space realization of the optimal Q can be obtained from the fol: owing state-space realization for the Schmidt pair (1, v) [286] -AT | -x Alw v= [of 44]. »- RY (6.169) where w is a right eigenvector of W.W, associated with the eigenvalue o? and where x £ 1/o,W,w. By using the formulas in Appendix B it is easily seen that UU = V-V. It follows then that V/U is allpas, which implies that the optimal error system G + 0 is also allpass. 1t can be shown (see Theorem 6:3 in [132]) that the previous results also hold when R is a square transfer matrix, that is infgc ys R+ Olle = Rll and (R-+ Qoge(R + Qogt) = 01. This case is known as the one-block approx- imation problem. We return now to the four-blocks approximation problem (6.162). Let yo denote the optimal approximation erzor. Since Q is now constrained to have a specific trutur, itis clear that y» > Ry. witha strict inequality holding most of the time. It should also be clear that {ite Rotts f[2"]]| } (6170) since these are submatrices of R + Q, and matrix compressions do not increase the norm. Thus the relative simple procedure used to solve the one-block approximation problem cannot, in general, be used to solve problem (6.162), Nevertheless, as we show next, this problem can be reduced to an equivalent one-block problem by using an allpass embedding argument (225]. The de- velopments in the sequel essentially follow the arguments presented in (182) and [135} Before attempting to derive a formal solution, itis instructive to present some intuitive arguments that justify the steps to be taken. Consider a one- block Nehari approximation problem of the form (ms ee] [2s 28] 0 = int 1 Oh | = [Rll am) Assume that the optimal solution is indeed of the form 00 2-[} ou] 17) and recall that, for the one-block case, the optimal error system is allpass, 198M, CONTROL that is, (R+ OFUR + Q) = (K+ OMR+ OF = x21. 1 follows then that (Ru Ral [RS] = vr ‘ean ae Rac [ RY] = ee As we will show later using an explicit state-space construction, the converse is also true in the sense that if this last equation holds then the one-block approximation problem (6.171) admits a solution of the form (6.172); that is, the one- and four-block problems are equivalent. Unfortunately, equation (6.173) does not hold in general. The key idea of the allpass embedding is that (6.173) can be forced to hold for some +, vem(ias matffEL} by introducing spectral factors L and N such that Rit (Rn Re L}) Ra) = 70 Ra (6174) Ry (Rit RF NI Ry | = 7T N provided that y is chosen large enough. This leads to an augmented one-block problem of the form Ry Re L] [Qu Qa Qe = dint |] Ra Ra +] +]Qu Ox Oss] =tRly 6.175) emt No « +] Lon On Ox) ]. Which admits a solution of the form oo oO O=|0 Ox Os (6.76) 0 Qn Os ‘The solution to the four-block problem is then obtained by discarding the augmentation, using the fact that matrix compressions do not increase the Next, we formalize these arguments. As before, for simplicity we will set ‘= 1 and we will seek suboptimal solutions to the problem. We begin by THE YOULA PARAMEIRIZATION APPROACH 199 finding necessary conditions for the existence of Qz» C RH, such that Ry Roy Assume that R € R74{, is strictly proper with the following state-space real: ization: ABB R=! ]Du De Dn 0 Ag | Bo On= al Co | Do such that (6.177) holds. Then by using standard dilation arguments (see ‘Theorem 5.2 in [132] and Lemma 3.1 in (135}) there exist matrices B,,C,, Bo, Cg, and Dy; such that <1 177) (6.178) and that there exists my im Pr Pa PPR Rp Ry 0 Pr |Rx En En Eu m, |Ru Ex En Ex m LO Ba Eo Ex. A 0 |B, B (6.179) 0 Ag | 0 Be CG Co | Dn Dg Dy Du Cy Ca} Dy Da Dy Dy 0 Ca] 0 De Ds Du =1, where Ey = Rij + Qij, Qij © RA, a [Ru Ra Ris Rn En En is allpass, that is, Fue Ry RHE, Let Ry Ea Es Since matrix compressions do not increase the norm, it follows that ||Exll~o < 1. Thus from Nehar’s theorem we have that |[Rely <1, where Ru Rn Ry 2 [Ra Rn Ro Ry Re Re By By By Dy Ds Dis i oo Dy + 4 (6.180) 200 Ha. CONTROL Note that since Ea, is allpass then Ris and Rs; can be found without explicit knowledge of by solving the following spectral factorization problems: RyRiF = F- RyRy — RyRi> 0 RyRy, = 1 RGR ~ RaiRa > 0 (6181) These spectral factors can be computed using standard techniques (see ‘Theorem 5.1 in 132), yielding DiDf, = I~ DyDf, - DDI >0 DhDay = 1 Di,Dy - Dia > 0 ay = (xeP 1% 21[D}]) oir (6.182) Gs = D5f (ary - (DR, DE {e)) where X and ¥ are the antistabitizing solutions to the Riccati equations XAT + AX ~[B, ani] ByBT =0 (6183) G GQ YA+ATY — [CT al |-da-o (6188) The construction of the remtining elements in R,, namely, Rey, Rip. and Res, can now be done using standard state-space manipulations (se (133), leading tova statespace realization ofthe form (6.180), Note that X and Y: are the Gramians for Ra. Since ||Rol|j < 1 it follows that p(XY) < 1. Finally, it can be shown (see proposition 3.3 in [135]) that these conditions can be strengthened to [|Rell <1 and p(XY) < 1. These results are summarized in the following. Jemma, Ry Ra Lamae it tan [BB | € RHE, be such that ||[Riy Rial <4 and { [Ih <1. If there exists Qz2 € RH ay such that (6.177) holds, then there exist X > 0 and ¥ > 0 satisfying the algebraic Riccati equations (6.183)- (6.184) and such that p(X¥) <1. As we will see next using an explicit state-space construction, it tums out that the necessary condition for the existence of a solution to the four-block approximation problem just obtained is also sufficient. Ry Ry Theorem 6.6 uae =[f ] 7 be sh har Ry Roalllas <1 [eI] <1. Then there exists Qro € RHa, such that (6.177) holds if THE YOULA PARAMETRIATION APPROACH 201 and only if p(XY) <1, where X > 0 and ¥ >0 are given in (6.183) and (6.184). Proof. Necessity has already been established in Lemma 6.11, Sufficiency will be established by constructing, a suitable Q. From the hypothesis it follows that 7Dy De) = [my ccan be embedded in a unitary matrix: moms pr pe mi [Du Da Dy 0 Pr |Dx Da Dy Day (6185) m [Dy Dn Dy Dy m 10 Da De Da For ease of reference we introduce the following notation: B= 1B B BO} CG =(C} cf cf oj” Be, = (0 Bo Bo Ball on) Co = (0 8 C8, CA) and define Re and 1x by the following state-space realizations mo LEHB] oH} A] com where Ag, Bg,, and Co, will be determined later. The corresponding state- space realization for the augmented error system Eay = Roe + Qus is given precisely by (6.179). We proceed now to select Quy $0 that Eqy is allpass. Since D, is unitary by construction, it follows from Theorem 5.1 in [132] that Euy is unitary if and only if there exist X,, and Y- such that fS hlseal Ae[i]oe aie om vfs 2][2 B]r-[Gfiq caro ows D[BE B8)+1C CouIX (6.190) 202 Hx. CONTROL DIIG Cyl+[Bl Bh] ¥.-0 (6191) MY. st (6.192) Motivated by Lemma 82 in (132] we will look for a solution to these equations of the form af-x 1 [7 v2] -y 2? i (6.193) [ Zx]=% 1-x¥ Note that since p(X¥) <1 then X; and ¥- are well defined. To solve for Co, and Bo, postmultiply equation (6.190) by (5 Z | to obtain Co, = CX ~ D_BT (6.194) Bo, = Z-"(YB.~ CID.) where we used the fact that Z is invertible and D, is unitary. Moreover, from (6.182) it follows that, 1 Du Z7Bo, 4 =YB,-CT a 0 (6.195) 0. 0 Simitay, 1 0 0 0]Co, =0. Tous Bg, and Co, have the required stuc- ture. Next, we choose Ag so that (6.188) holds. Premultiplying (6.188) by (1 0}and positing [9] st Ag = -AT ~ Bo,BE AT ~2-7(B, — CTD) BP = -2-T [1 YX)AT + B.BT ~ CID,B!] -Z-T(AT + YAX - CID,B?) (6.196) where the last line was obtained using the algebraic Riccati equation (6.183). Simple algebra shows that with these values for Ag, Bo,, and Co,, En Sais: fies equations (6.188)-(6.192) and hence itis unitary. To complete the proof wwe need (0 show that Qar € RHac. Using (6.196) itis easily seen that Ag satisfies the following Lyapunov equation AgYZ1+2-TYAR + Bo,B3, =0 (619) PROBLEMS 203 Since ¥Z-' > 0 and (Ag, Bg,) is stabilizabl it follows that Ag is Hurwitz a 6.7 PROBLEMS 1 In this problem we extend the proof of Lemma to the case where the pair (CA) is ony detectable rather than observable 22] witha unobservable and (2) Parton the states ofthe plantas x = [ (Ciu,An) observable. Assume that there éxists an internally stabil ing controller K rendering ||Tzullx <1. Write down the equations for the closed-loop system in terms of x1,x2 and &, the states of the controller. (b) Group these equations as follows: (i) a group containing only x,w and u and (ii) a group containing x,x,,2,w and u. Consider the first group as a “new” plant Goa, and the second group as a “new” controller Kot (©) Write a state-space realization for Gjp, and show that it satisfies the assumptions of the Fl problem. (@) Show that Ky. stabilizes Gy, and renders ||Tsullw <1. (€) Show that if X,p, > 0 satisfies the FI Riccati equation for Ga, then Ricltn) = Xa = [¥3 9] 20 exits for Gt follows that, with ‘out loss of generality (C,A) can be assumed observable, 2. Consider a stable strictly proper SISO transfer function G(s). Let wv denote the Schmidt pair of I defined in (6.23) and U(s), V(s) the cor responding Laplace Transforms. Show that UU = VV. Use this result to show that V/U is allpass. (Hint: Use the state-space realizations (6.169). 3. From (330), consider the following scalar 1, model-matching problem min {Ti + T2Qlloo: Q € RH} where Th, Tz € RM. are where Ta(ju) #0 for all w. (a) Show that for the optimal Q, T; + 720 is allpass (b) Show that if T(s) has N zeros in Re(s) > 0 then the degree of the optimal 7; + 7,Q is at most N 1, 4. Consider the problem of minimizing the Ma. norm of the sensitivity for ‘an open-loop unstable plant. Show that if the plant has an unstable pole ats = 5; then the closed-loop system has a pole at s = —5,. What conclu- sions can you draw about using 24. optimal controllers for plants with unstable poles very close to the ju axis? 204 H,, CONTROL 5. Nevalinna-Pick Interpolation. Given a set of m distinct points (ay, 42 4) with Re(a,) > 0 and aset of values {B1, Ba --,ba). bs € C, the classical ‘Nevalinna-Pick (NP) interpolation problem is to find a function G& RY sch that NGG) <1 Gla) = bj, F=1,2..n For simplicity, assume that the sets (a,} and {b,) have complex-conjugate symmetry. {(@) Show that the NP problem is equivalent to solving the optimal FL Woo problem for the following plant: (6.198) a, 0 0 by 1 0 0 be 1 “loo oft defer ted (6199) O00G bn 1 z=u (b) Show that if the Pick matrix defined by is positive definite then the NP problem is solvable, [Hint: Set up the FI Riccati equation and show that M~" is its solution.) (©) Give a LFT parametrization of all the solutions to the NP problem, 6. ‘The purpose of this problem is to compare the performance (in the ‘Ht sense) obtained by the optimal 712 controller against the optimal ‘Ha. performance. Given a function T € 71} it can be shown (132) that Tlau/|IU'rll < 2k, where k denotes the number of unstable poles of T. Use this result to show that the following bound holds (117) WT —TeOrles oy I= TeQrlin = where k denotes the number of zeros of 7s) in Re(s) > O and where Q> and Qz denote the solutions to the optimal % and MH, model matching problems, respectively af % control tools cannot be used when the plant has poles or zeros on the jig axis. This problem explores a technique for synthesizing suboptimal Hx. controller in these cases [78 (@) Consider the following bilinear transformation: Pipr>0 PROBLEMS 205 Show that this transformation maps the ja axis on the s-plane onto a RHP circle in the § plane. Find the center of this circle and show that it intersects the real axis at § = py and § = ps. (®) Show that the jo axis in the § plane is the image of a LHP circle in the s-plane (exactly the mirror image of the S-plane circle. (€) Suppose that a controller K(S) internally stabilizes a plant P(3) and ~ renders some ||7 (3) < 1. Show that the controller K(s) internally stabilizes P(s) and renders [I7(s)[lo < 1 (@) Based on these results, propose an algorithm for synthesizing subop: timal %. controllers for plants with poles or zeros on the ja axis. 8, Use the algorithm outlined in Problem 6.6 to synthesize a controller for the ACC benchmark problem (see Problem 5.3) subject to the additional specification that the closed-loop system should be stable for all values of k€ [0.5,2].[Hint: Try py ~ 0.5 and p2 ~ 100} ¥ Consider the problem of synthesizing a controller for fuel injection con- “" trol [177]. The nominal plant (at T= 25°C) has the following transfer funet 5.4986? + 400.7 — 444400, $34 9B.72s? + 9530s + 121400 ‘The control should operate correctly for fuel temperatures ranging from T =0°C to T = 60°C. The plants corresponding to these extreme tem- peratures are given by: 0.017368? + 493.95 ~ 313700 ‘9+ 98.345? + 92038 + 87710 4.6775? — 285.95 ~ 505300 535 915357 + 100805 + 176200 Py= (6.200) Pa = For the purpose of this problem we will assume that the frequency re- sponses of all plants corresponding to T € (0,60] are contained in the band determined by the frequency responses of the two extreme plants (2) Show that the family of plants can be modeled as the nominal plant Pa subject to multiplicative uncertainty & with |W; dln. < 1 where Wy6) = 2259 (b) Design an Hae controller that robustly stabilizes P.s (©) Plot the step response of the resulting closed-loop system for T (0,25, 60]. Is the tracking performance acceptable? Explain your re- sults, [Hint: The plant is open-loop stable} 206 24, CONTROL 10. (@) Design an H,, controller that robustly stabilizes Ps and renders [Wp Sla <1 where wis) <0 0 (e) Repeat part (c) with the new controller. The uncertain Linear system Alt) = [A+ w()DE] x (6201) where A € R™", D ER"! and £ € R!*" are known matrices and where (0) € [w" ,w*] represents memoryless time varying uncertainty is said to be quadratically stable if there exists a positive definite matrix @ € R™" and a scalar a > Osuch that the derivative of the Lyapunov function x'Qx along the trajectories of (6.201) satisfies: 4 : 4 (Fox) < ala (62m) for al posible values of w(). Use the Bounded Real emma (Lemma 68) to show that if G(s) 2 E(sf — A)"'D € Hy, and |\G(s)] < y then the system (6.201) is quadratically stable for w€ {—y-!.¥-1 STRUCTURED UNCERTAINTY 7.1 INTRODUCTION The objective of this chapter is to introduce the reader to more general uncertainty structures, which are often encountered in practical applications There are some topics in this chapter that are still an active area of research. Therefore, in these cases, the approach will be that of a survey rather than of a textbook. ‘The class of model uncertainty that has been considered in Chapters 2 and 4 is global dynamic. This name comes from the fact that the uncertainty is attributable to all the system and describes the unknown higher-order dy- namics of the plant, However, in many practical applications this type of un certainty description “covers” in a very conservative way the real uncertainty ff the plant’s model. This is the case when more “structured” information con the plant is available. Hence a nonconservative analysis o synthesis pro- cedure should take advantage of this extra information. Next, we mention some of these situations. ‘© Consider the total system as composed of individual subsystems, each ‘with its own dynamic uncertainty description, Take, for example, the actuators, the system itself, and the sensors, described by the following individual sets of models: Gals) = 4 AWals\Galsy Ave CH, Bl < 1 7) Gs) = [1+ BoWy(s)] Gis); Arye CH™, [dal E (72) Gls) = [4 DAWG), Ba E CHA, [sl] <1 (73) 207 208 STRUCTURED UNCERTAINTY The total interconnection yields Sts = Gils) Gls) - Gals) (4) The above set can be transformed to a LFT connection between @ nominal model G(s) and an uncertainty block in the set Aswoa © (diaglas a2 3), a, € C8, |a,l) < 1, i= 1,2,3). To this end, define the outputs from the uncertainty blocks as {u,, i = 1,2,3} and the out Puts from the weight functions as {y;, 7=1,2,3} as in Figure 7.1. In addition, the input and output of set S are defined as u and y, respee- lively. After some algebra, we obtain the following matrix G(s) 0 0 0 WG, Ys WG 00 W,GG. | [us [] ~ | W.6,6 W,G, 0 W,6,66, | @) GG G1 GG, Als 4 ws |r], w 76) a us Therefore the set {F, [Gi(s), 4], A € Aun} is equivalent tothe family of models S. This type of uncertainty sealed sirucured dynamic onthe other hand, the uncertainty of the plant is described as global dynamic, thats {F [6t6).8] eC, a < 1} (ey «hs + ke) distegarding the structural information will add more unnecessary mod- cls to the set. Hence the robustness analysis of a closed-loop system with Such an uncertainty description willbe conservative in general, In many cases, the plant has a well known mathematical mode! usually derived from physical equations. Tis isthe case of some applications from mechanical, aeronautical, and astronautical engineering, where the Tigid body model based on the Newton-Fuler equations provides a go0d enough description for mild performance spectications However the parameters ofthese models may not be known. As a consequence thelt values are estimated either by means of classical parameter identifica tion ({13, 184, 293) procedures or, more recently, using st member: Ship identification methods (155, 204)). As with any other mathematica Wal | Ay W,} | 4a As 3 TS ej} te Figure 7.1. System block diagram, uf oie], de 2 INTRODUCTION 209 procedure involving input data from physical measurements, the nom- inal values of these parameters will be known within some uncertainty bounds. When these bounds are deterministic worst-case bounds, this leads to a plant representation in terms of a family of models with a mathematical fixed structure and parameters that may take values within certain specified sets. This type of uncertainty description is called struc- tured parametric uncertainty or simply parametric uncertainty. Take, for ‘example, the following set of models, which represents a plant with fan uncertain real zero and uncertain natural frequency and damping coefficient: SPF Qos Fo} 2E ltt], on € lover), €€ [6.4] Hs,p) 7) p2[z one)” Here the structure is based on the fact that only specific elements of the plant’s model are uncertain (2, om Parametric uncertainty can be present in both state-space or transfer matrix representations. Inthe later case, there are basically (wo ways in Which the uncertain parameters are handled. The first adopts the stan- dara LFT structure with the parameters located in the upper uncertainty block A as in Figure 7.2. The second performs the analysis on the charac- teristic polynomial of the closed-loop system. In this example it is easier to put the uncertainty in the latter form in order to evaluate stability robustness: P(s,p) = 5? + (k+2un8)s+ (kz +02) 8) a) cai) where k is a constant (nominal) stabilizing controller. wi . va (a) | _.y Figure 7.2. UF uncertainty description, 210 STRUCTURED UNCERTAINTY ‘¢ In general, both parametric and structured dynamic uncertainty appear simultaneously. This is the case, for example, of large flexible space struc tures. They have a well known low-frequency model described in terms of the natural vibration modes, as well as unknown higher-order dynam: ics. The lower-order model is represented by several second-order modes with natural frequencies and damping coefficients within real intervals, that is, parametric uncertainty. The higher-order dynamics, on the other hand, can be represented more naturally as dynamic uncertainty. This is the so-called mixed type uncertainty. For all the above types of structured uncertainties, more general mathe- ‘matical tools than the ones described in previous chapters should be used. In this chapter, a brief description of some of the analysis and synthesis results {or these cases will be presented. Special emphasis will be placed on methods that use the standard LFT description for the uncertainty Recall from Chapter 2 that stability or performance robustness margins are related directly to the type of uncertainty present in the plant. As special cases we have mentioned the classical phase and gain margins. For struc- tured uncertainty the same concept holds; therefore a general definition of a robustness margin should be made. Next, we present such a definition, and we leave for the rest of the chapter the different methods to compute this ‘margin, according to the type of uncertainty structure 7.1.1. Stability Margin The margin of a particular property of a system (stability, performance, con- trollability, detectability, etc.) is always defined in terms of the distance of the nominal model, which represents the system, to the nearest model that lacks this property ((223)). To quantify this distance, a particular norm should be tused. Therefore the system is more robust if this margin is larger and vice versa. Furthermore, the nearest model that lacks this property will depend ‘on the type of uncertainty in the system. For this reason, as mentioned in Chapter 2, the concepts of uncertainty and margin computation are closely related. ‘We will be concerned in this subsection with the stability margin, although in many cases, as will be seen in further parts of this chapter, a simple ex- tension to performance can be made. Furthermose, due to the fact that the stability margin is an analysis tool, we assume that a stabilizing controller for the nominal model has already been designed and closes the loop. Characteristic Polynomial Framework A natural way to state the problem is in terms of the closed-loop characteristic polynomial (CLCP) This includes many practical applications, especially in cases where para metric uncertainty is involved. A general closed-loop polynomial for FDLT! IiRooucTON 211 systems can be defined as follows PUS,D)=3" Hep ips FE) +e), PElad) (79) where {o(), i= 1,.-.ym) ate real functions of the uncertainty vector p and PE [ab] represents p, & [abil = 1,-..ymeHere a nominal internally stabi lizing controlicr is assumed; therefore the nominal characteristic polynomial P(s,p!) has all its roots in C_ © {s € C; Re(s) <0}. An equivalent condition for robust stability i Robust stability —> P(s.p) #0, Vpela,b], VseC, (7.10) where C, & {5 €C; Re(s) > 0}. ‘The previous condition provides a qualitative answer, that is, a yes/no ver- ification of robust stability. On the other hand, a stability margin provides a {quantitative measure of the degree of stability. Obviously the former quali- tative answer can also be obtained using this margin. As a consequence we seek 10 define a computable robustness margin for stability. Checking condition (7.10) requires computing the roots of P(s,p) for all possible values of p € [a,b]. On the other hand (since the nominal system is stable), a robust stability margin only needs to indicate at which point the roots of P(s,p) cross over from C. to C, as the uncertainty around the nominal set of parameters p® is “increased.” In this situation, it seems natural that as poles move from C_ to C., the first unstable ones reach the jw axis before entering the interior of C.. In general this property is not guaranteed, however the conditions under which it holds are stated in the boundary crossing theorem presented next, To this end we first define a partition of the complex plane. Definition 7.1_An open region O, and a closed one O; define a partition of C if and only ifthe following two conditions are satisfied: = 0,U00,U0, qa) 0 = 0,0; = 0,0; = 30; 10; (7.12) where Oy is the interior of the closed set 02 and 80, is the boundary of O}. The conditions under which the imaginary axis crossing is achieved are as follows ([35, 47) ‘Theorem 7.4 (Boundary Crossing) Consider a polynomial P'(s,p) of fixed degree and continuous for all p {a,b), an open region Oy, and closed one Bs, which form a partion of ©. Then if P(s,a) has all its rooss in O; whereas 212 STRUCTURED UNCERTAINTY P(s,b) has at least one root in Da, there exists at least ome x € (a,b) such that P(s,x) has alls roots in ©; UO and a least one ro0t in OO) Proof. Based on the theorem of Rouché, the detailed proof can be found in (7) a The above result canbe interpreted in terms of robust stability by defining the open set ©; =C., the boundary 00, as the jw aris, and the closed set ©; =C,. Furthermore, the assumption that P(s.a) has all its roots in Oy ‘means that there exists a stabilizing nominal controller, where the nominal value of the closed-loop polynomial has been assumed at P(s,a). Under these conditions, the theorem states that to destabilize a nominally stable closed- Joop system by moving its parameters from a— b, the fst offending set of parameters will generate an unstable pote in the boundary jw and not in the interior of C... We define an offending vector x of parameters as the one that generates an unstable root of P(s, x). Therefore, under these conditions, robust stability can be verified by looking only at the s = jo axis. Similarly, for discrete time systems, we define the open set O, as the interior of the unit disk and O as its complement This result suggests another simplification to the robust stability equiv- alent condition stated in (7.10). If the characteristic polynomial meets the assumptions of the previous theorem, the necessary and sufficient condition is as follows Robust stability => Pye.) #0, ¥pelab), ¥w (713) Note that this is a maximum modulus ([332]) type theorem for the case of parametric uncertainty, LFT Framework In many cases, itis more convenient to structure the un- certainty a5 a LFT, a5 for additive or multiplicative dynamic uncertainty (see Chapter 2 and Appendix B). This is a natural way of evaluating robust sta- bility when dynamic uncertainty is involved, due to the fact that in these cases the model order is not fixed. In the parametric uncertainty case the LET setup includes only ¢,(p) functions that are polynomial in the parame- ters, Therefore the previous analysis based on the characteristic polynomial P(s,p) would be more general! Next, the analysis problem in a LFT setup is presented, as in Figure 7.3. Here T(3) includes the nominal model, the controller, and the performance and robustness weights. [Note that, if the inputs and outputs w and z are selected conveniently and the problem is well posed, the stability of F, [7(s) 4] for all in a prescribed "Historically, the CLCP has been used to compute the sabi margin for pure real param: ‘ter uncertainty: while the LFT formulation has been used forthe structured dynamic type uncertainty, due tothe abave reasons, INTRODUCTION 213 Ts) Figure 7.3. UT or robustness anaiysis. bounded set A guarantees robust internal stability of the closed-loop system it represents, Based on the fact that nominal internal stability guarantees T(s) stable—and hence the only element in F, (T(s),A) that may destabilize it {assuming 4 stable or constant) is {J — Tix(s)4]"'—the following conditions are equivalent to robust stability: F,|T(s),A] stable VAe A (74) <= I-Ti(s)Al"' stable VA A (748) +> det ll - TuG)d] #0, VAEA,VsEC, (7.16) ‘The above condition can be assumed without loss of generality due to the fact that w and z can always be selected so that input-output stability is equivalent to internal stability (Problem 1). The uncertainty structures A that will be used in the sequel are oO eo aadf [jase oan 0 Am a tor snctine dane uncer 59 n a8) 0 Bm ’ for parametric uncertainty; and the combination of both ave {{4 S]fseae sea} amy 214 STRUCTURED UNCERTAINTY for mixed uncertaimy descriptions, These are unbounded sets, hence we denote BA, as their corresponding unit (open) balls, that is, 7(,) < 1 and 3) <1.t=1,...,m,lor BAy and Bay, respectively. The notation for closed balls is BA, As in the previous ease, the robustness condition can be simplified to a test over the imaginary axis only. This maximum modulus type properly has been proved in [55] and can be used to establish the following result: Robust stability + det{—Tn(w)A] 40, VACA, Va (720) In the definitions of the stability margins presented next, we use the LFT. framework. However, in further subsections when considering specific classes of uncertainty structures, each corresponding type of formulation will be used. Stability Margin Definition The next two definitions take into account the previous discussion concerning the stability margins. These are formu Jated so that direct necessary and sufficient conditions for robust stability in the case of general types of uncertainty can be stated Definition 72 The multivariable stability margin ky (/257)) is defined as kno (M (je) © inf (k € (0,00)|det{d ~ kM (jw) Defi 0, for some A ¢ BA} jon 7.3 The structured singular value j1s ((102)) is defined as ss M03} © [ng f(a] det — M0) =o] and 1siM (ja)] = 0 if det{l — M(yw)S) 40 for all A € A. From these definitions itis clear that km (M (Jo)} = 43" [M(j)] when the uncertainty structure is the same. From both definitions, the following necessary and sufficient conditions for robust stability can be stated. Lemma 7.1 The following statements are equivalent: 1, The family of closed-loop systems {F, |T(s),], 4 € BA) of Figure 73 is robustly stable. 2 walTiGe)] <1 forall oe R (721) 3 ky (Tie) 21 forall eR (722) Here the set 4 may be any of the previously defined uncertainty structures” In further sections we replace Tyy(s) by M(s) for simplicity > Note that we have eoasidered the open hal in. Wis. The cane of choned Falls mae dificult ova ch produces the above nonstit inegua! STRUCTURED DYNAMIC UNCERTAINTY 215 As noted in the above discussion, these quantities have been defined to provide directly the equivalent robustness condition for these rather general types of uncertainty. The hard problem is the computation of these mar gins for all the previous uncertainty structures, In fact, it can be proved that the calculation of the exact stability margin (or even an approximation with guaranteed a priori bounds) of an uncertain model with a general paramet- ric uncertainty structure is a NP-hard problem ((60, 248)). To illustrate this point, note that this computation involves the mapping of a polynomial func- tion P(s,p) for all parameters in the set. For parameters defined within real intervals, this set is an m-dimensional hypercube at each 5 = ja, and for the general case its image produces a nonconvex set. The computation of the Image of the vertices alone involves 2” evaluations of the function. In addi tion, there is no way to determine, for a general P(s,p). a subset of these vertices, polynomial in m, which may define this image. It is well known that this is a NP-hard problem ((62)) Under this strong constraint on the computational aspect of the problem, it seems that research should be pointed at least in (wo directions. First, looking, for exact (or approximate with guaranteed bounds) polynomial time analy: sis of parametric uncertainty structures, which may not be general but can accommodate relevant practical applications; and second, studying approxi- ‘mate methods (branch and bound, heuristics) that can bound in polynomial time the stability margin for general caces and, although they may not have guaranteed a priori error bounds, work reasonably well in practical situations ((326)). The former approach has been pursued by the research area briefly described in Section 7.3 of this chapter, which seeks extensions to practical cases of analytical or tractable computational results. The formulation is in terms of the characteristic polynomial P(s, p) of the closed-loop system. The latter approach is based on the structured singular value applied in a LFT setup and will be explained in the last section of this chapter. ‘The formulation used for pure dynamic uncertainty has no computational hard constraints and will be presented in the next section in terms of the structured singular value in a LFT setup. In addition, the use of these margins as analysis and synthesis tools will be illustrated, 7.2. STRUCTURED DYNAMIC UNCERTAINTY 7.2.1. Computation In this section we restrict the uncertainty to the set Ay defined before; there- fore the definition of 4 will be considered with respect to such a structure. Historically, this problem has been developed in [102] using a LFT frame: work. An excellent tutorial on this subject can be found in [220} One of the reasons why this margin has been defined as the inverse of the destabilizing perturbation is that in this form it naturally generalizes the 216 STRUCTURED UNCERTAINTY notions of maximum singular value and spectral radius.’ This can be ex: plained by means of the following two examples, Example 7.1 Consider the set Ay © {81; 8 € C) Ay. By definition we have det [J — M(jw)5] = 0 for 6 = A>" |M (jw)), therefore the minimum |8.| = |Aal', Where Ay is the maximum magnitude eigenvalue, that is, p(M (jo) = [sh Hence for the uncertainty structure Ay, we have Ha, [M (jo)] = piMGw)] (7.23) Example 7.2 Take the set Ay © {A € C*'}, which includes the structure Ay when r= 1) +--+ Fm in (717). This 15 the global dynamic uncertainty case considered in Chapters 2 and 4. Therefore, according to the robust stability condition in that case, the following holds: Ha, [M (joo)] = [M (Jo)} (7.24) In fact, we can select the smallest perturbation as A. =VE,U", 34 = dinglo,?,0,..,0], where M (Jw) = U2V* and @(2) = 0}, Iris simple to verify that det{f — M(jw)A,) = 0. From the above examples and the fact that Ay C Ay C Ay, we obtain inf 70a) > inf Ba) > ing (a) (7.25) = p|M(w)] = Ha, [M(Uw)] $F [M yw)] (7.26) ‘This inequality presents the structured singular value as a generalization of the maximum singular value and the spectral radius. These measures will be instrumental in computing tight bounds on p, although a further step needs to be performed due to the fact that the above bounds may be conservative in general Example 73 Consider the constant matrix case (or, equivalently, a system “frozen” ata given frequency) R= M (1) x= [to] azn which has speciral radius p(R) = 0 and maximum singular value @(R) =1 Take the following uncertainty structure to compute w(-) a={[5 2]. acc} 728 The special rads i defined as p(A) # maxiA(A)l where A() ae the elgenvales of A ‘STRUCTURED DYNAMIC UNCERTAINTY 217 Then det (I~ RA) =1 for all B€ A and, by definition, p(R) =0. Hence p(R) = ws(R) =O < a(R) = 1, which by scaling can be made arbitrarily conservative. Example 74 Take instead the constant matrix 0s -05 x= [03 223] (729) with p(R) = Oand 3(R) = 1, and the same uncertainty structure ofthe previous example. Replacing A € A we obtain det (I — RA) = 1- (8) - &)=0 (7.30) Lo fraa[} F(R) =1, which By scaling can be made arbitrarily conservative Therefore ws(R)=1 and hence p(R)=0< a(R) = To improve these bounds in order to compute 41, the following properties will he nsed_ Far simplicity we drop the argument in the system matrix M(s) and the uncertainty set A, as the index of the structured singular value. 1. waM) =|alu(M), a € C. This is similar to the scaling property of norms (see Appendix A) and is due to the fact that det (I — aM) = det — M(ad)] (731) and (aA) = |ala(A). 2. w(A--B) < 7(A)q(B). This is based on the facts that det(/ — AB -A) = det — B(A- A)} and 7(MA) < (A) (A). 3. Consider a general uncertainty set A, with respect to which pis defined. Then: m(M) = max o(Ma) 732) ‘This can be used as an alternative definition for 4(.). From the scaling. property of i, it follows that we need to prove the above condition only for the two following cases: say p(MA)=1 > w(M)=1 (733) 218 STRUCTURED UNCERTAINTY which holds by definition, and det(- MA) 40, VOEd (734) = w(M)=0 (738) = sup p(MA) <1 (736) aed => sup p(Ma)~0 an acta Otherwise, if the last equation is not equal to zero, it may be sealed in order to contradict (736), Note that j4,(-) is continuous due to the continuity ofthe spectral radius and the max functions and the fact that BA is compact 4. n(MU) = w(M) for U €U, the latter defined as follows: U2 (Ue dy, U; unitary, i= 1,....m) (738) This can be proved with simple arguments (Problem 2). 5. w(DMD~) = u(M) for D € D, the latter defined as follows: ah 0 ps 1,2 identity in r, 739) 4,>0 0 dal which can also be proved using the definition of 4 (Problem 3), Equation (7.26) can be used to improve the bounds as follows: mage(MU) < max (MA) = u(M) < inf 7DMD-!) (7.40) Here the equality follows from property 3, Moreover, tighter bounds can be obtained using the following two results from {102]. ‘Theorem 7.2 maxp(MU) = w(M) (rat) Proof. The argument is based on the fact that for 4(M) = 1 there exists a U EU such that det(/ ~ MU) = 0, which in turn implies p(MU) > 1. Prop: enty (7.40) completes the proof, See [102] for details. 9 STRUCTURED DYNAMIC UNCERTAINTY 219 The fact that the optimum in (7.40) of the spectral radius over BA and U coincide means that the optimal uncertainty A in this problem can always be made unitary, ‘Theorem 73 jal 7(DMD~) = w(M) (7.42) for m <3 in the uncertainty structure Ag defined in (7.17), Proof. The proof exceeds the scope of this book and thus will not be pre- sented here. It can be found in [102] a The computation of the structured singular value by means of Theorem 7.2 involves a nonconvex optimization problem; thus there is no guarantee of reaching the global optimum. On the other hand, the optimization problem in Theorem 7.3 is convex ({259)) as will be proved next Lemma 72° The set {#(DY7MD-"") < , DED), for y>0 is convex inD. Proof. The following are all equivalent: w(DYPMD"'") = DVMDMD-?— 71 <0 <=> MDM -y'D<0 Which is convex in D € D (220) a ‘The latter is a linear matrix inequality (LMI), which can be computed efficiently ({59, 126)). From Theorem 7.3 we have that the upper bound is, tight only in the case of three or less uncertainty blocks, However, extensive ‘computational tests have shown that it does not exceed 15% in the case of Pure dynamic uncertainty ({113, 219)) 7.2.2 Analysis ond Design Robust Performance Analysis The condition presented in Lemma 7.1 ‘generalizes the robust stability theorems in Chapters 2 and 4, which apply only to the global dynamic uncertainty case, that is, m = 1 in the definition ‘of Ay. In addition ((102)), a necessary and sufficient condition for robust 220, STRUCTURED UNCERTAINTY performance can also be obtained. In fact, we will show next that robust performance for a family of models with a structured dynamic uncertainty set A, can also be interpreted as robust stability of a different uncertainty set (see Chapter 2} Based on the performance definition of Chapters 2 and 4, and using the condition for nominal performance based on the induced norm, the following conditions can be stated for the system in Figure 7.3: Robust performance <=> [z\j2 <1 Vllwlp <1, VA Ay (7.43) = IFITG) Me St, YEA, (7.44) From the robust stability condition obtained in Chapter 4 for plants described by a set of models with global dynamic uncertainty, it follows that equation (7.44) is necessary and sufficient for the system {Fi [T(s).8]. Ac Ag} to be robustly stable against (A © C’"’, a(8) < 1}, connected be- tween input w and output z (see Figure 7.4) Due to the fact that an interconnection of LFTS remains a LFT, Figure 74 can be transformed to Figure 7.5. Hence the new uncertainty structure can aa{(s paca.) os) ‘The above uncertainty set is qualitatively the same as Ay, but with an ad- ditional uncertainty block, usually called the performance block. However, the analysis tool remains the same, the structured singular value, but now with a slightly different uncertainty structure. Therefore, due to the equiva- lence between robust performance and structured robust stability depicted in ee YS) Figure 7.4, Equivalence between robust performance and stably SSIRUCTURED DYNAMIC UNCERTAINTY 221 T() Figure 7.5. Robus! performance analysis. Figures 7.3, 7.4, and 7.5, equivalent conditions for the former can be estab- lished as follows. Theorem 74 The set of models T(s,) of Figure 74 has robust performance if and only if ng(TOm)) <1, ve R 746) ‘There should be no confusion in the notation conceming the uncertainty structure and the system to which wis applied. Recall that in order to simplify notation, we have made M(3) = Tin(s); therefore it is clear that a, (M Ga] # uj (Tw). In addition, wa, (Ta) oF 45M Yw)] cannot be ‘computed due to a dimensional incompatibility between the system and the uncertainty structure. In Chapter 4a sufficient condition for robust performance of sets of models with global dynamic uncertainty was obtained. Since (7.46) is necessary and sulficient, the following holds inthe two-block case (m = 2 when defining 4) 44 (TU@)] < F(T Ge) +o [Tralyo)] (747) when Ty(s) and Ti(s) represent the robust stability and nominal per formance transfer matrices, respectively. Furthermore, in the SISO case (Chapter 2) the above inequality turns into an equality (Problem 4), ‘Synthesis When disturbances and errors are measured by energy content (il-) and uncertainty blocks by the equivalent induced norm ((j[|x.) for AE RH, and a) for A € C”*"), two design procedures have been presented so far: inf IFVIP(3),K(3)l [las € {2,00} (7.48) 222 STRUCTURED UNCERTAINTY The case a = 2 is the LOG or Ms optimal control problem addressed in Chapter 5. The ease @ = co leads to 7. optimal control covered in Chapter 6, It guarantees nominal internal stability and solves in an optimal way either robust stability for global dynamic uncertainty or nominal performance, The latter considers the more general case of norm-bounded disturbances. It can also solve, although not optimally, the robust performance problem known as mixed sensitivity ‘A more general procedure that seeks a controller that solves the optimal robust performance problem, in the case of structured dynamic uncertainty and norm-bounded disturbances, is known as -synthesis. It can be stated as the following optimization problem: “int, sup sa, (FeIPG), KOM 49) Due to the fact that the structured singular value is computed via its bounds, the following suboptimal (in the case of m <3 it is optimal) problem should be solved: (750) int, inf HOOF PE). KOO DEY" ‘A suboptimal solution to the latter problem can be obtained through the D~K iteration algorithm. This procedure “freezes” one optimization vari able at a time, say, Ki(s), and alternatively solves the other (convex) opti mization problem, say, Di.) Several remarks are in order: * The bound obtained in equation (7.42) holds for each s = jw, therefore in that case D € D is a constant matrix. On the other hand, the scaling ‘matrices D(s) in equation (7.50) are functions of s; hence they should be obtained at each frequency and fitted by a rational matrix function, for example, spline interpolation, For a fixed D(s) matrix, equation (7.50) is a weighted problem that is convex and has a well known set of solutions. In particular, for the case when D(s) equals the identity matrix, it is the PL. optimal control solution 10 the problem. In this sense the D — K iteration can be interpreted as an algorithm that solves scaled Hx. problems at each iteration. The order of the controller depends on the order of the augmented system, which in this case includes the weights D(s). Therefore the order of the matrix function D(s) that fits the optimal scales at each frequency should be made as low as possible. Otherwise, at each step, the order of the controller will increase significantly. For a fixed controller K(s), the optimization has been shown to he con vex: therefore a global minimum D(Je) can he found efficiently at cach Frequency point PARAMETRICUNCERTAINIY 223, However, itis well knowa that the combination of two convex problems is, not necessarily convex, As a consequence, if a fixed controller Ko(s) has been obtained and if by solving the optimal D-scale problem and interpolating it wwe obtain Dy(3), and so on, there is no guarantee that the optimals Kyyc(s) ‘and Dy»y(s) will be reached. A way to partially solve this is by checking both upper and lower bounds on In practice, many challenging application problems have been solved eff ciently using commercial software products ([20, 126, 262)), which implement this procedure, although research in this area is still active, Some examples will be presented in the last chapter of the book. 7.3. PARAMETRIC UNCERTAINTY 7.3.1. Introduction System models with parametric uncertainty originate either from physical laws or are the outcome of parametric identification algorithms. In the for mer case, the parameters may have a physical interpretation; in the latter, they usually do not. In general, parameter identification algorithms fit a given ‘mathematical structure that represents a certain physical phenomenon to a set of experimental data from this same phenomenon. This is accomplished through an opumuzation procedure with the parameters as its variables. The classical stochastic approach has been studied extensively ([13, 184, 293]), particularly in the area of adaptive control. More recently, set membership theory has been applied to model identification ((155, 204). In both ap- proaches, not only is a set of parameters obtained, but also error bounds for ‘each parameter, These bounds can be described stochastically (soft bounds) ‘or deterministically (hard bounds), In any case, in this section we assume a given mathematical FDLTI struc- ture in the frequency domain with unknown fixed parameters lying within real intervals. Gs.p). Pie (a,b. ‘This is consistent with the deterministic approach pursued in robust control. Once the above given family of models is assumed, the next step is to analyze the robust stability of this set for a given controller. A natural way to do this is by evaluating the set of roots of the characteristic polynomial of such a family, as mentioned at the beginning of this chapter: Pm (751) P(s,p)= pu pelae] (752) "4 ca alps” +4 cups + Cal) According 1 the boundary crossing theorem of Section 7.1.1, under certain conditions robust stability can be verified by looking only at the yo: axis. To 224 STRUCTURED UNCERTAINTY this end, we assume that the controller internally stabilizes at least a model with a (nominal) set of parameters, say, p =p" ja). In addition, the func: tions ¢,(p). = 1,...m, are continuous and do not vanish for p € a,b]. This accommodates many practical applications of interest. In this case, robust stability can be tested using the stability margins defined in Section 7.1.1, computed at each § = yoo. In particular, this structure includes many cases that may be recast in a LFT format as in Figure 72, with A ¢ Ay defined previously. These are ‘cases in which the characteristic polynomial coefficients ¢;(p) depend on the parameter vector p polynomially, for all i= 1,...,m. To this end a shifting and scaling of the uncertain parameters should be made: P= phehy 8 (733) PL = $(a+b,) (754) hy = Mbv—a) (758) Note that p; € (ai,bi] corresponds to 8; € [=1,1], and the nominal set of parameters to 6 = 0, i= 1,...,m_ The values of p’ and /h, can be incor Porated into the nominal model G(s). This is illustrated by the following example. Example 7S Consider the following set of models: ste Dee Goer) ZE[ez2 @€ laver] (7.56) which has an uncertain real pole and zero. This uncertainty structure can be transformed 10 fit the LET framework a follows 40 (s+ pQ)+ ids i - Ks 5] (5+ BN +2) [e( =I one fara (21 +22), 85,8) = PL = sai +2) thy = Hoya) The first term represents additive uncertainty and the second one quotient ype uncertain (see Chapter 2). Note that inthis case the weights are hy Wo) = = BT RereD sn) PARAMEIRICUNCERTAINTY 225 Defining the inputs and ouipats (u;.¥,), 1= 1,2, as the oupuas of the uncer tainties and weights as in Figure 7.1, we obtain the following for the system of Figure 72 0 fy Bae fy hy hols +p Gs) = | 2, He, tele pi (758) si pl 4p) + Bye pp —stet Gans PD with (49) the input and ouput to the system. The uncertainty structure is the set Ay as defined previously, now with m= 2 parameters (61,3). Therefore Gls, A) = FulGis).) 1 7.3.2. Research Directions The objective of this subsection is to provide a rough general idea of the main approaches pursued in this area of robust control. Due to the fact that this is still a very active research subject, the rest ofthis chapter will present only the basic results, from which many analytical and algonthmical tools for robustness analysis have been derived. A complete and very up-to-date tueatment of this prolific area of research can be found in the books by ‘Ackermana ((5}), Barmish ([35]) and Bhattacharyya and co-authors ((47}) Most of the research effort related to the robustness analysis of sys- tems with parameter uncertainty has taken place since the beginning of the 1980s. However, this problem had already been considered before by several researchers: Neimark ((213)), Horowitz (152), Sijak ((285)}, Zadeh and Desoer (328), and more recently Ackermann ([4). Horowitz templates gave origin to the value set approach ([35]), which uses the complex map of the characteristic polynomial at each s = jw to evaluate robustness against para metric uncertainty. The book by Zadeh and Desoer introduced the mapping theorem, which has been used extensively in computational solutions to the analysis problem. The more recent history of this area is closely related with the coefficient functions c(p) of the closed-loop characteristic polynomial P(s,p) P(s.p) Feip)s real), pe (a,b) ‘As mentioned before, when considering only parametric uncertainty, this is A general framework that includes the LFT formulation a5 a special case The complexity of the aforementioned functions of the uncertain parame- ters 6(p) determines the computational complexity of the solution, In addi tion, the type of functions considered leads {0 two clearly diferent research Sen Appts 226 STRUCTURED UNCERTAINTY Table 7.1. Coeticient functions, LFT structure, value sets and analysis results cLcP Framework ep) o=h Affine Multilinear Polynomial @) LET Framework 5, Structure M(s) rank 1 ‘M(s) rank 1 Independent 8, ‘M(s) general Repeated 8, Value Set Reciangle Polytope Nonconvex P(o.%)c eo1Plo, VI Nonconvex © IP, @ Result Kharitonov Edge theorem Analytical computational Computational M(s) general Pw, #) ¢ colP(w.} Figure 7.6, (a) ntowal,(b) ating, (¢) mutinect.and (4) polynomial. Here F = yw.) wih oR”. PARAMEIRICUNCERTANIY 227 approaches, Thisis illustrated in Table 7.1, which classifies the different tools according to these functions (see also Figure 7.6) The third row of this table, that is, the case of multilinear dependence of the coetficients with the parameters, establishes a boundary between two different research approaches. The first one seeks a small number of dis- tinguished models or test set of models ({35]), which may solve the analysis problem, hopefully with a small amount of computations. These are known as the Kharitonov type results, which start from the independent coefficient case and extend to particular multilinear dependence in the unknown parameters. ‘The second approach starts directly from the general multilinear dependence case and generalizes to polynomial functions c\(p). Many of these methods are based on the mapping theorem and have a clear computational basis. ‘As mentioned before, the general parametric analysis is NP-hard; therefore the procedures that are able to compute the stability margin exactly (or with guaranteed bounds) have exponential time complexity. Next, we comment in some detail on both these approaches. Kharitonov-like Results ‘The simplest case is when the parameters are the coefficients of the polynomial themselves, leading to the so-called interval polynomials. This also includes situations where the coefficients are functions fof independent sets of parameters, since this essentially reduces to the case 6. =p, As«will be shown in the sequel, in this ease the image af P(s, p) for all p © (a,b), defined as its value set, is a rectangle in C (see Figure 7.6a). The four vertices of this rectangle correspond to the only distinguished models that are necessary (and sufficient) to establish robust stability. This goes back to the 1978 result of Kharitonov (/172]), which was introduced to the control ‘community in 1983 ((27, 48)). In addition, the graphical result in [307] allows the computation of the stability margin for this class of uncertainty (see also f122)), ‘The first relevant generalization of the previous result considers the case where the coefficients are affine functions of the parameters. Therefore these coefficients are no longer independent, as in the previous case. The CLCP can be transformed to vols) + > pivi(s) (7.59) where v,(s) are known polynomials. The image in C at each w is a convex combination of certain vertex polynomials (P;(j0)....,Pijo)).€ <2”. This is a subset of the polynomials that correspond to the 2" vectors p « [a,b] of ‘the extreme values of all intervals. Hence the value set is a polytope in the complex plane (see Figute 7.6b). The first solution to this problem appeared jn 1988 ([3)) and is known as the edge theorem. The test for robust stability consists of a search over the CLCPS corresponding to all the exposed edges 228 STRUCTURED UNCERTAINTY of the value set, If these are stable, the whole set is stable. Further simpliica- tions of this result can be found in (31, 283] and [68]. The latter considers the characteristic polynomial of an interval plant, defined as a transfer function that has interval polynomials as its numerator and denominator, the set of parameters of each of them being independent. From the LFT perspective, it has been shown in [73] that both the interval and affine parameter depen- dence of the characteristic polynomials are special cases of det (7 — M(s)a\ A 4), when M(s) is a rank 1 matrix, as noted in Table 7.1 ‘The next natural step is to consider multilinear coefficient functions. The general multilinear case can be analyzed at the expense of an exponential growth in computations with the number of parameters. It is also well known ({35)) that there is no “edge” result in this case, that is, verifying the stability of the CLCPs of the boundary of the value set is not enough for the stability of the whole set. The worst parameter vector, in terms of stability, may be in the interior of the parameter set. In fact, interval matrices, realizable transfer functions, and random CLCPs may produce such examples ([35, 265)). However, there are results for specific multiinear structures such as the ones in (33, 71), which consider the CLCP corresponding to the series con. nections of interval plants. A particular case of this structure would have the following characteristic polynomial: Psp! spa?) = Ni(ssp!)-Na(s,p*) + Dils.q') - Da(s.q?) produced by the series connection of two plants of the form NGsypi) _ ST phy yo Le tps +P Dog) Hag Te gis ea per, der, prefab]. ae [da] $12, f=O..,m=1, k=0,0.n-1 B(s,p',4') ‘The results in [88, 111, 112, 270} provide polynomial time computation for systems with products and ratios of affine uncertainty structures, as follows: s,p) = Hes fils.p) R" ab), f= Ao-sm (7.60) Btn) Tey PERM BL (apd) P= Io. (7.60) where the (s,p), i=1,...4m represent polytopes of polynomials. This includes the case of products of uncertain poles and zeros, for example. ‘The general case of multilinear dependence with the parameters may have @ nonconvex value set in ©, Nevertheless, it can be proved that this value set (denoted as P(w,) in Table 7.1) is included in the convex polytape con: structed with the image of the 2” vertices of the uncertainty set {P(s,P). Pi 4, or b;, i = 1,...,2m) (defined as the convex hull and denoted as co[P(w ¥) in Table 7.1, see also Figure 7.6). This will be explained in further detail PARAMETRICUNCERTAINTY 229 when introducing the mapping theorem. Additional results in this area in: clude conditions under which the value set is a polytope, as if the uncertain coefficients were affine functions ({11, 150, 241]). This covers part of the third row of Table 7.1 When the CLCP has a polynomial dependence on the parameters, its value set cannot be included in the convex hull of the vertex polynomials, as in the previous case (see Figure 7.64). The model in (7.60) with numerator and denominator raised to given powers can be included as a special case of this uncertainty structure, A general setup that allows the analysis of complicated CLCPs in a polynomial computation time is the tree structured decomposition (30, 112). ‘Computational Results The computational approach is mainly based on the mapping theorem, which will be proved further along this section. A first step in this research direction was proposed in [254 based on the aforemen- tioned theorem. It was followed by the algorithm in {92}, which computes exactly the stability margin for systems in a LFT framework with multilinear parameter dependency. Note that the ¢j(A) are multilinear functions when the characteristic polynomial is det / ~ M(s)A], A € 4, and the parameters in 4 are independent. ‘A more general uncertainty structure isthe case of repeated parameters in A.€ Ap. This corresponds to CLCPs that have a polynomial dependence with the uncertain parameters. This fact suggests a simplification of polynomial dependence to multilinear dependence by adding equality constraints to the repeated parameters, Therefore, essentially the same procedures extend to the polynomial case ((266, 280)). Alll of the above algorithms are based on a branch and bound method over the two-dimensional value sets in the complex plane for each frequency «, A further simplification was introduced in [281], which applies the Routh Hurwitz procedure to eliminate the frequency dependency, therefore elimi nating the w-sweep for the computation of the stability margin. Furthermore, the value sets are reduced to one-dimensional intervals, which significantly simplifies the computations. Many other branch and bound procedures com: pute exactly the stability margin in these general cases ([17, 112, 250, 312)) However, the exact computation using these algorithms can be performed only for a small number of parameters, due to the combinatorial explosion with the number of parameters. There is yet another algorithmic approach that seeks an approximate tractable computation of the stability margin in the framework of the structured singular value ([326, 327]). The latter will be presented in the context of mixed parametric/dynamic uncertainty at the end of this chapter. The above description of this area of research is by no means complete And has only an introductory objective. A complete and up-to-date treatment of this subject can be found in [5, 35, 47] Inthe rest of this section we present Only the basic results, which initiated both of the approuches mentioned before: Kharitonov and the mapping theorems. 230 STRUCTURED UNCERTAINTY 7.3.3. Kharitonov's Theorem This result is the starting point of a prolific research area, which seeks the stability analysis of systems with parameter uncertainty based on the eval- uation of a finite (hopefully small) number of distinguished models in the class ‘The original result was presented in [172] and introduced in (27, 48] 10 the area of control systems. It considers the simplest case of uncertain character- istic polynomials, that is, interval polynomials. Although this is seldom the cease in practice, it generated a fruitful research area with generalizations of this theorem, which could be applied to practical cases, Consider the following family of closed-loop characteristic polynomials: Pls.c) = St eq ash tbo bes + ey (7.61) c= cH where 7 is an n-dimensional rectangle, as follows: S eR ne la.bl. OF lah), 1=0, .m=1) (762 The following notation will be used in the sequel to represent certain polynomials in the previous set, by means of their coefficients; by | (7.63) = [ag by as by (7.64) of bs ay (7.65) 0, = as by (7.66) eae oo) gag (768) where we have assumed without loss of generality that n is a multiple of 4. For example, a polynomial can be denoted as P(s, E}, 0"). Lemma 7.3. The image in C of the polynomial P (j«,c) for a given frequency w and all values © € 1 is.a rectangle with vertices P (yw, Ex, Os), P(}e, £y,02) P (ju, £3.01), and Pye, E>,02) PARAMETRIC UNCERTAINTY 231 Proof. The set of characteristic polynomials (7.61) can be decomposed as follows PUo.e) = Re|Pyo,6)] + m[P(e,¢)] (769) Bee [P (yu e)] = ca ~ c20? + cua" — eu’ + (7:70) Im|P(ya,0)] = cw ~ ew + es0* — Gu" + any Hence the real and imaginary parts of P (jw,-) are functions of independent subsets of parameters, and therefore their maximum and minimum values are independent as well. Due to the linearity of P(jw, ) in the parameters, Re [P (Jo, E>, 0)] < Re |P(jw,6)] < Re [P o, £1,0)) Im |P(jo, E,O2)] < Im|P(jw,)] < lm|P (jw, E,O1)) AS a consequence, the four extreme vertices of the image are P (jw, E3,01) , Pw, E102) , P(e, 2,01), Ply, E02) (7.72) a A basic element in the proof of Kharitonov's theorem is the interlacing property derived from the Hermite-Biehler theorem. This is a property of Hurwitz polynomials, defined as the ones with all their roots in _. To this end note that any polynomial with real coefficients can be put in the form P(s) = f(s?) + sgls*), where f includes all even coefficients and g all odd coefficients of P(s,c). In addition, a necessary condition for a polynomial to bbe Hurwitz is to have all of its coefficients non-zero and with the same sign. This will be assumed in the sequel. Theorem 7.5 (Hermite-Biehler) A polynomial P(s) = f(s?) + sg(s?) with real coefficients is Hurwitz if and only if the roots {x!,....xh} of f(x) and (xf, cxh} of g(x) are all real and negative and satisfy the interlacing property ahah cabcafcr ah <0 Proof. Sec {129) for a detailed proof a When the functions f and g have the above property they are called a Positive pair ({129)). “ Accunding. tothe order of Pts) the degree of ts) may be fon = 1. 232 STRUCTURED UNCERTAINTY Note that the functions /(s°) and sp(s*) at s = jw correspond to the real and imaginary parts of P(w,c). Furthermore, a negative root of f(x) corre sponds to an imaginary root of f(s®), and similarly for g. Next, we present the main result of this subsection. The proof is simpler than the original one in [172] and can be found in (322. Alternative proofs have been presented in (54, 69, 207] ‘Theorem 7.6 (Kharitonov) The set of interval polynomials in (7.61) is Hur- wite if and only if the following four elements of the set are Hurwitz: Puls) © P(s, 1,01) 073) Pyls) © PCs, 1,02) 774) Pals) © P(s, E201) (735) Pals) 2 P(s, Ex, 02) (776) Proof. Necessity is trivial, due tothe fact that all four polynomials belong to set (7.61). Sulficiency is based on the following argument. If P (ja, £*,0}) and P(jw, E*,02) are Hurwitz for a certain vector of even parameters E*, then any element of the set P(ja, E*,O) is Hurwitz, where O represents the set of all odd coefficients in %. This holds because P (so, E*,O1) = fw, E,) + jog(~w", 01) am P (Jo, E*, 02) = f(-w*, E,) + jog(—08, 02) a) being Hurwitz imply that f and g are positive pairs. From Theorem 7.5, P (J, E*,O;) and P (Jw, £*, 2) alternatively intersect the real and imaginary axes m times, the latter being the degree of f and g° Because g depends only on the odd coefficients, which affect only the imaginary part of set P (jw, E*,0), we have Im[P(ja, E*, O2}] P(jw, £),0) Hurwitz (7:9) Pow. , PUw, 5 } Hurwitz + P (jw, E,,0) Hurwitz (7.80) and finally, Pw, £,,0) er Plies Esoy Hurwite=> PUw,£,0) Hurwitz (781) equivalent to set (7.61) being Hurwitz, a 7.3.4 Mapping Theorem ‘The computational approach to the analysis problem can be stated in terms of the value set of the CLCP, represented by P(s,p), p € a,b, in the general case, or by det | -M(s)Al, A ¢ &y in a LFT framework. In any case, the objective is to compute the value set in € (or rather its boundary) to check whether or not it includes the origin z = 0. In such a case P(jw?, p*) = 0 for a \ertain combination of parameters p* © , which means that the system has an unstable pole at s = jw. As noted in Figure 7.6, except for the interval and. affine cases, the boundaries of these value sets cannot be defined with a finite ‘number of points, that is, the vertices, Nevertheless, in the case where P(s, p) is multilinearin p, the mapping theorem provides an important simplification, which allows the bounding of the value set To this end, consider the set of CLCPs {P(s,p), p € {a,b)} and define HE DERN, pela (7.82) VE {p=a oF py sm} (783) as the parameter domain } and the set of its 2” vertices, respectively. Note that here 2 is an n-dimensional rectangle that contains all possible uncertain Parameters p, and should not be confused with the domain of coefficients defined in the previous section. We are particularly interested in the image of all vertices in Y through the function P(jw,-), which we denote as P (jw, Y). In this case, the convex bull co| (jo, ¥)|*is defined by r vertices with r < 2”, as seen in Figure 7.6b. In the case of any multilinear complex function, the following result holds (328), "The conven hull, cS) of set i the smallest convex set that cans 234 STRUCTURED UNCERTAINTY ‘Theorem 7.7 (Mapping Theorem) The image of any multilinear complex function F :H + C is contained in the convex hull of the image of the vertices, Of H, that is, FH) © colF(V)} Proof, First, we show that the map of F along, a coordinate axis of set 1 is a line in C. This is clear considering that a coordinate axis in 2 fixes all parameters but one, that i, p; = x, for alli = 1,...,m.i 4 j. and p, € [aj,by) Therefore by the multilinearity of F, the image is affine in py Next, consider any point p* € 7 and a line that contains it and is parallel to the first axis of 2. This line intersects the opposite (n — 1)-dimensional “faces” of 7 at p™ and p*, where Ph =(b Pp} Pal” (7.84) pha (a Ph Pia (7.85) ‘The new points p™ and p** can be considered as generic points of the (rt — 1)- dimensional faces of H mentioned before. The same procedure can be applied to each of them, to generate four new points, each in a different (n~2)- dimensional face of 1 These are ph = [by be Ph Pol (7.86) oe a Pi Pal? (787) oe be Pi Pall (788) ae Po tml? 789) Each of the segments p™ "pe, and so on defines a linear image in C due to the fact that they are parallel to the coordinate axes ‘of 1. In addition, the image F(p") is contained in the segment F(p™p™). ‘The extremes of the latter are contained in the segments F(p*p') and F(p™ p=), respectively, and so on. Therefore the image of the first point @*) (which s generic is contained i the convex hull ofthe image ofthe vertices, that is, col (V)] In particular, this can be applied to the complex CLCP P(jw,p) over the set of parameters p € H. The above theorem provides an important tool for assessing robustness of a parametric family of models. A computational procedure applied to the image of the vertices P (jw, V) determines, at each frequency w, a sufficient condition for robust stability when this polytope does not contain the origin. ‘This is due to the fact that P(jw, V) includes the actual value set. However, this only provides a qualitative answer: stability instability PARAMEIRICUNCERTAINTY 235 If, instead, we seek a quantitative solution to the problem by means of the computation of the stability margin, the above theorem provides an upper bound in the case of 13, or a lower bound in the case of ky,, To compute either one of these margins we should scale the parameter domain # by k ¢ (0,20) until its image P (jw, k) intersects the origin. This scaling is exactly the stability margin (see Definitions 7.2 and 7.3). To this end, the following, property is useful, the proof being left as an exercise (Problem 5), Lemma74 The following holds when scaling a multilinear complex function. F: HC by real numbers ky > ky > 0: Flky-W) 3 F(ky Hh) (7.90) 0 [F ka 0] 9 eo fF Ukr 70) ay) ‘There are still certain issues that may limit the above result. The first one is the fact that the mapping theorem applies only to multilinear functions. The second one is the fact that it is useful only to bound the value set at each frequency w; therefore the procedure should be repeated over a grid of frequencies in w € {0,0} Next, we present further simplifications to the above problems. The first cone transforms a set of parameters that appear polynomially in P(s,p) to ‘a new set that appears multilinearly and has equality constraints. Therefore the above theorem can be applied. To this end, we consider for simplicity and without loss of generality the LFT framework. In the case of polynomial parametric dependence, this setup is equivalent to the CLCP provided that 4, is allowed to have repeated parameters. In the sequel we denote these As as ,,. Ibis easy to establish the equivalence between the following uncertainty sels Bye 2 | SiB( B15. Bip ey Biosys BmeeesBe) (7.99) BER, FH], and A, of dimension Mg, restricted by the following equality constraints; ee (793) The (closed) hall Bay is defined for |5,) 1.7 = 1... 236 STRUCTURED UNCERTAINTY ‘A second simplification consists in the elimination of the frequeney search. ‘This not only reduces the computational burden but also solves the problem Of gridding the jw axis, especially due to the fact that the stability margin for pure real parametric uncertainty can he discontinuous ((32, 221, 284)). One ‘way to perform this simplification is by applying the Routh-Hurwitz crite ria to evaluate the stability of the CLCP. In this way, n positivity conditions. should be verified to guarantee stability, n being the order of the polyno- mial P(s,p) in the variable s. In general, these inequalities have polynomial dependence with the unknown parameters but are real functions, indepen- dent from yw. Therefore the mapping theorem can be applied along with the previous simplification over the 1 positivity constraints in order to verify the intersection with the origin. This eliminates the frequency search and trans- forms the two-dimensional mapping of P(s,p) to n one-dimensional ones, which simplifies considerably the software and reduces the computation time. In fact, the convex hull of each of these functions is a real interval, which coincides with the actual value set of the same function (Problem 6), that is, F(M) = colF(V)) when F : + R. The increase in the number of functions to be evaluated is not a problem, because it is O(n), where n is independent of the number of parameters; as a consequence, it does not add complexity to the algorithmic solution, However, in this transformation of the problem, a symbolic manipulation procedure should be used in the case of large number of parameters ((281|). Another efficient method to eliminate the w-search, which directly deals with the state space realization of the plant in a LFT framework, can be found in (284), Finally, applying the mapping theorem with these simplifications in mind, fa program can be constructed to scale the parameter set} until its image (1) reaches the origin, The value of the scaling parameter at this point is the stability margin. To this end several branch and bound procedures have been proposed (17, 92, 250, 266, 281, 312, 326)). These iteratively compute ‘upper and lower bounds of the stability margin and, by different mechanisms, partition the parameter domain and compute bounds for the new subdomains. Finally, with the aid of certain criteria to eliminate some of these subdomains, the procedure continues until the bounds converge to a predetermined error. ‘One of the important issues in the implementation of these programs is the increment in the number of regions after each partition, which depend on the criteria adopted to eliminate some of these subdomains (|265, 325, 326)) ‘Another important point is the computation time of the bounds at cach subdomain, which in many cases seems to dominate the total computational time, for reasons we will explain next. In addition, it is important also to determine the “offending” parameter vector: this is the one that maps the scaled set km? to the origin, Here ky is the multivariable stability margin defined previously. Tis function could be either P.M) al each Feequency or the real-valued Reuth- Hurwitz funtion. MIXED TYPEUNCERTAINTY 237 In any case, the first step that maps the image of the vertices involves 2" computations of a function f(24), one for cach vertex of set 74, m being the number of parameters. This is related to the problem being, NP-hard. Therefore these algorithms will work reasonably well only for a small number of uncertain parameters, say, 10 to 20, Nevertheless, this may include many practical cases of interes, as, for example, the robustness analysis of aircrafts with uncertain aerodynamic coefficients. (See Chapter 11 and [265) for the analysis of the X-29 aircraft.) With the approach considered in this section, the only way to achieve tractable computational time would be with the use of some extra a priori knowledge of the model structure. This a priori information can be used to eliminate a number of vertices of %, in order to decrease the number of com: putations of f(74) to a polynomial function of m. For example, with m = 10 parameters the computation of all vertices involves 1024 evaluations of f(} ‘and an O(n?) algorithm would involve roughly the same number. Instead, for ‘m =, the image of the vertices involves approximately 10° computations of f() and an O(n) algorithm only 8000 evaluations. This extra information should therefore reduce the number of vertices to be evaluated by two orders ‘of magnitude in this last case. Alternatively, only approximate algorithms with no guarantee of conver- ‘gence to predefined error bounds can be constructed. These procedures could also use a branch and bound approach but with polynomial time computation at each subdomain, Therefore if the increase in subdomains remains poly ‘nomial, the complete computation is tractable, allowing for solving problems with fewer than 100 parameters. In this ease the objective is not a precise answer, but a fast one. It seems that a good approximation (10-20% er- ror) can be achieved in many practical problems. These algorithms will be presented in the next section in the context of mixed parametic/dynamic ‘uncertainty structures. 7.4 MIXED TYPE UNCERTAINTY 7.4.1. Introduction From the practical point of view, uncertainty is present in the plant in both the dynamics and the parameters. Usually the lower frequency part of the plant ‘may have a well known model structure, with parameters computed using identification methods. The upper-fequency portion may be unknown but bounded. This is typical for large mechanical flexible structures ({121), for Which well known second-order models fit the lower frequeney portion and where high-frequency oscillations, which have less detailed information, may be covered with dynamic uncertainty. The parameters, natural frequencies and damping coefficients are estimated within real bounds. Furthermore, even for @ very well known mathematical model (fixed 238 STRUCTURED UNCERTAINTY order), which may not have dynamic uncertainty, it has been shown in a previous section that performance can be inserted as an extra dynamic un: certainty performance block (see Figure 74) In addition, as mentioned in the last section, the pure parametric uncer- tainty case can be discontinuous in the problem data ([32, 284). As shown in (221), a regularization of the parameter uncertainty, by adding some phase to cach parameter, may prevent the stability margin from being discontinu: ous. In fac, it has been shown that adding at least one complex uncertainty block, for example the performance block, to the problem may eliminate the discontinuity as well. Ths is important when considering the gridding of the @ axis to compute the stability margin, either (jw) oF kn( jo). Therefore from a practical point of view, it seems natural to attempt the analysis and design considering a mixed type uncertainty structure. From the point of view of computational complexity, due to the fact that the computation of the stability margin for the pure parametric case is NP hard, it seems reasonable to expect the same limitation in the mixed ease. Nevertheless, the addition of complex dynamic uncertainty does not inere= ‘ment the complexity In the area of mixed parametric dynamic uncertainty several results have appeared since 1988; among others we can mention (28, 70, 149, 151, 282, 318] for the combination of parametric and global dynamic uncertainty and (118, 267, 325, 326, 327] for the general case. In can be shown that in the ‘mixed case, the minimization of the upper bound of a stil leads toa convex problem ((267, 325)). Therefore the addition of parametric uncertainty does not destroy the convexity properties of the structured pure dynamic case In the LFT framework, one approach to deal with this type of uncertainty structure is by partitioning the uncertainty matrix into two parts as indicated in Figure 7,7. Therefore, for the system M(s) to be robustly stable for the mixed type structure dy, itis necessary and sufficient to have ((267)): Ae =| M(s) M(s, 4p) Figure 7.7. LF approach for combined uncertaiy. MIXED TYPEUNCERTAINTY 239° 1s, 8,) © FIM (9). sable for a8, € Bay 2 us, [1Ge.8)] <1 fora, 84, anda Condition 1 is a necessary condition and establishes that the “nominal” model M(s,4p) should be robustly stable, when only considering parametric ‘uncertainty: Under certain conditions it can be verified by checking whether oF not Ha, [Muay] <1 (794) for all. If this holds, then condition 2, which represents the stability margin with respect to the pure complex case, is necessary and sufficient. As before, the latter is checked via its bounds. Thus, a sufficient condition for robust stability in the mixed case is: int 7 [patue, AyD} <1, vw (795) for all dy € Bay. Here D is the set of scaling matrices introduced in Sec- tion 771) Fauation (7.5) is alsa a margin computation with respect to pure parametric uncertainty. In this way. we can transform the problem of mixed Uncertainty into two stability margin computations with pure parametric un- certainty. This also supports the fact that the mixed case is NP-hard, due exclusively t0 the pure parametric case. As a consequence, the same branch and bound algorithms constructed for the parametric case ({267]) could be used in the mixed case as well, bearing in mind the limitations due to the computational complexity of the problem. On the other hand, for a large number of uncertain parameters (up to 100) there is no other way to approach the problem but by bounding the stability ‘margin in polynomial time, although these bounds will have no guarantee of reaching a prescribed error. A branch and bound algorithm that proceeds along these lines is available in a commercial package ((20, 325}). Next, we present the basic results that are used in such an algorithm. 7.4.2 Mixed w This section follows the research effort aimed to construct bounds com Putable in polynomial time for the stability margin in the case of mixed {ype uncertainty. For further details the reader is referred to (38, 216. 302. 325, 326, 327]. The uncertainty set for mixed parametric-structured dynamic uncertainty has been defined in (7.19). For this same structure, here we also specify the repeated parametric uncertainty case, already defined in (7.92) 240 STRUCTURED UNCERTAINTY and the repeated complex scalar ease defined as follows: , 4 ding (BB. Bed (7:96) ec, em ‘Therefore the mixed uncertainty structure is as follows: block diag (Ai, 42,45) BE yy ae) cay, a ase ay Next, define the following sets of matrices: ae (1,-1), a 24 facay| 5) = m+ me aay ” block diag (D,, dd) Di>0, DEC, 71cm D.>0, Die, fam, +1,..-yme d)>0, djeR, j=mtme+ly..m a} Dlock diag (Diy... Drvames demerit: -sdml) G ig Block diag (Gi,..-s Grn Geers, ded) 40, Dee ™ det(D,) 40, DiC, i= m+1,...4m,+me 4;40, djeC, j=m,tmerlesm [meson Bq, "| BER, 7=1,...,Mp Here Mg has been defined in (7.93) and 0 and / are the null and iden- tity matrices, respectively, of appropiate dimensions so that the elements of the previous sets may be multiplied by the ones in Ay. In addition, define MIXED NPEUNCERIAINTY 247 (A) © max{|Ai, Ay is a real eigenvalue of A} and the structured singular ‘value as in Section 7.1.1, now with respect to dy Property 5 of Section 7.21 still holds for this more general structure. In addition, using the definition of y, the following is also true (see property 3 in Section 72.1 ay M) = max pa( Od) (758) It is clearly easier to search over the boundary of the set BAy than over the whole set. For this reason, taking into account that the boundary Qc BAy, the following bounds are obtained: maxpa(MQ) < say(M) < inf @(DMD~) (7.99) ‘Next, two theorems are presented that prove that the lower bound is tight and a generalization of the upper bound can be made tight as well, in the case of less than three complex blocks. ‘Theorem 78 max pn(MQ) = way(M) (7.100) Proof. See {325} o Theorem 7.9 Let: 4, = inf [min (a € R| [M"DM +)(GM~M"G)~ aD) <0}] (7101) Then if a <0, Hay = 0, otherwise Hay S Vee (7102) Proof, See (115, a ‘The lower bound is a nonconvex optimization, as with the pure dynamic uncertainty structure. The upper bound is a convex optimization problem due to the fact that tis a LMI in both G and D. Therefore the global minimum can be computed efficiently with the use of commercial software packages (20, 126) Note that inthe case of only nonrepeated structured dynamic uncertainty we have m, = me~0, hich imply G = {0). Therefore the previous theorem reduces to a, = inf min {er | (M*DM ~ aDj <0}] (7.103) 242 STRUCTURED UNCERTAINTY which can be shown is equivalent (0 the upper bound in equation (7.40), Section 7.2.1 (see Theorem 7.3 and Lemma 7.2). In [526 for computational purposes the following alternative characteri- zation of the upper bound was introduced: ‘Theorem 7.10 Consider a constant matrix M € C"*' and a real scalar > 0; then there exist matrices D € D and G € G such that a[M-DM + (om M°G) Po] <0 (7.104) if and onty if there exist mares B € B and G e G such that z [wn e) (GoMo” -18) (4) “| <1 (7108) From this theorem itis clear thatthe minimization problem to compute jy ccan be carried out alternatively as the computation of the minimal 6 > 0, such that 3 [Gy (Zdmb~*— 6) ue wet 106 Although the above probleut is uu longer convex in D and G, the former enters the problem as in the pure complex case, the latter is now a real diag- ‘onal matrix, and the minimization of the norm 7() offers certain numerical advantages. It seems that a good choice is initially to solve the previous min- mization, which provides an initial guess for the LMI problem in (7.101). ‘This has been implemented in commercial packages ({20)) that compute the mixed y stability margin. 7.5 PROBLEMS 1. Show, in a loop block diagram example, the signals w and z that guarantee that stability of T,.(s) is equivalent to nominal internal stability. In the same way show for which signals this is not the case. 2 Prove tha u(MU) = w(M) for U €U, with U8 (Ue de, for U, unitary i= I,m} 23 Prove that x(DMD°) = u(M) for D € D, with ano 4 Jy: identity in 7, z der. Oe ale PROBLEMS — 243 4, Prove that (7.47) holds when the uncertainty structure is global dynamic and Ti(s) and Tza(s) represent the transfer matrices to be tested for robust stability and nominal performance, respectively. Verify that equality hnolds in the SISO case. 5. Prove Lemma 7.4. 6. Consider a real polynomial P(p) that is multilinear in p € 71. Prove that cofP(0)] = PCH). c! CONTROL &1 INTRODUCTION Up to now, most of the material in this book has been devoted to the case where the signals involved are £2 signals Thus, in this contest, itis natural to assess performance in terms of the L>-induced norm (ie., the Hx, norm) and to consider the effects of £ bounded model uncertainty. However, in several situations of practical relevance the £2 norm is usually not the most adequate to capture the features of the problem. Examples of this situation are cases involving time-domain specifications and persistent, unknown but bounded exogenous signals, such as when it is desired to minimize the worst tracking error in the presence of persistent, bounded noise, while, at the same time, keeping the control action below a given value. Another common situation where the need for a worst-case time-domain norm arises is the case where the plant model being used for synthesis has been obtained from the linearization of a nonlinear plant at a given operating point. In this case the states of the plant should be kept confined to a region around this nominal condition where the linearized model is known to be accurate enough. ‘While conceivably these problems could be addressed by using the H and Yq, methodologies discussed in Chapters 5 and 6, this will entail the use of Weighting functions in an attempt to recast the time-domain specifications in terms of the £; norm of a suitably weighted output. Thus we can expect at best a lengthy trial-and-esrortype process without guarantee of success. Rather than pursuing this approach, in this chapter we will develop tools to handle directly cases involving persistent bounded signals and performance specifications given in terms of peak time-domain values, that is, the La. 245 246 Cc! CONTROL Consider the prototype control loop used in Chapters 5 and 6 and shown again in Figure 8.1. As before, the goals to synthesize an internally stabilizing controller K(s) such that the worst-case value ofthe output 2(0) (measured now in terms of its peak time-domain value) to persistent exogenous distur- bances w(t), |w(0)| < 1, is optimized. Denoting by [l|z,--c,, the C.-induced norm of a given mapping, this problem can be stated precisely as 1g FHC Rec 1) Since for LTI stable systems the £._ 10 Ca-induced norm coincides with the £; norm of the impulse response matrix (see Appendix A), this problem is known as the £; control problem. The problem of rejecting persistent disturbances has been considered as far back as the early 1970s, [39, 40, 131) These papers addressed the problem of finding a static state-feedback control, possibly under control input con straint, guaranteeing the permanence of the state in a given time-dependent set, in the presence of unknown but bounded persistent disturbances. The problem was solved by finding a sequence of sets (the reduced target tube) in which the state could be confined by means of an appropriate control action. While some progress was made in this direction, this line of research was abandoned, probably in view of the computational complexity involved in implementing these controllers, not compatible with the computer tech. nology available then, in all but the simplest cases Interest in this problem arose again in the early and mid-1980s, when it was shown that in the case of linear time invariant plants it could be recast into an equivalent C, model matching problem and solved using linear programming (26, 81, 82, 315} The approach that we pursue inthe fist part of this chapter follows closely after the work of Dahleh and Pearson ((81, 82)). For reasons o 0) ’ Kl Figure 8.1. General loop stuctutefor he £, problem, ROBUSTSTABLITY REVISED 247 that will become clear latter, we also will depart from the approach taken so far and consider discrete, rather than continuous, time systems first. The continuous time case will be reduced to an equivalent discrete time problem using the simple change of variable proposed in [41). Before proceeding to solve the £; control problem, we explore the con: nections between £; norm minimization and robust stability. 2, ROBUST STABILITY REVISITED Up to the present point, the only class of dynamic model uncertainty that wwe have considered has been linear time invariant with bounded £;-induced norm. In this context we developed in Chapter 2 (for the SISO ease) and Chapter 4 (for MIMO systems) necessary and sufficient conditions for £> stability of the form SBR | Wile <1 (2) ‘Next, we want to extend the class of admissible perturbations to include non- linear, possibly time varying operators. To this end we begin by introducing the concept of causal operators. Definition 81 An operator M:tf +, is causal if PM ~My for all k (83) where Px denotes the standard truncation operation, that is, Pa Cop Xt os ¥4-) = (ho 284,050.) 4) y PiM = PMP... for all k (65) then M is said to be strictly causal, That is, an operator is causal if its present output does not depend on future inputs. If, in addition, the present output does not depend on the present input, then itis strictly causal Definition 82 A mapping H:° —+ €P is said to be & stable if 1. whenever x © € then Hx € €?, and 2. there exist finite constants k and b such that [aly < kijllp +b, Vx ee” 248 ¢! conTROL Note that in the case of arbitrary nonlinear operators both conditions are required. For instance, the operator H(x) defined by (Hx)(é) = x(i)? maps € into itself. However, the second condition fails since clearly no suitable constants & and 6 can be found. When p = oo, Definition 82 reduces to the usual bounded-input bounded- ‘output (BIBO) stability. If, in addition, we restrict our attention to linear time invariant mappings, a necessary and sufficient condition for é stability is given by the following well known result. Lemma 8.1 Consider a LTI operator Hit —+ & and let {hy} denote its impulse response. Then H is € stable if and only if {h,} € € Proof. (Sulliciency) For any x € &* we have that (see Appendix A) WHeloo < WH l= 1 loo 86 =U Therefore Hx € & and condition? in Definition 82 and b= 0. (Necessity) Assume that h ¢ £1 => S2%5 [hi] = oo. Consider the following input x = {x(9},x(0) = sign|A(N ~ fn this case we have that (Hx)(N)| itolful + 90. Thus Hx g &. . s satisfied with k = Ji Define now the class of nonlinear, causal perturbations with bounded induced ¢” norm as! BNCTY = [dlp « 2 le induced < i} A: —+ 6”, causa: sup { vo Pl where + denotes convolution. Our goal is to assess internal stability of the interconnection shown in Figure 82 when A is restricted to some appropri. ately chosen subset SC BN'LTV. The first step is to extend the concept of internal stability introduced in Chapter 3 Definition 83 The feedback loop shown in Figure 82 is internally & stable if whenever the inputs |uy(s),12(s)} are in & the corresponding out- ‘Puts [e\(s),€2(s)] are also in &?, and, in addition, there exist finite constants k ‘and b such that Hetllp $k (lel + ally) + (8.8) Heal Sey + Ie) +6 6) " Without los of generality we tke this Bound tobe ROBUST STABLITYREVISIED 249 w Ate) a-s ow Figure 8.2. Prototype loop interconnection. A sufficient condition for stability is given by the well known small gain theorem. ‘Theorem 8.1 Consider the interconnection shown in Figure 82. Assume that M:0 — & is & stable and that the loop is well posed for all S © BNCTY. If |Milee-ue < 1, then the interconnection is stable for all 4 © BNCTY. Proof, From the block diagram of Figure 82 we have that Neill 0. a In the case of linear operators, the small gain condition guarantees that the Joop is well posed. However, this is not necessarily true in the case of general NCTY operators. The following lemma provides a sufficient condition for well posedness in this case. Lemma 82 Consider the interconnection shown in the block diagram of Figure 82, where M is a proper linear system and A is @ NCTV operator. If is swicly causal then the interconnection is well pose. 250! CONTROL Proof. Since M is proper and A is strictly causal, it follows that the operators ‘MA and AM are both strictly causal. Thus given the input sequences u; and tus, the equations e)=uy ~ AME; + Aue =u ~ Muy ~ Mdez ccan be solved recursively to obtain the (unique) solution ¢,, 3 (un) Given that the small gain theorem provides only a sufficient condition for stability, question that arises is whether or not this condition is conservative. In the case of & stability, since the set BN'CTY includes all perturbations & € ‘BY.., it follows from equation (82) that the small gain is both necessary and sufficient for stability. Following along these lines, itis natural to investigate whether or not necessity holds in other situations and, in particular, in the case of £ stability. As we show next the answer is positive, provided that the set of admissible perturbations is enlarged to include either linear time varying or nonlinear operators. 1 Robust Stability Under LTV Perturbations Consider tinear time varying causal operator L.:( —+€ defined by its emel L(i,j), that is, ify = Lu then yt) = OG kate) 2) ‘An alternative representation of this operator is via an infinite-dimensional Tower block triangular matrix £: 10,0) 14,0) 10,1) ‘ i (13) LK) LRA). LKR) Clearly a necessary and sufficient condition for the operator L to have bounded induced norm is sup [E(k [p(t BC bil > Pal 6 jn smallest integer such that [y(in)] > [y(n] Consider the following operator defined in terms ofits matrix representation: ft) ta) o sta) ye) (817) 0 § 0 oa) Sis) (is) By construction A is a LTV causal operator such that € = Ay. In addition, since each row of & has at most one nonzero element itis easily seen that Wage < [Alex = sup Mel Le Pia (818) and thus A € BCTY. a With this result we are ready to show now that the small gain theorem is both necessary and sufficient for © stability provided that the set of admis- 252! CONTROL sible perturbations is enlarged to BCT. This fact was originally established in [84]. The simpler proof that we provide here follows after [86] and con- sists of two parts: (i) construction of a suitable input signal resulting in an unbounded output; and (ii) contruction of an admissible perturbation that ‘maps this output to the input. Combination of (i) and (ii) results in a loop where a bounded input generates an unbounded internal signal, thus violating internal stability, ‘Theorem 82 Consider the loop interconnection of Figure 8.2. Assume that M isa LTI stable system and that the loop is well posed. Then the interconnection is stable for all 8 € BCTV if and only if UM llmee 1 (19) Proof. Sulliciency follows from the small gain theorem. To establish necessity assume that [Mlle~e > > 1. We will show that in this ease there exists & ‘destabilizing A € BCTV. Proceeding as outlined before, we frst construct an input sequence uy and a signal e, such that 1. fim [P(e lle — 20. 1 2. TllPeealle 2 l\Peeille for all k y For simplicity in the sequel we will assume that M has a finite impulse response of length N2 Select the first N elements of the sequence e; s0 that |IPu_t¢i [luo = 1 and [yy(N ~ 1)| 2 +. Select the corresponding sequence uy a8 u2(8) = 7 sign(ya(i)). With these choices we have that le(Kl=r+bil 2% & lea(W — 1) = y+ bn(N -1)) > 27 ihe? (820) Select now the next NV elements ofthe sequence eso that [Payee = 2 and |y(2N —1)] > 2y, with up(i) = —y sign(y,(i)). With this choice we have now that |P,(es) 2 vlPseilin, K< 2N—1 and [ex(2N~ 1) > 3y. Continuing along these lines, pick the mth chunk ofthe sequence e, so that [WPmn-1(€1 )llac = m and |y(mN — 1)| > mry. It is clear that with these choices both conditions 1 and 2 above are satisfied. From Lemma 83 it follows that there exists a causal operator 8 € BCTV such that ¢) ~ Aep, Moreover, by selecting the sequences so that ey(0) = Oand es(0) #0 (eg. by setting us(1) 7) it follows that A is strictly causal and thus the loop is well posed. To 2 This assumption can be removed by using the fac tha since M is stable, for any € > 0 there exists N such that J JM] <¢, where Mf denote the Maskov parameters of Mf (se 8 for etait. ROBUSTSIABIUTYREVISIED 253 complete the proof we need to show that this operator A destabilizes the loop in Figure 8 2. This follows immediately from the fact that for this choice ‘of 4 the bounded input signal up results in an unbounded internal signal e, a 2 Stability Under Time Invariant Perturbations In the last section we have shown that the small gain theorem is necessary and sufficient for € stability under linear time varying perturbations. Next, we show that this is also true for the time invariant case, provided that the perturbations are allowed to be nonlinear causal mappings. The proof is simi lar to the proof of Theorem 8.2 except that the destabilizing perturbation that maps ¢2 to ¢; is selected to be NLTI rather than LTY. To this effect, denote by 5; the “shift-ight by i" operator, and given any sequence f = {f(k)} € © consider the following mapping: ane {e* 7A, CPUS = Pasi for vome { coy It is easily seen that this perturbation maps e, to #2. Moreover, since by con- struction "Piero > [|Peeil}xx it follows that Alm r= < ¥! < 1 Finally, it can also be shown that S\A = AS; and thus A is time invariant. Combining this result with Theorem 8.2 yields the following wiollaty. Corollary 8.1 The condition |(M Ij~e~ <1 is necessary and sufficient for ‘he stability of the the loop interconnection of Figure 8.2 for all causal S with [dlenve= <1 whem & is either a LTV or NLTI mapping, Finally, we address the issue of the conservativeness of the small gain theorem for systems subject to LTI perturbations, Consider an ¢™ stable LTT system Hand let h(i) denote its impulse response, Since H is & stable, the sequence (,} ¢ € and therefore H(2) = SE ght()z~" is analytic on the exterior of the unit disk and continuous on the unit circle, It follows that H(2) € Hay. Moreover, for any «w Wate) <0 Uh tlh (822) Hence |\H ln. S ht. Thus, for LTT systems, &* stability implies & stability 1e converse, however, is not true as illustrated by the following example, Example 8:1 Consider a LTI system having as z transform H(z) = exp {z/(1 ~z)}. Ie can be shown that Wl, = ess sup exn( |[-ex0-05) (823) 284! CONTROL BLTT, @ () Figure 8.3. Comparison of “=-and (4nduced unit balsa) Lt case and (b) LTV cate. Therefore H(z) € Hoo and the system is © stable. On the other hand, since H(z) isnot continuous on the unit circle its Laurent series expansion does not converge absolutely there. Hence h(i) ¢€ and the system is not € stable. From this discussion it follows that in the ease of LT perturbations the &. induced unit ball BCTT., is strictly contained inside the ¢-induced unit ball RTT, as illustrated in Fighre 83. Thus condition (8.19) ie more conservative than (82). Moreover, given any 1, itis relatively simple to construct a LTT stable operator H € BLTZ; satisfying WAlle Qn+ WMA, = Qn +1) (824) For example, consider a system having the following z transform: Gta) = FIST, O 1 have multiplicity 1 Assumption (A4) is made for convenience, in order to simplify the subse- {quent developments. We will briefly indicate how to relax it at the end of the section. On the other hand, assumption (A3) is required for the solution to the problem to be well behaved. As shown in [317], if this assumption fails the optimal approximation is not a continuous function of the problem data, Lemma 84 Assume that T:(z) € RE" has n single zeros 24,22, ..+.2m,|2il > 1, and no zeros on |z| =1. Then, given Re RE", there exists Q RE! such that R = T;«Q if and only if Riz) =0, ores (8.29) Proof. (Necessity) If QeRE then G(z) is analytic in fe|> 1. Thus ix(z,)O(z,) = 0, which implies (8.29). (Sufficiency) Since R = T; + Q € Reé', the only situation where Q ¢ ¢ is the case where one of the unstable 2er0s of 7s(z), sayz, cancels an unstable pole of Q(z) at that location, However, since all the nonminimum phase 2er0s of fy have multiplicity 1, this implies that 73(z)}03(z) # 0, contradicting (829), a Let S={RERERZ) 12cm} (830) 1m terms of this set the ¢” model matching problem can be recast as the following minimum norm probiem: int in — Rll (331) Note that this problem has the same form as the minimum norm problems set in general & spaces discussed in Appendix A. As we show there, these problems can be solved by exploiting the following duality principle stating the equivalence of the original problem (the primal) and an optimization problem set in the dual space (the dual problem) ‘Theorem 83 [187] Consider a real normed linear space X, its dual X*, a subspace MCX, and its annihilator M* © X*. Let x* © X* be at a distance i from M+. Then B= min fe vmip let = sup (xe) ASOLUTIONTO THESISO F CONTROL PROBLEM 257 where the minimum is achieved for some rg, € M* If the supremum on the right is achieved for some x9 © M, then (x" ~ ri.) =x" ~ 15h ul (he x7 15 is aligned with x9) A special case of the above theorem is the case when M is finite dimen- sional. In this situation, the supremum on the right will always exist, and hence both problems have solutions. As we show next this is precisely the case for problem (831), In order to apply Theorem 8.3 to this problem we need to identify the spaces X and X" and the corresponding subspaces M and M. Clearly in this case, we need X" = ¢! and $= M+. Thus we should select X= cy, the subspace of # formed by sequences converging to zero, since c; = ¢! {187 With these choices, given any r= f(D} EME CE and x= [} EM fo, We have that (nx) = Verda) (8.32) The only remaining task is to characterize M. To this effect consider the ‘onminimum phase zeros of 73(z), 2, and define the following sequences: \ (833) In terms of these sequences M can be characterized as follows Lemma 85 Let M = Span {zf, 2{} then $= M+. Proof. For any element R € S we have that RQ) = Vor)" =0 (8.34) where r = {r()} denotes the impulse response sequence of R. Thus the fol- Towing holds forall Re [A(z)] = (7.27) =0 (835) Im [Ata] =(.2/) =0 Hence (r,m) = 0 for all m © M, which implies that $C M+, A similar argu- ™ment shows that if some sequence r é" satisfies (r,m) =0 for all m © M. 258 C' CONTROL then its z transform R satisfies R(z,) = 0, i = 1,2,...,. This implies that M® CS, thus establishing the desired result a Combining Theorem 83 with Lemma 8.5 yields the following result, Theorem 84 Let 2(2) haven simple zeros z, ouside the unit disk and mo zeros on the unit circle. Then Ho = min F(z) ~ Fa(z)Ote)In = nap {Soame [Red] [real (836) subj 0 bot [Scene [24] + Soesnlm rf] <2, VeRO... 37) Furthermore, denote by &(2) the resting optimal closed-loop system, by dr € the corresponding impulse response sequence, and by % € €' the optimal solution tothe dual problem. Then & has the following properties (a) (Kk) =0 whenever |z(K)] <1. () eGR) 20. © EleOl=H0 (a) Loe (en), = Ao Proof. Let = {U(i)) denote the impulse response sequence of T). Then (ZR) = ERe [2;'] = Re [F(z] Zh.) Eom = im [F(z] 2 Tom Lemna 85 We have hat any Mean be tena re Dazh+ye 639 Thucorcadh 26M, we bavetat ene Larethie) + Davalmlften) (840) [ASOLUTIONTO THE S150 F! CONTROL PROBLEM 259 Equation (8.36) follows now immediately from Theorem 8:3 and Lemma 85. ‘The constraint (8.37) is simply a restatement of [il] < 1. Next, we establish the properties of the optimal solution. Properties (a), (b), and (c) all follow from the alignment condition in Theorem 83. Here this condition takes the form (6.4 = WPlhilhillse = Ho A = (841) YeOW = lolli Clearly this condition holds if and only if (a), (b), and (¢) hold, Finally, to establish (4) note that © ~ 7; € $. Thus (z,) ~ fy(z,) =0, = 1,2,...4m, a Example 82 Consider the simple SISO stabilization problem shown in Fig- ture 8.4, where the goal is to design a controller K(s) that internally stabilizes the loop for all A € yBLTY, where ¥ is a scaling factor to be maximized. Assume that the nominal plant has the following transfer function: x Bea Gale) a>i The problem can be solved by designing K(a) such that it minimizes | Ty The first step is to invoke the Youla parametrization to recast Tyy, a8 an affine function of Q-€ €!, Since the plant is open-loop unstable, we need 10 use the general form of the parametrization, obiained in Theorem 3.4. Begin ‘by factoring the plant as 1 Go=N@MQ)y N= 2, (8.42) Figure 8.4. S150 robust sabilizction problem, 260! CONTROL can easily be checked that the let coprime factors Vo=1 (843) satisfy the Berout equation NU, + MV, = 1. Hence the set of all controllers stabilizing Gy can be parametrized as K=(U,-MOWV.+NOy", Gee (844) with the corresponding complementary sensitivity function given by 228) oo Ty = Ur — MQ) ~ ¥( Note thatthe € norm is invariant under muliplication by 1/2 (since this corresponds to a unit delay). It follows that minimizing Ton is equivalent to the following £ optimal control problem: Ho = min |T\(2) ~ Te) 0G (846) where (847) Since T; has two nonminimum phase zeros at 2 = have that problem (8.46) is equivalent 10 a, from Theorem 8.4 we max(a; ~ ay)a subject 10 ) |arRe[(ja)-*] + aRe[(—ja)-*] + asim[(ja)*] + aalm|(—ja)*]] <1, Clery is senile to mee subject (8.49) ASOLUIIONTO THESISO F CONTROLPROBLEM 261 where we have defined , = 0, +02 and Bx = as — au. The solution to this problem is 8, =0, By =a. Thus the minimum achievable norm is ho = 0? Next, we construct the optimal solution as follows. Note that the constraints in (8-44) satura only for k = 1. From property (a) in Theorem 8. it follows that 8(K) =0 forall k # 1, Combining this fact with property (a) yields 4) = ma". Hence (2) = -© + Qo =0 (850) Finally, replacing Ong, = 0 in (8:44) yields Kop = UoVi5! = 0/2 8.51 Tray = 2/2? Hence this controller guarantees robust stability against all perturbations hav ing |Aliemee 0 such that $(k) = VEEN, Proof. For simplicity assume that all the nonminimum phase zeros of T(z) are real. For a given m, consider the following matrix: leon Qj on a od (852) seen geen Since Fy is a Vandermonde matrix, it has full column rank for all m > n and therefore its eft inverse F,,! is well defined. Let ‘qn terms of F and a the constraints (8.37) can be rewritten as (\Facetllac <1 (8.53) Next, we show that there exists N such that if the first NV constraints in (8.37) are satisfied, then for all k > N we have that - Fa] «<1 (8.54) 262! conTROL where Fi! denotes the kth row of Fi.. Note that Law] < llth max" 855) ‘A bound on lal} for alla satisfying (8.37) can be found as follows. Assume thatthe fist constraints in (8:37) are satisfied. Then a = Fy 'Fya W < ltl MlllFnetle < Fy (856) a Halls < lla < mle It follows that if N is chosen such that nliFe"Inmaxz; "| <1 (657) then (854) is satisfied for all k > N. The proof ofthe theorem follows now from property (a) in Theorem 8.4. a Corollary 82 Let b = [T)(z1), Ty(22),.-., Ti(2n)]". Then problem (8.36) (837) is equivalent 10 the following fintedimensional linear programming problem: nin, Desde @) eo5td' ooo 3 subject to (858) FY (@A — Oy where oa) oa) (8.59) o"(N) o-(N). Proof. The proof follows from combining Theorems 8.4 and 8.5 with standard linear programming arguments. 0 3.1 Properties of the Solution In the last section we have shown that the optimal" (SISO) control problem can be solved by recasting it into a model matching form that can be solved via finite-dimensional linear programming. In this section we explore some [ASOLUTIONTO THE S150 ¢ CONTROL PROBLEM — 263 of the properties of the solution to this matching problem. We will show that, contrary to the 7, and Mx. cases, this solution is not unique and its order cannot be a priori bounded in terms of the order of the plant. In fact, we will illustrate with a simple example that even in the state-feedback case (where ‘optimal static controllers exist in both the Hp and 7% cases), the solution may be dynamic, with arbitrarily high order, We first address the issue of the nonuniqueness of the solution to the ‘model matching problem (827). Consider the following example [203] Example 8.3 Suppose that T(z) and Tx(2) are given by 492432 +2 NO=Gopae | RO= (860) Applying Theorem 84 we have thatthe dual problem is given by max t+ 2 mat 9 subject 10 (ot) Jaye ee3-A] <1, =01 ‘Note that the lefi-hand side of the constraint corresponding to the value k =2 ‘coincides precisely with the objective function. Thus the maximum value of the objective is 1, and dhe constraint corresponding to k = 2 is active. t can easily be seen that the constraints corresponding to k = and k = 1 are also active, while all the constraints corresponding to k > 3 are inactive. Moreover, the corresponding dual functional is of form r = {—1,1,1,4,+,-.} where the + denotes elements with absolute valueless than 1. From the alignment conditions in Theorem 84 we have that the optimal approximation has the form cree ee where $y <0 and ¢y, 62 > 0. From the interpolation constraints at z = 2 and at z =3 we have that 4Q)= =} ‘ (8.63) d@)= a Finally, the constraint \\6\|, = 1 yields [ls =~ 8 + rtd =1 (8.64) 264! CONTROL The system of equations (8.63) and (864) admits infinite solutions that can be arametrized as (865) where a is a real such that Next, we show that, even in the full state-feedback case, the order of the controller can be arbitrarily high. To this effect we need to introduce the following preliminary result (93) that provides a sufficient condition for the order of the closed-loop system to be bounded by the number of unstable zeros of the plant, Lemma 86 Let n denote the number of unstable zeros of Ts, {zi}. The optimal approximation ® is a FIR of degree n~1 if Dial r, ‘Thus with this choice of f, Ty, has also a finite impulse response (of order +1). It follows that the model matching problem (8.26) is equivalent to min 16 (2)Ih = min||2-" [os ~ &) + eflh_,(2 - 2402) wo =min|iy + where Teal -&) heezM. (2-2) (@70) q=ahy ‘¢) and ¢) are constants, and where we used the fact that the ¢” norm is invariant under multiplication by z~* Lemma 87 If the nonminimum phase zeros of Giz satisfy condition (8.66) then the state-feedback gain f is €* optimal. Proof. From Lemma 86 we have that the optimal ¢ is a FIR of order r — 1. Since 7, and Fp in (8.69) are both FIRs of order r, it follows that (2) must be a constant qy and such that ° pA "(Abi + bod) = GANA + bagedbs en epAjby — cA‘ "bal f ~ go)b = cp *ba(f — qo)bi since cyA¥, = 0. Obviously qo = f satisfies this condition. Thus, the controller k= FU.) =f is optimal. A Example 8.4 [87] Consider the following third-order system, where x > 0 is 4@ parameter: 27 235 46] 11 wy.| 1% oO foo Ge=| 9 + oloo (872) a 266 c!cONTROL Table 8.1. denotes 1st 1301 1500 1499 The zeros of Gra are given by the roots of z2 ~ 2.5z + x. Thus, for x > 35, condition (8.66) holds and the optimal €' state-feedback coniroller is static (f=[-27 235. -46)). On the other hand, for 1.5 < « <35, the optimal sate-feedback controller is dynamic. As « | 1.5, one of the nonminimum phase zeros moves toward the unit circle, while the controller order Ng 1 00 and the optimal cost ||: | 3. Table 8.1 shows the optimal € norm and the order of the corresponding controller for different values of x. If x < 1.5, Giz has only ‘ne nonminimum phase zero. Therefore condition (8:66) is trivially satisfied and the optimal controller becomes static again. It is also worth noting the abrupt change in the value of the optimal F norm as this zero moves across the unit circle 3.2 The MIMO Case ‘As we have shown in the last section, the SISO £' optimal control prob- lem leads to an €! model matching problem that can be solved via finite- dimensional linear programming. A key clement in this derivation is the characterization of the annibilator space M+ obtained in Lemma 85, used to establish that the dual problem is finite-dimensional. While the general idea ‘of recasting the problem into a model matching form and exploiting duality is still valid in the MIMO case, here the problem is considerably more involved. Consider the MIMO model matching problem (826). Proceeding as in Section 83, the first step is to obtain an equivalent to Lemma 84: that is, given RERE,,., Tir € REl,,,, and Tm € RE, find necessary and suffi- cient conditions for the existence of Q € Rél,,, such that R= Tia+Q+ Tay Recall that in the SISO case these conditions had the form R(z,) = 0, where 2 denote the unstable zer0s of Tiz(z) + Tz(2). However, in the MIMO case these interpolation constraints must take into account, in addition to preser vation of the unstable zero structure of Ti2 and Ty, the directional infor ‘mation associated with these zeros. This leads to two sets of constraints: () the zero interpolation conditions (essentially a generalization of the SISO conditions) and (ii) the rank interpolation conditions [87 Consider first the so-called one-block case, where the plant has as many APPROXIMATESOLUTIONS 267 controls as regulated outputs and as many measurements as exogenous distur: bhances, that is, ny =z and ny = nq. These problems are the closest MIMO equivalent to the SISO case, sharing many of its properties, For instance, it can be shown that these problems require considering only the zero inter: polation constraints. This results in a primal problem having a finite number of constraints and in a dual problem having only finitely many variables, Moreover, by exploiting the alignment conditions it can be shown that, as in the SISO case, the optimal £ closed-loop system ¢ has a finite impulse response. It follows that the one-block MIMO case can be reduced to a fnite- dimensional linear programming problem, proceeding as in Theorem 8:5 and Corollary 82 (see [81)) Problems having more disturbances than measurements or more regulated variables than control inputs (nq > my or n > n,) ate known as multiblock (or bad rank) problems. Here both sets of interpolation conditions are re quired, with the rank interpolation constraints leading to infinite-dimensional linear programs in both the primal and dual spaces. One can attempt to ob: tain approximate solutions to these problems either by constraining the closed loop to have (a sufficiently large) finite support (the finitely many variables approximation) or by retaining only a finite set of constraints in the primal problem (the finitely many equations approximation). Alternatively, one can attempt to embed the problem into a one-block problem by augmenting the matrices 72 and T>, with pure delays. The solution to the original problem is then obtained through a compression, by simply discarding the additional inputs. A complete treatment of the multiblock MIMO #! control problem is beyond the scope of this book (interested readers are referred to [87] for details). Rather, in the next section we will introduce a method for obtain- ing controlters tht minimize the peak time-domain value of the euclidean ‘norm, While these controllers are suboptimal for the ¢! problem, they have advantages in terms of both their order (bounded by the order of the plant) and the computational complexity of the synthesis process. APPROXIMATE SOLUTIONS ‘As we have shown in Section 83.1, " optimal control theory may result in large-order controllers, even in the state-feedback case’ This feature limits the applicability of the theory in the case of relatively fast plants having a limited amount of computational power availabe (e.., see Section 11.1). In Principle this problem can be solved by applying some of the model reduction techniques covered in Chapter 9 to the optimal €! controller. However, this ‘model reduction step is far from trivial. Table 82 shows the results obtained ‘when attempting to reduce the order of the controller corresponding to x = STs sndvuntage x shared hy ther widely ane design methods, such as -syess that will stso produce contllers wath no guaranteed complesty hound 268! CONTROL Tob mance of reducedorder controllers, « ~ 1.801 N 6 14 2 a ieh 301 38 an Unstable 1501 in Example 8.4 using balanced truncation, While the controller can easily be reduced to 15 order, attempting to reduce the order below 12 yields controllers that do not stabilize the plant ‘As an alternative to model reduction, low-order suboptimal controllers can be obtained by minimizing an upper bound of the #' norm, obtained from geometrical considerations. As we show in the sequel, computing this ‘upper bound only entails solving a LMI optimization problem, followed by ‘@ one-dimensional quasi-convex optimization, Both problems can be solved efficiently with existing algorithms [3, 59]. Proceeding along these lines, sub- ‘optimal controllers can be synthesized by minimizing this bound, rather than the actual cost. As shown in {3] for the continuous time case and in (64) for the discrete time case, this approach has the advantage that, in the state feedback case, the resulting controller is static, while in the output-feedback cease it has the same order as the plant. 8.4.1 An Upper Bound of the ¢' Norm As a first step in synthesizing these suboptimal controllers, we introduce a geometrical interpretation of the é norm, based on the concept of invariant sels Definition 8.4 Consider the discrete time dynamic system x(k 41) = Ax(k) + Bw(k) (873) where x(k) € R® and w(k) € RY, |w(k) hse < 1A set P is said 10 be positively invariant for this system if for all x € P we have Ax + Bw € P. Definition 85 Consider the system (8.73). Given a sequence = (#0), #(D),--) ard an initial condition xo, denote by $(k, Xo, ()) the corresponding trajec- tory. The origin-reachable set Ra, is defined as Ray © (££ = $(k,0,w)} for some finite time k and some sequence w(k). w(K) <1 It can easily be shown that the set Rx. is the smallest invariant set com taining the origin in its interior. Moreover, consider a stable system having & APPROXIMATESOLUTIONS 269 Figure 8.5. Geomettic interpretation of he ¢! norm and ts upper bound. state-space realization o-[444] oo {€: Cel < wi — 8} (8:75) where i= 1 1s 1]? €R and 6 € RP is the vector whose ith com: fe is given by 5 © [Dyll,. Then |Gljr 0h (8.77) ” 1 G bo ee where Q > 0 satisfies the following linear matrix inequaliy 1 npr [-@oze) 2 ] a) ga" (a-1)0 Then the following properties hold: 1. The ellipsoid {x:37Q"'x <1} is am invariant set for G. 2. Ihe Via)" 3. V(aq) is quasiconver for a © (0, dmax), where amay = pA). In order to prove Lemma 88 we need the following preliminary result (3. Lemma 89 Given P > 0, then rman, ex Del oP 0 CT Sint )9:30 >0 such har] 0 (n-o) D'|>0} (879) cop ot APPROXIMATESOLUTIONS 271 Proof of Lemma 8.8. Part I: Since A is stable, it follows that for every a € (Ong) there exists a symmetsic matrix Q > 0 satisfying 4087-04! nB" <0 0) Define the set Vg © {x :x7Q"'x <1}. Then, from Section 43:2 in (276), i follows that for all w(k), x{K) such that w! (k)w(K) <1 and x1 (A) he(k 1. x(k +1) satisies x(k TT x(k +1) < where = (2g) 40a? « Laat Since (880) implies Q~! 0 c Dee is the solution to the discrete Lyapunov equation epowa” 0.+ ant = as) From (8.77) it follows that V(a) < Uy. If V(a) < Ug, there exists 7 such that V(a) < 1 < Ug. Thus, from (8.77) it follows that there exist Q > 0 and > such that [1/(1 ~ ajAQAT — + (1/a)BB? <0 and oo! 0 cr 0 (yoy DT] >0 (883) cop 4 272 C!contROL is eayy to show that 0 < Q, | 0 >0 (B84) eo c Hence Ux 0 such, that oO ct 0° (=e) DT} >0 (685) cop ot Pre- and postmultiplying by 1 OQ 0 0 “4 var o 0 "0 vat yields the equivalent condition a [@ 9 cr M(ao)2} 0 fy oD" | >0, where f(a) £ a(n ~ 0) CQ, “Dol Given any ay oy € Sy, and A € (0,1) define & 8 Aa + (1 ~ Aap. Suppose that there exist oy and oy such that May, ey) >Oand Mao.) = 0. Then, forany 4c (.1),AM(ano;) + (-A)M(ay on) > or equivalently, @ 0 ocr 8 flor) +0—AVfalos))! EDT) >0 (887) co =D al where Q = AQ(a) + (1 = A)O(ar) and & = Ag + (1 — A)oy, Note that the function f,(o) = (nm ~ 0) is concave in o, that is, fa(0) > Afylor) + (I — AVfalan). Therefore 0 fel DT (888) cg oD al Qo o "| APPROXIMATESOLUTONS 273, Multiplying both sides by oo 10 (8.89) or yields oo ct 0 flat aDT| >0 (8.90) Cc aD at Using the explicit expression for the solution of the Lyapunov equation (8.82) A'BBT AT it can easily be established that Q, is a convex function of «. Thus O(a@)-! > O and | Hence a € 5, which implies that S, is convex on a. Thus V(a) is quasicon vex on the interval (0 dng). o 0 cr @' oO c s(oyt ad") >| 0 flat avr} >u 91) oD at c oD “a This lemma suggests that an upper bound of Gli can be computed by minimizing V(a). Following the approach in {3} we will define this upper bound as the * norm of G, that is, inf V(a)!? (892) It can easily be shown that the ||. satisfies all the properties of @ norm. Moreover, computing its value only entails solving a LMI optimization prob- lem and the scalar minimization of the function V(a) in the interval (0,4). 8.4.3 Full State Feedback In this section we consider the problem of synthesizing full state feedback controllers that minimize the * norm of the closed-loop system. The main result shows that these controllers are static and can be found by combining @ LMI optimization with a scalar optimization in (0,1) 274 £! conROL ‘Theorem 86 Assume shat G has the following state-space realization: A|By Bs GO Da 4893) where the pair (A, By) is stabilizable. Then the following statements are equiv- alent 1, There exists a finite-dimensional, ful state-feedback internally stabilizing LTI controller such that |\Toulle <7. 2. There exists a static contro! law u = Kx such that |Tewle < ¥. 3. There exists a scalar a, 0 3) Suppose that there exists a finite-dimensional LTI controller K that internally stabilizes the system and renders |'Tzn!|. < 7. Assume @ realization of K is given by ‘The corresponding realization of Tey is given by ABD, BC.|B1) [ale Tw=| Be Ac [ol = [a4] (895) C+ D2D_e Di2G 0 clo From Lemma 88, there exist a scalar « and a symmetric matrix Q > 0 such that AQAT ~ 0+ 188" <0 and Cole vt APPROXIMATESOLUTIONS 275, By Schur complements, this is equivalent to [ere 4g |: o co) ear (a1) ” oe ° 897 i ol > 97) Partition Q as (8 ee Oh On Since Q > 0, it follows that Q,, > 0, Multiplying on the left-hand side of (8.96) by [le o} 0 | 0 Fo and on the right by its transpose yields <0 (898) On BBE AQn +BY OuaT eVTBE (a~1)0n where V = DyQy; + C.Of, Pre- and postmultiplying (897) by (4) one obtains : [4 %]>° 99) Finally, multiplying on the left by [!" 6] ‘| oor and on the right by its transpose yields (eocte aa mame les (100) (3+ 2 and 3 1) Suppose there exist Q > 0 and V that satisfy the LMI (894), With the controller given by u = VQ-'x, the closed-loop system be- comes Ay} a (8.101) cto _ [As Bvo" [Bi] _ T= Les paveto | 276 C' conTRot Table 8.3, Optimal versus staticeedback “” norm for Example 8.4 « Na 1eh, Taw Gap y z 2 421 465 10% 52 ist 1 304 348, 14% 391 1301 16 301 346 13% 390 1500 300 346 15%, 390 Since Q and V satisy the LMI (8.4), it can easily be verified that 4 Ajat -0+1B,BF <0 and COC} -¥71<0 ‘The proof follows now from Lemma 88 o Nest, we illustrate these results by synthesizing «norm controllers for the system used in Example 84 and comparing their performance against that of the optimal é controller. Table 83 shows a comparison of the optimal # norm corresponding to different values of x versus the ¢” norm achieved by the static optimal s-norm controtier. Its worth noting that the gap 1e- ‘mains constant at about 15%, even when the order ofthe optimal controller approaches co. 8.4.4 All Output Feedback Controllers for ‘Optimal »-Norm Problems In this section we establish necessary and sufficient conditions for the exis- tence of y-suboptimal »-norm output-feedback controllers. ‘Theorem 8.7 Consider a discrete time FDLTI plant G of McMillan degree with a minimal realization: A| BB G=|G)0 Dr (8.102) C:| Dy Da Assume that the pairs (A, Bz) and (A,C3) are stabilizable and detectable, re- spectively. The suboptimal s-norm control problem with parameter y, that i { Toulle < i8 solvable if and only if there exist pairs of symmetric matrices (R,S) in °°" and a real number > 0 such that the following inequalities APPROXIMATESOLUTIONS 277 are feasible: wi (5 1 spat = Re LinBT) wi <0 yfATSA~Q-a)S ATSB, . nl BISA wel s6, [No o RI), [F s}2° ion WY (ol ~ GRCT)W, >0 Sct le, Si]>o where Wi, No, and Ws are any matrices whose columns span the null spaces of BE, (Cz Dz), and Df, respectively. Moreover, the set of y-suboptimal controllers of order k is nonempty if and only if the LMIs (8.103) hold for some R, S, which further satisfy he rank constraint rank (I ~ RS) < k Proof. Assume, without toss of generality, that D2: = 0.7 Proceeding. as in Section 64, given a controller with state-space realization: ul x= [Ale] (e109 combine its parameters into the single variable o-[% 3] As before, in the sequel we use the following shorthand notation: Ay ¢ a): a= [%]: CHIC, Ol: oo ° ok (6 4] e-(2 4] [] ho " Recal hat this assumption can always be removed thoagh a loop shiing transformation Dis= [0 Dys|: 278 Cc‘ CONTROL In terms of these variables the closed-loop matrices A,j, By. Ca and Dy corresponding to the controller K can be written as Ay = Au + BOC, By =B, + BOD, (8.105) Cut D2@C, Dy = Dx®Px From Lemma 838 we have that the + norm of the closed-loop transfer function is less than y if there exists X, > 0 in RU" such that exer Och [‘ vrcor Bi) >0 (106 er XA By ai (@-)x, 0 | co (107) Bho al Rewriting these inequalities in terms of the matrices (8.105) yields + PTOQ+OT@TP <0 (8.108) ¥+PTOQ+ OTOP 0 where a Ag By oF Al (a -1)X 0 BE Oe ai. pal Me Srximty Oram “lero 0 0 hoo 0 Dn (8209) os (7 -o)lm, 0 In nm 0 Dh 0 Oem, 0 Dy 0 APPROXIMATESOLUTIONS 279 From Lemma 6.9 it follows that (8.108) is feasible in @ if and only if WEOWe <0: WEbWy <0 Cees (8.110) WIVWe > 0: WEWWo>0 where Wp, Wo, Wp, and Wo denote bases of Ker P, Ker Q, Ker P, and Ker , respectively I can easily be seen that Wp has the form wo 0 0 0 Ona) 0 7a a (an) oe where Ws denotes any basis ofthe null space of BY. Proceeding as in Theo: rem 6.3, partition X, and X>! as y{| AT 0 Be oo alm, Using Schur's complement formula we have that this is equivalent to aoe our (Lara -eeLmay)m, 6m By ct in sia sion, can be eae he neil Wao cObs cout Nb [* SA ~(1-a)S ATSB, BISA atvatse,] lo<0 @B.116) 280! contRot where Ny denotes any basis of the null space of [Cy Dy). The condition X, © ROK 5 (is equivalent to RS satisfying Ba Jo and santas) 0 reduces i eo we) Gren, Eo! |meze eum Ic, 0} 0 | where af tu 0 AO as) wm 3 8 ‘After rearranging rows and columns this can be rewritten as, a 0. io me & Y we 0 | igo (8.119) a where : _ [Ww 0 0 M210 Im 0 (120) 00 Ing Using Schur complements once more, this is equivalent to 1 wr oo crcl wo] Hl wore J>o 2) PE adPsee oy ][F Adee Carrying out the block multiplication explicitly yields WE (of -GRCL)Ws> 0 Fo (1m) APPROXIMATESOLUIONS 281 Similarly, from W5¥We > 0, one can obtain that [é S]>o To recap, if X, > 0 of dimension m+ k solves (8.106) and (8.107), then the inequalities (8.103) hold for some R, S, and 0. Conversely, ifthe set (8.103) admits a solution (R, S, c), then a matrix X, >0 satisfying the inequalities (8.106) and (8.107) can be feconstructed from R, S. o Remark Theorem 8.7 implies that if ||7;.||. < y is achieved by some con- troller of order k > n, there exists a controller of order m also rendering Tew. < ¥. I follows that in the output-feedback case the optimal + norm can always be achieved with controllers having the same order as the gener- alized plant The LMl-based approach introduced in Theorem 8.7 is also useful for synthesizing reduced-order controllers. These y-suboptimal controllers of order k 0 (ie. A= AM x £4(R,)), equip ped with the norm l[élia = tll, + lll. A denotes the space of Laplace transforms of elements in A with the norm defined as Ala = Wella where His) is the Laplace transform of A). B(R.) denotes the space of bounded functions on Ry equipped withthe nor fia = sucn COL Tecan be shown (82) that A = BUR.) x C=(R,). Each element (fs) € A" defines a bounded linea functional on A, with its vluc given by (Chan Rio) = Soha) + [COC a Proceeding as in Section 83, the first step toward applying duality isto iden- tify the sets M and M+. To that effect, we need the following continuous time equivalent of Lemma 84. TE CONTINUOUSTIME CASE 283 Lemma810 Assume tht T3(s) © A havm single cers 2i,20:+-y2m Ree) > 0, and no zeros om the jo axis Then given RCA, there exists Oe A such that R=; Q ifand only if R@)=O, 1212.40 (8.124) Let $= {REA:RE)=0. i= 1.n} en with this definition the continuous time C' problem can be reduced (via the Youla parametrization) to Ho = ink IT + Ria (6.126) ‘We will solve this model matching problem by exploiting the dual of Theorem 83. In this case it is lear that X = A and S = M. Moreover, lel File) =Re {e™"} (= Im fe} ¢ sequences Mis given by M! Span {fy fs(68))} Combining these results with the dual of Theorem 83 yields [82] the fol- lowing. (8.127) It can easily be shown that in terms of th ‘Theorem 88 Let T3(s) have m zeros z; in the open right-half plane and no zeros on the jw axis. Then uo = max Seameiriso) + So matmtrit)] (8.128) subject 10 been Lareley+Paadmley] ct veer. — (6129) Furthermore, the following facts hold: (i) the oprimal extremal functional r°(0) equals at only finite points: t,....tm; (i) an optimal solution (s) = Ty(s) — T,(3)Q(s) (0 problem (8.126) always exists; and (iii) the optimal @ has the following form: = S450), HER. finite (8130) and satisfies the following conditions: 284 C' CONTROL @ dr) 20. 0) El = Ho (©) Vide = Ta), k=l. un. Remark It can be shown [82] that the constraints (8.129) need to be satisfied only for all < tan, Where fmas is finite and can be determined a priori. Even 50, there are still infinite constraints, and therefore the dual problem is a semi-infinite linear programming problem, Example 8.6 Consider the following unstable, nonminimum phase plant The control objective is to minimize {bly = ||PC(L + PC)". By using the Youla parametrization, this problem can be recast into the form (8.27) with sth (8131) The dual problem ts guven by Ho = Maxey (6.132) subject IN{t,a)| = |e + ape") <1 for allt e Ry (8.133) By differentiating (8.133) it can easily be shown that r(t,<2) achieves local extrema at the points t = 7, where Te me eO (8.134) with the corresponding value r(r,a) = ~a/4ap. Thus problem (81132) is equivalent to do = maxes (6.135) subject 10 lay +091 <1 fax] <* (8.136) a) <0 THE CONTINUOUS TIME CASE 285, The solution to this last problem is given by era -2+2V%), ma 3e2V (137 The constraint (8133) saturates only at two points, 1=0 and t= 7 = ~In(t + V2)/t3-+2v2). Thus the optimal closed-loop system has the form b= bo S(t) + H(t 7) Finally, the values of ¢, and ¢ can be obtained from the interpolation con- straints at § = Land § ~2 yielding i+ v2 342V2 b, a vi sith bull, = a2 = 5.8284. The corresponding controller is given by (5 = 290.7071 — 4.1213 (e=D(-O70T1 + 4.1213 HY 55) Ke 8.5.2. Rational Approximations to the Optimal £, Controller From equation (8.130) It follows thar, unlike in the discrete-time case, the y-optimal controller is irrational even if the plant is rational. Prompted by the difficulty in physically implementing a controller with a nonrational transfer function, a rational approximation method was developed in [41]. In the sequel we will briefly review this result, based on the observation that the C; norm of a continuous time system is (uniformly) upper bounded by the £! norm of its discretization using the forward Euler approximation, Definition 86 Consider the continuous time system G(s). Its Euler opproxi- ‘mating system (EAS) is defined as the following discrete time system: otean= [A I From this definition it is easily seen that we can obtain the EAS of G(s} by the simple variable transformation s = (z ~ 1}/7. that is, ofan =6(5*) ‘On the other hand, for any given 7 we can relate a discrete time system to a continuous system by the inverse transformation z = 1 +75. Itis obvious that the discrete time system is, in fact, the EAS of the continuous time system ‘obtained in this form. (8.138) 286 C' contRoL Next, we recall the main result of [41] showing that the £! norm of a stable transfer function is bounded above by the £' norm of its Euler approximating system (EAS). Moreover, this bound can be made arbitrarily tight by taking the parameter + in (8.138) small enough. This result is the basis for the approximation procedure proposed in (41) Theorem 8.9 Consider a continuous time system with rational Laplace trans: form ®(s) and its EAS, (2,1). If © (2,7) is asympuotically stable, then ®(s) is also asymptotically stable and such that HOG) SHOE Cz, 7h Conversely if (s) is asymptotically stable and such that |§P(s)I]v © po, then for all u > we there exists 1° >0 such that for all 0< 7 <7, OF(2,7) is asymptotically stable and such that \}"(z, rl < w- ‘Theorem 8.10 [41] Consider a strictly decreasing sequence 7, +0, and define HS iat, HOGG 7h where (2,7) denotes the closed-loop transfer function. Then the sequence Jus 18 noninereasing and such that a, the optima Cy cost Corollary 83 A suboptimal rational solution to the £; optimal control prob= lem for continuous time systems, with cost arbitrarily close to the optimal cost, can be obtained by solving a discrete time £' optimal control problem for the corresponding EAS. Moreover, if K(z) denotes the optimal &, compensator for the EAS, the suboptimal £, compensator is given by K(rs +1). 1 Example 8.7 °Consider again Example 8.6. For 011 the EAS is given by 120 01 T (8.139) 1A 2.9091 and L = 0.3667 yields GF (2,01) [ ‘and the Youla parametrization with F = 0.9091 0.2909 | 0 0.0909 0 0.8333 |0.3667 0 TOOT 2509170 TT 0 -0.8333/0.8333 0 (8.140) z 00900 | 1.6091 | ~0.3667_~0,2824 yale 1.5909] 0.8333 0.7576 THE CONTINUOUS TIMECASE 287 Hence we have that 16(2 — 1.1) ‘s(t - N2e—) (cn Liye = 12) = Le "= ig ze) Solving for the optimal € compensator yields optimal cost jy = 6.184 and ‘optimal error: o(e) = 1.8414 — 4.342327 ‘The corresponding optimal Q and compensator Keas are given by Q(z) = 2.4309~ 0.05252"! + 0.06072? +0.20892-* + 0.40042-* + 0.654225 + 0.95542°%4+ 1.34582"? + 1.83432-* — 3.28052? (3142) Keas = FiO) Finally, the transformation z = 7s +1 yields the corresponding compensator for the continuous time system. Figure 86 shows |\6\|4 versus 7. Note that as + ~» 0, \blla —* Hor Moreover, the error —+0-as (7) [316]. ‘Acnorm 5 ° 107 107 10) Figure 8.6. Optimcl cost ebtoined using he EAS method versus > 288 C' conTROL 8.6 PROBLEMS 1. Consider the uncertain linear system 2(k +1) = [Ay + w(K)DE) x(k) A,=05+[ 1 HE o- [i] E={1 oO) G43) Jw(k)| < Wem where w(K) represents memoryless (arbitrarily fast) time varying uncer tainty (a) Use Theorem 8.2 to show that if wma. = 0.6 then the system (8.143) is BIBO stable (b) Define G(z) = E(zl ~ A,)-"D and recall from Problem 6.10 that if Yan S (Glen)! then there exist P > 0 such that V(x) = x" Px is a Lyapunov function for the system (8.143), Use this result to reduce the conservatism in the estimate of the maximum value of Way guar anteeing stability (© Consider now the following polyhedral Lyapunov function: Vp(xe) = max | P82 (8.144) me: where oo4 1 1 0 1 ee ies (8.145) 1 0.3390. oon. ‘ Show that if wus < 0.975 then V,(x) is a Lyapunov function for the system (8.143). In fact, using similar techniques it can be shown that (8.143) is stable if and only if wax < 1. (See [42] for a thorough treat- ment of polyhedral Lyapunov functions). 2 The purpose of this problem is to investigate ¢" controt problems with zeros on the unit crcl (317). Consider the SISO é" model-matching prob- lem: 0 = inf, T(z) ~ Ta(2)QIh Ho = jaf ITi(2) ~ Ta(2)h ‘Assume that T,(2) has two complex conjugate zeros z; = e!® and 22 = e-!# 6 the unit circle and no other unstable zeros. (2) Show that bo = inf Ti — Tr lh on = max [arRe{ Fi(21)} + atm{y(z2)}] mons 289 subject |r = fv onthe) ~ sink] <1 £=0.1,2 (8.147) (by Show that for the case @ = yx where y is a rational number the infimum in the primal problem is always achieved and the optimization problem can be solved via finite-dimensional linear programming. (©) Show that if @ = yx where 7 isan irrational number, there exist values ‘of 7; such thatthe infimum in the primal problem can not be achieved (a) Show that when 7 is irrational, the set of «satisfying the constraints (6.147) 5, does not depend on the exact value of y (i. Sy is the same for all irrational y). Denote this as 8}. Similarly, let Si (y) denote the constrained area of the dual problem for the case where y = m/n Show that $#(y) does not depend on m and that S89) 28h, for any of rational y (6.148) (e) Use these results to conclude that the optimal approximation error j., does not change continuously when the zeros of T(z) move continu- ously on the unit circle. 3. Consider the pitch axis control of a forward-swept wing X29 aircraft [95]. ‘A simplified model of the plant is given by: (s+3)__ 20 (5-26) PO) = By 106 =6) G+ 20) (5 +26) (8.149) We are interested in synthesizing a controller to minimize (|W(s)S(s)|) where 5 denotes the closed-loop sensitivity function and where (+1) WO) = Bor (8.150) ‘Assume that the plant is discretized via a zero-order hold at the inputs ‘and sampling at T, = 4 (a) Design a controller to minimize ||W,(z)S(z)Jh where the weighting function W,(z) is obtained from W/(s) using a bilinear transformation with 7, = 30-1 (b) Use a CAD package to set up a simulation diagram where the con- tinuous time plant (8.149) is connected via sample- and hold-devices to the discrete-time controiler found in part (b). Find the impulse re- sponse of the resulting sampled-data system and compare it against the discrete impulse response of the system obtained by connecting the discretized version of (8.149) and the controller found in (b), 290! contRoL 4. Consider the following discrete time plant P: A| BB P=|G]Dn Dn | (151) | Du 0 where (A, B:) and (A2, C2) are stabilizable and detectable. (a) Proceeding as in Section 8.4.4 show that the suboptimal «norm control problem with parameter y is solvable, if and only if, there exist pairs of symmetric matrices (R, S) in R"*" and a real number o > Oso that the following inequalities are feasible (66} RAT — Re TBBT) W) <0 7 [ATSA —(1-a@)S ATSB, Ng BISA at +B7sB,|No< RI > [i see (9152) WYO) [1-ZGRCT Du | |W, Ol Lg or Di, (-ollo 7 7 [os ce, Cyn MDE oa acorn) N02 where W, No, Wo ate any matrices whose columns form bases of the null spaces of BY, (Cz D2], and DF, respectively (©) Show that if Dy, =0 then the inequalities (8.152) reduce to the LMIs (8.103) ‘5. Consider the optimal closed-loop system found in Problem 3. (2) Find the worst ease peak of the control action when the input to the system isa persistent signal wth magnitude bounded by one. What is the worstease input signal? (©) The peak control action can be reduced by synthesizing a controller K(z) that minimizes WaG)Ste) | 01K G)St2) This leads to a two-block MIMO ¢' problem. Find a suboptimal solution to this problem using the inequalities (8.152) and compare its performance against that of the controller found in Problem 3 PROBLEMS 291 6. In order to obtain a rational controller, the X29 plant in Problem 3 was discretized using sample-and-hold elements, and the resulting discrete- time problem solved using €! theory. However, this approach could result in significant intersampling ripple. To avoid this, use the EAS technique to synthesize a suboptimal, continuous-time rational £' controller for the plant (8.149) and the weight (8.150), MODEL ORDER REDUCTION 9.1 INTRODUCTION ‘Computers play an important role in the modeling, analysis, and design of ‘modern control systems. Thus, it is necessary to understand not only theit advantages but also the new problems that appear with their use. These problems will be considered in this chapter. In particular we will concentrate ‘on problems that concern state-space realizations of FDLTI systems, since this is how they are represented in a computer. I is easy to understand in other fields the problems to which we are referring. Take, for example, matrix inversion. A nonsingular matrix can be theoretically well defined by means of its detéemmimant, However, front a numerical point of view the situation is not as clear, That is, a badly condi tioned matrix can be cheoretically nonsingular but numerically singular as will be seen in the next example. This is a simple consequence of using truncated rational numbers in a computer instead of real numbers, as in the supporting theory, Example 9.1 Consider the following matrix: Mt 1 A= eR” (1) 1 00.05 Its eigenvalues, hy =---= Ay =—1, and its determinant, det(A) = (-1)", indicate that in theory its a nonsingular matrix. Nevertheless, if we multiply 293 294 MODEL ORDER REDUCTION it by vector u = [11/2 1/2")? we obtain Au = —2'-"|L yy Ay the size of A increases, ||Aullz tends 10 zero. This implies thar oumeri cally this is a “near singular” mairix. The latter is confirmed by the fact that aA) x2" Therefore a theorctca zero may be represented computationally by avery small (nonzero) number that depends onthe existing computing technology ‘The problem canbe solved by quantifying the meaning of “small” and “large” in a computational numerical sense. In this example, its well Known that this can be accomplished using the singular valves ofthe matrix. Te smallest singular from singularity depending on this value, a quantitative property. Instead, theoretically, it can be either singular or nonsingular, a qualitative distinction. The same problem appears when we work with FDLTI dynamical sys tems, represented in a computer by a state-space realization. The basic the- ‘ory of minimal realizations (165, 166)) suffers from the problem of structural instability. This means that “small” numerical perturbations can change the properties of a system. The properties we are interested in are controllabil- ity, observabilty, stabilizabilty, and detectability. As in the case of matrix singularity, measures of each of these properties can be computed, using the singular values of specific matrices (see (67, 108, 171, 202, 223, 269, 311) {As in the matrix inversion case, each of these values measures the minimal numerical perturbation that cancels a particular propery. For example, py provides the smallest complex perturbation (8, 8B) on the stabilizable sy- tem (A, B) such that it sno longer stabilizable. It ean be computed as follows (269) os inf @{[(A~s1) B)} (92) ‘This and other similar measures quantify a property of a system but do not distinguish separately each of its states. In this chapter, we present a way to distinguish the states of a given system according to a property of interest. For example, we might want to order the states according to their controllability, with the first state being the “most” controllable and the last one the “least” controllable. This provides a real- ization useful when the less controllable states need to be eliminated, From all possible realizations of a certain system whose order is fixed, itis often convenient to select the one having its states ordered by their relative com trollability and observability. This is the case when we need to verify how “minimal” a realization is from the numerical point of view, or when pet forming model reduction, The latter problem is of fundamental practical importance. The following is a list, by no means extensive, of the need for model order reduction. GEOMETRY OF STATESPACE REALIZATIONS 295 ‘# The series. parallel, or feedback interconnection of systems (see Appendix B). even when these are represented by state-space minimal realizations, may produce a nonminimal one, ‘© Many design methodologies (e.g. u-synthesis, € optimal control), result, in high-order controllers, + The state-space realizations of time-delayed systems or systems described by partial differential equations are infinite-dimensional, There ‘are many methods that generate (high-order) finite-dimensional approximations. Thus sometimes a further model reduction step needs to be performed, ‘¢ In many cases. when working with a high-dimensional model, the numerical errors may distort the analysis or simulation results ‘* Because of computation time, limited memory, or simplification of the controller hardware, itis important to have a small number of states in its realization, In all the above cases, itis also important to have a measure of the error between the nominal and reduced-order model, which ean be considered as ‘model uncertainty in a robust control context 9.2 GEOMETRY OF STATE-SPACE REALIZATIONS 9.2.1. Controllable/Unobservable Spaces Consider the following FDLTI realization: 8) = Ax() 4 Bul), x) Py 3) ye) = Cx (04) Define in R" the controllable and unobservable subspaces as follows ([61). Definition 91 The controllable subspace Xe contains all final states corre sponding to all piecewise continuous inputs w €U and zero taital condiiions, Definition 92 The unobservable subspace Xz contains all initial condivions 4 € RY, which produce null ouiputs, y@) 7 Itcan be proved that Xc = R(e*’B) = R(W,)-and Xz = N(Ce**) = N(W,) (see [61]. Lemma 1, page 75). Here W, and W, ate the controllability and observability Gramians, respectively: This has been fepresenied graphically in Figure9.1, 296 MODEL ORDER REDUCTION Re Xo = Re) “BE u ett {0} Xy=N(CeM) Figure 9.1. Contiollable and unobservable spaces ‘The state space R" can be divided as follows: R" = Xco@Xéo 05) Xco = Xen (Xe Xp) 06) ‘with Xco the minimalsubspace needed to describe the input-output behavior of the system. Therefore each state can be represented as the sum of two orthogonal vectors—one in Xco, the other in Xzo, Note that we have not define (Xe NXp)!. This is due 10 the fat pservable sats is no a vector space. Take, for example, two observable initial states x,(0) and ‘xp(0) with Zero input (u(r) = 0) and outputs y(r) and —y(1), respectively. The sum of both initial conditions xy = x4(0) + x9(0) gives a null output and, as @ consequence, is not observable. In general, the sum of observable states of the form x4(0) = x,+m and xp(0) = ~x, + m, where y(t) = Cex, 40 and mim € Xqu may not be observable. ‘The subspace X 4 (XcMXq)* contains the zero element as well as observable states_Italso contains unobservable and wwe prove next. Divide the space R" in X and X' and consider nonzero states %y EX* and xp CX. For any state x — xq +xp, we have Cet!x = CeM'xp. If Ce*"xy 40, then it is observable. On the other hand, if Ce*'xp = 0, then xp € Xz and xp ¢ Xc because it cannot be in X and X simultaneously un- less xp ~ 0. Therefore all unobservable states in X are also uncontrollable Asa consequence Xco ineludes only controllable and observable sates ‘The subspace Xco contains all states of a minimal realization of the system and therefore completely characterizes the input-output behavior from the state-space standpoint. Related with this input-output behavior, we can also describe both subspaces X'c and Xin terms of input signals, as follows ontralable sates 25” [GEOMETRY OF STATESPACE REALIZATIONS 297 # Define 21(1), # = 1,...,p, as the state response to input vectors u(t) = e7B(0), = 1,...,p, where ey is the Fth column of the identity matrix in RP? and 5(:) is Dirac’s delta distribution, Then Xe can be defined as the smallest dimension subspace that contains the range of all outputs, that is, RUX(®); > 0), with xo= [kOe] a) Define»), €= I... as the output response to iil condition vectors x(t) = er, € = 1,...,m, with ey the fth column of the identity matrix in R"*". Then X,; can be defined as the largest dimension sub- space contained in the null space of all outputs, that is, N[Y (0); « > O}, with yo = po yo 8) 9.2.2 Principal Components In this subsection we describe the geometrical characteristics of the state space in terms of the input-to-state transformation and state-to-output trans- formation. For simplicity we define an operator and its corresponding, Gramian, Which can be interpreted as any of the latter transformations. In particular, the input-to-state transformation is related to the controllability Gramian and_ uutput transformation to the observability Gramian, ‘continuous Tunction F : RR" and its associated we [Faron 9) ‘The above matrix is symmetric and positive semidefinite (prove); therefore it has nonnegative real eigenvalues. As a consequence we can select the following set of orthonormal eigenvectors (Problem 1) W = v3tvT (9.10) oe Ba (uy 0 where the eigenvalues have been ordered for convenience (o > @ > += > o) and V is a unitary matrix whose columns are the corresponding eigen: vectors. To clarify, we present next specific examples of the above and the State-space interpretation of the corresponding 298 MODEL ORDER REDUCTION Example 92 Consider the problem of reaching a final controllable) state 2) 20 al Lime 1, =0, starting from zero initial conditions at ty» —>e with a ‘minimum energy input signal u © £x(-~D0,0) iin, Lllssth that (0) ~ ay 912) This problem is similar to the following leas squares one ina finite dimensional space: min ula subject to 2 = Mue BY (ony sin ll sub ) with M/€ R™" and m > n. Denote by M* the adjoint operator of M (in this case, the transpose). Using the geometrical fact that R can be separated into the direct sum of R(M") and N(M), it is clear that the minimum norm typ, should be in the range of M". The fact that B" can he separated into the direct sum of R(M) and N(M*) implies tap = MC, 89) = M27, x, CRM), sy CNM) (9.18) y= Muy = MM, (915) Its easy to prove that R(M) = R(MM") as follows If y C RIMM") + yc R(M) Because y = MM "xy = Maz for x2 = Mn. On the other hand, ify R(MM*) + y © N(MM") therefore MM"y =0. For 2 = M"y we obtain 2 € N(M) and z ¢ ROM"), thavis, 2 =O. This in urn implies that y belongs tothe Kernel of M* and as a consequence is notin the range of M. When M has full rank, this fact combined with equation (9.15) yields: Hayy = MMM bay (9.16) leo = x8 (MM xa (9.17) where M>(MM")"" is known asthe pseudoinverse 0 “taney Sean eat be oben oho ihe input space is L(-,0) (infinite-dimensional, although the operator is different. In this case we define the infinite-dimensional operator M as soa) = MW © fe Buln)de (9.18) = [er ou-nar (919) Therefore in (9.9) F(t) = eB and all eigenvalues of A should be in the open lef-half plane for W to exist. In this case W is exactly the controllability GEOMETRY OF STATESPACE REALIZATIONS 299 Gramian W.. The minimum energy input is Hop) = BIE WEN ay (920) ney Ieopil = xg We xo (921) In this case BT e""W-" is the pseudoinverse of the operator M. It is clear from the above equation that if W-" is “large” (a [W.| “small”), certain states will be reached only by high-energy inputs, being usp, the optimal one. This gives a quantitative measure of controllability, since the most controllable states ‘an be reached with the least input energy, and vice versa. In the limit, as W, tends to singularity, [Wd ~ 0 (922) = [luapll> + 00 (923) = 49 > noncontrollable (924) Example 9.3 The dual problem is as follows. Using the output equation (9.4) with u(t) = 0, € > 0, we want to find the inital condition xo that mavimizes the energy of the output: xe vo Ax(0), x(0) = x0 apy be a ar [20020 (925) ‘In this case the transformation can be defined as y(t) © M(x) = CeA'x. Using duality and the solution of the last example, we obtain (026) As before, a quantitative measure of observability can be obtained from the size” of the Gramian W,, If itis near singularity (|We] 0), rio matter how large certain nil states may Be, He oul energy wil edt zero, that fs, this wal é aaa a Next we represent the function F(t) in terms of the orthonormal basis formed by the eigenvectors of the Gramian W. Feo = Soro 20 fy © oT FO (9.28) 300 sooeL onoER DUCTION Vector functions v;f! (0) are called the principal components and fF(t) are defined Ue urcional somone 909) The Totowing property's Uefa aS an interpretation of the optimal projection and “energy” distribution of this operator (for simplicity we use fp = 0 from now on). Lemma 9.1 The set {f\(0),---»falt)} 18 an orthogonal basis of functions. The norm of each element of the set is elated tothe singular values of the Gramian W, as follows a hh (9.29) If + 00 and the following integral converges, then it coincides with the norm PIE: - trace [FTO] de = (930) For each 1, the argument of the integral is defined as the Frobenius matrix norm |IF (OI Proof. ronoa otf ror wade of VE2VT wy, of of, oe {a wie) oan which proves (9.29). Since the trace of a matrix is the sum of its eigenvalues, we obtain trace(W) = FIR (932) [fs [Porro] ar ‘The above result can be interpreted a follows. The total the operator Fl), that is, [Fis distributed spacially accord éalions given by THE constant 3 uj, The relative import ‘each direction is quantified by the eigenvalues a. Therefore the Gramian W. (GEOMETRY OF STATESPACE REALIZATIONS 301 ‘through its eigenvectors and values plays an important role in the ‘geometrical interpretation of F(¢): in particular, as will be seen next, in the description of the controllable and unobservable subspaces and also in the input-output priority of each state (see Section 9.4) ‘To this end define the following subspace of R Se = ROW) (9.33) which can be spanned by vectors v; corresponding to all g; > 0. Assume Without loss of generality, that the whole space can be spanned, that is, S- = R*. Then the optimal projection satisfies the following property. Lemma 92 Take 1 <1 0) > --- > oy > 0 in (9.11). Then there exists q = VE~!p such that 2 = Wa = V3p € §, with [pl = 1. Therefore the input signal a) = FT ng a vey (a1) =a [fl Ferner od =4"Wa [pl = 1 (9.42) takes the state 2(0) = 0 to 2(s) = Sos. In Example 92 we have proved that the above input u(t) is the minimum ‘energy one! satisfying 2 = fy F( ~ u(t) dr. Due to linearity and jus = 1. this means that there is no juz <1 that generates a final state outside the ‘boundary of the hyperellipsoid S: therefore this boundary is S. 2 € S (Problem 2). This proves that From the above results we can note the following: Notes ri.then q = W-!2 = 9;"'v, which in turn implies Tr=du _ filr-1) un =F (043) ‘Therefore the set {a)v;} generates the reachable states in R" and the functions /,} generate the orthonormal set of corresponding input signals. 2. When F(1) = eB. the boundary represents the reachable'states start ing from x(0) = Owith normalized inp li» ~ 1. The boundary of the hyperellipsoid in the dual case—F(t) = e*"'CT—represents the initial "Take the final state #45 4p and W = W. Then Ill? = 4 Wig = af Wy tay = TWH (GEOMETRY OF STATESPACEREAUZATIONS 303 conditions from which normalized outputs [yj =1 canbe obtained, with zero input. In each of these cases, the values (04 /0,,) reflect the condition numb of control and observation, respectively. By this we mean the relati ontrollability and’ observability with Tespect to all directions 3. The specific coordinates and magnitudes for controllability and obsery- ability depend on the Gramians. which in turn depend on the specific set of selected states; that is they are not invariant under state trans- formations. Therefore it is important to select a particular state trans. formation that can be useful to reflect the relative controllability and observability “measures” simultaneously. To this end, note the following fact. A state transformation changes the system matrices as follows: A|B Tr TAT"! | TB LEE] ow ‘The Gramians Mo) ate changed to (TW, TT, T-TW,T-), respec- tively: Except for special cases (unitary T), the Gramian eigenvalues will ange_under sate tansfomatons, Neverteles, ie /, femain unchanged, since the transformed product is TW.WaT~! Example94 Consider the following realization of the system 1/(s + 2)(s + 1): = [? S]ao [Se] (045) YQ) = [10% 10-4] 20) (046) For this realization, the Gramians are [Osx g mem OO paste] om 025x108 = [PAT os do] os with similar condition numbers in the order of fe = ky ~ 10". It is clear that for this realization x is “almost” uncontrollable and x “almost” unobservable. With the following state transformation, ‘ wo 6.49 10 =['% Jao 304 MODEL ORDER REDUCTION ve obtain the following realization iW = [e ‘Jim [ 4] (950) yo = (1 10) (ast) with a better “balance” between both condition numbers, which have decreased (0 R. = Ry = 38. Clearly, this new realization has much better numerical prop: erties. From the geometrical point of view, the highly eccentric ellipsoids of controllability and observabilty in R? have now a more “spherical” descrip: tion 9.3 HANKEL SINGULAR VALUES 9.3.1 Continuous Systems As indicated in the last section, itis useful to represent the dynamic system in terms of a particular set of states, which “balances” the quantitative measures of controllability and observability. In this way it is easy to order the states from the most controllable observable to the least one. This is useful from the point of view of model reduction because this order provides the relative priority of each state in terms of its input-output effect ‘The quantitative measure of relative controllability observability should depend now on both Gramians W, and W,. In particular, these measures should be a property of the system, that is, they should be independent of the specific realization selected, According to the notes mentioned in the last section, the eigenvalues of the product of Gramians are invariant under state transformations. This is coherent with the definition of the Hankel singular values of Chapter 6 (Definition 6.2), as the square roots of the eigenvalues of the product WW, (or W,W.), indicated as (wWees = (a) en =the (932) For convenience, they are usually ordered as follows of! > off > ---> of! In (209) these values are called second-order modes. ‘They Fepresent the fundamental measures of gain and complexity of a LTI system ((132)). The preferred state realization, which we seek for, should make both Gramians ‘equal and diagonal (see Section 9.4) such that of 0 w= We (053) oo HANKELSINGULAR VALUES 305 Hankel Operator As mentioned in Section 62.1, the Hankel singular vat interpreted in terms of the singular values of the Mankel operator fin Detintian-B1-This operator provides an intresting Tlation between the iaput-outpat pains of a system and the controllability and ob- servability of its states. I is easly seen that Definition 6.1 is equivalent to: y(Q) =Tev {Rrcemnnude, 130 cE 1<0, where v € £(0, 0) ‘The interpretation of this operator in terms of inputs and outputs is as fol- lows, Take the input u(t) = o(-1), ¢ < 0, therefore the input-output transfor. ‘mation can be performed in terms of the Hankel operator: y(¢) = (F@v)(t) (g + 4)(0). This is the convolution of a system whose input is defined for neg- ative times (in £3(—00,0)) and its output for positive times. The final state at r= 0 for input u(t) is x) = f. eM Bus) dz (955) 1 corresponds (0 the initial state for the unforced system whose output is (0), > 0. This operator transforms past inputs into future outputs through the state (0). ‘The Hankel singular values can be defined either way, as in Chapter 6 (Definition 62) or in terms of the Hankel operator. Furthermore, it can be shown that the Smith-MeMillan order of a system is equal to the number of positive Hankel singular values. Recall that the Hankel norm of a system G(s) has been defined (Defi tion 6.3) as the largest singular value of its corresponding Hankel operator 1G(S)lln = @ ['c}. The Hankel norm can be interpreted as an induced norm between past inputs and future outputs [see equation (6.19)}: max WPale* wdla welieol Tle NGC) = (056) which also equals /p(W.W.)- Note that the Hankel operator, its singular values, and the Hankel norm are all independent of the constant D term of the realization, This is the reason why, without loss of generality, we have considered only strictly proper systems Next, we present the results for discrete time systems. In turns out that if the latter is obtained as a bilinear transformation of a continuous system, the Hankel singular values and Hankel norm remain unchanged. 306 MODEL ORDER REDUCTION 9.3.2 Discrete Systems Consider the discrete time, linear, shift invariant, asymptotically stable system: AUK T] = AgakT) + Bul kT) (951) WAT) = Cyx(kT) + DgulkT) (9.58) Since [A,(Ay)| <1, Vi = 1,...41, the Gramians can be defined as we © atauat (a4)! (959) 2 (As)! cfcuat (960) Both are obtained also as solutions ofthe discrete Lyapunov equations we — A,WiAl — B,BT = 0 (961) Wy -AlWiAy—C}C, = 0 (9.62) respectively. Similar to the continuous case, the Hankel singular values are defined as follows. Definition 93 Consider the asymptotically stable system Gaz) = [ eit | 063) with controllability and observability Gramians Wand Wi, respectively. The set of Hankel singular values of G(s) are the square roots of the eigenvalues of the product WW, indicated as (wewan, = (fo, v.20 (964) For convenience they are usually ordered as follows aff > af >. > of! Next, we show that the equivalent continuous system, obtaines bilinear transform of the above, preserves the controllability and observability Gramians. Therefore the Hankel operator, singular values, and norm ave preserved as well. Also, due to the fact that the McMillan degree can be defined in terms of the Hankel singular values, the order of both discrete and continuous systems is the same WANKEL SINGULAR VALUES 307 ‘The bilinear function is (965) (9.66) which maps the interior of the unit disk |z| < 1 into the open left-half plane oi) = Gx(e= 44) a and for the state-space realization matrices (Problem 3) A= (4 Ag Ag—D) (9.68) B= VUI4 Ag 'By (9.69) C = VIC {I+ AG (9:70) D = Dy Gilde Agy'Bg (on) Next, pre-and postmltiply the continuous controllability Lyapunov equa- tion by (J+ Ay) and (1+ Af) as follows (14+ AAW + WAT + BBTVI+ AT on) By substituting (9.68) and (9,69) we obtain equation (9.61) (Problem 3). Its unique solution therefore coincides with the continuous case, that is, We = W4, Similarly, we can prove that W, = W. Furthermore, if we evaluate both systems in the boundary of their re- spective stability regions (|z| = 1 and s = jw), we obtain from the bilinear ofrm(4)] on This indicates that both frequency responses coincide in @ € (0, 7), and there fore their respective infinity norms. Due to the fact that the McMillan degree ccan be defined in terms of the Hankel singular values, the order of both dis crete and continuous systems is the same, Gale" 2 For simplicity consider the sa 308 MODEL ORDER REDUCTION From the above, we conclude that any model approximation of systems G(s) and Gq(2) of arder n by others G,(s) and G¥(z) of order r 0) can be obtained by combining the above equations: (0) (1) HT)) = KoKe | M27) (979) ae As before, the Hankel singular values of = 0,(H) provide a measure of the relative controllability observability of each state x(0). Geomet- rically, we can represent the above equation as a hyperellipsoid with each axis representing an eigenvector of Hf with a relative magnitude given by off. Therefore the relative controllability /observability of x(0) depends on its direction relative to the principal axes of the hyper ellipsoid. Finally, we define the Hankel norm of a diserete time system. Definition 95 The Hankel norm of a stable system Ges) is the largest singu lar value of its corresponding Hankel operator: \Gals)\n = GH) (or equiv alently the largest Hankel singular value of") 310 MODEL ORDER REDUCTION 9.4 MODEL REDUCTION 9. 1 Introduction The purpose ofthis chaper is to provide @ model reduction procedure sich that, given a model Gl) of MatMilan depee n, we btan a reduced ——r——e——e T_T Pen. Woate ako interested in aving 9 hound on the appx G(s) — G.(s)la in terms of a meaningful norm a. By meaningful we mean that the specific norm chosen should have relevance from the modeling and contel pint of view "ake for example, the Hankel norm defined inthe latter section It po r-rel mmentl or model reduction, is main Objective bel relevant sates from the input-OUIput perspecive. that i. iy” oF & fealization from the numerical point of view. a r—t——r—r— be explained quite imply, we ise a reducedorder mode! o design acon —— se _ it fits naturally as (global) dynamic uncertainty. As we have seen Fnchapters2 and the obust stay coniton for his ype of uncertain brebocedon te M-norm in lace of the rests weave so far (Gee Lemma 92) the wuncation of states that correspond to small Hankel singular sales proves an optimal projection in the norm Tn th section me describe the balanced-truncation model reduction a )— Clr ———§__ Cro in terms ofthe infinity and Hankel norms Although the procedure oes not sve the eptimal problem, min||G(s) — Gy(sNlas @ = Hyco (980) it guarantees an approximation error that is very close to the optimal, with a smaller computational effort. The optimal Hankel norm approximation prob- lem has been solved in (132). An extension of the balanced-truncation method ‘and the optimal Hankel norm approximation applied to the approximation Of infinite dimensional systems can be found in [79] 9.4.2, Hankel Operator Reduction For simplicity, consider the discrete time system described by matrices (Ag, By, Ca,Dq) with Hankel operator H of rank n. Due to the fact that the rank of the latter is defined as the order of the model. we can state the MODEL REDUCTION 317 ‘model order reduction problem in terms of this operator as follows |H — Helle (ai) where H, is a Hankel operator of rank r. This problem applied to finite dimensional matrices was one of the early uses of the singular value decom- position. The solution is due to Mirsky ((208)) and is easy to prove. Theorem 9.1 Take MeO" with a singular value decomposition M URV*, 3 = diaglay, =. -50] such that 0) > --- > om >0. Also, con: sider a unitarily invariant matrix norm, that is, |(M lly = ||UM Ilo = IM Ul with unitary U. Then fon 0 int {\\a# — atl, , ROM) aXe”) (993) and rearranging, we obtain K(WeW,) = 2 (994) Be) gclT)Ko(T) (0.95) (We) 314 MODEL ORDER REDUCTION where we have used in (994) the symmetry of W, and W,, to obtain 03 (W.W,) = 4 (W,W2W,) = AZ (WW, ). Furthermore, we have: max{xe(T),o(T)] > Yre(Fwo(T) = (3) (996) ‘Therefore the minimum is achieved by T = T.y for which x¢(Tin) = wo(Tis) = (of! jaf!) a ‘This property of balanced realizations indicates that they offer the best comproinise in terms of conditioning of controllability and abservability mea sures Graphically, this means that they Tend to make both controllability and 0 and ~2BB" z, = WB" |} <0, we obtain Re(A.) <0. (9.99) MODEL REDUCTION 315 If (A, B) is not controllable, using the Hautus test we have (9.100) and using (998), we obtain Re(A,) = 05 that i, Ais unstable with imaginary unstable eigenvalues. Now assume A is nt stable and agai, using (295) all unstable eigenvalues are located inthe jo axis, say, at A= yoy-From (9.98) we obtain |B" sy], = 0 which implies (4, 8) is not completly controllable. a By duality we can prove a similar result for observability using the corre- sponding Lyapunov equation. Next, we prove the main result Theorem 92 Consider a minimal, asymptotically stable balanced realization with states x(1) € RY, ¢ > 0: (9.101) ow = L244 with Gramians W. = Wy =% = diag (of! into 1wo subvectors as follows oft). Divide the sate veetor 10)= [20] wie nt emt i= 12, 20 such that ny +nq =n, and partion the system and Gramians accordingly An Av By a-[4 4]: 5=(3] c=[a a] (0.102) xy 0 eo (9.103) Assume the Gramians of both subsystems do not share common diagonal ‘elements. Under these conditions both realizations, [He ‘are minimal, internally balanced, and asymptotically stable Gis) 2 (9.108) Proof. Separating both Lyapunov equations into two subsystems, we obtain for the first block-diagonal entries Bay +AhE+ CFO = 0 (2.05) An Si +3,Ah + BBP = 0 (0.106) Br. to, a8" 316 MODEL ORDER REDUCTION and for the lower-eft block entries AB, +EAh+ By = 0 (@.107) ALE, +¥adn+ CFC = 0 (0.108) Both (9.105) and (9.106) imply that subsystem (Ay), B1,C;) is balanced, ‘Assume it is not stable; therefore by the last lemma Ay; has at least one imaginary egenave sy ay yo, Then (An poh¥ =0 ox) V" (Ale sot) <0 ox) where the columns of V span the right mullspace of (Ay) ~ jo). In (9.105) and (9,106) add and subtract jw. Then pre- and postmultiply the former by V* and V, respectively, and the latter by V*E) and 5\V, respectively, Then: VIS (Ay —s01)V + VAT, + J01)EV =-Viclov =0 ot) 2CV 205 (Ah + 0D3,V (0.112) VeE(An ~0DEV+VSHAT + 10)EV (0.13) S1B\BTEAV =0 (9.14) 3 BIZV = 0 (Ay ~ jo)2V =0 (9.115) Therefore the columns of 32V are included in the nullspace spanned by V and we can always find a matrix 3, such that 32V = VS}. In particular, V can be selected so that Eis diagonal with a subset of the diagonal entries of. Now postmultiply (9.107) by 3,V and (9.108) by V to get An3V +E2AR3V = 0 (9.116) AREV +2,Anv = 0 on) ‘which in turn implies BAnv (8) Because £, and ¥; do not share common entries, it follows that A2)V ‘Therefore rearranging equation (9.118) and the eigenvalue equation for Ai1. MODELREDUCTON 317 Ay An [a a2] [5]-[0] one which contradicts the stability of the original system, Therefore (Aj), B,C) is asymptotically stable. Similarly, we can prove the stability and balance of system (Azo, Bp, Cy) as well. Using the previous lemma, both subsystems are controllable and observable, therefore minimal o The above result is useful as a model reduction method. The objective is to separate the full model G(s) into two subsystems G,(s) and Gy-r(3). Cate must be taken to order the states according fo Their Hankel singular Values, such that of > --a!" > off, ---> a In this way, the truncated model with the first r states G(s) contams the str Hankel singular values, Therefore the mosf controllable and observable states are contained in this reduced system, which are more relevant in an input-output sense. The assumption that both subsystems should not share common Han- kel singular values is extical. The following example (240]) indicates that, otherwise, one of the subsystems could be uncontrollable, unobservable, or unstable. Consider the balanced system xo= [22 Eas [ha w= [1 0] xt) Both Hankel singular values are 0.5 and, clearly, the subsystem (0,0,0) is neither controllable, nor observable, nor stable. From the practical point of view it is not a problem, due to the fact that, generically, any matrix has distinct eigenvalues, Also, for model reduction, it is important to eliminate ‘the subsystem that has very small Hankel singular values (a/ > of,), which is not this case. This last point is important when considering tie model reduction error. It can be proved ({132)) that the approximation error for the balanced and truncated method, defined as the norm of the residual system Gy-,(s), is given by IG(s) - G()Ila <2 of, «= H,c0 (0.120) for both the Hankel and infinity norms, when all Hankel singular values are distinct (otherwise we may have an unstable error). Although this error is Rot minimal, it is close to the optimal solution. In (132) the family of optimal reduced. order models, measured in terms of the Hankel norm approximation, is obtained. The corresponding minimum approximation error is min |IG(s) — Gils)Ilw = off, (9.121) 318 MODEL ORDER REDUCTION It is possible to find a D, matrix for the Hankel optimal reduced-order model G,{s) such that G(s) ~ (Gels) + Dill = So of (9.122) This is so because the Hankel norm does not depend on the constant D term of the realization Therefore the balanced truncation algorithm produces a solution subopti- ‘mal up to a factor of 2 in the infinity norm, with less computational work 9.5 ALGORITHMS ‘Moore's algorithm ((209)) for balancing an asymptotically stable system [42 aw = [244] (9123) is a follows. 1, Compute the controllability Gramian W, from the corresponding Lya- punov equation, 2. Compute its Jordan form with ordered eigenvalues (largest first) and orthogonal eigenvectors: Ve (9.124) This is used to form the transformation matrix T) 2 3c'Vz. which produces the new system matrices A = TAT; ', B= T,B, C= CT," ‘The new controllability Gramian is W. = 1 3. Compute the eigenvalue decomposition of the new observability Gramian as before, Wy = Vo3Vg. This is used to form the transfor- mation matrix Vg, which produces the balanced realization: Ay = TATs! = TAT, "Ts! (9.125) By = nb =r. iB (9.126) Gye ETS C7 S'T5! a2) liza under It is not difficult to verify that this is_an internally hatance tion and that the transformation matrix T = T>T is nonsingular certain assumptions (Prableni'8)~~—~ ALGORTHMS 319 Another version of this algorithm was developed by Laub {180] and Glover (133) 1. Solve the corresponding Lyapunov equations to compute the Gramians W. and W,. 2. Compute the Cholesky form W, = R'R (also known as the matrix square root). 3. Compute the Jordan form of RW.R* = US*U", selecting U to be unitary and ordering the eigenvalues as before. The transformation ‘matrix that takes the system to its balanced form is T= 3-"?U"R, with Ay = TAT!, B, = TB, Cy = CT Its also easy to verify that this is an internally balanced realization and that the transformation matrix T = 737; is nonsingular under certain assumptions (Problem 9), In both cases, after a balanced realization is obtained, a truncation step needs to be performed to reduce the system's order, A, Ap B, we[h a me [E] C= [a CG) Gls) = i Be | (0:28) 5.1, Approximation Error Selecting the order r of the reduced model entails trading-off model size versus approximation error. It is desirable to have a small number of states due to computer speed and memory size constraints, but this may increase the approximation error. To this end it is important to consider the following, ‘bounds, taking into account that the r selected states corsespond to the larger Hankel singular values {0}, i = 1,7) NGC) ~ Gels)Ily <2 of! (9.129) G(s) ~ G6} <2 of! (9.130) 320 MODEL ORDER REDUCTION If the original system is not strictly proper, that is, D # 0, we select for the reduced-order system D, © D. This choice does not modify the Hankel sin- {gular values or the approximation error bound and causes the approximation error at higher frequencies tend to zero, that is, lim,» |G(s) ~ Grls))la — 0. If the system has a stable and an unstable part, we can reduce the order of the stable part and retain the unstable one Gls) = Gls) + Guls) (9.131) Gls) = Gals) + Gals) (9.132) G(s) ~ Gio = G8) ~ Garb)» Hoo (9.133) Again, the approximation error remains unchanged as well asthe number of unstable poles. This is an important assumption for robust stability analysis (Chapters 2 and 4). I we consider G,(s) as the nominal model from which a robust controller will be designed, we may interpret the error (Gu) ~ G, (jo) as frequency-weighted additive dynamic uncertainty. This family of models should “cover” the plant G(s) by means of a nominal reduced-order model G,(s) and a frequency-weighted bound. Additionally, the numberof unstable poles of each member ofthis family should remain the same, in order forthe necessary and sufficient condition that guarantees robustness ofthis class of (additive) dynamic uncertainty sets to be vali ‘Equivalently. if G, (3) isinverible, we may consider the trequency-weighted ‘multiplicative uncertainty [G(Jw) ~ G,()|G, "Gu. In most cases, the weight for multiplicative uncertainty is lower at low ftequencies and increases at higher ones, up to a eitcal « for which it increases above 100%.-On the other hand, the behavior of the balanced-truncated approximation error is just the opposite: its higher at lower frequencies and goes to zer0 as w ~ 00. ‘Therefore, for robust control applications, its important to invert this fre- quency characteristic of the approximation error. This can be achieved by balancing the realization of G(s), but truncating @ new system G(s) = G(i/s). The reduced model is piven by G,(s) = G,(1/s). ‘This transforms lower into higher frequencies and preserves stability, Gra- mians, and, asa consequence, also balancing (see [24). Furthermore, the error bounds remain the same, although the error asa function of frequency changes, Specifically, the algorithm proceeds as follows: 1, Transform the balanced system G(s) to G(s): Het] [eatoteee] ono ave [A 2)'-[ ee] ons PROBLEMS 321 ‘based on the partition of the original balanced system in (9.102). Then the truncated system G(s) is 6.0) [Reon Enea GRi-GRa | D-Cate | 136) 3. Finally, invert the frequencies on the reduced-order model G,(s). Sim- Plifying we obtain (prove) Au ~AvAg An | Bi-AnAgB (9.137) Gis) = | Ae] 08 C= GAAn | D- GAB: a Note that the matrix Az, is always invertible, because it is stable, Thus Gy(s) is well defined, There are other algorithms that solve some of the numerical problems inherent in obtaining balanced realizations ((263)).It is also possible to per form balanced truncation over badly conditioned realizations or nonminimal cones. These algorithms are implemented in commercial CAD tools: Mu-Tools and Robust Toolbox from MathWorks ([20, 262). 9.6 PROBLEMS Prove that the Gramian Win (99) is symmetric and postive semi- ‘definite, and therefore its eigenvalues are all real and nonnegative, 2,/In the proof of Lemma 9.3, verify that signal u(t) takes the initial state ~' 2(0) to 2(7) with minimum energy. 3. Obtain the expressions in (9.68) and (9,69) as functions of the matrices of the equivalent discrete time system. Prove that substituting in (9.72) ‘we obtain the controllability Gramian Lyapunov equation (9.61). 4. Show, by means of an example, that the Hankel singular values of the difference of a system and its balanced-truncated version are not equal to the Hankel singular values of the difference of their Hankel operators, In which case could they be equal? ‘&/ Prove the relation between matrices Tip, Tiqy and Ton, a8 indicated in (9:88) and (989). \@ Consider a stable, strictly proper minimal realization represented by ma- trices A, B, and C. Prove that any balanced realization of this system can be transformed to another balanced realization by means of a ut satisfying TE = 37, anced Testization tary ith W, = W, = 3. Thus there is not a unique bal- 322 MODEL ORDER REDUCTION 7. Prove tha Moore’s algorithm transforms a realization into its balanced form, Which is the assumption that guarantees Tis a transformation matrix? 8, Prove that the Laub/Glover algorithm transforms a realization into its balanced form. Which is the assumption that guarantees T is a transfor- mation matrix? From the point of view of computational cost, what is the difference between both algorithms? 10, Consider the Jordan form of the product of Gramians ofa system, W.Wo = TET" with unitary 7. Can T be used asthe transformation matrix which takes the realization to its internally balanced form? IL. Verify the equations for the algorithm that provides zero error at « = 0 presented in Section 95.1, ROBUST IDENTIFICATION 10.1 INTRODUCTION In the first part of this book, the analysis and synthesis of robust controllers were considered. A basic point that has not yet been addressed is the method by which a set of models that represents a specinc physical process can be obtained. Before the appearance of systematic methods, families of models were obtained by ad hoc procedures (see Example 11.2). At the end of the 1980s, the first algorithmic strategies were introduced, based on approxima: tion techniques that provide a uniform error bound ((134, 139, 321)). Classical identification procedures (184, 293]) are not adequate for ro- bust control, because usually they identify, using stochastic methods, a set of parameters of a fixed mathematical structure. These procedures are used to identify parametric models in adaptive control. Within this framework, a fixed model order is assumed, which is not the case when using robust design and analysis tools. Even if families of models with parametric uncertainty could be obtained in this way, there is a limited design machinery for robust analysis and synthesis of this class of uncertain plants. In general, the analysis of parametric sets of models is a NP-hard problem ([60, 248). Robust control analysis and design tools are based on a deterministic worst-case approach, with no previous assumption on the order of the system. For this reason intense research was initiated in the 1990s on deterministic identification procedures, which can be used as a first step in a robust control design methodology. This area is called robust identification, ‘The robust identification problem has been posed in (143, 144, 145, 146] and has attracted considerable attention since then. It considers model uncer- tainty due to two different sources: measurement noise and lack of knowledge 323 324 ROBUST IDENTIFICATION of the system itself due to the limited information supplied by the experimen- tal data. Therefore these new identification procedures are based not only ‘on the experimental data (a posteriori information) but also on the a priori assumptions on the class of systems to be identified. The algorithms produce ‘2 nominal model based on the experimental information and a worst-case bound over the set of models defined by the a priori information. If time response experiments are performed, an é; error bound can be de- fined ((75, 193, 194, 197), which is adequate for & optimal control. Further- ‘more, these types of experiments are useful in verifying the (approximate) stability, ime invariance, and linearity of the systems ((272), for example, before a frequency response experiment! is performed. ‘The uncertainty bounds can also be measured in terms of the Ma, norm (9, 137, 138, 146, 147, 191, 196)), which is more adequate for 7. optimal control or w-synthesis procedures. In this case, the experimental data are obtained from the frequency response of the physical system, whichis also assumed to be linear and time invariant ‘A general abstract formulation of the problem will be presented next. We postpone the specific algorithms and convergence results using Mx. and metrics to later sections of this chapter. 10.2 GENERAL SETUP In this chapter, we present algorithms for solving the problem of (robust) control-oriented identification of physical plants. The outcome of these pro- cedures is a family of models, represented by a nominal one and an identifi- cation error. We start with an abstract and general formulation ((304)), which does not take into consideration specific metrics (used to evaluate identifi cation errors and convergence) or model and noise sets. The objective of presenting a general formulation is to point out the differences with classical parametric identification methods ((184, 293]). It is also useful to understand the fundamental limitations that appear when working with corrupted and incomplete data and model ser inclusion, from an information theoretic point of view. As a consequence, the structure of the information on the plant will be stressed. Furthermore, this general setup is useful to indicate the similarities with approximation procedures of infinite-dimensional models. These will be in- cluded as particular cases ofthe identification ones, Recall that approximation is loosely defined as a procedure that represents a known large-order mode! (even infinite-dimensional) by a simplified smaller-order one. On the other hhand, identification is meant as model computation from (possibly) a priori information and experimental data, "Athough all experiments are performed in the time domain, «sinusoidal swcep atthe input ofthe plant obtain an output magnitode and phase response i deiaed here as» equeney ‘esponte experiment (sce Seton 1032), GENERALSEIUP 325 10.2.1 Input Data ‘The outcome of a robust identification procedure is a family of models that should include the real physical plant behavior. This family is specified by 2 nominal model and an uncertainty error measured in a certain norm. For the problem to be nontrivial, classes of candidate models and measurement noise should be assumed? as will be explained later in greater detail. The uncertainty bound should be valid for any plant included in the class of mod: els or noise inside the class of measurement noise. This naturally produces a worst-case approach. Therefore the input data to the problem arc the class of candidate models and measurement noise assumed, called a priori infor- ‘mation, and the experimental data, called a posterior’ information, ‘The a posteriori information is a vector y € C™ of experimental data cor: rupted by noise 7. The data can be frequency and/or time noisy samples of the system to be identified. For a model g and a given noise vector 7, the experiment can be defined in terms of the operator E(-,-) as follows: y= E(g.n) (10a) For simplicity we assume E(.,-) linear with respect to both variables, the noise being additive, Note that this is not an injective operator because the same outcome y may de produced by different combinations of mode! and noise. This 18 4 restatement of the fact that the information provided by y is incomplete (M samples) and corrupt (noise »). Therefore the operator is not invertible and 1no direct operation over y will provide the mode! g. Instead, a type of set inversion will be attempted. By this we mean the computation of a set of models, described by a (central) nominal one gy and an uncertainty bound ¢y- This set “covers” all possible sets of plants in the a priori class, which could have produced the a posteriori information y. In order 10 make this a nontrivial problem, itis necessary to add extra information on the class of models and measurement noise. Otherwise the combinations of models g and measurement noises 7, which could have produced y, would not only be infinite but would also form unbounded sets. The set of candidate models would have an uncertainty bound ¢y —+ e. This last point can be exemplified as follows. Assume we have an ex- periment that produces the vector of real numbers y = [yo dual" ‘These represent the (truncated) impulse response of a certain (unknown) system g, measured with (unknown) noise m, at times ty = KAY, that is, Yi = Bulls) +76), k=0,...,M — I. If the only available information isthe vector y € RY, there is no way to identify plant g.. Note that we can always Fepresent every measurement point by yx = glt4) + m(t). =0,--.)M— 1, where g(t) = yu-+z and ns.) = —2, for any arbitrary real value z. There: Otherwise the uncetsnty bound could be infrite and both the maxing and contr problems would not make seme 326 ROBUST IDENTIFICATION fore any (possibly unbounded) model or noise could generate this same set of data points Even without measurement noise, that is, 7, =0, the candidate models _g() could be almost anything, as long as yy ~ gli). k = 0,..-5M ~ 1. We can interpret this result in terms of the more familiar aliasing concept. Specifically it is related to Shannon's theorem, which poses an upper bound on the class of plants that could be identified, using this experimental information y. Only if g, belongs to the FDLTI class of models with bandwidth less than 1/24 ccan it be computed from the available data. If the bandwidth is greater than 1/241 but bounded, a smaller (and bounded) set of candidate models (which include g,) could be determined, The above argument justifies the need to define a class S of models and N of measurement noise. Specific classes will be described in further sections of this chapter, according to the norms used to bound the uncertainty error. Hence the input data to a robust identification algorithm are composed of both @ priori (S,N) and a posteriori (y) information. Duc to the fact that the assumed @ priori information is a quantification of the engineering common sense or simply a “leap of faith,” there is no guarantee that it will be coherent with the experimental a posteriori data. Therefore consistency of both types of information should be tested before an algorithmic procedure can be applied to compute the family of models. 10.2.2 Consistency Consistency is a toncept that can easily be understood if we first define the set of all possible models, which could have produced the @ posteriori information y. in accordance with the class of measurement noise: SY) = {gE S | y= Elen), EN} (102) ‘Therefore S(y) CS isthe smallest set of models, according to all the available input data (a priori and a posteriori), which are indistinguishable from the point of view of the input information. This means that with the knowledge of (y,5,.N) there is no way to select a smaller set of candidate models. Note that the above set is independent of the metric selected and depends on all the input information. ‘As we mentioned in the last section, for a robust identification problem to make sense, the identification error should be bounded. Since the “size” of the family of models is defined by this error bound and should include the set S(y), this set should be bounded as well. For this itis sufficient to have a bounded class $ of models (see Figure 10.1). Furthermore, the “size” of the set S(y) places a lower bound on the identification error, which cannot be decreased unless we add some extra information to the problem. The latter can take the form of new experimental data or more detailed information on. the sets S and/or NV. This lower bound on the uncertainty error holds for GENERALSEIUP 327 noise (corruptness) Son . $0) gorithm 2'@ System partins) —— a Figure 10.1. Gonotal statement of he problem any identification algorithm and represents a type of uncertainty principle of identification theory (see Lemma 10.4), It is clear from the above arguments that consistency of @ priori and a posteriori information can be defined as follows. Definition 10.1 The a priori information (S,N) is consistent with the exper- imental a posteriori information y if and only if the set S{y) is nonempty. It should be noted that once consistency between « priori and a posteriori information is established, both are equally reliable. Nevertheless, this docs not imply that the “real” plant g. belongs to S. This statement deserves a more detailed explanation, which will be presented next. 1k could happen that, although (S,V) and y are consistent, the set S does not include the real plant g., that is, the @ prior’ information is false. This could be the case of very noisy measurements (A’“large"), which allows the set S(y) not to be empty, although g, ¢ S(y), that is, there exists gy € S(y) such that E(@1,m) = Elgasm) =y with m,n EN. From @ practical point of view we can detect this problem by performing more experiments, say, {y',92,...9"}, ¥ ERY .i= 1,...,N. If we define this set of experiments a6 A, the new consistency set is Sa)={eeS | y= Elen!) a eNandye a} (103) Its clear that if one of the experiments in A is y in equation (10.2), then S(A) € S(y). In fact, we can equivalently define S(A) =n yS{y'). If the sets $ and AV meet certain conditions (detailed in further sections) and the experimental operator E(-,-) makes sense, as we increase the number of experiment, the consistency set might get “smaller.” Also, the a posteriori local error bound may decrease (see equation (10.5) Nevertheless, from @ worst-case point of view, although N’ new exper iments ate performed, we could always obtain the same set of measure- ments y. In this sense, the new consistency set would not “decrease,” that is, S(A) = Sty). This can be interpreted in terms of the a priori lowest ertor bound (see Lemma 104). Because this a priari errar is based on the worst- case experiment, it will not change with new experimental data, Therefore to 328 ROBUST IDENTIFICATION decrease the consistency set (and the a priori lower error bound) new types of experiments need to be performed, for example, increasing the number M of measurements, decreasing the measurement noise bound, and incorporating ‘2 new type of experimental data (time and/or frequency). However, and this concerns any empirical theory, no matter how many experiments we perform there is always the possibility of having the real plant excluded from the set S, even for consistent input information, This is ‘an epistemological problem that relates to the following facts: '* Any empirical theory (or model) can only be falsified (see (245) ‘+ The number of experiments is always finite At this point, the following practical remark can be made: To reduce the possibilities of having the real plant g. excluded from S, as many different types of experiments (time, frequency, changing sensors, and/or data points) as possible should be performed. The final verification is always through a new consistency test. This isthe best that can be done in the framework of the scientific methodology. In the case of approximation, the set S has only a single element (usually infinite-dimensional). Therefore the “experiment” that consists of @ partial sample of this system (§M data points), “measured” without noise, is always consistent. The uncertainty in this case arises only from the partialness (M data points) of the information. The only element of S (the “real” system) cannot be used as the desired model g,, because of computational reasons. In later sections, the consistency problem will be stated as determining the existence of a model g € S, such that the difference between its output and y is included in set NV, that is, it is only due to measurement noise, This may be solved as an interpolation problem. The latter holds for additive measurement noise and linearity ofthe operator E(-,-) with respect to model and noise, which we have assumed. 10.2.3. Identification Error The a priori knowledge of the real system and measurement noise present in the experiment y is stated in terms of sets S and AV. The statement of the problem does not assign probabilities to particular models or noises; therefore it is deterministic in nature. In addition, the modeling error should be valid no matter which model g S is the real plant (or € NV the noise vector) leading to a worst-case approach. In this deterministic worst-case framework, the identification error should “cover” all models g © S that, combined with all possible noise vectors 7 € JV, are consistent with the experiments. Ideally then, the error should only include the models g € S(y). In practice, however, the family of models conservatively covers this “tight” uncertainty set. Hence it provides an upper bound for the distance (in a certain metric) from 2 model (consistent or not) to the seal plant. In this framework, the worst-case ertor GENERAL SETUP 329 is defined as follows: A) E sup. 8 ATE(S, MSN} (104) where £(-,:) is the experiment operator and m(.,:) a specific metric. The identification algorithm A maps both a priori and a posteriori infor: mation to a candidate nominal model. In this case the algorithm is said to be tuned to the a priori information. On the other hand, i it only depends ‘on the experimental a posteriori data, itis called untuned. Almost all class cal parameter identification algorithms belong to the latter class ({184)). For simplicity, in the sequel, the identification algorithm will be denoted as A, without specifying its arguments unless needed, ‘The identification error (10.4) can be considered as @ priori, in the sense that it takes into account all possible experimental outcomes consistent with the classes A’ and S, before the actual experiment is performed, Since it considers all possible combinations of models g and noises 7, that is, all experimental data y, it is a global identification error in terms of the ex: Periments. A focal error that applies only to a specific experiment y can be defined as follows: Ay) 2 sp mig. ASN) (105) From the above, it can be shown that the global worst-case error can be defined equivalently as follows, Lemma 10.1 (A) = sup e(A.y) (106) ye¥ where Y is the set of all possible experimental data, consistent with sets S and Nv ee By definition ¥ © {E(g, n)|g € S, 1 € N)therefore the set {g € S(y) Y€ Y} is equivalent to S, From the definition of Y and replacing y by E(g, n) in (105), we obtain sup sup m (a, A[E(G,n),5,N1) = sup sup mg. Aly,S,]} wee yeveesiy Obviously the algorithm always depends onthe posterior’ information forthe computation fan empirical model 330 ROBUST IDENTIFICATION Clearly we always have e(A,y) < ¢(A). To decrease the local error more experiments need to be performed, whereas to decrease the global error new types of experiments, compatible with new a priori classes, should be performed, for example, reducing the experimental noise and changing ” accordingly. ‘As mentioned before, there is an uncertainty principle in identification, which produces a lower limit to the possible global error, closely related to the consistency set S(y). To this end we first define the radius and diameter Of a set and prove several properties. Definition 10.2. The radius and diameter of a subset A of a metric space (Xm), are 1A) = ink sup m(x,a) (107) (103) ‘The radius can be interpreted as the maximum error, measured in a certain ‘metric, when considering the set A as represented by a single “central” point (which may not belong to A). The diameter is the maximum distance between any two points in the set. From these definitions, the following properties can be proved. Lemma 10.2 1A) < (A) < 2r(A) (109) Proof. From Definition 102 we have (A) 0, VEX | YEA m(x,y) 4100) (1016) ye ‘As a consequence, there exists an experiment y., a pair of models gy, 2 € S(y,), and noises m,m € N, such that: E(u.) = Elen.) = ys (10.17) m(gi.g2) > a1S(yo)] (10.8) By symmetry, the element (2es ~ g1) € S, and by convexity the element ty © J Igo + (cs — gi)] € S as well. Hence the element hy 2 cs + $(gi ~ 82) belongs to the set S, due to its symmetry. Replacing $ by AV we can prove in a similar way that the elements jx © ew + }(m1— m2) and #2 2 ew + Hm — 1m;) belong to NV. Using the linearity of the experiment operator E(-,-), we have Yo = 5 (Ete. m) ~ Ear, m)] + Eles.cw) (10.19) ~ E(u in) (1020) and similarly E(/p, ja) = yo This proves that hy, fz € (yo), with m(h, a) = ‘m(gi, 2), which contradicts (10.18), Tis easy to verify that S(yp) is a symmetric set, and therefore supm(sy) < maxes) + mses) (1021) i 2max [m(x,cs),m(y,c5)] (10.22) which achieves equality when y = —x. The diameter is computed among the symmetric elements, which proves (10.15). o Ifthe sets belong to normed linear spaces, we may use the metric mab lla- BI, ‘The above proves that the experimental output achieved by the centers of both « prior sets provides the least amount of information (largest diameter of indistinguishable set). In the case where these centers are the nul elements, it means that the experimental instance with null output provides minimal information, which in some sense is intuitively clear. From the previous discussions, it seems important to have identification procedures that can compute models inside the consistency set. These are called interpolation algorithms and are defined as those that produce models GENERAL SETUP 333 included in the set S(y) for all possible experiments y € ¥. The name comes. from the fact that algorithms that interpolate the experimental data produce a model g that satishes y = E(g,0). and therefore g € S{y) if g € S. ‘A lower bound for the worst-case error of any identification algorithm and fan upper bound for any interpolation algorithm are presented next Lemma 10.4 The worst-case identification error defined in (10:6) satisfies the following inequality e(A) 2 R(T) > 5010) (1023) for any algorithm A The following upper bound holds for any interpolation algorithm Ay Diz) > eA) (1024) Proof. Its clear that sup e(Ay) > e(Ay) > inf _e(A.y) (10.25) agree) ates Next, take the supremum over all experiments y € ¥ on each member of the above inequality. Based on definitions for radius and diameter of information and equation (10.6), we obtain for the first inequality DU) > eA) (1026) ‘The above holds only for interpolation algorithms because the supremum has been taken over all Ay(y) € S(y). Using Lemma 10.2, we obtain for the last inequality (A) 2 RG) > Spa) (1027) Which holds for any algorithm A o ‘As mentioned before, e(A), D(Z), and R(Z) are global in terms of the experimental outcome y. Therefore the above a priori bounds hold for any experiment (consistent with S and NV) and may be used before any mea: surement is performed. As a consequence of the above lemma, interpolation algorithms can obtain the minimal error bound R(Z) up to a factor of 2, that is, D(Z) > e( As) > $D(Z), “almost” optimal In terms of local errors, an important subclass of interpolation procedures, called central algorithms, generate a model g. that satisfies (Ay) = sup m(g.g.) =r 509)) (1028) ws) 334 ROBUST IDENTIFICATION ‘This “central” clement may not belong to the set itself (see Definition 10.2 may be nonunique, or may not exist. If it does exist, then itis called Cheby- shev center and represents the best approximation by a single element of the smallest indistinguishable set S(y). From Lemma 1044 it follows that these al gorithms provide the smallest possible worst-case error. Therefore they may be considered optimal in that sense. The importance of symmetric sets is that, the symmetry point is a Chebyshev center of the set Optimal Algorithms 1: is important to consider algorithms that make an clfcient use of the availabe information and produce the smallest possible identification errors. These optimal algorithms ({75) are defined next. Definition 105 The local optimal identification error of a particular experi- ment y cis ©) 3 infe(ALy) (1029) An algorithm A* is called locally optimal if and onty if A, )= ey) Definition 10.6 The global optimal identification error is defined as nf e(A) (1030) An algorithm AY is called globally optimal if and only if e(*) = e*. Further ‘more, an algorithm is called strongly optimal if itis locally optimal for all experiments y € Y. Note that strong optimality is much stronger than just global optimality. I implies that, regardless of the set of available data, the identification pro- ‘cedure provides an estimation of the model, which minimizes the worst-case error. The central algorithms defined before are strongly optimal 10.2.4 Convergence ‘A question that might be posed is the following: What happens with the family of models when the amount of information increases? It would be desirable to produce a “smaller” set of models as input data increase, that is, model uncertainty should decrease. This could be interpreted as having ‘monotonically decreasing uncertainty bound with inereasing input informa- tion, Certainly, we would expect the set of models to tend to the real system (monotonically or not) when the uncertainty of the input information goes to zero. This latter property can be stated in terms of convergence as follows’ ‘An identification algorithm A is said to be convergent when the worst case lobal identification error e( 4) in (10.4) goes to zer0 as the input information GENERALSETUP 335, tends to be “completed.” The latter means that the “partialness” and “cor ruption” of the available information, both a priori and a posteriori tend to zero simultaneously ((146) Input information is corrupted solely by measurement noise inthis scheme, ‘Therefore “corruption” tends to zero when the set NV’ only contains a single known element, say, the zero element. On the other hand, partialness of information can disappear in two different ways: by a priori assumptions or 4 posteriori measurements. In the first case, partialness disappears when the set S tends to have only one element (the real system); in the second case, when the amount of experimental information is completed by the remaining (usually infinite) data points It is convenient to unify the above concepts as follows. The available information (a priori and a posteriori) is completed when the consistency set Sty) tends to only one element: the real system. When this happens there is ‘no uncertainty that is, the input information is complete.* Hence converger can be defined as follows: Definition 10.7. An identification algorithm A converges if and only if ii nea) =0 (1031) The diameter of information tends to zero when the noise set A’ reduces toa single element (usually zero noise) and either the set S tends to the real plant or the experiment operator provides all necessary data. Note that as the consistency set S(y) reduces to a single element, the experiment operator tends to be invertible. Since the identification error is defined in a worst-case sense, its convergence is uniform with respect to the a priori sets N’ and S. The definition of convergence that has been presented reflects the in: formation structure of the problem and generalizes the ones that will be stated in the following sections, When considering convergence in terms of completing the a posteriori experimental information, care must be taken because infinite data points (time or frequency response experiments) do rot in general guarantee complete information ((236)). Take, for example, a frequency response experiment with data only at integer frequencies {o, = k, k=0,...,M~—1). By making M —+ 00 the @ posteriori information will not determine the unique real system since D(Z) does not vanish (see Deft- nition 10-7). that is, information is still incomplete. For this reason, all neces: sary (infinite) data points to complete the information should be considered. For example. it takes an infinitely countable number of points in the time response or an infinitely uncountable number of points for the frequency response of a discrete IIR time system, According to Lemma 1044, all interpolation algorithms are convergent. “Here we aesome that the tea plant ways Relongs tthe consistency set. Aocording tothe comments the end of Section 10.22. th wil evenly he the cise a= kong. 38 the et SY) 336 ROBUST IDENTIFICATION Morcover, since these procedures generate a nominal model in the consis. tency set, they are always tuned to the a priori information. In later sections, convergent noninterpolating algorithms, which may be untuned, will also be sought. Convergent untuned algorithms are said to be robusily convergent, This is an important property because algorithmic convergence is dissociated from a priori assumptions. Therefore convergence is guaranteed, even if the a priori information is false, although the resulting bounds are no longer valid, This has a practical implication because most of the a priori information is bbased on a “leap of faith” or a combination of engineering “intuition” and ‘As mentioned before, model approximation can be considered as a special case of identification. In that case, the sets $ and N already have single known, clements: the real model and zero noise, respectively. Uncertainty is only due to the partialness of the a posteriori data. In model approximation the objective is not to exactly invert the experiment but to produce a lower-order ‘model (due to computational tractability). Convergence is therefore defined in the limit as the number of data points tends to be completed (M —+ 0c), rather than with the diameter of information going to zero. Convergence and a priori class S The worst-case error defined by equation (10.4) depends on the classes of models and measurement noise. In turn, the convergence of this error to zero also depends on these classes, ‘Thus there are necessary conditions that the sets S and A’ should meet in order to guarantee the etistence of convergent algorithms, Next, we present a condition on S that precludes the existence of these types of algorithms (1194). Lemma 10.5 _[f the a priori class 5, which belongs to metric space (Vym), con- ‘ains a ball of size y, say, By(go) = {g © V, m(g,go) < y} CS, then R(T) > y {for any number of measurements and any noise level. Proof. The experimental evaluation of the plant g produces y = E(go,n) = [9b + nf BRT Wl") for m €.N’. We can always find a model g. such that gf = gh, k =0,...,M—1 and m(go,g.) = 7, that is, go € By(@0) ‘Therefore independently of the value of M, we have two indistinguishable elements gp and g, inside the ball. Thus these elements should be included in the consistency set. Since m(go,g.) ~ y then r[S(y)] > a Several comments concerning the above result are presented next. «Ifa set is defined only in terms of the distance among its elements, there is no relation between the number of measurements M and the “size” Of the set. Therefore the @ priori class should not contain this type of set if convergence is desired. ‘* The element g, described above belongs to the kernel of the operator [E(,-m) ~ y). Therefore the “size” of this kernel should decrease to 2er0 GENERALSETUP 337 as M — oo and 1 ~ 0. This is related to the inversion of the experiment operator £(-,-) when we have complete information, as mentioned in Section 102.1 « In [194] and [229] specific plants «, are provided for the metrics defined by the f; and %4,, norms. + From Lemma 105 it follows that S should not have interior points if we expect to have convergent algorithms. There are yet two other elements that should be limited by the a priori information: the intersample behavior and the “tail” of the experiment, that is. (94, & > M). Both are a consequence ofthe finite numberof experimental data points obtained from systems with infinite time oF frequency responses. To this end define the following: Definition 10.8 Consider a metric space (Vm) and a family of functions Az) —+V. The family A is equicontinuous if and only if for any ¢ >0, there exits 6 > 0 such that if x,y € Y, m(x,y) < 8 implies miUf(x), f0)] < ¢ for all fea The above is a generalization of the uniform continuity of a function, extended to a whole family A. Such a family has a uniform bound over the intersample behavior of each of its functions, An equivalent characterization of bounded equicontinuous sets is given by the Arzela-Ascoli theorem ((201)), ‘Theorem 10.1 The following statements are equivalent (a) The family A is equicontinuous and uniformly bounded. (b) The closure A is compact. ‘Therefore selecting compact sets for the a priori class of models will limit the intersample behavior. To limit the “tail” of the experimental output, which is unknown due to the finite number of samples, additional restrictions should be added to the a Priori set of models. When considering the last portion of the time response, this condition is related to the stability of the class of plants. On the other hand, when considering the frequency response the “tail” of the experiment is restricted by the “properness” of the models. To recap, in order to have convergent algorithms, we need the a priori class of models 5 to have the following properties: 1. It should not contain interior points so that it converges as M —+ oo. 2. It should be an equicontinuous set to limit the intersample behavior. It should also be uniformly bounded in order to have a bounded consis: tency set, therefore it should be compact. 3. It should inelude only stable and proper models so that the (unknown) tail portions of the (time and/or frequency) experiments are bounded, 338 ROBUSTIDENTIICATION For the sets considered above it can be proved that pointwise convergence is sufficient for uniform (worst-case) convergence ((19). Lemma 106 For a set uf M measurements and noise level €, any untuned denufication algorithm A satisfying im tim Els, m))~ gh =0 (10.32) for all g € V, also sauisfies (1033) lim limsup AEC.) ~ for any relaively compact set? 5 CY. Proof. See {196 c 10.2.5 Validation ‘The main objective of robust identification is 10 compute a set of models that can be used in a robust control setup. A general framework to deseribe robust control problems uses LFTs, as described in the initial chapters of this hook fn this setup, there is a block that represents the nominal model (including robustness and performance weights), a controller block, a model uncertainty one, uncertain disturbance inputs, and error outputs, Both the ‘model uncertainty block and the uncertain disturbance inputs are usually defined as bounded sets, ‘Similarly to the consistency test that verifies that a priori and a posteriori information are coherent, another procedure should establish whether of not the outcome of the identification procedure is coherent with the setup in which it will be used by being able to “explain” all available experimental information. Therefore, if a set of models is to be used in the context of a Robust Control framework, it seems reasonable to verily the existence of an element of the model uncertainty block and an uncertain input signal (both included in their respective sets) that reproduces the experimental data. This is a reasonable test irrespective of how the set of models has been obtained. This same procedure is used in general to test an empirical theory. In a scientific framework, an empirical theory must be capable of being falsified ((245)). Similarly, an empirical model® should be tested against all available experimental data. If positive, this provides a verification that the model is rnot false. Obviously this test does not prove that the model is completely valid, because a new set of experimental data may invalidate it. For this Teason, a model or set of models cannot be validated, only invalidated. The 5 sets sid to Be relatively compact itt closure compact "11 not necessary hat the mode! computation sned the sorementioned experimental date ow the mrael as heen obaned st relevant st this pnt FREQUENCY DOMAIN IDENTIFICATION 339 procedure called invulidation tests the model or set of models against all available experimental information. It provides a sufficient condition for the model to be invalid, if particular experiment does not fit in the framework. Otherwise it only provides a necessary condition for the model to be valid when all experiments can be explained in this setup. For different reasons, the problem of invalidation has usually been called the validation problem. It is not the purpose of this section to justify the use of this terminology (which can be explained probably better by psychological reasons) but only to introduce this concept, which will not be developed further. Nevertheless, care must be taken when using the word validation in this context. Ifthe analysis framework is identical to the one used for the experiment, a set of models computed using a robust identification algorithm will be always validated by the experimental data used to compute it. Therefore the only two reasons by which a robustly identified set of models may be invalidated © A different structural setup + A new set of experimental data. ‘The usual framework in which the robust idemtification experiments are performed is coherent with the use of the LFTs as a general control frame: ‘work. The consistency among both structures should include the norms used to bound disturbance inputs, model uncertainty, and output errors as well In practice, the invalidation problem involves computing the minimum norm bound on the disturbances and mode! uncertainty to validate the ex- periments. Its clear that with large enough disturbances or uncertainty, any experiment can be validated, Hence this minimization problem also measures the conservativeness of the set of models. ‘The problem of validation has been initiated in {292}, for models given in a LFT form, using the 1, norm to bound uncertainty and disturbances. The problem has a solution in terms of the structured singular value and includes structured dynamic model uncertainty in the model description. The statement of the problem is given entirely in the frequency domain, For a practical application see (291)) ‘The use of a time-domain description and bounds in the & norm have been proposed in [242]. The solution involves a convex matrix optimization problem based on the experimental data and the proposed set of models. 10.3. FREQUENCY DOMAIN IDENTIFICATION In this section we present Robust Identification procedures for a priori classes oof models in 71. and experiments performed in the frequency domain. The scheme follows the general framework presented in the last section, although specific issues pertaining to this metric will be stressed. 340. ROBUSTIDENTIFICATION 10.3.1 Preliminaries We consider causal, linear time invariant, stable either continuous or dscrote time models. For simplicity we consider only SISO systems, although all results thoughout this setion can be applied to MIMO systems aswell, either clement by element or asin 76. To unify the treatment we denote the models as H(z) = HAS), A > O in the continuous time ease or H(z) ~ Hall/2), for discrete time systems, with z € C. Note thatthe latter (known a5 the transform in the bibliography (87) is the inverse of the usual z transform. Therefore in the Sequel, causal stable models H(2) wil be analytic inside the unit disk, In this context the idemtification problem reduces 0 that of computing a model, analytic in a certain region, based on partial and conrupt data measured atthe boundary of ths region In order to define the a posterior information, we introduce the following sets Mas {io analyte ll | and finite K, the following holds: Moolp, K) C Hoop © A(D) C Hoe (10.39) Note that %a.(9,K) isa strict subset of A(D) because there are elements in the subalgebra that do not belong to (p,K) (give an example) The a priori class of models is defined as the set S = Hy.(p, K). This cor- responds to exponentially stable systems, not necessarily finite dimensional FREQUENCY DOMAIN IDENTIFICATION 341 From a practical viewpoint, the models defined by the @ prior’ information have @ stability margin of (p ~ 1) and a peak response to complex exponential inputs of K. The following bound is satisfied in the time domain (prove) [mi] < Kp (1040) The a posteriori experimental information is the set of M samples of the frequency response of the system, measured with additive bounded noise: Y= HC) +n, k=, MAI cost) which represent the components of vector y € C™, The measurement noise is bounded by in the f, norm, To recap, the a priori and a posteriori input data are, respectively, (He HK), p>1, Ko) (108 Wa (ne imbce k=OouM 1} 088 y = EH.) =Hlel™) +m, k=0, M (10.44) Since the a priori class S is relatively compact, Lemma 10.6 applies. Also note that both sets, Sand A’, are symmetric with centers cs and cy in the zero model and noise vector, respectively. Based on the general statement of the last section, the global error and convergence criteria for algorithm A are as follows: (A) = sup AE(H.n),5,M]—Hln (1045) lim eA) = 0 (10.46) According to Lemma 106 if A is untuned, that is, Aly). the convergence criteria need only be applied to the error corresponding to plant H: |.A(y) — Hljx instead of to the above uniform error. Using the results of Lemmas 10.2 and 103, a simple lower bound for the information radius can be found. Lemma 108 The radius of information cannot be smaller than min(e, Proof. Define C= min(e, K) and consider the constant models Hy(2) =Ha(2) = C and noises 7) =~» = —C, which belong to the a priori classes S and N, respectively. Models H, and Hy also belong to the consistency set ‘S(y) for the null experiment yp = 0, that is, Hy 4 1) = Hz + m = 0. Hence. 342 ROBUST IDENTIFICATION from Lemmas 10.2 and 103. RID) 2 |DIL) = Lals(oy) min(e.K) (10.47) o Again, this bound is global and holds for all possible experiments. A feature of Hx. identification is that it assumes that the experimental data is in the frequency domain, Real world experiments are always per- formed in the time domain. For this reason, the time-domain data needs to be pre-processed in order to obtain the required complex frequency-domain 4 posteriori information. In the next subsection, two convergent algorithms, which compute the data vector y and the equivalent frequency measurement error bound e, are presented. 10.3.2 Sampling Procedure To obtain the a posterior’ information, an experiment should be performed to compute the frequency response of the system and the data vector y € C¥ For practical reasons. this experiment involves bounded sequences of inputs ‘and outputs measured in the time domain, The usual procedure is to input, sinusoidal signals and measure the phase and magnitude of the output in relation to the input. The a priori information indicates the time at which the transient response is irrelevant and only the steady-state values are significant. Additional practical considerations involve verifying the a prior’ assumptions. (272), that is, stability, linearity, and time invariance, The computation of the a posteriori data points can be performed using the same framework as, the model identification itself, as shown next Consider the system initially at rest, a sequence of sinusoidal input signals bounded by a and measurement noise bounded in magnitude by y. Inputs. The complex sinusoidal sequence ono c . 4 w= {6 Ben (1048) for the set of frequencies (4: k= 0,...M=1} COupus. The measured vector x € CY. with components ee ee ye foals, N coe) ‘Two possible algorithms ((146}) that identify the values yy in (1041) are the following FREQUENCY DOMAIN IDENTIRCAION 343, Algorithms. The data points at each frequency are computed by algo: rithms P, or P, as follows: Pek =m = 1050) LS oe Prk =e = oy Yael (ost) Errors, The above values have the fllowing error bounds Method 1: Im| < MBp 42 2e (1052) Kelp") ya Method 2: Im) < M226 oss ints KeGme), 2 (1083) forall k= 0,...,M-1 a8 shown in (146), Note that v isthe time domain measurement noise bounded by from which a bound € over the complex frequency-domain noise m,k =0,...,M ~ 1, is calculated, For the real system the inputs are replaced by sinusoids, although the bounds and outputs are similar. For simplicity the frequency points are equally spaced , = 2/M_ Note that the first term of both error bounds is related to the transient response, which vanishes as time increases (N — 00). The second term, on the other hand, is constant and represents a signal-to-noise ratio that can only be decreased with a better measurement procedure, that is, decreasing, y ‘The above frequency point identification procedures can be recast in the same framework as model identification. The a priori information are the noise bound y and the values (p, K) of the plant set Hax(p, K). The a pos: teriori data are the time measurements x. The error is defined in a worst-case sense over the set of measurement noises v, defined above and the class of models H.u(p, K). AS in the general case, there is also a lower bound to the identification error, which can be specified in terms of the radius of infor- ‘mation (see Section 10.2.3). Specifically for all frequency point identification procedures, the following result holds Lemma 109 The worst-case error for any frequency point identfcaion al gorithm P is always greater than min(K, ¥/a. Proof. The proof is similar to the one in Lemma 10.8 by defining the con ‘models as + min(K, y/a); therefore it will not be repeated here, 3 344 ROBUST IDENTIFICATION I is clear from the bounds of Methods | and 2 that both are convergent In fact, since they are untuned with respect to the a prior’ information, they are also robustly convergent im tim ec) = 0 (05a) Furthermore, according to the above lemma, both algorithms converge when K > y/a, to the (optimal) lower bound y/a as N —+ 00. Therefore these pro- cedures are robustly convergent and asymptotically optimal. We can always assume that K > y/a, since otherwise the input signal magnitude a can be scaled such that this condition is guaranteed. For practical reasons, itis clear that the frequencies cannot be gen- crated exactly. Nevertheless, the frequency error 5, can be included in the bounds as well ((229)) 10.3.3. Consistency Recall from Section 10.2 that consistency of a priori and a posteriori in- formation establishes the validity of the identification error bounds. As a consequence, the covergence of tuned algorithms depends on the solution to this problem. Although untuned identification procedures converge even for inconsistent input data, the use of the set of models in @ robust control framework is based on the validity of these bounds and hence also depends ‘on the consistency test. Given the experimental data {yx = H(e) + ng k=0,...4M —1} and the a priori sets S and N defined before, the consistency set S(y) can be specified using (10.2). The consistency problem can be rephrased as the ex- istence of a model H € S whose difference with y atthe sampling points can be attributed solely to measurement noise in A, that is, lite ye] Se for all k=0,...,M—1 (1055) We refer to this problem as an approximate interpolation one because /7(e!) matches y, up to an error of size €. For 7 identification, the consistency problem can be cast as an approximate interpolation problem and solved via the Nevanlinna-Pick procedure ({74]). A brief statement of this latter result follows ((21)) - Consider M values w, measured in the interior of the unit disk Dat points zy. There exists a function f € BM, that satisfies f(z) = my for all k=0,...,M-~1 if and only ifthe Pick matrix P, 1 wy Py = a” rat i,j =0,...M— (1056) FREQUENCY DOMAIN IDENTIFICATION 345 is powitive semidefinite. If P > 0 then there exist infinite interpolation fune: tions f, which can be described as a linear fractional function of a free parameter in BH... These solutions will be described in further detail in ‘the context of interpolation algorithms (see Section 10.3.4 and Problem 6 5). This result can be extended to Huu(p,) using the fact that f(z) € BR, + Kf(c/p) € Mou(p,K) (verify). Hence the rescaled set of Pick matrices can be defined as follows: 1 Dyn LR G+ 1 La) 1,7 =0,...M— (1037) for any vector of measurement noise 9 € NV. Therefore an equivalent char- acterization of the consistency problem can be stated as follows Lemma 10.10 A set of experimental data y ¢ ©™ is consistent with the a priori sets S and N’ defined in terms of constants (K, p,e) if and only if there exists a vector n €N such that the Pick maurix in (10.57) is positive definite! Proof. Follows from the fact that f(z) € BH, => Kflz/p) € Sand the classical Nevanlinna-Pick theory. 0 From the above lemma it follows that a necessary condition for consistency is that all experimental samples remain below K + €. This is based on only ppart of the a prior’ information and stems from the fact that for all diagonal clements in P to be positive, itis necessary to have Wet ml lal —Iml Inl< Kee (10.58) for all =0,...,M ~1 and 9 €.M/, where a signal-to-noise ratio greater than 1 has been assumed, that i, y4| > /n| Clearly the above isan intuitive upper bound for the experimental data based only on the magnitude of the signal and the measurement noise Finally, to test consistency, at least one noise vector 7. € N’ should be found. A’computational practical solution to this problem can be obtained by solving a minimization problem ([74): Lemma 10.11 The following statements are equivalent: 1. The a posteriori data y € © and the a priori sets (S\N) defined by (K,p,€) are consistent. The pongeneric case of P can be found in = 0, which has a unique solution, will not be considered here hu 346 RORUST IDENTIFICATION 2 ring. MR) <0 (1059) o: 4 XC) : Fe lavaxem “0 aes where Y= diap[ye --- yas fe X= dal ms mes] and 04, = [1~ th jp?) 3. The linear matric inequality (LMI) problem RIXmIl > 0 1061) tex) ne [xin “?] >0 (10482) thas a feasible solution for a vector 9€ CM 4 iminyea {QY2[¥ + X(N] OF} < K 1063) Proof. Alter some straightforward algebra, the Pick matrix can be written pag reer +X)" (1064) K Using the Schur complements, P > 0 <=» R > Oand the equivalence of state ments 1 and 2 follows. Matrix R(X) is linear in X, which in turn is linear in n, The positiveness condition in equation (10.62) is equivalent to |7)| < €, i =0,...,M 1, Therefore the LMI feasibility problem 3 is equivalent to consistency. Finally the equivalence between statements 4 and 1 is derived in (74 0 Statement 2 is a standard generalized eigenvalue minimization problem that can be solved via the methods developed in [214] or by semidefinite programming ({308)). Problem 3 is a standard LMI and can be solved using available commercial software ({126)). Finally, convex programming can solve problem 4, ‘An advantage of testing consistency through an approximate interpolation problem is that, from the parametrization of all interpolating functions, an interpolatory identification algorithm can be obtained dircctly. This will be detailed in the next subsection, 10.3.4. Identification Procedures ‘Once consistency has been established, the computation of a nominal model and a valid model error bound can be attempted. In this section two different FREQUENCY DOMAIN IDENTIFICATION 347 types of algorithms will be presented, as well as their corresponding worst: case model error bounds. The frst type of procedure, initiated in [144] and extended to H.. identifi cation in [137], consists of two steps, the second of which is an optimal Nehari approximation. In this approach most of the research effort has been concen: trated on the first step, aimed towards improving the algorithms perform- ance. These procedures can be untuned to the a priori information: therefore they can be robustly convergent when the worst-case error vanishes as the input information is completed. This means that even if the model error bounds are invalid (if a priori and a posteriori information are inconsistent), the convergence of the algorithm is guaranteed. On the other hand, the nom- inal model generated by these two-step procedures is not guaranteed to be in the consistency set, which in principle may lead to a larger model error bound. ‘The second type of algorithm presented in this section is interpolatory, hence the nominal model is always in the consistency set. From the results presented in Section 10.2, it follows that the algorithm is always tuned and ‘convergent, Furthermore, this procedure is also strongly optimal, in terms of the identification error, to within a factor of 2. This interpolatory algorithm is based on the consistency test performed in Section 10.3.3 and can be solved tusing the parametrization of all solutions of the same Nevanlinna-Pick prob- lem. This interpolatory algorithm was first presented in [74]. Another type of interpolation algorithm has been reported in [233, 236), but will not be presented here. As a comparison, from a computational point of view, the first step of the two-stage algorithms can be solved via a FFT for equidistant experimental data and may take longer in the case of nonuniformly spaced data ([8, 127)). The second step (Nehari approximation) has a computational cost similar to the Nevanlinna-Pick problem. Finally, note that the above identification algorithms produce a constant ‘uncertainty bound as a function of frequency. Therefore the description of the set of models does not admit any frequency-dependent uncertainty weight, In the case of model approximation since the “real” system is available, the difference with the nominal model can be computed explicitly as a function of frequency. The uncertainty weight can be computed much in the same way as the nominal model. In this process the order of both the nominal model and the uncertainty weight that constitute the augmented plant will be the same as that of the nominal model itself. This will not be explained here but is detailed in [127 In the sequel we illustrate the identification algorithms through the fol- owing example Example 10.1 The physical system under consideration is a Euler-Bernoulli beam with viscous damping, which can be described using the following phys: ‘cal model that relates the vertical displacement y to time t and the longitudinal 348. ROBUSTIDENTIFICATION coordinate x: Oy ep LOY, MY wad 1 TY ocr Zh 0, ve (Ly (1065) with boundary conditions ay Bane Facey =0; aan=0 (10.66) Here ais the cross-sectional area, p the mass density of the beam, I the moment of inertia, E the Young modulus of elasticity, and E* the normal sirain rate The values adopted in this case are: ap = 46, E*1 = 0.46, and El = $5.2, and the dynamics are evaluated at x = 0. Since the PDE (10.65) is linear, by using the Laplace transform we obtain ‘he following (infinte-dimensional) transfer function: ao WCET+ ED Hs) = (sim sinh A ~ 008A cosh A ~ 1) cos Ax SinbAx~ sin Ax + "cos Asin A + sin A cosh & (cin Asinh A + cos A cosh A + 1) cosh Ax ‘cos A sinh A + sin A cosh A oe (10.68) Due to the rigid mode of the structure, this transfer function has an unstable part 1/2ps. Although in practice itis nox possible to physically separate the tuistable and stable responses, the purpose ofthis example is only 10 test the different idenification methods over the stable portion. Thus in the sequel we assume exact knowledge of the unstable portion of the model, and we subtract its response from the experimental data inorder 10 identify only the sable part At the end of this chapter some comments on identification of unstable systems in closed-loop experiments are presented. In this case, we have generated three “experimental” data sets of 180 points each from the frequency response of the flexible structure using the software package (173). Figure 102 shows the total magnitude response of the sytem (stablesunstable), the stable part, and three sets of simulated “measurement” noise in f(€) with € =8 x 10" The three sets of “experimental” measurements can be combined to pro- dluce an equivalent set of data points and a new measurement error bound. This can be achieved by considering each data point as the center of a ball of radius «, in C for each experiment i = 1,2,3. At each frequency, he center and radius of a new ball should be computed such tha t “covers” the inter FREQUENCY DOMAIN IDENTIFICATION 349 otal system x10) Stable part Nose scrote requency Figure 10.2. (0) Total froquoncy rosponse, (b) Stable frequency response. (c) Thee ses of “oxparimental” measurement noise. section of all the balls corresponding to the experiments to be combined. This produces an equivalent set of data points and a frequency-dependent error Dowd that includes all previous experimental data. This frequency-dependent noise bound should be smaller than the individual bounds for each exper ‘ment and could provide a less conservative description of the noise, that i, €(1@) < miner, 4 6) In this particular example, the equivalent set of measurements has been “superimposed on the thre inital datasets in the first llustration of Figure 10.3 The frequency-dependent error bound of this new measurement set appears in the second part of Figure 103. For simplicity throughout this chapter we will use the constant upper bound € = 8 x 10°*. Two-Stage Algorithms A first intuitive algorithmic approach to obtain a model from the a posteriori measurements would use the discrete Fourier transform (DFT) and its inverse (IDET), as follows: rt = 5 ee, ae 0 360 ROBUST IDENTIFICATION x10 7| es IL — inversion ql Experments | 5 ee ai 2| o 05 1 7s 2 25 3 x10 a ° 05 1 15 2 25 3 scree frequency Figure 10:3. (2) Superpostion of three measurement sats ond their combination. (©) Equivalent measurement noise ot each Frequency, Hala) = > halkyz* (10.70) Here hy, corresponds to an “equivalent” FIR impulse response obtained from the experimental samples y and H, is the identified model. Model Hy,(2) interpolates the experimental data at M frequencies in the unit circle. that 4 = Hy (@), k=0,...,M — 1. If the real system has a finite impulse response (FIR) with less than or equal to M nonzero samples and in the absence of measurement noise, then H,,(2) coincides with the real system and only the M values H,y(e), k = 0,...,M ~ 1, are necessary to represent it If, instead, the real system has an infinite impulse response (IR) and/or the samples are affected by noise, the following two problems appear: ‘© The equivalent impulse response ht, suffers from sine aliasing. Thus the ‘identified model H(z) could be unstable, even if the real plant is stable This becomes clear by shifting the sum in (10.70) by (M ~ 1)/2 samples* Assume without los of penerahy thal A iso FREQUENCY DOMAIN IDENTIFICATION 381 0 compute the (periodic) function H(z) as follows Ha) = YI hatkyat (071) aya Hence if the (periodic) function hy1(k) # 0 for —(M ~ 1)/2< k <0 (the same as for (M + 1)/2 < & 0, 26 Mii Mee 2 Flog) +01) (10.76) AS a consequence, the worst-case interpolating function diverges as the number of samples increases. ‘The above problems motivate the 1wo-siage algorithms ({146, 137]) 1. Using the IDFT, an equivalent impulse response is obtained from the ‘experimental data y. This response, suitably weighted so that the con vergence is uniform as M — oo and ¢ —+0, is used to compute a model H(z). This part of the procedure is linear in the a posteriori data Yes k=0,.-.5M—1 2, The model H(z) may have an unstable part, although the original plant was stable, Therefore a projection of the model over the stable set 7 should be performed. ‘The details of each step are described next. ‘Stage J, The input is the set of @ posteriori information {yn, Ory M1), from which hy is computed 1S ost ath) = By me, 077) ‘The proidentfied model /1, is calculated using the following weighted Hult) = 2 w(k dha (De (10.78) FREQUENCY DOMAIN IDENIICATION 353. Here the weight values w(k) ate suitably selected to have the required convergence characteristics. Because of its reminiscence to signal pro- cessing procedures, itis often called “window” w. By replacing hy, in the last equation it is clear that this part of the algorithm is linear in the input data y, Hate) © tly) = ye So eKulze™) (10279) where Kwlz)~ whet (1080) whee is the kernel ofthe interpolation operator fy : C™(Fx.) —+ La (the input data are equipped with the f, norm). The design of the algorithm depends on the selection of the weight w or, equivalently, the Kernel Ky or the interpolation operator fy. Convergence is guaranteed for certain classes of weights w or operators fy. This will be discussed at the end of this section in further detail Stage 2. The optimal projection from Hy; € Ca, to a stable causal model Hig € Hoy ean be found by solving: War ~ Hille (1081) ‘This is precisely the Nehari problem solved in Section 6.6 (Theorem 655), The second step differentiates this identification procedure from the standard fitering problem, which arises in digital signal processing, where the noncausality of the filter can be solved via a delay. ‘The projection from £._ to Ha, of H.: should be performed only on its antistable (anticausal) part. To see this, consider H,y(2) in (10.78), which can be separated into its stable and antistable? parts H.a(z) = Hj(2)+ Ha(2). Is clear that the optimal projection only depends on H(z) because the stable partis already in 7,. and can be subtracted. AS in Section 6.6, a solution to the Nehari problem can be calculated in terms of the singular value decomposition of the Hankel matrix for the antistable part H(z) as follows FER dy, TE that Hidl2) = Hale) (10.82) * Recall thatthe anisole part hs its ples inside the unit circle inthis ese at = 354 ROBUST IDENTIFICATION where d is the maximum singular value of the Hankel matrix of H,(2), (6,2) its corresponding right and left singular vectors, and N= (M 1)/2. This optimal approximation is rational and has degree less than ‘or equal to I. Nevertheless, care must be taken when using the above equation, due to the fact that the second term should cancel exactly the antistable part of H,i(z). More efficient numerical solutions pre sented in state-space form can be found in [21] and [332] for both the continuous and discrete time cases, The above procedures are untuned, since the only input data are the a pos terior’ information. The second stage is nonlinear in y, as opposed to the first stage, which is linear; hence the complete algorithm is nonlinear in the input data. In fact, it can be proved that there are no robustly convergent Ma. identification procedures that map the input data y linearly to the nominal ‘model ([234)) When solving a model approximation problem ({127)), the second step is not necessary if the number of samples is high enough, so that the val- tues of {i(k}, k= =I,...,-(M —1)/2} are negligible. This is due to the fact that there is no measurement noise. In that case only the causal values {hy(k), k=0,...,(M ~1)/2} need to be kept to compute the approximate ‘model, which corresponds to a rectangular unitary one-sided window. There- fore the whole procedure is linear in the input data, Moreover, it can be shown that itis robustly convergent ((139, 321)) In a similar way, in many practical identification applications when mea- surement noise is small enough only the first step of the identification algo- rithm needs to be applied ({9, 138]). In this later case, the divergence of the identification error due to measurement noise is Ole log( Convergence and Error Bounds 11 is possible to consider the identifica- tion errors at each stage in order to evaluate their convergence separately. It can be proved that the convergence of the whole procedure depends solely fon the convergence of the first stage. Due to the optimality of the Nehari solution to problem (10.81) we have: Wy ~ Hall < Waa ~ HL (10.83) where H € S is the system to be identified. Using the triangle inequality and the above equation, the total error can be bounded as follows: WH = Hull < Hos — Hil + WHat ~ Meal, 2 Hull, (1084) ‘Therefore the whole error converges if the identification error of the (linear) first step does, as M — 2c and ¢ — 0. Furthermore, the total worst-case error is bounded by twice the first-stage error. This indicates that the design of the algorithm and its convergence depend only on the selection of a convenient window function w for the first stage FREQUENCY DOMAIN IDENTIFICATION 355 Necessary and sufficient conditions for stage 1 to be convergent are pre sented next ({137). Lemma 10.12 The first step of the wo-stage identification algorithm is ro- busily convergent if and only if the following two conditions are satisfied: im sup pL = (1085) fim sup pill. = (10.86) where the preidentified model of step 1 (see (10.79)) has been separated into Hy, = Hps + ty, using the fact that the noise is additive. Proof. Both conditions are necessary by selecting 9 = 0 € N and H =0€ S, respectively. Sufficiency can be proved by applying the triangle inequality to Hy = Hpi + ni a8 follows sup Wa ~ Hil $ sup |p ~ Hl. + sup ll The convergence of these two conditions guarantees the convergence of stage 1. u Note that the above two conditions are met by any plant H € $ mea sured without noise and by the null plant H'=0 being robustly identified, respectively, Tn the case of model approximation, only condition (10.85) needs to hold for convergence, where Hy = Hy = Hl. Therefore the worst-case ident- fication noise-free error supycs [Hpi — H||,, i8 called approximation error. Similarly, sup yey [il is termed the nove esto, The uncertainty bound of the identincation algorithms and their convergence will depend on both these errs “There are many diferent windows that may guarantee convergence of the fst stage, and hence ofthe algorithm. For example, the triangular window, Known in digital signal processing asthe Bartlett window, given by fs ikl i087) { 0 KEN where N has been defined in stage 2. In this case, the sum in (10.78) cor- responds to the average of partial sums of the f,i(k), also known as their Cesaro sum, which guarantees uniform convergence ((139, 321) as N > oo. Iv can be proved that this type of window produces a noise error bound of wk 356 ROBUST IDENTIFICATION «Hence it is optimal, because this is the lowest possible uncertainty bound, when K > € (see Lemma 10.8). Another rapidly convergent procedure uses the De La Vallée Poussin sums ((137, 233) In (137), necessary and sufficient conditions are given, guaranteeing that 4 particular window will generate a (robust) convergent identification al gorithm. These conditions therefore define a class of two-stage algorithms that includes, as special cases, all other two-stage algorithms cited in the bibliography. For cach different algorithm, there is a compromise between the approximation and the noise identification errors. From the many possi- ble convergent windows that may be generated, a subset that quantifies this ‘compromise can be parametrized as a function of an integer q as follows: ho naka nae asksn m4 w(k) (10.88) nak nq 0, otherwise with 0.< q 0) can be found in (21, 74) The algorithm depends on the noise sequence n, computed in Section 10.33. to establish consistency. Recall that if the a priori and a posteriori information are consistent, a noise vector 7. € A’ must be available such that the conditions in Lemma 10.11 hold. This is important not only to initiate the procedure. but also to guarantee that the idemification error bounds are valid, since the algorithm is tuned to the @ priori information, The algorithm proceeds as follows: ‘* Compute the Pick matrix P in (10.57) (which should be positive definite. replacing 7 by ‘# Define z, 2 et and wy 2 H(e!") for k= 0,....M ~ 1 and compute Ti) = Rep FREQUENCY DOMAINIDENTICATION 359 where po tw Ke= tu) ‘© The interpolating function is Ty(2)al2) + Trl) Hyle) = KB (1090) Tal 3)ale) + Taal) where g(z) € Bw isa free parameter In particular, if g(2) is constant, then Hj4(z) is of order M. ‘The degrees of freedom available suggest a further optimization step to select the optimal 7. and q(2), for instance to minimize certain criteria, The computational size of the problem corresponds to the number of experimen- tal data points As for the identification error, in (74] the following bound has been deter- ‘mined for the diameter of information: esky PO)

2minte, Ke) (2099) where Ky = (0). Proof. Consider the models h = mine, Ko), consistent withthe null ouput Equation (10.99) follows immediately from the general results in Section 102 a ‘The above bound is independent of the duration of the experiment and the nature of the input sequence; therefore it represents an absolute lower limit on the error. Next, we obtain bounds for the impulse response experiment that depend on the number of samples taken (230) Lemma 10.14 The information diameter for the &, identification procedure with an impulse response experiment is given by: Dy(Z) =2[ nine #00) + $290] (10.100) Proof. Define the following: fay = {Bh ESA nt = me) Next, consider the pair of models and measurement noise vectors: {mer= ay and {rns =to nk) = —n{k) math) = n(k) = hymeS, mmen Hence the above produces zer0 outputs to an impulse excitation, that is, Ehy.m) = Ella, m2) = 0. Note that this isthe worst-case model (maximum f, norm) compatible with the experimental output and in the consistency set S(0). Equation (10.100) follows now from Lemma 103 and the fact that $ and N’ are symmetric and convex by considering d(S(0)] = Wy ~ fal, = 2. (A) and « being positive a The following comments are in order: # There is a significant quantitative difference between the bound in (10.100) and the global one in (10,99), the latter being a conservative lower bound for the identification error. TIMEDOMAIN IDENTIFICATION 365 + The diameter of information tends to zero as M — 90 and € -» 0. This ‘guarantees the existence of convergent algorithms, for example, inter- polation procedures. ‘© For a given noise bound level, there exists 4 maximum number of ex: perimental samples M,, which decreases the diameter of information. In other words, for M >'M,, the diameter remains constant, which means ‘hat any extra experimental sample will not increase the amount of useful information. This fact will be used in the formulation of an interpolatory convergent algorithm ((230)) in the next subsection, Under certain conditions ((163}) it can be proved that the impulse input sequence is optimal in terms of the diameter of information, for a given experiment duration M. These conditions are related to the fact that the 4@ priori information does not exclude models in the consistency set of the null experimental outcome. The optimality stems from the fact that this in- put produces outputs directly related to each time coefficient. Therefore all the uncertainties are decoupled, that is, each measurement noise element is related to only one model coefficient. For example, for the step response experiment, the reader may verify that the diameter of information satisfies Dy (Z) ~ minle, (0)]+2 > minle,@]+ S24) (10.101) ‘The proof is based on the same arguments as before, but using the following model: min{(0),€), k=0 i(k) = } 2(—1) minf(K),€], 1k 2. Since AC) is based only on the experimental information, it cannot distinguish between both plants, but should converge for all plants in S(O), Therefore if sy lim, AQ) ~ hol = 0 366 ROBUST IDENTICATION = Mo, such that [AQ <1, M > Moy € <6 Asa consequence for M > My and € < y, WAG) — Ath > Wall ~ AQIIh 2 1 (10.108) which contradicts the assumption that A converges to a plant in S(0) (see also (160) a In [237] the following necessary and sufficient condition for the existence of a robustly convergent ¢; identification algorithm is established: 36 >O-such that lh «ull > Sli, Yh Eb (10.104) ‘here lulas < 1 without loss of generality. The above condition excludes non persistent inputs for which u(t) — 0 as f — oo. Specific untuned algorithms, based on experiments performed with inputs known as Galois sequences, have been proposed in (193). Furthermore, in [238 it is proved that there are no linear robust convergent algorithms in f, as was the case for Ha, identification. 10.4.2 Identification Procedures Next, several tuned convergent algorithms for f, robust identification are presented. These are based on the impulse response experiment although, with minor changes, can be applied to the step response experiment as well. ‘The fact that these algorithms produce a nominal model in the consistency set makes it unnecessary to test consistency, as long as such a model exists. ‘Untuned convergent algorithms for additional input signals, can be found in [141, 193, 194, 306} Based on the Firstn ine, (K)] (10.106) Proof. This proof follows a similar one in [146]. For every h € S,n € N and for any algorithm A, the following holds: We AQ) < eA) (10.107) where e(A) Is the worst-case local error defined in (10.108). Consider the following models and measurement noises {nw = min((k), ¢} {ne —hy(k) m(k) = -fu() am{k) = ~ha{k) => nS, mmeN, Eth m) = Elta, m) Now, applying equation (10.107), Wry AO), < eC) and Way + A(O)I, < (A) => [hall <5 lr ~ ADI, + $s + ACI, < CA) To obtain (10.106), simply take the infimum over A. a Lemma 10.17 For the identified model Higa) = yz’, nM the following ienifation error bound holds: Wh Hh) < ne + S01) (20108) 368 ROBUST IDENTIFICATION The value of n, that minimizes the above bound is the smallest integer n achieving ®(n) © € Proof. The first part follows from: We Ah = SAG) = vel + S100) < nes 20k) (10.109) For the second pat, define a(n) =ne + 370K) ‘Therefore a(t +1) ~ a(n) = € ~ O(n) is nondecreasing because © is nonin creasing. As a consequence, the minimum of a(n) is achieved at the smallest rn, satisfying a(n+ 1) > a(n) a ‘An intuitwe interpretation of the above result is the following. Since VA(K)| < (k) and & is nonincreasing, for values of k beyond the point for which (K) < ¢, there is no useful information on the system which can be obtained from the data It is clear from (10.109) that the value of n should depend not only on M but aiso on ¢ in order for the algorithm to converge. This is supported by the fact that there are no untuned idemtfication algorithms based on impulse response data ((160, 193, 194)) as proved before. There are many values of ‘n that may be selected for the algorithm to converge. In the sequel, three of them are presented. Lemma 10.18 The algorithms obtained by selecting n as either m(M,«) = min(M ,¢p ~ cy Ine) or na(M ,€) = min(M, co + c1€-"),1 € (0,1) and ¢a,¢; real constants, are tuned to € and convergent. A third, asymptotically globally op- imal algorithm tuned 10 (@,€) can be obtained using ns = min(M,n.), Me defined in Lemma 10.17. Proof, For n= ni, a(n) < (ca ~ eyln€ + Ie + Dig, PCR). The fist term vanishes as ¢ -+0, and ny 00 as M 0 and-e'~~0. Since € &), this implies that the second term also goes to ze¥0. Convergence for n =n) can bbe proved in a similar way. For m= ns, when M— 00 then 1+, and therefore the upper bound in (10.108) achieves the optimal value in (10.106) Also, for any given 8 > 0, there is always aq such that S22 y,, (Kk) < 8/2. TIMEDOMAINIDENTIICATION 369 and € > 0 such that (g + 1}e < 8/2. Because the lower bound in (10.106) is alobal, the last procedure is asymptotically globally optimal. a ‘As a special case of the above results, consider again the set of models ‘Yoolp, K), which includes the exponentially stable discrete time linear sys- tems considered in the Ha. identification setup ((146)). For p > 1 and K > 0, Hoo(p, K) € A(D), the disk algebra of functions analytic in the open unit disk and continuous on the unit circle. It is well known that the impulse response ‘of this class of systems is bounded above by the sequence {Xp-*, k > 0} This leads to the following corollary. Corollary 10.1 For the identified model #1", defined in (10.108), the identifi cation error can be bounded as follows: Kp Wh Hay < nes SO (10.110) This bound is minimized by selecting the smallest integer n > ma, with pst Inp (20111) Proof. Follows immediately from Lemma 10.17 by noting that %x(), K) CS when selecting (k) = Kp™* o Notes 1. Depending on the resulting value of n, it may be necessary to balance and truncate the identified model H",, to obtain a smaller-order model ‘This is a standard procedure, which may be simplified, avoiding the computation of Gramians, using the algorithm in [45, 127]. The new error bounds are given now in terms of the |: oF the Hankel norm, ‘Similarly although computationally more expensive, an optimal Hankel ‘model reduction may be performed, 2. For the case of systems that have all positive impulse response values, the &; bound coincides with the Ha. bound. Therefore the asymptoti: cally &; optimal algorithm is also asymptotically Hx, optimal. 3. When the value of n, is very high, for example in cases where M is high and ¢ is very small, care must be taken when balancing, in particular, when the first singular vector of the Hankel matrix is computed ([45, 127)), to avoid numerical errors that may increase the approximation 4. The identified system can be represented by a family of models, in terms of the nominal H, and the identification error bound, This family could be unnecessarily Conservative if the convergence of the algorithm 370 ROBUST IDENTIFICATION is not monotonically decreasing with € -+0, In this case there might exist ¢ > « for which the identification error is smaller than the one considered and still “covers” the class of models S and noise N= Gel€) C fu(€,). All the algorithms presented here have monotonically decreasing convergence. Based on all Samples Next we present, a strongly optimal convergent algorithm, that is, locally optimal for any set of experimental data. This al- gorithm is based on the one in {160} Consider the vector of experimental data y of length M and di ho (k) = min yy + €, (8) (wo.12) fi (k) = max ly, — €, -(K)] (10.13) Note that Jy(k) is the lowest value of h(k) compatible with the a priori information, therefore consistent. Similarly, hy(K) is the upper bound on. ‘h(k) for each k, so that model h belongs to the consistency set. In addition, the consistency interval {hy (k), hu (k)] has length 2min [e, (k)} The algorithm is as follows: siy.0,0), = {AMM KEM oar The above algorithm is called central because it selects the model in the center of the consistency set; therefore it is also strongly optimal. This lat property can be proved as follows. Lemma 1019 The worst-case error for algorithm (10.114) is e[Ay, ®,€)] = > min fe, (4)] + JO, - > G9] (10.115) and tends 10 zero in the limit M +00 and € ~+0. The algorithms is strongly optimal Proof. The bound can be obtained as in Lemma 10.17, using the fact that AY. ®, €), — A(R] < mine, P(K)], & Mad since the model is in the center of the interval iz(k),/u(&)]. The above bound can be achieved with a suitable choice of system and measurement noise, It is clear that the consistency set is symmetric with respect to the ‘model Hiy = Aly, , €); therefore the algorithm is also central. TIMEDOMAIN IDENTIFICATION 371 Note that the first term of the error is equivalent to ner [6] (10.116) with n, defined in Lemma 10.17. This error converges to zero as (1/e,M) — oo. Using the above equation, it is easily seen that the total error can be rewritten as the global lower bound in equation (10.106). It follows that the algorithm is globally optimal. o In the special case where (k) = Kp-*, the bound (10.115) reduces to ((160)): we ie e[Aly,®,€)] = Zemin Ol SH (10.117) ‘The above algorithm is clearly tuned to all a prior’ information. Next, an- other interpolatory algorithm, which is tuned only to the model information, that is, ©, will be presented. This algorithm is given by: sign (4) min|ly«{,(0)] Os ks M— 0, kom AQ.) = { (10.118) Lemma 10.20. The worst-case error for algorithm (10.118) is: 14(9,0)) = 5 ming, 20(0] + ¥ ee] (20.119) which tends 19 zero inthe limit M oo and € ~+0. Proof. The proof is similar to the one above, considering the fact that [Ay ~ A(k)| < min fe, 20(6)] when [yx| < (If, however, [| 2 (k), the following holds: LA, ®)e = A(K)] < 26(4) Since the experiment and the noise have the same sign, in this ease LAly. ®), — AC} < WH) ~ yu] < € In both cases this leads to the desired bound. ° 372 ROBUST IDENTIFICATION For the particular case when ¢(k) = Kp-*, the bound (10.119) simplifies 10 ({166)} elAly,)] ~ Y minfe, 2049) y 10.5 FURTHER RESEARCH TOPICS Due to the fact that robust identification is a currently active research area, there are yet many theoretical and computational aspects that have not been developed fully. In this section, several items that have not been included in the main part of this chapter will be commented on and referenced. An excellent recent survey of the area of robust identification can be found in {199}, which includes an extensive list of references 10.5.1, Unstable Systems ‘The methods developed in this chapter apply to stable LTI systems. For slowly divergent systems, that is, with slow unstable poles, care must be taken when carrying out the experiment, especially when considering its duration ({127}). In general, neither time nor frequency response experiments can be Performed over open-loop unstable systems. Rather, they should be carried ut after 2 stabilizing controller has been applied to the plant. In that ease, the closed-loop input/output data should be used to identify the open-loop plant. The identification error should also take into account the fact that both. ‘model and plant are unstable. Thus the 7, norm should be replaced by a suitable distance function. In [196] a complete framework for the robust identification of strongly'® stabilizable plants is presented. The measure of identification error is com- puted in terms of the gap, graph, or chordal metrics. 10.5.2, Nonuniformly Spaced Experimental Points In the 2. identification two-stage procedures, it is assumed that the fre- quency experimental data points are equally spaced. Since this may not be the case in practical situations, the standard algorithm should be modified ac- cordingly. In (8] and (235] interpolation is used to perform identification with nonuniformly spaced experimental data points. In [127] a procedure that uses explicitly the unequally separated data points is presented, in the context of, model approximation, "A sronely sabiizable plant isthe one that can be nabilized hy a stable conte FURTHER RESEARCH TOPICS 373 10.5.3 Model Reduction In many of the algorithms presented in this chapter, the model order re- mains high. As a consequence, a further model reduction step needs to be performed and the model error bounds should be modified accordingly. This hhas been attempted in (142] for é identification and in (198, 253] for Hu. identification, 10.5.4 Continuous Time Plants ‘The robust identification framework presented so far considers both discrete and continuous time systems represented by H(z) in Section 10.3. Never- theless, in this last case, some extra considerations related to the frequency aliasing should be taken into account, Specifically, strictly proper stable LTT plants with an a priori upper bound on their roll-off frequencies should be considered. This problem has been treated in (9, 147) 10.5.5 Sample Complexity For application of the identification techniques to practical problems, it is important to know whether the experiments can be performed in a reasonable amount of time. Specifically, the following question may be asked. Which is the minimum time length of an experiment in order to arrive at a certain degree of accuracy, measured by the worst-case identification error? A general framework for this problem has been considered in [183] and specific results have been obtained in (163, 85, 243, 303]. These results are presented next Consider the FIR systems of length M, bounded experimental inputs [luli <1, and a (optimal over all algorithms and inputs) worst-case iden- tification error bounded by nf theca 20 as where € is the measurement noise uniform upper bound and the parame ter Q > 1 defines the level of uncertainty (Q = 1 corresponds to the optimal error bound 2e). For this case, the minimum length of the experiment dura- tion to achieve this bound is O@//2), with f() a strictly monotonically increasing function in (0,1), with f(0) = 0 and fim f(x) = 1 as x —» 1. Itean bbe shown that this isa tight bound, A consequence of this result is that the worst case clas of measurement noise (uniformly bounded by e) considered in time-domain experiments may be very conservative. Therefore, as suggested in [243], the class of measure~ ‘ment noise should be relaxed or noise should be considered in the average ‘On the other hand, the worst-case class of models does not need to be mod: ified. Tn the case of M., identification, the sample complexity is only quadratic in at ((199). 374 ROBUST IDENTIFICATION 10.5.6 Mixed Time/Frequency Experiments ‘When applying robust identification techniques to either frequency- or time: domain experiments separately, it may happen that a model that efficiently fits the data in one domain may have poor performance in the other do- ‘main, For example, with only frequency experimental data points, a “good” frequency response fitting (small %,, error norm) may lead to a “poor” fit ting in the impulse response. Additionally, from an information theoretic viewpoint, more experiments produce a smaller consistency set of indistin- ‘guishable models, and as a consequence a smaller worst-case error, From a practical standpoint, in many cases robust identification algorithms are ap- plied to systems that may not be exactly LTI. Under these circumstances itis desirable to perform both time and frequency response experiments ((272)) to assess the validity of these assumptions. Thus, in these cases, an algorithm that combines both types of experimental data could take advantage of the additional data available Recent papers [77, 333] have proposed interpolatory algorithms that use data obtained from time-domain experiments and generate a nominal model together with an a, bound on the identification error. Therefore, although they combine time domain measurements with frequency domain error bounds, they do not use both types of experimental information simulta: neously, Alternatively, some recently propased algorithms explait hoth time and frequency domain experiments at the same time. In {273 necessary and suf- ficient conditions for the consistency of mixed time/ frequency experimental data for FIR systems are presented and an interpolatory identification algo- rithm is proposed, In [231] the above result was extended to IIR systems, based on a general interpolation theorem ([21}), which allows the simultan- cous interpolation of both the time and frequency experiments. The solution to this problem leads to a convex LMIls optimization, 10.5.7 Mixed Parametric/Nonparametric Models All ofthe results in this chapter address nonparametric identification of mod- els with a worst-case global bound. In many cases, part of the model has @ clear parametric structure. In these cases disregarding this information may lead to conservative results (see Example 10.4). This is usually the case in mechanical flexible structures, which have a well defined parametric model for the lower-frequency modes and an unknown higher-frequency behavior ‘that naturally leads to a nonparametric identification ((121)) In general, the parametric information does not appear explicitly in the 4 priori knowledge (K,p) usually considered. Therefore it is important to include it, if available, so that less conservative a priori estimates of (Kp) for the nonparametric portion can be derived. This is the case of systems ex- hibiting large peaks in the frequency response (for instance, a poorly damped FURTHER RESEARCH TOPICS 378 flexible structure), which must be “covered” by large values of K ({121)). This is a consequence of the fact that the usual a prion information characterizes only the smoothness and magnitude of the whole class of models but cannot distinguish among other properties, for example, low frequency model struc: ture. Instead, the a priori parametric knowledge provides more “structured” information, ‘There are several results that address this problem, for example, in [121. 175, 205, 232, 333]. In particular, the identification of affine parametric and nonparametric models using simultaneous time and frequency response ex- perimental data can be solved as a convex optimization problem ([232]). This includes cases of practical interest such as flexible mechanical structures, as well as models that can be described concisely in terms of a set of Laguerre ‘or Kautz functions of, more generally, any other basis of 2 ((148)) APPLICATION EXAMPLES 11.1 SAC-C ATTITUDE CONTROL ANALYSIS. n Introduction The SAC-C satellite is a joint project between the Natioual Conunission of Space Activities (Argentina) and NASA (United States). It is essentially an Earth observation satellite that carries an Argentine CCD camera for Earth resources observation, a Global Positioning System (GPS) experiment from JPL (NASA), and a magnetic sensor from Denmark. The latter has a Star Imager and two Magnetometers mounted on an 8 meter boom, which will bbe deployed in orbit. The purpose of the boom is to magnetically isolate the ‘magnetometers from the spacecraft. The length of this boom makes it very flexible; therefore the model of the spacecraft cannot be considered as a rigid body (see Figure 11.1) The attitude objective is to point the camera toward the vertical with a precision of +1° around the roll and pitch axis; that is, the line of sight of the camera should be within a 1° square with respect to the vertical direction. Based on optical considerations, the jitter of the control should be on the order of 0.05%/s, Due to several practical limitations, it is important to have the simplest possible controller and the least amount of hardware, These limitations stem from the fact that the microprocessor onboard should not only implement the mission contro! mode but also several alternative safe-hold modes, as well as failure logics. Therefore a low-order controller is desirable. In addition, sensor and actuator hardware adds weight to the spacecraft, which in turn decreases the available payload. For this reason, only the minimum required rotations along the axes are sensed. Similarly only the minimum number 377 378 APPUCATIONEXAMPLES Of torques are applied by the actuators. Thus only the roll and pitch axes are sensed by means of a scan wheel horizon sensor (SW) and the control torques are applied only in the pitch and yaw axes. These control torques act through magnetic torque coils (MTCs) and momentum wheels, the latter ‘built into the SW. In the design, the torque bounds of both actuators should bbe considered, due to limitations in the maximum magnetic moment of the MTC and the maximum size and acceleration of the momentum wheels ‘The desired attitude of the satellite in Mission Mode (MM) is the fol- lowing: roll along the velocity vector, yaw opposite to the vertical, and pitch perpendicular to the orbital plane (See Figure 11.1). Our goal is to design ‘@ minimal order controller to achieve this nominal attitude, This controller feeds back the roll and pitch angles and rates as well as the momentum wheel velocities, using static LOR feedback. The only dynamics involved in the controller arise from a yaw angle and rate observer (the only axis without sensor). Therefore the complete controller has only two states, is applied to only two of the three axes, and senses two out of three axes: as simple as it can be. The possibility of obtaining such a low-order controller and not hhaving to sense or actuate over all axes is due to the fact that the spacecraft and the orbital dynamics are coupled. ‘The attitude for the Earth Pointing Safe Hold Mode (EPSHM) is the same as for MM, except that the momentum wheels are kept at nominal constant speed. Here the control is applied using only the Earth magnetic held, through the MICs, The controller design is carried-out using a very simple procedure, that takes into consideration the limitations in the controller order, sensors, and actuators. This design is validated in Section 11.1.4 through an analysis involv: ing all the performance objectives as well as model uncertainties. Using the structured singular value, we assess the robustness of the controller against the unknown high-order dynamics of the boom, considered as global uncer- tainty. Robust performance is guaranteed using 2 for both modes: MM and EPSHM. Several time simulations of the complete system are presented in- volving random measurement noise and torque perturbations. Further details of this analysis are reported in [274] 11.1.2. Linear Model ‘The general nonlinear dynamic equation of the system is: Tut Teyt Ty + Ty = [pW 4 W x pW W xh tlle] SAC ATITUDE CONTROLANALYS'S 379. ‘ORBIT Xa FORWARD VELOCITY << A EARTH Figure 11.1, SAC-C Eorth observation satelite Here W is the angular rate of the spacecraft with respect to inertial space, represented as the rotation of a frame attached to the spacecraft with respect to the inertial reference frame, The inertia of the rigid part of the spacecraft is Fr and h represents the wheels momentum in the spacecraft frame. This frame is the usual roll (x), pitch (y), and yaw (z) axes mentioned before (see Figure 11.1), The external torques consist of the perturbation (7,), gravitational (T;), and magneticcontrol (Tx), that is, T = Ty + Ty + Tm The gravitational torque is due to the effect of the rigid body model of the boom and the magnetic control is produced by the MTCs. The perturbation torques are due to the aerodynamic drag (almost constant in pitch and yaw along the orbit) and the 380 APPLICATION EXAMPLES magnetic and gravitational residual torques, both periodic and negligible as compared with the former: The wheels and magnetic torque col torques are used for control purposes the first one depends on the velocity change of the momentum exchange wheels and therefore does not change the total momentum of the spacecraft, the latter is an external torque that does change the total momentum. Note that these torques are applied only in the pitch and yaw axes, The reason is that a control torque inthe roll axis is unnecessary due to the orbital coupling between roll and yaw, as will be seen next. This reduces the number of actuators, therefore simplifying the design and minimizing costs The external perturbation torques are bounded but otherwise unknown Next, linearization is computed around the equilibrium conditions, cor responding to the nominal attitude of the satelite: the roll and yaw axes colinear with the velocity and local vertical vectors, respectively, and pitch perpendicular to the orbital plane, which completes the spacecraft frame (boom opposite to the local vertical). The roll, pitch, and yaw angles are de- noted as Op, Up, and Oy, respectively. The incremental values of the wheels momenta are 8h, in pitch and dh, in the yaw direction. For simplicity we can assume that Jy = diag [/, ly. Ie] Since the roll and yaw dynamics are coupled but independent from the pitch, we define the system state as xT = [x5 x7] ay ty = [Oe Oe Oy By dhe]", —xp=[ 6p Op Bhy|” ‘Adding the equations for the wheels torques Tey = —Siy and Tue = we obtain the state equations for both the {roll-yaw] and [pitch] dynamics: = Apt + Bry lies | +B.Tpe (12) pean «Bows [72 | + BTyy (13) { (0) = Ax(0) + Bult) + B,T, y= [22 Te] + aa xl + Dute) Oona Tent where x(1) has been defined in (11.1) and where: 1, wy 2 | SAC-C ATITUDE CONTROLANALYSS 381 Note that the system output consists only of the measured states, that is roll and pitch, Mattices Ay, Bry, Ap, Bpisn, By. and B, have been defined in [274 11.1.3: Design Constraints To simplify the controller as much as possible, a state feedback has been designed. The only controller dynamics will be the ones from the yaw angle and rate observer, due to the fact that only roll and pitch sensors are available (scan wheels) ‘A static LOR controller has been designed, taking into account the desired bounds for the different states and actuators when computing the weights to solve the minimization problem. These constraints are as follows Tue] [0012sing N-m d 2610- Nem Tr | | otzacose Nm aus) te) | foi Nm ‘This is due to the fact that the torque bound in each wheel is 12 mNm and that in pitch both add their torques projected through cosa (a < 10°). In yaw, on the other hand, they subtract both torques projected by sina. The bbound on the magnetic torque coils is computed as the average torque along the orbit. Each MTC has 15 Am? and the average magnetic field along the orbit is (V3/2) 0.4» 10-' Tesla, Therefore the actuator weight in the LOR problem is a diagonal matrix that has the inverse of these hounds as each element (Bryson's Rule, see Chapter 5). “The state weighting matrix is also diagonal with the inverses of the maxi- ‘mum allowable bound in each of the states. These are given by: bn oss Om r by o/s by r ah. |<] 0.06N-m-s co bp O1°/s op r ah] [006N-m-s The bounds in both 5h; and 8h, correspond to @ maximum deviation from the nominal wheel speed of 400 rpm, For a total momentum of 3.2 Nims. (2000 rpm, J = 0.0077) along the pitch axis, these bounds are 0.11 and 0.64 Noms, respectively. The observer involves only the [roll-yaw] dynamics, due to the fact that pitch is uncoupled, Furthermore, a reduced order observer will be used so 382 APPLICATION EXAMPLES that the controller has the minimum possible order. The measured, estimated, ‘and input variables to this observer are, respectively, i a = af by 8 | Tw: melon 8 [#) 2 [Be] [* le rm Closed-Loop System The observer spstem is given by Ao | Bo Ao | Bor Bor fo] | au) Co | Do Co | Dor x2 with output fy and input [x2 7", In this case the observer gain can be computed analytically ([274). ‘The control signals, LOR gain matrix, and state feedback are oR Ts i. 4 | Tme| [Fe Fy Fie 03] | gy, a le Gir Oar Ons Fe |B] O18) h o ™ or a, where the feedback matrix has been partitioned according to the roll, yaw, {5h,, and pitch variables, From the above equation and the observer dynamics we compute the controller as follows: (Ao ~ Bo2FyCo)z + {Bor ~ Bor (Fy Dor + (Fe Fre )} %m Du O FrCo2~ (FrDoi + [Fe AD™[7 | =] or equivalently, a9) SAC-C ATITUDECONIROL ANALYSIS. 383, Note that the controller has the minumum number of states, given the avail able sensors (roll and pitch) and the minimum number of actuator outputs (yaw and pitch) given the the coupled dynamics. Using the above definitions for the controller, input, state, and output variables, the closed-loop system is given by: (A+ BD,C)x + BC yz + BpT» + BDyn 2 = Bre Az + Bin Ei Thus the state-space realization of the generalized plant G,(3) with states [27 27], disturbance inputs p 2 [ TF n? ]’, and e the output to be mini- mized is A+BD,C BC, | By BD, BC AL | Oe Be |_| Ac| & Teen Opn2 O54 Ones Ce | De DCG | Oe Dy 11.1.4 Robustness Anatysis The closed-loop dynamics computed in the last section correspond to the nominal system, which takes into account only the rigid body modes of both satellite and boom. Nevertheless, the flexible modes of the boom, should be taken into account explicitly in the robustness analysis, ‘These modes can be modeled as uncertainty in the high-order dynamics, using experimental data provided by the boom manufacturer ((295)), to com- pute the frequency distribution of the uncertainty. The boom dynamics is considered in this setup as follows: Hs = Grontsdrar v2 [ TH], ys 2[M] ‘The reason i that the diving input, which iis the Rexble dynamics Osilations the angla rotation of the spacecraft in rll and ich (he torsion around the yaw axis eligible) One other andthe pemtubation ofthe boom onthe stele sructre represented by the (intemal) trgues Applied nthe roland pth anc Gusto these esllation. The experiment in [295 determined lower requency yeamics ot! Hz. Therefore a convenient Weigh that “cover” the uncertain high fequency Ojai of the boom and assumes a 20% modeling eror at lower frequencies Seon Wa(s) = 04752 1,,5 384 APPLICATION EXAMPLES The performance specifications call for minimizing the states and bounding the control torques of MWs and MTCs. Therefore a suitable performance ‘weight is a constant diagonal matrix with the inverse of the upper bounds in (115) and (11.6) as its elements. To perform the robustness analysis ofthe satelite the nominal and uncer. tain portions, and weights for robustness and performance are connected in the LFT form shown in Figure 11.2 ‘The closed-loop nominal system computed in Section 11.1.3 corresponds to the lower portion of the general system R(s) = Fe(P, K), connecting the inputs p to the outputs e, that is, Gy(s) = Ron(s). The total system, with- ‘out weights, connects the inputs (us,p) to the outputs (¥s,e) and has the following state-space realization: Ac Ris) = o10000000 000000100 G “feta vm To perform the robustness analysis the output weights must be added to the system as indicated in Figure 11.2. Furthermore, to normalize the anal- ysis, we scale the inputs in vector p so that the magnitude of the worst-case perturbations has unity norm. To this end we use a constant input weighting ‘matrix having the maximum perturbation torques and measurement noise as its elements. These perturbation torques are due mainly to the atmospheric mae va —Awnj2—1 Pe) ¢ fim} d y K(s) Figure 11.2. Unear toctional tanslormation setup for robustness analysis. SAC ATITUDECONTROL ANALYSIS 385, drag effect in yaw and pitch with a maximum of 10-° Nn, and tothe residual ‘magnetic torques in roll with a maximum of 10-® Nem. The angular measure iment error due to the scan wheels is 0.3°in roll and pitch, the only measured axes (recall yaw is estimated with an observer). The angular rate measure ment error in the same axes stems from the error in the LSB of the word that spans the total angle, this is 0.01". The rate is computed as the difference between two correlative angle samples, divided by the sampling time 7. For this reason, the rate “measurement” noise weight is similar to a notch fier centered al w = 2/T and scaled by 0.01*/s ols We (s+ 2/5)(s/207 +1) sels) au Finally, the rate measurement errors of the wheel velocity, which con- tribute to the error in their angular momentum in yaw (8h) and pitch (5h,), amount to a maximum of 2% of hyo. ‘The necessary and sufficient conditions to guarantee robust (internal) sta- bility, nominal performance, and robust performance are RS > [Ws(s)Ru (Wye <1 (a2) NP = [Wp(s)Ra2(5)Wir(5)I, SL Gusy RP e=> sup Ha [Wo(s)R(s)W)(s)] <1 ais) where Wo(s) = diag|Wa(s) Wp(s)], Wr(s) = diag[Wi(s) Wra(s)],and the ‘uncertainty structure A contains two full blocks, one for the boom high-order ‘dynamics, the second for the performance objectives. 5 Simulations ‘The following are simulated outputs of the main variables to be minimized, in Mission Mode (MM). Figures 11.3 to 11.6 show the angles, angular rates, and momentum wheel and magnetic control torques, all below the desired bounds. Figure 11.7 shows the robust stability and nominal and robust per- formance conditions, just below the value I ‘A different controller has been designed with the same methodology to control the spacecraft only with the MTCs, called Earth Pointing Safe-Hold Mode (EPSHM). In this case the wheels are kept at a constant speed and provide the necessary angular mmomentum to the system. In particular, the ‘weight on the MW control torques for the design is large enough so that the wheel torques are below 6 x 10° Nm, that is, almost constant speed, The robust analysis curves based on the infinity norm and the structured singular values are presented in Figure 11.8. 386 ogress “3900828003605 aa09 000 000 ine) Figure 11.3. Rotation angus rors MM) otc) Yaw) Pa ete cons oo 0.009 t 009 oo oo A ! 0 360 20003004008 S003 tooo via) APPLICATION EXAMPLES Fo (i), Yam (Pach) ii Torque (nN) SAC ATITUDE CONTROL ANALYSIS. 387 i ores in Yaw (0) & Pa (-) sine 6) Figure 11.5. Momeniurn wheel contol foraues (MM), MIC toques in Yaw so) & Pe) Figure 11.4. Rotation ate erors (MM). sine (6) Figure 11.6. Magnetic contol lorques (MM). 388 APPUCATIONEXAMPLES jj Nominal ana Robust Performance sis ah Robust Stabilty ESO RSEEE HEEL Av w (adie) 4 w (cadis) Figure 11.8, Robust stabilty and nominal and robust performance conaiions (EPSHM). CONTROLLER DESIGN FOR AD.O PLANT 389 11.2 CONTROLLER DESIGN FOR A D,O PLANT In this section we illustrate by step, the robust control methodology through an application (0 a D;O plant. The mathematical model of the plant is a linearized and simplified version of a nonlinear system with a delay. This particular application is a standard problem in the process control area. The Purpose of this example is to assess the incidence of model and disturbance luncertainty in this typical application. The model uncertainty arises from the time delay and the gains and poles of the transfer function, which depend on the particular linearization point. The usual control structure to handle systems with time delays is the Smith predictor ([289)). This classical method in process control has the advantage that, when the delay is known exactly, the controller designed to achieve (nominal) stability needs to take into account only the rational (nondelayed) part of the plant. In this example, based on the Smith predictor, we perform the robust analysis and design of a controller for this particular uncertain system, which exhibits all the relevant issues of process control problems. The following example is based on [268), and the interpretation of Smith predictors with uncertainty is described in [211] 11.2.1, Model of the Plant ‘The plant to be considered is an experimental process that obtains D;O by isotopic exchange between H;O and HyS (see (128, 206)). The plant has already been constructed but is not yet operational. ‘A general diagram of the process can be seen in Figure 11.9. This diagram is a simplification of the process flow sheet (there are more cold and hot towers in the actual plant) but retains the main issues of interest, The input water flow and liquid flows are marked with full lines and the gas flows with dashed ones. Two compressors have been included, as well as three heat exchangers that control the temperature in the process. The greatest D;O concentration in each stage is obtained between the hot and cold towers. A fraction of the partially enriched water of the first stage is sent to the second stage for further enrichment, The product is extracted from the second stage after the H3S stripper eliminates the gas. The isotopic exchange process that takes place in the cold and hot columns, respectively, is the following H,0+HDS “© HDO+H,S HDO+H,S C H,0+HDS The controlled variable is the ratio L/G, with L and G representing the liquid and gas operational flows, respectively. The value of G remains almost constant duc to the fact that it circulates on a closed circuit, as seen in the same figure; therefore the control is performed by changing the value of L 390 APPLICATION EXAMPLES sho) Figure 11.9. Schematic dlagram of the D,0 plant. ‘A general nonlinear model of the whole process has been implemented ({206, 228}) for simulation purposes. This general model has been obtained from physical considerations and is too complicated for both analysis and control design. Therefore the controller analysis and synthesis are based on a linear model that takes differential H,O flow as an input and differential D;O ‘outlet concentration as the output. The linearization has been made at 0.01%, 0.3%, and 1% disturbances in the input water flow, under the restriction of 15% outlet concentration of D,O. The structure of the linear model remains the same for all three linearization points, but the parameters (gains and poles) differ from one point to another. Assuming continuity between the linearization points (water flow disturbance) and the model parameters, the linear process may be described by a set of models with the same linear structure but with parameters lying in real bounded intervals, ‘The delay is a consequence of the measurement procedure for the out- put variable. It consists of the collection of samples at each column, which takes around 15 minutes per column, with a transportation time between columns of another 15 minutes. All these samples are measured with a mass spectrometer, to obtain, after approximately a 4 hour test, the DO concen: LE CONTROLERDESIGNFOR ADO PLANT 391 tration. Depending on particular conditions, the elapsed time between the sample extraction and the final output value may differ significantly. There. fore for analysis purposes, we consider a nominal delay of 5 hours, obtained as an average value, with an uncertainty of £1 hour, In this linear model we also consider additive disturbances in the output. ‘These disturbances include several liquid and gas flows, which go through part of the process and which, for operational purposes, are bounded. If these bounds are exceeded, the required level of D;O extraction will not be ‘met (or we will have no extraction at all). Due to the fact that these flows circulate through most of the columns, we may assume that the dynamics of the disturbances is very slow (see Figure 11.9). For the above reasons, we ‘model the additive disturbances at the output as a set of bounded signals with unknown and slow dynamics, ‘The set of linear models representing the process is the following _— «fe, Gls) = 0 |e ot ae tye | Teme) Kee Miu 4 71 €lypwjh, vj >0, 4 ais) T=%+8, [8l<1, 8ER, where the nominal delay is x» = 5 hours. The uncertainty intervals for poles and gains are obtained from their values at the three linearization points mentioned before ([228)). The disturbances are represented as a weighted and bounded set, that is, {We(s)d{s)|ldl2 < 1). Here W,(s) takes care ofthe fact that the distur bances have a higher energy content in the lower frequencies (slow dynam ies). For the controller design step, a rational approximation of the nominal time delay needs to be made. Many different methods exiss, but @ standard simple procedure is the use of Padé approximants, Nevertheless, we will verify the robust performance condition with the exact values of the delay, without approximation errors, 11.2.2 Robustness Analysis In this section we obtain conditions on the controller such that stability, performance, and robustness are achieved. Due to the fact that the system hhas a delay, we work with the Smith predictor control structure, but including the uncertainty in the delay, poles, and gains. The classical Smith predictor structure ([28]) is presented in Figure 11.10, The stable plant G(s) consists of a rational term Gy(s,p) with a vector 392 APPLICATION EXAMPLES of uncertain parameters p. p; € j.b,) (in this case the poles and gains in (11.15)) and an uncertain delay 7 = 7, + 8,[6] < 1. Define the nominal model 88 Gals) = € "*Gyy(s,po) which wil be used for the design. Next, we analyze step by step nominal and robust stability and performance Nominal Stability The Smith predictor operates by subtracting from the feedback signal the difference between the nominal model and the core. sponding minimal phase part (Figure 1110) This is equivalent tothe incl sion ofa prediction in the sytem since the feedback signal f() is exaely the output ata future time (1+), This procedure i elective only when theres Complete knowledge of the system (he parameters ake their nominal values +=, and p= pp) otherwise we will not be able to havea corect prediction oy From the same figure, we observe that, the nominal transfer function be- tween u and the feedback f, equals Gy, © Gy(s,p.). Therefore in terms of nominal stability, only a controller K(s) that stabilizes the nominal minimum phase part Gwo(s), regardless ofthe time delay 1», needs to be designed This isthe main advantage ofthe classical Smith predictor, but this esti to longer valid when considering performance of robustness issues, a5 will be seen nex Nominal Performance Based on the results in previous :chapters, nom- inal performance depends on the transfer function between the disturbance and the output error signal. Figure 11.11 yields the following transfer function in the nominal case, that is, G(s) = G(s) vals) = [10% Tu(s)] Walsdd(s) (1.16) AO Figure 11.10. Smith precictr. CONTROLER DESIGNFORAD.O PLANT 393 where we have defined the complementary sensitivity as Tals) © Gass) K(s) [1 + Gael )K(P The necessary and suffiient condition for nominal performance is Nominal performance => jlya(slz $1, ¥lld(s)I <1 si qt) Tu (s}] Wal) This leads to the following 21. control problem: Ip- sary hll “°° TaCO WAC oy An important point is that the nominal performance condition includes the nominal delay 7,.. As we mentioned before, the classical Smith predictor does ‘ot simplify the analysis step for performance, as it did with nominal stability. A controller that solves equation (11.18) can be obtained by a standard Ha. ‘optimal control algorithm, with a previous approximation of the time delay. The optimal solution without approximating the delay can be obtained from [116], but this approach cannot solve the robust performance prahlem Robust Stability In this section we analize the closed-loop stability for all Possible delays between 4 and 6 hours and all possible parameters p [poles and gains in (11.15)] of the minimum phase part of the plant. Wale)d A lyautye Ks) [ber ou(s,p) Figure 11.11. Smim predictor wit: uncertainty. 394 APPLICATIONEXAMPLES A practical! way to analyze models with parametric uncertainty is to “cover” them with dynamic uncertainty. In this case we have used a mult plicative dynamic description for the set of plants, as follows: ME {[1+A.Wa(s)Gmols), eC y (Sul <1} (1119) In this application, there are two different sources of parameter uncer. tainty, which in principle could be covered separately. This will produce a ‘smaller” set of plants than in the case where the coverage is performed over both parametric uncertainties together. As a consequence, the analysis results for the “smaller” set of models will always be less conservative than in the case where both parametric uncertainties are globally covered, For the delay, we transform the dashed block of Figure 11.1] into a nom: inal system with global multiplicative uncertainty, with p = py. To this end, consider the family of models GF Gs) — Gals) + Gols) (11.20) = eles -re out, BER, jl mate =| at2t) Figure 11.12 shows the above condition in a magnitude Bode plot. A rational " ama hoc analysis considering directly parametric uncertain could be made, bu iis ot the purpose of thin example. The resus on analysis of parametnc uncertainty srutures ae ee? Snslyieal methods that solve particular eases (and donot i hs spication) oF algorithmic ‘methods that solve general cases but ae computationally involved, infact « NP-complete prob Jem (jf), 24). There i 9 yet a sates method that supports a peneral parametric anal ‘Am cxcllent survey ena he found [35] CONTROLERDESIGNFORAD,O PANT 395 7 10 10 10% 10 107 10% wads) ‘Figure 11.12. Dynamic weight W(s} covering delay uncertainty. approximation Ws(s) for the weight can be computed from the same figure, Note that the set of models that covers the delay uncertainty is independent of the nominal delay 7, Following a similar procedure, a set of models with dynamic multiplicative uncertainty, which covers the parametric uncertainty in gains and poles can be obtained. For briefness it will not be repeated here, but the corresponding, ‘weighting function may be seen in Figure 11.13. ‘The necessary and sufficient condition to guarantee the stability of all ‘members of a family of models with multiplicative uncertainty” is IMs) Tus) <1 (122) From the above equation, we observe that the critical point in terms of robust stability i the frequency w. for which |W,(ja.)| > 1. From (11.22) itis clear that, for w > a, the values of |Tiy(jw)| should be less than one to achieve the robust stability condition. In some sense this frequency defines an upper bound for the closed-loop bandwidth, the latter interpreted in terms of Ty (5) 2 For the family M the condition is necessary and sufficient when replacing W with Wy, but since @'C AM forthe actual set of models 8 only sufeient 396 APPLICATIONEXAMPLES Weights for uncertain delay & poles 10! Total uncertainty weight +04 : a Uncertain pote-gain weight 107 10 (ee ay w(radis) Figure 11.13. Dynamic weights covering pole gain and delay uncertainties, Tota Uncertainty wight is Wo(5) ‘The above condition could be useful in determining the relative impor- tance of both types of uncertainties for the robust stability of the plant. By comparing the weights in Figure 11.13 we observe that, for the frequency range of interest, the curve due to the pole and gain uncertainty remains be- low one. Therefore the limiting factor in the robust stability condition is the delay uncertainty, Nevertheless, in the lower-frequency range, the uncertainty in poles and gains is larger than the one due to the delay. Therefore, from a practical point of view, we should consider the total uncertainty weight W.(s), which covers both types of parametric uncertainty (see Figure 11.13). Finally, it should be noted that the Smith predictor structure allows this analysis to be independent of the nominal delay 7. Only the uncertainty is taken into account, which appears indirectly in the weights Wa(s) (see equation (11.21)) and W, (3) Robust Performance For robust performance we need to achieve the performance objective for all models in the set {[1 + &,Wu(s)] Ga(S), [Mul < 1). From Figure 11.11 we obtain the following transfer function between d and the output signal yals) = [10 Ty (s)] [1+ AaWals)Tu(s))! Walsddts) (11.23) ‘CONTROLLER DESIGN FOR AD, PLANT 397 where we have replaced G(s) = Gy(s) + 4uWal)Grols)- A necessary and sufficient condition for robust performance can be obtained from the above equation in the following way: | et Tue) + WaT)" Wa] SA, VS = Je, Nal <1 = |(l- Ty) Wal < inf 10+ WaT wl V5 = Jo ) Wal < inf 10 DI J ee |(1— Ty) Wel +1WuTul <1, Ys = jo (1.24) where we have selected a worst-case || < 1 at each frequency. Equation (11.24) is a combination of both the robust stability and nom- inal performance conditions of previous sections. The main difference with a standard mixed sensitivity problem is that both terms of this equation are applied to different plants—one to Gyzo(s) and the other to a combination Of Gyo(s) and Go(s). To recast it as a standard mixed sensitivity problem ‘we apply the following transformation Le Ty(s) = (Ie) +S u(s) (1125) where Sys) © 1+ Gols) )}" isthe sensitivity funtion, The goal is ow {o design a controler that solves: ani ggg Sm COWA)| + Wels) Ta] (1126) with Wy =e" Wa(s), The factor y not only provides relative weight be- tween the nominal performance and robust stability conditions, Dut also compensates forthe extra term (1-~€""*) in equation (11.28). At the de sign stage in the next section, the value of 7 is computed so tht the robust performance condition (11.24) is achieved inthe least conservative way. 11.2.3 Controller Design ‘The mixed sensitivity problem of equation (11.26) can be solved in two ways. ‘Hoe optimal control theory can be used to design a controller that solves: Wa(s)Ta(s) ll Wals)Suls) a that is, it minimizes the norm of the transfer matrix from the input v2 10 the outputs (z1,22) in Figure 11.14, Recall from Section 48 that the optimal value of (11.27) is at most twice the value of (11.26). 398 APPLICATION EXAMPLES wile as wy cd K Lew. Figure 11.14. Mixed sonstivy probier, Alternatively synthesis can be used to design a controller that minimizes directly the value in (11.26). This is equivalent to cosy xe) —Wals)Tu(s)— Wals)Tu(s) {[ yWals)Su(s) yWe(s)Sw(s) } a the argument being the transfer function matrix between inputs (11,3) and outputs (21,22) of Figure 11.14. ‘The sensitivity Sy (s) and its complement Ty(s) are both functions of the ‘nominal minimum phase part of the plant Gy(s), which has the following state-space description: she 0 0 o 0.395 x 10° Oe nn 0 08 | ome ings) ORT gO nace 0 [*T gar xcao-e | 0 0 oo 0.166 x 10° we =[1 11 axe ‘The uncertainty weight, obtained from Figure 11.13, and the performance ‘weight from design considerations are, respectively, ) = 191 86% 1088 41 Wee 3x 10541 (129) ‘When equation (11.26) is solved using M4. optimal control, the resulting. ‘optimal solution does not necessarily achieve the optimal robust performance in (11.24), where we have a combination of minimum and nonminimum phase models. Therefore we have iterated over until the design achieves the robust performance condition (11.24) in the least conservative way. In fact, CONTROLER DESIGN FOR AD.O PLANT 399 for this case, the best performance has been obtained for the first iteration of a w-synthesis procedure, that is, the 74. optimal design.’ and a value of Youn = 1.58. For the design, we used a third order Padé approximation of the nominal, delay. Nevertheless, 10 verify robust performance in (11.24). the exact val ues of the delay have been taken into account. The plots of the robust and nominal performance and robust stability can be seen in Figure 11.15. In the same figure we can see the structured singular value 4 of the closed-loop system. We remark that, since the design has been made using the minimum phase part of the system Gyo(s), the value of is below the actual robust performance condition (11.24), which includes the delay. For the value Yaa the latter is very close to unity at w, = 2-10~ rad/s. In Figure 11.16 we can see the response of the system to an 11 hour sinusoidal disturbance of unit magnitude for 4,5. and 6 hour time delays. The frequency of the disturbance has been selected as a worst-case value according to the peak on robust performance in Figure 11.15 (w, = 2-10 rad/s, worst-case disturbance) Due to the fact that the present design optimizes condition (11.26), a 09 08| 07| o6| os| o4 03 02 oat ogee Te wiradts) Figure 11.18, Robust and nominal performance, robust stabilly, and y (forthe design sructute) “yma sandard problem, where the semsitvity and ts complement correspond tthe same mode! the performance peadoce hy synthesis design should have ben Petter than or east equal teste 0 optimal one 400 APPLICATION EXAMPLES (Oe i0 eco eeeon] 40 60 6070 80 time (hours) Figure 11.16. Time response for worst.case disturbance (tobus disturbance rejection), possible improvement could be obtained by the use of methods that con- sider directly parametric uncertainty and the fact that the system is infinite dimensional. This is beyond the scope of this example, only intended to illus trate a complete step-by-step robust control analysis and design procedure ‘on a typical process control application. 11.3 X-29 PARAMETRIC ANALYSIS In this section we perform a robust stability analysis of the Jateral-directional controlled dynamics of NASAYs X-29 aircraft. This is a true multivariable control problem due to the coupling of the roll and yaw axes. The rigid body ‘model of the aircraft with a controller in place has been taken into considera- tion and the analysis consists of computing the multivariable stability margin (or equivalently j.) for the uncertainty in the aerodynamic coefficients. Fur- ther details of this example can be found in (265). 11.3.1 Linear Model First we' introduce the notation used throughout this section ({189]) in Table 11.1, corresponding to Figure 11.17. An illustration of the aircraft is shown in Figure 11.18. X29 PARAMEIRIC ANALYSIS 401 Table 11.1. Aiteraft parameters = Pitch rate ® + Lateral ‘acceleration () + Roll rate (p) + Longitudinal acceleration (a,) Figure 11.17, Lateral acceleration (@) Reference span (0) Roll and yaw moment coefficients Lateral force coefficient Gravity acceleration (322 fy/¢) Roll, piteh, yaw axes moments of inertia (slug) Roll and yaw axis cross inertia (slug) Altitude (ft) Mass (slugs) Roll, pitch, yaw rates (rad/s) Dynamic pressure (Ib/t) Reference area (ft) Sampling period (0.025 5) Velocity (ft/s) ‘Angle of attack (rad) Sideslip angle (rad) Aileron and rudder deflection (rad) Pitch and roll attitude (rad) Rudder Elevator * Normal ‘accelerator (@n) 4 ED + Yaw ae (0) Parameters of e generic aircraft. 402 APPLICATIONEXAMPLES X29 PARAMETRIC ANALYSIS 403. 4 = pegtandsind +rtand cose (11.33) ee dy = 4 y Se (11.34) eam with the aerodynamic coefficients given by Cy = CyB +B Cr, 9, a3 : _ pb eth nse GB + C84 C1,8 + 6,20 + 2? 1136) oe eb E Cu = CyB Cag Ba+ Cn Bs CBP 4 Cy (37) The subscripts fi, ,, and 8, indicate the variables with respect to which the derivatives Cy,C), and C, are taken; for example, Cy, represents the deriva. tive of the lateral force with respect to the sideslip angle. These derivatives are computed by parameter identification procedures applied to the experi mental data. In addition, error bounds on these parameters are obtained as well ‘The linearized state-space equations, considering the input from the dif ferential aileron and flap w= (8, J” and the output y= |p r ¢ ay]? a= EArt el Bu (11.38) y = Cr+Du 2139) where the state is x=[ pr 6]!. Matrices (A,B) explicitly contain the uncertain aerodynamic coeficiems Ci, Cys Ch Cras Oy and Ca Gn, Sina cosa (4/¥)e0s0 Dace ae ane 20 7 an40) nT Figure 1.18, NASAXZ ocr Ini nr, B ly Cry M41. a GC, Cay, : y The lateral-iectional linear mode of the atralt (189) is presented next oD 4 ‘ additional matrices are B= Bey epsina—reosa+ £singeose (11.30) oe 7 v 10 0 0 Ply — Hes = GSbC) + rqlly ~ Ie) + pals 131) _|o Qo -he 0 fl a O -he hb 0 11.42) He bs ~ §80Ce+ path ~h) rah 0132) 00 04 406 APPUICATIONEXAMPLES o 1 0 0 0) Or 8 i, to c o 0 014 (143) Cry ~Za/8 Xo le 0 o 0 0 0 lol 0 14a) Cr. Cr, Here we have defined mn = GS/mV, 4 =4S/mg, gy = 45d (1145) 4 = GSP LV, n= (gC, rane) (1146) dy = aC, +4(ly— 1), da =a, #g(le—4) (11.87) dx = = (Gn, ~ Glas), dae tan 6 (a148) ‘This model has four open-loop poles: the dutch roll, a complex conjugate pair that applies to the roll and yaw dynamics; the roll mode, a real stable pole; and the spiral mode, a low frequency unstable pole. ‘The controller structure for all flight conditions is the following: (5) = Ka(s) Kyls) (i149) where Ke(s) is: [ore Mins «Mua /2+ Xan] one (1150) and the elements of Ky(s) are given by (Ky) = Pils) Sp(s) [Ke —a-Blend-L(s)-K3] (1151) (Kyra = Ks S365) -Fo(5) Ps) L(s) (11.52) (Ks)is = —Ks-Po(s)-L(s)-g- Blend/V (113) (Kyu = Ke Fels) Pus) (assy (Kyla: = Puls) Sp(5) -[Kig ~ a Blend - L(s)] (iss) (Kolar = Kir Sis) Fe(3)- Pals) L(s) (1156) (Ky)zs = —Kir- Pals) L(3) -g- Blend/V (asa) (Ky)nu = Kun Fala) Pals) (1138) where Blend is a parameter. The elements of both matrices are shown in Table 11.2 X29 PARAMETRIC ANALYSIS 405 Table 11.2. Mattices Ko(s) ane Ko(3) Aye) Fourth-order differential ap actuator model AiG) Fourth-order rudder actuator model Zs) Padé approximation due to Sample & Hold delay Pus) Roll, yaw, and lateral acceleration prefiters Pus) Roll attitude prefiter 5,18) Second-order Roll rate sensor model Sis) Second-order yaw rate sensor model FAs) Second-order yaw notch fter Fis) Second-order lateral acceleration notch filter Lis) ‘Analog equivalence of digital filter L(z) 11.3.2 Results ‘The robust stability analysis was performed over two different linearization points. The first one corresponds to the nominal design flight conditions, the second one to a critical condition at sea level and with large angle of attack. These conditions are shown in Table 11.3 ‘The mathematical structure of the controller and nominal model of the plant remain unchanged, although the parameters have different values ac- cording to the fight condition. In the controller these parameters are Kz to Kis, Xepav Xin, Xzpas and Blend; in the nominal plant model the parameters are the aerodynamic coefficient derivatives. In this last case not only do the ‘nominal values of the parameters change but also their uncertainty bounds. Before a branch-and-bound analysis algorithm is applied to this example ((266, 281]), we proceed to compute the structure of the characteristic poly nomial f (5,4) as a function of the uncertain parameters mentioned before: this is G(sa) = D+C [se 40a] 8) 1159 f(s,A) = det [I + K(s)G(s,4)] (11.60) {eble 1.3. Unwotionconatons _ No Nea Mach No. 09 04 was som seni ° oe 0 ¢ sss als : Bat a A os 2a | ssn ar 406 APPLICATIONEXAMPLES A = diag|5, 54) 161) with 6 the bounded uncertainty of parameter p, around its nominal value Pov hati, p= Pa +5, [8 < 8, A'symbolic manipulation solver has been used to determine the mathematical structure of this equation, in particular the exponents of the uncertain parameters, to determine which are repeated ‘ones. The mapping theorem can be applied to the parameters that appear linearly. On the other hand the ones raised to powers larger than one appear in a polynomial fashion in the characteristic equation and therefore should be restricted by the equality contrain's mentioned in Chapter 7, reducing them to a moltiinear dependence. in this case, the characteristic polynomial /() 's multilinear in all parameters but to, in this latter case with a quadiatie dependence. The new set of parameters generated by the adition of equality restrictions (see Chapter 7) Pi=P2 = Cip~ Poi = Peat (11.62) >= Pa = Cay = Pos + 8s = Poa + 5s (11.3) Ps = C= Post 8s (11.64) Po = Eng = Poo + 8 (11.65) b= C= Path (166) Ps = Cny = Pow + 8x (11.67), ‘The bounds on the uncertainties are 8) = 5) = 333%, 5 =3,=137,% K= 128% By = 58%, 5, = 189%, By = 14.3% (11.68) {or the fist linearization point and Bb =39%, == 111%, H=DO% iggy A= 636%, = 685%, | = 226% for the second. ‘The multivariable stability margin Km(jw) [1a,(Jw)] Was computed for a certain range of frequencies. The minimum values for each flight condition were km(Ja) = 3.15 [ia,(s0) = 0.317) at w = 3.5 rad/s in the fist case, and Kyg(Jo) = 2.89 [19, (Jen) = 0.346] at « = 10 rad/s for the second case. Thus the design is robustly stable. ‘The worst-case parameter combination, from the stability viewpoint, can also be obtained from the algorithm. This is the vector p* whose image frst reaches the origin of the complex plane for the smallest scaling factor k. This is a very important element in the analysis and redesign (if necessary) of the controller CONTROLOF A DCTOE RESONANTCONVERTER 407 This worst-case combination of parameters for flight conditions | and 2 were, inthis case, Ai = diag[) By Be 086-5) - i &) a1) ds = diag [ds 8 8-8-8. (O18-]) Be ] LTH The design of K(s) was made at both linearization points for the set of ‘nominal parameters py = py + 4y. with Ay as follows: ay = diag(—5 -& 5B -3 -& ] -&] C1172) {As observed, this disagrees with the values obtained in (11.70) and (11.71) for both fight conditions. This is a consequence of the fact that itis difficult to develop a physical iuurion on how the parameter variations will influence the stability of the system, especially in MIMO plants with a large number of uncertain parameters, As stated previously (see Chapter 7). the exact computation of ky (ja) or 443,(70) in the cases of parameter uncertainty has an exponential time compttation with the number of parameters, Nevertheless. in many cases, as the one presented here, it can be computed in a practical way by branch-and: bound procedures. Further details of the algorithm used in this ease ean be found in (266, 281). and this particular example has been developed in (265) under a NASA grant. 11.4 CONTROL OF A DC-TO-DC RESONANT CONVERTER 11.4.1 Introduction DC-to-DC resonant converters have been the object of much attention late since they have the potential to provide high-performance conversion. al- lowing for smaller, lighter power supplies. However. they require using a control circuit capable of maintaining the desired output voltage under dif- ferent operating conditions. In this section (based on the recent paper [65)) ‘we illustrate the z-synthesis technique by designing a controller for a paral- lel resonant converter. In addition to guaranteeing stability for a wide range of load conditions, this controller rejects disturbances at the converter input while keeping the control input and the settling time within values compatible with a practical implementation, 11.4.2. The Conventional Parallel Resonant Converter Figure 11.19 shows a diagram of a conventional second-order parallel res: nant converter. The combinations of the diodes and transistors form bi- directional switches operating at 50% duty ratio, Thus. in cach switching period. the resonant circuit L-C is alternatively excited by +Vy and — Vy 408 APPLICATIONEXAMPLES [ts rtp EY Figure 11 Conventional second-order PRC cireutchagram, ‘The large output inductor L,, and capacitor C, are used to minimize the load effect on the resonant capacitor voltage and to ensure the constant out put voltage through the output circuit (37). As for notation, the resistor Ry and the voltages V_ and V,, represent the load, the line (input), and the output, respectively. Throughout this section we use as nominal parameters the following val- vves, taken from the design example in Chapter 2 of (37) L = 4.18 aH C = 11300F Ry = 208.330 = 10mH C= 47 nF V, = 100V Vv, = 250V f, > 200kHz VEC = 60.390 CONTROL OF A DC-TO.DC RESONANICONVERTER 409 For convenience, we introduce the following normalized variables: v, where the resonant frequency f, 11.4.3 Small Signal Mode! Under steady-state conditions it can be shown that, for a PRC operating in the continuous conduction mode [37], there are four circuit modes in each ‘witching period. Thus the converter is a nonlinear, variable structure system, with its steady-state trajectory uniquely determined by the normalized switch: ing frequency F,, and the load condition Q,. For a given operating point, a discrete time, small signal model of the converter can be obtained by using a perturbation method. The sampling time for this discrete time model is equal to T,/2, where T, = 1/f, is the switching period. Therefore it follows that this model is correct under small-signal perturbations with frequencies up to the operating switching frequency 1, = 2nf, = 1.26 x 10®radjs ‘The discrete time model from the normalized switching frequency Fyy and the normalized line Vag to the normalized output Vp (10 simplify the nota- tion, we use the same variables for both the steady state and its perturbation) at the nominal operating point is given by the following state-space rcaliza- tion [37|: X(k +1) = AX(k) + BUR) Volk) = CX(K) where 0.2767 0.6108 0.6644 0.0053 0.0075 0.9387 [ 0.8219 0.5504 -2 “| 6.4684 0.4834 B= | 106774 1.9499 = 0.0002. 0.0162 410 _-APPUCATION EXAMPLES C= |0 0 345] The state variables and inputs are defined as X= [iath) vad) tal)" UCR) = [Fal Vag] where iy, 3nd Jy, are the normalized resonant inductor current, capacitor voltage, and output current, respectively. 11.4.4 Control Objectives Figure 11.20 illustrates the diagram used for control design. In the small signal model of the converter there ate two inputs: line voltage Vy and switching frequency F,,. The objective is to synthesize a controller having 4 input the error signal (obtained by comparing the output voltage versus the reference input r) and as output the switching frequency, F,, such that the output voltage is kept at a prescribed level (in our case V,, ~ 250 Vie. Vo = 2.5) at all operating points, This problem can further be divided into four parts ‘© Line Regulation (Nominal Performance). The line voltage is often unreg, tlated and could have a substantial range of variation, typically around + 20%. This variation will be modeled as an external disturbance, thus leading to a disturbance rejection problem. Performance specifications for this type of problem are usually given in terms of time-domain quan- tities, such as: 1. Zero steady-state error 2, Small overshoot at output (usually less than 10% for reference input step response). 3. Appropriate settling times for both fine and reference input step responses (5 ms at most in our case). Line Volage 7 CONVERTER ouput Switching Frequency Reference. ——] CONTROLLER Figure 11.20. the dlogrom tor contol design CONTROL OF ADCTODC RESONANT CONVERTER 411 4. A closed-loop bandwidth of at least 360 Hz. in order 1o successfully suppress line ripple. © Load Regulation (Robust Stability). Ow the other hand, the load con- dition could also vary over a wide range. Since the load R,, enters the ‘dynamics of the model, load variations will appear as model uncertainty and could possibly lead to stability problems. Normally the load changes from 10% at low load to 90% at full load condition, Other model uncer tainties, such as unmodeled high-frequency dynamics and uncertainties in the resonant inductor £. and capacitor C, will also be considered, Robust Performance. Since the converter operates over a wide range of load conditions, the performance requirements must be satisfied at all operating points. This is equivalent to requiring satisfactory response under both line and load variations. 11.4.5 Analysis of the Plant Control Characteristics For a PRC converter operating under steady- state conditions, the input-output relationships can be represented by the control characteristics curves, relating the output voltage to the load and switching frequency. Given any two variables among the normalized output Vo, Switching frequency ratio Fy,, and output load Q,. the third variable can be determined from the curves, Thus these curves allow one to casily visualize the effects of the switching frequency and load on the converter ‘output. From a control point of view, the control characteristics curves allow us to make an initial estimate of the load change that can be tolerated and to see some of the difficulties inherent in the load regulation problem. Figure 11.21 shows the control characteristics curves for various output loads Qp, obtained analytically from the steady-state analysis. To maintain the output voltage constant in the presence of perturbations, the controller should adjust the switching frequency to keep the converter operating along the dashed line indicated in Figure 11.21. As the converter is perturbed away from the nominal operating point (marked with an asterisk in the figure) the plant dynamics may vary significantly, resulting ina difficult control problem, Remark Note that from Figure 11.21 it follows that at lighter loads (higher Ry, larger Qp, and lower J,), a small frequency change will result in larger output changes. Thus we should expect that the control problem will become more difficult at larger Q, values, In the next section we will show. through a frequency-domain analysis, that this is exaclly the case. Frequency Responses From the discrete time state-space model, we can easily get the z-transfer functions from the normalized switching, frequency 412 APPUCATIONEXAMPLES 19-— O5 055 06 065 07 075 08 085 09 095 1 Fre Figure 11.21. The conventional PRC contol characterises curves Fax and the normalized line input Vrg to the normalized output Vn Vno(2) Vnol2) Fus(2) Vrg(2) | ~ [o02) Ge(2) Ct-ap"B (11.73) Following a common approach, we will carry out the analysis of the plant and the synthesis of a digital controller using a w-plane approach. To this ‘effect, the bilinear transformation _Lash/4 ran (7) is used to obtain the transfer functions in the frequency domain s. These transfer functions, still denoted as G(s) and G,(), are given by (5+ 798041)(5 ~ 792481)(s — 800003) 2.652 x 10-2 + 7950415 = 792431)(s — 800003) ot (5 +29167)(s + 83363 + 202487i) Coe) (s = 800003)(s + 484950 + 2902534) Gals) = 136710 poreny(s R365 20D8T|) LT) as) ‘The above transfer functions correspond to the nominal load R, = 208.330. AS stated before, since the load enters the dynamics of the converter, load variations result in different transfer functions, Figures 11.22 and 11.23 show ‘CONTROL OF A OC-TO-OC RESONANT CONVERTER 10 104 NOE ak at aa at Al yor 1020 10* tPF C10? wirads) Figure 11.22, Frequency responses G(s) a ferent load conditions. ior igs i iv wieadls) Figure 11.28. Frequency responses G(s) at diferent load conditions. as. 414 APPLICATIONEXAMPLES the frequency responses of G(s) and G,(s) corresponding to several different load conditions, respectively. These figures show that as the load becomes lighter (larger R,), the over shoot inereases, leading to a more difficult control problem. This conclusion is consistent with the conclusion drawn in the last section from the study of the control characteristics. On the other hand, the control characteristics require that Q» be greater than Vqp in order to get the prescribed output voltage. Since in our design example the value of Vn» is chosen to be 2.5, it follows that R, should be greater than 151. Therefore, in the sequel, we will assume that R, varies within the range 151 to 1200 0. 11.4.6 Control Design ‘As mentioned before, our goal is to design a controller that satisfies the performance specifications listed in Section 11.4.4 for all load conditions in the range 151.0 < R, < 12000, assuming that the values of the components of the resonant tank are known within a 10% tolerance. In the sequel we solve this problem by recasting it into a robust performance synthesis form and using synthesis. To this effect we need first to describe the family of plants corresponding to different values of the load as a nominal plant subject to uncertainty. Plant Description and Uncertainty Weight Selection In this exam- ple we choose to model the uncertainty caused by load variations by using a single, norm-bounded, multiplicative uncertainty that covers all possible plants. Let G¥(s) and G(s) denote the transfer functions from the control input and line input to the output at operating points other than the nominal point (Ry 208.332), respectively: In the sequel we represent these transfer functions as GMs) = Gls) [t+ 8/(5)Wi)] au) GPe(s) = Gas) [1 + Bg(5)65)] (11.78) where G(s) and G;(s) are the nominal transfer functions given in equations (11.75) and (11.76), respectively, W;(s) and We(s) are fixed weighting fune- tions containing all the information available about the frequency distribution of the uncertainty, and 4y(s) and A,(s) are stable transfer functions represent- ing model uncertainty. Furthermore, without loss of generality (by absorbing any scaling factor into W,(s) and W(s) if necessary), it can be assumed that [Sr(s)lx <1 and liMe(s)[le <1. Thus W;(s) and W,(s) are such that their respective magnitude Bode plots cover the Bode plots of all possible plants. Some sample uncertainties corresponding to different values of the load R, are shown in Figures 11.24 and 11.25. We can see that in both figures the ‘multiplicative uncertainties have a peak around the resonant frequency. This Peak becomes larger and steeper as the load resistance R, increases. Magnitude Figure 11.24, Multiplicative uncertainties (contol output and weight Magnitude 108 ‘CONTROL OF A DC-IO-DC RESONANT CONVERTER 10 104 107! 10" Cn a a nT) wiradis) 10 10 108 10F 10 wiradis) Figure 11.25, Mutipicatve uncertainties (ne fo output) and weight as 416 APPUCATION EXAMPLES Based on these plots, the following multiplicative uncertainty weights were chosen for control design: 4x 10-45 +065 Wy) = AXIO 06 (1179) 104s Wo) = ee (1180) The magnitude frequency responses of W,(s) and We(s) are also shown in Figures 11.24 and 11.25, respectively. These figures clearly show that attempt- ing to cover the sharp peak around the resonant frequency will result in large gaps between the weight and the uncertainty at high frequencies, introducing conservatism at that frequency range. On the other hand, a tighter fit at high frequencies using higher-order functions will result in high-order controller. The weights (11.79) and (11.80) used in our design provide a good trade-off between robustness and controller complexity. ‘We turn our attention now to the effects of changes in the values of L and CC, the resonant tank components. Since these changes affect the location of the resonant peak, they could conceivably destabilize any controller design based on its nominal location. Figure 11.26 shows the changes in the trans: fer functions due to +10% changes in the values of L and/or C. It is worth noticing that our choice of weighting functions W; and W, will also cover this family of plants, even at the extreme load conditions Ry = 12000 and 151.0. Thus a robust controller designed using these weighting functions will be able to accommodate both changes in the load condition and uncertainty in L and C. Figure 11.27 shows a block diagram of the converter, taking into account the uncertainty. Here A; and Ay are scalar blocks, representing the model uncertainty perturbations from the control and line inputs, respectively, and W; and We are the corresponding uncertainty weights. Performance Weight Selection As discussed in Chapter 7, in order to guarantee robust performance we need to add to the structure shown in Figure 11.27 an additional (fictitious) uncertainty block 4p, along with the corresponding performance weights W, and W,, associated with the track ing/regulation error and the control effort, respectively, resulting in the block diagram shown in Figure 11.28. The selection of the performance weights W, and W, entails a trade-off among good regulation versus peak control action. The weight on the control error W,(s) is usually selected to be very large at low frequencies in order to get good tracking and regulation, Additionally, the order of the weights should be kept low in order to keep the overall controller complexity low. A g00d compromise between performance and compleaity is given by weighting CONTROL OF ADC-TODC RESONANTCONVERTER 417 Control to Output fe eee ee 107 10? 109 10% 10° 108 107 10% wads) Line 1 Output 10? us —— 10! ww fo 5 F151, L-9THH,C=t00F 20? 10 xo! 10) = = “107 10 70 “6 79 10! wired) Figure 11.26. Uncertoinies due to 410% changes of L ond/or C of extreme loos conditions Re = 120010 and 155, 418 APPUCATIONEXAMPLES Figure 11.27. the block diagram of the converter including the uncertainty due to 00d and component variations. functions of the form W.(s) (181) BSA where A is the desired steady-state error (A will be zero if zero steady state is required); T, approximately determines the bandwidth (w, = 1/T) and Figure 11.28. The block diagtam for synthe ‘CONTROL OF A DC-IO‘OC RESONANTCONVERER 419 hence the rising time and settling time; where the ratio 73/7; is associated with performance requirements against high-frequency noise ((188]). Note that there is no exact relationship between the parameters 7; and T> and time-domain performance specifications given in terms of rise-time, setling: time, and overshoot. The design of multiobjective robust controllers subject to both time- and frequency-domain specifications is, to a large extent, an ‘open problem, although some progress has been made recently [301] In this particular design example we selected the following weights: w= ae ns 104s we) = cay The weight on the control input W,(s) was chosen close to a differentiator to penalize fast changes and large overshoot in the control input, The weight W(s) guarantees zero steady-state error. The frequency responses of W.(s) and W,(s) are shown in Figure 11.29. The closed-loop bandwidth correspond: ing to this choice of weights is approximately 1/0.0006 = 1700rad/s. Log Magnitude 10° cr 104 Frequency (radiansisec) Figure 11.28. Performance weights We(s) and Wu(s) 10 420 APPUICATIONEXAMPLES 11.4.7, Controller Synthesis Combining the uncertainty description with the performance weights yields aan uncertainty structure 4 consisting of two scalar blocks (corresponding to the robust stability requirements) and a 2x2 block (corresponding to the ro- bust performance requirements). Since the A structure has only three blocks, the upper bound of 1, inf pep IDM D> |, coincides with its exact value. The robust controller was synthesized using the . Analysis and Synthesis Toolbox (20), applied to the block diagram shown in Figure 11.28. After four D-K itera: tions with third-order D-scalings, we obtained a 18th-order controller yielding veep = 0.9721. Finally, Hankel norm model reduction yielded a sixih-order controller with virtually no performance degradation (size = 0.9753 <1). ‘The state-space description of this reduced-order controller is given by K=Oyst~A)"'By + De (11.84) where 025 1008 -1as¢ nats nis 1.296, 1708 1320045 998e+4 3190045 1460044 —2213045 Aga | HIS 9980044 —16%e+4 S2MMeS 1200004 1.983665 w=] chats 34mess “320805 “2erdews 172905 4053045, FUN6L 14600 +4 1200644 hm8645 -2664e83 LOIBer S 1296 2213e05 “LBBerS “408305 “101965 “RORSe rs BY ~(-2338 7.983 5.345 —6.610 0.543 6.060) 10-2 G.=(-0.935 3.193 -2.138 2.644 0.217 2.424) x 10 De=0 In order to benchmark the performance of the robust controller, we also designed a phase-lag controller using classical design tools, based on the plant frequency responses at the various operating points shown in Figure 11.22. To improve performance, this controller was further tuned by trial and error. ‘The transfer function of the final controller is given by 0.025 + 200, 3402 The frequency responses of both the 4. and the phase-lag controllers are shown in Figure 11.30, Both controllers have similar responses at low fre- quencies, while at high frequencies the gain of the u controller decays faster im order to accommodate the model uncertainties at high frequencies. Figure 11.31 shows the closed-loop frequency responses for the nominal plant and for the lightest load considered in the design. Note that in both cases the y controller provides lower gain and better roll-off at high frequencies. Moreover, while the response corresponding to the phase-lag controller is acceptable for the nominal plant, it exhibits a large peak at the resonant frequency at light loads. As we show next, this peak results in significant performance deterioration at these loads. K, = (115) CONTROL OF A DC-TODC RESONANTCONVERIER 421 Log Magnitude 10 1 oe 10 rg 108 Frequency radianssec) 9 7 = & so é 09 50nd 10 10 10 10 6 10 Froquency (rasianaee) Figure 11.20. Frequency responses o 1 (solid) and phase-og (dashed) controls. 11.4.8 Simulation Results Linear Simulations The closed-loop system corresponding to the j. con- troller was simulated at the nominal operating point R, = 2080) and at two ‘extreme cases R, = 151 and 12000, using the corresponding linear model of the plant. The time responses 10 20% step change in line voltage Vag and reference input r are shown in Figure 11.32, For the nominal case R, = 2080, the settling time is about 2.5 ms for both line voltage change and reference input change. The output responses are satisfactory since the settling time is smaller than the required 5 ms, with ro overshoot. The control action in these responses is also adequate, without overshoot or abrupt change. This is due to the choice of the weight Wi, penalizing fast changes and overshoots in the control action ‘When the operating point moves to R, = 151.0, the setting times for both step changes are about 4 ms, This increase is due mainly to the significant decrease in plant static gain (see Figure 11.22). The u controllers undertuned at this operating point in order to achieve robust performance. When the operating, point moves toward lighter loads, the responses are almost the same as for the nominal plant, except that for the case R, = 12000 (note 422 APPLICATIONEXAMPLES 10° — _ : ~ ~~ Phase-iag contalier Gain 104 p-controlle oe 108 10 10° Frequency (rad) ° Phase-ag controller /\ 1ot w-conteoller 108 10° 10" 10 10° Frequency (ad) Figure 11.31. Closedoop trequency rospontes for x (sold) ond phaselag (dashed) controllers: (4) Ry = 20881 and (0) Re ~ 1200, CONTROL OF ADCO-0C RESONANTCONVERIER 423 409) aa x10 085 oe 0G Time (582) 10? Time (see) 492 @) (b) Figure 11.82. Linear simulation resus with 4. conkolier at diferent operating points Fo = 208 1 (oid), 181 12 (dashed), and 1200 1) (does): (a) relerence input stop change (20%) and (bine vollage step change (20%) that this load is the lightest Joad considered in our design). some chattering in both the output and control input starts to show up at the beginning of the responses. The occurrence of the chattering is linked to the large peak in the plant frequency response at lighter loads, barely covered by the uncertainty weights W; and Wy From the simulation results it follows that the controller achieves robust, performance, since all performance specifications are satisfied at all operat: ing points of interest. However, significant variation of performance is also observed. This isa direct result of the large variation in the plant dynamics. and any fixed linear controller can do very litle in this respect, To reduce ‘this variation will require using a nonlinear, gain scheduling controller. The same simulation was performed for the closed-loop system corre sponding to the phase-lag controller. The time responses to 20% step change im line voltage Vy and feference input r at three different operating points — R, = 1510, 2080, and 1200—are shown in Figure 11.33. They are 424 APPLICATIONEXAMPLES om 07 aso ano _ 5 5 aso | asl 0 0) aay 0 1 2 3 4 rece) gs : —— om a 2 34 2 3 Time (660) y Tere 600) yo! (a) (b) Figure 11.38. Linea simulation resus wih phaselag controler at diferent operating pins Re = 208 0 (soli), 181.0 (dashed), ond 1200 £1 (coted): (a) reference Input step. ‘change (20%) ond (b) ne volage step change (20%), similar to the responses with the 1 controller except that the performance is far worse for Ry = 12000. This is due to the phase-lag controllers inability to provide enough attenuation to counteract the increment in the magnitude of the resonant peak of the plant at heavy loads, as shown in Figure 11.31. Furthermore, as shown in Figure 11.34, at R, = 240010 the phase-lag con- troller fails to stabilize the system, while the controller can still produce acceptable performance. Nonlinear Simulations and Validation While simulations using the linearized model of the plant corresponding to different load conditions can usually provide an approximate evaluation of load regulation performance, this is usually insufficient to assess the performance of a highly nonlinear system such as the converter. Thus to further validate the controller, a non- linear simulation of the PRC circuit was performed using P-Spice. Figure 11.35, shows the responses due to reference input and line voltage step changes. Note that these results are similar to those obtained using a linear simulation CONTROL OF ADCIO.DC RESONANICONVERIER 425, 1 har om; Oem oa 0 foo 3 od aa rr os Cectmenianate aes ee nes) yg? me ees) yg? Input 2 Opt Int 22 4 eos 0s é q $4 $004 4 i a Eo q ons| oi se 8% Os se Tine 08 49? Time ane ag? ® 4 4 co a 2] = a ce lteeas Te see9} 40? “we sees) ap? 10" _Iput2 vet ‘to? _ hpi 2 Ou? «poe meen) , E os| il i a} 2 2 os ay Tre sos) 9? Figure 11.34, Stop responses of he closecioop systom at = 2400.2: (6) 4 contoler (slab) and (©) phoseag controller (unsiabi), 426 APPLICATIONEXAMPLES 220) 200] S280 260) 2405 SO nce ses) Time (560) 97 095; og toss 03] 88 Time ($20) 49? Time (see) 30? (a) (by Figure 11.38. Nonlinear simulation resus wih. contoler a ferent operating points p= 208 1 (sold), 181 11 (dashed) and 1200 11 (coted) s: (a) reference input sep ‘change (20%) and (bine vollage sep change (20%). as shown in Figure 11.32. In the responses to line voltage step change and reference step change, settling times are slightly larger than those in the li car simulation, The chattering observed in the output voltage is substantial, due to the periodic switching behavior of the converter (periodic charge and discharge of capacitor). i ro i ‘6 fu bi BIBLIOGRAPHY ‘Abdallah C. "On the robust control of uncertain nonlinear systems,” Proceed ings American Control Conference, San Diego, 1990. Abdallah C., Dawson D., Dorato P, Jamshidi M, “Survey of robust control for rigid rob0W," IEEE Control Sysiems Magazine, Vol 11, NO. 2, 1991 ‘Abedor J., Nagpal K., Poola K.,"A linear matrix inequality approach to peak to peak gain minimization,” International Journal of Robust and Nonlinear Con: ‘rol, Yo. 6, pp. 99-927, 1996. ‘Ackermann J, "Parameter space design of robust control systems,” IEEE Trans actions on Auiomatic Control, Vol 25, pp. 1058-1072, 1980. ‘Ackermann J., Robust Control: Systems with Uncertain Physical Parameters Springer-Verlag, New York, 1993, ‘Adamjan VM,, Arov DZ., Krein MG., "Analytic properties of Schmidt pairs for a Hankel operator and the generalized Schur-Takagi problem,” Math USSR Sbornik, Vol. 15, pp. 31-73, 1971 Adamjan VM. Arov D.Z., Krein M.G., “Infinite Hankel block matrices and related extension problems,” AMS Translations, Vol. 111, No. 133, 1978, Akcay H., Gu G., Khargonekar P, “Identification in ., with nonuniformly spaced frequency response measurements," Proceedings American Control Con- ference, Chicago, 1992 ‘Akcay H., Gu G., Khargonekar P, “A class of algorithms for identification in Ex: continuous-time case,” JEEE Transactions on Automatic Control, Vol. 38, No. 2, 1993, ‘Anderson B.D O., VongpanitlerdS., Network Analysis and Synthesis: A Modern Systems Theory Approach, Prentice-Hall, Englewood Clif, NI, 1973. ‘Anderson B., Kraus F, Mansour M., Dasgupta S., “Easily testable sufficient conditions for the robust stability of systems with muliaffine parameter de- pendence,” in Robustness of Dynamic Systems with Parameter Uncertainties, M Mansour, S. Baler, W. Trudl, Eds. Birkhauser, Berlin, 192 427 428 BIBLIOGRAPHY 12] Anderson B.D.O., Moore. JB. Optimal Conrot: Linear Quadratic Methods, Prentice-Hall, Englewood Cis, NI, 1990. 113) Astrom K., Eykhoft P, “System identication: survey.” Automatica, Vl 7, pp. 123-162, 197. [14) Astrom K. Witenmark B.. Computer Controled Systems: Theory and Design, Prentice-Hall, Englewood Cli, NJ, 1984 115) Athans M., Falb P, primal Control: An Inroduction 1 the Theory and Is ‘Applications, MeGraw Hil, New York, 1966, [16] Athans M, KapasourisP, Kappos E, Spang H.A. IM, “Linear quadratic gaus- Sian with Toop transfer recovery methodology for the F100 engine.” AIAA Journal of Guidance, Control and Dynamic, Vo. 9, No. 1, 1986 17] Balakrishnan VS., Boyd S., Balem S, “Branch and bound algorithm for com- Puting the minimum stability degree of parameter dependent linea systems” ‘Technical Report. Information Systems Laboratory Stanford University, 1991 (18) Balarishaan VS.. Boyd .P, “On computing the worst case peak gain of linear systems," Systems and Controls Leters, Vo. 19,192 119) Balas G., Doyle J.C, "Identiication of flexible structures for robust contol,” IEEE Comiro Systems Magazine, Vo. 10, No.4 199. [20) Balas G., Doyle .C. Glover K, Packard A, Smith R, p-Analysis and Syuess Toolbox. The MathWorks Ine, Natick, MA, 1991 [21] Ball.A., Gotberg I, Rodman L, Interpolation of Rational Matrix Functions, Operator Theory: Advances and Applications, Vol. 4, Birkhauser, Beri, 1990 [22] Ball 1A, Helton J.W, Walker MLL, "7, control for nonlinear systems wth ‘output feeaback.” IEEE Transactions on Automatic Control, Vo. 38, No. 4 1998, [23] Bancroft A.R, “Heavy water G-S process research and development achiewe- ments,” Atomic Energy Canada Limited (AECL) Report No. 6215, 197. 24} Banks HLT, Burns).A., “Hereditary contol problems: numerical methods based ‘on averaging approximation,” STAM Journal of Conirol and Optimization, No. 16, pp. 169-208, 1978 [25] Banks FLT, Kappel F, “Spline approximations fo functional differential equa tions,” Journal of Differential Equations, Vo. 34, pp. 496-822, 1979. [26] Barabsnov AE. Granichin OLN. “Optimal controler fo afinear plant with bounded noise,” Aulomation and Remote Control, Vol. $, pp. 39-46, 1984 [27] Barmish BR., “Invariance of the strict Hurwite property for polynomials with perturbed cosficients," IEEE Transactions on Automatic Control, Vo. 29,1984 {First presented atthe 1983 Conference on Decision and Control, Tex.) Barmish B. Khargonckar PP. “Robust stability of feedback control systems sth uncertain parameters and unmodelled dynamics.” Proceedings American Control Conference, Georgia, 1988 (29) Barmish B R, “New tools fr robustness analysis," Plenary Session, Proceedings Conference on Decision and Control, Austin, Texas, 1988. Barmish B, Ackermann JE, Hu HZ. “The tee structured decomposition: a new approach to robust stability analysis,” Proceedings Confrence on Informa ton Setences and Systems, Johns Hopkins University, Baltimore, 1989 es} [p9) BRUOGRAPKY 429 [31] Barmish B.R., "A generalization of Kharitonov’ four polynomial concept for ‘robust stability problems with linearly dependent coefficient perturbations,” IEEE Transactions on Automatic Control, Vol. 34, No, 2, 1989, [22] Barmish B., Khargonekar PP, Shi Z.C., Tempo R., "Robustness margin need not be a continuous function of the problem data,” System and Control Letters, Vol. 15, 196. [53] Barmish B.R,, ShiZ., "Robust stability of class of polynomials with coefficient depending mulilinearly on perturbations,” JEEE Transactions on Automatic Conirol, Vol. 35, No.9, 1990, [34] Barmish B., Kang H., “A survey of extreme point results for robustness of control systems,” Automatica, Vol. 29, No, 1, 1993, [35] Barmish BR. New Tools for Robusiness Analysis, Memillan, London, 1994, [36] Bartiet A.C, HollotC, Lin H., “Root locations of an entire polytope of poly ‘nomials it suffices to check the edges.” Mathematics of Control, Signals and Systems, Vol. 1, No.1, 1988. [57] Batarseh L., Steady Site Analysis and Design of High Order Resonant Convert rs, PhD. Thesis, University of Illinois, Chicago, June 1990 [38] Beck C., Doyle JC. “Mixed s2 upper bound computation,” Proceedings Con: {ference on Decision and Control, Tucson, 1992 [89] Bertsekas DR, Rhodes LB., “On the minmax reachability of target set and target tubes,” Automatica, Vo. 7, pp. 233-247, 1971 (40) Bersekas DP, “Infinie-cime reachability of state-space reglons by using feed: back control," JEBE Transactions on Automatic Control, Vol. AC-17, No. 4, PP. 604-613, 1972 [a1] Blanchini F, Sanaier M., “Rational £” suboptimal compensators fr continuous- time systems.” JEEE Transactions on Automatic Control, Vol. 39, No. 7, pp. 1487-1492, 1994, [42] Blanchini, F, “Non-quadratic Lyapunov funetion for robust contra” Automat ‘ca, Vol 31, No. 3, 451-461, 1994 (43) Blanchini F, Sznaier M. “Persistent disturbance rejection via static state feed back," IEEE Transactions on Automatic Control, Vol. 4, No. 6, pp. 1127-1131, une 1995, [44] Becker N., Grimm WM. “On £; and £,, stability approaches for the robust control of robot manipulators,” IEEE Transactions on Auomatic Control, Nol 33, No.1, 1988. (45) BetczynskiB., Kale 1, Cain G.D., “Approximation of FIR by IR digital filters: an algorithm based on balanced model reduction,” IEEE Transactions on Signal Processing, Vol. 40, No. 3, 992. {46] Bernstein, D., Wie, B., Organizers, Invited Session on “A Benchmark Prob- lem for Robust Ms / 7. Control Design,” Session WP-1S, Proceedings of the American Control Conference, San Diego, pp. 961-973, 1950. [47] Bhattacharyya SP, Chapellat H., Keel LH, Robust Convol: The Parametric Approach, Prentice-Hall, Englewood Cliff, NJ, 1995 (48) BialasS.,“A necessary and sufficient condition for stability of interval matrices, International Journal of Control, Vol. 37, pp. 117-122, 1983, 430 19 150) 154] [ssi 6 1371 Iss 9 (60) (on (62) (63) (4), (6s) (6s) 67) BisLIOGRAPHY DBialas S. Garloff J, “Stability of polynomials under coefficient perturbations,” IEEE Transactions on Automatic Control, Vol. 30, 1985. Biernacki RM., Hwang H., BhattacharyyaS-P, "Robust stability with structured real parameter perturbations,” JEEE Transactions on Automatic Control, Vol. 32, 1987 Blondel V. and Gevers, M.. "Simultaneous stabibzation of three linear systems is rationally undecidable,” Mathematics of Control, Signals and Systems, Vol. 6, pp. 135-145, 1993, Bode H.W, Network Analysis and Feedback Amplifier Design, Van Nostrand, Princeton, NJ, 1945, Bodson M., Chiasson J.N., Novotnak RT, Rekowski R.B., “High-performance nonlinear feedback control ofa permanent magnet stepper motor,” IEEE Trans. actions on Control Systems Technology, Vol. 1, No. 1, 1993, Bose NK, “A system theoretic approach to stability of sets of polynomials,” Contemporary Mathematice, Vol. $7, pp. 25-38, 1985, Boyd S., Desoer C.A., “Subharmonic functions and performance bounds in linear time invariant feedback systems,” IMA Jounal of Mathematics, Control and Information, Vol 2, pp. 153-170, 1985. Boyd SP, Doyle J.C, "Comparison of peak and RMS gains for discrete-time systems,” Systems and Controls Lentrs, Vol 21, 1983. Boyd S. Balakrishnan V., Kabamba P, “A bisection method for computing the ‘Yo norm of a transfer matrix and related problems.” Mathematics of Control, Signals and Systems, No, 2, 1989. Boyd S,,Barrat CH., Linear Controller Design Limits of Performance, Prentice: Hall Information and System Sciences Series, Prentice-Hall, Englewood Clif, NJ, 1991, Boyd S., EI Ghaoui L., Feron E., Balakrishnan V, Linear Matrix Inequalities in Spstems and Control Theory, SIAM Studies in Applied Mathematics, Vol. SIAM, Philadelphia, 1994 Braatz R.D., Young PM., Doyle .C., Morari M. "Computational complexity of ju calevlation,” IEEE Transactions on Automatic Control, Vol. 38, No, 5, 1998. Brockett RW, Finite Dimensional Linear Systems, Wiley, New York, 1970. Bruck J, Blaum M., “Neural networks, errr-correctng codes and polynomials over the binary r-cube,” IEEE Transactions on Information Theory,No.9, 1989. Bryson AE,, Ho ¥.C,, Applied Optimal Control, Hemisphere Publishing, New York, 1975. Bu J., Sznaier M,, Holmes M., "A linear matrix inequality approach or synthe- sizing low order ! controllers,” Proceedings 38th IEEE Conference on Decision and Cantrl, Kobe, Japan, pp- 1875-1880, 1996. Bu J., Senaier M. Batarseh 1, Wang Z.Q., “Robust controller design for @ parallel resonant converter using, w-synthesis.” [EEE Transactions on Power Electronics, Vol. 12, No. 5, pp- 837-853, September 1997 Bu J., MultObjectve Robust Control: Design of Mixed €°/P,. Controllers, PhD. Dissertation, The Pennsylvania State University, 1997. Burns J, Peichl G., “On robustness of controllability for nite dimensional approximations of distributed parameter systems,” Technical Report. Dept. of (63) (69 (70 m 7) 73) (74) i (76) m (78) 9) (80) isi (82) 83] aauocRarHy 431 Mathematics. Virginia Polytechnic Institute and State University. Blacksburg, 1985, Chapeliat H., Bhattacharyya SP. °A generalization of Khatitonov’s theorem: robust stability of interval plants." IEEE Transactions on Automatic Control. Vol 34, No. 3 1989 CChapeliat H.. Bhattacharyya S.P, “An alternative proof of Kharitonov’s theo- rem." IEEE Transactions on Automatic Conwrol, Vol. 4. No. 4.1989. CChapellat H, Dahleh M. Bhattacharyya SP, “Robust stability under structured ‘and unstructured perturbations." IEEE Transactions on Automatic Con. Vol 35, No. 10, 1990 Chapellat H., Dahleh M, Bhattacharyya SP, “Robust stability manifolds for ‘multilinear interval systems.” IEEE Transactions on Automatic Conirol. Vol. 38 No. 2, 1993, Chen CT. Linear System Theory and Design, Holt, Rinehart & Winston. New York, 1984, Chen J., Fan M. Nett C, “The structured singular value and stability of un- certain polynomials: @ missing link,” ASME Annual Wier Meeiig, Atlanta, 1991, Chen J. Nett C. Fan M, “Worst-case system identification in W.: validation of priori information, essentially optimal algorithms. and ertor bounds.” Proceed: ings American Control Conference, Chicago. 1992; also in IEEE Transactions ‘on Automatic Control. Vol. 40, No. 7. 1995 Chen J. Nett C, Fan M.. “Optimal non-parametric system identification from arbitrary corrupt finite time series: a control-oriented approach.” Proceedings “American Conirol Conference, Chicago, 1992. Chen J, Farrell, Nett Zhou K.,"% identification of multivariable systems by tangential interpolation methods.” Proceedings Conference on Decision and Control, Florida, 1994 Chen J., Nett C., “The Carathéodory-Fejér problem and the 1. /t! identi cation: 9 time domain approach,” JEEE Transactions on Automatic Control Vol. 40, No. 4, 1995 Chiang, R, and Safonov, MG., “Ha. Robust control synthesis for an undamped. hon-colocatedspring-mass system.” Praceedings American Control Conference. PP. 966-967, 1990. Curtain R., Glover K., “Balanced realisations for infinite dimensional systems.” Workshop on Operator Theory and lis Applications. Amsterdam. 1985, Dahleh MA., Pearson J../-Optimal controllers for discrete time systems, Proceedings American Control Conference. Seattle. pp. 1964-1968. 1986, Dahlen M.A.. Pearson J. "f"-Optimal controllers for MIMO discrete time sys tems.” IEEE Transactions on Automatic Control, Vol 32. No. 4.1987 Dahlen M.A,, Boyd Pearson J. °'-Optimal compensators fr continuous time systems.” [EEE Transactions on Automatic Conirol, Nol. 32, No. 10. pp. 88— 895, 1987, Dahlch MA.. Pearson J.. ‘Optimal rejection of persistent disturbances. robust stability and mixed sensitivity minimization.” FEEE Transactions om Autonnatic Control. Nol 33, No. 8, 1988. 432 BIBLIOGRAPHY WN) Dableh M.A.. Ohta Y.,"A necessary and suflicient condition for robust BIBO. stability” Systems and Control Lens. Vel, 11. pp. 271-275, 198 {Ks} Dalen M.A... TheodosopoulosT. Tsitsiklis J, ~The sample complexity of worst ease [dentcation of FLR. linear systems,” Systems and Control Leter. Vo. 20. Dp. 157-166, 1993 'Ko) Dahleh M.A.. Khammash M.. “Controller design for plants with structured un certainty.” Automatica, Vol 29. No. 1. pp. 37-56. 193. {87| Dahleh M.A.. Diaz-Bobillo LJ... Control of Uncertain Svstems: A Linear Pro gramming Approach, Prentice-Hall, Englewood Cis. NI 1995 (88) Fu M.. Dasgupta S.. Blondel V.. “Robust stability under a class of nonlinear parametric perturbations.” IEEE Transactions on Auwomatic Consrol, Vol. 4. No, 2.1985, [89] D’Auetis CE. ouroduccién a los Sistemas Nolineales de Control y sus Aplica (a&b) € F. Operation & is associative, that is, (@& b)&e= a&(b&C) = akb&e for a,bceF Operation & is commutative, that i, a & b = b &a for a,b F. Set F contains the neutral element ng with respect 10 &, that is, there exists ng © F such that (a &ng) =a for all ac F 44s 446 MATHEMATICAL BACKGROUND 5. Set F contains the inverse element al, with respect to &, that is, for all ae F, there exists aly € F such that (a8 al.) = ne Set F is closed with respect to Operation + is associative Set F contains the neutral element n, with respect to * Set F contains the inverse element al with respect t0 Operation + is distributive with respect to &, that is, (a& b) we = (a & (x0) for a,b,c F. Take, for example, set R equipped with operations (+, x) as (8, +) respec: tively, then mg =0.,= 1, ah = ~a, and al = a7! (a0), A.1.2. Linear Vector Space Definition A2._A set V is a linear vector space over the field (F,+,*) if and only if the following properties are satisfied: (Im the sequel the elements of F and V will be called scalars and vectors, respectively) Set V is closed with respect to + Operation + is associative in V. Operation + is commutative nV. Set V contains the neutral element with respect 10 + Set V contains the inverse element with respect 10 + V is closed with respect to operation x between scalars and vectors. Operation x among scalars and vectors is associative in the scalars, that is, (ax b) xv =ax (b xv) =ax bx for abe F and vey. 8. Disiributive I: (a+b) x v = (a xv) +(b xv) for a,b € F and ve ¥. 9. Distributive 2: (u+v} x a= (ux a) +(v xa) fora € F and uv € V. 10. Field F contains the neutral element of operation x between vectors and scalars, that in x0 =v form. € F and v€ V. In the sequel, equivalently. inear vector space, linear space, or vector space will be used Definition A.3_ The set of vectors {04,...,9} in the linear space V, defined over the field F, is linearly independent if and only if the linear combination ev bo ay = a = 0, with a, € Fi =1,...,m. Other wise they are called linearly dependent, Definition A.4 A set of linearly independent vectors {Uy,..-,tn} of @ linear space V constinue a base of V if and only if any v € V can be represented ‘as a linear combination of these vectors. The dimension of V is the minimum number of vectors that constitute a base, in this case n. ALGEBRAIC STRUCTURES 447 Linear spaces may have finite or infinite dimension, as will be shown next bby means of several examples. Example Al The marrices in ""' or C°!, constirue finite-dimensional lin ear spaces. The representations of position in physical space with respect 10 a certain reference frame also constitute a finite-dimensional linear space. The sequences of real numibers Sz. oF the set of continuous functions defined in a real interval Cig, form infinite-dimensional linear spaces. These last two cases may represent the pulse sequences in digital systems or certain classes of signals in analog circuits, respectively A.1.3. Metric, Norm, and Inner Products Definition A.S A metric space (V,d\-,.)) is defined in terms of a linear vec~ tor space V and a real function d{-,»):V x V = R,, satisfying the following conditions: 1 diy) 20, Vay ev. 2dyy)=0 =e xray: 3. dry) = dla), Yayev. 4 dena) x=0 3. lex||=lal-ix], Vx Via F. 4, xe yll) is defined in terms of 4 linear vector space V and a function <_,- >: x V+ F that satisfies the following conditions ¥ x,y,z €V and a ¢ F: kxtye>acne>+ FS. 20, Vr 60 i 2, 3. a. 4 S=0 oox=0. Here @ represents the complex conjugate of element a, Example A.3- The three normed spaces from the last example can be made inner product spaces with the following definitions: 1, Take vectors in C* and define t [Foire (ayy ‘An inner product space generates a normed vector space when using its inner product (o define a norm, as follows: |x\|= =X. >. In addition, a normed vector space generates a metric space when using its norm to define the distance function, as follows: d(x, y) = lx ~ yl Definition A.B Two vectors x,y in (V,< +>) are said 10 be orthogonal if and only if =0. If, in addition, =1 and < y,y >= 1, they are called orthonormal ‘A.2 FUNCTION SPACES 4.2.1 Introduction Next, some important properties of infinite-dimensional vector spaces are briefly reviewed. Definition A.9 A sequence {x,} in a metric space (V, 4) is said to be a Cauchy sequence if and only if Ve > 0 AN such that nm > N > dn, %m) < € Definition AO A sequence {9} in a metric space (V,d) converges to x, €V if and only if ¥€ > 0 AN such that n > N => d(tq%.) < € IL is easy to prove that any convergent sequence is a Cauchy sequence. (On the other hand, the converse, is not necessarily true. Example Ad Consider among the rational numbers Q the sequence {3 3.14 34415. This isa Cauchy sequence, although it does not con: verge because = ¢ Definition A.11 A metric space (V,d) is said 10 be complete if and only if all Cauchy sequences are convergent in V. A.2.2 Banach and Hilbert spaces jon A.12 A complete normed vector space (V, | -|\) is ealled a Banach ition A.I3 A complete inner product space (V.<-,->) is called a Hilbert space 450. MATHEMATICAL BACKGROUND x(t) Figure A.1. Nonconvergent sequence in Cra Example AS The infinite-dimensional vector space of continuous functions defined in the unit interval C3, #8 not complete when using the metries (x,y) = |x = ylh oF daxyy ilo, where these norms have been de: {fined in (AS). ‘Note in Figure A.J thatthe sequence X(t) of continuous functions converges ww a discontinuous one. Therefore the Space of continuous functions is not complete, under these metrics 2.3. Operator and Signal Spaces Many ofthe useful function spaces, which willbe described inthis subsection, include other than continuous functions. In these cases integration should be understood inthe Lebesgue integral sense (Riemann integral is valid ony for continuous functions). In addition, the concepts of minimum and maximum of a function (min, max) are replaced by their generalizations: infimum and supremum (int, sup). Finally, essential supremum, denoted as ess sup, refers to the supremum of a function for al but sets of measure zero in its domain These functional analysis concepts can be found in greater detail in 174) Time Domain ‘* C\(R,R) is the space of absolutely integrable scalar functions, equipped with the norm (a0) FUNCHIONSPACES 451 © £\(R,R"™*) isthe space of absolutely integrable matrix functions, equip. ped with the norm fil mol wn (Any 1* £:(R,R™") is the space of square integrable matrix functions of 1 € R, with inner product - + 74(B, RY") isthe subspace of £(R, R°*") formed by causal functions, that is, zero for: <0, The notation £,((0, 20), R”™) is sed equivalently + 715 (8, RP") isthe subspace of £3(R,R”") consisting of anricausalfonc- tions, that is, zero for 1 > 0. The notation £((—o0,0],R"*™) is used equivalently. Next, we show that for a causal stable system G, ||Gilz can be intexpreted as the RMS value ofits output z = G+ u when the input u isa white noise signal with unit covariance. To this effect, given a wide sense stationary random vector signal u(r) define its autocorrelation as ate) =€ ote] cary where £ denotes expectation. The Fourier transform of Ru(1) is called the spectral density Suu(jw), that is, Saja) = J Rane ar (a4) follows that €(u w) = trace [E(uu)] = trace [Ry(0)] Ibis a standard result that if z= Gu then Sj) = G(jo)Su(ja)G"Gw). Thus aeto-meettool= 2 ["wowtutwide Ly eS 52 = ay [Trace 610)" a) = 16 where we used the facts that Sya(jiw) = I and that trace(GG") = trace(G"). 452 MATHEMATICAL BACKGROUND Frequency Domain # C)R,C"™") is the space of square integrable matrix functions of jo, @ € R, with inner product, = a | se (FUG y0y"}de (Ais) and norm ||F\|2° V. Here we have defined G'(s) £ GT-s), + 7(UR,C™") is the subspace of £(jR, C""") composed of analytic func- tions in Re(s) > 0 such that spf trace|Fjw+o)Fyw+o)'|de 0 = 1,UR,C™™) is the subspace of £..()R,C™") of analytic and (essen- tially) bounded functions in Re(s) <0. £{R,-) and £,GR,-) are Hilbert spaces; £,.(JR,-) is a Banach space. In many cases, when these spaces are restricted to the set of rational transfer functions or matrices, an R is added to the notation, that is, RL3, RLao. RH, RM. In the main part of the book, the arguments of these spaces will be dropped for simplicity, these being clear from the context A.2.4 lsomorphism Definition A.14 Two vector spaces V and W are said to be isomorphic if and only if there exists a linear transformation L.: V —+ W, satisfying the following conditions: 1. L(x) = LO) > x= y EV, that is, N(L) = {Q}. This property is defined as monomorphism or one to one. 2. w= L(x) for all w € W with x € V, that is, R(L) = W. This property is called epimorphism or onto. FUNCIIONSPACES 453 Here N() and R(.) represent the kernel and range spaces of an operator, respectively. Definition AS Two normed vector spaces (V, |v) and (W,||- lw) are said to be isomette if and only if they are isomorphic and, in addition, the @ linear transformation L satisfies lly = WL(0)|w for any vector v € V. Example A The normed linear vector spaces £3(R,R) and £3(R,©) are isomorphic and isomeric under the linear transformation F : g(t) Ga) defined as the Fourier transform: Vf oper J 16003? ao This is also known as Parseval’s theorem. Furthermore, the inner product ‘among vectors in both spaces is also preserved, isle Giz (A.19) y — (A21) As a.consequence, the “algebraic structure” of both spaces is similar. In addi- tion, orthonormal vectors in one space will remain orthonormal in the trans- formed space, that is, “geometry” and “proportion” are preserved from one vector space to the other. Due to the isometry that the Fourier transform establishes between £(R) and £,()R), their subspaces can be related in the same way. Therefore %9(R) and Ha()R) are isometric, as well as 43 (R) and 13 ()R). Hence causal signals (or systems) ate isometric with stable transfer matrices and anticausal signals (or systems) with antistable ones. A.2.5 Induced Norms Definition A16 Consider two normed linear vector spaces (V,\\- Iv) and (WV, Iv) and a linear transformation L : V — W. The induced norm of the transformation is defined as follows: A IL()lw Wy 2 sup WOOl (a2) vive Telly It can be proved that equivalently the induced norm can be computed as SUP yy, 1 HEC)lw. 454 MATHEMATICAL BACKGROUND Example A.7 1. Consider the operator A € R"”" and use the Euclidean vector norm js in both R" and R” A. ieTATAX Malia © 9B Tid sel V aT = YEE) | aaa) (ax) The latter will be defined in the next section as the maximum singular value of A, 2. If the norm |) is used instead in RY and R, the following induced norm is obtained: Ath x = max Yai (A2a 3. When || - oo is used in R" and R" the following induced norm is obtained: All ow = max lanl (a2s) i ‘The induced norm quantifies the maximum “amplification” of a linear operator, 8 generalization of the concept of gain in SISO transfer functions. ‘Take, for example, the following matrix: f [om 10° 0 0.001 (a26) which can be interpreted as a linear operator A :R? ~ R?. Its eigenvalues Ay 0.001 do not measure the maximum amplification of vectors in RE when multiplied by the matrix. On the other hand, its induced norm (Alla 2 = (A) ~ 10? does. For example the input w= 0. 1)", produces an ‘output having Aullp = a(A) = 108 ‘An important property of induced norms is the fact that they are sub- smulipicative, that is, [> Lill w < [!Lalv—wlLi yw Proof. From (A.22) we have that jLyllw < hLllw-—wlyllw- Ht follows that MaLayllw < Waliw av hLaylw < Law ovliLallw-awllyllw— (A27) Equivalently, Mataylh Se < Uallw_wllLslww A28) ly = Weal wile (A28) ‘The proof follows now directly from the definition of induced norms, © FUNCTION SPACES 458 A.2.6 Some Important Induced System Norms In this section we consider the £* and £? induced system norms and we show that they coincide with the €, and H.. norms introduced in Section ALB The £*-+£ Case For simplicity consider the SISO case. For any in put u(t) € £*, lull = 1, we have that the corresponding output y = g + w satisfies boots fee sandr s [enue s [nian =i ‘This shows that lglle.c. < ligll. To show that the equality holds, assume that lg = Ho. Then, given € > 0, there exist T such that (A29) bom es fella sto (a30) ider che following signal aoe {omar osrsT 0 otherwi (ey) I is easily seen that the corresponding output y(¢) satisfies jy > |y(T)] > Bo ~ €. Thus sup gulls 2 Ho ~€ (ax) Halles ex ‘The proof follows now from the fact that ¢ is arbitrary, The CC Case Consider any input w € £2,;Iuj2 = 1 and let z = Gu Using the frequency-domain definition of |- >. we have that Hells = WGu(s)3 (A33) =¢ [ L2G Yay yaya) ds ve Eff stove genie de 1p (Ge 3 fw e)W10) deo GON. (As) 456 MATHEMATICAL BACKGROUND ‘This shows that the Mq norm is an upper bound of the £> induced norm, To show equality we need to find some w. € £2, |Iu.ll2 = 1, such that [22 = UiGllx. Assume that sup, @ [Gla)] is achieved at the frequency w.. Consider the input 1a(t) = k(e}e-* £05 (wal)man, WHETE Vas iS the Unitary right sin gular vector that corresponds to 4 (GUw,)] and k(e) is such that |4.(3)h for € > 0. As ¢ +0 we have that leallS = Tim G(s)u.(5)IH 5 i [GUoomek(e)[6(w +0.) + 5(w— a)? do + HG, )omex|” = 2 (Gye)] =I), (35) The 4 Norm ‘The Hz norm does not admit an interpretation as a norm induced by a single space (and thus it does not have the submuliplicaive property). Nevertheless, it i instructive to show that it can be considered fs an induced system norm, if the input and ouput spaces are allowed to have different norms. Given a signal u(t) we can define its autocorrelation and spectral density proceeding exactly as in Section A23. If Ryi(2) and Sy(jw) exist and Ry, is bounded, then w is said to be a power signal, with power (semi)norm given by |lull = trace [R,,(0)]. A power signal v is said to have bounded spectral density if ||Si«(J0}l|~ <0. In this case we can define its spectral density norm |jlls = |Sve(Jw)|}ac- Proceeding as in Section ‘A.23 it can easly be shown (by considering a signal u(?) having Sy, = 1) that Gl = IIIs Finally, in the SISO case, |j.\|2 can be interpreted as the £; — £o. norm. ‘To establish this fact consider € Cx, Illy 1. The corresponding output satisies viol = [f° ete ~ nara « ([Lat-otae)” (fiwerae)" = Isl (A36) where the second line follows from the Cauchy-Schwartz inequality, To com: plete the proof we need to show that there exists at least one signal u(t) € Lays lhully = 1 such that lly > gla To this effect take a-d Osh (a37) DUALITY AND DUAL SPACES 487 Obviously falls = 1 and Ae lla > Iy(OW TED Wolke 2 DOl= Tei Fa A.3 DUALITY AND DUAL SPACES (Aan) 4.3.1 The Dual Space Consider a linear vector space X. A functional on X is a function f:X'— R, Definition AAT Consider a normed linear space X. Its dual space X" is defined as the space of all bounded linear functionals on X, equipped with the norm | fl = stPjx4<1 If) It can be shown (187 that X" isa Banach space. In the sequel, given x € X and r © X", (x,r) denotes the value of the linear functional r at the point x. Example A.7: The Dual of f°, 1< p< oo Consider the space #” of all real sequences h = {,} for which >", /hi)? < 90, equipped with the norm Welle = COR Mul?" (i p =e then [Ase = supy !hu|)- Given 1 < p < 00, define the conjugate index g by the equation! (A39) Next, we show that every linear functional fon €? has a unique representation of the form A) = Show (A40) where = {n} € €*. Assume first that 1

0, there exists my € M such that ie ml) QandU ecm, are unitary matrices. ‘crow pan "Thun npone ocelot ast of rbonomategemesr satay DE o viacaw = [8 8] ass Let = YEA. Without tos of generaiy we ssume thatthe eigen vale of WPA are ode. that 2 Dene Vand te itr anv -n columnsof Vesecvey Repiang m (A00) min {\jrl>, Ax =b, Ae C™™, n < mh (as) The solution to these problems can be obtained by means of the matrix pseudoinverse or generalized inverse A*. The latter can efficiently be com puted with the use of the singular value decomposition as follows: =z 0 a-u[% 8 : : EO] y. Je eer fs Qo ay ere (Ae) ‘The input vector x = (1 1/2 L2"" yields Ax = —2""1 ay that tends to the zero vector as n —+ 90. Hence for a large enough m this ‘matrix is computationally singular, although itis theoretically not. ‘The eigenvalues are Ay = --- = An = —1 and the determinant is det(A) = (-1)", none of which provide a clear measure ofthe computational singularity of this matrix. In this last case, the reason is that det(A) = []} A,, which therefore depends on all eigenvalues, not only on the smallest one. In the former case, the measure provided by the smallest magnitude eigenvalue only quantifies the effect of the matrix on the eigenvectors, but not on any arbitrary vector in the input space ‘On the other hand, by considering the minimum value [Axl can take over all possible input vectors x (normalized by {illp 4.0 10 avoid trivial solutions), the measure of singularity can be computed as Axi RTATAR int Mele ing iehzo [ie] neo (EPID way twat which is directly the minimum singular value oA) = 2!" SINGULAR VALUES 463, Condition Number Consider the linear matsix equation Ax = b. The ‘maximum relative error 6b due to an error 5x, can be found as follows 86H yp Adele ner [Ole ~ Se ART =1 with axils = lal sup IAB inf, [Axl with [Sxl = [lela = 1 A) 2 7 wa!) (Ass) where 5x and x are independent vectors. The problem is equivalent to the one of finding the maximum relative error 8x of the solution for @ given error 8b in the data vector b. The condition number x(:) provides this measure and quantifies the relation between the maximum and mini- mum “gains” of A (according to different input vector “directions”). For ‘example, orthogonal (rotation) or unitary matrices are “spherical” because U'U = 1 = oy == oy = 1s hence in this ease «(U) = 1. On the other hhand, for singular matrices « —+ oo. APPENDIX B SYSTEM COMPUTATIONS In the following sections we present the state-space representations of the ‘most usual interconnections between linear systems. These representations are used to manipulate systems in a computer code. There is standard soft ware that implements these interconnections in state-space form, as in the case of the j-Toolbox and the Robust Toolbox of Matlab. Consider the following linear systems ous) = 54] (Ba) 2 | Be | (62) Gals) = |-p with states, inputs, and outputs (x1,4,y1) and (x, u2,92), Fespe series connection can be seen in Figure B.1, where uu). yam ‘The objective is to compute a state-space realization (not necessarily min- imal) of the product system. To this end, consider the state of Ga(s}Gu(s) as ely. The and x= [xf 21)’, and proceed as follows: AU = Arante) + Bini) (B3) 465

You might also like