Nonlinear Systems
Nonlinear Systems
Nonlinear and
Adaptive Control
Systems
Ding
Zhengtao Ding
04/03/2013 09:32:20
Nonlinear and
Adaptive Control
Systems
Nonlinear and
Adaptive Control
Systems
Zhengtao Ding
Contents
Preface
1 Introduction to nonlinear and adaptive systems
1.1
Nonlinear functions and nonlinearities
1.2
Common nonlinear systems behaviours
1.3
Stability and control of nonlinear systems
ix
1
1
4
5
9
9
11
12
18
3 Describing functions
3.1
Fundamentals
3.2
Describing functions for common nonlinear components
3.3
Describing function analysis of nonlinear systems
25
26
29
34
4 Stability theory
4.1
Basic denitions
4.2
Linearisation and local stability
4.3
Lyapunovs direct method
4.4
Lyapunov analysis of linear time-invariant systems
41
41
45
46
51
55
55
59
65
71
6 Feedback linearisation
6.1
Inputoutput linearisation
6.2
Full-state feedback linearisation
75
75
83
89
90
94
99
102
103
viii
109
109
111
120
122
127
136
9 Backstepping design
9.1
Integrator backstepping
9.2
Iterative backstepping
9.3
Observer backstepping
9.4
Backstepping with ltered transformation
9.5
Adaptive backstepping
9.6
Adaptive observer backstepping
141
141
144
147
152
159
167
10
175
175
186
194
204
11
Control applications
11.1 Harmonics estimation and rejection in power distribution
systems
11.1.1 System model
11.1.2 Iterative observer design for estimating frequency
modes in input
11.1.3 Estimation of specic frequency modes in input
11.1.4 Rejection of frequency modes
11.1.5 Example
11.2 Observer and control design for circadian rhythms
11.2.1 Circadian model
11.2.2 Lipschitz observer design
11.2.3 Phase control of circadian rhythms
11.3 Sampled-data control of nonlinear systems
11.3.1 System model and sampled-data control
11.3.2 Stability analysis of sampled-data systems
11.3.3 Simulation
219
219
220
224
232
234
235
238
239
241
243
247
249
251
260
Bibliographical Notes
263
References
268
Index
275
Preface
This book is intended for the use as a textbook at MSc and senior undergraduate
level in control engineering and related disciplines such as electrical, mechanical,
chemical and aerospace engineering and applied mathematics. It can also be used as
a reference book by control engineers in industry and research students in automation
and control. It is largely, although not entirely, based on the course unit bearing the
same name as the book title that I have been teaching for several years for the MSc
course at Control Systems Centre, School of Electrical and Electronic Engineering,
The University of Manchester. The beginning chapters cover fundamental concepts
in nonlinear control at moderate mathematical level suitable for students with a rst
degree in engineering disciplines. Simple examples are used to illustrate important
concepts, such as the difference between exponential stability and asymptotic stability.
Some advanced and recent stability concepts such as input-to-state stability are also
included, mainly as an introduction at a less-demanding mathematical level compared
with their normal descriptions in the existing books, to research students who may
encounter those concepts in literature. Most of the theorems in the beginning chapters
are introduced with the proofs, and some of the theorems are simplied with less
general scopes, but without loss of rigour. The later chapters cover several topics
which are closely related to my own research activities, such as nonlinear observer
design and asymptotic disturbance rejection of nonlinear systems. They are included
to demonstrate the applications of fundamental concepts in nonlinear and adaptive
control to MSc and research students, and to bridge the gap between a normal textbook
treatment of control concepts and that of research articles published in academic
journals. They can also be used as references for the students who are working on
the related topics. At the end of the book, applications to less traditional areas such
as control of circadian rhythms are also shown, to encourage readers to explore new
applied areas of nonlinear and adaptive control.
This book aims at a unied treatment of adaptive and nonlinear control.
It is well known that the dynamics of an adaptive control system for a linear dynamic
system with unknown parameters are nonlinear. The analysis of such adaptive systems requires similar techniques to the analysis for nonlinear systems. Some more
recent control design techniques such as backstepping relies on Lyapunov functions
to establish the stability, and they can be directly extended to adaptive control of nonlinear systems. These techniques further reduce the traditional gap between adaptive
control and nonlinear control. Therefore, it is now natural to treat adaptive control
as a part of nonlinear control systems. The foundation for linear adaptive control
and nonlinear adaptive control is the positive real lemma, which is related to passive
systems in nonlinear control and Lyapunov analysis. It is decided to use the positive
real lemma and related results in adaptive control and nonlinear control as the
main theme of the book, together with Lyapunov analysis. Other important results
such as circle criterion and backstepping are introduced as extensions and further
developments from this main theme.
For a course unit of 15 credits on nonlinear and adaptive control at the Control
Systems Centre, I normally cover Chapters 14, 6 and 7, and most of the contents of
Chapter 5, and about half of the materials in Chapter 9. Most of the topics covered
in Chapters 8, 10 and 11 have been used as MSc dissertation projects and some of
them as PhD projects. The contents may also be used for an introductory course
on nonlinear control systems, by including Chapters 15, 8 and the rst half of
Chapter 9, and possibly Chapter 6. For a course on adaptive control of nonlinear
systems, an instructor may include Chapters 1, 2, 4, 5, 7 and 9. Chapter 8 may be
used alone as a brief introduction course to nonlinear observer design. Some results
shown in Chapters 8, 10 and 11 are recently published, and can be used as references
for the latest developments in related areas.
Nonlinear and adaptive control is still a very active research area in automation
and control, with many new theoretic results and applications continuing to merge.
I hope that the publication of this work will have a good impact, however small,
on students interests to the subject. I have been beneted from my students, both
undergraduate and MSc students, through my teaching and other interactions with
them, in particular, their questions to ask me to explain many of the topics covered
in this book with simple languages and examples. My research collaborators and
PhD students have contributed to several topics covered in the book through joint
journal publications, whose names may be found in the references cited at the end of
the book. I would like to thank all the researchers in the area who contributed to the
topics covered in the book, who are the very people that make this subject fascinating.
Chapter 1
Nonlinearity is ubiquitous, and almost all the systems are nonlinear systems. Many
of them can be approximated by linear dynamic systems, and signicant amount of
analysis and control design tools can then be applied. However, there are intrinsic
nonlinear behaviours which cannot be described using linear systems, and analysis
and control are necessarily based on nonlinear systems. Even for a linear system,
if there are uncertainties, nonlinear control strategies such as adaptive control may
have to be used. In the last two decades, there have been signicant developments in
nonlinear system analysis and control design. Some of them are covered in this book.
In this chapter, we will discuss typical nonlinearities and nonlinear behaviours, and
introduce some basic concepts for nonlinear system analysis and control.
1.1
A dynamic system has its origin from dynamics in classic mechanics. The behaviour of
a dynamic system is often specied by differential equations. Variables in a dynamic
system are referred to as states, and they can be used to determine the status of a
system. Without external inuences, the state variables are sufcient to determine the
future status for a dynamic system. For a dynamic system described by continuous
differential equations, state variables cannot be changed instantly, and this reects the
physical reality. Many physical and engineering systems can be modelled as dynamic
systems, using ordinary differential equations. Application areas of dynamic systems
have expanded rapidly to other areas such as biological systems, nancial systems,
etc. Analysis of the behaviours of dynamic systems is essential to the understanding
of various applications in many science and engineering disciplines. The behaviour
of a dynamic system may be altered by exerting external inuences, and quite often
this kind of inuences is based on knowledge of the current state. In this sense, the
dynamic system is controlled to achieve certain behaviours.
State variables are denoted by a vector in an appropriate dimension for convenience, and the dynamic systems are described by rst-order differential equations of
the state vector. A linear dynamic system is described by
x = Ax + Bu
y = Cx + Du,
(1.1)
(1.2)
1
(u) = u
for u < 1,
otherwise,
for u > 1.
(1.3)
The only difference between the systems (1.2) and (1.1) is the saturation function .
It is clear that the saturation function is a nonlinear function, and therefore this
system is a nonlinear system. Indeed, it can be seen that the superposition principle
does not apply, because after the input saturation, any increase in the input amplitude
does not change the system response at all.
A general nonlinear system is often described by
x = f (x, u, t)
y = h(x, u, t),
(1.4)
where x Rn , u Rm and y Rs are the system state, input and output respectively,
and f : Rn Rm R Rn and h : Rn Rm R Rs are nonlinear functions.
Nonlinearities of a dynamic system are described by nonlinear functions. We may
roughly classify nonlinear functions in nonlinear dynamic systems into two types.
The rst type of nonlinear functions are analytical functions such as polynomials,
sinusoidal functions and exponential functions, or composition of these functions.
The derivatives of these functions exist, and their Taylor series can be used to obtain
good approximations at any points. These nonlinearities may arise from physical
modelling of actual systems, such as nonlinear springs and nonlinear resistors, or due
to nonlinear control design, such as nonlinear damping and parameter adaptation law
for adaptive control. There are nonlinear control methods such as backstepping which
requires the existence of derivatives up to certain orders.
discussed in Chapter 2, and limit cycles also appear in other problems considered
in this book. High-order harmonics are discussed in disturbance rejection. When the
disturbance is a harmonic signal, the internal model for disturbance rejection has to
consider the high-order harmonics generated due to nonlinearities.
People often start with linearisation of a nonlinear system. If the control design
based on a linearised model works, then there is no need to worry about nonlinear
control design. Linearised models depend on operating points, and a switching strategy might be needed to move from one operating point to the others. Gain scheduling
and linear parameter variation (LPV) methods are also closely related to linearisation
around operating points.
Linearisation can also be achieved for certain class of nonlinear systems through
a nonlinear state transformation and feedback. This linearisation is very much different from linearisation around operating points. As shown later in Chapter 6, a number
of geometric conditions must be satised for the existence of such a nonlinear transformation. The linearisation obtained in this way works globally in the state space,
not just at one operating point. Once the linearised model is obtained, further control
design can be carried out using design methods for linear systems.
Nonlinear functions can be approximated using articial neuron networks, and
fuzzy systems and control methods have been developed using these approximation
methods. The stability analysis of such systems is often similar to Lyapunov functionbased design method and adaptive control. We will not cover them in this book. Other
nonlinear control design methods such as bandband control and sliding mode control
are also not covered.
In the last two decades, there were developments for some more systematic
control design methods, such as backstepping and forwarding. They require the system
to have certain structures so that these iterative control designs can be carried out.
Among them, backstepping method is perhaps the most popular one. As shown in
Chapter 9 in the book, it requires the system state space function in a sort of lowertriangular form so that at each step a virtual control input can be designed. Signicant
amount of coverage of this topic can be found in this book. Forwarding control design
can be interpreted as a counterpart of backstepping in principle, but it is not covered
in this book.
When there are parametric uncertainties, adaptive control can be introduced to
tackle the uncertainty. As shown in a simple example earlier, an adaptive control system is nonlinear, even for a linear system. Adaptive technique can also be introduced
together with other nonlinear control design methods, such as backstepping method.
In such a case, people often give it a name, adaptive backstepping. Adaptive control
for linear systems and adaptive backstepping for nonlinear systems are covered in
details in Chapter 7 and Chapter 9 in this book.
Similar to linear control system design, nonlinear control design methods can
also be grouped as state-feedback control design and output-feedback control design.
The difference is that the separation principle is not valid for nonlinear control design
in general, that is if we replace the state in the control input by its estimate, we would
not be able to guarantee the stability of the closed-loop system using state estimate.
Often state estimation must be integrated in the control design, such as observer
backstepping method.
State estimation is an important topic for nonlinear systems on its own. Over the
last three decades, various observer design methods have been introduced. Some of
them may have their counterparts in control design. Design methods are developed
for different nonlinearities. One of them is for systems with Lipschitz nonlinearity,
as shown in Chapter 8. A very neat nonlinear observer design is the observer design
with output injection, which can be applied to a class of nonlinear systems whose
nonlinearities are only of the system output.
In recent years, the concept of semi-global stability is getting more popular.
Semi-global stability is not as good as global stability, but the domain of attraction
can be as big as you can specify. The relaxation in the global domain of attraction does
give control design more freedom in choosing control laws. One common strategy is
to use high gain control together with saturation. We will not cover it in this book,
but the design methods in semi-global stability can be easily followed once a reader
is familiar with the control design and analysis methods introduced in this book.
Chapter 2
The nonlinear systems under consideration in this book are described by differential
equations. In the same way as for linear systems, we have system state variables, inputs
and outputs. In this chapter, we will provide basic denitions for state space models of
nonlinear systems, and tools for preliminary analysis, including linearisation around
operating points. Typical nonlinear behaviours such as limit cycles and chaos will
also be discussed with examples.
x(0) = x0 ,
(2.1)
(2.2)
with > 0.
Note that Lipschitz condition implies continuity with respect to x. The existence
and uniqueness of a solution for (2.1) are guaranteed by the function f being Lipschitz
and being continuous with respect to t.
10
Remark 2.1. The continuity of f with respect to t and state variable x might be
stronger than we have in real applications. For example, a step function is not continuous in time. In the case that there are nite number of discontinuities in a given
interval, we can solve the equation of a solution in each of the continuous region, and
join them together. There are situations of discontinuity with state variable, such as
an ideal relay. In such a case, the uniqueness of the solution can be an issue. Further
discussion on this is beyond the scope of this book. In the systems considered in
the book, we would assume that there would be no problem with the uniqueness of
a solution.
The system state contains the whole information of the behaviour. However,
for a particular application, only a subset of the state variables or a function of state
variables is of interest, which can be denoted as y = h(x, u, t) with h : Rn Rm
R Rs , normally with s < n. We often refer to y as the output of the system. To
write them together with the system dynamics, we have
x = f (x, u, t),
x(0) = x0
y = h(x, u, t).
In this book, we mainly deal with time-invariant systems. Hence, we can drop the
variable t in f and h and write the system as
x = f (x, u),
x(0) = x0
(2.3)
y = h(x, u),
where x Rn is the state of the system, y Rs and u Rm are the output and the
input of the system respectively, and f : Rn Rm Rn and h : Rn Rm Rs are
continuous functions.
Nonlinear system dynamics are much more complex than linear systems in
general. However, when the state variables are subject to small variations, we would
expect the behaviours for small variations to be similar to linear systems, based on
the fact that
f (x + x, u + u) f (x, u) +
f
f
(x, u)x +
(x, u)u,
x
u
11
(2.4)
where A Rnn , B Rnm , C Rsn , D Rsm matrices with elements ai,j , bi,j , ci,j
and di,j respectively shown by, assuming that f and h are differentiable,
ai,j =
fi
(xe , ue ),
xj
bi,j =
fi
(xe , ue ),
uj
ci,j =
hi
(xe , ue ),
xj
di,j =
hi
(xe , ue ).
uj
Remark 2.2. For a practical system, a control input can keep the state in an equilibrium point, i.e., at a point such that x = 0, and therefore it is natural to look at the
linearisation around this point. However, we can obtain linearised model at points that
are not at equilibrium. If (xe , ue ) is not an equilibrium point, we have f (xe , ue ) = 0.
We can carry out the linearisation in the same way, but the resultant linearised system
is given by
x = Ax + Bu + d
y = C x + Du ,
where d = f (xe , ue ) is a constant vector.
(2.5)
12
Remark 2.3. It is easy to see that for a system in (2.3), if the control input remains
constant, then it is an autonomous system. We only need to re-dene the function f as
fa (x) := f (x, uc ) where uc is a constant input. Even if the inputs are polynomials and
sinusoidal functions of time, we can convert the system to the autonomous system by
modelling the sinusoidal and polynomial functions as the state variables of a linear
dynamic system, and integrate this system into the original system. The augmented
system is then an autonomous system.
Denition 2.3. For an autonomous system (2.5), a point xe Rn is a singular point
if f (xe ) = 0.
It is easy to see that singular points are equilibrium points. Singular points are
more preferred for autonomous systems, especially for second-order systems.
Since autonomous system do not have external input, the set of all the trajectories provides a complete geometrical representation of the dynamic behaviour.
This is often referred to as the phase portrait, especially for second-order systems
in the format
x 1 = x2
x 2 = (x1 , x2 ).
In the above system, if we interpret x1 as the displacement, then x2 is the velocity. The
state variables often have clear physical meanings. Phase portraits can be obtained by
a number of methods, including analysing the behaviours near the singular points. In
fact, singular points might get the name from their positions in the phase portrait. In a
phase portrait, the lines usually do not intercept each other due to the uniqueness of the
solutions. However, they meet at the points where f (x) = 0, seemingly intercepting
each other. Those points are singular in this sense.
f1
x1 (x1e , x2e )
Ae =
f2
(x1e , x2e )
x1
f1
(x1e , x2e )
x2
.
f2
(x1e , x2e )
x2
(2.6)
13
x1
Stable node
Unstable node, for 1 > 0, 2 > 0. This singular point is unstable, and the trajectories
diverge from the point, but not spiral around it.
x2
x1
Unstable node
Saddle point, for 1 < 0, 2 > 0. With one positive and one negative eigenvalues, the
hyperplane in three dimensions may look like a saddle. Some trajectories converge
to the singular point, and others diverge, depending on the directions of approaching
the point.
x2
x1
Saddle point
Stable focus, for 1,2 = j, ( < 0). With a negative real part for a pair of conjugate poles, the singular point is stable. Trajectories converge to the singular point,
14
spiralling around. In time domain, the solutions are similar to decayed sinusoidal
functions.
x2
x1
Stable focus
Unstable focus, for 1,2 = j, ( > 0). The real part is positive, and therefore
the singular point is unstable, with the trajectories spiralling out from the singular
point.
x2
x1
Unstable focus
Centre, for 1,2 = j, ( > 0). For the linearised model, when the real part is zero,
the norm of the state is constant. In phase portrait, there are closed orbits around the
singular point.
x2
x1
Centre
To draw a phase portrait, the analysis of singular points is the rst step. Based
on the classication of the singular points, the behaviours in neighbourhoods of these
points are more or less determined. For other regions, we can calculate the directions
of the movement from the directives.
At any point, the slope of trajectory can be computed by
f2 (x1 , x2 )
dx2
.
=
dx1
f1 (x1 , x2 )
15
With enough points in the plane, we should be able to sketch phase portraits connecting the points in the directions determined by the slopes.
Indeed, we can even obtain curves with constant slopes, which are named as
isoclines. An isocline is a curve on which (f2 (x1 , x2 )/f1 (x1 , x2 )) is constant. This, again,
can be useful in sketching a phase portrait for a second-order nonlinear system.
It should be noted that modern computer simulation can provide accurate solutions to many nonlinear differential equations. For this reason, we will not go to
further details of drawing phase portraits based on calculating slope of trajectories
and isoclines.
Example 2.1. Consider a second-order nonlinear system
x 1 = x2
x 2 = x2 2x1 + x12 .
Setting
0 = x2 ,
0 = x2 2x1 + x12 ,
we obtain two singular points (0, 0) and (2, 0).
Linearised system matrix for the singular point (0, 0) is obtained as
0
1
A=
,
1 2
and the eigenvalues are obtained as 1 = 2 = 1. Hence, this singular point is a
stable node.
For the singular point (2, 0), the linearised system matrix is obtained as
0
1
A=
,
2 1
and the eigenvalues are 1 = 2 and 2 = 1. Hence, this singular point is a saddle
point. It is useful to obtain the corresponding eigenvectors to determine which direction is converging and which is diverging. The eigenvalues v1 and v2 , for 1 = 2
and 2 = 1, are obtained as
1
1
v1 =
, v2 =
2
1
This suggests that along the direction of v1 , relative to the singular point, the state
converges to the singular point, while along v2 , the state diverges. We can clearly
see from Figure 2.1 that there is a stable region near the singular point (0, 0).
However, in the neighbourhood of (2, 1) one part of it is stable, and the other part
is unstable.
16
x2
1
1.5
2
2.5
3
3.5
4
4
x1
(2.7)
where H is the inertia, is the rotor angle, Pm is the mechanical power and Pe is the
maximum electrical power generated. We may view Pm as the input and Pe sin as
the output. For the convenience of presentation, we take H = 1, Pm = 1 and Pe = 2.
The state space model is obtained by letting x1 = and x2 = as
x 1 = x2
x 2 = 1 2 sin (x1 ).
For the singular points, we obtain
5
1
or ,
6
6
= 0.
x1e =
x2e
Note that there are an innite number of singular points, as x1e = 2k + 16 and
x1e = 2k + 56 are also solutions for any integer value of k.
17
Let us concentrate on the analysis of the two singular points ( 16 , 0) and ( 56 , 0).
The linearised system matrix is obtained as
A=
1
.
0
0
2 cos (x1e )
For ( 61 , 0), the eigenvalues are 1,2 = 31/4 j, and therefore this singular point it a
centre.
For ( 56 , 0), the eigenvalues are 1 = 31/4 and 2 = 31/4 . Hence, this singular
point it a saddle point. The eigenvalues v1 and v2 , for 1 = 31/4 and 2 = 31/4 , are
obtained as
v1 =
1
,
31/4
v2 =
.
31/4
x2
3
2
x1
18
(2.8)
19
From this state space realisation, it can be seen that when = 0, van der Pol oscillator
is the same as a harmonic oscillator. For with small values, one would expect that
it behaves like a harmonic oscillator.
A more revealing state transformation for with big values is given by
x1 = y
1
x2 = y + f (y),
where f (y) = y3 /3 y. Under the above transformation, we have the system as
x 1 = (x2 f (x1 ))
x 2 = 1 x1 .
(2.9)
(2.10)
2
= 0 or x2 = f (x1 ). This can be
This equation suggests that as , we have dx
dx1
seen from the phase portrait for very big values of in Figure 2.4.
Let us stick with the state space model (2.9). The only singular point is at the
origin (0, 0). The linearised system matrix at the origin is obtained as
A=
1
.
0
From the eigenvalues of A, we can see that this singular point is either an unstable
node or an unstable focus, depending on the value of . Phase portrait of van der
Pol oscillator with = 1 is shown in Figure 2.3 for two trajectories, one with initial
condition outside the limit cycle and one from inside. The broken line shows x2 =
f (x1 ). Figure 2.4 shows the phase portrait with = 10. It is clear from Figure 2.4 that
the trajectory sticks with the line x2 = f (x1 ) along the outside and then moves almost
horizontally to the other side, as predicted in the analysis earlier.
Limit cycles also exist in high-order nonlinear systems. As seen later in Chapter 11, circadian rhythms can also be modelled as limit cycles of nonlinear dynamic
systems. For second-order autonomous systems, limit cycles are very typical trajectories. The following theorem, PoincareBendixson theorem, describes the features
of trajectories of the second-order systems, from which a condition on the existence
of a limit cycle can be drawn.
Theorem 2.1. If a trajectory of the second-order autonomous system remains in a
nite region, then one of the following is true:
x2
3
2.5
1.5
0.5
0
x1
0.5
1.5
2.5
x2
20
3
2.5
1.5
0.5
0
x1
0.5
1.5
2.5
21
For high-order nonlinear systems, there are more complicated features if the
trajectories remain in a bounded region. For the asymptotic behaviours of dynamic
systems, we dene positive limit sets.
Denition 2.5. Positive limit set of a trajectory is the set of all the points for which
the trajectory converges to, as t .
Positive limit sets are also referred to as -limit sets, as is the last letter of
Greek letters. Similarly, we can dene negative limit sets, and they are called -limit
sets accordingly. Stable limit cycles are positive limit sets, so do stable equilibrium
points. The dimension for -limit sets is zero or one, depending on singular points or
limit cycles.
Strange limit sets are those limit sets which may or may not be asymptotically
attractive to the neighbouring trajectories. The trajectories they contain may be locally
divergent from each other, within the attracting set. Their dimensions might be fractional. Such structures are associated with the quasi-random behaviour of solutions
called chaos.
Example 2.4. The Lorenz attractor. This is one of the most widely studied examples
of strange behaviour in ordinary differential equations, which is originated from
studies of turbulent convection by Lorenz. The equation is in the form
x 1 = (x2 x1 )
x 2 = (1 + x3 )x1 x2
(2.11)
x 3 = x1 x2 bx3 ,
where
, and b are
positive
constants. There are three equilibrium points (0, 0, 0),
( b, b, ) and ( b, b, ). The linearised system matrix around the origin
is obtained as
A = +1
0
1
0
0
0 ,
b
x3
20
15
10
5
0
20
10
0
x2
10
20
10
15
10
15
x1
20
15
10
5
x2
22
0
5
10
15
20
15
10
0
x1
10
15
30
25
x3
20
15
10
0
15
10
0
x1
10
15
10
x1
10
15
20
40
60
Time (s)
80
100
23
24
Chapter 3
Describing functions
In classical control, frequency response is a powerful tool for analysis and control
design of linear dynamic systems. It provides graphical presentation of system dynamics and often can reect certain physical features of engineering systems. The basic
concept of frequency response is that for a linear system, if the input is a sinusoidal
function, the steady-state response will still be a sinusoidal function, but with a
different amplitude and a different phase. The ratio of the input and output amplitudes and the difference in the phase angles are determined by the system dynamics.
When there is a nonlinear element in a control loop, frequency response methods
cannot be directly applied. When a nonlinear element is a static component, i.e., the
input and output relationship can be described by an algebraic function, its output
to any periodic function will be a periodic function, with the same period as the
input signal. Hence, the output of a static nonlinear element is a periodic function
when the input is a sinusoidal function. It is well known that any periodic function with piece-wise continuity has its Fourier series which consists of sinusoidal
functions with the same period or frequency as the input with a constant bias, and
other sinusoidal functions with high multiple frequencies. If we take the term with
the fundamental frequency, i.e., the same frequency as the input, as an approximation, the performance of the entire dynamic system may be analysed using frequency
response techniques. Describing functions are the frequency response functions of
nonlinear components with their fundamental frequency terms as their approximate
outputs. In this sense, describing functions are rst-order approximation in frequency
domain. It can also be viewed as a linearisation method in frequency domain for
nonlinear components.
Describing function analysis remains as an important tool for analysis of nonlinear systems with static components despite several more recent developments in
nonlinear control and design. It is relatively easy to use, and closely related to frequency response analysis of linear systems. It is often used to predict the existence
of limit cycles in a nonlinear system, and it can also be used for prediction of subharmonics and jump phenomena of nonlinear systems. In this chapter, we will present
basic concept of describing functions, calculation of describing functions of common
nonlinear elements and how to use describing functions to predict the existence of
limit cycles.
26
3.1 Fundamentals
For a nonlinear component described by a nonlinear function f : R R, its
A sin(wt)
f (x)
w(t)
output
w(t) = f (A sin(t))
to a sinusoidal input A sin (t) is a periodical function, although it may not be sinusoidal in general. Assuming that the function f is piecewise-continuous, w(t) is a
piecewise-continuous periodic function with the same period as the input signal. A
piecewise periodical function can be expanded in Fourier series
a0
(an cos(nt) + bn sin(nt)),
+
2
n=1
w(t) =
where
1
a0 =
an =
bn =
(3.1)
w(t)d(t)
w(t) cos(nt)d(t)
w(t) sin(nt)d(t).
Remark 3.1. For a piecewise-continuous function w(t), the Fourier series on the
right-hand side of (3.1) converges to w(t) at any continuous point, and to the average
of two values obtained by taking limits from both sides at a dis-continuous point. If
we truncate the series up to order k,
a0
wk (t) =
(an cos(nt) + bn sin(nt)),
+
2
n=1
k
a0
+ a1 cos(t) + b1 sin(t).
2
(3.2)
Describing functions
27
signal A sin(t), is a sinusoidal function in (3.2) with the Fourier coefcients a1 and
b1 shown in (3.1). Hence, we can analyse the frequency response of this nonlinear
component.
We can rewrite w1 in (3.2) as
w1 = M sin(t + ),
where
M (A, ) =
(3.3)
a21 + b21 ,
Mejt+
b1 + ja1
=
.
Aejt
A
(3.4)
Remark 3.2. A clear difference between the describing function of a nonlinear element and the frequency response of a linear system is that the describing function
depends on the input amplitude. This reects the nonlinear nature of the describing
function.
Remark 3.3. If f is a single-valued odd function, i.e., f (x) = f (x), we have
1
a1 =
f (A sin(t)) cos(t)d(t)
1
1 0
f (A sin(t) cos(t)d(t) +
f (A sin(t) cos(t)d(t)
=
0
1
1
f (A sin(t) cos(t)d(t) +
f (A sin(t) cos(t)d(t)
=
0
0
= 0.
If a1 = 0, the describing function is a real value.
Example 3.1. The characteristics of a hardening spring are given by
x3
.
2
Given the input A sin(t), the output is
f (x) = x +
w(t) = f (A sin(t))
= A sin(t) +
A3
sin3 (t).
2
28
for n > 2,
we have
3
b1 = A + A3 .
8
Therefore, the describing function is
N (A, ) = N (A) =
3
b1
= 1 + A2 .
A
8
w(t) = A sin(t) +
Through the above discussion, describing functions are well dened for nonlinear
components whose inputoutput relationship can be well dened by piecewisecontinuous functions. These functions are time-invariant, i.e., the properties of
nonlinear elements do not vary with time. This is in line with the assumption for frequency response analysis, which can only be applied to time-invariant linear systems.
We treat describing functions as the approximations at the fundamental frequencies,
and therefore in our analysis, we require a0 = 0 which is guaranteed by odd functions
for the nonlinear components. With the describing function of a nonlinear component, we can then apply analysis in frequency responses for the entire system. For the
convenience of this kind of analysis, we often assume that the nonlinear component
for which the describing function is used to approximate its behaviours is the only
Describing functions
r +
f (x)
29
G(s)
kx,
sign(x)ka,
otherwise.
(3.5)
The output to the input A sin(t), for A > a, is symmetric over quarters of a period,
and in the rst quarter,
f(x)
a
a
kA sin(t),
0 t ,
ka,
< t /2,
(3.6)
where = sin1 (a/A). The function is odd, hence we have a1 = 0, and the symmetry
of w1 (t) implies that
30
/2
w1 sin(t)d(t)
0
kA sin2 (t)d(t) +
/2
ka sin(t)d(t)
1
4ka
2kA
sin(2 ) +
cos( )
=
4ka
2kA
a
=
cos( ) +
cos( )
2kA
a
=
+ cos( )
A
2kA
a
a2
=
+
1 2 .
A
A
Note that we have used sin =
describing function is given by
2k
b1
=
N (A) =
A
sin
a
A
and cos =
1
a
a2
a
1 2 .
+
A A
A
a2
.
A2
Therefore, the
(3.7)
Example 3.3. Ideal relay. The output from the ideal relay shown in Figure 3.3
(signum function) is described by, with M > 0,
M , t < 0,
w(t) =
(3.8)
M,
0 t < .
It is again an odd function, hence we have a1 = 0. The coefcient b1 is given by
b1 =
0
M sin(t)d(t) =
4M
f(x)
M
x
M
Describing functions
31
f(x)
a
a
4M
.
A
(3.9)
Example 3.4. Dead zone. A dead zone is a complement to saturation. A dead zone
shown in Figure 3.4 can be described by a nonlinear function
(3.10)
The output to the input A sin(t), for A > a, is symmetric over quarters of a
period, and in the rst quarter,
0,
0 t ,
w(x) =
(3.11)
k(A sin(t) a), < t /2,
where = sin1 (a/A). The function is odd, hence we have a1 = 0, and the symmetry
of w(t) implies that
4
b1 =
/2
w(t) sin(t)d(t)
0
/2
1
4ka
2kA
+ sin(2 )
cos( )
2
2
2kA
a
= kA
+ cos( )
A
2kA
a2
a
= kA
1 2 .
+
A
A
32
sin1
a
a2
a
1 2 .
+
A A
A
(3.12)
Remark 3.4. The dead-zone function shown in (3.10) complements the saturation
function shown in (3.5) in the sense that if we use fs and fd to denote the saturation
function and dead-zone function, we have fs + fd = k for the describing functions
shown in (3.7) and (3.12), the same relationship holds.
Example 3.5. Relay with hysteresis. Consider a case when there is a delay in the
ideal relay as shown in Figure 3.5. The nonlinear function for relay with hysteresis
can be described by
M,
for x a,
M , for x a.
(3.13)
When this nonlinear component takes A sin (t) as the input with A > a, the
output w(t) is given by
for t < ( ),
M,
w(t) = M , for ( ) t < ,
M,
for t < ,
(3.14)
where = sin1 ( Aa ). In this case, we still have a0 = 0, but not a1 . For a1 we have
f(x)
M
x
M
Describing functions
a1 =
( )
1
+
M cos(t)d(t) +
( )
33
M cos(t)d(t)
M cos(t)d(t)
4M
sin( )
4M a
=
.
A
Similarly, we have
b1 =
( )
1
+
M sin(t)d(t) +
( )
M sin(t)d(t)
M sin(t)d(t)
4M
cos( )
4M
a2
=
1 2.
From
N (A, ) =
b1 + ja1
,
A
we have
4M
N (A) =
A
a2
a
1 2 j
.
A
A
(3.15)
Using the identity cos( ) + j sin( ) = ejr , we can rewrite the describing function as
N (A) =
4M j arcsin(a/A)
e
.
A
Remark 3.5. Comparing the describing function of the relay with hysteresis with
that of ideal relay in (3.9), the describing functions indicate that there is a delay in the
relay with hysteresis by arcsin(a/A) in terms of phase angle. There is indeed a delay
of = arcsin(a/A) in the time response w(t) shown in (3.14) with that of the ideal
relay. In fact, we could use this fact to obtain the describing function for the relay
with hysteresis.
34
(3.16)
or
G(j) =
1
.
N (A, )
(3.17)
Therefore, the amplitude A and frequency of the limit cycle must satisfy the above
equation. Equation (3.17) is difcult to solve in general. Graphic solutions can be
found by plotting G(j) and 1/N (A, ) on the same graph to see if they intersect
each other. The intersection points are the solutions, from which the amplitude and
frequency of the oscillation can be obtained.
Remark 3.6. The above discussion is based on the assumption that the oscillation,
or limit cycle, can be well approximated by a sinusoidal function, and the nonlinear
component is well approximated by its describing function. The describing function
analysis is an approximate method in nature.
Only a stable limit cycle may exist in real applications. When we say stable limit
cycle, we mean that if the state deviates a little from the limit cycle, it should come
back. With the amplitude as an example, if A is perturbed from its steady condition,
say with a very small increase in the amplitude, for a stable limit cycle, the system
will decay to its steady condition.
Describing functions
35
1
.
K
In this case, the Nyquist criterion can be extended to determine the stability of the
closed loop by counting the encirclements of the Nyquist plot around (1/K, 0) in the
complex plain in the same way as around (1, 0). The Nyquist criterion for non-unity
forward path gain K is also referred to as the extended Nyquist criterion. The same
argument holds when k is a complex number.
We can apply the extended Nyquist criterion to determine the stability of a limit
cycle. When the condition specied in (3.17) is satised for some (A0 , 0 ), A0 and
0 are the amplitude and frequency of the limit cycle respectively, and N (A0 , 0 )
is a complex number. We can use the extended Nyquist criterion to determine the
stability of the limit cycle with the amplitude A0 and frequency 0 by considering a
perturbation of A around A0 .
To simplify our discussion, let us assume that G(s) is stable and minimum phase.
It is known from the Nyquist criterion that the closed-loop system with constant
gain K is stable if the Nyquist plot does not encircle (1/K, 0). Let us consider a
perturbation in A to A+ with A+ > A0 . In such a case, 1/N (A+ , 0 ) is a complex
number in general. If the Nyquist plot does not encircle the point 1/N (A+ , 0 ), we
conclude that the closed-loop system is stable with the complex gain 1/N (A+ , 0 ).
Therefore, in a stable closed-loop system, the oscillation amplitude decays, which
makes A+ return to A0 . This implies that the limit cycle (A0 , 0 ) is stable. Alternatively,
if the Nyquist plot encircles the point 1/N (A+ , 0 ), we conclude that the closedloop system is unstable with the complex gain 1/N (A+ , 0 ). In such a case, the
oscillation amplitude may grow even further, and does not return to A0 . Therefore,
the limit cycle is unstable.
Similar arguments can be made for the perturbation to a smaller amplitude. For
an A < A0 , if the Nyquist plot does encircle the point 1/N (A , 0 ), the limit cycle
is stable. If the Nyquist plot does not encircle the point 1/N (A , 0 ), the limit cycle
is unstable.
When we plot 1/N (A, 0 ) in the complex plane with A as a variable, we obtain
a line with direction of the increment of A. Based on the discussion above, the way
36
Stable
N(A,w 0)
N(A,w 0)
Describing functions
37
the case when G(s) is unstable or has unstable zeros by using corresponding stability
conditions based on the Nyquist criterion. For example if G(s) is stable and has one
unstable zero, then the stability criterion for the limit cycle will be opposite to the
condition stated in the theorem, i.e., the limit cycle is stable if the plot of 1/N (A, 0 )
crosses the Nyquist plot from the outside of the encirclement to the inside of the
encirclement as A increases.
K
with K a
s(s + 1)(s + 2)
positive constant and an ideal relay in a closed loop, as shown in Figure 3.8. We will
determine if there exists a limit cycle and analyse the stability of the limit cycle.
K
s(s + 1)(s + 2)
4M
. For the transfer function, we can
A
K
j(j + 1)(j + 2)
=K
32 j(2 2 )
.
(32 )2 + 2 (2 2 )2
From
G(j) =
1
,
N
(2 2 )
= 0,
(32 )2 + 2 (2 2 )2
A
32
,
=
2
2
2
2
2
(3 ) + (2 )
4M
38
(3.18)
We will use describing function analysis to predict the existence of a limit cycle, and
compare the predicted amplitudes and periods for different values with the simulated
ones.
To use the describing analysis, we need to formulate the system in the format of
one linear transfer function and a nonlinear element. Rearranging (3.18), we have
y y + y =
d 3
y .
dt
0.25
0.2
0.15
0.1
0.05
0
0.05
0.1
0.15
0.2
0.25
34
36
38
40
42
44
Time (s)
46
Describing functions
39
Hence, the system (3.18) can be described by a closed-loop system with a nonlinear
component
f (x) = x3
and a linear transfer function
G(s) =
s
.
s2 s + 1
3 2
A.
4
Setting
G(j) =
1
N
= 0,
we have
(1 2 )
= 0,
(1 2 )2 + 2 2
which gives = 1. From the equation for the real part, we obtain
4
2 2
= 2
(1 2 )2 + 2 2
3A
0.5
0.5
10
15
20
25
30
35
Time (s)
0.5
40
0.5
30
40
50
60
70
80
90
100
110
120
Chapter 4
Stability theory
For control systems, design, one important objective is to ensure the stability of the
closed-loop system. For a linear system, the stability can be evaluated in time domain
or frequency domain, by checking the eigenvalues of the system matrix or the poles
of the transfer function. For nonlinear systems, the dynamics of the system cannot be
described by equations in linear state space or transfer functions in general. We need
more general denitions about the stability of nonlinear systems. In this chapter, we
will introduce basic concepts of stability theorems based on Lyapunov functions.
(4.1)
x = f (x, u),
(4.2)
where u R is the control input. With u as an external signal, the system (4.2) is
not autonomous. However, for such a system, if we design a feedback control law
u = g(x) with g : Rn Rm as a continuous function, the closed-loop system
m
x = f (x, g(x))
becomes an autonomous system.
42
In this chapter, we will present basic denitions and results for stability of
autonomous systems. As discussed above, control systems can be converted to
autonomous systems by state feedback control laws.
There are many different denitions of stability for dynamics systems. Often
different denitions are needed for different purposes, and many of them are actually
the same when the system is linear. Among different denitions, the most fundamental
one is the Lyapunov stability.
Denition 4.1 (Lyapunov stability). For the system (4.1), the equilibrium point x = 0
is said to be Lyapunov stable if for any given positive real number R, there exists a
positive real number r to ensure that x(t) < R for all t 0 if x(0) < r. Otherwise
the equilibrium point is unstable.
x(t)
r
x(0)
x(0) = x0 ,
where
A=
with > 0. For this linear system, we can explicitly solve the differential equation
to obtain
x(t) =
cos t
sin t
sin t
x.
cos t 0
Stability theory
43
It is easy to check that we have x(t) = x0 . Hence, to ensure that x(t) R, we
only need to set r = R, i.e., if x0 R, we have x(t) R for all t > 0.
Note that for the system in Example 4.1, the system matrix has two eigenvalues
on the imaginary axis, and this kind of systems is referred to as critically stable in
many undergraduate texts. As shown in the example, this system is Lyapunov stable.
It can also be shown that for a linear system, if all the eigenvalues of the system
matrix A are in the closed left half of the complex plane, and the eigenvalues on the
imaginary axis are simple, the system is Lyapunov stable. However, if the system
matrix has multiple poles on the imaginary axis, the system is not Lyapunov stable.
For example let x 1 = x2 , and x 2 = 0 with x1 (0) = x1,0 , x2 (0) = x2,0 . It is easy to obtain
that x1 (t) = x1,0 + x2,0 t and x2 (t) = x2,0 . If we want x(t) R, there does not exist
a positive r for x(0) r to guarantee x(t) R. Therefore, this system is not
Lyapunov stable.
For linear systems, when a system is stable, the solution will converge to the equilibrium point. This is not required by Lyapunov stability. For more general dynamic
systems, we have the following denition concerning with the convergence to the
equilibrium.
Denition 4.2 (Asymptotic stability). For the system (4.1), the equilibrium point
x = 0 is asymptotically stable if it is stable (Lyapunov) and furthermore limt
x(t) = 0.
x(t)
r
x(0)
(4.3)
44
For linear systems, the stability properties are relatively simple. If a linear system
is asymptotically stable, it can be shown that it is exponentially stable. Of course,
for nonlinear systems, we may have a system that is asymptotically stable, but not
exponentially stable.
Example 4.2. Consider a nonlinear system
x(0) = x0 > 0,
x = x3 ,
where x R. Let us solve this differential equation. From the system equation we
have
dx
= dt,
x3
which gives
1
1
= 2t
x2 (t) x02
and
x(t) =
x0
1 + 2x02 t
It is easy to see that x(t) decreases as t increases, and also limt x(t) = 0. Therefore,
this system is asymptotically stable. However, this system is not exponentially stable,
as there does not exist a pair of a and to satisfy
x0
1+
2x02 t
ax0 e t .
1 + 2x02 te t
1
,
a
which is not satised for any choices of a and , because the left-hand side converges
to zero. Hence, the system considered in this example is asymptotically stable, but
not exponentially stable.
Lyapunov, asymptotic and exponential, stabilities are dened around equilibrium
points. If the properties hold for any initial points in the entire state space, they are
Stability theory
45
If the linearised system has all the systems poles in the open left half of the complex
plane, the equilibrium point is asymptotically stable for the actual nonlinear
system.
If the linearised system has poles in the open right half of the complex plane, then
the equilibrium point is unstable.
If the linearised system has poles on the imaginary axis, then the stability of the
original system cannot be concluded using the linearised model.
We do not show a proof of this theorem here. It is clear that this theorem can
be applied to check local stabilities of nonlinear systems around equilibrium points.
For the case that the linearised model has poles on the imaginary axis, this theorem
cannot give conclusive result about the stability. This is not a surprise, because stable
and unstable systems can have the same linearised model. For example the systems
x = x3 and x = x3 have the same linearised model at x = 0, that is x = 0, which is
marginally stable. However, as we have seen in Example 4.2, the system x = x3 is
asymptotically stable, and it is not difcult to see that x = x3 is unstable. For both the
stable and unstable cases of linearised models, the linearised model approximates the
original system better when the domain around the equilibrium point gets smaller.
Hence, the linearised model is expected to reect on the stability behaviours around
the equilibrium point.
46
1
1
1
0
(4.4)
Stability theory
47
Since V (x) is positive denite, we have a > 0. We then dene the level set within BR
c := {x BR |V (x) < c},
where c is a positive real constant and c < a. The existence of such a positive real
constant c is guaranteed by the continuity and positive deniteness of V . From the
denition of c , x c implies that x < R. Since V 0, we have V (x(t))
V (x(0)). Hence, for any x(0) c , we have
V (x(t)) V (x(0)) < c,
which implies
x(t) < R.
Since c contains 0 as an interior point, and V is a continuous function, there
must exist a positive real r such that
Br := {x|x < r} c .
Hence, we have
Br c BR .
48
where := {x D|V (x) < }. Since V is negative denite, we have > 0. From
the denition of , we have
V (x(t)) V (x(0)) t.
The right-hand side turns to negative when t is big enough, which is a contradiction.
Therefore, we can conclude that limt V (x(t)) = 0, which implies limt x(t) = 0.
2
Example 4.4. A pendulum can be described by
+ + sin = 0,
we re-write the dynamic system as
where is the angle. If we let x1 = and x2 = ,
x 1 = x2
x 2 = sin x1 x2 .
Consider the scalar function
V (x) = (1 cos x1 ) +
x22
.
2
(4.5)
The rst term (1 cos x1 ) in (4.5) can be viewed as the potential energy and the
x2
second term 22 as the kinetic energy. This function is positive denite in the domain
D = {|x1 | , x2 R}. A direct evaluation gives
Stability theory
49
V (x) = sin x1 x 1 + x2 x 2
= x22 .
Hence, the system is stable at x = 0. However, we cannot conclude the asymptotic
stability of the system from Theorem 4.2. This system is in fact asymptotically stable
by using more advanced stability theorem such as invariant set theorem, which is not
covered in this book.
When establishing global stability using Lyapunov functions, we need the function V (x) to be unbounded as x tends to innity. This may sound strange. The reason
behind this point is that we need the property that if V (x) is bounded, then x is
bounded, in order to conclude the boundedness of x from the boundedness of V (x).
This property is dened in the following function as the radial unboundedness of V .
Denition 4.8 (Radially unbounded function). A positive denite function V (x) :
Rn R is said to be radially unbounded if V (x) as x .
Theorem 4.3 (Lyapunov theorem for global stability). For the system (4.1) with D =
Rn , if there exists a function V (x) : Rn R with continuous rst order derivatives
such that
x(0) = x0 > 0,
50
1 2
x
2
and it is easy to see that this function is globally positive denite. Its derivative is
given by
V = x4
which is negative denite. Hence, from Theorem 4.3, we conclude x = 0 is
asymptotically stable.
To conclude this section, we have another result for exponential stability.
Theorem 4.4 (Exponential stability). For the system (4.1), if there exists a function
V (x) : D Rn R with continuous rst-order derivatives such that
a1 xb V (x) a2 xb ,
V
f (x) a3 xb ,
x
(4.6)
(4.7)
t 0
(4.8)
implies that
V (t) eat V (0) +
t
0
e(t ) g( )d ,
t 0
(4.9)
Stability theory
51
(4.10)
(4.11)
Multiplying both sides of (4.11) by ea gives (4.8). This completes the proof.
V (0)
e(a3 /a2 b)t
a1
1/b
a2
(4.12)
a3
,
a2 b
(4.13)
52
where x Rn and A Rnn , x is the state variable, and A is a constant matrix. From
linear system theory, we know that this system is stable if all the eigenvalues of A are
in the open left half of the complex plane. Such a matrix is referred to as a Hurwitz
matrix. Here, we would like to carry out the stability analysis using a Lyapunov
function. We can state the stability in the following theorem.
Theorem 4.6. For the linear system shown in (4.13), the equilibrium x = 0 is globally
and exponentially stable if and only if there exist positive denite matrices P and Q
such that
AT P + PA = Q
(4.14)
holds.
Proof. For sufciency, let
V (x) = xT Px,
(4.15)
(4.16)
Let us use max () and min () to denote maximum and minimum eigenvalues of a
positive denite matrix. From (4.15), we have
min (P)x2 V (x) max (P)x2 .
(4.17)
min (Q)
x2 .
max (P)
(4.18)
Now we can apply Theorem 4.4 with (4.17) and (4.18) to conclude that the equilibrium
point is globally and exponentially stable. Furthermore, we can identify a1 = min (P),
a2 = max (P), a3 = min (Q) and b = 2. Following the proof of Theorem 4.4, we have
x(t)
(Q)
max (P)
min
t
x(0)e 2max (P) .
min (P)
(4.19)
Stability theory
53
for some positive real constants a and , which implies limt x(t) = 0. Since
x(t) = eAt x(0),
we can conclude limt eAt = 0. In such a case, for a positive denite matrix Q, we
can write
(4.20)
= AT
exp (AT t)Q exp (At)t +
Let
P=
and if we can show that P is positive denite, then we obtain (4.14), and hence
complete the proof. Indeed, for any z Rn = 0, we have
z T Pz =
Since Q is positive denite, and eAt is non-singular for any t, we have z T Pz > 0, and
therefore P is positive denite.
2
Chapter 5
Lyapunov direct method provides a tool to check the stability of a nonlinear system
if a Lyapunov function can be found. For linear systems, a Lyapunov function can
always be constructed if the system is asymptotically stable. In many nonlinear systems, a part of the system may be linear, such as linear systems with memoryless
nonlinear components and linear systems with adaptive control laws. For such a system, a Lyapunov function for the linear part may be very useful in the construction
for the Lyapunov function for the entire nonlinear system. In this chapter, we will
introduce one specic class of linear systems, strict positive real systems, for which,
an important result, KalmanYakubovich lemma, is often used to guarantee a choice
of the Lyapunov function for stability analysis of several types of nonlinear systems.
The application of KalmanYakubovich lemma to analysis of adaptive control systems will be shown in later chapters, while in this chapter, this lemma is used for
stability analysis of systems containing memoryless nonlinear components and the
related circle criterion. In Section 5.3 of this chapter, input-to-state stability (ISS) is
briey introduced.
For analysis of adaptive control systems, strictly positive real systems are more
widely used than the positive real systems.
56
Denition 5.2. A proper rational transfer function G(s) is strictly positive real if
there exists a positive real constant such that G(s ) is positive real.
Example 5.1. For the transfer function G(s) =
s = + j,
G(s) =
1
,
s+a
1
a + j
=
a + + j
(a + )2 + 2
and
(G(s)) =
Hence, G(s) =
a+
> 0.
(a + )2 + 2
1
s+a
(G(s )) =
and therefore G(s) =
a+
>0
(a + )2 + 2
1
s+a
Denition 5.1 shows that a positive real transfer function maps the closed right
half of the complex plane to itself. Based on complex analysis, we can obtain the
following result.
Proposition 5.1. A proper rational transfer function G(s) is positive real if
all the poles of G(s) are in the closed left half of the complex plane
any poles on the imaginary axis are simple and their residues are non-negative
for all R, (G(j)) 0 when j is not a pole of G(s)
1
It can be seen that G(s) = s+a
, with a < 0, is not positive real. If G(s) is positive real, we must have (G(j)) 0. Similarly, other necessary conditions can be
obtained for a transfer function G(s) to be positive real. We can state those conditions
in an opposite way.
Proposition 5.2. A transfer function G(s) cannot be positive real if one of the
following conditions is satised:
s2
57
s1
, G2 =
+ as + b
s+1
1
and G3 = 2
are not positive real, for any real numbers a
s2 s + 1
s + as + b
and b, because they are non-minimum phase, unstable, and with relative degree 2
s+4
respectively. It can also be shown that G(s) = 2
is not positive real as
s + 3s + 2
1
= 2
.
(G(s)) =
+ j
+ 2
Therefore, G(s) =
1
s
1
s
For the stability analysis later in the book, we only need the result on strictly
positive real transfer functions.
Lemma 5.3. A proper rational transfer function G(s) is strictly positive real if and
only if
G(s) is Hurwitz, i.e., all the poles of G(s) are in the open left half of the complex
plane.
The real part of G(s) is strictly positive along the j axis, i.e.,
0,
(G(j)) > 0,
Proof. We show the proof for sufciency here, and omit the necessity, as it is more
involved. For sufciency, we only need to show that there exists a positive real constant
such that G(s ) is positive real.
Since G(s) is Hurwitz, there must exist a positive real constant such that for
(0, ], G(s ) is Hurwitz. Suppose (A, b, cT , d) is a minimum state space
realisation for G(s), i.e.,
G(s) = cT (sI A)1 b + d.
We have
G(s ) = cT (sI I A)1 b + d
= cT (sI A)1 ((sI I A) + I )(sI I A)1 b + d
= G(s) + E(s),
(5.1)
58
where
E(s) = cT (sI A)1 (sI I A)1 b.
E(s) is Hurwitz, and strictly proper. Therefore, we have
(E(j)) < r1 ,
R, (0, ]
(5.2)
for some positive real r1 and the existence of lim 2 (E(j)), which implies
2 (E(j)) < r2 ,
(5.3)
(5.4)
for some r3 > 0. Hence, combining (5.2) and (5.4), we obtain, from (5.1), that
(G(j )) > r3 r1 ,
R.
(5.5)
implies that
2 (G(j)) > r4 ,
for || > 2
(5.6)
for some positive reals r4 and 2 . From (5.1), (5.3) and (5.6), we obtain that
2 (G(j )) > r4 r2 ,
for || > 3
(5.7)
where 3 = max{1 , 2 }. From the second condition of the lemma, we have, for some
positive real constant r5 ,
(G(j)) > r5 ,
for || 3 .
(5.8)
for || 3 .
(5.9)
Combining the results in (5.7) and (5.9), we obtain that (G(j )) > 0 by setting
= min{ rr42 , rr51 }. Therefore, we have shown that there exists a positive real such that
G(s ) is positive real.
2
59
The main purpose of introducing strictly positive real systems is for the following
result, which characterises the systems using matrices in time domain.
Lemma 5.4 (KalmanYakubovich lemma). Consider a dynamic system
x = Ax + bu
y = cT x,
(5.10)
where x Rn is the state variable; y and u R are the output and input respectively;
and A, b and c are constant matrices with proper dimensions, and (A, b, cT ) is
controllable and observable. Its transfer function G(s) = cT (sI A)1 b is strictly
positive real if and only if there exist positive denite matrices P and Q such that
AT P + PA = Q,
Pb = c.
(5.11)
Remark 5.1. We do not provide a proof here, because the technical details in the
proof such as nding the positive denite P and the format of Q are beyond the scope
of this book. In the subsequent applications for stability analysis, we only need to
know the existence of P and Q, not their actual values for a given system. For example
in the stability analysis for adaptive control systems in Chapter 7, we only need to
make sure that the reference model is strictly positive real, which then implies the
existence of P and Q to satisfy (5.11).
(5.12)
u = F(y)y,
where x Rn is the state variable; y and u R are the output and input respectively;
and A, b and c are constant matrices with proper dimensions. The nonlinear component
is in the feedback law. Similar systems have been considered earlier using describing
functions for approximation to predict the existence of limit cycles. Nonlinear elements considered in this section are sector-bounded, i.e., the nonlinear feedback gain
can be expressed as
< F(y) <
for some constants and , as shown in Figure 5.2.
(5.13)
60
.
x = Ax + bu
y = cTx
F(y)
F(y)y
b
61
Note that the conditions specied in Lemma 5.4 are the sufcient conditions.
With the result shown in Lemma 5.5, we are ready to consider the general case
for < F(y) < . Consider the function dened by
F
F =
F
(5.14)
and obviously we have F > 0. How to use this transformation for analysis of systems
stability?
With G(s) = cT (sI A)1 b, the characteristic equation of (5.12) can be
written as
G(s)F + 1 = 0.
(5.15)
(5.16)
G(s)( F) = G + 1.
(5.17)
+ 1 = 0.
1 + G F
(5.18)
:= 1 + G
G
1 + G
(5.19)
Let
(5.20)
62
which implies that the stability of the system (5.12) with the nonlinear gain shown in
(5.13) is equivalent to the stability of the system with the forward transfer function G
and the feedback gain F. Based on Lemma 5.5 and (5.20), we can see that the system
is strictly positive real.
(5.12) is stable if G
in (5.19) cannot deal with the case = .
The expressions of F in (5.14) and G
In such a case, we re-dene
F = F
(5.21)
G
.
1 + G
(5.22)
G(s)
=
1 + G(s)
or in case of = , by
G(s)
=
G(s)
1 + G(s)
is strictly positive real, with G(s) = cT (sI A)1 b, the system is absolutely stable.
is strictly positive real? Let us assume that
What is the condition of G if G
> > 0. Other cases can be analysed similarly. From Lemma 5.3, we know that
to be strictly positive real, we need G
to be Hurwitz, and (G(j))
for G
> 0, that is
1 + G(j)
1 + G(j)
> 0,
R,
which is equivalent to
1/ + G(j)
1/ + G(j)
> 0,
R.
(5.23)
63
< 1 2 < ,
2
2
which is equivalent to the point G(j) that lies outside the circle centered at
( 12 (1/ + 1/), 0) with radius of 12 (1/ 1/) in the complex plane. This circle intersects the real axis at ( 1 , 0) and ( 1 , 0). Indeed, 1 + G(j) is represented
as a vector from the point ( 1 , 0) to G(j), and 1 + G(j) as a vector from the
point ( 1 , 0) to G(j). The angle between the two vectors will be less than 2 when
G(j) is outside the circle, as shown in Figure 5.3. Since the condition must hold
1
G
.
(5.24)
The mapping shown in (5.24) is a bilinear transformation, and it maps a line to a line
or circle. For the case of > > 0, we have
1
1
1
/
G=
.
(5.25)
G
The function
/
G /
Im
G( jw)
q2
1
a
q1
Re
1
b
64
G( jw)
1
a
1
b
Re
G
maps the imaginary axis to a circle with the diameter on the line from (0, 0) to
( 1 1 , 0) on the complex plane. Finally, it can be seen from (5.25) that this map of G
1
to G maps the imaginary axis to the circle with the diameter on the line from ( , 0)
to ( 1 , 0), or in other words, the circle centered as ( 12 (1/ + 1/), 0) with radius
of 12 (1/ 1/). It can also be shown that the function maps the open left-hand
complex plane to the domain inside the circle.
Indeed, we can evaluate the circle directly from (5.24). Let u and v denote the
real and imaginary parts of the mapping of the imaginary axis, and we have
u =
j 1
+ 2
= 2
,
j
+ 2 2
v =
j 1
( )
.
= 2
j
+ 2 2
Denoting = , we obtain
1/ + (1/)2
1
1 1
1 1 2
1 1
+
+
,
1 + 2
2
2
1 + 2
(1/ 1/)
1 1
1
2
v =
=
.
2
1+
2
1 + 2
u =
65
1
u+
2
1
1
+
2
2
1 1
1
+v =
.
(5.26)
x(0) = 0
From the above example, it can be seen that even the corresponding autonomous
system is asymptotically stable, the state may not remain bounded subject to a bounded
input. We introduce a denition for systems with the property of bounded state with
bounded input.
To show a denition of input-to-state stable (ISS), we need to use comparison
functions, which are dened below.
66
Denition 5.3. A function : [0, a) [0, ) is a class K function if is continuous, and strictly increasing with (0) = 0. If a = and limr (r) = , the
function is a class K function.
Denition 5.4. A function : [0, a) [0, ) [0, ) is a class KL function if it
is continuous, for a xed t = t0 , (, t0 ) is a class K function, and for a xed x(0),
limt (x(0), t) = 0.
Denition 5.5. The system (5.26) is ISS if there exist a class KL function and a
class K function such that
x(t) (x(0), t) + (u ),
t > 0.
(5.27)
t > 0.
(5.28)
eA(t ) Bu( )d ,
67
we have
eA(t ) d Bu .
Since A is Hurwitz, there exist positive real constants a and such that eA(t)
aet . Hence, we can obtain
t
x(t) aet x(0) +
ae(t ) d Bu
0
ae
a
x(0) + Bu
It is easy to see that the rst term in the above expression is a KL function of t and
x(0) and the second term is a K function of u . Therefore, the linear system with
Hurwitz system matrix A is ISS.
Next, we show a result on establishing ISS property from a Lyapunov function.
Theorem 5.8. For the system (5.26), if there exists a function V (x) : Rn R with
continuous rst-order derivatives such that
a1 xb V (x) a2 xb ,
V
f (x, u) a3 xb ,
x
x (u)
(5.29)
(5.30)
where a1 , a2 , a3 and b are positive real constants and is a class K function, the
system (5.26) is ISS.
Proof. From Theorem 4.4, we can see that the system is exponentially stable when
u = 0, and V (x) is a Lyapunov function for the autonomous system x = f (x, 0). To
consider the case for non-zero input, let us nd a level set based on V , dened by
c := {x|V (x) c}
and determine the constant c such that for x
/ c , we have x > (u ). Indeed,
let
c = a2 ((u ))b .
In this case, from V (x) > c, we have
V (x) > a2 ((u ))b ,
which implies that
a2 xb > a2 ((u ))b
and hence x > (u ).
68
Therefore, for any x outside c , we can obtain from (5.29) and (5.30), in a similar
way to the proof of Theorem 4.4,
a3
V V
a2
and then
x(t)
a2
a1
1/b
x(0)e
a
2
a 3b t
a2
a1
1/b
(u ).
a2
a1
1/b
x(0)e
a
2
a 3b t
a2
,
a1
1/b
(u ) .
Hence, the system is ISS with the gain function () = ( aa21 )1/b ().
There is a more general result than Theorem 5.8 that requires x = f (x, 0) to be
asymptotically stable, not necessarily exponential stable. The proof of that theorem
is beyond the level of this text, and we include it here for completeness.
Theorem 5.9. For the system (5.26), if there exists a function V (x) : Rn R with
continuous rst-order derivatives such that
1 (x) V (x) 2 (x),
V
f (x, u) 3 (x),
x
x (u),
(5.31)
(5.32)
69
(5.33)
V
f (x) (x) + (u),
x
(5.34)
1 2
x .
2
1 4 3 4/3
|x| + |u|
4
4
70
which gives
3
3
V x4 + |u|4/3 .
4
4
Therefore, the system is ISS based on Corollary 5.10 with () = 34 ()4 and
() = 34 ()4/3 .
ISS property is useful in establishing the stability of interconnected systems. We
include two results here to end this section.
Theorem 5.11. If for the cascade connected system
x 1 = f1 (x1 , x2 ),
(5.35)
x 2 = f1 (x2 , u),
(5.36)
the subsystem (5.35) is ISS with x2 as the input, and the subsystem (5.36) is ISS with
u as input, the overall system with state x = [x1T , x2T ]T and input u is ISS.
Theorem 5.12 (ISS small gain theorem). If for the interconnected system
x 1 = f1 (x1 , x2 ),
(5.37)
x 2 = f1 (x2 , x2 , u),
(5.38)
the subsystem (5.37) is ISS with x2 as the input with 1 as the ISS gain for x2 , and
the subsystem (5.38) is ISS by viewing x1 and u as the inputs, with the ISS input gain
function 2 for x1 , the overall system with state x = [x1T , x2T ]T and input u is ISS if
1 (2 (r)) < r,
r > 0.
(5.39)
From Theorem 5.11, it can be seen that if the subsystem x2 is globally asymptotically stable when u = 0, the overall system is globally asymptotically stable. Similarly,
Theorem 5.12 can be used to establish the stability of the following system:
x 1 = f1 (x1 , x2 )
x 2 = f1 (x2 , x2 ),
and the global and asymptotic stability of the entire system can be concluded if the
gain condition shown in Theorem 5.12 is satised.
Example 5.6. Consider the second-order nonlinear system
x 1 = x13 + x2 ,
2/3
x 2 = x1 x2 3x2 .
71
From Example 5.5, we know that x1 -subsystem is ISS with the gain function
1 () = (2)1/3 . For the x2 -subsystem, we choose
V2 =
1 2
x
2 2
and we have
5/3
V 2 = 3x22 + x1 x2
(5.40)
(5.41)
72
Remark 5.2. The conditions specied in (5.41) are useful for observer design, in
particular, for the stability analysis of the reduced-order observers in Chapter 8.
A similar denition to differential stability is incremental stability. However, the
conditions specied in (5.41) are not always satised by the systems with incremental
stability. When x = 0, i.e., in the case for one system only, the conditions specied
in (5.41) are then similar to the properties of the nonlinear systems with exponential
stability. The last condition in (5.41) is specied for interactions with other systems.
This condition is similar to the conditions for the existence of changing the supply
functions for inter-connection of ISS systems.
We include two illustrative examples below for the properties of differential
stability.
Example 5.7. A linear system is differentially stable if the system is asymptotically
stable. Consider
x = Ax,
where A Rnn . If the system is asymptotically stable, A must be Hurwitz. Therefore
there exist positive denite matrices P and Q such that
AT P + PA = Q.
Let V (x) = xT Px. In this case, the conditions (5.41) are satised with
1 (x) = min (P)x2 ,
2 (x) = max (P)x2 ,
3 (x) = min (Q)x2 ,
c1 =
min (Q)
, c2 = 2,
4(max (P))2
where min () and max () denote the minimum and maximum eigenvalues of a
positive denite matrix.
The differential stability is closely related to observer design. Consider a system
with input u Rs
x = Ax + Bu.
(5.42)
x (0) = 0.
(5.43)
73
2
1
x2 .
V 1
x2 .
Hence, the system is exponentially stable. But this system is not differentially stable.
Indeed, let e = x x . We have
e = e 2( sin (x) sin (x + e)).
By linearising the system at x = and x = , and denoting the error at this point by
el , we have
e l = el 2 cos (x)|x= (x ) + 2 cos (x)|x = (x )
= el
and the system is unstable in a neighbourhood of this point.
Chapter 6
Feedback linearisation
Nonlinear systems can be linearised around operating points and the behaviours in
the neighbourhoods of the operating points are then approximated by their linearised
models. The domain for a locally linearised model can be fairly small, and this may
result in that a number of linearised models are needed to cover an operating range
of a system. In this chapter, we will introduce another method to obtain a linear
model for nonlinear systems via feedback control design. The aim is to convert a
nonlinear system to a linear one by state transformation and redening the control
input. The resultant linear model describes the system dynamics globally. Of course,
there are certain conditions for the nonlinear systems to satisfy so that this feedback
linearisation method can be applied.
(6.1)
76
Viewing v as the new control input, we see that the system is linearised. Indeed, let
us introduce a state transformation
1 := y = x1 ,
2 := y = x2 + x13 .
We then obtain a linear system
1 = 2
2 = v.
We can design a state feedback law as
v = a1 1 a2 2
to stabilise the system with a1 > 0 and a2 > 0. The control input of the original system
is given by
u = 3x12 (x2 x13 ) x12 + v.
(6.2)
We say that the system (6.1) is linearised by the feedback control law (6.2). Notice
that this linearisation works for the entire state space.
As shown in the previous example, we can keep taking the derivatives of the
output y until the input u appears in the derivative, and then a feedback linearisation
law can be introduced. The derivatives of the output also introduce a natural state
transformation.
Consider a nonlinear system
x = f (x) + g(x)u
y = h(x),
(6.3)
where x D Rn is the state of the system; y and u R are output and input respectively; and f and g : D Rn Rn are smooth functions and h : D Rn R is a
smooth function.
Remark 6.1. The functions f (x) and g(x) are vectors for a given point x in the state
space, and they are often referred to as vector elds. All the functions in (6.3) are
required to be smooth in the sense that they have continuous derivatives up to certain
orders when required. We use the smoothness of functions in the remaining part of
the chapter in this way.
The inputoutput feedback linearisation problem is to design a feedback control
law
u = (x) + (x)v
with (x) = 0 for x D such that the inputoutput dynamics of the system
(6.4)
Feedback linearisation
x = f (x) + g(x)(x) + g(x)(x)v
y = h(x)
77
(6.5)
are described by
y() = v
(6.6)
for 1 n.
For the system (6.3), the rst-order derivative of the output y is given by
h(x)
(f (x) + g(x)u)
x
:= Lf h(x) + Lg h(x)u,
y =
h(x)
f (x).
x
for k = 0, . . . , 2,
Lg Lf1 h(x) = 0.
Example 6.2. Consider the system (6.1). Comparing it with the format shown in
(6.3), we have
3
x1 + x 2
0
f (x) =
, g(x) =
, h(x) = x1 .
1
x12
78
For SISO linear systems, the relative degree is the difference between the orders
of the polynomials in the numerator and denominator of the transfer function. With
the denition of the relative degree, we can present the inputoutput feedback
linearisation using Lie derivatives.
Example 6.3. Consider the system (6.1) again, and continue from Example 6.2. With
Lg h(x) = 0, we have
y = Lf h(x)
where
Lf h(x) = x13 + x2 .
Taking the derivative of Lf h(x), we have
y = Lf2 h(x) + Lg Lf h(x)u
where
Lf2 h(x) = 3x12 (x2 + x13 ) + x12 ,
Lg Lf h(x) = 1.
Therefore, we have
v = Lf2 h(x) + Lg Lf h(x)u
or
u=
Lf2 h(x)
Lg Lf h(x)
1
v
Lg Lf h(x)
The procedure shown in Example 6.3 works for systems with any relative degrees.
Suppose that the relative degree for (6.3) is , which implies that Lg Lfk h(x) = 0 for
k = 0, . . . , 2. Therefore, we have the derivatives of y expressed by
y(k) = Lfk h(x),
for k = 0, . . . , 1,
(6.7)
(6.8)
Feedback linearisation
79
1
Lg Lf1 h(x)
(Lf h(x) + v)
it results in
y() = v.
We can consider to use i := y(i1) = Lfi1 h as coordinates for the linearised
has full
inputoutput dynamics. The only remaining issue is to establish that
x
rank. To do that, we need to introduce a few notations.
For any smooth functions f , g : D Rn Rn , the Lie bracket [f , g] is
dened by
[f , g](x) =
g(x)
g(x)
f (x)
f (x)
x
x
and we can introduce a notation which is more convenient for high-order Lie
brackets as
adf0 g(x) = g(x),
adf1 g(x) = [f , g](x),
adfk g(x) = [f , adfk1 g](x).
For the convenient of presentation, let us denote
dh =
h
x
(6.9)
80
for k = 0, . . . , 2
for k + l 2,
(6.10)
and
< dLfk h, adfl g > = (1)l < dLf1 h, g >,
for k + l = 1.
(6.11)
dh(x)
dLf h(x)
1
dLf h
0
...
0
(1)1 r(x)
0
. . . (1)2 r(x)
= .
.
.
.
..
..
..
..
r(x) . . .
(6.12)
dh(x)
dLf h(x)
..
=
x
.
dLf1 h
has full rank. We summarise the result about inputoutput feedback linearisation in
the following theorem.
Theorem 6.1. If the system in (6.3) has a well-dened relative degree in D, the
inputoutput dynamics of the system can be linearised by the feedback control law
u=
1
Lg Lf1 h(x)
(Lf h(x) + v)
(6.13)
Feedback linearisation
1 = 2
..
.
1 =
81
(6.14)
= v
with a partial state transformation
i = Lfi1 h(x) for i = 1, . . . , .
(6.15)
Remark 6.2. The inputoutput dynamics can be linearised based on Theorem 6.1. In
the case of < n, the system for < n can be transformed under certain conditions
to the normal form
z = f0 (z, ),
1 = 2 ,
..
.
1 = ,
= Lf h + uLg Lf1 h,
y = 1
where z Rn is the part of the state variables which are not in the inputoutput
dynamics of the system, and f0 : Rn Rn is a smooth function. It is clear that
when < n, the inputoutput linearisation does not linearise the dynamics
z = f0 (z, ). Also note that the dynamics z = f0 (z, 0) are referred to as the zero
dynamics of the system.
To conclude this section, we use an example to demonstrate the inputoutput
linearisation for the case < n.
82
3
x1 + x2
f (x) = x12 + x3 ,
x12
0
g(x) = 1 ,
1
h(x) = x1 .
1
(Lf2 h + v)
Lg Lf h
Feedback linearisation
83
With the coordinates z, 1 and 2 , we have the system in the normal form
z = z 2 + 13
1 = 2
2 = z + 2 + 12 13 + 312 2 + u.
It is clear that the inputoutput linearisation does not linearise the dynamics of z. Also
note that the zero dynamics for this system are described by
z = z.
(6.16)
(6.17)
with (x) = 0 for x D such that the entire system dynamics are linearised by a state
transformation with v as the new control input.
It is clear from the results shown in the inputoutput linearisation that the complete linearisation can only be achieved when the relative degree of the system equals
the order of the system. If we can nd an output function h(x) for the system (6.16),
the inputoutput linearisation result shown in the previous section can be applied
to solve the full-state feedback linearisation problem. Therefore, we need to nd an
output function h(x) such that
Lg Lfk h(x) = 0
for k = 0, . . . , n 2
Lg Lfn1 h(x) = 0
(6.18)
(6.19)
Based on Lemma 6.2, the condition specied in (6.18) is equivalent to the condition
Lad k g h(x) = 0
f
for k = 0, . . . , n 2
(6.20)
(6.21)
The output function h(x) that satises the condition shown in (6.20) is a solution of
the partial differential equation
[g, adf g, . . . , adfn2 g]
h
= 0.
x
(6.22)
84
To discuss the solution of this partial differential equation, we need a few notations
and results. We refer to a collection of vector elds as a distribution. For example if
f1 (x), . . . , fk (x) are vector elds, with k a positive integer,
= span{f1 (x), . . . , fk (x)}
is a distribution. The dimension of distribution is dened as
dim ((x)) = rank[f1 (x), . . . , fn (x)].
A distribution is said to be involutive, if for any two vector elds f1 , f2 , we
have [f1 , f2 ] . Note that not all the distributions are involutive, as shown in the
following example.
Example 6.5. Consider the distribution
= span{f1 (x), f2 (x)}
where
2x2
f1 (x) = 1 ,
0
1
f2 (x) = 0 .
x2
0 0
= 0 0
0 1
0
= 0.
1
[f1 , f2 ] =
f1
f2
x
0
2x2
0
0 1 0
0
0
0
2
0
0
0
1
0 0
0
x2
1
0
x2
0
0 = 1.
1
Feedback linearisation
85
the matrix G = [g(x), adf g(x), . . . , adfn1 g(x)] has full rank
the distribution Gn1 = span{g(x), adf g(x), . . . , adfn2 g(x)} is involutive
Proof. For the sufciency, we only need to show that there exists a function h(x) such
that the relative degree of the system by viewing h(x) as the output is n, and the rest
follows from Theorem 6.1. From the second condition that Gn1 is involutive, and
Frobenius theorem, there exists a function h(x) such that
h
[g, adf g, . . . , adfn2 g] = 0
x
which is equivalent to, from Lemma 6.2,
Lg Lfk h(x) = 0,
for k = 0, . . . , n 2.
for
k = 0, . . . , n 2,
which implies
h
[g, adf g, . . . , adfn2 g] = 0.
x
86
From Frobenius theorem, we conclude that Gn1 is involutive. Furthermore, from the
fact that the system with h(x) as the output has a relative degree n, we can show, in
the same way as the discussion that leading to Lemma 6.2, that
dh(x)
dLf h(x)
n1
..
.
dLfn1 h
0
...
0
(1)n1 r(x)
0
. . . (1)n2 r(x)
= .
.
.
.
..
..
..
..
r(x) . . .
where r(x) = Lg Lfn1 h(x). This implies that G has rank n. This concludes the
proof.
2
Remark 6.3. Let us see the conditions in Theorem 6.4 for linear systems with
f (x) = Ax, g(x) = b.
where A is a constant matrix and b is a constant vector. A direct evaluation gives
[f , g] = Ab,
and
adfk g = (1)k Ak b
for k > 0. Therefore, we have
G = [b, Ab, . . . , (1)n1 An1 b].
It can be seen that the full rank condition of G is equivalent to the full controllability
of the linear system.
In the next example, we consider the dynamics of the system that was considered in Example 6.4 for inputoutput linearisation for the input h(x) = x1 . We will
show that the full-state linearisation can be achieved by nding a suitable output
function h(x).
Feedback linearisation
87
3
0
x1 + x2
f (x) = x12 + x3 , g(x) = 1 ,
1
x12
we have
1
[f , g] = 1 ,
0
and
3x12 + 1
adf2 g = 2x1 .
2x1
Hence, we have
1
0
G2 = span 1 , 1
1
0
and
0 1
G = 1 1
1 0
3x12 + 1
2x1 .
2x1
0 1
h
1 1 = 0.
x 1 0
For this h(x), it is easy to check that
88
1
v (15x14 + 6x1 x2 )(x13 + x2 ) 3x12 (x3 + x12 ) x12 .
+1
3x12
Chapter 7
90
(7.1)
where y and u R are the system output and input respectively, and ap and bp are
unknown constant parameters with sgn(bp ) known. The output y is to follow the output
of the reference model
y m + am ym = bm r.
(7.2)
The reference model is stable, i.e., am > 0. The signal r is the reference input. The
design objective is to make the tracking error e = y ym converge to 0.
Let us rst design a Model Reference Control (MRC), that is, the control design
assuming all the parameters are known, to ensure that the output y follows ym .
Rearrange the system model as
a p am
y + am y = bp u
y
bp
and therefore we obtain
a p am
bm
y r
e + am e = bp u
bp
bp
:= bp (u au y ar r) ,
where
ay =
ap a m
,
bp
ar =
bm
.
bp
(7.3)
(7.4)
91
Note that the parameters ar and ay are the parameters of the controllers, and they
are related to the original system parameters ap and bp , but not the original system
parameters themselves.
The certainty equivalence principle only suggests a way to design the adaptive
control input, not how to update the parameter estimates. Stability issues must be
considered when deciding the adaptive laws, i.e., the way how estimated parameters
are updated. For rst-order systems, the adaptive laws can be decided from Lyapunov
function analysis.
With the proposed adaptive control input (7.4), the closed-loop system dynamics
are described by
e + am e = bp ( a y y a r r),
(7.5)
1 2 |bp | 2 |bp | 2
a +
a ,
e +
2
2r r
2y y
(7.6)
where r and y are constant positive real design parameters. Its derivative along the
trajectory (7.5) is given by
y
r
a
a
V = am e2 + a r |bp | ebp r + a y |bp | ebp y .
r
y
If we can set
a r
|bp | ebp r = 0,
r
a y
|bp | ebp y = 0,
y
(7.7)
(7.8)
we have
V = am e2 .
(7.9)
Noting that a r = a r and a y = a y , the conditions in (7.7) and (7.8) can be satised
by setting the adaptive laws as
a r = sgn(bp )r er,
(7.10)
a y = sgn(bp )y ey.
(7.11)
The positive real design parameters r and y are often referred to as adaptive gains,
as they can affect the speed of parameter adaptation.
From (7.9) and Theorem 4.2, we conclude that the system is Lyapunov stable
with all the variables e, a r and a y bounded, and hence the boundedness of a r and a y .
However, based on the stability theorems introduced in Chapter 4, we cannot
conclude anything about the tracking error e other than its boundedness. In order
to do it, we need to introduce an important lemma for stability analysis of adaptive
control systems.
92
(7.12)
1
.
s+2
We can directly use the result presented in Lemma 7.2, i.e., we use the adaptive
laws (7.10) and (7.11) and the control input (7.4). Since b is known, we only have one
unknown parameter, and it is possible to design a simpler control based on the same
design principle.
From the system model, we have
Gm =
y + ay = u,
which can be changed to
y + 2y = u (a 2)y.
93
y
ym
y and ym
0.8
0.6
0.4
0.2
0
0
10
10
t (s)
Estimated parameter
0
1
2
3
4
4
t (s)
94
Zp (s)
u(s),
Rp (s)
(7.13)
where y(s) and u(s) denote the system output and input in frequency domain; kp is the
high frequency gain; and Zp and Rp are monic polynomials with orders of n and
n respectively with as the relative degree. The reference model is chosen to have
the same relative degree of the system, and is described by
ym (s) = km
Zm (s)
r(s),
Rm (s)
(7.14)
where ym (s) is the reference output for y(s) to follow; r(s) is a reference input; and
km > 0 and Zm and Rm are monic Hurwitz polynomials.
Remark 7.2. A monic polynomial is a polynomial whose leading coefcient, the
coefcient of the highest power, is 1. A polynomial is said to be Hurwitz if all its roots
are with negative real parts, i.e., its roots locate in the open left half of the complex
plane. The high-frequency gain is the leading coefcient of the numerator of a transfer
function.
The objective of MRC is to design a control input u such that the output of the
system asymptotically follows the output of the reference model, i.e., limt (y(t)
ym (t)) = 0.
Note that in this chapter, we abuse the notations of y, u and r by using same
notations for the functions in time domain and their Laplace transformed functions
in the frequency domain. It should be clear from the notations that y(s) is the Laplace
transform of y(t) and similarly for u and r.
To design MRC for systems with = 1, we follow a similar manipulation to the
rst-order system by manipulating the transfer functions. We start with
y(s)Rp (s) = kp Zp (s)u(s)
and then
y(s)Rm (s) = kp Zp (s)u(s) (Rp (s) Rm (s))y(s).
95
R (s) R (s)
Note that Rp (s) Rm (s) is a polynomial with order n 1, and m Zm (s) p is a proper
transfer function, as Rp (s) and Rm (s) are monic polynomials. Hence, we can write
Zp (s)
Rm (s) Rp (s)
y(s)Rm (s) = kp Zm (s)
u(s) +
y(s) .
Zm (s)
Zm (s)
If we parameterise the transfer functions as
Zp (s)
T (s)
= 1 1
,
Zm (s)
Zm (s)
Rm (s) Rp (s)
T (s)
y(s) = 2
y(s) 3 ,
Zm (s)
Zm (s)
where 1 Rn1 , 2 Rn1 and 3 R are constants and
(s) = [sn2 , . . . , 1]T ,
we obtain that
Zm (s)
1T (s)
2T (s)
u(s)
u(s)
y(s) 3 y(s) .
y(s) = kp
Rm (s)
Zm (s)
Zm (s)
(7.15)
(7.16)
T (s)
1T (s)
u(s) + 2
y + 3 y + 4 r(s)
Zm (s)
Zm (s)
:= T ,
where
T = [1T , 4T , 3 , 4 ],
= [1T , 2T , y, r]T ,
with
1 =
(s)
u,
Zm (s)
2 =
(s)
y.
Zm (s)
(7.17)
96
Remark 7.3. The control design shown in (7.17) is a dynamic feedback con(s)
troller. Each element in the transfer matrix
is strictly proper, i.e.,
Zm (s)
with relative degree greater than or equal to 1. The total number of parameters in
equals 2n.
Lemma 7.3. For the system (7.13) with relative degree 1, the control input (7.17)
solves MRC problem with the reference model (7.14) and limt (y(t) ym (t)) = 0.
Proof. With the control input (7.17), the closed-loop dynamics are given by
e1 (s) = kp
Zm (s)
(s),
Rm (s)
where (s) denotes exponentially convergent signals due to non-zero initial values.
The reference model is stable, and then the track error e1 (t) converges to zero
exponentially.
2
Example 7.2. Design MRC for the system
y(s) =
s+1
u(s)
s2 2s + 1
s2
s+3
r(s).
+ 2s + 3
We follow the procedures shown early to obtain the MRC control. From the transfer
function of the system, we have
y(s)(s2 + 2s + 3) = (s + 1)u(s) + (4s + 2)y(s),
which leads to
y(s) =
s+3
2
s + 2s + 3
s+1
4s + 2
u(s) +
y(s)
s+3
s+3
10
s+3
2
u(s)
y(s) + 4y(s) .
= 2
u(s)
s + 2s + 3
s+3
s+3
Subtracting it by the reference model, we have
s+3
2
10
e1 (s) = 2
u(s)
u(s)
y(s) + 4y(s) r(s) ,
s + 2s + 3
s+3
s+3
which leads to the MRC control input
u(s) =
2
10
u(s) +
y(s) 4y(s) + r(s)
s+3
s+3
97
where
1
u(s),
s+3
1
2 (s) =
y(s).
s+3
1 (s) =
2T (s)
3
Zm (s)P(s)
98
Zm (s)
T (s)
T (s)
u(s) 1
u(s) 2
y(s) 3 y(s) .
Rm (s)
Zm (s)P(s)
Zm (s)P(s)
km
. The
kp
1T (s)
T (s)
u+ 2
y + 3 y + 4 r
Zm (s)P(s)
Zm (s)P(s)
:= T
(7.19)
(s)
u,
Zm (s)P(s)
2 =
(s)
y.
Zm (s)P(s)
Remark 7.4. The nal control input is in the same format as shown for the case
= 1. The lters for w1 and w2 are in the same order as in the case for = 1, as the
order of Zm (s)P(s) is still n 1.
Lemma 7.4. For the system (7.13) with relative degree > 1, the control input (7.19)
solves MRC problem with the reference model (7.14) and limt (y(t) ym (t)) = 0.
The proof is the same as the proof for Lemma 7.3.
Example 7.3. Design MRC for the system
y(s) =
s2
1
u
2s + 1
1
r.
s2 + 2s + 3
99
1
s2 + 2s + 3
(s + 5)(s2 2s + 1)y(s) + (14s 2)y(s)
s+1
1
(s + 5)u(s) + (14s 2)y(s)
= 2
s + 2s + 3
s+1
1
4
1
= 2
u(s) +
u(s) 16
y(s) + 14y(s) .
s + 2s + 3
s+1
s+1
1
2
s + 2s + 3
(s2 + 2s + 3)(s + 1)
y(s)
s+1
In this section, we present MRAC design for linear systems with relative
degree 1.
Consider an nth-order system with the transfer function
y(s) = kp
Zp (s)
u(s),
Rp (s)
(7.20)
100
where y(s) and u(s) denote the system output and input in frequency domain; kp is
the high frequency gain; and Zp and Rp are monic polynomials with orders of n 1
and n respectively. This system is assumed to be minimum phase, i.e., Zp (s) is a
Hurwitz polynomial, and the sign of the high-frequency gain, sgn(kp ), is known. The
coefcients of the polynomials and the value of kp are constants and unknown. The
reference model is chosen to have the relative degree 1 and strictly positive real, and
is described by
ym (s) = km
Zm (s)
r(s),
Rm (s)
(7.21)
where ym (s) is the reference output for y(s) to follow, r(s) is a reference input, and
Zm (s) and Rm (s) are monic polynomials and km > 0. Since the reference model is
strictly positive real, Zm and Rm are Hurwitz polynomials.
MRC shown in the previous section gives the control design in (7.17). Based on
the certainty equivalence principle, we design the adaptive control input as
u(s) = T ,
(7.22)
(s)
u,
Zm (s)
2 =
(s)
y.
Zm (s)
With the designed adaptive control input, it can be obtained, from the tracking
error dynamics shown in (7.16), that
Zm (s) T
( T )
Rm (s)
kp T
Zm (s)
= km
Rm (s)
km
e1 (s) = kp
(7.23)
where = .
To analyse the stability using a Lyapunov function, we put the error dynamics in
the state space form as
k
e = Am e + bm kmp T
(7.24)
e1 = cmT e
where (Am , bm , cm ) is a minimum state space realisation of km RZmm (s)
, i.e.,
(s)
cmT (sI Am )1 bm = km
Zm (s)
.
Rm (s)
101
(7.25)
(7.26)
1 T
1
kp
T 1
V = e Pm e +
,
2
2 km
where R2n is a positive denite matrix. Its derivative is given by
kp
kp T
1 T T
T
V = e (Am Pm + Pm Am )e + e Pm bm +
T 1
2
km
km
Using the results from (7.25) and (7.26), we have
kp
kp
1
V = eT Qm e + e1 T +
T 1
2
km
km
kp
1
= eT Qm e +
T 1 sgn(kp )e1 .
2
km
Hence, the adaptive law is designed as
= sgn(kp ) e1 ,
(7.27)
which results in
1
V = eT Qm e.
2
Furthermore it can be shown that
We can now conclude the boundedness of e and .
e L2 and e 1 L . Therefore, from Barbalats lemma we have limt e1 (t) = 0.
The boundedness of other system state variables can be established from the minimumphase property of the system.
We summarise the stability analysis for MRAC of linear systems with relative
degree 1 in the following theorem.
Theorem 7.5. For the rst-order system (7.20) and the reference model (7.21), the
adaptive control input (7.22) together with the adaptive law (7.27) ensures the boundedness of all the variables in the closed-loop system, and the convergence to zero of
the tracking error.
Remark 7.5. The stability result shown in Theorem 7.5 only guarantees the convergence of the tracking error to zero, not the convergence of the estimated parameters.
In the stability analysis, we use KalmanYakubovich lemma for the denition of
Lyapunov function and the stability proof. That is why we choose the reference model
to be strictly positive real. From the control design point of view, we do not need
to know the actual values of Pm and Qm , as long as they exist, which is guaranteed
102
by the selection of a strictly positive real model. Also it is clear from the stability
analysis, that the unknown parameters must be constant. Otherwise, we would not
have = .
Zp (s)
u(s),
Rp (s)
(7.28)
where y(s) and u(s) denote the system output and input in frequency domain, kp is
the high frequency gain, Zp and Rp are monic polynomials with orders of n and
n respectively, with > 1 being the relative degree of the system. This system is
assumed to be minimum phase, i.e., Zp (s) is Hurwitz polynomial, and the sign of the
high-frequency gain, sgn(kp ), is known. The coefcients of the polynomials and the
value of kp are constants and unknown. The reference model is chosen as
ym (s) = km
Zm (s)
r(s)
Rm (s)
(7.29)
where ym (s) is the reference output for y(s) to follow; r(s) is a reference input; and
Zm (s) and Rm (s) are monic polynomials with orders n and n respectively and
km > 0. The reference model (7.29) is required to satisfy an additional condition that
there exists a monic and Hurwitz polynomial P(s) of order n 1 such that
ym (s) = km
Zm (s)P(s)
r(s)
Rm (s)
(7.30)
is strictly positive real. This condition also implies that Zm and Rm are Hurwitz
polynomials.
MRC shown in the previous section gives the control design in (7.19). We design
the adaptive control input, again using the certainty equivalence principle, as
u = T ,
where is an estimate of the unknown vector R2n , and is given by
= [1T , 2T , y, r]T
(7.31)
103
with
1 =
(s)
u,
Zm (s)P(s)
2 =
(s)
y.
Zm (s)P(s)
The design of adaptive law is more involved, and we need to examine the dynamics
of the tacking error, which are given by
Zm
(u T )
Rm
Zm P(s)
= km
k(uf T ) ,
Rm
e1 = kp
(7.32)
where
k=
kp
,
km
uf =
1
u
P(s)
and =
1
.
P(s)
(7.33)
(7.34)
k = (uf T ).
(7.35)
With these adaptive laws, a stability result can be obtained for the boundedness of
parameter estimates and the convergence of the tracking error. For the completeness,
we state the theorem below without giving the proof.
Theorem 7.6. For the system (7.28) and the reference model (7.29), the adaptive
control input (7.31) together with the adaptive laws (7.34) and (7.35) ensures the
boundedness of all the variables in the closed-loop system, and the convergence to
zero of the tracking error.
104
(7.36)
(7.37)
where
= y
will render the convergence of the estimate by taking
V =
1 2
(7.38)
The boundedness of can then be concluded, no matter what the signal is.
Now, if the signal is corrupted by some unknown bounded disturbance d(t),
i.e.,
y = + d(t).
105
(7.39)
In this example, we have observed that adaptive law designed for the disturbancefree system fails to remain bounded even though the disturbance is bounded and
converges to zero as t tends to innity.
Remark 7.6. If is a constant, then from (7.38) we can show that the estimate
exponentially converges to the true value. In this case, there is only one unknown
parameter. If is a vector, the requirement for the convergence is much stronger. In
the above example, is bounded, but not in a persistent way. It does demonstrate that
even a bounded disturbance can cause the estimated parameter divergent.
Robust adaptive control issue is often addressed by modifying parameter adaptive
laws to ensure the boundedness of estimated parameters. It is clear from the example
shown above that bounded disturbance can cause estimated parameters unbounded.
Various robust adaptive laws have been introduced to keep estimated parameters
bounded in the presence of bounded disturbances. We will show two strategies using
the following simple model:
y = + d(t)
(7.40)
and
with d as a bounded disturbance. In the following we keep using = y
2
V = 2 . Once the basic ideas are introduced, it is not difcult to extend the robust
adaptive laws to adaptive control of dynamic systems.
Dead-zone modication is a modication to the parameter adaptive law to stop
parameter adaptation when the error is very close to zero. The adaptive law is
modied as
106
0
|| > g
|| g
(7.41)
where g is a constant satisfying g > |d(t)| for all t. For > g, we have
V =
= ( )
= (y d(t) )
= ( d(t))
< 0.
Therefore, we have
< 0, || > g
V
= 0, || g
and we can conclude that V is bounded. Intuitively, when the error is small, the
bounded disturbance can be more dominant, and therefore, the correct adaptation
direction is corrupted by the disturbance. In such a case, a simple strategy would be
just to stop parameter adaptation. The parameter adaptation stops in the range || g,
and for this reason, this modication takes the name dead-zone modication. The
size of the dead zone depends on the size of the bounded disturbances. One problem
with the dead-zone modication is that the adaptive law is discontinuous, and this
may not be desirable in some applications.
-Modication is another strategy to ensure the boundedness of estimated parameters. The adaptive law is modied by adding an additional term to the normal
adaptive law as
=
(7.42)
(7.43)
107
and therefore we can conclude that V L , which implies the boundedness of the
estimated parameter. A bound can be obtained for the bounded parameter as
2
1 d02
+
.
(7.44)
V ()
2
2
Note that this modication does not need a bound for the bounded disturbances, and
also it provides a continuous adaptive law. For these reasons, -modication is one
of the most widely used modications for parameter adaptation.
Remark 7.7. We re-arrange the adaptive law (7.42) as
+ = .
Since ( ) is a positive constant, the adaptive law can be viewed as a stable rst-order
dynamic system with () as the input and as the output. With a bounded input,
obviously remains bounded.
The robust adaptive laws introduced here can be applied to various adaptive
control schemes. We demonstrate the application of a robust adaptive law to MRAC
with = 1. We start directly from the error model (7.24) with an additional bounded
disturbance
e = Am e + bm (k T + d(t))
e1 = cmT e,
(7.45)
where k = kp /km and d(t) are a bounded disturbance with the bound d0 , which represents the non-parametric uncertainty in the system. As discussed earlier, we need a
robust adaptive law to deal with the bounded disturbances. If we take -modication,
then the robust adaptive law is
= sgn(kp ) e1 .
(7.46)
We will show that this adaptive law will ensure the boundedness of the variables.
Let
V =
1 T
1
e Pe + |k| T 1 .
2
2
108
Note that
1
d02
min (Q)
e
2 +
,
4
min (Q)
1
1
| T |
2 +
2 .
2
2
Hence, we have
|e1 d|
|k|
d02
|k|
1
2 +
+
2
V min (Q)
e
2
4
2
min (Q)
2
V +
|k|
d02
+
2 ,
min (Q)
2
estimate .
From the above analysis, it is clear that the adaptive law with -modication
ensures the boundedness of all the variables in the closed-loop adaptive control system.
It is worth noting that the output tracking error e1 will not asymptotically converge
to zero, even though the bounded disturbance d(t) becomes zero. That is the price to
pay for the robust adaptive scheme.
Chapter 8
(8.1)
where x Rn is the state; y Rm is the system output with m < n; and A Rnn
and C Rmn are constant matrices. From linear system theory, we know that this
system, or the pair (A, C), is observable if the matrix
C
CA
Po = .
..
CAn1
has rank n. The observability condition is equivalent to that the matrix
I A
C
has rank n for any value of C.
When the system is observable, an observer can be designed as
x = Ax + L(y C x ),
(8.2)
110
where x Rn is the estimate of the state x, and L Rnm is the observer gain such
that (A LC) is Hurwitz.
For the observer (8.2), it is easy to see the estimate x converges to x asymptotically. Let x = x x , and we can obtain
x = (A LC)x.
Remark 8.1. In the observer design, we do not consider control input terms in the
system in (8.1), as they do not affect the observer design for linear systems. In fact, if
Bu term is added to the right-hand side of the system (8.1), for the observer design,
we can simply add it to the right-hand side of the observer in (8.2), and the observer
error will still converge to zero exponentially.
Remark 8.2. The observability condition for the observer design of (8.1) can be
relaxed to the detectability of the system, or the condition that (A, C) is detectable,
for the existence of an observer gain L such that (A LC) is Hurwitz. Detectability is
weaker than observability, and it basically requires the unstable modes of the system
observable. The pair (A, C) is detectable if the matrix
I A
C
has rank n for any in the closed right half of the complex plan. Some other design
methods shown in this chapter also need only the condition of detectability, although,
for simplicity, we state the requirement for the observability.
There is another approach to full-state observer design for linear systems (8.1).
Consider a dynamic system
z = Fz + Gy,
(8.3)
111
Lemma 8.1. The dynamic system (8.3) is an observer for the system (8.1) if and only
if F is Hurwitz and there exists an invertible matrix T such that
TA FT = GC.
(8.4)
Proof. The sufciency has been shown in the above analysis. For necessity, we only
need to observe that if any of the conditions is not satised, we cannot guarantee the convergence of e to zero for a general linear system (8.1). Indeed, if (8.4)
is not satised, then e will be a state variable with a non-zero input, and we can set
up a case such that e does not converge to zero. So does for the condition that F is
Hurwitz.
2
How to nd matrices F and G such that the condition (8.4) is satised? We list
the result in the following lemma without the proof.
Lemma 8.2. Suppose that F and A have exclusively different eigenvalues. The necessary condition for the existence of a non-singular solution T to the matrix equation
(8.4) is that the pair (A, C) is observable and the pair (F, G) is controllable. This
condition is also sufcient when the system (8.1) is single output, i.e., m = 1.
This lemma suggests that we can choose a controllable pair (F, G) and make sure
that the eigenvalues of F are different from those of A. An observer can be designed if
there is a solution of T from (8.4). For single output system, the solution is guaranteed.
(8.5)
(8.6)
where x Rn is the estimate of the state x and L Rnm is the observer gain such
that (A LC) is Hurwitz. The only difference between this observer and the one in
(8.2) is the nonlinear term (y, u). It can be seen that the observer error still satises
x = (A LC)x.
112
Note that even the system (8.5) is nonlinear, the observer error dynamics are
linear. The system in the format of (8.5) is referred to as the system with linear
observer errors. More specically, if we drop the control input u in the function ,
the system is referred to as the output injection form for observer design.
Let us summarise the result in the following proposition.
Proposition 8.3. For the nonlinear system (8.5), a full-state observer can be designed
as in (8.6) if (A, C) is observable. Furthermore, the observer error dynamics are linear
and exponentially stable.
For a nonlinear system in a more general form, there may exist a state transformation to put the system in the format of (8.5), and then the proposed observer can
be applied.
We will introduce the conditions for the existence of a nonlinear state transformation to put the system in the format shown in (8.6). Here, we only consider single
output case, and for the simplicity, we do not consider the system with a control input.
The system under consideration is described by
x = f (x)
y = h(x),
(8.7)
(8.8)
(8.9)
(8.10)
This system is in the output injection form for nonlinear observer design.
Remark 8.3. Assume that the system (8.9) is different from (8.10) with (A, C) as
a general observable pair, instead of having the special format implied by (8.10).
113
In this case, we use a different variable z to denote the state for (8.10), and it can be
written as
z + (y)
z = A
(8.11)
y = C z .
It is clear that there exists a linear transformation
z = Tz,
which transforms the system (8.9) to the system (8.11), because (A, C) is observable.
The transformation from (8.7) to (8.11) is given by
z = T (x) := (x).
Therefore, if there exists a nonlinear transformation from (8.7) to (8.9), there must
exist a nonlinear transformation from (8.7) to (8.11), and vise versa. That is why
we can consider the transformed system in the format of (8.10) without loss of
generality.
If the state transformation transforms the system (8.7) to (8.9), we must have
(x)
f (x)
= Az + (Cz)
x
x=(z)
h((z)) = Cz,
h(z)
= Cz.
From the structure shown in (8.10), we can obtain
h(z)
= z1 ,
Lf h(z)
= z2 + 1 (z1 ),
1
(z2 + 1 (z1 ))
z1
:= z3 + 2 (z1 , z2 ),
Lf2 h(z)
= z3 +
...
Lfn1
h(z) = zn +
n2
n2
k=1
zk
(zk+1 + k (z1 ))
:= zn + n1 (z1 , . . . , zn1 ).
(8.12)
114
h
z
Lf h(z)
..
n1
Lf h(z)
=.
.
.
0
1
..
.
0
0
..
.
...
...
..
.
...
z
This implies that
dLf h,
. . . , dLn1 h
d h,
f
are linearly independent. This property is invariant under state transformation, and
therefore, we need the condition under the coordinate x, that is
dh, dLf h, . . . , dLfn1 h
are linearly independent. Indeed, we have
h(x)
x
Lf h(x)
x
..
.
Ln1 h(x)
f
x
h(z)
z
Lf h(z)
=
z
..
.
n1
Lf h(z)
z
for k = 0, 1, . . . , n 1
(x)
.
x
z=(x)
The linear independence of dh, dLf h, . . . , dLfn1 h is a consequence of the observability of (A, C) in (8.9). In some literature, this linear independence condition is
dened as the observability condition for nonlinear system (8.7). Unlike linear systems, this condition is not enough to design a nonlinear observer. We state necessary
and sufcient conditions for the transformation to the output injection form in the
following theorem.
Theorem 8.4. The nonlinear system (8.7) can be transformed to the output injection
form in (8.10) if and only if
115
(8.13)
dh
dLf h
..
.
dLfn1 h
0
..
= . .
0
1
(8.14)
Proof. Sufciency. From the rst condition and (8.14), we can show, in a similar way
as for (6.12), that
dh(x)
dLf h(x)
n1
..
.
dLf1 h
=.
..
0
1
..
.
...
...
..
.
...
0
0
.. .
.
(8.15)
n1
r(x) . . . adf r(x) r(x) has full rank. This implies that there exists
Therefore, adf
an inverse mapping for . Let us denote it as = 1 , and hence we have
(x)
n1
adf r(x) . . . adf r(x) r(x) = I .
(8.16)
x
Let us dene the state transformation as z = (x) and denote the functions after this
transformation as
f (z) = (x) f (x)
,
x
x=(z)
h(z)
= h((z)).
We need to show that the functions f and h are in the format of the output injection
form as in (8.10). From (8.16), we have
(x) nk
adf r(x) = ek ,
x
for k = 1, . . . , n,
where ek denotes the kth column of the identity matrix. Hence, we have, for
k = 1, . . . , n 1,
116
(x) nk
adf r(x)
x
=
x=(z)
(x)
n(k+1)
[ f (x), adf
r(x)]
x
x=(z)
(x)
(x) n(k+1)
=
r(x)
f (x),
adf
x
x
= f (z), ek+1
=
f (z)
.
zk+1
for k = 1, . . . , n 1,
f (z) .
= ..
1
0
..
.
0
1
..
.
0
0
0
0
0
0
.. .
.
1
0
...
...
..
.
...
...
for k = 1, . . . , n.
h(z)
(z)
h(x)
=
zk
x x=(z) zk
h(x)
nk
=
[adf
(x) r(x)]x=(z)
x x=(z)
= [Lad nk
f (x) r(x)
h(x)]x=(z) .
h(x) = 1,
Lad nk
h(x) = 0,
f (x) r(x)
f (x) r(x)
for k = 2, . . . , n.
Therefore, we have
h(z)
= [1, 0, . . . , 0].
z
This concludes the proof for sufciency.
x=(z)
117
Necessity. The discussion prior to this theorem shows that the rst condition is
necessary. Assume that there exists a state transformation z = (x) to put the system
in the output injection form, and once again, we denote
f (z) =
(x)
f (x)
x
x=(z)
h(z)
= h((z)),
where = 1 . We need to show that when the functions f and h are in the format
of the output injection form, the second condition must hold. Let
g(x) =
(z)
zn
.
z=(x)
(z)
f (x) =
f (z)
z
.
z=(x)
(z)
(z)
,
f (z)
z
zn z=(x)
z=(x)
(z)
f (z), en z=(x)
z z=(x)
f (z)
(z)
z z=(x) zn
(z)
en1
z z=(x)
(z)
.
zn1 z=(x)
(z)
znk
for k = n 2, . . . , 1.
z=(x)
(8.17)
118
The remaining part of the proof is to show that g(x) coincides with r(x) in (8.14).
From (8.17), we have
h(x) (z)
h(z)
=
z
x
z
h(x)
n1
=
adf g, . . . , adf g, g x=(z)
x
= [Lad n1 g h(x), . . . , Ladf g h(x), Lg h(x)]x=(z) .
f
Lad nk g h(x) = 0,
f
for k = 2, . . . , n,
dh
dLf h
..
.
dLfn1 h
for k = 2, . . . , n,
0
..
= . .
0
1
Remark 8.4. It would be interesting to revisit the transformation for linear singleoutput systems to the observer canonical form, to reveal the similarities between the
linear case and the conditions stated in Theorem 8.4. For a single output system
x = Ax
y = Cx,
(8.18)
where x Rn is the state; y R is the system output; and A Rnn and C R1n
are constant matrices. When the system is observable, we have Po full rank.
Solving r from Po r = e1 , i.e.,
0
C
..
CA
.. r = . ,
0
.
n1
1
CA
and the state transformation matrix T is then given by
T 1 = [An1 r, . . . , Ar, r].
(8.19)
119
We can show that the transformation z = Tx which transforms the system to the
observer canonical form
z = TAT 1 z := Az
1
y = CT z := Cz
where
a1
a2
..
.
A =
an1
an
1
0
..
.
0
1
..
.
0
0
0
0
...
...
..
.
...
...
0
0
.. ,
.
1
0
C = [1 0 . . . 0 0],
with constants ai , i = 1, . . . , n, being the coefcients of the characteristic polynomial
|sI A| = sn + a1 sn1 + + an1 s + an .
Indeed, from (8.19), we have
AT 1 = [An r, . . . , A2 r, Ar]
(8.20)
a1 1 0 . . . 0
a2 0 1 . . . 0
.. .. . .
.. = T 1 A,
1
n1
AT = [A r, . . . , Ar, r] ...
. .
. .
an1 0 0 . . . 1
an 0 0 . . . 0
which gives
TAT 1 = A.
Again from (8.19), we have
(8.21)
120
(8.22)
(8.23)
(8.24)
for a chosen controllable pair (F, G) with F Rnn Hurwitz, and G Rn . Comparing
(8.24) with (8.9), the matrices F and G in (8.24) are chosen for the observer design,
while in (8.9), A and C are any observable pair which depends on the original system.
Therefore, the transformation in (8.24) is more specic. There is an extra benet
gained from this restriction in the observer design as shown later.
121
From (8.22) and (8.24), the nonlinear transformation must satisfy the following
partial differential equation
(x)
f (x) = F(x) + Gh(x).
x
(8.25)
qi i (A).
i=1
The following theorem states a result concerning with the existence of a nonlinear
state transformation around the origin for (8.22).
Theorem 8.6. For the nonlinear system (8.22), there exists a state transformation,
i.e., a locally invertible solution to the partial differential equation (8.25) if
Proof. From the non-resonant condition and the exclusion of the origin of the convex
hall of the eigenvalues of f
(0), we can establish the existence of a solution to the
x
partial differential equation (8.25) by invoking Lyapunov Auxiliary Theorem. That
the function is invertible around the origin is guaranteed by the observability of the
linearised model
and the the controllability of (F, G). Indeed, with the observability of
f
h
(0), x (0) and the controllability of (F, G), we can apply Lemma 8.2 to establish
x
that
(0)
x
is invertible.
With the existence of the nonlinear transformation to put the system in the form
of (8.24), an observer can be designed as
z = F z + Gy
x = 1 (z ),
(8.26)
where is the inverse transformation of . It is easy to see that the observer error
dynamics are linear as
z = F z .
122
Note that once the transformation is obtained, the observer is directly given without
designing observer gain, unlike the observer design based on the output injection
form.
The observer can also be implemented directly in the original state as
x = f (x) +
1
(x)
G(y h(x)),
x
(8.27)
which is in the same structure as the standard Luenberger observer for linear systems
1
by viewing
(x) G as the observer gain. In the following theorem, we show that
x
this observer also provides an asymptotic estimate of the system state.
Theorem 8.7. For the nonlinear system (8.22), if the state transformation in (8.25)
exists, the observer (8.27) provides an asymptotic estimate. Furthermore, the
dynamics of the transformed observer error ((x) (x)) are linear.
Proof. Let e = (x) (x). Direct evaluation gives
1
(x)
(x)
e =
f (x)
f (x) +
G(y h(x))
(x)
x
x
x
(x)
(x)
f (x)
f (x) G(y h(x))
x
x
= F(x) + Gy (F(x) + Gh(x)) G(y h(x))
=
= F((x) (x))
= Fe.
Therefore, the dynamics of the transformed observer error are linear, and the transformed observer error converges to zero exponentially, which implies the asymptotic
convergence of x to x.
2
Remark 8.5. For linear systems, we have briey introduced two ways to design
observers. For the observer shown in (8.2), we have introduced the observer shown
in (8.6) to deal with nonlinear systems. The nonlinear observer in (8.27) can be
viewed as a nonlinear version of (8.3). For both cases, the observer error dynamics
are linear.
123
(8.28)
with > 0.
Once again we consider linear systems perturbed by nonlinear terms as
x = Ax + (x, u)
y = Cx,
(8.29)
where x Rn is the state; y Rm is the system output with m < n; u Rs is the control
input, or other known variables, A Rnn and C Rmn are constant matrices with
(A, C) observable; and : Rn Rs Rn is a continuous function with Lipschitz
constant with respect to the state variable x. Comparing with the system (8.5), the
only difference is the nonlinear term (x, u). Here, it is a function of state variable,
not only the output, as in (8.5), and therefore the observer design by output injection
does not work.
We can still design an observer based on the linear part of the system, and replace
the unknown state in the nonlinear function by its estimate, that is,
x = Ax + L(y C x ) + (x, u),
(8.30)
where x Rn is the estimate of the state x and L Rnm is the observer gain. However,
the condition that (A LC) is Hurwitz is not enough to guarantee the convergence
of the observer error to zero. Indeed, a stronger condition is needed, as shown in the
following theorem.
Theorem 8.8. The observer (8.30) provides an exponentially convergent state estimate if for the observer gain L, there exists a positive denite matrix P Rnn such
that
(A LC)T P + P(A LC) + 2 PP + I + I = 0,
where is the Lipschitz constant of and is any positive real constant.
Proof. Let x = x x . From (8.29) and (8.31), we have
x = (A LC)x + (x, u) (x, u).
Let
V = x T P x .
Its derivative along the observer error dynamics is obtained as
V = x T ((A LC)T P + P(A LC))x + 2xT P((x, u) (x, u)).
(8.31)
124
1
= 2 x T PP x + 2 (x, u) (x, u)2
2 x T PP x + x2
= x T ( 2 PP + I )x.
Applying the condition shown in (8.31), we have the derivative of V satisfying
V x T ((A LC)T P + P(A LC) + 2 PP + I )x
= x T x .
This implies that the observer error converges to zero exponentially.
(8.32)
Using the same Lyapunov function candidate as in the proof of Theorem 8.8, we can
show that
V < 0
which implies that the observer error converges to zero asymptotically. We summarise
this result below.
Corollary 8.9. The observer (8.30) provides an asymptotically convergent state estimate if for the observer gain L, there exists a positive denite matrix P Rnn that
satises the inequality (8.32).
Remark 8.6. By using the inequality (8.32) instead of the equality (8.31), we only
establish the asymptotic convergence to zero of the observer error, not the exponential
convergence that is established inTheorem 8.8 using (8.31). Furthermore, establishing
the asymptotic convergence of the observer error from V < 0 requires the stability
theorems based on invariant sets, which are not covered in this book.
The condition shown in (8.32) can be relaxed if one-side Lipschitz constant is
used instead of the Lipschitz constant.
Denition 8.3. A function : Rn Rs Rn is one-sided Lipschitz with respect to
P and one-sided Lipschitz constant if for any vectors x, x Rn and u Rs
(x x )T P((x, u) (x, u)) x x 2
where P R
nn
(8.33)
125
Note that the one-sided Lipschitz constant can be negative. It is easy to see from
the denition of the one-sided Lipschitz condition that the term (x x )T P((x, u)
(x, u)) is exactly the cross-term in the proof Theorem 8.8 which causes the term
2 PP + I in (8.32). Hence, with the Lipschitz constant with respect to P, the
condition shown in (8.32) can be replaced by
(A LC)T P + P(A LC) + 2I < 0.
(8.34)
This condition can be further manipulated to obtain the result shown in the following
theorem.
Theorem 8.10. The observer (8.30) provides an asymptotically convergent state
estimate if the following conditions hold:
L = P 1 C T ,
(8.35)
A P + PA + 2I 2 C C < 0,
T
(8.36)
where P Rnn is a positive real matrix, is a positive real constant and is the
one-sided Lipschitz constant of with respect to x and P.
Proof. From (8.35), we have
T
LT
LT
1
1
CP
CP
= 0,
which gives
P 1 C T LT + LCP 1 = P 1 C T CP 1 +
LLT
.
Using (8.35) and multiplying the above equation by P on both sides, we obtain the
identity
C T LT P + PLC = 2 C T C.
From this identity and (8.36), we can easily obtain the inequality (8.34). Similar to
the proof of Theorem 8.8, we let
V = x T P x ,
and obtain, using the one-sided Lipschitz condition of ,
V = x T ((A LC)T P + P(A LC))x + 2xT P((x, u) (x, u))
x T ((A LC)T P + P(A LC))x. + 2 x T x .
Applying the inequality (8.34) to the above expression, we have
V < 0,
which implies that the observer error asymptotically converges to zero.
126
To end this section, we consider a class of systems with nonlinear Lipschitz output
function
x = Ax
(8.37)
y = h(x),
where x Rn is the state vector, y Rm is the output, A Rnn is a constant matrix
and h : Rn Rm is a continuous function. We can write the nonlinear function h as
h = Hx + h1 (x) with Hx denoting a linear part of the output, and the nonlinear part
h1 with Lipschitz constant .
An observer can be designed as
x = Ax + L(y h(x)),
(8.38)
where the observer gain L Rnm is a constant matrix.
Theorem 8.11. The observer (8.38) provides an exponentially convergent state
estimate of (8.37) if the observer gain L can be chosen to satisfy the following
conditions:
1
L = 2 P 1 H T ,
(8.39)
PA + AT P
HTH
+ (1 + )I = 0,
2
(8.40)
+
x
P
P
x
L
2
= x T x .
Therefore, we can conclude that x converges to zero exponentially.
Remark 8.7. The nonlinearity in the output function with linear dynamics may
occur in some special cases such as modelling a periodic signal as the output of a
second-order linear system. This kind of formulation is useful for internal model
design to asymptotically reject some general periodic disturbances, as shown in
Chapter 10.
127
(8.41)
y = h(x),
h(x)
g(x)
is a diffeomorphism, where h(x) is the output function. The transformed states are
referred to as output-complement states.
Clearly an output-complement transformation denes a state transformation
together with the output function. If the output-complement states are known, then
the state variables can be fully determined. In fact, if we can design an observer for
the output-complement state, this observer is a reduced-order observer.
Denition 8.5. The dynamic model
z = p(z, y) + q(y, u)
(8.42)
is referred to as reduced-order observer form for the system (8.41) if the z = g(x) is
output-complement transformation, and the dynamic system
z = p(z, y)
is differentially stable.
128
(8.43)
Theorem 8.12. If the system (8.41) can be transformed to the reduced-order observer
form (8.42), the state estimate x provided by the reduced-order observer in (8.43)
asymptotically converges to the state variable of (8.41).
Proof. First, let us establish the boundedness of z. Since z = p(z, y) is differentially
stable, from the denition of the differential stability in Chapter 5, there exists a
Lyapunov function, V (z), such that the conditions (5.41) are satised. Take the same
function V (z) here as the Lyapunov function candidate with z = 0. From (5.41) and
(8.42), we have
V
V 3 (z) +
q(y, u)
z
V
3 (z) +
q(y, u).
(8.44)
z
Let us recallYoungs inequality in a simplied form that for any a R and b R,
and a pair of constants p > 1 and q > 1 with 1p + q1 = 1, we have
|ab|
p p
1
|a| + q |b|q
p
q
where c5 =
3
1 1
( c c ) c2
c3 2 1 2
1
1
V 3 (z) + c5 q(y, u)c3 .
(8.47)
2
2
Since q(y, u) is continuous, there exists a class K function g such that, for all y Rm
and u Rs , we have
q(y, u) g(y
+ u).
129
c3
+ u))c3
3 (z) 2c5 (g(y
2c5 q(y, u)c3 ,
which further implies that
1
V 3 (z).
4
(8.48)
130
Therefore, the system satises the conditions specied in (5.41). We design the
reduced-order observer as
z = (1 + z 2 )z + y
x 2 = z + y.
Simulation study has been carried out, and the simulation results are shown in
Figures 8.1 and 8.2.
0.7
x1
x2
0.6
x1 and x2
0.5
0.4
0.3
0.2
0.1
Time (s)
10
131
0.4
x2
Estimate of x2
0.35
x2 and estimate
0.3
0.25
0.2
0.15
0.1
0.05
0
10
Time (s)
There are lack of systematic design methods for nonlinear observers with global
convergence when there are general nonlinear terms of unmeasured state variables
in the systems. With the introduction of the reduced-order observer form, we like to
further explore the class of nonlinear systems which can be transformed to the format
shown in (8.42), and therefore a nonlinear observer can then be designed accordingly.
We consider a multi-output (MO) nonlinear system
x = Ax + (y, u) + E(x, u)
y = Cx,
(8.49)
132
Without loss of generality, we can assume that C has full row rank. There exists
a nonsingular state transformation M such that
CM 1 = [Im , 0m(nm) ].
If span{E} is a complement subspace of ker{C} in Rn , we have (CM 1 )(ME) = CE
invertible. If we partition the matrix ME as
E1
ME :=
E2
with E1 Rmm , then we have CE = E1 . Furthermore, if we partition Mx as
1
Mx :=
2
with 1 Rm , we have
z = g(x) = 2 E2 E11 1 .
(8.50)
M :=
1
,
2
(8.51)
where
q(y, u) = (A2,2 E2 E11 A1,2 )E2 E11 y + (A2,1 E2 E11 A1,1 )y
+ 2 (y, u) E2 E11 1 (y, u).
Note that the nonlinear function (x, u) does not appear in the dynamics of z
in (8.51) due to the particular choice of z in (8.50).
Remark 8.8. After the state transformation, (8.51) is in the same format as (8.42).
Therefore, we can design a reduced-order observer if z = (A2,2 E2 E11 A1,2 )z is differentially stable. Notice that it is a linear system. Hence, it is differentially stable if it
133
is asymptotically stable, which means the eigenvalues of (A2,2 E2 E11 A1,2 ) are with
negative real parts.
Following the format shown in (8.43), a reduced-order observer can then be
designed as
z = (A2,2 E2 E11 A1,2 )z + q(y, u)
and the estimate of x is given by
y
.
x = M 1
z + E2 E11 y
(8.52)
(8.53)
sIm A1,1
A1,2
E1
sInm A2,2 E2 .
= A2,1
(8.54)
Im
0
0
Let us multiply the above matrix in the left by the following matrix to perform a row
operation:
0
0
Im
E2 E11 Inm 0 ,
0
0
Im
134
A1,2
E1
sIm A1,1
sInm (A2,2 E2 E11 A1,2 ) 0
Im
0
0
with
= E2 E11 (sIm A1,1 ) A2,1 .
Since E1 is invertible, any values of s which make the matrix
sI A E
C
0
rank decient must be the eigenvalues of (A2,2 E2 E11 A1,2 ). From the second condition that all the invariant zeros of (A, E, C) are with negative real part, we can conclude
that (A2,2 E2 E11 A1,2 ) is Hurwitz.
2
Example 8.2. Consider a third-order system
x 1 = x1 + x2 y1 + u + x2 x3 x1 x3
x 2 = x1 + x2 + x3 2y1 + u + y1 y2 + x2 x3
x 3 = y12 + x1 x3 + x2 x3
y1 = x 1
y2 = x1 + x2 .
We can identify
y1 + u
= 2y1 + y1 y2 + u ,
y12
and we have
A = 1
0
1
1
0
0
1,
0
=
E= 1
1
x2 x3 x1 x3
,
x1 x3
0
1,
2
C=
(8.55)
1
1
0
1
0
.
0
(8.56)
It can be easily checked that (A, C) is observable and the invariant zero of
(A, E, C) is at 2. Therefore, the conditions in Theorem 8.13 are satised. It is
also easy to see that x2 = y2 + y1 . Therefore, the only unmeasured state variable is
x3 . Indeed, following the procedures introduced earlier, we have
1
0 0
1 0
, E2 1 2 ,
M = 1 1 0 , E1 =
0 1
0
0 1
x1
, 2 = x3 ,
1 =
x2 x 1
A1,1 =
1 =
0
0
1
,
0
A1,2
y1 + u
,
y1 + y1 y2
0
=
,
1
A2,1 0
0 ,
A2,2 = 0,
2 = y12 ,
1
x1
x2
x3
6
0
10
12
Time (s)
14
16
18
20
135
136
Estimate of x3
0.2
x3 and estimate
0
0.2
0.4
0.6
0.8
1
1.2
0
10
12
14
16
18
20
Time (s)
(8.57)
137
(8.58)
(8.59)
(8.60)
where = .
Consider a Lyapunov function candidate
V = x T P x + T 1 ,
where Rnn is a positive denite matrix. From (8.58) and (8.60), we have
V = x T ((A LC)T P + P(A LC))x + 2xT Pb(y, u)T + 2 T 1
= xT Qx + 2xT C T (y, u)T 2 T 1
= xT Qx 2( (y C x )(y, u))T 1
If we set the adaptive law as
= (y C x )(y, u),
we obtain
V = xT Qx.
Then similar to stability analysis of adaptive control systems, we can conclude that
limt x (t) = 0, and is bounded. The above analysis leads to the following theorem.
Theorem 8.14. For the observable single-output system (8.57), if the linear system
characterised by (A, b, C) is minimum phase and has relative degree 1, there exists an
138
observer gain L Rn that satises the conditions in (8.58) and an adaptive observer
designed as
x = Ax + 0 (y, u) + b T (y, u) + L(y C x )
= (y C x )(y, u),
(8.61)
0 1 0 ... 0
b1
0 0 1 ... 0
b2
..
..
..
.. , b = .. ,
..
A=.
.
.
.
.
.
0 0 0 ... 1
bn1
bn
0 0 0 ... 0
C = [1
...
0],
where b1 = 0, since the relative degree is 1. Indeed, for an observer system {A, b, C}
with relative degree 1, there exists a state transformation, as shown in Remark 8.4, to
transform the system to the observable canonical form. Furthermore, if we move the
rst column of A in the canonical form, and combine it with 0 (y, u), we have A as in
the format shown above. Since the system is minimum phase, b is Hurwitz, i.e.,
B(s) := b1 sn1 + b2 sn2 + + bn1 s + bn = 0
has all the solutions in the left half of the complex plane. In this case, we can design
the observer gain to cancel all the zeros of the system. If we denote L = [l1 , l2 , . . . , ln ],
we can choose L to satisfy
sn + l1 sn1 + l2 sn2 + + ln1 s + ln = B(s)(s + ),
where is a positive real constant. The above polynomial equation implies
L = (I + A)b.
This observer gain ensures
1
,
s+
which is a strict positive real transfer function.
C (sI (A LC))1 b =
Now we will consider adaptive observer design for a class of Lipschitz nonlinear
systems. Consider
x = Ax + 0 (x, u) + b T (x, u)
y = Cx,
(8.62)
139
(8.63)
(8.64)
+ b T (x, u),
where = . Consider a Lyapunov function candidate
V = x T P x + T 1 .
Its derivative along the dynamics in (8.65) is obtained as
V = x T ((A LC)T P + P(A LC))x + 2xT P(0 (x, u) 0 (x, u))
(8.65)
140
Chapter 9
Backstepping design
For a nonlinear system, the stability around an equilibrium point can be established
if one can nd a Lyapunov function. Nonlinear control design can be carried out by
exploring the possibility of making a Lyapunov function candidate as a Lyapunov
function through control design. In Chapter 7, parameter adaptive laws are designed
in this way by setting the parameter adaptive laws to make the derivative of a Lyapunov
function candidate negative semi-denite. Backstepping is a nonlinear control design
method based on Lyapunov functions. It enables a designed control to be extended to
an augmented system, provided that the system is augmented in some specic way.
One scheme is so-called adding an integrator in the sense that if a control input is
designed for a nonlinear system, then one can design a control input for the augmented
system of which an integrator is added between the original system input and the
input to be designed. This design strategy can be applied iteratively. There are a few
systematic control design methods for nonlinear systems, and backstepping is one
of them. In this chapter, we start with the fundamental form of adding an integrator,
and then introduce the method for iterative backstepping with state feedback. We also
introduce backstepping using output feedback, and adaptive backstepping for certain
nonlinear systems with unknown parameters.
(9.1)
where x Rn and R are the state variables; u R is the control input; and f :
Rn Rn with f (0) = 0 and g : Rn Rn are continuous functions. Viewing the
rst equation of (9.1) as the original system with x as the state and as the input,
the integration of u gives , which means that an integrator is added to the original
system.
We consider the control design problem under the assumption that a known
control input exists for the original system. Furthermore, we assume that the Lyapunov
function is also known, associated with the known control for x-subsystem. Suppose
that control input for the x-subsystem is (x) with differentiable and (0) = 0, and
the associated Lyapunov function is V (x). We assume that
142
V
(9.2)
(f (x) + g(x)(x)) W (x),
x
where W (x) is positive denite.
Condition (9.2) implies that the system x = f (x) + g(x)(x) is asymptotically
stable. Consider
x = f (x) + g(x)(x) + g(x)( (x)).
Intuitively, if we can design a control input u = u(x, ) to force to converge to (x),
we have a good chance to ensure the stability of the entire system. Let us dene
z = (x).
It is easy to obtain the dynamics under the coordinates (x, z) as
x = f (x) + g(x)(x) + g(x)z
z = u = u
(f (x) + g(x) ).
x
Consider a Lyapunov function candidate
1
Vc (x, z) = V (x) + z 2 .
2
Its derivative is given by
(9.3)
V
V
V c =
g(x)z
(f (x) + g(x)(x)) +
x
x
+z u
(f (x) + g(x) )
x
V
= W (x) + z u +
g(x)
(f (x) + g(x) ) .
x
x
Let
V
g(x) +
(f (x) + g(x) )
x
x
with c > 0 which results in
u = cz
V c = W (x) cz 2 .
(9.4)
(9.5)
It is clear that W (x) cz is negative denite with respect to variables (x, z). Hence,
we can conclude that Vc (x, z) is a Lyapunov function, and that (0, 0) in the coordinates
(x, z) is a globally asymptotic equilibrium. From (0) = 0, we can conclude that (0, 0)
in the coordinates (x, ) is also a globally asymptotic equilibrium, which means that
the system (9.1) is globally asymptotically stable under the control input (9.4). We
summarise the above result in the following lemma.
2
Lemma 9.1. For a system described in (9.1), if there exist differentiable function (x)
and a positive-denite function V (x) such that (9.2) holds, the control design given
in (9.4) ensures the global asymptotic stability of the closed-loop system.
Backstepping design
143
Remark 9.1. Considering the structure of (9.1), if is viewed as the control input for
the x-subsystem, = (x) is the desired control, ignoring the -system. This is why
can be referred to as a virtual control input for the x-subsystem. The control input
u for the overall system is designed with the consideration of the dynamics back to
the control design for the x-subsystem, and it may suggest the name of this particular
design method as backstepping.
Example 9.1. Consider
x 1 = x12 + x2
x 2 = u.
We design a control input using backstepping. Comparing with (9.1), we can identify
x x1 ,
x2 ,
f (x) x12 ,
g(x) 1.
1 2
x
2 1
with
V (x1 ) = c1 x12 .
Therefore, the condition specied in Lemma 9.1 is satised with (x1 ) = c1 x1
x12 , V (x1 ) = 12 x12 and W (x1 ) = c1 x12 . The control input u can then be obtained
from (9.4) by substituting proper functions in the equation. Alternatively, we can
obtain the control input by directly following the backstepping method. Indeed, let
z = x2 (x1 ). The dynamics of the system in coordinate (x1 , z) are obtained as
x 1 = c1 x1 + z
(x1 ) 2
z = u
(x1 + x2 ),
x1
where
(x1 )
= c1 2x1 .
x1
Let
1
1
Vc (x1 , z) = x12 + z 2 .
2
2
144
Vc (x1 , z) = c1 x1 + x1 z + z u
(x1 + x2 ) .
x1
Designing the control u as
u = x1 c2 z +
(x1 ) 2
(x1 + x2 )
x1
results in
V c (x1 , z) = c1 x12 c2 z 2 .
Hence, the system is asymptotically stable with (x1 , z). As (0) = 0, we conclude limt x2 (t) = limt (z(t) + (x1 (t))) = 0, which implies that the system
is asymptotically stable in the equilibrium (0, 0) in (x1 , x2 ). Note that the closed-loop
system in (x1 , z) is written as
x 1
1
x1
c1
=
.
1 c2
z
z
In Example 9.1, backstepping design has been used to design a control input for a
nonlinear system with unmatched nonlinearities. When a nonlinear function appears
in the same line as the control input, it is referred as a matched nonlinear function,
and it can be cancelled by adding the same term in u with an opposite sign. From the
system considered in Example 9.1, the nonlinear function x12 does not appear in the
same line as the control input u, and therefore it is unmatched. However, it is in the
same line as x2 , which is viewed as a virtual control. As a consequence, (x1 ), which
is often referred to as a stabilising function, can be designed to cancel the nonlinearity
which matches with the virtual control, and backstepping method enables the control
in the next line to be designed to stabilise the entire system. This process can be
repeated by identifying a virtual control, designing a stabilising function and using
backstepping to design control input for more complicated nonlinear systems.
(9.6)
Backstepping design
145
where xi R for i = 1, . . . , n are state variables; i : R R R for
i = 1, . . . , n are differentiable functions up to the order n i with i (0, . . . , 0) = 0;
and u R is the control input.
When xi+1 in the ith equation of (9.6) is identied as the virtual control, the nonlinear function i is then matched with respect to the virtual control, and backstepping
method can be applied to move down the control design to (i + 1)th equation with
xi+2 as the next virtual control. This process starts from i = 1 and can be repeated
until i = n 1 when u is reached. We will present this iterative backstepping design
for the system (9.6) in n steps. In each step, we could show the Lyapunov function
and other details for which Lemma 9.1 can be applied to. Although, for the simplicity
of the control design, we leave the stability analysis to the end, stability is considered
in designing the stabilising function at each step. The control design will be shown
in n steps.
Let
z1 = x1 ,
zi = xi i1 (x1 , . . . , xi1 ),
for i = 2, . . . , n,
(9.7)
(9.8)
= x3 + 2 (x1 , x2 )
Design 2 as
2 = z1 c2 z2 2 (x1 , x2 ) +
1
(x2 + 1 (x1 )).
x1
(9.9)
(9.10)
146
Note that the term z1 in 2 is used to tackle a cross-term caused by z2 in the dynamics
of z1 in the stability analysis. Other terms in 2 are taken to cancel the nonlinear terms
and stabilise the dynamics of z2 .
Step i. For 2 < i < n, the dynamics of zi are given by
zi = x i i1 (x1 , . . . , xi1 )
= xi+1 + i (x1 , . . . , xi )
i1
i1
j=1
xj
(xj+1 + j (x1 , . . . , xj ))
= zi+1 + i + i (x1 , . . . , xi )
i1
i1
j=1
xj
Design i as
i = zi1 ci zi i (x1 , . . . , xi )
+
i1
i1
j=1
xj
(9.11)
(9.12)
Note that similar to the design of 2 , the term zi1 is used to tackle a cross-term
caused by zi in the dynamics of zi1 in the stability analysis, and the other terms in i
are used to stabilise the dynamics of zi .
Step n. At the nal step, we have
zn = x n n1 (x1 , . . . , xn1 )
= u + n (x1 , . . . , xn )
n1
n1
j=1
xj
n1
n1
j=1
xj
(9.13)
(9.14)
Backstepping design
147
z =
c1
c2
0
..
.
1
..
.
c3
..
.
...
0
..
.
0
..
.
0
z := Az z.
..
.
1
..
. cn
1 T
z z.
2
(9.15)
ci zi2 2 min ci V .
i=1
i=1
(9.16)
(9.17)
148
0 1 0 ...
0 0 1 ...
0 0 0 ...
0 0 0 ...
0
0
.. ,
.
1
0
0
..
.
0
b=
b ,
.
..
bn
T
1
0
C=. ,
..
0
(9.18)
where x Rn is the state estimate and L Rn is an observer gain designed such that
(A LC) is Hurwitz. Let x = x x , and it is easy to see
x = (Ac LC)x.
(9.19)
The backstepping design can be carried out with the state estimate x in steps.
From the structure of the system (9.17) we have y = x1 . The backstepping design will
start with the dynamics of y. In the following design, we assume > 1. In the case
of = 1, control input can be designed directly without using backstepping.
To apply the observer backstepping through x in (9.18), we dene
z1 = y,
zi = x i i1 , i = 2, . . . , ,
(9.20)
z+1 = x +1 + b u ,
where i , i = 1, . . . , , are stabilising functions decided in the control design.
Consider the dynamics of z1
z1 = x2 + 1 (y).
(9.21)
Backstepping design
149
(9.22)
We design 1 as
1 = c1 z1 k1 z1 1 (y),
(9.23)
where ci and ki for i = 1, . . . , are positive real design parameters. Comparing the
backstepping design using the output feedback with the one using state feedback, we
have one additional term, k1 z1 , which is used to tackle the observer error x 2 in the
closed-loop system dynamics. Then from (9.22) and (9.23), we have
z1 = z2 c1 z1 k1 z1 + x 2 .
(9.24)
where l2 is the second element of the observer gain L, and in the subsequent design,
li is the ith element of L. We design 2 as
2 = z1 c2 z2 k2
+
1
y
2
z2 2 (y) l2 (y x 1 )
1 (y)
(x2 + 1 (y)).
y
(9.25)
1
y
2
z2 + z 3
1
x 2 .
y
j
j=1
i1
150
We design i , 2 < i , as
i = zi1 ci zi ki
i1
y
2
zi i (y) li (y x 1 )
i1
i1
(x2 + 1 (y)) +
x j .
+
y
x j
j=1
i1
(9.26)
(y, x 1 , . . . , x ) x +1
.
b
(9.28)
The stability result of the above control design is given in the following theorem.
Theorem 9.3. For a system in the form of (9.17), the dynamic output feedback control
with the input (9.28) and the observer (9.18) asymptotically stabilise the system.
Proof. From the observer error dynamics, we know that the error exponentially converges to zero. Since (Ac LC) is Hurwitz, there exists a positive denite matrix
P Rnn such that
(Ac LC)T P + P(Ac LC) = I .
This implies that for
Ve = x T P x ,
we have
V e = x 2 .
(9.29)
Let
Vz =
zi2 .
i=1
i1
i1
2
2
zi x 2 ,
ci zi ki
zi
V z =
y
y
i=1
where we dene 0 = y for notational convenience. Note that if we ignore the two
terms concerning with ki and x 2 in the dynamics of zi , the evaluation of the derivative
Backstepping design
151
of Vz will be exactly the same as the stability analysis that leads to Theorem 9.2. For
the cross-term concerning with x 2 , we have, from Youngs inequality,
i1
i1 2 2
1 2
k
z
x
zi +
x .
i
y i 2
y
4ki 2
Hence, we obtain that
1 2
2
ci zi +
x .
Vz
4ki 2
i=1
(9.30)
Let
1
Ve ,
V = Vz + 1 +
4d
ci zi +
x 2 1 +
x 2
V
4k
4d
i
i=1
ci zi2 x 2 .
i=1
b+1
x+1
x
= ... ... ,
b
xn
bn
and from the system dynamics (9.17), it can be shown that
b+1
b+1
+1 (y)
x
.. (y)
..
+ B ... ,
= B +
.
.
b
b
n (y)
bn
bn
where
b+1 /b
b+2 /b
..
B=
bn1 /b
bn /b
1
0
..
.
0
0
... 0
..
. 0
..
..
.
.
... 1
... 0
(9.31)
152
(9.32)
y = Cx.
Dene an input lter
1 = 1 1 + 2
..
.
(9.33)
1 = 1 1 + u,
where i > 0 for i = 1, . . . , 1 are the design parameters. Dene the ltered
transformation
= x
d i i ,
(9.34)
i=1
for i = 2, . . . , 1.
Backstepping design
153
d i (i i + i+1 ) d 1 (i 1 + u)
i=1
= Ac +
Ad i i + (y) +
i=1
= Ac +
d i i i
i=1
(A + i I )d i i + (y)
d i i+1
i=1
2
i=1
d i i+1
i=1
= Ac + (y) + d1 .
The output under the coordinate is given by
y = C +
C d i i
i=1
= C
because C d i = 0 for i = 1, . . . , 1, from the fact that d i,j = 0 for i = 1, . . . ,
1, 1 j i. Hence, under the ltered transformation, the system (9.32) is then
transformed to
= Ac + (y) + d1 ,
(9.35)
y = C .
Let us nd a bit more information of d. From the denition
d 2 = (Ac + 1 I )b
we have
n
d 2,i sni = (s + 1 )
bi sni .
i=1
di sni =
1
i=1
(s + i )
bi sni
(9.36)
which implies that d1 = b and that d is Hurwitz if b Hurwitz. In the special form
of Ac and C used here, b and d decide the zeros of the linear systems characterised
by (Ac , b, C) and (Ac , d, C) respectively as the solutions to the following polynomial
equations:
154
bi sni = 0,
di sni = 0.
i=1
Hence, the invariant zeros of (Ac , d, C) are the invariant zeros of (Ac , b, C) plus i for
i = 1, . . . , 1. For the transformed system, 1 can be viewed as the new input. In
this case, the relative degree with 1 as the input is 1. The ltered transformation lifts
the relative degree from to 1.
As the ltered transformation may have its use independent of backstepping
design shown here, we summarise the property of the ltered transformation in the
following lemma.
Lemma 9.4. For a system in the form of (9.32) with relative degree , the ltered
transformation dened in (9.34) transforms the system to (9.35) of relative degree 1,
with the same high frequency gain. Furthermore, the zeros of (9.35) consist of the
zeros of the original system (9.32) and i for i = 1, . . . , 1.
We introduce another state transform to extract the internal dynamics of (9.35)
with Rn1 given by
d2:n
= 2:n
y,
(9.37)
d1
where Rn1 forms the state variable of the transformed system together with y,
the notation ()2:n refers to the vector or matrix formed by the 2nd row to the nth row.
With the coordinates ( , y), (9.35) is rewritten as
= D + (y)
y = 1 + y (y) + b 1 ,
where D is the companion matrix of d given by
d2 /d1 1 . . . 0
d3 /d1 0 . . . 0
..
..
..
.. ,
D=
.
.
.
.
dn1 /d1 0 . . . 1
dn /d1 0 . . . 0
and
(y) = D
y (y) =
d2:n
d2:n
y + 2:n (y)
1 (y),
d1
d1
d2
y + 1 (y).
d1
(9.38)
Backstepping design
155
If we view 1 as the input, the system (9.35) is of relative degree 1 with the stable
zero dynamics. For such a system, there exists an output feedback law to globally and
exponentially stabilise the system.
For this, we have the following lemma, stating in a more stand alone manner.
Lemma 9.5. For a nonlinear system (9.32), if the relative degree is 1, there exist
a continuous function : R R with (0) = 0 and a positive real constant c such
that the control input in the form of
u = cy (y)
(9.39)
DT P + PD = 3I .
(9.40)
We set
(y) = y (y) +
P 2 (y) 2
.
y
(9.41)
Note that (0) = 0, and therefore (y) 2 /y is well dened. The closed-loop system
is then obtained as
= D + (y),
y = 1 cy
P 2 (y) 2
.
y
Let
1
V = T P + y2 .
2
We have
V = cy2 3 2 + 2 T P(y) + y1 P 2 (y) 2 .
With the inequalities of the cross terms
|2 T P(y)| 2 + P 2 (y) 2 ,
1
|y1 | y2 + 2 ,
4
we have
1 2
y 2 .
V c
4
156
Therefore, the proposed control design with c > 14 and (9.41) exponentially stabilises
the system in the coordinate ( , y), which implies the exponential stability of the
closed-loop system in x coordinate, because the transformation from ( , y) to x is
linear.
2
From Lemma 9.5, we know the desired value of 1 . But we cannot directly assign
a function to 1 as it is not the actual control input. Here backstepping can be applied
to design control input based on the desired function of 1 . Together with the ltered
transformation, the overall system is given by
= D + (y)
y = 1 + y (y) + b 1
1 = 1 1 + 2
(9.42)
...
1 = 1 1 + u,
to which the backstepping design is then applied. Indeed, in the backstepping design,
i for i = 1, . . . , 1 can be viewed as virtual controls.
Let
z1 = y,
zi = i1 i1 ,
for i = 2, . . . ,
z+1 = u ,
where i for i = 2, . . . , are stabilising functions to be designed. We also use the
positive real design parameters ci and ki for i = 1, . . . , and > 0.
Based on the result shown in Lemma 9.5, we have
1 = c1 z1 k1 z1 + y (y)
P 2 (y) 2
y
(9.43)
and
z1 = z2 c1 z1 k1 z1 P 2 (y) 2 .
(9.44)
z2 = 1 1 + 2
(9.45)
Backstepping design
The resultant dynamics of z2 is obtained as
1 2
1
z2 = z1 c2 z2 k2
z2 + z 3
1 .
y
y
157
(9.46)
zi = i1 i1 + i
= zi+1 + i i1 i1
i2
i1
j=1
i1
(1 + (y))
y
(j i1 + j+1 ).
i2
i1
j=1
i1
y
2
zi +
i1
(y)
y
(j j + j+1 ) + i1 i1 .
(9.47)
(9.48)
i2
1
j=1
(j j + j+1 ) + 1 1 .
(9.49)
For the control design parameters, ci , i = 1, . . . , and can be any positive, and for
di , the following condition must be satised:
1
.
4ki
1=1
The stability result of the above control design is given in the following
theorem.
158
Theorem 9.6. For a system in the form of (9.32), the dynamic output feedback
control (9.49) obtained through backstepping with the input ltered transformation
asymptotically stabilises the system.
Proof. Let
Vz =
zi2 .
i=1
From the dynamics for zi shown in (9.44), (9.46) and (9.48) we can obtain that
2
i1
i1
2
2
ci zi ki
zi
zi 1 P 2 (y) 2 ,
V z =
y
y
i=1
where we dene 0 = y for notational convenience. For the cross-term concerning
with 1 , we have
i1
i1 2 2
1 2
zi +
.
y zi 1 ki
y
4ki 1
Hence, we obtain that
1 2
ci zi2 +
1 P 2 (y) 2 .
V z
4ki
i=1
(9.50)
Let
V = T P
and we can obtain, similar to the proof of Lemma 9.5,
V 2 2 + P 2 (y) 2 .
Let
V = Vz + V .
and we have
V
ci zi2 +
i=1
1 2
1 2 2
4ki
ci zi2 2 .
(9.51)
(9.52)
i=1
Therefore, we have shown that the system (9.42) is exponentially stable under the
coordinate ( , z1 , . . . , z1 ). With y = z1 , we can conclude limt (t) = 0. From
y = z1 , and 1 (0) = 0, we can conclude that limt 1 (t) = 0. Following the same
process, we can show that limt (t) = 0 for i = 1, . . . , 1. Finally from the
ltered transformation (9.34), we can establish limt x(t) = 0.
2
Backstepping design
159
(9.53)
(9.54)
(9.55)
where c is a positive real constant, and
Rpp is a positive denite gain matrix.
The closed-loop dynamics is given by
y = cy + T (y)
with the usual notation = .
For stability analysis, let
V =
1 2 1 T 1
y +
2
2
160
(9.56)
x n1 = xn + n1 (x1 , x1 , . . . , xn1 )
T
x n = u + n (x1 , x2 , . . . , xn )T ,
where xi R for i = 1, . . . , n are state variables; Rp is an unknown vector of
i
constant parameters; i : R R Rp for i = 1, . . . , n are differentiable
functions up to the order n i with i (0, . . . , 0) = 0; and u R is the control input.
Note that this system is exactly the same as (9.6) if the parameter vector is
known.
The backstepping design method will be applied iteratively in a similar way to
the control design for (9.6), with only difference of including adaptive parameters in
the control design.
Let
z1 = x1 ,
zi = xi i1 (xi , . . . , xi1 , ),
for i = 2, . . . , n,
zn+1 = u n (xi , . . . , xn , ),
1 = 1 ,
i = i
i1
i1
j=1
xj
j ,
for i = 2, . . . , n,
1 = c1 z1 1T ,
(9.57)
z1 = c1 z1 + z2 + 1T ,
where = .
Backstepping design
161
x1
1
1
.
= z3 + 2 + 2T
x2
x1
= x3 + 2T
The dynamics of z2 involve the adaptive law . Even though the adaptive law has
not been designed, it is surely known to the control design, and can be used in the
control input. However, the dynamics of zi with i > 2 will also affect the design of the
adaptive law. For this reason, we would like to leave the design of the adaptive law to
the end. Inevitably, the adaptive law will include zi for i > 2. This causes a problem,
if we use 2 to cancel at this step, because z3 depends on 2 . Instead, we only deal
with the part of the adaptive law that depends on z1 and z2 at this step, and we denote
that as 2 , which is a function of z1 , z2 and . In the subsequent steps, we use notations
i = i (z1 , . . . , zi , ) for i = 3, . . . , n, which are often referred to as tuning functions.
Based on the above discussion, we design 2 as
1
1
x2 +
2 .
x1
The closed-loop dynamics are obtained as
2 = z1 c2 z2 2T +
z2 = z1 c2 z2 + z3 + 2T
(9.58)
1
( 2 ).
Then the adaptive backstepping can be carried on for zi with 2 < i n. The
dynamics of zi can be written as
zi = x i i1
= xi+1 + iT
i1
i1
j=1
= zi+1 + i +
iT
xj
x j
i1
i1
j=1
xj
i1
xj+1
i1
i1
i1
j=1
xj
xj+1 +
i1
i + i ,
(9.59)
i1
( i ) + i .
162
The control input u appears in the dynamics zn , in the term zn+1 . When i = n,
we have i = by denition. We obtain the control input by setting zn+1 = 0, which
gives
u = n
= zn1 cn zn nT
+
n1
n1
xj
j=1
xj+1 +
n1
+ n .
(9.60)
We need to design the adaptive law, tuning functions and i to complete the
control design. We will do it based on Lyapunov analysis. For notational convenience,
we set 1 = 2 = 0.
Let
n
1 2 1 T 1
V =
z +
,
(9.61)
2 i=1 i
2
where
Rpp is a positive denite matrix. From the closed-loop dynamics of zi ,
for i = 1, . . . , n, we obtain
n
i1
zi
( i ) + T
1
i=1
i=2
n
n
i1
2
zi i zi
ci zi +
( i )
=
i=1
i=2
n
T
1
+
z
.
V =
(ci zi2 + zi iT + zi i )
i i
i=1
zi i .
(9.62)
i=1
zi i z i
( i ) = 0.
i=2
Substituting the adaptive law in the above equation, and with some manipulation of
index, we obtain that
n
n
zi i zi i1
0=
zj j i
i=2
n
i=2
j=1
z i i
n
n
i=2 j=2
i1
i1
j +
(i z1
1 )
zi zj
zi
n
i=2
Backstepping design
=
zi i
n
n
i=2
i=2 j=i+1
i
n
zi i
i1
j
i1
i1
zi
j +
(i z1
1 )
n
zi zj
i=2 j=2
zi zj
163
i=2
j1
n
i=2
j=3 i=2
zi zj
i1
j
n
i
i1
+
zi
i z1
1
zj
j
i=2
j=2
i1
j1
=
i
zj
i=3
j=2
n
i
i1
+
i z1
1
zi
zj
j .
zi i
i=2
j=2
Hence, we obtain
i1
j1
i =
i ,
zj
j=2
i =
zj j ,
for i = 3, . . . , n,
for i = 2, . . . , n.
(9.63)
(9.64)
j=1
ci zi2 .
i=1
164
1
1
(x2 + (ex1 1) )
,
x1
where
1
= c1 ex1 ,
x1
1
= (ex1 1).
(x2 + (ex1 1) ) +
.
x1
Note that the control input u contains , which is to be designed later, as a function
of the state variables and . The resultant dynamics of z2 are obtained as
u = z1 c2 z2 +
z2 = z1 c2 z2
Let
1 x1
(e 1) .
x1
1 2
2
2
V =
z + z2 +
.
2 1
We obtain that
1 x1
1
2
2
x1
V = c1 z1 c2 z2 + z1 (e 1) z2
(e 1) .
x1
Backstepping design
165
Example 9.3. In Example 9.2, the nonlinear system is in the standard format as
shown in (9.56). In this example, we show adaptive control design for a system which
is slightly different from the standard form (9.56), but the same design procedure can
be applied with some modications.
Consider a second-order nonlinear system
x 1 = x2 + x13 + x12
x 2 = (1 + x12 )u + x12 ,
where R is the only unknown parameter.
Let z1 = x1 and z2 = x2 1 . The stabilising function 1 is designed as
1 = c1 z1 x13 x12 ,
which results in the dynamics of z1 as
z1 = c1 z1 + z2 + x13 .
The dynamics of z2 are obtained as
z2 = (1 + x12 )u + x12
1
1
where
1
1
= x13 .
= c1 3x12 2x1 ,
x1
Therefore, we design the control input u as
1
1
1
2
3
2
u=
z1 c2 z2 x1 +
(x2 + x1 + x1 ) +
.
x1
1 + x12
The resultant dynamics of z2 are obtained as
z2 = z1 c2 z2 + x12
Let
1 3
x .
x1 1
1 2
2
2
V =
z + z2 +
.
2 1
We obtain that
1
1 3
166
x2
State variables
0
1
2
3
4
5
6
0
10
10
t (s)
Estimated parameter
3
2.5
2
1.5
1
0.5
0
6
t (s)
Backstepping design
167
The rest part of the stability analysis follows Theorem 9.7. Simulation results are
shown in Figures 9.1 and 9.2 with x(0) = [1, 1]T , c1 = c2 = = = 1. The state
variables converge to zero as expected, and the estimated parameter converges to
a constant, but not to the correct value = 1. In general, estimated parameters in
adaptive control are not guaranteed to converge to their actual values.
i (y)ai
i=1
(9.65)
y = Cx
with
T
0 1 0 ... 0
1
0 0 1 ... 0
Ac = ... ... ... . . . ... , C = . ,
..
0 0 0 ... 1
0
0 0 0 ... 0
0
..
a1
.
a2
0
:= 0(1)1 ,
a = . , b =
b
b
..
.
ap
..
bn
168
are unknown parameters, we will not be able to design an observer in the same way
as in the observer backstepping. Instead, we can design lters similar to observers,
and obtain an expression of state estimation which contains unknown parameters. The
unknown parameters in the state estimation are then tackled by adaptive backstepping.
For the state estimation, we re-arrange the system as
x = Ac x + 0 (y) + F T (y, u) ,
(9.66)
F(y, u) =
T
0(1)(n+1)
(y)u, (y) .
In+1
(9.67)
= A0 + F(y, u) ,
(9.68)
(9.69)
Let
= x x
and from direct evaluation, we have
= A0 .
(9.70)
for j = , . . . , n,
where ej denotes jth column of identity matrix I in Rn . For 1 < j < n, we have
A0 ej = (Ac LC)ej = Ac ej = ej+1 .
Backstepping design
169
vj = A0 vn .
Finally, we summarise the lters for T as
T
vj
=
=
=
=
[v , . . . , vn , ],
A0 + (y),
A0 + en (y)u,
nj
A0 , for j = , . . . , n.
(9.71)
With the lters being designed, control design can be carried out using adaptive
backstepping in a similar way as the observer backstepping, by combining the design
of tuning functions in parameter adaptation. During the control design, the state
variable x will be replaced by
x = + T + ,
and in particular, whenever we encounter x2 , we will replace it by
x2 = 2 + T(2) + 2 ,
where the subscript (i) denotes the ith row of a matrix. In the following control design,
we will consider the tracking control instead of stabilisation. The output y is designed
to track a trajectory yr with its derivatives available for the control design.
Let us dene a number of notations:
z1 = y yr ,
zi = v,i y
r(i1) i1 , i = 2, . . . , ,
z+1 = (y)u + v,+1 y
r() ,
0 = 2 + 0,1 ,
= [v,2 , . . . , vn,2 , (1) + (2) ]T ,
= [0, v+1,2 , . . . , vn,2 , (1) + (2) ]T ,
i = [1 , . . . , i ]T ,
y i = [yr , y r , . . . , yr(i) ]T ,
Xi = [ T , vec()T , ,
Ti , y iT ]T ,
j,i =
j1 i1
,
y
where ,
an estimate of = 1/b , and i , i = 1, . . . , , are the stabilising functions
to be designed.
Consider the dynamics of z1
z1 = x2 + 0,1 (y) + (1) y r
= 2 + T(2) + 2 + 0,1 (y) + (1) y r
= b v,2 + 0 + T + 2 y r .
(9.72)
170
Note that b is unknown. To deal with unknown control coefcient, we often estimate
its reciprocal, instead of itself, to avoid using the reciprocal of an estimate. This is the
reason why we dene
yr
z2 = v,2 1
:= v,2 1
yr .
From b = 1, we have
b = 1 b ,
where = .
Then from (9.72), we have
yr ) + 0 + T + 2 y r
z1 = b (z2 + 1 +
= b z2 b (
1 + y r ) + 1 + 0 + T + 2 .
(9.73)
Hence, we design
1 = c1 z1 k1 z1 0 T ,
(9.74)
where ci and ki for i = 1, . . . , are positive real design parameters. Note that with
= 1 , we have 1 = 1 (y, X1 , ).
The resultant closed-loop dynamics are obtained as
1 + y r ) + T + 2
z1 = c1 z1 k1 z1 + b z2 b (
= c1 z1 k1 z1 + b z2 + b z2 b (
1 + y r ) + T + 2
= c1 z1 k1 z1 + b z2 b (
1 + y r )
+ ( (
1 + y r )e1 )T + 2 ,
(9.75)
Note that b = T e1 .
where = .
As shown in the observer backstepping, the term k1 z1 is used to tackle the
error term 2 . In the subsequent steps, the terms headed by ki are used to deal with
the terms caused by 2 in stability analysis. Indeed, the method of tackling observer
errors is exactly the same as in the observer backstepping when all the parameters are
known. The adaptive law for can be designed in this step, as it will not appear in
the subsequent steps,
= sgn(b )( 1 + y r ))z1 ,
(9.76)
i = 2, . . . , ,
(9.77)
Backstepping design
171
1
1
1
X1
yr
y
yr .
X1
y
With
y = 0 + T + 2 ,
z3 = v,3 2
yr ,
we obtain that
z2 = z3 + 2 l2 v,1
1
1
1
X1
yr ,
(0 + T + 2 )
X1
y
z2 + l2 v,1 +
2 = b z1 c2 z2 k2
y
+
1
1
+ 1 2 .
X1 +
(0 + T )
X1
y
(9.78)
1 T
1
1
2
( 2 ).
y
y
(9.79)
i1
i1
Xi1
(0 + T + 2 )
Xi1
y
i1 (i1)
y
r .
We design i , 2 < i , as
i = zi1 ci zi ki
(i1) +
+ y
r
i1
y
2
zi + li v,1
i1
i1
Xi1 +
(0 + T )
Xi1
y
i1
i
j,i zj ,
i1
i = 3, . . . , ,
(9.80)
j=2
where the last term i1
j=2 j,i zj is similar to the term i in the adaptive backstepping
with tuning functions. The resultant dynamics of zi are obtained as
172
i1
y
2
zi + zi+1
i1 T
i1
j,i zj
2
y
y
j=2
i1
( i )
i1
i = 3, . . . , .
(9.81)
(9.82)
1
r() ).
( v,+1 + y
(y)
(9.83)
For the adaptive observer backstepping, we have the following stability result.
Theorem 9.8. For a system in the form of (9.65), the control input (9.83) and adaptive laws (9.76) and (9.82) designed by adaptive observer backstepping ensure the
boundedness of all the variables and limt (y(t) yr (t)) = 0.
Proof. With the control design and adaptive laws presented earlier for the case > 1,
the dynamics of zi , for i = 1, . . . , can be written as
z1 = c1 z1 k1 z1 + 2 + ( (
yr + 1 )e1 )T
b (yr + 1 ) + b z2 ,
z2 = b z1 c2 z2 k2
1
y
(9.84)
2
z2 + z 3
1
1 T
2 +
2,j zj ,
y
y
j=3
zi = zi1 ci zi ki
i1
y
(9.85)
2
zi + zi+1
i1 T
i1
2 +
i,j zj
y
y
j=i+1
i1
j=2
j,i zj
i = 3, . . . , .
(9.86)
Backstepping design
173
Let
1 2 1 T 1
|b | 2 1 T
V =
zi +
+
P,
+
2 i=1
2
2
4ki
i=1
(9.87)
i1 2 2 i1
1
V =
ci + k i
zi
zi
2
2 ,
y
y
4k
i
i=1
i=1
i=1
0
= 1. Noting
where we set
y
2
i1
zi
1 2 2 + ki i1 z 2 ,
2
i
y
4k
y
i
(9.88)
(9.89)
we have
V
ci zi2 .
(9.90)
i=1
Chapter 10
Disturbances are often inevitable in control system design. A well-performed controller is expected to suppress undesirable effects of the disturbances in the system.
There are various types of disturbances in physical systems, from random disturbances, wide-band, narrow-band disturbances, in terms of disturbance power spectra,
to deterministic disturbances that include harmonic disturbances, i.e., sinusoidal functions, general periodic disturbances and other deterministic signals generated from
nonlinear dynamic systems such as limit cycles. The spectral information of random
disturbances may be considered in loop-shaping and other conventional design methods. In this chapter, we concentrate on suppression and rejection of deterministic
periodic disturbances. One control design objective is to track a specic signal. If we
take the tracking error as the state variable, the tracking problem could be converted
to a stabilisation problem. Indeed, we can formulate both the disturbance rejection
and output tracing problems in terms of output regulation. This will be discussed later
in this chapter.
Disturbance rejection and output regulation are big topics in control systems,
design. We have to be selective to limit the contents in one chapter. In this chapter, we
will concentrate on rejection of deterministic disturbances in a class of nonlinear output feedback systems. As for disturbances, we will start from sinusoidal disturbances
with unknown frequencies, then disturbances generated from nonlinear exosystems
and then general periodical disturbances, etc. Adaptive control techniques are used
to deal with the unknown disturbance frequencies and unknown parameters in the
system. The presented design concepts can be applied to other classes of nonlinear
systems.
(10.1)
176
with
0 1 0 ...
0 0 1 ...
0 0 0 ...
0 0 0 ...
0
0
.. ,
.
1
0
0
..
.
0
b=
b ,
.
..
bn
T
1
0
C=. ,
..
0
177
(10.3)
(10.4)
n
n
d ni
d n+1
d ni
b
(w)
=
Ei w.
i
dt ni
dt n+1
dt ni
i=
i=
(10.5)
A solution of (w) can always be found from (10.5). With (w), we can write, for
< i n,
i =
d
i1 Ei1 w bi1 (w).
dt
(10.6)
2
(10.7)
(10.8)
(10.9)
(10.10)
178
(10.11)
(10.12)
(10.14)
q2 = T ,
with the initial value (0) dependent on exogenerous variables.
Remark 10.4. The importance of (10.14) compared with (10.13) is the re-formulation
of the uncertainty caused by the unknown exosystem. The uncertainty in (10.13)
parameterised by unknown S and l is represented by a single vector in (10.14). The
relation between the two parameterisations is discussed here. Suppose that M Rmm
is the unique solution of
MS FM = Gl T .
(10.15)
The existence of a non-singular M is ensured by the fact that S and F have exclusively different eigenvalues, and (S, l) and (F, G) are observable and controllable
respectively. From (10.15), we have
MSM 1 = F + Gl T M 1 ,
(10.16)
1
179
(10.17)
where p is generated from (10.9) with q and satisfying (10.10) and (10.12)
respectively. In particular,
x2 = p2 T + 2 ,
(10.18)
(10.19)
(10.20)
i = 2, . . . , ,
z+1 = b u + p+1 ,
(10.21)
(10.22)
(10.23)
(10.24)
We design 1 as
1 = c1 z1 k1 z1 1 (y) + T .
(10.25)
180
(10.26)
After the design of 1 , the remaining stabilising functions can be designed in a similar
way to the standard adaptive backstepping with tuning functions shown in the previous
chapter. Omitting the deriving procedures, we briey describe the nal results
i1 2
zi li (y p1 ) i (y)
i = zi1 ci zi ki
y
+
i1
i1
i1
i1
(p2 T + 1 ) +
p j +
kj +
j
y
p
k
j=1
j=1
j1 i1
i1
i +
zj ,
y
i1
i1
i1
i = 2, . . . , ,
(10.27)
j=2
where li are the ith element of the observer gain L in (10.9), is a positive denite
matrix, i , i = 2, . . . , , are the tuning functions dened by
i =
i
j1
zj ,
y
j=1
where we set
0
y
i = 2, . . . , ,
(10.28)
= .
The adaptive laws for ki , i = 1, . . . , , are given by
i1 2 2
zi ,
ki = i
y
(10.29)
(10.30)
where i is a positive real design parameter. The control input is obtained by setting
z+1 = 0 as
u=
1
( p+1 ).
b
(10.31)
(10.32)
(10.33)
(10.34)
181
(10.35)
(10.36)
i1
y
2
zi + zi+1
i1
i1 j
2
zj
y
y
j=i+1
i1
j1 i1
zj ,
y
j=2
i = 2, . . . , ,
i1 T
( T )
y
(10.37)
(10.38)
with = .
In the following analysis, we denote 0 , i,j , i = 1, . . . , 3, and j = 1, . . . , , as
constant positive reals, which satisfy
0 +
j=1
3,j <
1
.
2
(10.39)
182
2,i | T G|2 +
j=2
ki > 1,i +
1
2
+
42,i
43,i
2
,
43,1
i = 2, . . . , .
(10.40)
(10.41)
1 T
2
1 2
T
1
T
V =
e Pe +
zi +
i ki + + P ,
2
i=1
i=1
(10.42)
>
PG2 1
+
.
40
41,i
j=1
(10.43)
i1
i1
2
ci + k i
zi
zi
2
V =
y
y
i=1
i=1
i1 T
zi
( e + T Gz1 )
y
i=1
eT e eT PG(z2 + 2 ) T .
(10.44)
2
i1
i1
1 2
zi
zi2 +
,
i = 1, . . . , ,
y 2
< 1,i
y
41,i 2
i1 T
i1 2 2
1
T
2 2
zi ,
i = 2, . . . , ,
y Gz1 zi
< 2,i | G| z1 + 4
y
2,i
i1 T
2 i1 2 2
T
zi ,
i = 1, . . . , ,
y ezi
< 3,i e e + 4
y
3,i
e
|eT PGz2 | < eT + PG2 z22 /2.
2
183
Then based on the conditions specied in (10.32), (10.39), (10.40), (10.41) and
(10.43), we can conclude that there exist positive real constants i , for i = 1, 2, 3,
such that
V 1
zi2 2 eT e 3 T .
(10.45)
i=1
= limt (t).
From (10.14), (10.19) and (10.35), we obtain
d
( ) = F( ) + G( )T + (t),
dt
(10.46)
where
T
T + G
(t) = G( )
(e + Gy) .
(10.47)
184
where w1 R is a disturbance with two sinusoidal components at unknown frequencies 1 and 2 . An augmented disturbance vector w = [w1 , w2 , w3 , w4 ]T satises
w = Sw with eigenvalues of S at {j1 , j2 }. It is easy to obtain that = [0, w1 ]T
and = w1 w 1 . Using the state transform x = , we have
x 1 = x2 + (ey 1)
x 2 = u
(10.48)
y = x1 .
A transform w = T w can also be introduced to the disturbance model so that the
disturbance is generated by
0
1
022
1 0
w := S w
w =
0
2
022
2 0
w1 = [t11 , t12 , t13 , t14 ]w,
0
1
0
0
0
0
0
1
0
, G = 0 ,
F =
0
0
0
1
0
1 4 6 4
1/10
which is controllable with the eigenvalues of F at {1, 1, 1, 1} and PG < 1.
We set c1 = c2 = 10, with the condition (10.32) being satised. Other parameters
in the control design were set as 1 = 2 = 1, = 1000I , L1 = 3 and L2 = 2. The
disturbance was set as
w1 = 4 sin 1 t + 4 sin 2 t
with 1 = 1 and
2,
250 > t 0,
2 =
1.5, t 250.
A set of simulation results are presented here. Figure 10.1 shows the system output
together with the control input. Figure 10.2 shows the estimates of . The ideal values
of are [30, 40, 10, 40]T and [12.5, 40, 27.5, 40]T for 1 = 1, 2 = 2 and 1 = 1,
0.4
0.2
0
0.2
50
100
150
200
250
300
350
400
450
500
300
350
400
450
500
450
500
Time (s)
30
20
10
0
10
20
0
50
100
150
200
250
Time (s)
80
60
Estimated parameters
40
20
20
40
60
80
50
100
150
200
250
Time (s)
300
350
400
185
186
2 = 1.5 respectively. Under both sets of the frequencies, converged to the ideal
values. It can be seen through this example that the disturbance of two unknown
frequencies has been rejected completely.
(10.49)
e = y q(w),
with
0 1 0 ...
0 0 1 ...
0 0 0 ...
0 0 0 ...
0
0
.. ,
.
1
0
0
..
.
0
b=
b ,
.
..
bn
T
1
0
C=. ,
..
0
187
188
where i > 0, for i = 1, . . . , 1, are the design parameters, and the ltered
transformation
z = x [d 1 , . . . , d 1 ] ,
(10.51)
(10.52)
y = C z ,
where d = (Ac + 1 I )d 1 . It has been shown in (9.36) that d1 = b and
n
di s
ni
i=1
1
(s + i )
i=1
n
bi sni .
(10.53)
With 1 as the input, the system (10.52) is with relative degree 1 and minimum phase.
We introduce another state transform to extract the internal dynamics of (10.52) with
z Rn1 given by
z = z2:n
d2:n
y,
d1
(10.54)
where ()2:n refers to the vector or matrix formed by the 2nd row to the nth row. With
the coordinates (z, y), (10.52) is rewritten as
z = Dz + (y, w, )
y = z1 + y (y, w, ) + b 1 ,
(10.55)
where the unknown parameter vector = [aT , bT ]T , and D is the left companion
matrix of d given by
d2 /d1 1 . . . 0
d3 /d1 0 . . . 0
..
..
..
.. ,
(10.56)
D=
.
.
.
.
dn1 /d1 0 . . . 1
dn /d1 0 . . . 0
and
d2:n
d2:n
y + 2:n (y, w, a)
1 (y, w, a),
d1
d1
d2
d2:n
y (y, w, ) =
y+
1 (y, w, a).
d1
d1
(y, w, ) = D
Notice that D is Hurwitz, from (9.36), and that the dependence of d on b is reected in
the parameter in (y, w, ) and y (y, w, ), and it is easy to check that (0, w, ) = 0
and y (0, w, ) = 0.
189
The solution of the output regulation problem depends on the existence of certain
invariant manifold and feedforward input. For this problem, we have the following
result.
Proposition 10.4. Suppose that an invariant manifold (w) Rn1 satises
(w)
S( )w = D (w) + (q(w), w, ).
w
(10.57)
(w, , ) =
b1
q(w)
S( )w 1 (w) y (q(w), w, ) .
w
(10.58)
(10.59)
Proof. With 1 being viewed as the input, is the feedforward term used for output
regulation to tackle the disturbances, and from the second equation of (10.55), we
have
q(w)
(w, , ) = b1
(w)
(q(w),
w,
)
.
S(
)w
1
y
w
From the structure of the exosystem, the disturbances are sinusoidal functions.
Polynomials of sinusoidal functions are still sinusoidal functions, but with some highfrequency terms. Since all the nonlinear functions involved in the system (10.49) are
polynomials of their variables, the immersion in (10.58) always exists. For a controllable pair (F, G), M is an invertible solution of (10.59) if (, ) is observable, which
is guaranteed by the immersion.
2
We now introduce the last transformation based on the invariant manifold with
z = z
(10.60)
190
(10.61)
...
1 = 1 1 + u,
where
= (y, w, ) (q(w), w, )
and
y = y (y, w, ) y (q(w), w, ).
Since the state in the internal model is unknown, we design the adaptive internal
model
= F + G1 .
(10.62)
(10.63)
z1 + b1
= F FGb1
e + b G
G y .
(10.64)
(10.65)
where the Nussbaum gain N is a function (e.g. N () = 2 cos ) which satises the
two-sided Nussbaum properties
1
N (s)ds = +,
(10.66)
lim sup
0
1
lim inf
N (s)ds = ,
(10.67)
0
where denotes + and respectively. From (10.61)
and the denition of the Nussbaum gain, we have
e = z1 + (b N 1)1 + 1 + b 1 + b 1 lbT + y ,
191
(10.68)
k0 = e2 + e2p ,
b,0 = 1 e,
(10.69)
l,0 = e,
where b,0 and l,0 denote the rst tuning functions in adaptive backstepping design
for the nal adaptive laws for b and lb . If the relative degree = 1, we set u = 1 .
For > 1, adaptive backstepping can be used to obtain the following results:
2 = b e c1 1 k1
1
e
2
1
1
1
(b 1 lbT )
+
e
1
k0 +
k0
1
l,1 ,
lb
(10.70)
192
i1
i1
(b 1 lbT )
+
i1
i1
i1
b,i1 +
k0 +
l,1
k0
b
lb
i
i1 j2
j=4
i
i1 j2
j=3
lb
1 j2
j2
for i = 2, . . . , ,
(10.71)
b,i = b,i1
l,i
for i = 1, . . . , 1,
for i = 1, . . . , 1,
b = b,1 ,
(10.72)
lb = l,1 .
(10.73)
(10.74)
For the proposed control design, we have the following result for stability.
Theorem 10.5. For a system (10.49) satisfying the invariant manifold condition
(10.57), the adaptive output regulation problem is globally solved by the feedback
control system consisting the -lters (10.50), the adaptive internal model (10.62),
Nussbaum gain parameter (10.65), the parameter adaptive laws (10.69), (10.72),
(10.73) and the feedback control (10.74), which ensures the convergence to zero of
the regulated measurement, and the boundedness of all the variables in the closed-loop
system.
Proof. Dene a Lyapunov function candidate
V = 1 T P + 2 z T Pz z
1
1 2 2
2
2
T
+
e +
i + (k0 k0 ) + b + lb lb ,
2
i=1
193
where 1 and 2 are two positive reals and Pz and P are positive denite matrices
satisfying
Pz D + DT Pz = I ,
P F + F T P = I .
With the design of i , for i = 1, . . . , , the dynamics of i can be easily evaluated.
From the dynamics of z in (10.61) and the dynamics of in (10.64), virtual controls
and adaptive laws designed earlier, we have the derivative of V as
V = 1 ( T 2 T P Fb1
T P b1
z1 + 2 T P b1
Ge + 2
G
G y )
+ (k0 k0 )(e2 + e2p )
+ 2 (z T z + 2z T Pz )
c0 e2 k0 (e2 + e2p ) + (b N 1)e1
+ ez1 + e y elbT + l T Ge2
2
1
i
A
i
2
ci i ki
+
i2 i
z1
e
e
i=1
i
i T
i T
y + i
l i
l Ge .
i
e
e b
e
The stability analysis can be proceeded by using the inequalities 2xy < rx2 + y2 /r
or xy < rx2 + y2 /(4r) for x > 0, y > 0 and r being any positive real, to tackle the
cross-terms between the variables z , ,
e, i , for i = 1, . . . , 1. It can be shown
that there exist sufciently big positive real 1 , and then sufciently big positive real
2 , and nally the sufcient big k0 such that the following result holds:
1
1
1
ci i2 .
V (b N () 1) 1 T 2 z T z c0 e2
3
4
i=1
(10.75)
(t)
(10.76)
If (t), t R+ , is not bounded from above or below, then from (10.66) and (10.67)
it can be shown that the right-hand side of (10.76) will be negative at some instances
of time, which is a contradiction, since the left-hand side of (10.76) is non-negative.
194
z , e, i L2 L and Barbalats
i
lemma, limt = 0, limt z = 0, limt e(t) = 0 and limt i = 0 for
i = 1, . . . , 1.
2
(10.77)
0 1 0 ...
0 0 1 ...
0 0 0 ...
0 0 0 ...
0
0
.. ,
.
1
0
195
0
..
.
0
b=
b ,
.
..
T
1
0
C=. ,
..
0
bn
w = s(w)
(10.78)
di+1
di+2
di+1 d2
di+1
zi =
y + (i+1 (y)
z1 + zi+1 +
1 (y))a
d1
d1
d1
d12
di+1
E1 (w), i = 1, . . . , n 2,
(10.79)
d1
dn
d n d2
dn
dn
= z1 2 y + n (y) 1 (y) a + En (w) E1 (w),
d1
d1
d1
d1
+ Ei+1 (w)
zn1
y = z1 +
d2
y + 1 (y)a + E1 (w) + b 1
d1
196
i (w)
di+2
di+1 d2
di+1
1 (w) + i+1 (w) + q(w)
s(w) =
w
d1
d1
d12
+ Ei+1 (w) E1 (w)
di+1
+ (i+1 (q(w))
d1
di+1
1 (q(w)))a, i = 1, . . . , n 2,
d1
n1 (w)
dn
dn d2
s(w) = 1 (w) 2 q(w) + (n (q(w))
w
d1
d1
dn
dn
1 (q(w)))a + En (w) E1 (w).
d1
d1
Proof. Since the last equation of input lter (10.50) used for the ltered transformation
is an asymptotically stable linear system, there is a static response for every external
input u(w), i.e., there exists a function 1 (w) such that
1 (w)
s(w) = 1 1 (w) + (w).
w
Recursively, if there exists i (w) such that
i (w)
s(w) = i i (w) + i+1 (w),
w
then there exists i1 (w) such that
i1 (w)
s(w) = i1 i1 (w) + i (w).
w
Dene
(w)
= Da ( (w) [d 1 , . . . , d 1 ] ),
q(w)
where = [1 , . . . , 1 ]T
d2 /d1 1
..
..
.
.
Da =
dn /d1 0
1
0
and
... 0
.
..
. ..
.
... 1
... 0
197
It can be seen that (w) satises the dynamics of z along the trajectories of (10.78)
as shown in (10.80), and hence the proposition is proved.
2
Based on the above lemma, we have
d2
q(w)
s(w) = 1 (w) + q(w) + 1 (q(w))a + E1 (w) + b (w),
w
d1
where (w) = 1 (w). With 1 being viewed as the input, (w) is the feedforward term
used for output regulation to tackle the disturbances, and it is given by
q(w)
d2
(w)
q(w)
(q(w))a
E
(w)
.
s(w)
= b1
1
1
1
w
d1
We now introduce the last transformation based on the invariant manifold with
z = z (w(t)).
Finally we have the model for the control design
zi = di+1 z1 + zi+1 + di+2 di+1 d2 e
d1
d1
d12
+ (i+1 (y) i+1 (q(w)))a
di+1
(1 (y) 1 (q(w)))a, i = 1, . . . , n 2
d1
dn
d n d2
= z1 2 e + (n (y) n (q(w)))a
d1
d1
zn1
dn
(1 (y) 1 (q(w)))a
d1
d2
e = z1 + e + (1 (y) 1 (q(w)))a + b (1 (w)),
d1
d2
e + (1 (y) 1 (q(w)))a + b (1 (w)),
d1
T
d3
d2
dn
dn1 d2 dn d2
=
22 , . . . ,
,
d1
d1
d1
d12
d12
d2
2 (y) 2 (q(w)) (1 (y) 1 (q(w)))
d1
.
.
..
(y, w, d) =
dn
n (y) n (q(w)) (1 (y) 1 (q(w)))
d1
(10.81)
198
Let
Vz = z T Pd z ,
where
Pd D + DT Pd = I .
Then using 2ab ca2 + c1 b2 and 2 (|e|) 2 (1 + e2 ), there exist unknown positive
real constants 1 and 2 such that
V z = z T z + 2z T Pd (e + (y, w, d)a)
3
z T z + 1 e2 + 2 e2 2 (1 + e2 ),
4
noting that
1
2z T Pd e z T z + 8eT T Pd2 e
8
1
z T z + 1 e2 ,
8
and
1
2z T Pd (y, w, d)a z T z + 8aT T Pd2 a
8
1
z T z + 12 ||2
8
1
z T z + 12 2 |e|2 2 (|e|)
8
1 T
z z + 2 e2 2 (1 + e2 ),
8
where 12 is an unknown positive real constant.
(10.82)
199
Now let us consider the internal model design. We need an internal model to
produce a feedforward input that converges to the ideal feedforward control term
(w), which can be viewed as the output of the exosystem as
w = s(w)
= (w).
Suppose that there exists an immersion of the exosystem
= F + G (J )
= H ,
(10.83)
(10.84)
Pf F0 + F0T Pf = Q
Pf G + J T = 0
(10.85)
T Q 0 |1 |2 , 0 > 0, Rr
span(P K) span(Q).
F
Remark 10.15. It reminds us a challenging problem to design internal models for
output regulation with nonlinear exosystems. It is not clear at the moment what
general conditions are needed to guarantee the existence of an internal model for
output regulation of nonlinear systems with nonlinear internal models. Here we use
the condition of the existence of an immersion (10.83) for the internal model design.
Also note that even when the exosystem is linear, an internal model can be nonlinear
for a nonlinear dynamic system.
Remark 10.16. Note the condition specied in (10.85) is weaker than the condition
that there exist Pf > 0 and Q > 0 satisfying
PF F0 + F0T PF = Q
(10.86)
PF G + J T = 0,
which can be checked by LMI. This will be seen in the example later in this section. In
particular, if G and J T are two column vectors, (F0 , G) controllable, (J , F0 ) observable
and Re[J (jI F0 )1 G] > 0, R, then there exists a solution of (10.86) from
KalmanYacubovich lemma.
200
d2
1
+ b K z1 + e + (1 (y) 1 (q(w)))a .
d1
Let
F .
V = P
Then following the spirit of (10.82), there exist unknown positive real constants 1
and 2 such that
d2
T
T
1
V = Q + 2 PF b K z1 + e
d1
+ 2 T PF b1
K(1 (y) 1 (q(w)))a
+ 2 T PF G( (J ) (J ( b1
Ke))
3
12
0 | 1 |2 + b2
z 2 + 1 e2 + 2 e2 2 (1 + e2 ).
4
0 1
(10.87)
d2
e + (1 (y) 1 (q(w)))a + 1 + b (1 1 1 + b1
K1 e),
d1
where 1 = 1 1 and
1 = b1
1 .
(10.88)
(10.89)
1 2
e,
2
d2
e le(1 + 2 (1 + e2 )) + (1 (y) 1 (q(w)))a + b (1 1 ).
d1
and
2,
201
d2
V e = c0 e2 + ez1 + e2 + eb (1 1 )
d1
+ e(1 (y) 1 (q(w)))a le2 (1 + 2 (1 + e2 ))
1
1
c0 e2 + z12 + 0 12 + 1 e2 +
8
4
2
2
le (1 + (1 + e2 )) + b e1 .
2e
(1 + e2 )
(10.90)
Let
1
V0 = Vz + V + Ve + 1 (l l)2 ,
2
where 96
b2 is chosen and l =
0
unknown constant. Let
+ 1 + 2 + (1 + 2 ) is an
l = e2 (1 + 2 (1 + e2 )).
Then, it can be obtained that
1
1
V 0 z T z 0 | 1 |2 c0 e2 + b e1 .
2
2
If the system (10.78) has relative degree 1, the virtual control 1 shown in (10.88)
together with 1 in (10.88) gives the input, i.e., u = 1 . For the system with higher
relative degrees, the control design can be proceeded with backstepping using (10.50)
in the same way as the adaptive backstepping with ltered transformation shown in the
previous section for adaptive output regulation with linear exosystems. We summarise
the stability result in the following theorem.
Theorem 10.8. For the system (10.78) with the nonlinear exosystem (10.78), if there
exists an invariant manifold (10.80) and an immersion (10.83), then there exists
K Rr such that F0 = F KH is Hurwitz and there exist a positive denite matrix PF
and a semi-positive denite matrix Q satisfying (10.85), and there exists a controller
to solve the output regulation in the sense the regulated measurement converges to
zero asymptotically while other variables remain bounded.
We use an example to illustrate the proposed control design, concentrating on
the design of nonlinear internal model.
Example 10.2. Consider a rst-order system
y = 2y + sin y y3 sin w1 + w2 + u
e = y w1 ,
202
when |w1 | 1,
and that
q(w) = w1 ,
= w1 ,
(w) = w1 .
From the exosystem and the desired feedforward input , it can be seen that the
condition specied in (10.85) is satised with = w and
1 1
1 0
F = 1 0 , G =
0 1
1
0
.
1 (s) = 2 (s) = s , J =
0 1
Let K = [2, 0]T . Then with
1 1
F0 =
, PF = I ,
1 0
Q = diag(2, 0),
l = e2 (1 + (e2 + 1)2 ).
For simulation study, we set c = 1, = 1, = 1, and the initial states are
y(0) = 1, w1 (0) = 2 and w2 (0) = 2. The initial state of dynamic controller is zero.
The system output and input are shown in Figure 10.3, while the feedforward term
and its estimation are shown in Figure 10.4 and the portrait of the exosystem is
shown in Figure 10.5. As shown in the gures, the internal model successfully reproduces the feedforward control needed after a transient period, and the system output
measurement is regulated to zero, as required.
203
0.4
Tracking error e
0.2
0
0.2
0.4
0.6
0.8
1
10
15
10
15
Time t (s)
5
4
2
1
0
1
2
5
Time t (s)
0.5
0
h1 and h1
Control u
0.5
1.5
10
15
Time t (s)
204
1.5
w2
0.5
0.5
1
1
0.5
0.5
w1
1.5
205
(10.91)
y(y)
with being a continuous function; b is a known constant; u R is the input;
n
z R is the internal state variable; and f : Rn Rm R Rn is a continuous
function.
Remark 10.17. For the convenience of presentation, we only consider the system
with relative degree 1 as in (10.91). The systems with higher relative degrees
can be dealt with similarly by invoking backstepping. The second equation in
(10.91) describes the internal dynamics of the system states, and if we set v = 0
and y = 0, z = f (z, 0, 0) denotes the zero dynamics of this system.
Remark 10.18. The system in (10.91) species a kind of standard form for
asymptotic rejection of general periodic disturbances. For example, consider
x = Ax + (y, v) + bu
y = cT x,
206
with b, c Rn and
a1
a2
A=
.
..
an
... 0
..
. 0
0
.. . .
.. ,
. .
.
... ... 0
1
1
b1
0
b2
b = . , c = . ,
..
..
bn
where x Rn is the state vector; y and u R are the output and input respectively
of the system; v Rm denotes general periodic disturbances; and : R Rm Rn
is a nonlinear smooth vector eld in Rn with (0, 0) = 0. This system is similar to
the system (10.49) with q(w) = 0. For this class of nonlinear systems, the asymptotic
disturbance rejection depends on the existence of state transform to put the systems
in the form shown in (10.91), and it has been shown in Section 10.2 that such a
transformation exists under some mild assumptions.
The wave prole information of a general periodic disturbance is used to construct
a nonlinear function as the output function for a linear exosystem, for the generation
of the desired feedforward input. By doing that, an observer with nonlinear output
function can be designed, viewing the feedforward input as the output, and an internal
model can then be designed based on the nonlinear observer. The problem to solve in
this section is to design a control scheme, using a nonlinear observer-based internal
model, to asymptotically reject a class of general periodic disturbances for the system
in (10.91).
We start with modelling general periodic disturbances as the outputs of a linear
oscillator with nonlinear output functions, and then propose nonlinear observer design
for such a system, for the preparation of internal model design.
Many periodic functions with period T can be modelled as outputs of a secondorder system
0
w = Aw, with A =
0
(v) = h(w),
where = 2
. Here, the desired feedforward input (v) is modelled as the nonlinear
T
output h(w) of the second-order system. With
cos t sin t
eAt =
,
sin t cos t
the linear part of the output h(w), Hw, is always in the form of a sin (t + ) where
a and denote the amplitude and phase respectively. Hence, we can set H = [1 0]
without loss of generality, as the amplitude and the phase can be decided by the initial
value with
w(0) = [a sin () a cos ()]T .
207
Based on the above discussion, the dynamic model for general periodic disturbance
is described by
w 1 = w2
w 2 = w1
(10.92)
= w1 + h1 (w1 , w2 ),
where h1 (w1 , w2 ) is a Lipschitz nonlinear function with Lipschitz constant .
Remark 10.19. General periodic disturbances can be modelled as af (t + ) with a
and for the amplitude and phase of a disturbance, and the wave prole is specied by
a periodic function f . In the model shown in (10.93), the amplitude and phase of the
disturbance are determined by the system state variables w1 and w2 , and the prole is
determined by the nonlinear output function. In some results shown in literature, the
phase and amplitude are obtained by delay and half-period integral operations. Here,
we use nonlinear observers for the estimation of phases and amplitudes of general
periodic disturbances.
For the model shown in (10.93), the dynamics are linear, but the output function
is nonlinear. Many results in literature on observer design for nonlinear Lipschitz
systems are for the system with nonlinearities in the system dynamics while the output
functions are linear. Here we need the results for observer design with nonlinear output
functions. Similar techniques to the observer design of nonlinearities in dynamics can
be applied to the case when the output functions are nonlinear.
We have shown the observer design for a linear dynamic system with a nonlinear
Lipschitz output function in Section 8.4 with the observer format in (8.38) and gain
in Theorem 8.11. Now we can apply this result to observer design for the model of
general periodic disturbances. Forthe model shown in (10.93), the observer shown
0
in (8.38) can be applied with A =
and H = [1 0]. We have the following
0
lemma for the stability of this observer.
Lemma 10.9. An observer in the form of (8.38) can be designed to provide an exponentially convergent state estimate for the general periodic disturbance model (10.93)
if the Lipschitz constant for h1 satises < 12 .
Proof. Our proof is constructive. Let
1
p
4 2
,
P=
1
2
p
4
where p >
1
.
4 2
PA + AT P
HTH
1
= 2 I.
2
208
Therefore, the second condition in (8.40) is satised. Following the rst condition
specied in (8.39), we set
4(4p 2 )
(4p 2 )2 1
.
L=
(10.93)
4
(4p 2 )2 1
The rest part of the proof can be completed by invoking Theorem 8.11.
Hence from the above lemma, we design the observer for the general disturbance
model as
x = Ax + L(y h(x)),
(10.94)
0
where A =
and H = [1 0] with the observer gain L as shown in (10.93).
0
Before introducing the control design, we need to examine the stability issues of
the zsubsystem, and hence introduce a number of functions that are needed later
for the control design and stability analysis of the entire system.
Lemma 10.10. Assuming that the subsystem
z = f (z, v, y)
is ISS with state z and input y, characterised by an ISS pair (, ), and furthermore,
(s) = O(a2 (s)) as s 0, there exist a differentiable positive denite function V (z)
and a K function satisfying (z) a2 (z) such that
V (z) (z) + (y).
(10.95)
(10.96)
209
Dene
V (z) := (V (z)),
and it can be obtained that
V (z) q(V (z))(z) + q(V (z)) (|y|)
1
q(V (z))(z) + q( (|y|)) (|y|)
2
1
q(1 (z))(z) + q( (|y|)) (|y|)
2
1
= q (z)(z) + q((|y|)) (|y|),
2
where is dened as
(r) := 2 ( 1 (2 (r))
for r R+ . Let us dene a smooth function such that
(r) q( (|r|)) (|r|)
for r R and (0) = 0, and then we have established (10.95).
(10.97)
(y)
u = b1 0 (y) + k0 y + k1 y + k2
+ h( b1 Ly), (10.98)
+ k3 (y)
y
where k0 is a positive real constant, and
3
k1 = 1 b2 ( + H )2 + ,
4
k2 = 4 1 ||b1 PL2 + 2,
1
k3 = 4 1 ||b1 PL2 + .
2
For the stability of the closed-loop system, we have the following theorem.
Theorem 10.11. For a system in the form shown in (10.91), if
feedforward term (v) can be modelled as the output of a system in the format
shown in (10.93) and the Lipschitz constant of the output nonlinear function
satises < 12
210
the output feedback control design with the internal model (10.97) and the control
input (10.98) ensures the boundedness of all the variables of the closed-loop system
and the asymptotic convergence to zero of the state variables z and y and the estimation
error (w + b1 Ly).
Proof. Let
= w + b1 Ly.
It can be obtained from (10.97) that
= (A LH ) + b1 L(h1 (w) h1 (w )) + b1 La(z) + b1 L(y, v).
Let Vw = T P . It can be obtained that
V w ( ) 2 + 2| T b1 PLa(z)| + 2| T b1 PL(y, v)|
1
2 + 2 1 ||b1 PL2 (a2 (z) + (y, v)|2 )
2
1
1
2
y(y)
+ (k2 2)(z) + k3
2
2
(10.99)
where = 21 2 1.
Based on the control input (10.98), we have
y = k0 y k1 y k2
(y)
k3 (y)
+ a(z) + (y, v) + b(h(w ) h(w)).
y
1
1
2
k0 y k2 (y) k3
y(y)
+ (z) + 2 . (10.100)
2
4
Let us dene a Lyapunov function candidate for the entire closed-loop system as
V = Vy + Vw + k2 V z .
Following the results shown in (10.96), (10.99) and (10.100), we have
1
V k0 y2 2 (z).
4
Therefore, we can conclude that closed-loop system is asymptotically stable with
respect to the state variables y, z and the estimation error .
Several types of disturbance rejection and output regulation problems can be
converted to the form (10.91). In this section, we show two examples. The rst
example deals with rejection of general periodic disturbances, and the second example
demonstrates how the proposed method can be used for output regulation.
211
(10.101)
b2
x1 .
b1
b2
y + 1 (y) + b1 u
b1
b2
b2
z = z + 2 (y) 1 (y)
b1
b1
b2
b1
2
y + (w).
Consider
z =
b2
z + (w).
b1
It can be shown that there exists a steady-state solution, and furthermore, we can
express the solution as a nonlinear function of w, denoted by z (w). Let us introduce
another state transformation with z = z z (w). We then have
b2
y + 1 (y) + b1 (u + b1
1 z (w))
b1
2
b2
b2
b2
z = z + 2 (y) 1 (y)
y.
b1
b1
b1
y = z +
(10.102)
212
(y) =
h(w) = b1
1 z (w),
f (z, v, y) = 2 (y)
b2
1 (y)
b1
b2
b1
2
y.
b1
b2
b2
2 (y) 1 (y)
b1
b2
b1
2 2
y .
(10.103)
It can be seen that there exists a class K function (|y|) to dominate the second term
on the right-hand side of (10.103), and the z-subsystem is ISS. For the control design,
we can take
2 2
2
b2
b2
b1
2 (y) 1 (y)
y .
(y) =
b2
b1
b1
The rest part of the control design follows the steps shown earlier.
For the simulation study, we set the periodic disturbance as a square wave. For
convenience, we abuse the notations of (w(t)) and h(w(t)) as (t) and h(t). For
with t in one period, we have
d, 0 t < ,
2
(10.104)
=
T
d,
t < T,
2
213
h = d h(t),
where
b
1
T
1 bb2 t
T b2
b2 t
1
1
, 0t< ,
)+ e
tanh
b (1 e
b2
4 b1
2
2
=
h(t)
(10.105)
b
b
b
T
2
2
2
1
T
1 t
T b2
t
( t)
,
t < T.
(1 + e b1 2e b1 2 ) + e b1 tanh
b2
b2
4 b1
2
Eventually we have the matched periodic disturbance h(w) given by
w2
2
2
h(w) = w1 + w2 h arctan
.
w1
Note that w12 + w22 decides the amplitude, which can be determined by the initial
state of w.
In the simulation study, we set T = 1, d = 10, 1 = y3 , 2 = y2 and b1 = b2 = 1.
The simulation results are shown in Figures 10.610.9. It can be seen from Figure 10.6
that the measurement output converges to zero and the control input converges to a
periodic function. In fact, the control input converges to h(w) as shown in Figure 10.7.
0.2
0.15
0.1
0.05
0
0.05
10
15
10
15
Time (s)
5
0
5
10
15
20
0
5
Time (s)
214
u and h
10
15
20
10
15
Time (s)
8
h
Estimate of h
6
h and estimate
4
0
10
Time (s)
15
215
40
w1
w2
h1
h2
30
w and estimates
20
10
10
20
10
15
Time (s)
Figure 10.9 The exosystem states and the internal model states
As for the internal model and state estimation, it is clear from Figure 10.8 that the
estimated equivalent input disturbance converges to h(w), and converges to w.
Example 10.4. In this example, we briey show that an output regulation problem
can also be converted to the form in (10.91). Consider
x 1 = x2 + (ey 1) + u
x 2 = (ey 1) + 2w1 + u
w = Aw
y = x1 w1 ,
(10.106)
where y R is the measurement output and w1 = [1 0]w. In this example, the measured output contains the unknown disturbance, unlike Example 10.3. The control
objective remains the same, to design an output feedback control law to ensure the
overall stability of the system and the convergence to zero of the measured output.
The key step in the control design is to show that the system shown in (10.106) can
be converted to the form as shown in (10.91).
Let
z =
1
[1 ]w,
1 + 2
216
2 + 2
[1 ]w (ew1 1).
1 + 2
It can be seen that we have transformed the system to the format as shown in
(10.91) with (y, v) = ew1 (ey 1).
To make H = [1 0], we introduce a state transform for the disturbance model as
2 + 2
=
1 + 2
w.
1
1
2 + 2
1
.
1
(10.107)
where
h( ) = 1 (e1/(2+
2 )[1]
1).
y
y
0.1
0.05
0
0.05
0
0.5
1.5
2.5
Time (s)
3.5
4.5
0.5
1.5
2.5
Time (s)
3.5
4.5
50
0
50
u
100
150
200
250
300
50
u
h
0
u and h
50
100
150
200
250
300
0
0.5
1.5
2.5
3.5
4.5
Time (s)
217
h and estimate
5
0
5
10
15
20
25
0.5
1.5
2.5
Time (s)
3.5
4.5
30
20
m and estimates
218
10
10
20
30
0.5
1.5
2.5
3.5
4.5
Time (s)
Figure 10.13 The exosystem states and the internal model states
Chapter 11
Control applications
In this chapter, we will address a few issues about control applications. Several
methods of disturbance rejection are presented in Chapter 10, including rejection
of general periodic disturbances. A potential application can be the estimation and
rejection of undesirable harmonics in power systems. Harmonics, often referred to
high-order harmonics in power systems, are caused by nonlinearities in power systems,
and the successful rejection depends on accurate estimation of amplitudes and phase
of harmonics. We will show an iterative estimation method based on a new observer
design method.
There are tremendous nonlinearities in biological systems, and there have been
some signicant applications of nonlinear system analysis and control methods in
system biology. We will show a case that nonlinear observer and control are applied
to circadian rhythms. A Lipschitz observer is used to estimate unknown states, and
backstepping control design is then applied to restore circadian rhythms.
Most of the control systems are implemented in computers or other digital devices
which are in discrete-time in nature. Control implementation using digital devices
inevitably ends with sample-data control. For linear systems, the sampled systems are
still linear, and the stability of the sampled-date system can be resolved in stability
analysis using standard tools of linear systems in discrete-time. However, when a
nonlinear system is sampled, the system description may not have a closed form, and
the structure cannot be preserved. The stability cannot be assumed for a sampled-data
implementation of nonlinear control strategy. We will show that for certain nonlinear
control schemes, the stability can be preserved by fast sampling in the last section.
220
0 1 0 ...
0 0 1 ...
0 0 0 ...
0 0 0 ...
0
0
.. ,
.
1
0
0
..
.
0
b=
b ,
.
..
bn
T
1
0
C=. ,
..
0
d1
d = ... ,
dn
Control applications
221
(11.2)
222
NT +t
eA bw( )d
N 1
eA bw( )d
iT
i=0
NT +t
(i+1)T
eA bw( )d ,
NT
(i+1)T
eA bw( )d =
iT
(i+1)T
eA bw( iT )d
iT
=
eA(iT + ) bw( )d
= eiAT
eA bw( )d.
Therefore, we have
x(NT + t) = eA(NT +t) x(0) + eAt
+
N 1
eA(N i)T
eA bw( )d
i=0
t
(11.4)
The steady-state response in (11.3) is obtained by taking the limit of (11.4) for
t .
2
To obtain the equivalent input disturbance, we need to introduce state transformation to (11.1). To extract the zero dynamics, we introduce a partial state
transformation for system (11.1) as
+1
B bi ,
z = ...
n
i=1
where
b+1 /b
..
.
B=
bn1 /b
bn /b
... 0
.
..
. ..
,
... 1
... 0
1
..
.
b+1 /b
..
b =
.
.
bn /b
Control applications
223
i = 1, . . . , 1
= z1 +
ri i + (y) + d w + b u
(11.5)
i=1
z = Bz + z (y) + dz w,
where
i , for i = 1, . . . , ,
ri = (Bi b)
+1
z (y) = ...
B bi + B by,
n
and
i=1
d+1
i
dz = ...
B bdi .
dn
i=1
The periodic trajectory and the equivalent input disturbance can be found using
the system in the coordinate (1 , . . . , , z). Since the system output y does not
contain the periodic disturbance, we have the invariant manifold for 1 = 0. From
Lemma 11.1 which is used for the result of the steady-state response of linear systems
to the periodic input, we have, for 0 t < T ,
z (t) =
T
with WT = 0 eB dz w( )d . From the rst equation of (11.6), we have, for
i = 1, . . . , 1
i+1 (t) =
di (t)
di w.
dt
Based on the state transformation introduced earlier, we can use its inverse
transformation to obtain
+1
..
i.
Bi b
. = z +
n
i=1
(11.6)
224
d (t)
z,1
ri z,i d w .
dt
i=1
(11.7)
Let x = denote the difference between the state variable and the periodic
trajectory.
The periodic trajectory, , plays a similar role as the invariant manifold in the
set-up for the rejection of disturbances generated from linear exosystems. For this,
we have the following result.
Theorem 11.2. For the general periodic disturbance w in (11.1), the periodic trajectory given in (11.6) and the equivalent input disturbance given in (11.7) are well
dened and continuous, and the difference between the state variable (11.1) and the
periodic trajectory, denoted by x = , satises the following equation:
x = Ac x + (y) + b(u )
y = Cx.
(11.8)
The control design and disturbance rejection will be based on (11.8) instead
of (11.1).
Remark 11.4. The control design and disturbance rejection only use the output y,
with no reference to any of other state of the system. Therefore, there is no difference
whether we refer to (11.1) or (11.8) for the system, because they have the same output.
The format in (11.8) shows that there exist an invariant manifold and an equivalent
input disturbance. However, the proposed control design does not depend on any
information of , other than its period, which is the same as the period of w. In other
words, control design only relies on the form shown in (11.1). The form shown in
(11.8) is useful for the analysis of the performance of the proposed control design,
including the stability. In this section, we start our presentation from (11.1) rather
than (11.8) in order to clearly indicate the class of the systems to which the proposed
control design can be applied, without the restriction to the rejection of matched
disturbances.
Control applications
225
0
T
f ( ) cos k d = 0.
(11.10)
lim
t+T
f ( ) cos k d = 0.
(11.12)
(11.13)
where x Rn is the state variable; and y and u R are output and input, the matrices
A, b and C are with proper dimensions; and the system is stable and observable with
the transfer function Q(s) = C(sI A)1 b. The problem considered in this subsection
is to create a signal such that specic frequency modes can then be removed from the
input , which is a general periodic disturbance described by
(t) =
ak sin (k t + k ),
k=1
and ak and k are the amplitude and phase angle of the mode for
where k = 2k
T
frequency k . The dynamics for a single frequency mode can be described as
w k = Sk wk ,
where
Sk =
0
k
(11.14)
k
0
.
For a single frequency mode with the frequency k as the input denoted by
k = ak sin (k t + k ), its output in y, denoted by yk , is a sinusoidal function with
the same frequency, if we only consider the steady-state response. In fact, based on
the frequency response, we have the following result.
Lemma 11.3. Consider a stable linear system (A, b, C) with no zero at jk for any
integer k. For the output yk of a single frequency mode with the frequency k , there
exists an initial state wk (0) such that
yk = g T wk ,
(11.15)
226
where g = [1 0]T and wk is the state of (11.14). Furthermore, the input k for this
frequency mode can be expressed by
k = gkT wk ,
where
1
gk =
mk
(11.16)
cos k
sin k
sin k cos k
g := Qk g
(11.17)
wk (t) = eSk t wk (0) =
mk ak sin (k t + k + k )
mk ak cos (k t + k + k )
such that
yk = g T eSk t wk (0) = g T wk (t).
Considering the gain and phase shift of Q(s), we have
1 T Sk (t(k /k ))
g e
wk (0)
mk
1 T (k /k )Sk Sk t
=
g e
e wk (0).
mk
k =
Hence, we have
gk =
1 (k /k )S T
k g
e
mk
and therefore
1 (k /k )S T
1 (k /k )Sk
1
k =
e
e
=
Qk =
mk
mk
mk
cos k
sin k
sin k
cos k
.
Control applications
227
For a stable single-input linear system, if the input is a T -periodic signal that is
orthogonal to a frequency k , the steady state, as shown earlier, is also T -periodic.
Furthermore, we have the following results.
Lemma 11.4. If the input to a stable single-input linear system (A, b) is T -periodic
signal that is orthogonal to a frequency mode k , for any positive integer k, the steady
state is orthogonal to the frequency mode and the state variable is asymptotically
orthogonal to the frequency mode. Furthermore, if the linear system (A, b, C) has
no zero at jk , the steady-state output is orthogonal to the frequency mode k if and
only if the input to the system is orthogonal to the frequency mode.
Proof. We denote the input as , and the state variable x satises
x = Ax + b.
Since is T -periodic, the steady-state solution of the above state equation, denoted
by xs , is also T -periodic and
x s = Axs + b.
Let
Jk =
xs ( ) sin k d.
k1 A
= k2 A
x s cos k d
0
T
xs cos k d
xs d sin k
= k2 A
= k2 A2k Jk .
Hence, we have
(k2 I + A2 )Jk = 0.
228
xs ( ) cos k d = 0
(11.18)
where lk is chosen such that Sk lk g T is Hurwitz. For this observer, we have a useful
result stated in the following lemma.
Lemma 11.5. For any positive integer k, with the observer as designed in (11.18),
(( ) gkT w k ) is asymptotically orthogonal to the frequency mode k , i.e.,
t+T
lim
lim
t+T
(( ) gkT w k ) sin k d = 0,
(11.19)
(( ) gkT w k ) cos k d = 0.
(11.20)
Control applications
229
Proof. Consider the steady-state output y of input . From Lemma 11.3, there exists
an initial state wk (0) such that
T
(y( ) g T wk ( )) sin k d = 0,
0
T
(y( ) g T wk ( )) cos k d = 0,
which implies that gkT wk is orthogonal to the frequency mode k , again based on
Lemma 11.3.
Let w k = wk w k . The dynamics of w k can be obtained from (11.14) and
(11.18) as
w k = S k w k lk (y g T wk ),
(11.21)
The result in Lemma 11.5 shows how an individual frequency mode can be
removed with the observer designed in the way as if the output would not contain
other frequency modes. From the proof of Lemma 11.5, it can be seen that there
is an asymptotic error, k , between the observer state and the actual state variables
associated with the frequency mode k . Although k is orthogonal to the frequency
mode k , it does in general contain components generated from all the other frequency
modes. Because of this, a set of observers of the same form as shown in (11.18) would
not be able to extract multiple frequency modes simultaneously. To remove multiple
frequency modes, it is essential to nd an estimate which is asymptotically orthogonal
to the multiple frequency modes. For this, the interactions between the observers must
be dealt with.
Suppose that we need to remove a number of frequency modes k for all the k
in a nite set of positive integers K = {ki }, for = 1, . . . , m. To estimate the frequency
modes for k,i , i = 1, . . . , m, we propose a sequence of observers,
w k,1 = Sk,1 w k,1 + lk,1 (y g T w k,1 )
(11.22)
230
and, for i = 2, . . . , m,
T
w k,i1
k,i1 = Ak,i1 + bgk,i1
i1
Ck,j g T w k,i ,
w k,i = Sk,i w k,i + lk,i y
(11.23)
(11.24)
j=1
where lk,i , for i = 1, . . . , m, are designed such that S k,i := Sk,i lk,i g T are Hurwitz,
and
1
cos k,i sin k,i
g := Qk,i g
gk,i =
mk,i sin k,i cos k,i
with mk,i = |C(jwk,i A)1 b| and k,i = C(jwk,i A)1 b.
The estimate for the input disturbance which contains the required frequency
modes for asymptotic rejection is given by
m =
T
gk,i
w k,i .
(11.25)
i=1
The estimate m contains all the frequency modes k,i , for i = 1, . . . , m. The useful
property of the estimate is given in the following theorem.
Theorem 11.6. For the estimate m given in (11.25), m is asymptotically
orthogonal to the frequency modes k,i for i = 1, . . . , m.
Proof. In the proof, we will show how to establish the asymptotic orthogonality in
detail by induction.
We introduce the notations w k,i = wk,i w k,i . We use k,i to denote the steadystate solutions of w k,i and ek,i = w k,i k,i , for i = 1, . . . , m.
Lemma 11.5 shows that the results hold for m = 1. Let
T
(wk,1 k,1 )
1 = gk,1
(11.26)
such that y Cqk,1 g wk,2 is orthogonal to the frequency mode k,2 where qk,i , for
i = 1, . . . , m 1, denote the steady-state solution of
T
T
(wk,i k,i ).
q k,i = Aqk,i + bgk,i
T
wk,2 to the system
Note that g T wk,2 can be viewed as the output for the input gk,2
T
(A, b, C), based on Lemma 11.3. Hence, y Cqk,1 g wk,2 is the output for the input
T
T
gk,1
(wk,1 k,1 ) gk,2
wk,2 to the system (A, b, C). Therefore, from Lemma 11.4,
T
T
gk,1 (wk,1 k,1 ) gk,2
wk,2 is orthogonal to k,2 . Furthermore, from the previous
Control applications
231
T
T
(wk,1 k,1 ) gk,2
wk,2 is orthogonal
step k,1 is orthogonal to k,1 . Hence, gk,1
to the frequency modes k,j for j = 1, 2.
The dynamics of w k,2 and k,2 are obtained as
(11.27)
(11.28)
T
gk,j
(wk,j k,j w k,j )
j=1
T
gk,1
ek,j .
j=1
T
gk,j
(wk,j k,j ).
j=1
232
lk,i+1 y C
k,j g T wk,i+1 ,
(11.29)
j=1
lk,i+1 y C
qk,j g T wk,i+1 .
(11.30)
j=1
(qk,j k,j ).
j=1
Since qk,j k,j , for j = 1, . . . , i, exponentially converge to zero, so does ek,i+1 . With
i+1 i+1 =
i+1
T
gk,j
(wk,j k,j w k,j )
j=1
T
gk,j
ek,j ,
j=1
(11.31)
Control applications
233
If the disturbance does not exist in (11.31), the system (11.31) is in the linear
observer error with output injection that is shown in Chapter 8. In that case, we can
design a state observer as
p = (Ac LC)p + (y) + bu + Ly,
(11.32)
(11.33)
where q denotes the steady-state solution. Such a solution exists, and an explicit
solution is given earlier. Each element of q is a periodic function, as is periodic.
We have an important property for p and q stated in the following lemma.
Lemma 11.7. The state variable x can be expressed as
x = p q + ,
(11.34)
(11.35)
(11.36)
and, for i = 2, . . . , m,
T
k,i1 = (Ac kC)k,i1 + bgk,i1
w k,i1 ,
i1
Ck,j g T w k,i ,
w k,i = Sk,i w k,i + lk,i Cp y
(11.37)
(11.38)
j=1
where lk,i , for i = 1, . . . , m, are designed such that S k,i := Sk,i lk,i g T are Hurwitz,
and
cos k,i sin k,i
1
gk,i =
g := Qk,i g
(11.39)
mk,i sin k,i cos k,i
with mk,i = |C(jwk,i (A kC))1 b| and k,i = C(jwk,i (A kC))1 b.
234
The estimate for the input disturbance which contains the required frequency
modes for asymptotic rejection is given by
m
m =
T
gk,i
w k,i .
(11.40)
i=1
The estimate m contains all the frequency modes k,i , for i = 1, . . . , m. The useful
property of the estimate is given in the following theorem.
Theorem 11.8. The estimate m given in (11.40) is bounded and satises the
following:
t+T
[( ) m ( )] sin k,i d = 0,
(11.41)
lim
t
lim
t+T
[( ) m ( )] cos k,i d = 0,
(11.42)
for i = 1, . . . , m.
Proof. From Theorem 11.6, the results in (11.41) and (11.42) hold if the input
Cp y in (11.36), (11.37) and (11.38) equals Cq. Since Cp y Cq converges
exponentially to 0, and (11.36)(11.38) are stable linear systems, the errors in
the estimation of caused by replacing Cq by Cp y in (11.36)(11.38) are also
exponentially convergent to zero. Therefore, (11.41) and (11.42) can be established
from Theorem 11.6.
2
(11.43)
For the stability of the closed-loop system with disturbance rejection, we have the
following theorem.
Control applications
235
Theorem 11.9. The closed-loop system of (11.1) under the control input (11.43)
ensures the boundedness of all the state variables, and the disturbance modes of
specied frequencies k,i , for i = 1, . . . , m, are asymptotically rejected from the
system in the sense that all the variables of the closed-loop system is bounded and
the system is asymptotically driven by the frequency modes other than the specied
modes.
Proof. For the standard backstepping design for the nonlinear systems, we establish
the stability by considering the Lyapunov function
1
1
1 2
,
V = z T Pz + y2 +
2
2 i=1 i
(11.44)
where is a constant and i = i i and P is a positive denite matrix satisfying DT P + PD = I . The standard backstepping design ensures that there exists a
positive real constant 1 such that
V 1 V [2z T Pbz + yby ]
1 V 2 V
3 V + 4 2
(11.45)
11.1.5 Example
In this section, we use a simple example to demonstrate estimation of harmonics using
the method proposed earlier in this section.
We consider a voltage signal with diados as shown in Figure 11.1. We are
going to use iterative observers to estimate harmonics in this signal. To simplify
the presentation, we assume that the signal is directly measurable. In this case,
there are no dynamics between the measurement and the point of estimation.
Hence, we can simplify the algorithms presented in Subsection 11.1.2 ((11.22),
236
Voltage
0.5
0.5
1
0.7
0.72
0.74
0.76
0.78
0.8
t (s)
(11.23) and (11.24)). In this case, there is no need to use the observer model in
(11.23). The simplied algorithm for direct estimation of harmonics is presented
below.
To estimate the frequency modes for k,i , i = 1, . . . , m, the iterative observers
are given by
w k,1 = Sk,1 w k,1 + lk,1 (y g T w k,1 )
(11.46)
and, for i = 2, . . . , m,
g T w k,j ,
(11.47)
j=1
where lk,i , for i = 1, . . . , m, are designed such that S k,i := Sk,i lk,i g T are Hurwitz.
For the electricity, the base frequency is 50 Hz. We will show the estimates
of harmonics from the base frequency, that is, we set k,1 = 100 , k,2 = 200 ,
k,3 = 300, etc. For these frequency modes, we have
0
i100
.
Sk,i =
i100
0
Control applications
237
The observer gain lk,i can be easily designed to place the close-loop poles at any
specied positions. For the simulation study, we place the poles at {200, 200}. For
such pole positions, we have
lk,i =
400
40000
200
i100
.
The harmonics of second and third orders are plotted with the original signal and
the estimated base frequency component in Figure 11.2. The approximations of
accumulative harmonics to the original signal are shown in Figure 11.3.
In this example, we have shown the estimation of harmonics from direct
measurements. The advantage of this estimation is that the estimation is on-line
and can be implemented in real time. Also, the estimation using the proposed
method can provide the phase information, which is very important for harmonic
rejection. We did not show the rejection of harmonics in this example. However,
it is not difcult to see that rejection can be simulated easily with the proposed
method.
1.5
Original
1st order
2nd order
3rd order
0.5
0.5
1
0.74
0.745
0.75
0.755
0.76
0.765
t (s)
0.77
0.775
0.78
238
0.5
0.5
1
0.74
0.745
0.75
0.755
0.76
t (s)
0.765
0.77
0.775
0.78
Control applications
239
x 1 = vs
(11.48)
240
6
x1
x2
x3
0
0
10
20
30
40
50
60
70
80
90
100
Time (h)
110
120
Control applications
241
The model (11.48) is not in the right format for control and observer design. We
introduce a state transformation as
z = Tx
with
0
T = 0
k1 ks
(11.49)
1
0,
0
0
k1
0
z2 = z3 k1 z2 + k1 k2 z1 vd
(11.50)
(11.51)
provided that the conditions specied in Theorem 8.10 are satised. That suggests
that we need nd the Lipschitz constant together with a positive denite matrix P
for nonlinearities in the circadian model (11.48) such that the inequality (8.36)
AT P + PA + 2I 2 C T C < 0
is satised, and then we obtain the observer gain as
L = P 1 C T .
The one-sided Lipschitz condition shown in (8.36) can be checked using the
Lipschitz constant for individual elements in the vector . Let us denote the Lipschitz
constants of i by i . The condition shown in (8.36) can be guaranteed by
AT P + PA + 2n
n
i=1
i i I 2 C T C < 0,
(11.52)
242
where i are positive real constants, and P ni=1 P n (i, i ). The denition of
P ni=1 P n (i, i ) is given by
P n (i, i ) = {P : |pji | < i , for j = 1, 2, . . . , n}.
For the system (11.50), we set
C1 = 1 0 0
and therefore state variables z2 and z3 are unknown. Comparing with the structure in
(8.29), we have
0
,
2
(z, u) =
3a + 3b
where clearly 1 = 0, and
2 = vd
3a = vs
k 1 z2
,
k1 K d + z 2
k1 ks Kin
,
Kin + z1n
3b = vm
k 1 k s z3
.
K M k1 k s + z 3
Control applications
243
3.9887
= 67.8032, L = 2.6067
1.3469
for the observer (11.51). Simulation study for the observer has been carried out with
the results shown in Figures 11.5 and 11.6.
4.5
z2
Estimate of z2
4
3.5
3
2.5
2
1.5
1
0.5
0
0
10
20
30
40
50
60
70
Time (h)
80
90
100
110
120
244
1.6
1.4
1.2
1
0.8
0.6
0.4
0.2
0
10
20
30
40
50
60
70
Time (h)
80
90
100
110
120
where u is the control input. The system model after the state transformation (11.49)
is then obtained as
z1 = z2 k2 z1
k 1 z2
k1 K d + z 2
k1 ks Kin
k1 ks z 3
k1 ks Kin
z3 = vs n
v
+
u
.
m
Ki + z1n
K M k1 k s + z 3
Kin + z1n
z2 = z3 k1 z2 + k1 k2 z1 vd
(11.53)
This model (11.53) is then used for phase control design. We use q to denote the state
variable for the target circadian model
q 1 = q2 k2 q1
q 2 = q3 k1 q2 + k1 k2 q1 vd
q 3 = vs
k1 q2
k1 K d + q 2
k1 ks Kin
k1 k s q 3
vm
,
Kin + q1n
K M k1 ks + q 3
(11.54)
Control applications
245
(11.55)
e3 = z3 q3 ,
where ei , for i = 1, 2 and 3, are the tracking errors. The error dynamics are obtained as
e 1 = e2 k2 e1
e 2 = e3 k1 e2 + k1 k2 e1
e 3 = vs
vd k12 Kd e2
(k1 Kd + z2 ) (k1 Kd + q2 )
k1 ks Kin
k1 ks Kin
vm KM (k1 ks )2 e3
k1 ks Kin
.
+
u
s
Kin + z1n
Kin + q1n
Kin + z1n
(KM k1 ks + z3 ) (KM k1 ks + q3 )
(11.56)
It is easy to see that the model (11.56) is still in the triangular form. The phase resetting
is achieved if we can ensure that the errors ei , i = 1, 2, 3, converge to zero. To the
model (11.56), iterative backstepping can be applied. Following the procedures shown
in Section 9.2, we dene
w1 = e1 ,
w2 = e2 1 ,
(11.57)
w 3 = e3 2 ,
where 1 and 2 are stabilising functions to be designed.
From the rst equation of (11.56), we obtain dynamics of w1 as
w 1 = e2 k2 e1
= w 2 + 1 k 2 e1 .
The stabilising function 1 is then designed as
1 = c1 w1 + k2 e1 ,
(11.58)
(11.59)
vd k12 Kd e2
1
e 1
e1
(k1 Kd + z2 ) (k1 Kd + q2 )
= w 3 + 2 k 1 e2 + k 1 k 2 e 1
vd k12 Kd e2
1
e 1 .
e1
(k1 Kd + z2 ) (k1 Kd + q2 )
k1 K d e 2
1
e 1
+
e1
(k1 Kd + z2 ) (k1 Kd + q2 )
246
(11.60)
k1 ks Kin
k1 ks Kin
k1 ks Kin
v
+
u
s
Kin + z1n
Kin + q1n
Kin + z1n
vm KM (k1 ks )2 e3
2
2
e 1
e 2 ,
k
k
+
z
k
k
+
q
e
e2
(KM 1 s
3 ) (KM 1 s
3)
1
c
w
v
+
v
+
u=
(e2 k2 e1 )
2
3
3
s
s
n
n
n
n
n
k1 k s K i
Ki + z1
Ki + q1
e1
vd k12 Kd e2
2
2
(e3 k1 e2 + k1 k2 e1 )
+
,
(11.61)
e2
e2 (k1 Kd + z2 ) (k1 Kd + q2 )
with c3 as a positive real constant. The resultant dynamics of w3 are then obtained as
w 3 = w2 c3 w3 .
(11.62)
The stability analysis of the proposed control design can be established in the
same way as the proof shown in Section 9.2 for iterative backstepping control design.
Indeed, consider a Lyapunov function candidate
V =
1 2
w1 + w22 + w32 .
2
(11.63)
Control applications
247
4.5
q2
z2
4
Control input applied
3.5
q2 and z2
3
2.5
2
1.5
1
0.5
0
0
10
20
30
40
50
60
70
Time (h)
80
90
100
110
120
1.8
q3
z3
1.6
Control input applied
1.4
q3 and z3
1.2
1
0.8
0.6
0.4
0.2
0
10
20
30
40
50
60
70
Time (h)
80
90
100
110
120
248
Control input
5
10
15
20
25
30
0
10
20
30
40
50
60
70
Time (h)
80
90
100
110
120
Control applications
249
period and initial condition is established, which suggests that for a given domain
of initial conditions, there always exists a sampling time such that if the sampling is
faster than that time, the stability of the sampled-data system is guaranteed.
(11.64)
y = Cx,
with
0 1 0 ...
0 0 1 ...
0 0 0 ...
0 0 0 ...
0
0
.. ,
.
1
0
0
..
.
0
b=
b ,
.
..
bn
T
1
0
C=. ,
..
0
(11.65)
u = uc (y, ),
(11.66)
250
where
1
0
= .
..
1
2
..
.
0 ...
0
1 ...
0
..
..
.. ,
.
.
.
0 . . . 1
0
..
bf = . .
0
1
t [mT , mT + T ),
(11.67)
T
(11.68)
where y(mT ) is obtained by sampling y(t) at each sampling instant; (mT ) is the
discrete-time implementation of the lter shown in (11.65); T is the xed sampling
period; and m is the discrete-time index, starting from 0.
Before the analysis of the sampled-data control, we briey review the control
design in continuous-time.
As shown in Section 9.4, through the state transformations, we obtain
= D + (y)
y = 1 + y (y) + b 1 .
(11.69)
For the system with relative degree = 1, as shown in Lemma 9.5 and its proof,
the continuous-time control uc1 can be designed as
1
P
2
(y)
2
y y (y)
,
(11.70)
uc1 = c0 +
4
y
where c0 is a positive real constant and P is a positive real constant that satises
DT P + PD = 3I .
For the control (11.70), the stability of the continuous-time system can be established
with the Lyapunov function
1
V = T P + y2
2
and its derivative
V c0 y2
2 .
For the case of > 1, backstepping is used to obtain the nal control input uc2 .
We introduce the same notations zi , for i = 1, . . . , , as in Section 9.4 for backstepping
z1 = y,
zi = i1 i1 ,
z+1 = u ,
(11.71)
for i = 2, . . . , ,
(11.72)
(11.73)
Control applications
251
The control design has been shown in Section 9.4. We list a few key steps here
for the convenience of the stability analysis. The stabilising functions are designed as
1 = c1 z1 k1 z1 + y (y)
P
2
(y)
2
,
y
1 2
1
2 = z1 c2 z2 k2
z2 +
(y) + 1 1
y
y
i1 2
i1
i = zi1 ci zi ki
zi +
(y) + i1 i1
y
y
i1
+ i2
(j j + j+1 ) for i = 3, . . . , .
j=1
j
(11.74)
i2
1
j=1
1
y
2
z +
1
(y)
y
(j j + j+1 ) + 1 1 .
(11.75)
For the control input (11.75) in the continuous-time, the stability result has been
shown in Theorem 9.6. In the stability analysis, the Lyapunov function candidate is
chosen as
V =
zi2 + T P
i=1
ci zi2 2 .
(11.76)
i=1
t (mT , (m + 1)T ],
(11.77)
252
hold for all (mT ) D, where , are any given positive reals with > , T > 0 the
xed sampling period and Vm := V ( (mT )). If (0) D, then the following holds:
lim (t) = 0.
Proof. Since (0) D, then (11.77) holds for t (0, T ] with the following form:
V V + V ( (0)).
Using the comparison lemma (Lemma 4.5), it is easy to obtain from the above that
for t (0, T ]
V ( (t)) et V0 +
1 et
V0 = q(t)V0 ,
(11.78)
where q(t) := (et + (1 et )). Since > > 0, then q(t) (0, 1), t
(0, T ]. Then we have
V ((t)) < V0 ,
t (0, T ].
(11.79)
(11.80)
t (mT , (m + 1)T ].
(11.81)
(11.82)
indicating that V decreases at two consecutive sampling points with a xed ratio.
From (11.82),
Vm q(T )Vm1 qm (T )V0 ,
(11.83)
which implies that limm Vm = 0. The conclusion then follows from (11.81), which
completes the proof.
2
Remark 11.6. Lemma 11.10 plays an important role in stability analysis for the
sampled-data system considered in this section. When sampled-data systems are
analysed in discrete-time in literature, only the performances at sampling instances are
considered. In this section, Lemma 11.10 provides an integrated analysis framework
where the behaviour of the sampled-data system both at and between sampling instants
can be characterised. In particular, the systems behaviour at sampling instants is
portrayed in (11.82), and (11.81) shows that the inter-sample behaviour of the system
is well bounded by a compact set dened by Vm .
Control applications
253
For the case of relative degree 1, the sampled-data system takes the following
form:
= D + (y)
(11.84)
y = 1 + y (y) + ud1 ,
where ud1 is the sampled-data controller and can be simply implemented via a zeroorder hold device as the following:
ud1 (t) = uc1 (y(m)),
t [mT , mT + T ).
(11.85)
t [0, T ], (0) c .
(11.86)
254
y(t) = y(0) +
+
ud1 ( )d
0
t
1 ( )d +
0
y (y) y (y(0)) d +
y (y(0)) d.
Then we have
|y(t) y(0)|
( )
d +
0
!
+
Lu1 |y(0)|d
1
L1 |y( ) y(0)|d +
L1 |y(0)|d.
(11.87)
We rst calculate the integral 1 . From the rst equation of system (11.69), we obtain
t
Dt
(t) = e (0) +
eDt (y( ))d.
(11.88)
0
Since D is a Hurwitz matrix, there exist positive reals 1 and 1 such that
eDt
1 e1 t . Thus, from (11.88)
t
1 t
(t)
1 e
(0)
+
1 e1 (t )
(y( )) (y(0))
d
+
0
t
1 e1 (t ) (y(0)) d
1 e
1 t
(0)
+ L2
+ L2
1 e1 (t ) |y( ) y(0)|d
0
t
1 e1 (t ) |y(0)|d.
1 L2
1
(0)
|y(0)|t
(1 e1 t ) +
1
1
1 L2 t
+
|y( ) y(0)|d.
1 0
(11.89)
Control applications
255
Now we are ready to compute |y(t) y(0)|. In fact, we have from (11.87) and (11.89)
|y(t) y(0)| A1 (1 e1 t ) + B1 t + H
|y( ) y(0)|d ,
(11.90)
where
A1 = 11 1
(0)
,
B1 = Lu1 |y(0)| + L1 |y(0)| + 11 1 L2 |y(0)|,
H = 11 1 L2 + L1 .
Applying GronwallBellman inequality to (11.90) produces
|y(t) y(0)| A1 (1 e1 t ) +
B1 Ht
(e 1)
H
(11.91)
(11.92)
Note that 1 (T ) and 2 (T ) only depend on the sampling period T once the Lipschitz
constants and the control parameters are chosen.
Next we shall study the behaviour of the sampled-data system during each interval using a Lyapunov function candidate V (y, ) = T P + 12 y2 . Consider (0) =
[ (0), y(0)]T Br . When t (0, T ], its time derivative satises
V = 3
2 + 2 T P + y(1 + y + ud1 )
c0 y2
2 + |y||ud1 (y(0)) uc1 |
c0 y2
2 + Lu1 |y y(0)||y|
Lu1
(1 (T ) + 2 (T )) y2
2
c0
2
Lu1
Lu1
1 (T )|y(0)|2 +
2 (T )
(0)
2
2
2
1 (T )V + 1 (T )V ( (0), y(0)),
+
(11.93)
256
where
"
#
1
1 = min 2c0 Lu1 1 Lu1 2 (T ),
,
max (P)
"
#
Lu1 2 (T )
1 = max Lu1 1 (T ),
,
2min (P)
(11.94)
with max ( ) and min ( ) denoting the maximum and minimum eigenvalues of a matrix
respectively.
Next we shall show that there exists a constant T2 > 0 such that the condition 1 (T ) > 1 (T ) > 0 is satised for all 0 < T < T2 . Note from (11.92) that both
1 (T ) and 2 (T ) are actually the continuous functions of T with 1 (0) = 2 (0). Dene
e1 (T ) := 1 (T ) 1 (T ) and we have e1 (0) > 0. It can also be established from
(11.94) that e1 (T ) is a decreasing and continuous function of T , which asserts by
the continuity of e1 (T ) the existence of T2 so that for 0 < T < T2 , e1 (T ) > 0, that is
0 < 1 (T ) < 1 (T ).
Finally, set T1 = min (T1 , T2 ), and from Lemma 11.10, it is known that V1 c,
which means (T ) c , and subsequently, all the above analysis can be repeated for
every interval [mT , mT + T ]. Applying Lemma 11.10 completes the proof.
2
For systems with relative degree > 1, the implementation of the sampled-data
controller ud2 is given in (11.67) and (11.68). It is easy to see from (11.68) that (mT )
is the exact, discrete-time model of the lter
= + bf ud2
(11.95)
due to the fact that ud remains constant during each interval and the dynamics of
shown in (11.95) is linear. Then (11.68) and (11.95) are virtually equivalent at each
sampling instant. This indicates that we can use (11.95) instead of (11.68) for stability
analysis of the sampled-data system.
Let := [ T , z T ]T and we have the following result.
Theorem 11.12. For the extended system consisting (11.65), (11.69) and the
sampled-data controller ud2 shown in (11.67) and (11.68), and a given neighbourhood of the origin Br := { Rn |
r} with r any given positive real, there exists
a constant T2 > 0 such that, for all 0 < T < T2 and for all (0) Br , the system is
asymptotically stable.
Proof. The proof can be carried out in a similar way to that for the case = 1,
except that the effect of the dynamic lter of has to be dealt with.
For the overall sampled-data system, a Lyapunov function candidate is chosen
the same as for continuous-time case, that is
V = T P +
1 2
z .
2 i=1 i
Control applications
257
Similar to the case of = 1, the sets Br , c and Bl can also be dened such that
Br c Bl , and there exists a T3 > 0 such that for all T (0, T3 ), the following
holds:
(t) Bl ,
t (0, T ],
(0) c .
As in the proof for Theorem 11.11, we also aim to formulate the time derivative
of V ( ) into form (11.77). Next we shall derive the bounds for
(t) (0)
and
|y(t) y(0)| during t [0, T ] with 0 < T < T3 . Consider the case where (0) Br .
We have from (11.95)
t
e(t ) bf ud2 d.
(11.96)
(t) = et (0) +
0
Since is a Hurwitz matrix, there exist positive reals 2 , 3 and 2 such, that
et
2 e2 t ,
et I
3 (1 e2 t ),
where I is the identity matrix. Then, using the Lipschitz property of uc2 with respect
to the set Bl and the fact that ud2 (0, 0) = 0, it can be obtained from (11.96) that
t
2 Lu2
2
(0)
(11.97)
( )
d
1 e2 t +
(|y(0)| +
(0)
)t
2
2
0
and
(t) (0)
3
(0)
(1 e2 t ) +
ud2 (y(0), (0))
2 e2 (t ) d
3 (T )|y(0)| + 4 (T ) (0) ,
(11.98)
where
3 (T ) = 21 2 Lu2 (1 e2 T ),
4 (T ) = (3 + 21 2 Lu2 )(1 e2 T ),
and Lu2 is a Lipschitz constant of uc2 . As for |y y(0)|, we have from (11.69)
t
t
1 ( )d +
( )d
y(t) = y(0) +
+
0
t
y (y) y (y(0)) d +
y (y(0))d.
1
2
L1 |y( ) y(0)|d +
0
L1 |y(0)|d ,
(11.99)
258
where 1 is already shown in (11.89) and 2 in (11.97). With (11.89), (11.97) and
(11.99), it follows that
|y(t) y(0)| A1 (1 e1 t ) + A2 (1 e2 t ) + B2 t
t
+H
|y( ) y(0)|d
(11.100)
(11.101)
where
5 (T ) = H 1 (L1 + 11 1 L2 + 2 21 Lu2 )(eHT 1),
6 (T ) = 11 1 (0 eHT + He0 T (H + 0 ))(H + 0 )1 + 11 1 (1 e0 T ),
7 (T ) = 21 2 (1 e0 T ) + 21 2 Lu2 (eHT 1)
+ 21 2 Lu2 (0 eHT + He0 T (H + 0 ))(H + 0 )1 .
Note that
(0)
appears in (11.98) and (11.101) while for the analysis using Lyapunov
function to carry on,
z(0)
is needed. Therefore, it is necessary to nd out an
expression of
(0)
that exclusively involves
z(0)
, which is shown below.
Notice that due to the special structure of the lter (11.65) and the backstepping
technique, each stabilising function has the property that 1 = 1 (y), 1 (0) = 0, and
i = i (y, 1 , . . . , i1 ) and i (0, . . . , 0) = 0, i = 2, . . . , 1. From (11.74) we have
|1 (0)| |z2 (0)| + |1 (0)| z2 (0)| + L1 |y(0)|
|2 (0)| |z3 (0)| + |2 (0)| |z3 (0)| + L2 |y(0)| + L2 |1 (0)|
..
.
|1 (0)| |z (0)| + |1 (0)| |z (0)| + L1 |y(0)| + L1
|i (0)|
i=1
where with a bit abuse of notation, Li is the Lipschitz constant of i with respect to
the set Bl . Thus, a constant L0 can be found such that the following holds:
(0)
L0 (
z (0)
+ |y(0)|).
(11.102)
Control applications
259
1
1 2
Vd2 = T P + y2 +
z
2
2 i=2 i
during the interval (0, T ] satises
V d2
2
i=1
c1 y2 2 0
(11.103)
i=2
(11.104)
max (P)
#
, 2(0 4 (T )) ,
"
#
2 (T )
2 (T ) = max 2(1 (T ) + 2L20 3 (T )),
, 4L20 3 (T ) .
min (P)
260
11.3.3 Simulation
The following example is a simplied model of a jet engine without stall:
3
1
m = + m2 m3
2
2
= u,
where m is the mass ow and the pressure rise. Take m as the output and then the
above system is in the output feedback form. The lter = + u is introduced so
that the ltered transformation y = x1 and = x2 , and the state transformation
= y can render the system into the following form:
3 2 1 3
=
y y 2 y
2
2
y = + y +
3 2 1 3
y y + .
2
2
Control applications
261
1.2
0.8
x1
0.6
0.4
0.2
0.2
10
t (s)
0.5
x2
0.5
1.5
2
0.2
0.2
0.4
0.6
0.8
x1
1.2
262
4.5
4
3.5
3
2.5
2
1.5
1
0.5
0.01
0.01
0.02
0.03
0.04
0.05
t (s)
0.06
Bibliographical Notes
Chapter 1. The basic concepts of systems and states of nonlinear systems discussed in this chapter can be found in many books on nonlinear systems and
nonlinear control systems, for example Cook [1], Slotine and Li [2], Vidyasagar
[3], Khalil [4] and Verhulst [5]. For the background knowledge of linear systems,
readers can consult Ogata [6], Kailth [7]. Antsaklis and Michel [8], Zheng [9],
and Chen, Lin and Shamash [10]. The existence of unique solutions of differential equations is discussed in the references such as Arnold [11] and Borrelli and
Coleman [12]. Many concepts such as stability and backstepping control design are
covered in detail in later chapters of the book. For system with saturation, readers may consult Hu and Lin [13]. Sliding mode control is covered in detail in
Edwards and Spurgeon [14] and feedforward control in Isidori [15]. Limit cycles
and chaos are further discussed in Chapter 2. Semi-global stability is often related
to systems with saturation, as discussed in Hu and Lin [13], and Isidori [15].
Chapter 2. Lipschitz conditions and the existence of solutions for differential
equations are covered in detail in Arnold [11] and Borrelli and Coleman [12]. Classications of singular points of second-order nonlinear systems are covered in almost
all the introductory texts on nonlinear control systems, Cook [1], Slotine and Li
[2], Vidyasagar [3] and Khalil [4]. More phase portraits of systems similar to the
one in Example 2.1 can be found in Slotine and Li [2], Khalil [4] and Verhulst [5].
The swing machine equation in Example 2.2 was a simplied model of the system
shown in Elgerd [16] and a similar treatment is shown in Cook [1]. More examples
of similar systems can also be found in Verhulst [5]. The van der Pol oscillator was
rst discussed in his publication van der Pol [17], and perhaps it is one of the best
known systems with limit cycles. Detailed coverage of van der Pol oscillator can be
found in several books, including Verhulst [5] and Borrelli and Coleman [12]. Lorenz
attractor was rst shown in Lorenz [18] for his research on heat transfer. The boundedness of the trajectory shown in the book was taken from Borrelli and Coleman [12].
Chapter 3. Describing function analysis is covered in many classic texts on nonlinear control, such as Cook [1], Slotine and Li [2], Vidyasagar [3], and Khalil [4].
The stability criterion of limit cycles with describing functions is inuenced by the
treatment in Slotine and Li [2]. The describing function analysis for van der Pol systems is adapted from Cook [1]. More results on describing functions are to be found
in Atherton [19] and Gelb and Velde [20].
264
Bibliographical Notes
265
the model reference control is based on Ding [37]. Early work of using Lyapunov
function for stability analysis can be found in Parks [38]. A complete treatment of
MRAC of linear systems with high relative degrees can be found in Narendra and
Annaswamy [23] and Ioannou and Sun [39]. The example to show the divergence
of parameter estimation under a bounded disturbance is adapted from Ioannou and
Sun [39], where more robust adaptive laws for linear systems are to be found. Results
on robust adaptive control and on relaxing assumptions of the sign of high-frequency
gain, minimum-phase, etc., for nonlinear systems can be found in Ding [37, 4045].
Chapter 8. Linear observer design is covered in many books on linear systems.
A proof of Lemma 8.2 can be found in Zheng [9]. Early results on nonlinear
observer design can be traced back to 1970s, for example in Thau [46], and Kou,
Ellitt and Tarn [47]. The result on the output injection form was shown in Krener
and Isidori [48]. The geometric conditions for the existence of a state transformation to the output injection form can also be found in Isidori [33] and Marino and
Tomei [24]. Basic knowledge of differential manifolds can be found in Boothby
[32]. Linear observer error dynamics via direct state transformation was initially
shown by Kazantzis and Kravaris [49]. Further results on this topic can be found
in Xiao [50] and Ding [51]. In the latter one, polynomial approximation to nonlinear functions for a solution to a nonlinear state transformation is discussed and
an explicit region of convergence of such an approximation is given. The basic
idea of observer design for systems with Lipschitz nonlinearities is to use the linear part of the observer dynamics to dominate the nonlinear terms. This idea was
shown in some of the early results on nonlinear observers, but the systematic introduction to the topic was due to Rajamani [52], in which the condition (8.31) was
shown. Some later results on this topic are shown in Zhu and Han [53]. Onesided Lipschtiz condition for nonlinear observer design was shown in Zhao, Tao
and Shi [54]. The result on observer design for systems with Lipschitz output nonlinearities is adapted from Ding [55]. The result on reduced-order observer of linear
systems can be traced back to the very early stage of observer design in Luenberger
[56]. Early results on this topic were discussed in Kailath [7]. The reduced-order
observer design is closely related to observer design for systems with unknown
inputs, of which some results may be found in Hou and Muller [57]. For nonlinear systems, a reduced-order observer was used for estimation of unknown states
for control design of a class of nonlinear systems with nonlinearities of unmeasured states by Ding [58]. The result shown here is based on Ding [31]. An
early result on adaptive observer for nonlinear systems is reported in Bastin and
Gevers [59]. Results on adaptive observers with output nonlinearities are covered
in Marino and Tomei [24]. The results shown in Section 8.6 is mainly based on Cho
and Rajamani [60]. The method to construct a rst-order positive real system in
Remark 8.9 is based on Ding [40].
Chapter 9. Early ideas of backstepping appeared in a number of papers, such
as Tsinias [61]. Iterative backstepping for state feedback is shown in Kanellakopoulos, Kokotovic and Morse [62] with adaptive laws for unknown parameters.
266
Backstepping designs without adaptive laws shown in this chapter are basically simplied versions of their adaptive counterparts. Filtered transformations were used for
backstepping in Marino and Tomei [63], and the presentation in this chapter is different from the forms used in [63]. Initially, multiple estimation parameters are used
for adaptive control with backstepping, for example the adaptive laws in [63]. Tuning function method was rst introduced in Krstic, Kanellakopoulos and Kokotovic
[64] to remove multiple adaptive parameters for one unknown parameter vector with
backstepping. Adaptive backstepping with ltered transformation was based on Ding
[65], without the disturbance rejection part. Adaptive observer backstepping is largely
based on the presentation in Krstic, Kanellakopoulos and Kokotovic [66]. Nonlinear
adaptive control with backstepping is also discussed in details in Marino and Tomei
[24]. Some further developments of nonlinear adaptive control using backstepping can
be found in Ding [40] for robust adaptive control with dead-zone modication, robust
adaptive control with -modication in Ding [67], with unknown control directions
and unknown high-frequency gains in Ding [41, 42]. Adaptive backstepping was also
applied to nonlinear systems with nonlinear parameterisation in Ding [68].
Chapter 10. Early results on rejection of sinusoidal disturbances can be found
in Bodson, Sacks and Khosla [69] and Bodson and Douglas [70]. Asymptotic
rejection of sinusoidal disturbances for nonlinear systems in the output feedback
form is reported in Ding [71] when the frequencies are known. The result shown
in Section 10.1 is adapted mainly from Ding [72]. Asymptotic rejection of periodic disturbances can be formulated as an output regulation problem when the
periodic disturbances are generated from a dynamic system, which is known as
an exosystem. Output regulation for linear systems is shown in Davison [73]
and Francis [74], and for nonlinear systems with local stability in Huang and
Rugh [75], and Isidori and Byrnes [76]. More results on output regulation with
local stability can be found in the books by Isidori [33] and Huang [77]. The
result shown in Section 10.2 is based on Ding [65]. The format of the exosystem for internal model design was inspired by Nikiforov [78]. The Nussbaum
gain was rst used by Nussbaum [79], and the treatment here is similar to Ye
and Ding [80] used in the control design to tackle the unknown sign of high frequency gain. Early results on output regulation with nonlinear exosystems can
be found in Ding [81, 82], and Chen and Huang [83]. The result shown in
Section 10.3 is based on Xi and Ding [84]. Asymptotic rejection of general disturbances was studied as an extension of output regulation of nonlinear exosystems.
Waveform proles of the general periodic disturbances were used to estimate the
equivalent input disturbance in some early results in Ding [8587]. The approach
for asymptotic rejection with an observer-like internal model is adapted from
Ding [55, 88].
Chapter 11. Harmonic estimation is treated as a special case of state estimation.
The recursive estimation and rejection method in Section 11.1 is based on Ding
[89]. Practical issues on harmonics estimation and rejections are covered in Arrillaga,
Watson and Chen [90] and Wakileh [91].
Bibliographical Notes
267
Clock gene was reported in 1994 by Takahashi et al. [92]. Reports on key genes
in circadian oscillators are in Albrecht et al. [93], van der Horst et al. [94] and Bunger
et al. [95]. The effects of jet lags and sleep disorders to circadian rhythms are reported
in Sack et al. [96] and Sack et al. [97]. The effects of light and light treatments
to disorders of circadian rhythms are shown by Boulos et al. [98], Kurosawa and
Goldbeter [99] and Geier et al. [100]. The circadian model (11.48) is adapted from
Gonze, Leloup and Goldbeter [101]. The condition (11.52) is shown by Zhao, Tao
and Shi [54]. The observer design is adapted from That and Ding [102], and the control design is different from the version shown in that paper. More results on observer
design for a seventh-order circadian model are published by That and Ding [103].
There are many results in literature on sampled-data control of linear systems,
for example see Chen and Francis [104] and the references therein. The difculty
of obtaining exact discrete-time model for nonlinear systems is shown by Nesic and
Teel [105]. Structures in nonlinear systems cannot be preserved as shown in Grizzle
and Kokotovic [106]. Approximation to nonlinear discrete-time models is shown
by several results, for example Nesic, Teel and Kokotovic [107] and Nesic and Laila
[108]. The effect on fast-sampling of nonlinear static controllers is reported in Owens,
Zheng and Billings [109]. Several other results on emulation method for nonlinear
systems are shown in Laila, Nesic and Teel [110], Shim and Teel [111] and Bian
and French [112]. The results presented in Section 10.3 are mainly based on Wu and
Ding [113]. Sampled-data control for disturbance rejection of nonlinear systems is
shown by Wu and Ding [114]. A result on sampled-data adaptive control of a class of
nonlinear systems is to be found in Wu and Ding [115].
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
References
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
[34]
[35]
[36]
[37]
[38]
[39]
[40]
269
270
[41]
[42]
[43]
[44]
[45]
[46]
[47]
[48]
[49]
[50]
[51]
[52]
[53]
[54]
[55]
[56]
[57]
[58]
References
[59]
[60]
[61]
[62]
[63]
[64]
[65]
[66]
[67]
[68]
[69]
[70]
[71]
[72]
[73]
[74]
[75]
271
272
[76]
[77]
[78]
[79]
[80]
[81]
[82]
[83]
[84]
[85]
[86]
[87]
[88]
[89]
[90]
[91]
[92]
References
273
[93] Albrecht, U., Sun, Z., Eichele, G. and Lee, C. A differential response of two
putative mammalian circadian regulators, mper1 and mper2 to light, Cell
Press, vol. 91, pp. 10551064, 1997.
[94] van der Horst, G., Muijtjens, M., Kobayashi, K., Takano, R., Kanno,
S., Takao, M., de Wit, J., Verkerk, A., Eker, A., van Leenen, D., Buijs,
R., Bootsma, D., Hoeijmakers, J., and Yasui, A. Mammalian cry1 and
cry2 are essential for maintenance of circadian rhythms, Nature, vol. 398,
pp. 627630, 1999.
[95] Bunger, M., Wilsbacher, L., Moran, S., Clendenin, C., Radcliffe, L.,
Hogenesch, J., Simon, M., Takahashi, J., and Bradeld, C., Mop3 is an
essential component of the master circadian pacemaker in mammals, Cell,
vol. 103, pp. 10091017, 2000.
[96] Sack, R., Auckley, D., Auger, R., Carskadon, M., Wright, K., Vitiello, M.,
and Zhdanova, I. Circadian rhythm sleep disorders: part i, basic principles,
shift work and jet lag disorders, Sleep, vol. 30, pp. 14601483, 2007.
[97] Sack, R., Auckley, D., Auger, R., Carskadon, M., Wright, K., Vitiello, M.,
and Zhdanova, I. Circadian rhythm sleep disorders: Part ii, advanced sleep
phase disorder, delayed sleep phase disorder, free-running disorder, and
irregular sleep-wake rhythm, Sleep, vol. 30, pp. 14841501, 2007.
[98] Boulos, Z., Macchi, M., Sturchler, M., Stewart, K., Brainard, G., Suhner, A.,
Wallace, G., and Steffen, R., Light visor treatment for jet lag after west
ward travel across six time zones, Aviat. Space Environ. Med., vol. 73,
pp. 953963, 2002.
[99] Kurosawa, G. and Goldbeter, A. Amplitude of circadian oscillations
entrained by 24-h light-dark cycles, J. Theor. Biol., vol. 242, pp. 478488,
2006.
[100] Geier, F., Becker-Weimann, S., Kramer, K., and Herzel, H. Entrainment in
a model of the mammalian circadian oscillator, J. Biol. Rhythms, vol. 20,
pp. 8393, 2005.
[101] Gonze, D., Leloup, J., and Goldbeter, A., Theoretical models for circadian
rhythms in neurospora and drosophila, Comptes Rendus de lAcademie des
Sciences -Series III -Sciences de la Vie, vol. 323, pp. 5767, 2000.
[102] That, L. T. and Ding, Z. Circadian phase re-setting using nonlinear outputfeedback control, J. Biol. Syst., vol. 20, no. 1, pp. 119, 2012.
[103] That, L. T. and Ding, Z. Reduced-order observer design of multi-output
nonlinear systems with application to a circadian model, Trans. Inst. Meas.
Control, to appear, 2013.
[104] Chen, T. and Francis, B. Optimal Sampled-Data Control Systems. London:
Springer-Verlag, 1995.
[105] Nesic, D. and Teel, A. A framework for stabilization of nonlinear sampleddata systems based on their approximate discrete-time models, IEEE Trans.
Automat. Control, vol. 49, pp. 11031122, 2004.
[106] Grizzle, J. W. and Kokotovic, P. V. Feedback linearization of sampled-data
systems, IEEE Trans. Automat. Control, vol. 33, pp. 857859, 1988.
274
[107]
[108]
[109]
[110]
[111]
[112]
[113]
[114]
[115]
Nesic, D., Teel, A., and Kokotovic, P.V. Sufcient conditions for stabilization of sampled-data nonlinear systems via discrete-time approximations,
Syst. Control Lett., vol. 38, pp. 259270, 1999.
Nesic, D. and Laila, D. A note on input-to-state stabilization for nonlinear
sampled-data systems, IEEE Trans. Automat. Control, vol. 47, pp. 1153
1158, 2002.
Owens, D. H., Zheng, Y. and Billings, S. A. Fast sampling and stability of
nonlinear sampled-data systems: Part 1. Existing theorems, IMA J. Math.
Control Inf., vol. 7, pp. 111, 1990.
Laila, D. S., Nesic, D. and Teel, A. R. Open-and closed-loop dissipation inequalities under sampling and controller emulation, Eur. J. Control,
vol. 18, pp. 109125, 2002.
Shim, H. and Teel, A. R. Asymptotic controllability and observability
imply semiglobal practical asymptotic stabilizability by sampled-data output
feedback, Automatica, vol. 39, pp. 441454, 2003.
Bian, W. and French, M. General fast sampling theorems for nonlinear
systems, Syst. Control Lett., vol. 54, pp. 10371050, 2005.
Wu, B. and Ding, Z. Asymptotic stablilisation of a class of nonlinear systems
via sampled-data output feedback control, Int. J. Control, vol. 82, no. 9,
pp. 17381746, 2009.
Wu, B. and Ding, Z. Practical disturbance rejection of a class of nonlinear systems via sampled output, J. Control Theory Appl., vol. 8, no. 3,
pp. 382389, 2010.
Wu, B. and Ding, Z. Sampled-data adaptive control of a class of nonlinear
systems, Int. J. Adapt. Control Signal Process., vol. 25, pp. 10501060,
2011.
Index
-modication, 106
absolute stability, 60
adaptive control
direct, 89
indirect, 89
model reference, 89
nonlinear, 159
adaptive law
linear system, 101
nonlinear, 162, 172
robust, 107
asymptotic stability, 43
autonomous system, 11
backstepping
adaptive, 159
adaptive observer, 167
ltered transformation, 152
integrator, 141
iterative, 144
observer, 147
Barbalats lemma, 92
centre, 14
certainty equivalence principle, 90
chaos, 4, 21
circadian disorders, 238
circadian rhythm, 239
Neurospora control input, 246
Neurospora model, 239
circle criterion, 62
comparison functions
class K 66
class KL 66
class K , 66
comparison lemma, 50
dead zone, 31
dead-zone modication, 105
describing function, 27
dead zone, 32
ideal relay, 30
relay with hysteresis, 32
saturation, 30
stability of limit cycle, 35
detectability
linear system, 110
differential stability, 71
distribution, 84
involutive, 84
disturbance
bounded, 104, 107
equivalent input, 17677
general periodic, 204, 205
dynamic model, 207
generated from nonlinear
exosystem, 194
sinusoidal, 175
disturbance rejection
general periodic, 204
sinusoidal, 175
eigenvalue
resonant, 121
equivalent input disturbance, 17677
exosystem, 175, 176
linear, 176, 186
nonlinear, 195
exponential stability, 43
feedback linearisation, 76
feedforward
276
Lie derivative, 77
limit cycle, 4, 18, 34, 175, 195, 238, 240
linear observer error dynamics, 112,121
linearisation
around operation point, 10
feedback, 75, 76, 80
full state, 83, 85
linearised model, 10
Lipschitz constant, 9, 123, 207, 241
one-sided, 124, 241
Lorenz attractor, 21
Lyapunov function, 46
adaptive control, 101
adaptive observer, 139
ISS, 6770
linear systems, 52
Lyapunov stability, 42
Lyapunov theorem
global stability, 49
local stability, 47
Lyapunovs direct method, 46
Lyapunovs linearisation method, 45
model reference control
relative degree > 1, 9798
relative degree one, 9496
monic polynomial, 94
MRAC, 89
multiple equilibrium points, 4
nonlinearity, 1, 3
dead zone, 31
hard, 3
ideal relay, 30
Lipschitz, 7, 122
multi-valued, 3
relay with hysteresis, 32
saturation, 29
sector-bounded, 59
normal form, 81
Nussbaum gain, 190
observability
linear system, 109
Index
observer
general periodic disturbance, 207
observer canonical form
linear system, 118
observer design
adaptive, 137
direct state transformation, 121
Lipschitz nonlinearity, 122
output injection, 112
output nonlinearity, 126
reduced-order, 127
output feedback form, 148, 152
with disturbance, 175, 186, 194
with unknown parameter, 167
output feedback stabilisation, 158
output injection form, 112
state transformation, 114
output regulation, 186
adaptive, 186
nonlinear exosystem, 194
output-complement transformation, 127
pendulum model, 48
phase portrait, 12
Poincare-Bendixson theorem, 19
positive denite function, 46
positive limit set, 21
positive real transfer function, 55
strictly, 56
radially unbounded function, 49
relative degree, 77
relay with hysteresis, 32
robust adaptive law, 107
-modication, 106
dead-zone modication, 105
saddle point, 13
sampled-data
emulation method, 248
sampled-data controller, 250
saturation, 2, 29
sector-bounded nonlinearity, 59
self-tuning control, 89
semi-global stability, 3, 260
singular point, 12
small gain theorem
ISS, 70
stability
absolute, 60
asymptotic, 43
differential, 71
exponential, 43
globally asymptotic, 45
globally exponential, 45
ISS, 66
Lyapunov, 42
semi-global, 3, 260
stable focus, 1314
stable node, 13
STC, 89
transformation
ltered, 152, 187
output-complement, 127
tuning function, 161
unstable focus, 14
unstable node, 13
van der Pol oscillator, 18, 38
Youngs inequality, 128
zero dynamics, 81, 155
277
Nonlinear and
Adaptive Control
Systems
Ding
Zhengtao Ding
04/03/2013 09:32:20