On Neural Networks in Identification and Control of Dynamic Systems
On Neural Networks in Identification and Control of Dynamic Systems
R=19930021849 2020-05-14T18:51:00+00:00Z
Minh Phan
Jer-Nan Juang
David C. Hyland
G3/39 0177080
June 1993
1. Introduction
System identification and control are two related fields that have received
considerable development in the last few decades. System identification deals with the
problem of finding a mathematical description of a physical system from experimental data.
Control theory devises ways to influence the system in a desirable and predictable manner.
Typical control objectives are pointing control, vibration suppression, and tracking control.
System identification provides the necessary mathematical model of a system for a
particular control scheme to be designed. In turn, information gathered during the control
process can be used to evaluate the validity of the assumed model. Existing system
identification and control methods are based o n mathematical systems theory, which first
deals with deterministic then stochastic systems. For the most part, the systems under
study are idealized. They are linear, time-invariant, and often assumed to be noise-free.
When noises are present, they are assumed to be white, zero-mean, and with known
characteristics. These assumptions are often justified because less idealized assumptions
tend to render the analysis mathematically intractable.
2
2. The Neural Networks
A neural network is simply a set of interconnected individual units called neurons.
Depending on the connection between the neurons, there are two basic types of networks
known as the multi-layer feedforward networks and the recurrent networks, which will be
described in this section.
2.1 The Neuron. As a basic building block for a neural network, an individual
neuron has a finite number of scalar inputs and one scalar output. Associated with each
input is a scalar weighting value. The input signals are weighted by these values and added
at the summation junction. The combined signal is then passed through an activation
function producing the output signal. The activation function y ( x ) can take a variety of
forms, the most common one is a sigmoid function denoted by sigm(x),
1 - e-x
sigm(x) = -
1 + e-’
A plot of the sigmoid function is shown in Fig. 1 below. Generally, the activation function
can be any non-decreasing differentiable function which has finite limits at both ends as
shown in Fig. 1 below.
-2 I -
-10 -5 0 5 0
X
Figure 1: The sigmoid function.
Let r inputs of a neuron be denoted by ul, u2, ..., ur and the output denoted by y .
Let the r weights be denoted by wI, w2,...) w,. The output of the neuron can be
expressed mathematically as
The neuron is shown schematically in Fig. 2 below with the sigmoid function as the
activation function. For simplicity of notations, the network weights for the i-th layer may
sometimes be presented collectively as W = { w,,w 2 ,w3,...} .
3
Y
sigm ( x )
Remark 2.1.1. The activation function is a limiter that bounds the incoming
signal, which serves as a non-linear element in a neuron. The activation function given in
Eq. (1) has a linear range about the origin, and is bounded between -1 and 1. If output of
a neuron is bounded between -a and +a by taking y ( x ) = asigrn(x) as its activation
function then Eq. (2) becomes
jj
(ill )
=7 wiui = asigm C wiui = a y
(ill ) (3)
which is the same as the output of a neuron with the original activation function multiplied
by a constant gain a. Therefore, the activation function can be taken to be between -1 and
1 provided an additional factor is inserted after the neuron. In a network, this factor is
absorbed into the weights of the following neurons that directly receive the output of this
neuron as their inputs.
Remark 2.1.2. If the activation function is a linear function, y ( x ) = x , then the
neuron is a linear neuron. The input-output relationship of a linear neuron is
Equation (4) simply says that the output signal is a weighted (linear) combination of the
input signals with the weighting coefficients being the weights of the neuron.
2.2 Multi-Layer Feedforward Neural Network. A multi-layer feedforward
neural network consists of an input layer, a number of hidden layers, and an output layer.
In a fully connected feedforward network, every neuron in each layer accepts as its inputs
all signals coming from all neurons in the layer immediately preceding it (see Fig. 3). In a
partially connected network, some of these connections are missing. This is equivalent to
setting the corresponding network weights to zero. Figure 3 show a typical three-layer
three-input three-output feedforward network with two hidden layers.
Wl w2 w3
Yl
Y2
Y3
I
Figure 3: A three-layer three-input three-output feedforward neural network.
~-
2.3 Recurrent Network. A feedforward network with time delay feedback
elements is called a recurrent network. The delay elements take the outputs of certain
neurons in the network, delay them for a certain number of time steps, and feed back as
input to the neurons. In other words, in a recurrent network, time delayed outputs of a
certain number of neurons are the inputs to other neurons. A special one layer network
when the delayed outputs of the neurons are fed back as inputs to themselves is called a
Hopfield network (see Fig. 4).
W
Yl
Y2
Y3
5
Figure 5 : A two-layer three-input one-output feedforward network of linear neurons
The network weights between the individual connections are shown in the figure.
Since the neurons are linear, each neuron is represented by a summation junction and the
activation being a linear function is omitted. The output of the network in Fig. 5 is simply,
6
3.2. Feedforward Linear Network and the State Space Model. This
section describes the relationship between the feedforward linear network and the state
space model which is a common form of representing linear systems. The discrete-time
state space model of an n-th order, m-input, q-output system is a set of n simultaneous first
order difference equations of the form
x ( k + 1) = A x ( k ) + Bu(k)
(7)
y ( k ) = Cx(k)+ Du(k)
are the Markovparumeters of the system described by Eqs. (7), which are also the system
pulse response samples. The Markov parameters are expressed in terms of the system
discrete state space matrices A, B , C , D. Since the state vector is coordinate-dependent, the
state space matrices are not unique for a given system but the Markov parameters are
unique. Let the state vector be transformed by a coordinate transformation T , z ( k ) = T x ( k ) ,
then the relationship between u(k) and y(k) via a new state vector z(k) can be described by
a new state space representation TAT-', TB, CT-', D . The sy&em Markov parameters
computed using the new state space matrices are the same as before, i.e.,
For an asymptotically stable systems, the pulse response can be neglected after a finite
number of time steps, say p s . The input-output description in Eq. (8) can be approximated
by a finite number of Markov parameters
where pJ is sufficiently large such that CA'B = 0, k 2 pJ. Comparing Eq. (1 1) with the
structure of the linear neurons immediately leads to the following remarks.
Remark 3.2.1. The elements of the Markov parameters a-e simply the weights of a
single-layer linear network where inputs to the network include both current and past
values of the input signal. Note that the time delayed inputs do not affect the neuron
configuration because they are feedforward signals and thus can be treated as separate input
channels. This case is shown in Fig. 6.
7
Figure 6: Representation of linear systems by a feedforward network
with the system Markov parameters as network weights.
8
If M is a matrix such that A + MC is deadbeat of order p , i.e.,
(A+MC)"rO, k 2 p
The matrix M in the above development can be interpreted as an observer gain. The
system considered in Eqs. (7) has an observer of the form
Besides the effect of noises, j ( k ) may differ from y(k) if the actual initial condition x(0) is
not known and some different initial condition is assumed for i ( 0 ) . Defining the state
estimation error e(k)= x ( k ) - i ( k ) , the equation that governs e(k) is
e(k + 1) = ( A + M C ) e ( k ) (18)
For an observable system, the matrix M exists such that the eigenvalues of A + MC may be
placed in any desired (symmetric) configuration. If the matrix M is such that A + M C is
asymptotically stable, then the estimated state i ( k ) tends to the true state x(k) as k tends to
infinity for any initial difference between the assumed observer state and the actual system
state. The matrix M can therefore be interpreted as an observer gain. The parameters
defined as
F(k)=C(A+MC)'-'[B+MD , -MI
=[a 7
are the Markov parameters of an observer system, hence they are referred to as observer
Markov parameters. Like the system Markov parameters, the observer Markov
parameters are also invariant with respect to a coordinate transformation of the state vector.
To see this, again let the state vector be transformed by a coordinate transformation T ,
z ( k ) = T x ( k ) , then the observer is described by a new state space representation
TAT-', TB, CT-I, D and a new observer gain T M . The observer Markov parameters
computed using these new state space matrices and the new observer gain are the same as
before,
y(k)= CT-l(TAT-I+ TMCT-l)'-'[TB+ TMD , - TM]
(20)
=C(A+MC)'-'[B+MD, - M I , k = l , 2, 3, ...
Notice that in Eq. (15), the output y ( k ) is the open-loop response of the system, yet
the Coefficients ak, p k are related to an observer gain. Consider the special case where A4
is a deadbeat observer gain where all eigenvalues of A + MC are zero, the observer Markov
parameters will become identically zero after a finite number of terms. For lightly damped
stwtures, this means that the system can be described by a reduced number of observer
Markov parameters r ( k ) instead of an otherwise large number of the usual system Markov
parameters Y(k). For this reason, the observer Markov parameters are important in linear
system identification. By examining of the structure of Eq. (15), the following remarks
can be made.
Remark 3.3.1. The input-output equation given in Eq. (15) can be represented by
a recurrent network with a single layer of linear neurons. The number of neurons is equal
to the number of outputs of the system. The inputs to the neurons consists of both the
feedforward time-delayed input signals and the feedback time-delayed output signals.
Figure 7 shows the configuration of such a network for a single-output system.
Remark 3.3.2. The recurrent network weights are precisely the elements of the
observer Markov parameters. The relationship between the weights of a recurrent network
and an equivalent feedforward network is the same as that between the observer Markov
parameters and the system Markov parameters. It can be shown that the system Markov
parameters or the feedforward network weights are related to the recurrent network weights
by
where ak= 0, pk I 0 for k > p . To describe a system of order n, the number of observer
Markov parameters p must be such that qp 2 n where 4 is the number of outputs.
Furthermore, the maximum order of a system that can be described with p observer
Markov parameters is qp? The implication of this result to the network configuration is
that a recurrent network generally requires fewer number of parameters (or weights) than
that required by an equivalent feedforward network. The two equivalent networks,
however, have the same number of neurons. The minimum number of recurrent network
weight matrices that can describe the system is pmh,which is the smallest value of p such
that 4pAn 2 n.
Remark 3.3.3. As mentioned previously, to represent lightly damped structures,
the feedforward representation requires a large number of weights. Furthermore, it is not
possible to represent a marginally stable or unstable system by a feedforward network.
However, it is possible to represent such a system by a recurrent network. The implication
of this fact for the system identification problem will be discussed further in later sections.
4. Identification of Linear Systems using Neural Networks
It has been shown that a general network of linear neurons is equivalent to a single
neuron with appropriate weights. The problem of linear system identification using neural
network is therefore reduced to finding these network weights from input-output data. The
computation may be done off-line or on-line. In off-line computation, the input-output data
is already available and a network representing the system is to be determined. On-line
computation refers to the case where the network weights are continually updated as data is
made available.
4.1. Parallel vs. Series-Parallel Identification Models. In previous
consideration, it appears that the recurrent network is more advantageous in representing
certain systems than the feedforward network. To identify the recurrent network weights
one can simply use the feedforward network configuration with actual delayed system
outputs appeared as inputs to the feedforward network. Consider two identification models
shown in Figs. 8 and 9 below, which are known as parallel and series-parallel
identification model. The block denoted by D represents the time delay elements.
11
Figure 9: Identification using series-parallel model.
The basic difference between the two schemes is that in the parallel identification
model, the estimated output j ( k ) is computed based on the model own previous (estimated)
values whereas i n the series-parallel model, i t is based on actual output values.
Mathematically_, in the parallel model, the purpose of the identification is to obtain the
estimates &k, pk of the coefficients %, /$ that minimize the estimation error, e ( k ) =
y(k) - j ( k ) , where the estimated output j(k) is computed from
j(k)=&,y(k-l)+ + ~ , , ~ ( k - p ) + ~ ~ , u ( k ) + P I u ( k - l -..
) + +B,u(k-p) (23)
The difference between the two above equations is a subtle but important one. As
discussed in the previous section, the estimated output of the model in Eq. (22) is the
estimated open-loop prediction even though the coefficients of the model are related to an
observer. On the other hand, the estimated output of the model in Eq. (23) is that of an
observer. To see this, substitute the expression for j ( k ) to the estimated state equation in
(17) produces
Since j ( k ) = C.?(k)+ D u ( k ) , one can obtain Eq. (23) assuming zero initial conditions for
the observer. Therefore, j ( k ) in Eq. (23) represents the estimated output provided by the
observer. The estimation error 2(k) is the difference between the actual output and the
estimated output provided by the observer. On the other hand, if the actual response y(k) is
replaced by the estimated value j ( k ) in Eq. (23) then the terms involving the observer gain
Mcancel each other identically for any arbitrary initial condition i ( O ) ,
Z(k+I)=(A + MC)i(k)+(B+MD)u(k)-Mj(k)
= A i ( k ) + Bu(k)
12
Therefore, there is no longer any observer involved in the equation; ?(k) now plays the
role of the state vector x ( k ) as in Eq. (7) and the estimated output j ( k ) = C?(k)+ Du(k) is
the same as that produced by the open-loop model provided that the initial conditions for
i ( k ) and x ( k ) are identical. The quantity $ ( k ) now represents the predicted output
provided by the open-loop model alone, which is referred in this paper as open-loop
prediction. In this case, the error i ( k ) is the difference between the actual output and the
predicted output provided by the identified open-loop model.
Remark 4.1.1. First recall that the model structure in Eq. (15) subsumes an
observer. If the parallel identification model is used in conjunction with the model structure
of Eq. (15) then the prediction error that drives the parameter estimation scheme is simply
the open-loop prediction error not the observer (output) estimation error. Consequently,
the observer portion of the model cannot be identified. This fact accounts for the
difficulties encountered in parallel model identification, namely, the conditions for which
the scheme will converge are presently not known.
Remark 4.1.2. In the series-parallel identification model, since the actual instead of
(open-loop) predicted output enters the model, a feedforward network with delayed input
and actual output measurements can be used to identify the system. This consideration
eliminates the use of a recurrent network which would introduce additional but unnecessary
difficulties to the system identification problem, (see Fig. 10). Each output of the system is
represented by a single linear neuron. A multiple-output system is represented by a single
layer of neurons. The identified network can be used either as a feedforward or a recurrent
network. In the former case, the network provides estimation of the response by an
observer. In the latter case, it is an open-loop predictor. Again, this depends on whether
actual or predicted output is used in computing the response.
13
For simplicity, consider the case where the system starts from zero initial conditions.
Equation (15) can be written as
where network weight matrices pi , ai are defined in Eq. (16). Writing Eq. (26) in matrix
form for a set of input-output data N+1 samples long yields
y=YV
where
V=
? = yv+ (31)
, where (.)+ denotes the pseudo-inverse of the quantity in the parentheses. If the initial
conditions are not zero then a slightly differen-t equation muit be used to solve for the
network weights, that is
where yt and VI are obtained by deleting the first p columns in y and V, respectively.
Remark 4.2.1. The least-squares solution in Eq. (31) or (32) minimizes the enor
between the actual output and the estimated output computed using the actual input and
output data, i.e., theJeast-sqyares solution minimizes the residual e = y - $ where 9 is
computed from 9 = YV and Y is given in E% (31). If Eq. (32) is used instead, then the
least-squares solution minimizes 6, = y - y, where 9, = YV,. This computation,
therefore, corresponds to the series-parallel identification scheme that minimizes the
observer estimation error.
Remark 4.2.2. Ideally, the error between the actual output and the predicted output
provided by the identified open-loop model is the proper error to be minimized for the
identification of the system open-loop model. The above computation minimizes the
observer estimation error instead. For a linear system, it turns out that in the absence of
14
noises the open-loop system can be identified exactly by minimizing the observer
estimation error. In the presence of noises, however, minimizing the observer estimation
error does not necessarily implies that the open-loop prediction error is minimized.
Therefore, it is possible that the observer model fits the data well but the open-loop model
does not. Fortunately, if the order of the regression equation is chosen to be sufficiently
large then simultaneous observer and system identification will still be achieved in the limit
as the data record tends to infinity and the noises are white, Gaussian, and zero-mean (see
Ref. 9).
Remark 4.2.3. The least-squares solution in Eq. (31) can be obtained by an on-
line parameter estimation scheme. First, write each column in V as
1s
5.1. One-Step Ahead Predictor. To express explicitly the observer as an
one-step ahead predictor, one simply writes the observer equations as
i(k+l)=(A+MC)?(k)+(B+MD)u(k)-My(k)
(37)
j ( k + 1) = C i ( k + 1) + Du(k + 1)
As a predictor, the interested quantity is j(k + 1). One can therefore bypass the state
equation by writing
The following remarks can be made regarding the forms of Ey. (37) and Eq.(38).
Remark 5.1.1. In theory, if the state space model (A, B , C, D ) is known exactly
then one can design an observer gain M such that A + M C is asymptotically stable. To use
Eq.(38) as an output predictor, one need to include a sufficient number of terms such that
( A + MC)' is negligible for i 2 p. The state space representation in 3.(37) is a better
choice since it involves no such approximation. The above comment no longer holds true
if M is such that A + M C is deadbeat, i.e., ( A + MC)' 0, i 2 p since the approximation
becomes exact in this case.
Remark 5.1.2. In practice, the system model cannot be known exactly. To
identify the system from input-output data using the series-parallel structure, one in fact
computes directly the coefficients in Q. (38) rather than the state space mamces. To obtain
a minimal order state space representation from these coefficients, realization is rcquired.
As an output predictor, therefore, Q. (38) should be used directly because conversion to a
state space representation is not necessary for this purpose.
Remark 5.1.3. Equation (38) clearly indicates that the one-step ahead predictor
takes the form of a single layer network of linear neurons with actual input and output
signals entering the network, and the output of the network represents the one-step ahead
prediction. Schematically, this is the same as shown in Fig. 10.
5.2. A Two-step Ahead Predictor. This section derives the equations for a
two-step ahead predictor for linear systems, and shows that it also has a linear neural
network form. First, from Eq. (7). one can write
x(k+2)=Ax(k+l)+Bu(k+l)
+
= A * x ( k ) ABu(k) + Bu(k + 1) (39)
+
y(k + 2) = Cx(k + 2) Du(k + 2)
Adding and subtracting the term Gy(i) to the right hand side of the state equation yields
16
then the relationship between the input and output of the system can be described as a linear
combination of input-output data of the form
for k 2 2 p - 2 so that at sufficiently large time steps, terms involving the states x(0) and
x(1) vanish', Le., C(A2+ GC)'x(O)= 0, C(A2+ GC)'x(l) = 0, for i 2 p due to the
imposed deadbeat condition for A2 + GC. Furthermore, there is only a finite number of
coefficients that make up the linear combination in g(.), which are the predictor Markov
parameters of the form
Existence of the matrix G such that ( A 2 + GC)' E0 , k 2 p is assured if the pair (A2, C)
is observable.
Remark 5.2.1. The above derivation justifies the form of a two-step ahead
predictor. In fact, one can identify the coefficients of this predictor from input-output data
by minimizing the two-step prediction error. The procedure is similar to that discussed in
Section 4.
Remark 5.2.2. To obtain the two-step ahead prediction one can also propagate the
observer, which is a one-step ahead predictor, in two successive time steps by treating the
estimated output from the first time step as the actual output for the second time step
However, such a procedure would amount to performing open-loop prediction in the
second time step and is therefore sensitive to noises. On the other hand, if one uses the
predictor form with the coefficients directly identified from input-output data then only
actual data enter the computation and thus minimizes the errors due to noises.
Remark 5.2.3. Again, the predictor form can be represented by a single layer of
linear neurons, and the weights of this network are simply the elements of the predictor
Markov parameters shown i n Eq. (43). Results presented in this section can be easily
generalized to a general multi-step predictor. The relationship between such predictors and
the deadbeat control problem will be discussed in a later reference.
6. Control of Linear Systems using Neural Networks
As formulated in previous sections, linear systems can be represented by a single-
layer network of linear neurons. The weights of this network can be identified from input-
output data. Once identified, the network can be used as a one-step ahead predictor. This
section discusses the use of such network directly for control application without requiring
the state space model to be extrilcted from these weights.
6.1. A One-Step Ahead Controller. First, consider the case where the
linear system can be expressed in the form,
where the coefficients are assumed to be known. Let the desired response be denoted by
r(k). To obtain a controller directly from the above equation, one simply replaces y ( k + 1)
17
by its desired value r ( k + l ) and then solve for the control input u(k+l) to obtain the control
law,
If the coefficients in Eq. (44)are not known exactly then Eq. (46) represents an one-step
ahead estimate of what the system will produce based on current and past input-output data,
Define the tracking error to be the difference between the actual response and the desired
response, ~ ( k=) y ( k ) - r ( k ) , Eq. (50) reveals that
E(k + 1) = e ( k + 1) (51)
Therefore, if the predictor is such that its prediction error vanishes in the limit, then the
tracking error also vanishes in the limit, i.e.,
Remark 6.1.1. The one-step ahead control law has the property that the tracking
error is the same as the prediction error. The above analysis shows that accuracy of the
predictor model governs the accuracy of the tracking response. As long as the predictor
can perform a reasonably good one-step ahead prediction of the system response then the
control input can be computed to make the system track a desired trajectory. In the ideal
case where the system is linear and the data is noise-free, the prediction error and the
18
tracking error will be zero identically. Non-zero prediction and tracking error can only
come about during adaptation or when noises are present. This is different from the non-
linear case where both the estimation and tracking error are non-zero even when there are
no noises in the system. An important resmction of the one-step ahead controller is that the
open-loop system is required to be stably invertible, (ie., there are no unstable zeros in the
linear case). If this condition is not met, it is possible to have the controller producing
unbounded input while maintaining zero tracking error.
Remark 6.1.2. To obtain the result in Eq. (51), the controller coefficients must be
the same as those of the predictor model. In the event the coefficients of the predictor
model are updated at each time step, then the controller coefficients must also match those
of the predictor model. Mathematically, if at time step k, the predictor takes the form
Remark 6.1.3. The above controller can be implemented in neural network form.
Such a controller simply copies the weights of the feedforward predictor network to
generate the control input. This is shown schematically in Fig. 1 1 below.
/COPY
II
: CONTROLLER
I
I
Remark 6.1.4. Since the controller attempts to make the system track the desired
trajectory in one step, excessive control efforts are usually required. This makes the
approach unattractive in practice. To alleviate this problem, the weighted one-step ahead
19
controller is used, such that at each time step the control input minimizes the following
quadratic cost function
1 1
k 1) + -u(k
J ( k + 1) = - ~ ( k+ l ) T Q ~ ( + + l ) T S ~ ( +k 1) (55)
2 2
where the tracking error ~ ( +k1) = y ( k + 1) - r(k + 1). The weighting matrices Q and S are
required to be symmetric and positive definite. Substituting the expression for ~ ( +k1) and
y(k + 1) into the cost function and then performing the minimization produces
Setting the result to zero and solving for the control input yields
The above is known as a weighted one-step ahead controller in adaptive control literature.6
6.2. Model Reference Controller. A different way to avoid the requirement
that the system track a desired trajectory in one step is to use a control scheme known as
model reference control. Let the control law in Eq. (47) be modified as
Substituting Eq. (48) and Eq. (58) into Eq. (46) yields
The system response y ( k ) now no longer follows the reference input r(k) directly as in the
case in Eq. (49). Its behavior can be conveniently interpreted in terms of a reference
model. Define y,(k) as the response of a reference model when driven by the reference
input r(k),
20
and the tracking error E,(k) as the difference between the system response y ( k ) and the
reference model response y m ( k ) ,
I .
The equation that governs the behavior of this tracking error is obtained by subtracting Eq.
(61) from Eq. (60),
This requirement is easily satisfied since the coefficients y l , y2, ..., yp are the design
variables to be selected a priori.
Remark 6.2.1. The difference between this case and the previous case is that the
desired trajectory is not specified by the reference input r(k),but rather by the response of
the reference model. Since the reference model is known, the reference input r(k) that is
needed to make the reference model produces the desired response can be easily computed.
The introduction of the reference model is to slow down the convergence of the tracking
error so that excessive correction during the adaptation process does not occur.
Remark 6.2.2. The model reference control scheme can also be implemented in
neural network form. At any time step, the controller network copies the coefficients of the
predictor network.and uses them in the generation of the control input. The configuration
for this control scheme is shown in Fig. 12.
Remark 6.2.3. Equation (63) shows that the prediction error acts as a driving term
for the difference equation that governs the behavior of the tracking error. If the reference
model coefficients are designed such that the homogeneous solution is asymptotically stable
then the steady state tracking error is simply the particular solution of the difference
equation. One thus has the ability to affect the steady state tracking error through the
reference model coefficients. However, this freedom is constrained by the residual
dynamics of the prediction error that the steady state tracking error may be amplified or
reduced. Generally speaking, the natural frequencies of the reference model should be
placed away from those dominating the residual dynamics.
Remark 6.2.4. If the coefficients of the predictor model are updated at each time
step, then the controller coefficients must match those of the predictor model at each time
step. The resulting integration between parameter estimation and control computation is
known as model reference adaptive control. The adaptive scheme is summarized in the
following equations where the ordinary least-squares algorithm is used to perform the
parameter estimation step. Again, let Y ( k )denote the estimated coefficients of the predictor
model at time step k ,
21
7* SYSTEM
Y(k)
I I
II
/y COPY
- CONTROLLER 4
I
/y COPY
x'y
I I
I
.@
REFERENCE
MODEL
1
I
Ym(k)
starting with Y(0) as an arbitrary initial guess. The control input is computed from
where the reference model coefficients y l , y2, ..., yp are time-invariant and chosen a
priori. The above control input is applied to the system producing response y ( k + l ) . The
predictor coefficients are then updated according to the rule
staring with R ( 0 ) as any symmetric positive definite matrix. The newly estimated
parameters are then used to compute the control input for the next time step u(k+2).
Remark 6.2.5. The control schemes discussed in this section deals with a one-step
ahead predictor model of the form shown in Eq. (38). The previous section shows that a
two-step ahead predictor or a multi-step ahead predictor has the same linear form.
Therefore, the results presented in this section can be easily extended to these predictors.
22
For example, the two-step ahead controller will compute the control u(k + 2) requiring the
measurements upto y ( k ) only.
wherefl.) is some non-linear function of past input and output data. First, note that for the
response of the system to follow that of a reference model,
23
Therefore, at time step k + l , one wishes to determine the control input u(k+l) such that Eq.
(69) is satisfied. Since the relationship between y(k+l) and u(k+l) is non-linear and is not
known, one cannot solve for u(k+l) directly. However, if the non-linear system is such
that there exists a predictor of the form given in Eq.(44)such that jj(k + 1) = y(k + 1)then
one satisfies the following equation,
Y ( k + 1) = f ( y ( k ) , y ( k - I), ...
( u(k + l), u(k), u(k - l), ...)
= N ( Y ( k ) , Y ( k - 11, ..., u(k + l), u(k), u(k - l), ...) (74)
When a non-linear network of sufficiently large number of hidden layers is used, then it
may also qualify as an open-loop model of the non-linear system besides its being an one-
step head predictor. This is the fundamental difference between identification using a linear
network versus a non-linear network. Generally speaking, the theoretical advantage of
using a non-linear network for non-linear system identification is off-set by the difficulties
in finding such a network in practice. Neither the number of hidden layers nor the number
of neurons in each layer are known a priori. For a chosen network configuration, the back
propagation algorithm is often used to determine the network weights. Typically, the
convergence rate is slow and a large amount of data is needed. The back propagation
algorithm is well-known and discussed extensively in the literature.
In the model reference control problem, the theoretical advantage of a non-linear
network is somewhat diminished because the open-loop model need not be found for
purpose of tracking control. The model reference control scheme can accommodate a non-
linear network rather easily. Assume that the network representing the non-linear system
can be expressed in the form,
where e l ( k -t1) denotes the fitting error introduced with the separation of the u(k+l) term
from A'(.). The control input is computed from
24
where yl, y2,..., y p are the coefficients of the reference model representing the desired
response. The control input when applied to the system yields the closed-loop response,
The tracking error, &,(k) = y(k) - ym(k), where ym(k) is the response of the reference
model is governed by the difference equation
In practice, one identifies an approximation of N1(.) denoted by I?,(.). The control law is
then based on I?l(.) ,
Let ez(k + 1) denote the approximation error, e2(k + 1) = Nl(.)- I?,(.). The tracking error is
now governed by
25
C=[l.O -0:5 0.0 1.0 0.5 0.01 , 0~1.5
The system is excited by random input shown in Fig.13 producing the response shown in
Fig. 14.
4, 6,
-4 I J
0 4 8 12 16 20 0 4 8 12 16 20
Time (sec.) Time (steps)
Figure 13: Excitation input time history. Figure 14: System response time history.
Using the above time histories, the network weights can be identified using Eq. (31).
First, consider the case where p = 6, the following values for the network weights are
obtained.
The above results are checked against the data by performing an open-loop prediction of the
response using the input alone,
and an one-step ahead prediction (or observer estimation) using both actual input and
output data,
It can be verified that in both cases, both predicted responses match the actual data exactly.
Again, it should be emphasized that the result shown in Eqs. (83) represents a set of
weights that can be identified from any feedforward network that uses 6 past values of
input and output data to predict the current response. Specifically, if one uses a network
consisting of a single neuron, then the values listed in Eqs. (83) are precisely the weights
of this neuron. On the other hand, if a feedforward network consisting of several layers of
linear neurons is used to identify the system, then the values in Eqs. (83) are the weights of
a single neuron representation that is mathematically equivalent to the multi-layer network.
The system in Eqs. (82) in fact contains one uncontrollable mode as revealed by the
singular values of the controllability matrix, C = [ASB,A4B,..., A B , B ] ,
26
The model in a.
(84) is therefore an over-parameterizedmodel. The same system can be
modeled by using data from only 4 past time steps to predict the current response, i.e., p =
4. The corresponding weights are given below:
Note that the over-parameterization in Eq. (84) is in the form of having more distant past
input and output data to predict the current response, corresponding to the case of a neuron
having additional input channels. This is in contrast to the case where over-parameterization
is in the form of having additional neurons added to the network.
8.2. Model Reference Adaptive Control of A Linear System. Next, we
consider the application of the model reference adaptive control of the above system. The
goal is to have the system track a desired trajectory prescribed via the reference model,
where r ( k )= sin(k/21r). First, consider the ideal case where disturbance and noises are
not present. Since the system has a single output, the predictor network consists of only
one linear neuron. In this example, 6 past input and output values are used to predict the
current response. Recall that this is a case of over-parameterization since the effective order
of the system is only 4. The system is assumed to be unknown to the controller at the
beginning, and the weights are initially set to zero. Simultaneous prediction and control is
carried out producing the results shown in Figs. 15a-d below. Figure 15a shows that the
system response (dashed curve) quickly tracks the desired response (solid curve). The
time histones of the prediction error and of the tracking error during the process are shown
in Fig 15b and 15c, respectively. The control input time history is shown in Fig. 15d
revealing that the adaptive mechanism quickly produces the necessary control input to make
the system track the desired response.
4r----l
2
output 0
-2
-4 t J
0 4 8 12 16 20 0 4 8 12 16 20
Time (sec.) Time (sec.)
27
Tracking
Error
0.
-1.5 -2
0 4 8 12 16 20 0 4 8 12 16 20
Tme (sec.) Time (sec.)
Figure 1%: Tracking error. Figure 15b: Control input.
2
1
1
1.5 .
Tracking Control 0
-1.5
0 4 8 12 16 20 0 4 8 1 2 1 6 ,3
Time (sec.) Time (sec.)
Figure 1k : Tracking error. Figure 16b: Control input.
8.3. Identification and Prediction of a Non-Linear System. While it is
not possible to have a linear model that can reproduce the open-loop response of a non-
linear system, it is possible to have a linear predictor that can reasonably predict the non-
linear response. The predictor model uses actual input and output data to compute the
28
predicted response. Consider the system whose state space matrices are shown previously,
but the input and output are related by the following non-linear relationship,
--
2 2
output 0 output 0
-2 -2
I I I I
-4 -4
0 4 8 12 16 20 0 4 8 12 16 20
Time (sec.) Time (sec.)
29
linear system. Figures 18a-d show the tracking response, the prediction error, the tracking
error, and the control input time histories, respectively. Recall that the control method does
not require that the open-loop model be identified, but rather the predictor model that can
reasonably predict the response, which is the case illustrated in the previous example.
0.5
output -2 ~~~
-4 b
0 4 8 12 16 20
-1 [
0 4 8 12 16
I
20
Time (sec.) Time (sec.)
Figure 18a: Tracking response. Figure 18b: Prediction error.
(non-linear system)
3 2r
Tracking
Error
1.5
0
---I Control
Input
-1
1
-1.5 -2
4 8 12 16 20 0 4 8 12 16 0
Time (sec.) Time (sec.)
Figure 18c: Tracking error. Figure 18b: Control input.
When disturbance and noise are added to the system, the resulting behavior of the
system is shown in Figs. 19a-d. Again, this reveals a certain degree of stability robustness
of the adaptive scheme to possible disturbance and noises. This is due to the inherent
robustness in the ability of linear predictors that can predict the non-linear response.
4 1
2 0.5
output 0 Prediction
Error
-2 -0.5
-4 -1
0 4 8 12 16 20 Q 4 a 12 16 20
Time (sec.) Time (sec.)
Figure 19a: Tracking response with Figure 19b: Prediction error.
dist. and noise present (non-linear syst.)
30
Tracking
Error
0 .
-1.5 I
0. 4 8 12 16 20 0 4 8 12 16 20
Time (sec.) Time (sac.)
Figure 1%: Tracking error. Figure 19b: Control input.
31
This case requires additional theoretical treatment than that presented in this paper. Finally, \
the paper concerned mostly with stability rather than performance robustness issues.
Further work is required to assess this aspect of the problem.
9. References
1 Narendra, K.S. and Parthasarathy, K., "Identification and Control of Dynamical
Systems Using Neural Networks," IEEE Transactions on NeuraI Network, Vol. 1,
No. 1, March 1990.
2 Hornik, K., Stinchcombe, M., and White, H., "Multilayer Feedforward Neural
Networks Are Universal Approximators," Neural Networks, Vol. 2, No. 5, 1989.
3 Billings, S.A. and Leontaritis, I.J., "Input-Output Parametric Models for Non-Linear
Systems. Part 1: Deterministic Non-Linear Systems; Part 2: Stochastic Non-Linear
Systems, International Journal of Control, Vol. 41, 1985.
It
4 Chen, S., Billings, and Grant, P.M., "Non-Linear System Identification Using Neural
Network," International Journal of Control, Vol. 5 1, No. 6, 1990.
5 Hyland, D.C., "Neural Network Architectures for On-Line System Identification and
Adaptively Optimized Control," Proceedings of the IEEE Conference on Decision
and Control, Brighton, U.K., December 1991.
6 Goodwin, G.C. and Sin, K.S., Aduptive Filtering, Prediction, and Control, Prentice
Hall, Englewood Cliffs, New Jersey, 1984.
7 Ljung, L. and Siidertr6m, T., Theory and Practice of Recursive Identification, The
MIT Press, Cambridge, Massachusetts, 1983.
8 Chen, C.-W., Huang, J.-K., Phan, M. and Juang, J.-N., "Integrated System
Identification and Modal State Estimation for Control of Large Flexible Space
Structures," Journal of Guidance, Control, and Dynamics, Vol. 15, No. 1, pp. 88-95,
January-February 1992.
9 Juang, J.-N., Phan, M., Horta, L.G., and Longman, R.W., "Identification of
ObserverKalman Filter Markov Parameters: Theory and Experiments," Proceedings of
the AIAA Guidance, Navigation, and Control Conference, New Orleans, Louisiana,
August 1991; accepted for publication in the Journal of Guidance, Control, and
Dynamics.
10 Phan, M., Horta, L.G., Juang, J.-N., and Longman, R.W.,"Linear System
Identification Via An Asymptotically Stable Observer," Proceedings of the AIAA
Guidance, Navigation, and Control Conference, New Orleans, Louisiana, August
1991; also, accepted for publication in the Journal of Optimization Theory and
Applications.
REPORT DOCUMENTATION PAGE I F- Approved
0M8No.0~-0188
. AGENCY USE ONLY (leave Man&) 3. REPORT TYPE AND DATES COVERED
Technical Memorandum
I. TITLE AND SUBTITLE 5. FUNDING NUMBERS
On Neural Networks in Indentification and Control
of Dynamic Systems WU 585-03-11-09
.1 AUTHOR(5) I
Minh Phan*, Jer-Nan Juang, David C. Hyland**
Unclassified-Unlimited
Subject Category 39
I
The paper presents a discussion on the applicability of neural networks in the identification and control
of dynamic systems. Emphasis is placed on the understanding of how the neural networks handle linear
systems and how the new approach is related to conventional system identification and control methods.
Extensions of the approach to non-linear systems are then made. The paper explains the fundamental
concepts of neural networks in their simplest terms. Among the topics discussed are feedforward and
recurrent networks in relation to the standard state-space and observer models, linear and non-linear
auto-regressive models, linear, predictors, one-step ahead control, and model reference adaptive control
for linear and non-linear systems. Numerical examples are presented to illustrate the application of these
important concepts.