Error Analysis of Direct Methods of Matrix Inversion
Error Analysis of Direct Methods of Matrix Inversion
J. H. WILKINSON
National Physical Laboratory, Teddington, England
Introduction
1. I n order to assess the relative effectiveness of methods of inverting a matrix
it is useful to have a priori bounds for the errors in the computed inverses. I n
this paper we determine such error bounds for a number of the most effective
direct methods. To illustrate fully the techniques we have used, some of the
analysis has been done for floating-point computation and some for fixed-point.
In all cases it has been assumed t h a t the computation has been performed using
a precision of t binary places, though it should be appreciated t h a t on a com-
puter which has both fixed and floating-point facilities the number of permissible
digits in a fixed-point number is greater than the number of digits in the mantissa
of a floating-point number. The techniques used for analyzing floating-point
computation are essentially those of [8], and a familiarity with t h a t paper is
assumed.
2. The error bounds are most conveniently expressed in terms of vector and
matrix norms, and throughout we have used the Euclidean vector n o r m and the
spectral matrix norm except when explicit reference is made to the contrary.
For convenience the main properties of these norms are given in Section 9.
I n a recent paper [7] we analyzed the effect of the rounding errors made in
the solution of the equations
Ax = b (2.1)
( A + E ) x =- b + ~b (2.2)
and t h a t we could obtain bounds for E and 8b provided assumptions were made
about the ratio of the m a x i m u m element of the successive reduced matrices to the
m a x i m u m element of A. All the methods we discuss in this paper depend on the
(1) (2) (~)
successive transformation of the original matrix A into matrices A , A , • • • ,
(k) (8) (1) (k)
A such t h a t each A is equivalent to A and the final A is triangular. An
essential requirement in an a priom analysis of any of the methods, is a bound
for the quantity r, defined b y
(s)
R = maxla~3 I (s = 1, . . . , k ) (2.3)
max [ a~) I
* Received March, 1961.
281
282 J. H. WILKINSON
For methods based upon Gaussian elimination we have been able to obtain a
reasonable upper bound for general unsymmetric matrices only when pivoting
for size is used at each stage. This bound is far from sharp, but does at least
lead to a useable result, and further, the proof shows the factors which limit
the size of R. For a number of classes of matrices which are important in practice,
a much smaller bound is given for R, and this bound holds even when a limited
form of pivoting for size is employed. Finally, we show that for positive definite
matrices R ~ 1, even when pivoting for size is not used.
For comparison with methods of this Gaussian type, we give an analysis of
methods which are based on the reduction of A to triangular form by equivalent
transformations with orthogonal matrices. For these it is easy to show t h a t for
any matrix R < x / n , so t h a t control of size is assured. Because of this limitation
on the size of R, the error bounds obtained for orthogonal reductions of a general
matrix are smaller than those that have been obtained for elimination methods.
However, orthogonal reduction requires considerably more computation than
does elimination, and since in practice the value of R falls so far short of the
bound we obtained it is our opinion that the use of the orthogonal transforma-
tions is seldom, if ever, justified.
3. We do not claim that any of the a priori bounds we have given are in any
way "best possible", and in a number of cases we know that they could have
been reduced somewhat b y a more elaborate argument. We feel, however, that
there is a danger that the essential simplicity of the error analysis m a y be ob-
scured by an excess of detail. In any case the bounds take no account of the
statistical effect and, in practice, this invariably reduces the true error b y a far
more substantial factor than can be achieved by an elaboration of the arguments.
We also adduce arguments which partly explain why it is so often true in practice
t h a t computed inverses are far more accurate than might have been expected
even making allowances for the statistical effect.
All the a priori bounds for the error contain explicitly the factor [[ A -1 ]l • This
means that we will hardly ever be in a position to make use of the results before
obtaining the computed inverse. The nature of the analysis makes it obvious
how to use the computed inverse to obtain an a posteriori upper bound for its
error but the main value of the results is to indicate the range of applicability
of a given method for a given precision in the computation, and they should be
judged in this light.
A (n-2), . . . , A (1) ( s o that A (r) is a matrix of order r) and the modulus of the
pivotal element of A (r)by p~ so that
I a~;) I =< p , . (4.1)
Then we have
I d e t A (r) l = P ~ P ~ - ~ " ' P l (r = n , n - 1,...,2). (4.2)
On the other hand, Hadamard's Theorem gives
I d e t A (~)] =< [rp~]~ ( r = n, . . . , 2 ) (4.3)
since the length of every column of A (r) is bounded by r½p,. We write
p~p,-l'"p~_-< [rips] ~ (r = 2, 3 , . . . , n - - 1) (4.4)
but retain
p~p~-~ "'" pl = ]det A(~) ] . (4.5)
Taking logarithms of relations (4.4) and (4.5) and writing
log p~ = q, (4.6)
we have
r--1
~q<r
1 = ~ log r + (r -- 1)qr (r = 2, 3, . . . , n -- 1) (4.7)
< .1 log
. 213½4
. .t . (n 1
l)n--2 + 1
log [ d e t A (~) [ (4.10)
--2 n--1
q2 q3 q~-i
+2+3+'"+n-1'
giving
where
1
f(s) = [213½ . . . s~-~]½. (4.12)
Substituting for I det A (n) I from (4.3) (with r = n) we have
284 $. It. W I L K I N S O N
n nq~
q~..Fn q__21 =< l o g f ( n -- 1) + 2(n -- 1) log n + n - - ' - -1 (4.13)
l = max (a~))2
3 9=1
so that all columns of A (~) have lengths which are bounded by l, then
ql ~ l o g f ( n - 1) W Iogn Wlogl
2 (n - 1) (4.17)
= log f(n) + log l.
In particular if l =< 1 we have
pl <=f(n). (4.18)
The functions f(n) and g(n) increase comparatively slowly and it is easy to
show t h a t
f(n) N Cn il°g'~. (4.19)
In Table 1 we show the values of f(n) and g(n) for representative values of n.
The two forms of normalization
(I) ½ -_< max l a , ] _-< 1 (4.20)
9,$
(II) !2 =
< --,=la" -<_ 1 (j = 1,.--,n) (4.21)
will be used throughout this paper and will be called normalization (I) and (II)
respectively. The results (4.15) and (4.18) hold for any matrix normalized in
the appropriate way, and indeed their truth is in no way dependent on the left-
hand inequalities in (4.20) and (4.21). However, in our opinion, pivoting for
size is a reasonable strategy only if all rows and columns of the original matrix
have comparable norms.
For normalization (I) this m a y be achieved by scaling all rows and columns
so that they contain at least one element satisfying ½ ~ ] a , I < 1. For normali-
zation (II) this should be followed by an additional scaling of the columns so
that (4.21) is satisfied. F. L. Bauer has suggested that such matrices should
be termed "equilibrated" matrices. In practice it is unnecessary to ensure exact
E R R O R A N A L Y S I S OF D I R E C T M E T H O D S OF M A T R I X I N V E R S I O N 285
TABLE 1
n lO 20 50 100 200 1,000
Except for small r, this ratio is only slightly greater than unity. This means
t h a t the pivotal sequence pn, pn-1, " " increases quite slowly at first. Now
equality in (4.4) can hold only if all elements of A (~) are equal to ~ p , and all
columns are orthogonal. This is clearly not possible for some values of r, for
example, r = 3. Further if a,~.(r) = ::t:pr then all a (r-i)
, are ± 2 p r or zero. Hence
if at any stage the equality sign holds in (4.4) then either p,-z/pr = 2 or A (r-l) is
null, and in either case (5.2) cannot be satisfied.
These considerations suggest t h a t the least upper bound for the pivots m a y
be m u c h smaller than the one we have given. In our experience the last pivot
has fallen far below the given bound, and no matrix has been encountered in
practice for which pl/pn was as large as 8. Note t h a t we m u s t certainly have
p, _-< png(n + 1 - r) (5.3)
since p, is the final pivot in the reduction of a matrix of order (n + 1 -- r),
though if the m a x i m u m is to be approached b y pl then the initial rate of growth
m u s t be much slower than t h a t corresponding to (5.3).
and ~n"(~)is non-zero, because A (1) is positive definite. T h e elements of A (2) are
given b y
_(~) (~)
a~3( 2~) ¢b~ s -- m,lalj
a (~) a (~)
a(1)
_~3 _ 13
_(2)
- - ¢b3~
(6.2)
a(1)
11
~.(1) ~(1) ~1
~S X~ X s - - t,bll X l "Jr ~ X, k ~$ (1) / X~ X/
~1 ~1 ¢~2 t ~ l l • =2 j=~ all /
(6.3)
If A (2) is not positive definite there is a non-null set of x2, x3, • • • , x~ such t h a t
a ~ ) x , x , =< O. If we write
Xl = ~
•. +,1 ~
(x)
X~ , (6.4)
~2 t~ll
~ ( ' ).~
a~j ~ x s =< 0, (6.5)
t = l 351
Triangular Decomposition
7. For the symmetrical Cholesky decomposition of a positive definite sym-
metric matrix we may prove an even more satisfactory result. We show that
if A is positive definite and lanai ~ 1, then there exists a lower triangular
matrix L such that
LL r = A, (7.1)
[ 1 . ] _~ 1. (7.2)
A = [ "°
-~, - - 1 (7.3)
La ,aooJ
where B is a positive definite matrix of order (n -- 1) and a is a vector of order
(n - 1). By hypothesis there is an L satisfying
LL r = B, ] l , [ =< 1. (7.4)
If we choose 1 so that
L l = a, (7.5)
which can clearly be done since L is triangular and non-singular, then
and every l=, satisfies [ l~, I ~ 1. Equation (7.6) and (7.7) now give the appro-
priate decomposition of A and all elements of the triangular matrix are between
± 1 . A consequence of this is that an L L r decomposition m a y be performed in
fixed-point arithmetic. Incidentally we have proved that the sum of the squares
of the elements in any row of L is less than unity, and this result is used later.
288 J.H. WILKINSON
Hence
4,> ~ [I _41> + Ira,, I1 ~4,
Zla,, l< a,X
_-<
(1)
l a,, +Lal,41) l ~ l m , l L (8.5)
(1)
< ~21 -4" + lal)) i -- {a,, I-
Further
I _42)
~,, i>_ {a.(1) [-Im~,ila.(i) L
=> ~[a,,(1) I -- (1 -- ~}m,l})[ "41'~I,
{
(8.6)
(1) 41) (2)
E R R O R A N A L Y S I S OF D I R E C T M E T H O D S OF M A T R I X I N V E R S I O N 289
The diagonal terms therefore dominate in exactly the same way in all A (~) and,
from (8.5), the sum of the moduli of the elements of any column of A (~) de-
creases as r increases. Hence
(8.7)
(i)
=< max < 2 max [ a,,-(1)I.
l a,~ I =
and I det A (~) [ will usually be much smaller than p ~ if AC~) is ill-conditioned.
The behavior of the special matrices may give the impression that "partial"
pivoting for size is as effective as "complete" pivoting. The following example
shows that it may not be adequate. Consider matrices Bn of the type:
[+i 0 0 0 +'11
B:L;11 +1_1
+
--
+1
+I
+1+1
--1
0
+10
+I
0 -
.1/
+lJ
If we use partial pivoting, then the pivot is always in the top left-hand corner
and the sequence is + 1 , + 1 , - . . , + 1 , ( - - 2 ) ~-1. If complete pivoting is used
290 J . H . WILKINSON
(i) IIAATII=IIA115 .
(J) n max,,, l a,, 1 > II A II --> max,,, l a,, I.
(K) IIAII=<IIAII~ = [~ a,~]~,
2 For max X(AA r) ~ trace(AA T) = II A lie2.
(M) HA N --< III A III where I A I has elements I a,~ I • This follows from (B).
(N) If lA I =< B, then l[ A 1[ _<- Ill A Ill --< II B II.
(O) HI A Ill =< nt II A II • For NAII => n-½ II A lie and Ill A Itl --< III A II[E = II i liE.
T h e following results are f u n d a m e n t a l in the error analysis and will be used
repeatedly.
(P) II A II ~ max I M(A)[ (from (B)).
(Q) If ][ AII < 1, then I + A zs non-singular.
F o r X,(I + A ) = 1 + X,(A), i X , ( / + A) I ~ 1 -- I X~(A) I (9.1)
=> 1 - - l i A i t > 0 .
H e n c e det ( I + A ) = l I X , ( I + A) ~ 0. (9.2)
ERROR ANALYSIS OF DIRECT METHODS OF MATRIX INVERSION 291
f[ A If
(iii) H(I-bA) -*]] => 1
1 - II A I1"
These results follow from the identity
I = ( I --k A ) - I ( I --t- A ) , (9.3)
the existence of ( I W A) -1 following from (Q). Writing ( I -{- A) -1 = R,
I = R+RA (9.4)
1 =[[II[ a ]]R[[ -- lIRA II a [[R]] -- llRII IIAll (9.5)
which gives (i) since (1 - [[ A [[) is positive. (9.4) then gives
[[I- R]l = ]IRAi[ ~ IIRI] [[d[I (9.6)
which gives (ii) ; from (ii) we deduce
[[ I II -- II R II < li A ]] (9.7)
= 1 -II All
II A - 1 N II E [] (9.13)
II(A + E) -~ - A -~ II/li A - 1 II --< 1 - - II A -~ I[ II E ]l"
We shall refer to the expression on the left as the relative error in ( A + E ) -1.
Note that the bound we have obtained for this relative error contains II A-~ ]]
itself as a factor, while the bound for the absolute error contains ]1 A -~ l]2.
Whether this represents the true state of affairs depends critically upon whether
(9.12) is a reasonably sharp inequality.
The example in Table 2 illustrates the last point. We give there an ill-condi-
tioned matrix A, and show the effect on the inverse of two different perturbations
E~ and E2, in A. These perturbations are of the same order of magnitude, but E~
results in a perturbation in the inverse which is of order II A-i ]lel[ El [[ and the
second, in a perturbation of order ]] A -~ [[ [[ E2 ]l. Perturbations E will behave in
the same way as E2 whenever E can be expressed in the form
E = AG or E = GA (9.14)
where G is of the same order of magnitude as E. We shall see later t h a t such
perturbations are encountered in the error analysis. In applications of (S) we
will normally restrict ourselves ab initio b y some such condition as
II A-I }l 11E II < 0.1
since we will not otherwise be able to prove that the computed inverse is a reason-
able approximation to the true inverse.
TABLE 2
+. 45000 00000-
A = [[1 O00OOOO000
[_ .7777777800 •70000
; A -I = 109
F
L.+. 38888 88900 --• 50000 00OOO
Y - A -~ = A-i[P + Q + R] (9.18)
and taking norms of both sides. Provided all A -~ II and (bll A -~ II ÷ e) are appre-
ciably less than unity, (i) and (iii) show that Y is a good approximation to A -1.
In applications P is always present in (9.17~ but Q or R or both may be absent.
Again we stress that the result will be pessimistic if I] A-~ I] ]] P I] is a poor
approximation to [1 A-1p 1] and this will happen if P may be expressed in the
form A S where S is of the same order of magnitude us P (a normalized A is,
of course, implicit in this remark).
in the notation of [8]. From the result given there for an inner product, we have
ur,x,(1 ± e2-t) + Ur,,+lX,+l(1 ± O(n ÷ 1 -- r)2 -t)
0
÷ u~,~+2x~+2(1 ± 0(n ÷ 1 -- r)2 -t) ÷ . . .
÷ u . . . . ~x~_~(1 ± 04.2 -t) ÷ ur,~x,~(1 ± 03.2 -t) (10.3)
--= b ~ ( l ± 02-t)
where each 0 denotes a value, ordinarily different in each factor, satisfying
lel=<l.
294 J. H. WILKINSON
1b, I]
b~
I ~bl _-<2-' b3 ~ 2-'1 b l (10.5)
b4
0
]un[ 31u151]
51u12 [ 51u131 41u141
1~21 41u231 41u~41 31u~l|
[ 6U I < 2 - ' I l ussl 31u~41 31u~51| = 2 - i V (say), (10.6)
l u4,1 21 u451|
L ] us~[/
we m a y write this in the form
I Ux -- b l < 2-'[[ b [ + Vl x l]. (10.7)
For the solution of the n sets of equations
UX = B 10.8)
we have, for the computed X
[ U X - - B] < 2-'[[BI + V[X]] (lO.9)
and hence
then we have
II v II ~ I1 y II~ (10.17)
<= [ 1 + 3 ~ + £ + . . . + (n+l)~] ~ -<_0.6n 3/~ (n_>_ 10).
The result for the inverse is
II X - V -~ [1 ~ ( 0 . 6 ) 2 - " : l l V - ' IIV(1 - (0.6)2-"~ll U -~ II). (10.18)
I1. The corresponding results for fixed-point arithmetic are rather less defi-
nite because they depend on the details of the arithmetic facilities and on whether
we have a priori information on the size of the solution. For the sake of definite-
ness we assume that scalar products can be accumulated exactly in a double-
precision accumulator and that division into this double accumulator is possible.
These facilities are provided on all desk machines and m a n y high speed compu-
ters.
If we know in advance that the computed solution lies between 2 k-~ and 2 k,
then we may write
x = 2ky (11.1)
and solve
Uy = 2-kb. (11.2)
The variable y, is determined from the computed y,+l, • -. , y, by the relation
where fi means that we take the fixed-point result of evaluating the expression in
brackets. The numerator is accumulated exactly and the only error is the round-
ing error in the division. (We have tacitly assumed that intermediate values of
the scalar product do not require the introduction of a larger scale factor than is
necessary for the final value. On ACE this difficulty is avoided quite simply by
accumulating exactly 2 -p times the scalar product, where 2 p is greater than the
largest order matrix t h a t can be dealt with. Overspill in the scalar product can-
not therefore occur and the rounding errors introduced, 2 v-~t, are far below the
level of the others. We shall not refer to this point again.) Hence
Uy ~ 2-kb + ~b (11.4)
[~b~ [ =< ]u,~ 12-t-~ < 2 - H (11.5)
giving
Ux = b -+- 2~b. (11.6)
Applied to the solution of U X = B, we obtain an X satisfying
UX~B+C (11.7)
where
I C. [ < max Ix. 12-' II x 112 (11.8)
]] C H ~ n2-t[[ X H- (11.9)
When B = I, we have
UX = I + C (11.10)
where (11.9) is satisfied, giving
I1X -- U -1 ]l --< n2-tll U-~ I[~/( 1 - n2-'tl U -~ ]1). (11.11)
In this we have tacitly assumed that k ~ 0. If k < 0 no scale factor is necessary
and (11.9) becomes tl C II < n2-t-1. For the inversion problem k must neces-
sarily satisfy k -> 0, because one element of X is ( 1 / u , , ) and this is not less
than unity. The strictest application of (11.5) gives
variables x, b y x ~ "+1) (i = r + 1 , - . . , n ) so t h a t
x~ T+I) = 2k'+ly~r+'); [ y~+~) { < 1 (i = r --k 1, "." , n) ]
(12.1)
[ y~,+l) I ~ ½ for at least one i.
T h e next step is the c o m p u t a t i o n of x~ ") with the introduction of an increased
scale factor 2 k" if necessary, and we h a v e
UrrX~) q_ u~,.+ixr+~
(,+1) + "'" + u r . x .(,+1) ~ b. + ~,, I ~r I =
< 2k'-t+l . (12.2)
Matrix Multiplication
13. F o r m a t r i x multiplication in floating-point we write
C ~- f l ( A B ) . (13.1)
298 z.H. WILKINSON
Hence afortiori
<2-t
=
I41a211 41a2~1 31a231
4ta~l , 4laa2 I 3{a,, I
41 a41 1 4] a42 I 31 a43 I
2]a~4 | . I B I .
2[au,|
21 a44
(13.2)
a_(3)
,) ~ a ,(~) (~)
j - - m ~ 2 a ~ ) -9 e,~ (14.1)
(~) _ (i--i) ~ ~(*--1) (1)
a,; ~- ~j ~ :../t.l.i--l~z--l,.~ -9 e,.1
where e~ ) is the error made in computing a,,-(~-1) - m~.~_la~k_~] from the computed
a(k-~)
,; , me.k-1 and ak(~_~] . The element a,,(') is an element of the ith pivotal row
and undergoes no further change. Summing equations (14.1) and cancelling
common elements,
(ii) i > j. We now have the same equations as (14.1) terminating with
(:) ~
a~j ('-" -
a,~ m , , j - i a ~o-~)
- i , ~ JF e ~(:) . (1~.3)
The next value, a °+~)~, is taken to be exactly zero and remains zero for the rest
of the reduction. The element a~, (:) is used to calculate m,j from
aO)
" (14.4)
(~33
Note that equations (14.2) and (14.7) give a relation between elements a~:) of
the final triangular matrix, elements m,,, and elements a~2 with added perturba-
tions. Writing L for the lower triangular matrix formed by the m,, augmented
300 J. H. WILKINSON
by a unit diagonal, and U for the upper triangle formed by the pivotal rows,
(14.2) and (14.7) give
where E (k) is the matrix formed b y the e,`/(~).Note that this has null rows 1 to
(/c -- 1) and null columns 1 to (/c -- 2).
We m a y describe this result as follows. " T h e computed matrices L and U are
the matrices which would have been obtained by correct computation with
A m -t- E. Further A (k) is the matrix which would have resulted from exact
computation with
A m + E (2) W E (a) + . . . q - E (k) = A m + F (~) (say)." (14.9)
[_(i)a~ [ < g(n----)l_ 2-t I g ( 1 ) --~ 2g(2) + g2g(3)(n) -{- "'" '~ g(n)J , (15.1)
then
Hence
a(k)
~3 a(k-1)
(15.4)
=< 2 - t [ g(k) +g(n)
g(k-- 1)],
showing that, under our assumption (15.1), all elements of (A + E) are less
than 1 / g ( n ) .
Since we are assuming that all a~ ) are less than unity, it is clear that
fl r[a,i(k--l) -- m,,k-,ai~-7~l -- [a~-' --
k--1
m~,k--lak--l,j]
~
<
2--t.
(15.5)
2-t
i ooo!l
Hence a bound for the matrix E (k) is, typically for n = 5, k = 3,
0
0
1
1
0
1
1
0
1
1
(15.6)
~0 1 1 1 1_
and for E is
01 01 01 01 01i
2 -t 1 2 2 2 2 . (15.7)
1 2 3 3 3
1 2 3 4 4
Using the Euclidean norm as an approximation we have
II A -1 II 1 - o,41n~z~2 -' II a -1 II
1 --ff 0.34nm2 - ' II A-1 ]1
Hence
II BZ -- III = II A X - z li ~ (0.78) 2-'rim [1 A-1 II
(16.11)
(0.78)2-'nmg(n)[[ B-1 II .
The presence of the factor g(n) is inevitable, but notice that it should be
replaced by max [1, R] where R is the growth ratio, in each particular case and
we know in practice that R is often less than unity. For each of our special
types of matrix we have an a priori upper bound for R which is much smaller
than g(n).
Fixed-Point Computation
17. We do not wish to repeat the analysis for all varieties of fixed-point com-
putation and we restrict ourselves to one particular type. Before describing
this we justify its selection.
The alternative methods of solving unsymmetric equations for which an error
analysis has been performed are
(i) by reducing the problem to that of inverting A A r, a positive definite
matrix [3],
(ii) by orthogonal trriangularrizations of different types (§21-23).
None of these methods requires less than 2n 3 multiplications, whereas the
method we have just described requires n ~. If we are prepared to use 2n 3 multi-
plications, then the following strategy gives high accuracy.
(I) Invert the matrix as above, thereby determining the interchanges and the
required scaling factors.
(II) Repeat the computation in fixed point taking advantage of the informa-
tion gained in (I) and accumulating scalar products wherever possible.
In (II) we determine L and U directly from L U = A, where A has its rows
and columns permuted as determined by (I). We should compute ½L in case
some of the multipliers were just greater than unity as a result of the different
rounding errors. For this technique we have
from (11.8)
UX = Y + G (17.2)
I g,. I ~ 2-' max
./,./
I x,~l :1 I gr~ I ~ 2-' max
z,.~
(y,)
i
l
]le[I ~ n2-']]X]l : ]]GII =< n2-'[[ VII
ERROR ANALYSIS OF DIRECT METHODS OF MATRIX INVERSION 305
giving
II E X II ~ 2-~n II i II
NF ]l ~ (0.9) 2-tn~ HA-1 II (17.3)
I[ LG H =< (0.9) 2-tn2 II X [I II n v II ~ (0-7) 2-tn8 II A - ' II
provided 2 - t n 2 I] A-~ ]1 =< 0.1. (The constants have been taken to one decimal
only.) In both cases we have results of the type (T) of Section 9 with
a = 2-t[n + 0.8n ~] a = 2-tn
(17.4)
b = 2-t[0.9n 2] b = 2-t[0.9n 2 + 0.Tna]
In practice the second of these is very uncommon since the largest element of
X is usually larger than the largest element in Y, particularly if A is at all ill-
conditioned. For several of the special matrices we have already seen that
n--1
max~.~ I Y~J I = max~,~ I l~x I is equal to unity. In any case max,.~ [ y~, ] -_< 2 , so
that for small matrices II Y II cannot be large.
Hence we wifl usually have
In the alternative case we will not be able to guarantee that X is even an ap-
proximate inverse unless n32 -~ ]] A -1 II --< 0.1, and then we have
From (18.1) and our assumptions, A -q- Er. is a positive definite matrix with
elements bounded by unity.
Consider now the determination of la+l,r+l. The equation from which it is
computed is
ls+l,r+l
L l.+1,.+1 ] (18.4)
= q (say).
If this quotient were computed exactly it would be the (s q- 1, r + 1)-element
corresponding to the exact decomposition of (A q- Er~), and since this is a posi-
tive definite matrix with its elements bounded by unity, the exact quotient is
bounded by unity. Hence the computed l~+i,.+1 is also bounded by unity and
l.+1,,+1 = q -t- e; I e I =< 2 - H . (18.5)
Hence
/8+ml.+m + /8+1,2/r+1,2 -4- " " + 18+1,~+11r+1,.+1~ a8+1,.+1 --t- l.+l,.+le (18.6)
This completes our induction apart from the element/~+1,,+1 • This is computed
from the relation
18+1,,+1 "~- ~/a8+1,8+1 -- 2
18+u -- l2
~+1,~ . . . . . l~8+1,8 = ~ (say). (18.7)
Again if the square root were computed exactly it would be less than unity, since
it is an element in the exact decomposition of A q-- E8+1,8 which is positive definite
with elements bounded by unity.
We do not wish to enter into a detailed study of square root subroutines.
ERROR ANALYSIS OF DIRECT METHODS OF MATRIX INVERSION 307
where the three terms on the right come from the calculation of L, X and Y
respectively. By using fixed-point arithmetic in the calculation of X and Y we
can reduce the n 2 in the last two terms to n, but if A is very ill-conditioned, so
E R R O R ANALYSIS OF DIRECT METHODS OF MATRIX I N V E R S I O N 309
that [[ A -~ ]] is large, the errors made in the triangularization are in any case the
most important, and those made in the calculation of X X ~ are the least.
Instead of solving L T X = I we could solve L X = I and then calculate X T X .
The bound obtained for the relative error is identical in the two cases, though,
of course, the computed inverses are not identical.
The Residual M a t r i x
20. In most of the error analysis we have worked with the residual matrix
(AY -- I ) . I t is worth noting that we can obtain a far better bound for this
matrix if we solve L T X = I than if we solve L X = I. Further, the contributions
made to the residual matrix from the separate parts of the computation are
not in the same ratio as the contributions to the relative error.
Working with L r X = ( I + F ) we have
AY-- I = A [ X X T + G] - I
= (A +E)XX T-EXX r-I-AG-I
= LLTXX T-EXX r+AG-I
(20.1)
= L[I+FF]X r-EXX r +AG-I
= L[I+F][L -I+FrL -1] - - E X X r+AG-I
= L F L -~ + L F V L -~ -t- ( L F F r L -~) -- E X X r + AG.
The order of magnitudes of the bounds for the norms of the components arising
from E, F, G respectively are
1.3n2-~ ]] A-~ I], 1-5n22-t I1AII t }] A-1 II and 1.3 II A II n22-' 11A - ' II. (20.2)
The terms containing F and F r make the same contributions, and that of the
F F r term is negligible. Note that the contribution of G to the residual may
easily be the largest even when it has the smallest effect on the inverse itself.
On the other hand if we use
L X = I + F, Y = XrX-t- G,
then F and G are bounded as before but
A Y -- I = L L r F r ( L - 1 ) T L -1 + F
(20.3)
-t- L L r F T ( L - 1 ) r L - 1 F -- E X X r + AG.
where
S, = RiTR2 T . . . R, r. (21.11)
Since all the R , , and hence all the S , , are orthogonal, we have
[A1-F N]X ~ I + P 4- Q'~
s--1
II N ll =< ~ II g, ll
1
s--1
(21.12)
Il P ll ~ ~ IlK;I]
1
]]QII = I I M ] I .
(B) We may work instead with/~r and assume that we have bounds for Fr and
Gr of the form
II Fr [I ~ f(n, r)2-'; II G, II =< g(n, r)2-'. (21.13)
Premultiplying (21.9) now by/~-1/~-1 . . . /~:_11we have
where
S, = k~-1/~7t . - . /~:1. (21.15)
Now the S, are no longer exactly orthogonal, but suppose we can show that
]] Er 11 ~ a < 1. Then we have
since R;-t is exactly orthogonal. Hence we have a relation of the same form as
(21.12) in which:
S--1 1
11N II -<- ~ (1 - a) ' II F, II
8--1
1
I1Q II =< (1 - a) '-~ II M 11
312 J. H. WILKINSON
I
i a2~ a23 a24 a25/
0 a33 a~ a~ I (23.1)
a42 a4R a44 a4s ]
a52 a53 a54 a55_J
ERROR ANALYSIS OF DIRECT METHODS OF MATRIX INVERSION 313
The next step is to make the (r + 1, k -t- 1)-element zero by a rotation in the
(r + 1, k -t- 1)-plane. This modifies only the elements in rows (r --~ 1) and
(k -t- 1) and leaves the earlier zeros undisturbed. We shall assume that the
Euclidean norms of all columns of A and B at this stage are bounded by
(1 ± 9.2-t) p,
(b 2 + C2) ~ 1 + 3e6
which is certainly less than 2 -t. In replacing this b y zero we are committing an
error with a bound which is smaller than those we have obtained for other ele-
ments of rows (r + 1) and (k + 1). Combining these results we have for the
current F and G,
nfn--1)
II F ll < 2 - 2-' (1 + 9.2-')
(23.11)
n (n--l)
I]G]] < 2 V ~ 2 - ~ ( 1 + 9 . 2 -t) 2
Summing these results and taking ~ 2.3/2,
s½ -~ ~r ~ s3/2 ~. vr
2 5,3, we have
from (21.12) and (21.17),
(Ai + N)X ~ I + P + Q (23.12)
t n (n--l)
1 4V/2 n 5/2 2 -t (1 + 9.2- )
I[ N I] < n(n--i) 5
(1 -- 6.2 -t) 2
• t n(n--1)
1 n ( n -- 1) 2 V ~ 2 - , (1 + 9 . 2 - )--~---
[[ P ]1 < ~-') 2 (23.13)
(1 -- 6.2 -t) 2
1
II Q II < ~(~-1) IIM II
(1 - 6.2 -1)
n (n--l)
where we have replaced all 1/(1 -- a) * by 1/(1 -- a) --V--" We write for the
moment
t n (n--l) 1
(1 + 9 . 2 - ) ~ = l+d; ~(~-1) = l + e . (23.14)
(1 -- 6.2 -t) 2
Now M is the error made in solving the final triangular set of equations
A~X = B,, (23.15)
ERROR ANALYSIS OF DIRECT METHODS OF MATRIX INVERSION 315
so that we know from (10.10) and (10.17) that the computed X satisfies
A~X ~ B.+M (23.16)
where certainly
II M II ~ 2-~[n~ll B, II -4- 0.6(1 + d)n21I Z ill
(23.17)
< 2-71 + d)[,~ + 0.6,~11X ill
since the norm of each column of A. and B. is less than (1 -t- d). We have there-
fore
F4~/2 5/2
II A X - I II < 2-'(1 + d) (1 -4- e) [ _ - - ~ n ]l X II
(23.18)
+ -v'.qn ~/~ + n + o.6n ~ II x I1.]
-1
which is a result of the standard form. For a reasonably good inverse we shall
require
2 - i n 5/2 < 0.05 (say)
and assuming n >= 10 this certainly gives
(1-4-d) < 1.1
(l+e) <1.07 (23.19)
(1 "4- d)(1 -4- e) < 1.2.
We shall require further
2-tn51211 A-~ I] < 0.1
and if we assume this we certainly have when n > 10
I] A X -- I ]] < 2-~[1.61n5/2[[ X II + 1.71n5/~]• (23.20)
From (T) of Section 9 we obtain
l / z - A - ' l/
II a - ' II < 2-'n 5/2 [2.1 [[ A -1/I -4- 1.8]. (23.21)
The result is for a matrix with normalization (II). For normalization (I) we
have immediately, in the usual way,
accumulate (b 2 + c2) exactly and then scale it b y the largest 22k for which
22k(b2 -4- c2) < 1. We then calculate cos 0 and sin 0 from 2~b and 2%. The com-
puted matrix is rather closer to an orthogonal matrix t h a n for floating-point
computation and further, only one rounding error is made in computing each of
(x cos 0 + y sin ~) and ( - x sin 0 + y cos 0). The gain in accuracy is at best
equivalent to a small constant factor and we cannot gain b y a factor of~/n.
Householder's Reduction
24. Only (n -- l ) transformations are involved in the Householder reduction.
At the rth step we multiply b y a matrix of the form (I -- 2WrWrT) where wr r
is of the form
(O0...Ox~xr+l...x~); ~x, 2 = 1 (24.1)
and this produces zeros below the diagonal in the rth column. I t leaves unchanged
the first (r -- 1) rows of A and B and the first (r - 1) columns of A, so t h a t
(n - r + 1) 2 elements of A are modified and n ( n - r + 1) elements of B.
We are forced to round these elements before going on to the next stage so t h a t
we have no hope of having sharper bounds for F, and G~ t h a n
IIFrll ~ ( n - r + 1)2-t-1; IIGrll --< % / n ( n - r + 1)2 - H . (24.2)
This gives for N and P,
(INll < ~1 n2~--t--1
~ ; I l P l l ~ ~2 n2~--t--1
~ (24.3)
even if we take a, defined on p. 311, to be zero. The most we can hope to gain
over Givens therefore is a factor of .v/n, and perhaps a small constant factor,
if we compute (I - 2w~wrr)A~ and ( I - 2w~w~r)B~ with sufficient care.
If we work in conventional floating-point arithmetic, we do not gain even the
~,/n factor. As before it is easy to see t h a t the departure from orthogonality is
slight and therefore u n i m p o r t a n t and we m a y concentrate on the errors made in
computing Ar+~ itself.
We assume t h a t we have an exact w~ and estimate the errors made in computing
A~ ~w~w~A~ The first stage is typical so we deal with A~
-- 0 T .
2wtw~rA~.
We have
p~ ~ fl(w~A~) ~ w~TAi + q~ (say) (24.4)
and ql~ m a y be written in the form
qlk : xlaikel Jr x2a2kE2 -3F " ' " ~'~ Xnanken (24.5)
where
levi --< (n + 1 -- r)2 -t.
Now ~ x~~ = 1 and ~ , a 2~ = 1 (the latter persisting approximately at later
stages), but we cannot conclude anything sharper than
I q~k I < n 2-t]] q~ [[ < n3/22-t. (24.6)
E R R O R A N A L Y S I S OF D I R E C T M E T H O D S OF M A T R I X I N V E R S I O N 317
A2 = A1 -- KululrA1
ul r = (an :=i:: S, a2~, . . . , a ~ l ) ; S 2 = a~l + a~l2 + . - . + a,1
2i (24.7)
K = 1 / ( S 2 ± allS) where ~ S has the sign of a11.
This form avoids unnecessary square roots, but does not lead to a lower error
bound.
In fixed-point arithmetic we can indeed realize the gain of the factor nn½, but
only on a computer with good arithmetic facilities would this be worthwhile.
To ensure an accurate determination of w, S 2 should be accumulated exactly
and then multiplied b y 22k where/~ is the greatest integer for which 22kS2 is not
greater than zero. The vector w can then be calculated using 2ka~l instead of
a ~ . This gives a very accurate determination of w and we need only concern
ourselves with the errors made in computing (A1 -- 2wlwlrA~). In forming
pl = fi(wlrA1) we can accumulate scalar products, and since the columns of A1
have norms bounded by unity, the elements of w~rA~ are less than unity. If we
write
pl -~ fi(w~rA) ~ w~rA + q~, (24.8)
then all elements of q~ are less than 2 -t-~ and hence II q~ I] ~ x / n 2 - H . Although
the elements of 2w~w~rA~ may be greater than unity, we know that those of
A~ -- 2w~w~rA1 ure not, and therefore no harm is done by the overspill of the
former. To be safe throughout, we must scale A originally so that its columns are
sufficiently less than unity for those of A~ to be less than unity throughout.
With sufficient care we can obtain a computed inverse X satisfying
already seen [(17.7) and Section 9, (T)] that if we are prepared to perform Gaus-
sian elimination twice we can expect an inverse satisfying
[I X -- A -~ {]/H A-~ [] < (2.0)2-tn2RII A -~ {{.
This is more favorable than that for Householder, even when the computation
is carried out with extreme care, provided "R", the pivot ratio, is less than ~v/n.
In our experience this is almost invariably true and there are many important
types of matrix for which we know in advance that it will be true.
Comparing Gaussian elimination with Householder when both are performed
in floating-point, we have a relative error of 2-tnmrH A -~ ]] for the former and
2-tn3{{ A-1 H for the latter. In our view the presence of the factor n m springs
rather from the exigencies of the analysis than from any true shortcoming, and
the case for the Householder is slight.
The other method for which we have an error analysis is that of Goldstine and
yon Neumann based on the inversion of the positive definite matrix, A A r.
However, this requires 2n 3 multiplications and has ]] A -1 ]]2 in place of }] A -1 {{
in error terms. If A is at all ill-conditioned, this method has nothing to recom-
mend it.
and have then replaced the e, by their upper bounds, (n - i)2 -t. Now for these
bounds to be attained not only must all the individual rounding errors have
their greatest values but the a,b, must have a special distribution.
On the other hand the term [{ A -1 ]l which invariably occurs in the relative
error is not usually subject to any such considerations. I t is useful to have some
simple standard by which to judge the relative error and this is provided b y the
following considerations. Consider the effect on the inverse of perturbations of
order 2 -t in the elements of A. We have
We have therefore
The total effect of all the rounding errors made in the process of solution is far less
than those which come from the initial truncation.
There are two main phenomena which account for the remarkable accuracy
of the solution.
(i) We saw in Section 18 that L L r = I1 + E and that a typical element of
E is bounded by 2-t-11 l,~ I. If some of the elements [ l~ I are much less than unity,
the bounds for the corresponding elements of E are far smaller than 2 -t. Now
because B is so ill-conditioned, the l,, are small and become progressively smaller
with increasiug i. The matrix E is displayed and it will be seen that its elements
also become progressively smaller as we move to the bottom right-hand corner.
Here they are far smaller than ½.10 -8 (the optimum value which we might ex-
pect for 8 decimal computation). Note that L L r is so close t o / [ precisely be-
cause/~ is so ill-conditioned!
We now have to consider the effect of the errors E, on the inverse. The per-
turbation resulting from a change e in the (i, j)-element is
--e (ith column of/FF-1) (jth row o f / t - 1 ) / ( 1 + e l-l- -,1 );
now the denominator is approximately unity and the ith row and 3th column of
/1-1 increase generally in size with increasing i and 3. This means that the small-
est elements of E are in precisely those positions which have the most effect on
the inverse. The result is that we do not get the full effect of the ]]/1-1 ]] in the
computed inverse. The effect we have just noticed is very common in ill-condi-
tioned matrices with elements which are of a systematic nature. I t becomes even
more marked if we consider Hilbert segments of higher order. The presence of
very small elements in E is also quite common in ill-conditioned matrices which
are less obviously of this type.
(ii) The second phenomenon is of a more general nature and is associated
with the inversion of triangular matrices. We have seen that for a matrix A of
general form, the effect of perturbations in the end figures is to introduce a rela-
tive error which is usually proportional to ]] A -~ [I, and an absolute error which is
proportional to I] A-~ ]12.Such an effect is avoided only if the rounding errors are
peculiarly correlated. Now whereas for general matrices such a correlation is
very uncommon, it is so common for triangular matrices as to constitute the
rule rather than the exception (Sect. 28). In formal terms we have, quite com-
monly, for the computed inverse of a triangular matrix,
II Z - A -1 II < a ( n ) 2 - t (27.3)
II A-1 II
in contrast to the more common result for a general matrix,
÷7~7~
I+
oo
÷ ÷
÷lT÷l
' ~-7 7-
Illl
II
I++ +~ I+
+÷
+
÷
322 ~. H. WILKINSON
which is greater than one in the last figure retained in the computation. This means
that the errors made in the triangular inversion are negligible and we already
know that the errors involved in computing X X T are negligible. The combined
effect is to produce an inverse of quite astonishing accuracy. The total effect
of all the rounding errors is far smaller than that corresponding to a single per-
turbation of ½-10 -s in the (5, 5) element of/~.
Even for very high order matrices this gives a very good result. It shows that if
we use floating-point arithmetic, then even the small elements have a low relative
E R R O R ANALYSIS OF D I R E C T M E T H O D S OF M A T R I X I N V E R S I O N 323
error. Matrices of this kind are produced when Gaussian elimination is performed
on the matrices derived from finite-difference approximations to elliptic partial
differential equations.
When complete pivoting for size is used then we can prove a much weaker
result which is, however, independent of the sign distribution of the elements of L.
We show that if I l~, ] ~ I1, [, then for fixed-point computation with a fixed
scale factor,
I~. - z , t --< 2'+~-' max I x~ 12-~. (28.5)
k
The proof of this is immediate by induction, since we have, typically for the first
column,
2~ ~ -fi[(/n~n + 1~2:~21 + "-" + l . . . . ~:~r-L1)//~]
(28.6)
=-- --[(/r1:~1~ + Ir2~2~ + . . . + l . . . . x~r-~.~)/lr~] ± 2 p-t-~
ness in the last pivot since it means t h a t this last pivot has a high relative error.
In our opinion there is no substance in this belief. The emergence of a small last
pivot shows t h a t the matrix is ill-conditioned, since whatever pivoting m a y have
been done, the reciprocal of the last pivot is an element of the inverse matrix.
On a computer on which scalar products can be accumulated, it is usually better
to use a pivotal strategy which results in a very small last pivot t h a n one which
does not. In the example of the last section for instance it is better to work with
the matrix as it was, rather than working with the pivots in positions 4, 5, 3, 2
and 1 in t h a t order, in spite of the fact t h a t the former leads to the smaller value
for the last pivot. If scalar products cannot be accumulated, then there is a very
slight advantage in a strategy which does not make the last pivot very small.
I t should be realized t h a t the h a r m t h a t comes from a small last pivot is pri-
marily a property of the matrix and is not something induced b y the method of
solution. When scalar products are not accumulated exactly, the perturbation in
the original element of A which corresponds to the last pivot is the sum of
(n -- 1) rounding errors, whereas t h a t in the element corresponding to the rth
pivot is the sum of (r - 1) rounding errors. Ideally we should t r y to make the
last pivot correspond to the element i, j for which the product of the norms of
the ith column and the j t h row of A -1 is as small as possible. There does not seem
to be any practical strategy which assures this.
The triangular matrix of Section 27 does not belong to any of the classes we
nave mentioned so far. We now describe a much larger class of matrices for which
our result is true. Consider the solution of a lower triangular set of equations in
fixed-point arithmetic, using a constant scale factor 2 k. Let .X be the computed
inverse and X the true inverse. We have
L l~ l~ ..- l~,J
(28.10)
7
VI Xll I
IXEI ~ IXllEI • t E [. (28.11)
LlxllL Ix221 ..•
ERROR ANALYSIS OF DIRECT METHODS OF MATRIX INVERSION 325
giving
[ X E [ < C2 k-t-1
!i 1 1 (28.12)
I
|
L
" '011' '1.0'
'0.1' '1.0' I
'0.1' J
(28.14)
where we use 'a' to denote a number of the order of magnitude of a, but with
end figures which will lead to rounding errors. Now the inverse of this matrix
contains an element of order 10 ~. If n is at all large it is therefore very ill-condi-
tioned. However, it is obvious in the light of our previous analysis that the ele-
ments of the computed solution have a very low relative error.
An attempt to construct a matrix for which the result is not true reveals how
widespread the phenomenon is likely to be. An example is the matrix
- 1 0 -10 .9 --.4 ]
.9 ~0 ~ o j (28.15)
The last column of the exact inverse is [0,-~ 101°, 101°]. The last column of the
computed inverse is [4 X 109; 4444444444; 10 l°] SO that we really do have a
computed inverse with a relative error of order II A-1 I1 X 10-1°. A matrix of
low order must be very specially designed if it is to give such a result. If we have
used complete pivoting, then we cannot have such examples in matrices of low
order.
Selecting three matrices at random of order 18, 25 and 43 from moderately
ill-conditioned unsymmetric matrices inverted on DEUCE, we found in all cases
that the inversion of computed L and computed U and the computation of
L-1U -1 made practically no contribution to the errors in the inverse. This shows
the great importance of keeping the difference ( L U - A ) as small as possible
326 J . H . WILKINSON
since this is usually the main source of error. If partial pivoting for size is used
it is easy to ensure that I ( L U - A),~ ] < R2 - H where R is the pivotal growth
factor [7]. I t does not appear to be possible with complete pivoting, without
using twice as much storage as in the ordinary reduction or alternatively doing
the computation twice.
It is natural to ask why this accuracy of computed inverses of triangular
matrices was not revealed by our earlier analysis. The reason is that the analysis
is based on an assessment of [] X - L -1 [[ derived from a bound for [[ L X -- I II.
Now the norm of ( L X -- I ) is not significantly smaller in the cases when X
has the exceptional accuracy than in cases when it has not. The size of the resid-
ual gives a very crude estimate of the accuracy of X and the range covered b y
approximate inverses having residual matrices of a prescribed size is remarkably
wide when the original matrix is ill-conditioned. We are able to put the residual
matrix to such good use when analyzing computed inverses of general matrices,
only when the methods used are such as to bias the residual in such a way as to
be almost as small as possible having regard to the accuracy of the computed
inverse.
The really important norm is not that of ( L X -- I ) = E (say) but t h a t of
L - ~ E . Now it happens that for triangular matrices, E can usually be expressed
in the form L F where F has much the same norm as E. This gives L - 1 E much
the same norm as E itself. We can now see why it is that the progressive intro-
duction of scale factors in the inversion of a triangular matrrix usually gives a
better result than that in which the same scale factor is used throughout, al-
though the latter gives the smaller residual. We may illustrate it b y taking an
extreme case. Suppose no scale factor is required until the very last stage, and
this requires quite a large scale factor, 2*. The computed values x~, xn_~, • • •
x~ will then have to be rounded to k less figures to give x , / , x : - i , • • • , x2 p with
[x~ - x / I --< 2*-H. The contributions to the residuals corresponding to this
change are precisely L ( x - x').
Now although this may constitute the major part of the residual as far as
magnitude is concerned, it is obvious that the solution of L y = L ( x - x ' ) is
(x - x'), whereas if the other part of the residual is E1 the solution of L y = E~
is L - ~ E i and m a y be very large. The increased size of the residual is entirely
misleading. On the other hand, because the early stages have been done without
the scale factor, the residuals just before computing the last element x~ are
bounded by 2 - H I l,, I instead of 2k-*-~[ l,~ [.
When the residual matrix R, defined by ( A X -- I ) , is used to assess the error
in an approximate inverse X, then we can, in general, make no claim unless
I] R [[ < 1. Thus Goldstine and yon Neumann [3, p. 1080] say that unless
[I R [[ < 1, even the null matrix gives a better residual. Although this is true it is
reasonable to regard a matrix X, for which HX - A - ' [I/[[ A-1 [I (< 1, as a
much better inverse than the null matrix. Indeed if a normalized matrix A has
an inverse such that ]] A -111 is of order 102°, then if X is the matrix obtained
by rounding the elements of A -1 to 10 significant figures, the absolute errors in
some elements of X may be as large as ½ 101° and the residual matrix m a y have
ERROR ANALYSIS OF DIRECT METHODS OF MATRIX INVERSION 327
F+i
0 0 0 0 +1~
+ +1 0 0 0 - 1[
+1 +1 0 0 +1[ (29.1)
-I +1 +I 0 -- I|
+1 -1 +1 +1 + 1[
-1 +1 -1 +1 - 1 J|
The inverses are of the form illustrated in (29.2):
- '
+2-2 __2-3 +2 -4 _2-5 +2-5~
i+• + 2 -1
0
+ 2 -2
+ 2 -1
_2-3 + 2 -4
+ 2 -2 --2 -3
_2-4|
+ 2 -3 |
A;1 = i 0 0 + 2 - ' + 2 -2 --2-2| (29.2)
0 0 0 2 -1 +2-' /
L+2 -~ __2-2 + 2 -3 __2-4 + 2 -5 __2-5J
useless. Significant figure arguments are much too superficial in general to give
reliable indications, and reflection back into the original matrix is a much more
reliable guide.
Xl
fi [.88824 - (.81321).7 7 _-- r "318993] = .33116. (3o.3)
k .96326 J fi [_ .96326 .J
ERROR ANALYSIS OF DIRECT METHODS OF MATRIX INVERSION 329
Acknowledgments
The work described here has been carried out as part of the Research Pro-
gramme of the National Physical Laboratory and this paper is published by
permission of the Director of the Laboratory. The author wishes to thank G. E.
Forsythe and E. T. Goodwin for reading the manuscript and making many
valuable suggestions.
REFERENCES
1. BAUER,F.L. Sequential reduction to tridiagonal form J. Soc. Indust. Appl. Math. 7
(1959), 107.
2. GIVENS, W. The hnear equations prob|em Technical Report no. 3. Applied Mathe-
matics and Statistics Series, Nonr 225 (37). Stanford University, 1959.
3. GOLDSTINE, H. I I . ; AND VON NEUMANN, Z. Numerical inverting of matrices of high
order. B~dl. Amer. Math. Soc. 53 (1947), 1021.
330 J.H. WILKINSON