Basic Linear Algebra
Basic Linear Algebra
ALGEBRA
CEMALKO~
Department of Mathematics
Ankara
1996
VVeeks Sections
1.5,
4) 18 22 Cont: 1.5, 1.6
5) ')
_.') 29 Cont: 1.6. 1.7
6) April .) Cont: 1.7, 2.1
,-i
7) (. - 12 ') ')
By Prof.Dr.Cemal I\o<:;
.•
BASIC LINEAR
ALGEBRA
CEMAL KOC,
Department of Mathematics
Ti'c" "
. J\EJ\0~,
Ankara
1996
/
1. MATRICES AND SYSTEl\fS OF LINEAR EQUATIONS ........... 1
Chapter 1
MATRICES
AND
I l l SCALARS
~~=---~----~-------------------------------------------------
I
_j
(i)
theory the primary scalars to b0 us0d <11"<' 0 and 1 with binary addition a11d
mult.ipliration. For this rt'ason, we arr' ololig<'d to g<'neralize the concept of
scalar by iutroducing fi<'lds.
Definition 1.1.1. Dy, a Jilld we nwan a s<'l F together \Vit.h two operations
+ and · (usually ralh'd O]H'rations of addition and mu.llip/i('(l/ion) for which
the following properti<'S hold:
2. There exists a.u <'lelll<'llt 0 ( calkd the ::rm clrmcnt) snell 1h:d
a +0 = 0 +a =a for all a E F.
:3. For each element o ofF there 0xis1s an <'lcmen.t -a (call<'d I h<' addilit·e
im•crse of a) in r snell that
a + (-a) = (-a) + a = 0.
8. There exists an cl<'IIH'n1 1 (called 111<' unity or identity dement) such that
a· I =1·a = a for a 11 a m F
9. For each a different from th<' ZNO element 0 thNe exists an element a- 1
(called the multiplicative inl'rT.<~r' of a) such that
In our investigations in linear al~Phra we always fix a field and we call its
<'lements scalars.
Note that the properties list<•d ahove are natural properties of real and
complex numbers and thereby thos<' who are not familiar with other examples
of fields may be contented with considering scalars to be real or complex
numbers.
Example 1.1.2. Natural exampks of fields arc the following sets together
with the usual operation of addition and multiplication
+ 0 1 • 0 1
0 0 1 0 0 0
1 1 0 1 0 1
Similarly one can deftne Z 3 , Z 5 and in general Zp for any prime number p
which are the most elementary examples of fields.
181 EXERCISES
I. fn each cas<' ht•low dl'tPrmin<' wlwt l11•r t l1e given sd is a fi('ld or not.
(a.) The set of nonnPp;ative real nnmiH•rs Ulllh'r usual addition and nJul
tiplication.
(e) Tht• set of odd ini<')!;NS nnd<'r usual addition and nndtiplicat.ion.
(d) The set of rational numlwrs with odd denominator undPr usual
addition and multiplication.
2x + 3y + z = 4, 2;r + y- z = 0, x- y- z = -1
(j
llfJ MATRICES
A matrix is a rectangular array of clements taken from a field of scalars.
To make it precise we giw thl' follo\\'ing
into the field F. The image A(i.j) of (i,j) under A is called the (i,j)-entryof
A and it is usually cll•notPd hy .. \ij and A itsdf is denoted by
is called the ilh mw of J, and for any 1::; j::; n then X 1 matrix
Amj
is called the j-th column of A. The cntriC's A 11 , A 22 , .•• , Au, ... arc called
entries on the main diagonal.. \\'lH•n n = m. the matrix is said to ])(' square.
7
[-~ ~ l
is the 3 x 2 matrix which is denoted by the rectangular array
-1 0
-2 2 -1
[
1 0 1
we have
. .
o(z,]) = 0;3· ={ 1 if i =j
0 if i ::/= j
is referred to as Kronecker delta and the n x n matrix I defined by
1 0 0
0 1 0
I=
0 0 1
\ot<' that two matricP:-; .\and lJ an' <'qual ifl'tii('S<' I\\'O functions ha\'(' th0
sallll' domain and th<' s;tlll~> \·alll('s: in otll('r words
A = 11 ~ i\;j = /J;; for i = 1, 1, ... , m; j = I. 2, ... , 11
T
.I (i.))=.lji·
ii'J EXERCISES
i +j if i 2: j
1 if i =I= j
l
2. For each matrix below find the indicated value:
(b) A= [ -~ ~ ~
3 1 -1
l A12 + A21 + As2 =?
3. Find a if
[ x+y y+z]
z+t t+x -
[1 a]a .
3
5. For
10
G. Prm·p that (:IT j"l' = .·1 for a11y 1na1 rix :I.
1. Tl/( nodal inr·irh /IN 1/l(f/ri.r of" a diri'C11'd gr;q>ll ronsistinp; of" ~!IIIII' \'<'!"
tin's \'1. \~ ..... 1·i···· a11d Sllllll' f'dg1'~ (diri'Ctl'd pa1hs) r 1 .r~ ..... r 1 , ...
ir (j It'll ,.('S I.
I
·"ij =
{ -- I
()
if
ir
(
(.
J
J
. I'll t I' rs
<I lid I.
I
I .i
(a) v 3 (h)
/1e~
F 4 \' s \' 2
(~}~~
\. I
l
() -I
(ii) Ski'lrli II"' g1·a ph n>IT<''P"" d; 11g I" [ () -I
-I -I () ()
11
OPERATIONS ON MATRICES
Now we shall define some operations to produce new matrices out of some
given matrices and scalars.
Nate that this is in accordenc0 with the usual definition of the sum of real
valued functions, namely it is ess0ntia.lly in the form
[~ -~ ~ :l
1 -1 1 4
+[
-1
~ 1 :
0 1
=~4
l = [:
0 -1
~ ~ :
2 8
]·
The properties of the operation of addition on pnxn are very natural and
is given in
(2) For the m X n matrix 0 wl1ose entries are all zero we have
12
(3) For any m. X n matrix A the matrix -A defined hy (-A )ij = -A;i
satisfies
,.\ + (-A)= (-A)+ A= 0.
-A is called t.hc tugalit·(· of..-\.
(4) We have
Pmof: The proof is st raight.forward. For example to prow tl1e fl rst. prop
erty it is enough to V<'rif.'· that (A+ (13 + C'));j = ((il + IJ) + C')ij and it is
obta.hwd a.t once hy using 1h<> definition of addition:
(cf)(x) = cf(:l:)
..-\ = [ ·I
:J
0 2 3] c = -3
1 5 3 '
0 -1 2 8
JS
we have
r
-3A =
-0
-12
0
-3
-6
-15
-9]
-9 .
0 3 -6 -24
Exan1ples 1.3.3. a)
A= [
l I
.:
For tlH' matrices
-2 :1
.~
and IJ =[
:l:
z
11
y
t
v
l
we have
AJJ =[:
·1
-~ l[·: ~ l
:
·) II 1'
[
x - 2.:
:r+z+·tt
<Lr
+ 3u
+;:; + 5n
y - 2l
•!y
+ :"11·
'IJ+I+v
+ t + 5v
l
.
we have
From these examples w0 nhsPrw that 1\JJ and JJil may not be equal. r·1•rn
more one of them may not fJ( dr.fincd, and the pmduct of non.::cm ma I ric('H
need not be nonzero.
Now, consider three matrin•s ,\.JJ and C snch that the product (.\TJ)C is
defined, then since_,\[] is dPfined A and B must be m X nand n X r ma1rici'S
15
r r n r n
n n r r n
showing that (AB)C = A(BC). Thus we S('<' that if the product (AB)C is
defined, so is A(BC) and the equality A(RC) = (AR)C holds, that is matrix
multiplication is associative. In a similar way we prove distributive laws and
that the identity matrix acts as identity. So the following theorem has been
established:
Theorem 1.3.4. If A, B and C arf' matrices such that left sides of the
following are defined then right sides are also defined and the equalities hold:
1. (AB)C = A(BC).
2. fl(B +C)= AB +ftC and (.1 + /J)C = AC + BC.
2
-]
.I= 0
1
1
\V<> ran compute this product as (i\ll)C or il(DC). Not(' that tl1<' quicker
choic<' is ,\( /JC) and it gin's at once
2 2 10
1][_~]~~
-I -I -!i
,-\(TJ(') = (} ( [ :~ ·I 0 [!i] = 0
!")
;)
b) For A ~ [~ ~] n ~ [~ 0 ]
()
and (' = [ 0 0]
] 0
com
A II ~ ~ ~
[ ] "'"I AC ~ ~ ~
[ ]
and lwnc<'
Theorem 1.3.7. If A and Ban' matrices over the same field we have
(c) (ABf = BT AT
(d) (AT)T =A
where each Af;j has 1'i rows and 8j columns and each N1.:1 ha.s 8k rows and (]1
columns. Then for each pair ( i, j) the matrices
19
are defined and each of the entries of P;i is obtained by fixing a row of all
Nf;. 'sand a column of all N.j's, thus they constitute al1 entries of the product
AB. So we see that AB is partitioned by Pij 's, that is to say we have
AB= B=
R,
B11 Blr
B21 B2r
AB = [ (' 1 (''2 ... Cn ]
Bn1 Bnr
21
11@1 EXERCISES
1. Given
~
!;
----------------------------------------------------- ~
22
4. Th<' following is a list. of stat<'rn<'nts about squar<' rnat ricPs. In each case,
<'it h<'r show that the stal.('lll<'nt is ~<'IH'rally trne or find sp<'dfic matrices
for which th<' statellH'Ill is not t rll<':
(d) AB=0=>11:1=0
5. C:iven
-1
:J
l and X= [
:l't
.?·"
:ra
l
.
6. Given
A =[
-1
.j
G
-:3
1
1
2 -2
0
l
If B is a matrix v-:ith rows R 1 , 1? 2 , R 3 expr<'ss the S<'cond row of ;lfl in
terms of rows of B.
2:1
7. Compute ;F - A+ 3! for
9. Compute A 6 for
11. Given
-1 0 X y ..;,
0 -1 1l v 1l'
A= 0 0 1 0 0
0 0 0 1 0
0 0 0 0 1
A= [-I 0
(~ l
to show that A 2 =I.
24
=[
is to say AlJ = -11.1. Compn1<' (il + B) - A 2 - B 2 •
2
l
1 - 1 anti-commut<', that
E 11 =[ 1 0
0 0
l· F., 2 =[ 0 1
0 0
l, £ 21 =[ 0 0
1 0
l, En = [ 00 0
1
l
25
A. DIAGONAL MATRICES
A square matrix A for which Aii = 0 for all i ::J j is called a diagonal
matrix. So a diagonal matrix is one whose entries off the main diagonal art>
all equal to zero.
then we write
-1 0 0 0
0 1 0 0
0 0 0 0
[~ ~],
0 0 0 2
I
26
respectively. Note that any 11 x n zero matrix and identity matrix are diagonal.
When all diagonal <'111 ri!'s of a diagonal matrix are equal it takes the form
d
d 0
/)= = dl
0 d
d- dl
B. TRIANGULAR l\1:\TRTCES
27
For example
and [
-1 2l
0 0 .JI
f2
[ 4 0
1 -1
1 :1nd
l 0 0 ]
0 -1 0
0 0 3
are lower triangular. The last milt rix and in general any diagonal matrix is
both upper and lower triangular.
D. ECHELON T\[ATHICES
Let A be an m. X 11 matrix. The first non-z<>ro C'lltry (if any) of tb(• ith row
of ;1 is called the lcadiny (1//ry of the ith row. If the leading entries (if any) of
the rows of A arc of tlH' form
(ii) each leading entry is the unique non-zNo entry of its own column.
For example
0 0 0
000 0 0 0
000 0 0 0
[
~ 0 0 0
0 0 0 0
iitl EXERCISES
4. Let E;j denote tlw n X n matrix whose (i,j)-entry is 1, all other entries
are O.Prove that if an 11 Xn matrix A commutes with E;j ,then A,; = Aii
31
,and all other entries of its i-th column and all other entries of its j-th
row are zero.Use this to prove that if A is an n X n matrix commuting
with all n x n matrices , th('ll A is a sca.lar matrix.
-1
4
-7
-~ l'
0
(b) [ ~ ~~
-3
-
-7 0
l' (c)
0 -1
1
-1 -3
-1
0
2
1
3 -2
0
-4
1
4
0
y+z-2
-1
1
-1
1
z+:r-3
l.
9. Show that for any square matrix A the mat rices A+ AT, AAT and .r1? A
are symmetric and A- AT is skC'w-syrnnwtric.
10. Prove that every real or complex square matrix A can be written
uniquely in the form
A=S+I\
3.
'
~
A= [5-2
~
3]
-~ 0
11. Prm·e that the set of symmetrk mat rir<'s is closed und£>r <ultlit ion and
sra lar multiplication and that it. is not dosed under matrix mull i plica
l ion.
12. Fiud the number of2x3 cc:hdon mat riel's over Z 2 which ha\'(' two ]Pading
<'ntries. How many of them are row-r<'duced'?
Type I. Add c times_row ito row j for .~orne pair i i: j. (Notation: cRi + R1)
Examples 1.5.2.
(<1) [ -~
1
-1 0
3 -1
1 4 -2
3
l (-2)R 1 + Ra
-:
[-~ -1
10 -1
1 3
0
-2
-:]
2 -3 6 2 -3 6
-1 5 1 R2 <-+ Ra 0 1 -2
-!
(b)
0 1 -2 ~ -1 5 1
2 -5 1 2 -5 1
34
2 -3 6 2 -3 6
0 1 -2 5R3
0 1 -2
(c)
-1 5 1 -5 25 5
2 -5 1 2 -5 1
-~
~l u ~l [-~ :l
1 -1 -1
-1
(1 )Rt + R3 2R2
"2
(d) [
-----+
1 2 2
2 -3 6 -2R3 0 1 -2
(e)
0 1 -2 (-l)Rt + R4 2 -3 6
-1 5 1 Rt +-+ R2 2 -10 -2
2 -.5 1 0 -2 -5
Theorem 1.5.3. r:\'('ry <'lnwnt.a.ry row operation£: F"'x" - rmxn has
an inverse of the sam<' t~·p<'.
R; R; R;
rR; + Ri -cR; + Ri
-----+
Ri cR; + Ri R·J
Similarly the inverses of R; ,...... Rj and cR; a.re R; +-+ Rj a.nd c- 1 R; r0spectively:
35
Ri Ri
R-
1. f-+ Rj Ri R; - Ri
-:
Ri Ri
Ri:
[ R; I~ [ I,-, I cR; R, [ R,
0
2 G 0 -2
0 -2 2 6
and
-1 2 -2
2 1 0 -5
are row equivalent to one a.notlwr, IH•cause
-2R3
(-1)Rl + R4
2 6 R1 f-+ R2
0 -2
0 -2 --;. 2 6
-1 1 +---- 2 -2
2 1 R1·f--)- R2 0 -5
(1 )R1 + R4
-1/2R3
36
(c) If .-1 is row eqniYalPnt to B and JJ is row equivalent to C then .-l is row
equivaknt to C.
Proof: (a) Ohviou~ly <'arh matrix ..:1 is row equivalent to its<'lf, namely
the relation is refl<•xiw.
and
then
C - "
- (.._, e • • "
(•p• • • • ,. " (A).
(•:!L•l
This theorem allows us to usc th0 phrase "A and Bare equivalent" w]tl'll .-l
is row equivalent to B. Now we shall see that each class consisting of equivalent
matrices contains a row reduced Prlt<•lon matrix.
Example 1.5.8. Find a. row rt'dnced echelon matrix which is row equiv
.JB
alent to
0 -1 :3 -2
0 0 -:~ 1
0 2 -:3 4
0 1 -2 1
0 1 -!} 2
Dy applying the indicat<>d row opNations W<' obtain
0 -1 :3 -2 0 -1 3 -2
(2)U 1 + R:1
0 0 -:~ 0 0 -3
(I)R 1 + H..1
0 2 -:3 .J 0 0 3 0
(l)R 1 + Rr.
0 1 -2 () 0 1 -1
0 -!} 2 0 0 -2 0
() -1 :3 -2
(l)R~+Ra
0 0 -3 1
(Itnn:! + u..,
0 0 0
( -2j;3)R~ + R5
0 0 0 -2/3
() 0 0 -2/3
0 -1 3 -2
(-2j:3)Ra+ H.·t 0 0 -3 1
2/:H?a + R., 0 0 0 1
0 0 0 0
0 0 0 0
0 1 -3 2
(-l)R1 0 0 1 -1/3
(-Ij:l)R2 0 0 0 1
0 0 0 0
0 0 0 0
39
0 1 -3 0
(1/:l)R3+ R2 0 0 1 0
(-2)R3+Rl 0 0 0 1
---+ 0 0 0 0
0 0 0 0
0 1 0 0
0 0 1 0
(:1)R:! + R1
0 0 0 1
0 0 0 0
0 0 0 0
ELE/\IENTARY MATRICES
Lemma 1.5.10. For any matrix A and any elementary row operation [
we have
£(.-1) = [(I)A
40
Proof: The result can be seen imnwdiately by writing out the Nplality in
terms of rows R 1 , R 2 , • • ·, Rm of A and r 1 , r 2 , • ··,em of I for a specifir t :·pe of
operation. For example, if[. is the operation cR; + Ri then by using Example
1.3.7 we get
R; f'; e;A R,
£(A)= and f.(J)A = A=
cR;+Rj ce;+ej cc;A+eiA r·R;+ Rj
and the C'qnality follows. ThC' same' argument appliC's to other t.ypC':-;.0
n= P ..i
whC're
Proof: \Ve use induction on r,the nnmlH'r of operations. Let us prow th{'
particular case first. For r = 1 we haw f.d..l) = [ 1 (/)A and its validity follows
from the F)· '---------~
Assuming the equalit:·
The 'i'quality
[ 2 3 -1 06]
A= 1 -3 1 1 8
0 2 0 0 1
find a row reduced echolon matrix R and a product P of elementary matrices
such that R = P A. Reducing the partitioned matrix [Ali] we obtain
A=
[ 2
~ -3
3 -1 0 6 1 0
2
1 1 8 0 1
0 0 1 0 0 ~ l- [~
-3
1
9
1
0
-3 -2
1
0
8 0
1/2 0
-12
1
0
1 -2
1/; l
~ [~
-3 1 1 8 0 1
1 0 0 1/2 0 0 0]
1/2
0 -3 -2 -33/2 1 -2 -9/2
~
f~
~
0 1 1 19/2 0 1
3/2]
--+ [ 1 0 0 1/2 0 0 1/2
. 0 0 -3 -2 -33/2 1 -2 -9/2
INVERTIBLE 1\-IATRICES
and
(ii) A is invertible.
Pmof: We shall give a cydic proof: (i) :::> (ii) :::>(iii):::> (iv) :::> (i).
and using Example 1.3.7 we see that the i-th row of I is R;n- 1 = 0./J- 1 = 0
which is nonsense. So whPn A is invertible it cannot be row equiYaiPnt to a
matrix with a zero row.
(iii) :::> (iv): It follows from Theorem 1..5.7 that A is row equi\·aiPnl to a
row reduced echelon matrix R, and this matrix has no zero row by hypothesis.
So it has n leading entries R 1j,, R 2h, · · ·, R,j,. on its n nonz<'l"o rows. Since
R is row reduced echelon we have 1 ~ j1 < j 2 < · · · < jn ~ n and t !tis
is possible iff 1 = jll 2 = j 2, · · ·, n = jn. Thus tho loading entries or R are
R 11 = 1, R 22 = 1, · · ·, R,, = 1 and all other entries are necessarily Zf'ro, that
is to say R = I. Hence .4 is row equivalent t.o I.
We write down the partitioned matrix [Ali] and apply flcmentm·y rmr op
erations to nach mw reduced echelon matrix. Then either we ob/(Jin a ::rro
mw in the first pad at sonu· step and conrlude by (iii) that A is no! inlYrl
iblc or we reach the identity malri:r and conclude by {iv) that ;l is inl'erliblr
45
whose inverse is the matrix appra ring in the second block of the reduced matrix.
Symbolically we have either
Proof: =?:If A and B are row equivalent by Corollary 1.5.11 we can write
B = P A where P is a product of dementary matrices and thus invertible.
-¢::: If B = PA for some invertible matrix P, by the fourth statement
of the above theorem P is row equivalent to I, say P = fr · · ·£2 £1 (!) then
B =:= fr · · ·£~£1 (/)A and hence hy Corollary 1.5.11 B = fr · · ·£2 £1 (A) showing
that A and B are row equivalent. 0
Proof: Lemma 1.5.7 a.sserts the existence part and so it remains to prove
the uniqueness. Let A be row <'quivalent to the row reduced echelon matrices
R and R' with columns J( 1 , • · ·, /\'" and J(~, • · ·, J(~ respectively. From tran
sitivity it follows that R and R' arc equivalent and therefore by the above
·corollary R' = P R for some inwrtihle matrix P.Writing Rand R' in terms of
their columns this.yields:
(b) K, = 0 K; = 0 and h<>nre the first nonzNo columns (if any) of the
row reduced echelon matrices R and R' are the same.Thus this column
46
IS
0
= PE 1 = !\'
••
0
Further if the columns of Ih(' first k leading entries of Rand R' are identical,this
means that
0
0
1
1 0
0
E1= = PE 1 F.'2 = 0 p ~~-2 Ek = 1 = PEkl
0
0
0
0
satisfies
PI\= K = p- 1 I\·.
It follows from this thaI if one of our row n'd need echelon matrices has an
other column Ek+l contai11ing a leading entry the other one must also contain
47
Corollary 1.5.19. Two m x 11 mat rices are row equivalent iff they are
row equivalent to the same row reduced echelon matrix.
Proof. If A and B are row equiva.lent to the same row reduced echelon
matrix ,from transitivity, they are obviously row equivalent. Conversely assume
A and B are row equivalen(By the existence part of Theorem 1.5.18 ,A is row
equivalent to a row reduced echelon matrix R and B is row equivalent to a
row reduced echelon matrix R'. Tlwn W(' obtain
namely, A is row equivalent to both Rand R' .The uniqueness part of Theorem
1.5.18 shows that R = R' ,that is to sa.v .tl and B are row equivalent to the
same row reduced echelon matrix.
2
-1
1
2
are row equivalent matrices. vVhen we r<'d uce them we see that they are both
48
1 -2 0
[ 0 0 1 3
5]
0 0 0 0
119j EXERCISES
1. Let
0 0 1 3 5
0 -1 1 1 2
A=
0 2 -1 1 1
0 1 0 2 3
Find an echelon matrix E which is row equivalent to A.
1 -1 1 3 1
1 2 -1 1 0
A=
0 3 -2 -2 -1
2 0 1 7 2
c) R2 + R1
f) R2 ...-+ R3
l
5. Let
A= [ -: =: ~ -:
Find elementary mat ric('s F- 1 , ••• , Ek and a row redun•d <'clwlon matrix
R such that
R = Ek · · ·E1A.
[: -: l, [~
-1
a) h) -1
-2
:r
2
2 -1 1 -1 1 2
1 -1 2 2 1 -1 1
c) d)
0 1 -1 -;3 1 1 -1 1
1 2 -1 0 1 0 1
7. Determine wheth<'r
-1 1 2 1
1 -1 1 1
3 1 -1 1
1 1 2 -1
8. What must be k if
-1 k 2
2 -1 2k- 1 5
-1 1 1 -3
1 -1 k 4
is invertible. Find the inverse when k = -2.
9. If possible, write
[
1 -1
-~ -~ -~
0 l
as a product of elementary matrices corresponding to elementary row
operations of type cR; + Ri? Is it possible for elementary row operations
. of type cR; or R; .,..... Ri?
11. Let
~ -~ ~ .
1 l
Solve the equation
AX=B.
l
12. Let A be an invertible matrix with
A- 1 = [ ~ -~ ~
1 -2 .t
and B = [ -~ ~ ~
-1 -1 1
]·
52
2 :-1 1
A= [ 3 1 -1
-1 0 1
in the form
A= P
R
l
14. Prove tl1at
A~ [ _;
2
-2
2 0 -1
.] 0
4
-2 2
,: l and n =
[ -1
0
1
-2 0 1
2 1 2 -15
0 1 3 4
15. Prove tl1at if A= PH. whH<' R is a row reduced eclwlon matrix with r
leading entries t IH'n t h<' first r <:olumns of Pare the sa.nw as columns of
A wrresponding to rolnmns of le;:.ding 0ntrics of R.Use this to wmpute
Pin
[ (~ ~ -~ ~ ~]
I 0 2 1 0
=PR,
Two systems are said to be equivalent if they h aYe row equivalent augmented
matrices [AlB] and [rl'l B']. If this is the case arrording to the results of the
previous section we have [A'IB'] = P[AIB] = Q[A'IB'] for some
and [;\j B]
invertible matrices P and Q. Therefore, we have A' = PA, B' = P B, A = QA'
and B = Q B' showing that
x+y=2 x-y=2
and
X+ y::::: -2 X- y::::: -2
have the same solution set but they are not equivalent.
The fact 1hat equivalent systems have the sanw solutions allows us to reduce
systems to si m pier ones to determine their solutions. Given a system of linear
equations we reduce its augmented matrix to row reduced echelon form (or
simply to eriH'Ion form) whose system is of the form
Jl,.jr:t_'jr + ••• = br
0 = br+l
0=0
0=0
The method where the system is r('(ltH·<'d to a system with an echelon matrix
(or row reduced echelon matrix) i~ calkd GAUSSIAN elimination (or GA USB
JORDAN reduction.) Looking at 1h<' reduced system above we can get all the
information about solutions to th<' syst<'m:
(a) If br+l # 0 the equation 0 # br+l has no solution and hence the
system is inconsistent.
(b) If br+l = 0, we can expr0ss :rh, xh, · · ·, Xjr,( called basic variables) in
terms of other variables, that are call<'d free variables, by using back substitu
tion. Now there are two cases to consi<l<'r:
(i) In the case all unknowns ar<' ha.sic variables there is no free variable
to assign arbitrary values and h<'nce there is a unique solution.
(ii) In the case there is at l0ast. one free variable we can assign them
arbitrary elements of the fi<'ld and we obtain more than one solution,
further over the field of real or complex numbers we obtain infinitely
many solutions.
x+y-z+t+u= 1
-x + 2y + :3.: - t + 2u = -1
2x + y - z + 2t - u = 2
x + 6y + 4.: + t + 4u = 1
8y+7z+6u= 0
3x + 7y + 3.: + 3l + 3u = 3.
56
1 1 -1 1 1 1
-1 2 3 -1 2 -1
2 1 -1 2 -1 2
1 6 4 1 4 1
0 8 7 0 6 0
3 7 ;3 3 ;3 3
and by a.pplyi ng <'lcnH'Ii1 a ry row operations it can be reduc0d for <-'X am plc to
1 -1 1 1
0 -1 1 0 -3 0
0 0 5 0 -6 0
0 0 0 0 0 0
0 0 0 0 0 0
0 0 0 0 0 0
;r+y-z+t+u=1
-y + z- 3u = 0
5z- Gu = 0
0=0
0=0
0 = 0.
\Ve see that the syst0m is consistent with basic va.riabl0s x, y, z and free
57
z = 6/5u
y = z- 3u = (6j.5n)- 3u = -9/5u
x = 1- y + z- t - 11 = 1 - ( -9/5)u + (6j5u)- t- tt
= 1- t + 2u.
Thus the general solution of th0 syst0m is of the form
1- t + 2u
1 -1 2
-9/5u
0 0 -9/5
6/5u
0 +t 0 +u 6/5
t
0 1 0
u
0 0 1
Over the field of real numhNs, for any choice of t and tt we obtain a solu
tion, so the system has infinitely many solutions.
1 1 1 1
1 -1 1 0
X +y +z
3 1 3 2
1 1 -1 5
1 1 1 1
1 -1 1 0
3 1 3 2
1 1 -1 5
/j8
1 1 1 1
0 -2 0 -1
0 0 1 -2
0 0 0 0
:r+y+z=1
-2y = -1
z = -2,
and gives us
X= 5/2 , y = 1/2 , Z = -2.
has:
(c) No solution.
Forming its augmented matrix and applying elementary row operations one
g<•ts
1
a
a+1
VVe see that the system is inconsistent when a= -1, it is consistent and has
infinitely many solutions when o :;!:. -1. a has no value for which the system
has a unique solution.
+ y + zt = 0
-3~-r
3.1: + y(l - 1) - z( t + 1) = 0
<l.r + y + zt = t + 2
-3:r + y(l + 1) + zt = t + 5.
-3 1 0
3 (/- 1) -t- 1 0
-3 1 t t+2
-3 (I+ 1) t t+5
-3 1 t 0
0 -1
0 0 1 3
0 0 0 t + 2.
60
As is seen,the systnn is consistent only in the case t = -2. For this value we
ha.ve x = -5/2, y = -:l/2 and z = 3. Thus the unique solution of the system
is obtained as
0
0
Other solutions nam<'I.Y solutions having at kast one nonzero <'ntry are called
nontril'ial' solutions. i\on1 rivia.l solutions exist only in the case tlte system
has more than one solution, that is to say in the case there is at kast one free
variable. If this is the cas<' we fix each free variable once, assign it 1 and assign
all other free variahks z<>ro ~o get the so called fundamental .<;ofution.<:.
Example 1.6.5. Solve the homogeneous system
:r+y-z+t+'U=O
-x + 2y + 3z- t + 2'U = 0
2J' + y - z + 2t - 'U = 0
x + Gy + 4z + t + 4'U = 0
8y+7z+6'U=0
3.t: + 7y + 3z + 3t + 3'U = 0.
vVe sha.ll reduce its augmented matrix, but for a homogeneous system the last
column consists of zeros and does not change under elementary row opera.tons
and so may be neglected. Now, neglecting the last column we reduce
1 1 -1 1 1 1 1 -1 1 1
-1 2 3 -1 2 0 -1 1 0 -3
2 1 -1 2 -1 0 0 5 0 -6
to
1 6 4 1 4 0 0 0 0 0
0 8 7 0 G 0 0 0 0 0
3 7 3 3 3 0 0 0 0 0
x+y-z+t+u=O
-y + z- 3u = 0
5z- 61t =0
0=0
0=0
0 = 0.
giving
:::=6j':)u
y = -9/Gn x = -t + 2·u.
62
-t + 211 -1 2
-Hj':iu 0 -9j.5
Gj;iu = t 0 +u
6/5
1 0
'II 0 1.
-1
0
0
1
0
x- + 2z - t = 0
~:y
- 2.t + 2ky - 2z + 3t = 0
-:r + (~: + 1 )y + (k + 1 )t = 0
-:l.r + :Jky - 2z + (k + 4 )l = 0
68
1 -k 2 -1
0 1 2 k-1
0 0 2 1
0 0 0 k-1
This shows that the system has nontrivial solutions only in the case k = 1.
· By back substitution we obtain the solutions
X 3
y 1
.X= =t
z -1/2
1
(iii) A is invertible.
Proof. ( i) :::} ( ii) : Let A have a left inverse A' and let S be a solution of
A.Y = 0. Then AS= 0 implies A' AS= A'O =0 giving IS= S = 0 that is
to say I he only solution is the trivial solution.
64
(ii) ::::} ('iii) : Suppose the homogf'JH'ous sytem ha.s only the tri\·ial so
lution. Let R be the row reduced PrhPlon matrix which is row eqniYaJ,.nt to
A. Then RX =0 has is equivaknt to AX =0 and hence it also has only
the trivial solution; this m0ans that. all rows of the row reduced matrix R <nP
nonzero consequently R = I. Thus A is row equivalent to J and by Tlii'Ol"l'lll
1.5.16 ( iv) it is invertible.
(iv)::::} (iii): Let A hav0 a right inwrse A', then A' has kft inv0rs1' .-\.so
by the equivalence of (i) and (iii) proYNl above A' is invertible wi1l1 in\"I'I"SI'
say A". Now, we have
Corollary 1.6.8. For a. square matrix ..-1, the following are equvaJ,nt:
(i) .4 is invertible.
(iii) :::::? ( iv) : Supposing every system has at most one solution we con
clude that AX= 0 l1as a unique solution (i.e.1l1e trivial solution).
Example 1.6.9. tct us use Theorem 1.6.6 to prove that the matrix
l'n-1
~----------------------------------------------------------
~
i:"
~
66
these <'qua lit i<>s ran hc> put. into t.he form
or
t xk- 1(r·o + +r 1x + c x
Jo 2 2 + ·· · + Cn_ 1x"- 1)dJ: = 0.
1 1
J(:t)d.e = 0,1 1
J:f( :r )tl.r = 0,1 1
a.· 2 f( x )dx = 0, ···, 1 1
x"- 1 f(x )d:r = 0
Jv1ult.ipl,\"ing th<'m by r· 0 .c· 1.· .. ,c,_ 1 r<>spect.ively and adding we obtain
LU -f'ACTORIZATIO:\
A= LU
where L is a lower triangular matrix whose diagonal entries are all equal to
· 1 and F is an upp<'r triangular matrix. If this is the case A is said to have
an LU -far-tori:::ation or a11 ur -t!r-romposition. Note that, if i1 can he reduced
to an echPJ.on matrix F h~· applying elem<'nt.ary row operations .of the form
cR; + Ri wit It i >j t.hPn .-I has an LU decomposition, because then U =KA
where [,· is obtained frolll I by applying th<' same operations and hence J(
and its· invNS<' L = A'- 1 ar<' uppN triangular matrices whos<' diagonal entries
67
are equal to 1, and tlwy satisfy;\= l\"- 1 U = LU. Conversely, if A has an LU
decomposition A = LU then L ran h<' reduced to I by applying operations
of type cR; + Ri, (i > j), and hy th!.' algorithm following Theorem 1..5.16 the
same operations applied to I yields L- 1 • So, if we apply these operations to
A we get L- 1A = L- 1 LU = U,t.hat is to say A can be reduced to an (echelon)
upper triangular matrix by means of elementary row operations of the form
cRi + Ri with i > j. In this way \V<' also see that to determine L and U we
reduce the partition<'<! matrix (AJI] hy applying elementary row operations of
type indicated ahovf' and get [UIL- 1]. Inverting the second part we ohtain U.
They can also be computed effici<'ntly by writing out
LY =n UX=Y
of which the former can be soln'd hy back substitution and the latter by
forward substitution. For solving large systems computer programs for LU
decomposition are available.
r -~4 -~ ~ ]·
3 -1
and use this to solve
x- y + 2z = 4
-2x + y = -1
:t.r + 3y - z = 2.
To find the LU-decomposi1 ion we make the following reduction:
.
r -~
4
-~ ~ ~ ~ ~ 1 r ~ =~
3 -1 0 0 1 0 7
~]
1 -1 2 1 0
[ 0 -1 :J 2 1
0 0 19 10 7
Thus
r 10~ ~ ~]·
u = r ()
l -1 2]
-I -1
() 0 19 7 1
Inverting L- 1 we obtain
[, = r -~ ~ ~ J.
4 -7 1
Now, to solve AX= JJ W(' solve the equations LY =Band UX = Y, namely
u =4 X- y +2z = ll
EXERCISES ¥9
-3x + 2y - z + t = 5
x-y+z-t=1
2x- 3y + 3t = 0
-:r - y- z + 4t = 5
.1: - 4y - z + 7t = 5
x+y-z+t=2
2x - y + 2z - t = -1
-x - 4y + 5z - 4t = 0
:r - y + 4z - t = 1
1 1 1
-1 1 0
:I; +y +z
:3 1 3 2
1 -1 5
3 -2 3 -4 1 3
2 1 -1 -2 1 0 0
-1 4 1 -.5 3 G 1
2 8 -2 -4 0 7 4
5. Find all solutions of the following system over Z 2 :
:r+y+z=1
x+y+t=O
y+z+t=1
y+t=1
:r+z+t=O
6. Solve
x+y-z+t=2
2;t·- y + 2z - t = -1
-x-·1y+5z-4t=0
x - y + ·Lo - t = 1
7. \Vha.t must be the value of a if t hNe are numbers x, y and z such that
1 2
1 a+1
-1 0 a
X +y +z
3 -1
1 -1
2 0 -a
2x- y + 2az +t = b
-2x + ay- 3z = 4
2x- y + (2a + l)z +(a+ l)t = 0
-2;r + y + (1 - 2o )z- 2t = -2b- 2
has:
(c) No solution.
-X + 2y + Z - 2t = 0
-X- y + Z + t = 0
x+y+z+l=O
-:r + 5y + 4z - 5t = 0
2:1' - y + z +t = 0
10. Find the general sol11tion of the following homogeneous system in terms
of its fundamental solutions:
a.:+y-z+t+u=O
+ 2y - 2z + 3t = 0
x
- . T - 2y + 4z- 7l + u = 0
x-t+u=O
2.1' + 2y - 2z + 2/ + u = 0
+ x 2 - x 3 - x 4 + X5 = 0
:r 1
2xt + 2a:2 - + X5 = 0
X4
0 0 0 3 -4 1 3 0 0 0 0 1 -2 3 -1
0 0 0 -2 1 0 0 0 0 o· o 2 -4 6 -3
(a) (b)
0 0 0 -5 5 -1 -3 0 0 0 0 1 -2 1 2
0 0 0 -4 2 0 0 0 0 0 0 0 0 0 7
13. What must be the values oft if the following homogeneous syst('m has
nontrivial solutions?
+ tx 2 - 3x 4 =
x1 0
(t- 1)xl- (t + 1)x2 + 3x4 = 0
x 1 + tx2 + (t + 2)xa- 3x4 = 0
(t + 1)x + tx 2 + (t + 5)xa- 3x4 =
1 0
-1 /.; 3 -2
2 1 1 k
is invertible?
1 k+l -l k-2
2 1 1 k~ +1
15. Solve the equation AX = n for
-1 2 5 ;3 2 1 0 -1 1 2
5 3 1 -3 1 -1 -1 1 -2 1
A= fl=
1 -1 2 -2 1 1 1 -2 -2 -2
0 1 0 1 1 0 1 0 1 ·o
16. Prove:
2 1 -1 1
-6 0 -1 -4
A=
4 -1 2 3
0 -3 0 9
9
4
B=
-3
0
ID DETERMINANTS
Orw of the most useful devices oflinear algebra is the determinant function
that assings a scalar to each square matrix and even to some other concepts
wldch can be represented by matrices. vVe shall introduce them ind nctively
(in the way almost all readers are aquainted with from introductory courses
without using the adverb "inductively"), namely, first we definC' detNminants
of 1 X 1 matrices, then using this definition we define determinants of 2 X 2
matrices, then use this to define it for 3 X 3 matrices etc.... The JHOCC'dure will
bC' formalized below by considering submatrices obtained by delC'ting the first
row and one of the columns.
A~[
l
;\ ll A12 At3
.:121 An A23
l\ :n A32 A33
A_(3) =
Dl([Au]) = A11 and D11 (A) = 2::( -l)l+j A1jD 71 _ 1 (;1Ul) for 11 >1
j=l
det(A) , lA I,
Example 1.7.2. L<•t ns use this definition to give the well-known explicite
formulae for determinants of 2 X 2 and 3 X 3 matrices. From the defining
equality we obtain for a 2 x 1 matrix A that
Thus
77
Now using this determinant function for 2 X 2 matrices we shall obtain the
formula for the determinant function on 3 X 3 mat rices again by using the
defining relation above:
An A12 A13
A22 A23 A21 i\23 A21 An
A21 A22 A23 =An - A12 + A13
A32 A33 A31 A33 A31 A32
A31 A32 · A33
The computation can be made as the above arrows Rnggest, but this method
of computation, which is known as Sarrus' rule, is valid for the determinants
of 3 X 3 matrices only and it shouldn't be overeslimalcd.
Example 1.7.3. Let us use the definition to show that the determinant
of a triangular matrix is the product of its diagonal entries.We handle lower
triangular matrices and leave upper triangular matrices to the reader. The
statement is obvious for 1 x 1 matrices. Supposing it is true for ( n -1) x ( n -1)
lower triangular matrices we obtain
0
An 0
0
=An +0. +0. + ...
* *
A32 A33
An3
= Au(An···Ann) = AnAn···Ann as required. In par
ticular we have
det(J) = 1.
78
Now, we shall derive a. formula to express Dn(A), for n > 2, in terms of the
first two rows of A. Using the defining relation we get
n n n-1 - (k)
Dn(A) = L(-1) 1+i A1i Dn-d _:pn) = LH) t+i Ati(L(-1) 1+k(.:f(n)lkDn-1 (AU)) ).
i=l j=1 k=l
But by considering tlw f'nt ri<'s of .~Tu l we s<>e that the ('ntries of the first row
are the entries of the second row of A and so they can be expressed by
( .f':'l(j)). lk = ·t 2k
.f if k<j and ( -:-tU)) 1k-1 -- A 'lk
j' if k >j
Therefore we have
n j-1
D,(A) = L L( -1) 1 +i+l+~·AtjA'2kDn-'2(fp~·j))
j=l A·=l
n n
+L L (-1 )l+i+I+(A·- 1 ) A1jA2kDn-2(rT(kj))
i=l k=i+1
-- """'(
L..J - l)i+k .t-llj~'~2k
J " n· 11-2 (,--:--t'•il) + ""c·
L..J - l)l+Hk
. i t lj A 2k D n-2 (A-(kj))
·
k<j k>j
cR; + dR; +d R~
'
Proof: If we can prove (A) a11d (B) the last statement will follow at once
from
RI Rt R1
R;
R; R;
= c +
cR;+ Hj
R; R-J
Rn
R, Rn
"-v-"
,,!:;.$!
~
80
and
0= + + +
R;
R,
where undNhraccd dctNminants are equal to zero. So, all we have to do is to
prove (A) and (B). This \Vill be done by induction. The case n = 1 is obvious
for both statements. So to prove (A) assume Dn_ 1 is an (n-1)-lincar function
and take A to be an 11 X n matrix whose i-th row is of the form cR; + d R~. Let
B and B' be matrices obtained from A by replacing its i-th row by R; and R;
respectively. If i = 1 the defining relation takes the form
n
Dn(,.\) = 2)-1)1+i(cAli + dA~i)Dn-1(JIUl)
i==l
and decomposing it gin's D,(A) = cDn(B) + dD,(B') as required. If i > 1
one row of each fl(j) will he of the form cR; + dR~ and will he decomposed
into two parts of the required type, consequently (A) will follow.
To prove (B) forD" hy assuming it is true for Dn_ 1 we take ann X n matrix
whose i-th and k-th rows are identical and consider three cases separately:
(a) If i > 1 and k > 1 in the defining equality of Dn(A) each matrix JlU)
is to contain two identical rows and by induction their determinants are
all equal to zero, hence Dn (A) = 0.
(b) If i = 1, k = 2,t ll<'ll D,(A) = 0 follows at once from the formula estab
lished above givi11g the determinant in terms of its first t\VO rows and
81
their complements.
n n
Dn(B) = L( -l)l+i(Bli )Dn-t(_B(i)) = L( -1) 1+i(Ali)( -Dn-t(fi'U))) = -Dn(A)
i=1 i=1
since each _8(i) can be oht ained from A(j) by interchanging rowl and
row(k-1) and'since the assertion is assumed to be true for Dn-l· Now,
we get Dn(A) = -D,(B) = 0 hy the case (b). D
Especially the last part of this t h<'orem shows the effect of elementary row ·
showing that the validity of (a) and (c). As for (b), ET is the elementary
l
i
.
82
det(AT) = det(A).
8S
(b )If A and B are square matrices,t Ill'II for the matrices in block form we have
Theorem 1.7.'1 can also he used to give an explicit expression of th<' d<>
termina.nt of a. matrix in terms of its entries.In fact using this th<>on•m we
conclude that
* * * *
= Cl + ···C,
R,
Thus letting 1 1 , I".· · ·, ! 11 for the rows of the identity matrix I earl1 row of
an n X n matrix"' can be written in the form
n
Rk = .-1,.1/1 + Ak2f2 + · •· + Aknln = L Ak;J;k
ik=l
85
Rt
2:::7~=1 AJ;,l;. R2
R'2
R3
det(A) = Ra
Rn
R,
Tl
L·Ali 1 R3
it=l
L A!i,A2i2 R3
i1,i2=1
I;l
I;,
= L Ali,A;,···Anin J;3
i1,i2 1 ···,in
86
is either 0 or it is of the form ( -1 )111 .Ir there are any two identical indices ik
and i 1 ,then two rows are identical and t h<' determinant is O,if all indices arc
distinct by a number of interchan)!;<'S it can he converted to the determinant
of the i<l<'ntity matrix I and thus it is Pqual to ( -1 )"',where m is the numlwr
of interchanges to bring i 1 , £2 , • • ·, i" to t h<' nat ural order 1, 2, · · ·, n,for which
we also conclude that its parity is ind<'fli'IHknt of the way the interchanges are
carried out.Therefore b.Y coHsid<'rin)!; thi' function a from the set {1,2, .. ·,n}
into itself defined by
O"(n)
and letting
.sng( O") = (- 1)'"
we obtain
dct(A) = 2:sgn(O").I," 1 'l+la(:?)'''/l,,.(,)
"
where O" ranges over all functions from { l. '2, · · ·, 11} into itself such that a(k) f:.
O"(l) when k f:.l. Such functions are callPd jJ(I'mutations on {1,2, .. ·,n} and
.sng(O") is called the sign of the pcrnwfalion.ln order to indicate tlre images in
an explicit form we write
(]' = ( 1
Jl /2 ::)
87
Consider a square matrix rl and pick up one of its rows, say rowk. In
terchanging this row with the otw prior to it successively eventully we get a
matrix B whose first row is tlw k-1 h mw of 11, and its determinant becomes
(-1)k-l by Theorem 1.7.4 on one l1and and on the other hand it is
n -(A·j)
det(B) = L .\kj( -1) 1 +i det(A ),
j=l
for its first row consists of Akj's and their minors in Bare the same as tlH)se
in A. This shows that
n . -::_(kj)
det(A)=L·'kj(-1)"'+1 det(A )
_j=l
88
·\In
which is mnltipli<'<l I>~· . \kj is call<'d th<' <·ofl/("/or of Akj· The matrix adj(.·\)
ddi1wd by
is called the (l(Uoint of .I. In terms of tlt<'S<' rofactors the expansion takes the
form
11
<lPt( A) = L. \kj(/kj
j=l
\Yh<'ll the <:'ntrics of th<' /,·-throw of A is rPplacC'd hy :r1,:r 2 , ••• ,:r 11 th0 dctPr
minant of the r0su1t in!!; matrix h0comPs
1l
2:::: :rjakj
i=l
is the det<'rminant of a matrix with two id<'nt ira] rows R; = R.., IH'nce we have
11
L ..-\;jfl!-j = 0 for i =f k.
j=l
By intNpreting Corollary 1.7.G(d) that any property given for rows can aL<>o
be gircn for columns, we obtain the column expansions:
n
d('t (A) = ".L:>-hjakj
k=l
0 = LA;jakj,(i -:f. k)
k=l
In particular if det(A) -:f. 0 then matrix equalities tha.t we have obtained can
be put to the form
and it yields
A- 1 = (Ict~A) adj(A).
Thus we have proved the following
n { det(A) if i = k
'L:A;ja!·j =
j=l 0 if i -:f. k
equivalently
Aadj(A) = adj( A)A = det(A)I.
.?0
In particular the cofactor expansions with respPrt to any row (and any column
an') valid:
1 l OCJ(.!
,\ -1 = (f.r:l(.! [' l'). [=>
JOJ.~
)
Now, letting Dk = b 1aa + b'.!a:H: + · · · + bna,k and noting that it is tlw deter
minant of the matrix oht a i lll'd from A by n'placing the k- th col u 111 n by 17 W<'
k = 1,2, ... ,n
or
1);.
./k- / ) k = 1,2, ... ,n
2:r- y + z = 0
.r+y-z =G
l.r - !J.IJ + ;Jz = 28
91
We have
2 -1 1
D= 1 1 -1
= -6,
4 -5 3
0 -1 1
D!= 6 1 -1
= -12,
28 -5 3
2 0 1
D3= 1 6 -1
= 96,
4 28 3
2 -1 0
Da= 1 1 6
= 120.
4 -5 28
Therefore,
D1 -12 D., 96
X 0 0 0
2x x2 + 1 X
X x2 + 1 0 0
(a) =0 (b) X x2 + 1 X =0
3x- 5 x+1 X 0
3x- 5 x+1 4x- 5
1 -x -3 x+2
94
3. Given that
1 a 2
-1 1 b = 4,
a 2 3b
x a+ bx 2
(a) -x 1 b (b) -x 1-bx b
ax 2 3b ax 2 + abx 3b
-1 -2a -2 a+ 1 a+ 2 2 + 3b
(c) -2 4 2b (d) -1 1 b
a 4 3b a· 2 3b
0.1
0 0
1 0
l (b)
0 1 0 0
1 0 0 0
0 0 0 1
0 0 1 0
(c)
0 0 0 0 1
0 0 0 1 0
0 0
0 1 0 0 0
1 0 0
1 0 0 0 0
2 -5 4 -1 1
1 -3 3 -4
-7 3 -2 1 0
2 1 -1 -5
(a) 0 0 0 0 0 (b)
-3 9 -9 12
0 ·11 10 0 0
7 1 -1 5
12 10 0 0 -20
95
x+1 x+1 1
-2.1:-3 -X -3 = 3
1 0 1
10. Show that the determinant. of any skew-symmetric matrix of odd order
is zero.
b c o d e f
a 0 c X y z
b2 + ac be c2
0 a b u v w
ab 2ac be = b 0
., 0 0 0 c
a- ab b2 + ac
0 0 0 a 0 c
0 0 0 0 a b
13. Compute:
1 + :l"J a:1 Xt Xt
.7'2 1 + ,1:2 x2 x2
:Z':J x3 1 + X3 x3
;r" Xn Xn ... 1 + Xn
14. Compute:
a b b b
b a b b
b b ... a b
b b b a
15. Evaluate
0 1 n-2 n-1
1 0 n-3 n-2
n-2 n-1 0 1
11- 1 n-2 1 0
97
a b c d
b c d a
c d a b
d a b c
2 0 0 1
2 -3 1
-2 1 0 1
(a) 0 1 -1 (b)
0 0 1 -2
3 1 7
1 0 2 0
2 2 3 3
-1 -2 1 6
A-t=
1 -3 -2 5
4 4 5 4
(a) adj(AT) = (adj(A)f (b) adj(cA) = cn- 1adj(A) for any scalar.
1 -3 0 0 0
3 -6 0 0 0
2 -4 2 4 0
-2 3 5 1 -5
0 9 3 4 7
4 -3 9 2
1 -3 1 -1
0 0 1 1
0 0 2 -1
1 1 1 1
1 -1 'l -'l
1 1 -1 -1
1 -1 -/.
99
~ ~ - -
·c;_hapter 2 _
=, ;~ ;> ' ' ~ - ~ 1 ~. ' ~ ~ >
VECTOR SPACES
The n<'ccssity of vector qnantiti<'s as well as scalar quantities is very Wf'll
known to all readers. One usually nf'eds r<'sultants ( a.lgebraic sums) and scalar
multiples of vector quantities. The same sort of operations apply to many ot lwr
qnantitif's, for example matrices can 1><' considered to be vectors, althouglt it
docs not sound that natural. They ohcy all characteristic laws of vectors.
Den nition 2.1.1. Let F h<' a fi<>ld and V a nonempty set. If there is
(a) an operation + on ll
(b) a function
FxV ----+ V
(r, r) ~--+ cv
100
Ruch that the followin~ axioms are satisfied, then V is called a vcrlor spare
over F:
(V3 ) For cadt v E Y t hPn' exists an element v' E V such tlwt v + r' = r' + v =
0.
Elements of a vector spare are called l'ccfors whereas el<•nwnts of the field
Fare called scalars. Tlte n~apping (c, v) ~----+ cv is called scalar multiplication
or multiplication by s('([lnrs: and the eight axioms (VI)- (l&) arecalhl vector
space m:ioms. ThP particular type clement given in (V2 ) is unique (rf Exercise
7) and it is calkd tlt<' ::rro vector, it is denoted by 0 11 or simply by 0 if there
is no danger of confusion. For each v E V the element giH•n in (l;) is also
uniquely determined ( cf Exercise 7) and it is called the negative of v, it is
customary to denot<' it by -v. Vector spaces over the field of real numbers
are ca.JJed real vector spares and those OVE'r t]w field of complex nnmbcrs are
101
called complex vector spaces. In the subsequent sections our main concern trill
be real (or complex) vector spacc~'l.
(i) cO= Ov = 0,
(ii) ( -1)v = -v,
cO =c(O+O) by(V2)
=cO +cO by(Vs)
cO+ (-cO) =(cO+ cO)+ (-cO) by using(%)
0 =cO+ (cO+ (-cO)) by{Vt)
=c0+0 by(Va)
=cO by(V2)
and then
102
Here the reader is invitC'd to indicate the axiom to be used at each st0p.
and
A) The vector space of m x n matrices. For any field F tlw sC't pnxn
of m X n matrices is a \'0rtor space over F under usual matrix addition and
scalar multiplication sinrC' t h(' axioms (Vi)- (l18 ) are satisfied as we prov('d in
Theorems 1.3.1 and l.:l.2. In particular realm, X n matrices form a rC'a( V<'rtor
space and complex m X n matrices form a complex vector space.
Further, we can considPr m<1t.rices whose entries are vectors of a gi\'l'll VC'ctor
space V:
v= , vii E V.
All the statements givC'n in ThC'orems 1.3.1 and 1.3.2 hold for thesemat ricC's.
In addition if A is an p X rn matrix whose entries are scalars in the hase fip(<f
of lf and ifF is a matrix whosp entries arc vectors in V then thP product. A I'
can be defined in a natmal way an analog of Theorem 1.3.4 can bC' proved hy
making natural modifical inns.
------------------------------------------------~1Q1
l1 2 1
(1, 2)
(-1,3)
(2. 1)
(1,0)
+2
(2,0) (-2,1)
=
[
(-2,3) (-1,1)
(2,5) (4,3)
l[ +
(:l,-2) (2,-2)
(4,0) (-4,2)
l[
=
(2,1) (1,-1)
(6,5) (0,.5)
l .
B) The n-tuple space. For any field F the finite sequence (x 1 , •.•• ;r")
with (xb ... , Xn) with X; E F is called an n-tuple. These
Indeed, the axioms (Vi)- (V8 ) can E.'ither be verified directly or they can be
justified by identifying n-tuples (:~· 1 , • .. ,:tn) with 1 X n matrices [xt, ... ,x,,].
This vector space pn is known as the n-tuple space.
............................. ; (X I, X2)
(X I , X2, X3)
--0;-=::-----'-;-- X I
104
Th<' l'!'ad<'l' is advi<'ed not to try to <risurdi::e drmcnts of R" for n > 3 to
be "dircdrrlline segments'' or to be "some creatures whafe<,cr they are". They
are nothinp; but finite sequences of the form (x 1 , •.. , x,) with components
:r 1 , .•. , :r,. In various applications these components :z: 1 , ..• , :rn may appear
to be tlH' numlH'r of votes given to the parties in an election, they may appear
to be results of an exam, etc.
C) Field as vector space over itself. vVhen we take scalars and vectors
to be ('lC'm<'nis of a field F with operations ofF, the axioms (l'1 ) - (V8 ) are
obviously satisfied. Hence F is a. vector span' over ittwlf. Note that this vector
space can IH' id('ntified with the !-tuple space F 1 •
D) The space of functions. Let .F[a, b] be the set of all r0al valued
functions ddinPd on an int0rval [(/, b]. 1t forms a real v0ctor span' w.r. to
operations ddlned by
where f and g are functions ddined ou the interval [a, b] and cis a real number.
The verification of axioms (l1d- (l8) is straightforward and can b(' kft to the
reader.
namely polynomials
This function pis ca.lled the real polynom-ial function defined by the polynomial
P(x ). Note that real polynomial func1ions are defined in the interval ( -oo, oo)
and they form a real vector space.
SUBSPACES
Examples 2.1.4.( a) The set of vectors in the first quadrant of the plane
is closed under addition but it is not dosed under scalar multiplication because
106
as the figure shO\vs if Vl'ctors x and y lie in the first quadrant so does the
parallelogram with si<ks :r anJ. y and therefore its diagonal :t +y lies in the
first quadrant. Ilowewr ( -1 )x is not in the first quadrant. To see all this
algebraically we denot<' 1hP rirst quadrant by W and describe it by
It is closed under add i1ion l><'ra use if x = ( :rr, x 2 ) and y = (y 11 y~) hav<' positive
componf:'nts x t. x 2 , y 1 , lh t hPH the components x 1 + y 1 and :r:! + Y:2 oft he vector
:t +yare a.lso positiv0. Ilm\'('\'<'r, lr' is not closed under scalar multiplication,
for example the product ( -l )(5, 2) = ( -5, -2) is not in TV, although (5, 2) E
IV.
Proof. We first note that since lV is closed under addition and scalar
multiplication these operations can he considered to be operations of lV. For
108
Example 2.1.7. For every vector space ll the subsets 0 a11d l/ are
nonempty subsets \vhirh are closed under addition and scalar multiplication,
namely they are suhspac<'s.TIH'y arc called lr·ivial subspaccs.
of R 2 is a subspace ,for we have seen in Example 2.1.4 (c) that this subset
is closed under addition and scalar multiplica.tion.Ilowcver, the subsets given
in Example 2.1.4 (a) a11d (b) are not subspaccs, because one of the closure
properties is not sat bfiPd.
109
LINEAR SPAN
Example 2.1.10. Let S = {(3, -4, 1, 0), (0, 6, 1, -3)}.Regarding this set
as ~ subset of the space R 4 we can form the linear combinations
-8(3, -4, 1, 0)+4(0, 6, 1, -3)' 3(3, -4, 1, 0)+2(0, 6, 1, -3)' 0.(3, -4, 1, 0)+(0, 6, 1, -3)'
3 0 a
-4 6 b
1 1 c
0 -3 d
110
1 1 c
0 -3 d
0 0 a- 3c- d
0 0 b + 4c + 13°d
l
This is equivalent to sayin)!; that the system wllith augmented matrix
1 -3
[ 2 1 -32 (/b
-1 0 2 ('
I
[ 0
()
-3
1 -1
0
2
7 1 a- 3b- ic
b-_2aa
1
l
'and shows that this is in fact the case.
Theorem 2.1.12. If S' is any subset of a. vector space l' over a. field F
.then < S > is a subspace . .
111
(a) < S > is nonempty since the linear combination l.v is contained in
< S>.
(b) ·when we ta.ke two linear combinations of elements of S we can write
them as
< [ 12 -3]
1
[7
2 -1
3] [ -63 -51-5] >
by mf'ans of a homogeneous system.Now, a matrix
[: :l
112
1 7 3 a
-3 3 15 b
2 2 -6 c
1 -1 -5 d
has a solution.By reducing we sec that this is the case iff a.,b,c,d satisfy
-a + 2c - 3£1 = 0 , b + 3d = 0.
113
f'JII EXERCISES
1. The set of all symbols of the form << x, y >>,where x and y are real
numbers, together with the operations defined by
2. The set .of symbols of the form )x, y( where x and y are real numbers
together with the operations defined by
[ a b
c d
l +[ a' b'
c' d'
l[ a + a'
c + c'
b+ b'
d+ d'
l
c[ x Y] [ ex Y]
z t z ct
114
6. The set of positive 1·cal numbers together with operations EB and 8 defined
by
pEf)q=pq
c 8 p =PC
8. Is the subset
W = {(a,b,c) I a.,b,cE Z}
11. Which of the following are subspaces of the vector space of real n X n
matrices?
(a) Upper triangular matrices.
(b) Diagonal matrices.
(c) Invertible matrices.
(d) Symmetric matrices.
(e) Skew-symmetric matrices.
(f) Singular matrices.
115
12. Show that the following are subspaces of the real vector space of real
valued functions defined in ( -oo, oo) :
(a) Polynomial functions of the form J( x) = a + bx.
(h) Functions of the form acosx + bsinx.
(c) Functions satisfying y' + 2y = 0.
(d) Even functions.
(c) Odd functions.
(f) Periodic functions of period 1r.
1:3. Find the relation satisfied by a,b and c if (a,b,c) is a linear combination
of
(a) (1, -1, 1), (2, 1, -1) (4, -1, 1)
< (-1, 1, -1,1 ), (1, a, 1,-1 ), (-2a, -a, -2a+ 1,1 +2a), (-1, 3, a-1, 2) >
17. Can (1, -2, 4) he written as a linear combination of (1, 0, 1), (3, -4,1)
and (4, -6, -2)?
116
19. Find the relations satisfied by a,b and c if (a,b,c) is in the subspace of
R 3 spanned hy (1.-1.0),(0,1,1)<111<1(:3,-2,1)
20. Expr<'ss Span { (I, -I, 0 ), (0, 1, 1 ), (:J. -5, -2)} as the solution space of a
homogeneous s_vs1<'Jll of linear equations.
1 :J -2
-1 -2
2 -I 3
·~ 3
[ +4 --Lt-+ 2]
:r 10 '1 x G E<
[ 1 -x
:3 1 + .r
l[ '
-x
-:3.-r
a: 3
-:z: 2 - :z:
l[ l
'
3 2 - 3x
9 3:r +1 >
~~----~---- - -----~------~-~~---------~~-~-~-~-...-.,
117
Definition 2.2.1. Let Y h<' a vector space ov<.'f a field F and let
{ Vt. v2 , • • ·, Vm} be a finite set of vectors in V. \V<' say that v1 , v2 , • • ·, Vm
are linearly dependent if there <>xist scalars c1 , c2 , • • ·, cm, at least one of whkh
is nonzero, such that the (nontrivial) linear relation
implies
C1 = l'2 = ·· · = Cm = 0,
the vectors vb v2 , • • ·, Vm are said to be linearly ituh pf1ulent. A subset S of
V is called a· linearly dependent sf'f if it contains a. fl nite non empty subsetset
·consisting of linearly dependent v<>ctors. Otherwis<', it is called a. linearly
independent set.
Example 2.2.2. The vectors (1, -3, 2), (2, 2,-1 ). ( 1, 5, -3) in R 3 are lin
early dependent, because the existence of c1 , c2 , ca, not all zero, such that
x 1 +2:r 2 +:r 3 0
--:Lr 1 + 2x 2 + 5x 3 0
2:r 1 - x2 - 3.ra 0
is satisfied.
Example 2.2.4. For any vector spare t lw S<'t { 0} consisting of the zero
vector alon<' is lin<'arly dqwnd0nt:
1.0 = 0.
Further any set contait1ing the zero vector is a linearly depcnd<'nt set. A single
vector vis linearly d<'JWn<knt iff rv = 0 for some nonzero scalar c and hence
iff v = 0.
Example 2.2.5. Tit<' empty subset¢> of a vector space is linearly indepen
dent since it has no nonernpty linearly dependent subset.
is linearly independent. Indeed, if we have real numbers c0 , c" ···,en such that
then we have
(2)
Co = 0, Ct = 0, · · •, Cn = 0.
This example shows also tl1at two real polynomial functions are equal iff
the polynomials are equal. In fact, iff= g for
t}H'n
C=
are linearly dependent, b0cause they belong to the space of polynomials of the
form
121
and this space is spanned by 1, x, :z: 2 , x 3 , so the above theorem applies with
m = 3 and n = 4.
Example 2.2.9. Let us show that the vectors (1, -2, 4), (0, 1, 2), (0, 0, -3)
are linearly independent and deduce that neither of the sets
can span R 3 • Indeed, if we can show that the given vectors are linearly inde
pendent, the theorem says that R 3 can be spanned by at least three vectors,
however, the given s<'ts have fewN <'h•ments. To verify linear independence we
form
c 1 (1, -2,4) + c 2 (0, 1, 2) + c3 (0, 0, -3) = (0, 0, 0)
c1 0
-2c 1 + c2 = 0
4cl + 2c 2 - 3ca 0
namely c 1 = c2 = c3 = 0 as required.
Example 2.2.11. ThP Y<'ctors (1, 1, -1),(1,2,3) and (0, 1, 1) form a basis
for R. 3 . In order to verify tllis we must check conditions (a) and (h) above:
l~
001
0
1 b-
a a
0 -3 c + 5a- 4b
l
and hence the syst<'m has a solution. Therefore the given vectors span
v.
123
We solve the resulting homogeneous system and see that we must have
The basis for R 3 that we considered above is of course not the only one, for
instance it is very easy to verify that the set {(1,0,0),(0,1,0),(0,0,1)} is also
a basis and more natural than the one given above. In the subsection below
we describe this sort of natural bases.
Now we shall give examples of standard bases for natural vector spaces that
we usually need to consider.
Since we have
124
for each n-tuple (:r 1 , :r'.! .... , .r, ), on one hand we see that { r 1 , c 2 , ••• , c,} spans
is a basis for the n-tupl~' span•. This basis is known as the s!arulard (or usual)
bas.Zs for then-tuple span•.
As is done for the n-tupl(' space we can irnmediat<'ly verify that tlw matric<'S
00···000···0
00 000···0
whose (i,j)-entry is 1 all other entries are zero, form a basis which can he
considered to be the standard basis.
\Ve know that every polynomial has a unique expression in the form
l,:z·, .. . ,X 11 , •••
125
and these spanning vectors are lirwarly independent because any finite linear
relation
A vector space may have various bases but it turns out that all of them
have the same number of elements, as is established in the following theorem
·which also assures the existence of bases.
(i) V has a basis and it can be obtained by deleting generators which are lin
ear combinations of its pred<'!'essors. Further, any linearly independent
1:26
(iii) Any two bases for\' have the same number of elements.
appears can be wri 1(('n as a linear com hination of i 1s predecessors si nee Wm can
be replaced by a linear combination of them. If w, is not a linPar combination
of its predecessors WP ke('P it in the set of genNators. Thus in both cases we
pass to the set { w 1 , 1/':2, • • ·, W 711 _ 1 } and apply the same proced me. I 11 this way
we delete the generators which are linear combinations of th(' preceding ones
and get a generating set
·the last nonzero term would give one w;k as a linear combination of its prede
cessors, but this is impossible. To prove the rest of the sta.tPment it is enough
to note that given a linearly independent set and a set of generators their
union is also a set of generators and that if linearly independent vectors are
taken to be first vectors none of them will be omitted because none of them
will be linear combinations of their predecessors.
127
(ii) Since a basis is linearly independent this assertion follows at once from
Theorem 2.2.7.
(iii) Let B 1 and B2 be two basis for V containing n 1 and n 2 vectors respec
tively. They are linearly independent generating sets. Now, considering B 1
to be a basis and B 2 to be a generating set we have n 1 s; n 2 • Reversing the
argument we have n 2 s; n 1 and thus we get n 1 = n 2 by ( ii).
0
For example considering standard bases we observe that the n-tupk space
and the space of m. X n matrices arc finite dimensiona1 vector spaces with
dim(!R.") = n, dim(!R.mxn) = mn respectively. The space of polynomiaL how
ever, is infinite dimensional. The space of polynomials of degree s; n has a
(ii) If Sis a spanning set of n vectors, then by Theorem 2.2.12 (i) we find
a subset of S which is a. basis for V. Since this subset contains n = llim(l·r)
elements it cannot be proper and it must be S itself.
D
1':!.?
-1 :2 1 3 1
2 3 4 0
Xt = x2 = :r:1 =
'
:1.'4 =
'
J.~5 =
1 -1 0 -1 0
3 0 3 3 1
vVe shall use the above theorPill 1() find a basis. For this purpose we form the
matrix whose columns are tlH' ?;i\'('fl column matrices and we reduce it to gd.
for <'xarnplc the echelon ma1 rix
-I :2 :3 1
0 2 1
0 () 0 () 1
0 () 0 () 0
-(' 2
:2c 1
(' -1
:k ()
which is row ('quivalent to the ma1 rix obtain<'<~ by taking the first two rolnmns
1:'10
a hove namely to
-1 2
0 1
0 0
0 0
which is seen to be inconsistent. But x 3 is a linear combination of x 1 and x 2
since x 3 = c 1 x 1 + c2 x 2 leads us to the system with augmented matrix
-1 2 1
2 1 3
1 -1 0
3 0 3
-1 2 1
0 1 1
0 0 0
0 0 0
obtained by taking the first three columns of the above big echelon matrix.
As is seen the system has a solution (c 1 , c2 ).
1."11
Example 2.2.18. We know that R 3 has the standard basis of three ele
nwnts. So it has dimension three. Now, let us show that the sets
a) {(1,1,1),(1,-1, 1),(2,0,1)}
a11d
b) {(1,1,1),(1,1,0),(1,0,0)}
are has0s for JR 3 by using Corollary 2.2.1{) and verifying that the first set is
linearly independent and the second one spans R 3 •
In fact,
implies that
132
which yields c 1 = c2 = c3 = 0.
To verify that (1, 1, 1),(1,1,0) and (1,0,0) span R 3 , we write
c1 + c 2 + c3 a
c1 + c2 b
c1 = c, c2 =b- c, c3 =a - b.
COORDINATES
Proof: ::;.: Suppos<' B is a basis for V. Then each vis a linear combination
of vectors in B, say
Then we obtain
and hence
OVJ + · ••+ OV 11 = 0
[v]B = [ -:~· ]
134
so that [v + w] 13 = [v] 13 + [w] 13 and [cv] 13 = c[v] 13 • The matrix [v] 13 is called the
coordinate matrix of v relative to the basis B.
l
Thus we obtain a uniquely determin<'d n x 1 matrix
c1
[v]a = [ : .
c,.
l
If we take ~mother vector w with coordinate matrix
[w]Li = dt:
[
dn
then we have
and hence
l [l [l
Therefore
[v + w]s =[ c1 +: d1 c1
: + d1
: = [v]s + [w]s.
Cn + dn C,. dn
v ...._. [v]s = [
Cl
: . . . .,. (cl, ... ,c,).
c·,
for all r 111 V. This matrix is called the transition matrix (or the change of
coordinate rnatrLz:) from B to C.
Proof: In order to see the uniqueness first we take ·p to be v; in the equa.t ion
( 1) and ohtain
136
Noting that
0
0
[v;]B = 1
0
0
by using the product of partitioned matrices we obtain
To see that (1) is satisfied by the matrix in (2) we note that for any
·- --. 'I rc,i
"~ \~(
~ ' .u db lc
::::1
,,
we have
P[v]B
smce
1.'J7
by ( 1) and hence
_ [v]t., = QP[ I!]
implying that QP = I; ~)or all v/onsequently Q = p-l. Thus each chrmgf'
of coordinate matrix is invertible.
(a) from B = {(1, 1, 1 ), ( l, 1, 0), (1, 0, 0)} to [ = {(1, 0, 0), (0, 1, 0 ), ( 0. 0. 1)}
(b) from [ = {(1,0,0),(0.1.0).(0.0, 1)} to B = {(1, 1.1),(1, 1,0),(1,0.0)}
(c) from B = {( 1, 1, 1 ), ( l, l, 0 ), ( 1, 0, 0)} to C = {(1, 1, 1 ), ( 1, -1, 1 ). ( 1. 0. 0)}
l
inverting P. We have
p-1 = [ ~ -~
_:
c) Letting
1'1 wl
1 1
1'2 -w 1 - -w2
2 2
+ va
t'a wa.
· Therefore
P=
[~
1/2
-1/2
1 :]·
Now, we can explain how the change of coordinate matrix changes the
coordinates.
1:'1.9
e \Vrite the partitioned matrix [ w1, • • • Wn I v1 • • • vn] whose columns arc co
ordinate matrices of tv's and t' 's.
• Hednce it to row reduced ech<>lon form and get [liP] and thus obtaiu P
from the second part, because /) is the matrix of coefficients when v 's
are expressed in terms of tL' 's.
A) Row space
First we note that two row-cquivaknt matrices A and B have the same row
space, in particular they are of thC' same row rank. Because we have
n = PA
R~
B=
140
PuR1 + · · ·+ P1mRm
B=
P21R1 + · ··+ P2mRm
which shows that rows of B are linear combinations of rows of A in other words
row space o:C B is contained in the row space of A. Reversing the argument
we also obtain row space of A is contained in row space of B and the result
follows.
R=[~
... 0 A1h
*
... . .. 0
0
0 ... 00
A2h
*
0 Arir* .. ]
are linearly independrnl because when we multiply the rows by ch c2 , . . . , Cr
respectively and add we get
+ (0 .. ·0 0 .. ·0 c2A2h * ·· · )
+ (0 .. ·0 0 .. ·0 0···0 CrArir*···)=(O,O, ... ,O)
the subspace spanned by some n-luplrs, we consider the matrix whose rows are
given n-tuples and !.hen reduc£. it to an echelon matrix, and pick 11p non-zero
mws of this echelon matriJ:. Whr.'n l' is an arbitrary finite dimensional vector
space to apply the above pron'dure we use then-tuples of coordinates.
-1 2 0 -1
2 -I -2 0 1
-1 0 4 0 -2
0 -1 6 0 -3
and reduce it to
1 -1 2 0 -1
0 -6 0 3
() 0 0 0 0
0 0 0 0 0
Therefore a basis for H' is
we form
1 0 1 0 0
1 1 0 1 0
1 1 -1 1 0
-1 0 1 0 1
and find a basis for its row space by reducing it to
1 0 1 0 0 ,
0 1 -1 1 0
0 0 2 0 1
0 0 0 0 0
B=
where At, A 2 , ••• , An ar<' columns of A and c1 , c2 , ••• , Cn are scalars. \Vriting
out explicitely we see that it is equivalent to
lf.l
IJ=
AX= IJ
In order to show this we can form the matrix whose columns are coordinate
matrices of given vectors relatiw tot he standard basis and reduce it to echelon
form:
[-~ ~ ~] [:) ~ ~ ] [~ ~ ~ ]
2 1 -3 0 -3 -3 0 0 0
.
144
l
We see that the system with augmented matrix
1 2 0
[ -1 1 3
2 1 -3
l
is equivalent to the one with augmented matrix
1 2 0
[ 0 3 3
0 0 0
and hence it is consist0nt. According to the result above this means that
(0, 3, -3)is a linear combination of (1, -1, 2) and (2, 1, 1). Thus these three
vectors are linearly dl.'pendent.
We use Theorem 2.2.12 (i) to find a basis for the column space. \Ve apply
elementary row operations to reduce the matrix to a matrix in echelon form
and see in this way columns which are not linear combinations of their prede
cessors, they are exactly columns of leading entries. Thus bf colvmns of the
given matrix A (not columns of the resulting reduced matrix) corresponding to
columns of leading entries of the resulting matrix form a basis for the column
space of A. By using the coordinate matrices this procedure is used in the
construction of bases for subspaces spanned by a finite number of vectors in
an arbitrary vector space.
\Ve form the matrix by considPri n)!; coordinate matrices relative to the stan
dard basis and reduce to echelon form.
1 2 0 1 1 2 0 1 1 2 0 1
-1 1 3 0 () :3 3 1 0 1 1 0
1 -1 -3 1
---+ 0 -3 -3 0 ....... 0 0 0 1
0 1 1 0
0 1 1 0 0 0 0 0
1 1 -1 1
0 -1 -1 0 0 0 0 0
Thus the first, the second and th<> last columns are not linear combinations of
their predecessors namely v1 , 1' 2 and r· 4 form a hasis forth<' given subspace.
2) Find a basis containin~ :r and :z· + :z: 2 for the space of polynomials of
degree :s; 3 we know that { 1, x. :1· 2, .r:l} is a s<>t of genNa.tors for polynomials of
degree :s; 3. By adding a· + :r 2 to tlds s<'t we find a generating set
0 0 0 0
() () 0
0 0 1 0
() 0 () 0 1
l
which is row equivalent to
1 0 0 0
0 0 1 0
() () () 0
() 0 0 () 1
146
Thus the columns to be considered are columns 1, 2, 3 and 5 of the first matrix
a.nd the required basis is {:z·, x + x 2 , 1, x3 }.
C) The solution space of AX =0
Consider the homogeneous system AX = 0. If x 1 and x 2 are solutions of
this system then AX1 = 0 and AX2 = 0 an hence
= cAX1 =cO= 0
A(cX1 )
that is to say all solutions of AX = 0 form a subspace of n X 1 matrices. Recall
that to solve the homog('n<'ons system AX = 0 we reduce the coefficient matrix
A and we determine hasic variables and free variables, by a.ssigning 1 to each
free variable and 0 to all other variables we find fundamental solutions
* * *
0 0
0 1 0
x2 = , ... ,xs =
0
note that they are linearly independent and any solution is a linear combina
·. tion of them. So fundamC'nl al solutions form a basis for the solution space.
x+y+z-t 0
2x + 2y- z + t 0
-x - y - 4z + 4t = 0.
147
l [~ ~ l [~ ~ -: l
RPdncing the coefficicut matrix we have
[ 2
-1 -1
~ -: -:
-4 4
0 -3
0 0 -3
3 0 0 0
-11
0
.
x+y+z-t 0
-z+t 0
wit II frl'P variable y and t. Now, taking y = I and t = 0 W(' get the solution
-2
0
0
B=
1
0 1
148
SUMMARY
Let l' be a vector space over a field F .A subset S of V is said to be linearly
depend(n/ if we can find v 11 v2 , • • ·, Vm E V and c 11 c 2 , • • ·, Cm E F (not all
zero) such that
150
f»J EXERCISES
1. Determine whether (2, -1, 1, 4), ( -1, 1, 5, -2), (-3, 2, 4, -6) are linearly
independent .If not,write one vector as a linear cornbinetion of the others.
subspace'!
11. Find a basis for the row space of the following matrices
2 -1 1 3 1 -1 1 3 4
-~
l
1 3 1 4
4 1 1 -1 -1 1 1 4 1
(a) (b) (c) [ 1 -1 1 0
1
0
2
3
-1
-1
3
-7
3
3
-1
1 4
2 0 1
7 6
-1 4 8 4 12 1:
12. Find a basis for the column space of each of the following
2 5 3 1 0 -1 0 1 -1 3
-1 1 1 0 1 0 -1 2 1 1
(a) (b) (r-)
1 1 -1 2 1 -1 1 1 1 -1
4 -1 1 1 -1 1 1 1 -1 1
1 -1 1 1 4 2 1 -1 4
2 1 4 -5 6 1 1 3 4
(a) -1 4 1 -8 -6 (b) 3 2 2 8
2 3 6 -12 4 1 0 1 0
1 7 7 -20 -2 4 3 5 12
1 1 -1 1
4 2 1 0
2 0 3 -2
7 3 3 -1
(~ Let S = {(1, 1, 1), ( -1, 1, -1), (3, 2,3), ( 4, 6, 1), (1, 0, 4 )}. Find all subsets
of S which are bases for < S > by considering
(a) S as a subset of R 4
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ____!.l:i:l
(h) S as a subset of Zi
v'( c) S as a subset of Zj
19. Expr<'ss
Span { (1, -3, 2, 1), (1, 1, 1, 1), (3, -5, .5, 3), (2, -6, 4, 2), (2, -2, 3, 2)}
1 3 -2
-1 -2
2 -1 3
4 1 3
0 1 3 -2
2 4 0 4 2
1 4 1 4
1 1 -1 0 0
-1 1 -1 1
(a) -1 -1 1 0 0 (b)
1 -1 4 -4
3 1 -.5 -4 -2
-1 1 5 -.5
4 0 -8 -8 -4
3 0 1 2
154
22. Find a basis for the solution space of the homogeneous syst('m
x+y+z-t 0
2x - y - 2z + 3t 0
-x + 2y - 3z + 4t 0
2x + 2y - 4z + 6t 0
2x- y + 3z +t 0
-5x+ y + 4z- t 0
-x- y + 10z + t 0
3 -1 1 2 1 xl 0
4 4 -2 1 2 x2 0
0 1 0 1 -1 x3 0
1 5 -3 -1 1 x4 0
0 0 0 1 -1 x5 0
{(a,b,c,d): a-b+c+d=0,2a+3b-d=0}
-3x + y + 3z = 0
in space.
155
{a+ bx + cx 2 + dx 3 : a+ b + c = 0, a+ b + d = 0}
28. Find a basis for the space of 2x2 matrices
satisfying x + t = 0.
[: :l
29. Find the dimension of the vector space of polynomials generated by
x +x 2 , x - x2 +x3 , 2 - x - x3 , x + 1.
30. Find a basis for R 4 containing the vector (1,2,-3,0).
[ -1 2 1
4 1 -1
l and [ 1 0 1
0 1 0
l
33. Find a. basis for the vector space of polynomials of degree ::; 3 containing
1 +x + x 2 a.nd x + x 2 .
34. Find a. basis for R 3 containing {(1,-3,2),(4,-5,0)}
[
to a. basis of the space of 2
~1 : l,[~ ~ l,[~ -: l
X 2 matrices.
156
{(1, -2, 1), ( -2, 4, -2), (1, 1, 1), (2, -1, 2), (1, 1, 0)}
38. Find a basis for polynomials of d<'gree ::;3 consisting of polynomals from
(3~) Find the coordinate matrix of tlw function cos 2 x relative to tlH' ord<'fed
/ basis
are linearly dependent.For this valne express the last function as a linear
combination of the first two functions.
v+ w
cv
V·W
cos e = llvllllwll"
(see Figure).
X3
v \.llv-wll
\
All others are dedur('d from tlws<> and some other concepts can be introduced
by using them.
In this chapter we shall genNa.lb:e these ideas to arbitrary real vector spaces.
\Ve begin with g<>neralizing tlte concept of dot product as inner product.
Definition 3.1.1.
A) Let V be a real vector space. A function from V XV to R assigning ( vlw)
to the ordered pair ( 1.1, w) is cal!HI an inner product on V if it sati.<;fies the
following axioms:
B) A real vector space together wilh a specified inner product is called a real
inner product space.
Proof.
(i) Using (P2) and (Pl) W<' obtain
is an inner product on R 3 .
Finally
Solution. Let v = (x 1 , ••• , Xn), 11' = (yl, ... , Yn) and p = (z1, ... , Zn), then
n n
(2) ( wJv) = LY;X; = L x;y; = (~·Jw) ;
i=l i=l
(3) (vJv) =xi+···+ x;l, > 0 when X; =j:. 0 for some i, equivalently when
1J =1- 0.
162
Note that (AlB) is nothing but the standard inner product obta.in€'d by
considering the matrices A= [Ai~:] as vectors with mn components A;J:.
(XIY) = yrx
Example 3.1.7. Let C[a, b] be the real vector space of real continuous func
tions defined on a closed interval [a, b]. The following is an inner product on
V:
(Jig)= 1b f(x)g(x)dx.
This inner product will be taken to be the standard inner product on C[a, b].
In fact, if j, g and h are continuous functions on [a, b] and c and d are real
numbers we have
(1)
(cf+dglh) = 1b(cf(x)+dg(x))h(x)dx
= c 1b f(x)h(x)dx + d 1b g(x)h(x)dx
= c(flh) + d(glh)
16.1
(3) To \'<'rify ( P:l) take a continuous function on [a, b] such that f( :r 0 ) :j: 0
for some :r 0 E [a, b]. Letting L = f(x 0 ), from continuity off\ we see
that there exists o > 0 such that
~ . ~~
IJ-(:r)-L- < F·
2 for :ro-{!<x<x 0 +o
. /,'2 ., ., L:!. .
that 1s - - < j-(;r)- L- <-and eqtuvalcntly
2 2
J/ ., ;~U
-. < J-(:r) < - for :r 0 - b < :r < :r 0 + o.
2 2
Thus 1hNP is a subinterval [c. d] of the int<'rval [a, b] surh that c < d and
/} .,
T < J- (:r) if c ::; :r ::; d.
Th<'n
y
3 L2/2 1-----..,,c__ __:~
a b
164
fill EXERCISES
1. Show that the following functions from R 2 xR 2 into Rare inner products
on R 2 :
2. For each of the following, either show that the given function is a.n inner
product onR2 or provide a.n example violating one of the axioms (Pl),
(P2), (P3).
()
(
for v = (x 1,x 2) ami w = (y 1 ,,1}'2) is inner product on R 2 iffb = c, a> 0
and ad- be > 0.
.1, Given v = (1, -2, 1,3), w = (2. 0, 1, -1) in R 4 with the standard inner
product. Evaluate:
5. The trace of a matrix .-1 is t h<' sum of its diagonal entries Jlii and it is
dPnoted by Tr( A). For mat ric<'s
A=
1 -1
[ .
1 0
I
2
l and B = [1
2
1
-1
-1
1
l
evaluate:
a) Tr(,lT H) b) Tr(JJTt\) c) Tr(AT A)
8. Let V be any in n0r product space over F. If v f 0 and w an- two V<'ctors
in V determinE' c E F such that (cvlw- cv) = 0. For this value of c
prove that
(wlw) = (cvlcv) + (w- cvlw- cv).
Show that for Y = R 3 the vector cv is the orthogonal proj0rtion of w
along v.
Jfi7
Inner product axioms (PI), (P2), (P:J) and tlwir consequences giYPil 111
11 1 ·11=~.
If llrll = 1, then u is called a unit vector (or normalized vector).
(/'Ill')= 0.
c) For any sub,c;et S of V thr sri of all l'rcfors u•hich are orthogonal to all
vectors in S is called the· orthogonal complement of S and is denoted
by s1. (read ··s pcrp ").
Example 3.2.2. The norm of a \'('<"tor r = (J:I! ... , :r,) in R" with respect
to the standard innN product is
Solution. We have
a.)
= (sin xI sin x) =1 1r
sin 2 x dx = -111r (1 - ms 2.1· )dx
-1r 2 -1r
~ (x _ sin2x)l1r = 11 (l)
2 2 -1r '
The basic properties of norm and orthogonality are given in the following
theorem. These are the most important properties in all mathematics.
II vii = 0 iff v = 0.
llcvll = lclllvll·
(iii) If n and w are orthogonal vectors in F, then
16.9
Proof:
llvll = 0*) 2: 0.
(ii) By using the dPfinition of norm, inner product axioms and Theorem 1
we get
( vjw)
(v-nujnl') = d t•-cwJw) = c((vJu·)-c( wJw)) = c((vJw)---· (wJw)) = 0.
. llwll 2
170
(v) We have
that is to say
llv + wJJ ~ llvll + llwll.
(vi) By definition,
c) d(v,u):Sd(l',ll')+d(w,tt)
Proof. By using the properties (i), (ii) and (v) respectively we have
Corollary 3.2.6. If l' is a real inner product space, then for arbitrary non
zero oectors v and w in V, we hcwe
0 )-'-:-I
'l,---:
.,.:-I(..::..,.v,..:.,-1 < 1
llvllllwll-
Since ( vlw) is a r0al number it yields
(vlw)
-1 < < 1.
- llvllll·wll
0
172
This corollary shows that the following definition is reasonable in the real
case.
Definition 3.2. 7. Jf v and w are vectors of a real inner product spnce, the
angle e brlu•pen them is defined by
(vlw)
cos e = llvllllwll; 0 s () s 1f
Example 3.2.8. Find the angle between v = (2, -2, 0,-1) and
w = (1, -1, -1, 1) in R 4 with standard inner product.
(viw) = 2+2+0-1=3
llvll v'4+4+0+1=3
Thus e= ~·
Solution. Taking v =(a, b), w =(cosO, sinO) in R 2 (with the standard inner
prodnrt) and applying tl1e Cauchy-Schwarz inequality we get
(4)
Thus elements of Sl. are of the form (t, -t, -2t), i.e., Sl. is the subspac<'
spanned by {1, -1, -2).
174
BlfJ EXERCISES
A= [ 2
1 -1
-1
1
. -1 -2
3. Show that ll~ll v is a unit vector for any non-zero vector of a real inner
product space lf.
4. Normalize:
c) [ 10 -J3
-1 l in R2x2
8. If v and w are in a real inner product space prove that llvll = llwll iff
( v + wjv- w) = 0. Interpret this geometrically in R 3 •
12. If v and w are non-zero orthogonal vectors of an inner product space show
that llv + wll > ll·vll and llv + wll > llwll. Interpret this geometrically in
R3.
~~-----------------------~
I
176
13. Find the angle lwtweE>n the following vectors of the given real inner
product space:
b) v=t+1, w=tinC[-1,1]
c) A= [ 2 1
0 '-2
l' B = [ -1 1
1 1
l in R2x2
a) S = {(4,-3,2,-1),(8,-7,5,-30)} in R 4
b) S={(1,1),(1,2)}inR2
16. Let S 1 and S2 be suhspaces of an inner product space such that S1 ~ S2.
Prove that Sf ;;2 Sd-.
17. If 8 is a. subset of an inner product space V, prove that Sl. = (S)l. and
(S} ~ Sl.l. wherE' (.'i} is the subspace spanned by S.
/
177
s1 = {(L1,1),(-2,1,1),(o.-Ll)}, s2 = {(3,~,-1),(-1,2,-2)}
are orthogonal sets and the sets
s3-- {(-1
/3' _!__ J..-)
y'3' J3 '
84 = {(1.0,0),(0,1,0),(0,0,1)}
178
are orthonormal sets. The orthonormal set 8 3 has been obtained from S1 by
normalizing its non-zero vectors.
1, cos x, cos 2x, ... , cos 71 x, ... , a 71 d sin x, sin 2x, ... , sin mx, ...
1 1
-[cos(71 +m)x + cos(n- m)x]dx
_,. 2
~ [sin(71 + m)x + sin(n- m)],.
2 71 + m n- m _,.
0
1 [sin(n+m)x sin(n-m)x],. _
-- - -0.
2 71 + m n- m _,.
IIlii = V'fi and II cos n;rll = II sin nxll =Vi for each n EN+
Our next theorem shows that ort l10gonality implies linear independence (of
course not conversely). This is an ohvious fact in R 3 , for example three mutu
ally orthogonal vectors cannot lie i 11 the same plane, two non-zero orthogonal
vectors cannot be parallel to tlw sam<' line.
c, = 0.
180
form an orthogonal basis for R 3 with respect to the standard inner product.
Indeed, we have
Example 3.3.5. Let us use the orthogonality of sine and cosine functions in
C[-1r, 1r] to evaluate
The following theorem shows that every finite dL.nensional inner product
space has an orthonormal basis and gives us a procedure to construct it.
181
11'1
'll'tn
(1l'k'U' 111 ) =
Example 3.3.7. Consider the innN product space C[O, 1], find an orthogonal
basis for (1.:L"J 2 ).
(xl1) [1 1 2 [1 ( 1) 2 1
p 2 (x) = x - -
1- ·1=x- Jo xdx=x-2 with IIP2II = Jo x-2 d:r= 12
= x 2 -x+61
ORTHOGONAL PROJECTIONS
Given a subspace of an inner product space and a vector of the sparE', the
idea of finding a vector in the snhspar<' which '6;6 fits to the givC'n V<'rtor
/
has many useful applications. In this section we discuss the exist<>nre and
uniqueness of such a vector, and give some important applications.
We prove below that when l'V is finite dimensional the concepts of ort ]Jog
anal proj0ction and best approximation coincide and for each v E V t lH'r0 is
a unirp1e orthogonal projection in l¥. This fact is an obvious conS0<JI1f'lllf'
of geometric considerations in space but it is not obvious at all for arb it rar.v
inner product spaces.
(vlvt)
~Moreover,
(vlvJ) (vlvn)
(i) P = · I . V1 + · · · + (v,v,)
(v1v 1)
I . '1! 71 when B is orthogonal
Proof. Let { v1 , .•. , v,} be any basis for lY and let p = x 1 v1 + · · · + x, 1'" h<'
a vector in Hl. Then, v - p is orthogonal to all vectors of H' iff
This yields
Thus it remains to prove that, llvll ~ i!PII and the orthogonal projection p
is the same as the best approximation. In fact, for any w E TV, we have
(v- PIP - w) = 0 since p - w E W and hence
Proof. If we take lV = ( n1 , ••• , r,), then the last part of the theorem says
that llvll 2: IIPII and equa.lity holds iff
(rlr1) (vlun)
v=p=
(/'II I'll
7'1+ .. ·+'(1' lv,) L'n·
71
186
Example
3.3.11. Find the clos<'st vector to (5,2, -1,3,4) in the subspace generated
by the orthogonal vectors ( 1, 1, 0, 0, -1 ), (1, -1, 1, 0, 0), ( -1, 0, 1, 0, -1 ).
3 2 10 1
P = -tc 1 + -w.,- -w3 = -(15 -8 0 7)
:3 3-3 :3' ''
1: ex cos xdx = e"' cos :1' [1r + J: e"' sin xdx =- 2 cosh 1r + e"' sin x [1r-;_: cos xdx
!S7
we obtain
and thus
cosh 7r •
p( X) =- -(Slll X -
7r
COS X).
by a linear relation
(k = 1,2,···).
Y1 au a12 aln
Y2 an an a2n
=bG + b2 +···+ bn
Yrn aml arn2 a.mn
Yt
Y2
Y=
Ym
may not be in the subspace spanned by the vectors
bn (Yivn)
The solutions b17 b2, · · ·, bn are the values to be able to get the ](last value of
m
y = ax 1 + b
189
Xt y
1 2
2 1
0 2
1 1
1
we obtain
2
2 1 1
VJ () v2 = and Y= 2
1 1 1
1 1
that is to say
Ia + Sb G
.sa+ .Sb 7
. .
g iVlllg a
-
= --21 and b = -.
19
10
Thus the relation should be
y = -0.5x 1 + 1.9.
190
y
7
,.---·-·-·
j 6;
I I
:
!-1
!
!
_ __.:_-.+~--+X
-2! 2
'
f
;
l. -7
33a - b + 9c = 41
-a+ 9b- c -7
9a- b + 4c 12
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ___,_I.'JL
which has the unique solution (a, b, c)= (1, 1, 1). Thus the required parabola
JS
y = x2 + x + 1.
APPLICATION 2: FOURIER SERIES
fpt 8 = {<P 0 , <1> 11 •.• , <Pm, •..} be an orthogonal subset of the inner product
spac0 C[a, b] and let f be a. function in C[a, b].The orthogonal projection off
on t]Jf' subspace spanned by <P 0 ,<P 11 .•. ,<I>m is
The numbers
k = 0, 1, ... , m, ...
which e~ppea.r as coefficients in these projections, are called the Four·icr coeffi
r.inlfs off relative to the orthogonal set S and
00
L Ck<I>k(x)
k=O
f"" L Ck<l>k(x).
k=O
ao oo f(
f(:r)"" 2 + L)an cos nx + bn sin x)
n=l
1
an=-
1l"
f..
-.-
J(.r) cosnx dx, bn = -1
1l"
j . f( x) sin nx dx.
-7r
(-1)n-1
a,.= 0, b,. = -'---'-
n
· and thus
~ - ------------- -~-----
---------------------..
19.")
r
I
194
{[ -112]
5
' [ -31 02] [48 02] }
is an orthogonal set.. Find a matrix which is orthogonal to every element
of this set
(b) (-1,-1,-1.-1)
(a) (1,-1.1).(2.0,-1) in R 3
(b) (1,1,1.0.0.0),(1,1,1,1,1,1),(0,0,1,1,0,0)in R 6
196
-'-1
1 l in R2x2
x1 1 1 2 3 1 2
x2 1 2 1 1 2 1
y 221121
11. Find the Fourier cosine expansion of f(x) = sinx + cosx over [0,211"].
197
p- 1 AP=D
PD, namf'ly
Thus if A is diagonalizahlc then we can find scalars )q, ... , >., and linearly
indepencknt column matrices P1 , •.. , P, such that
Reversing th<' argumf'nt W<' can say that if we have scalars and column matrices
as we l1a\'0 just d0scrih<'d then
AX=.\X
are called charactrri.<;! ir· rrdors (or eigenvectors) corresponding (or belonging)
to >..
At this point we also note that ,\is a characteristic value iff (.\I- A)X = 0
has a nontrivial solution equival<'tllly ,\is a root of the polynomial
dPt(Al- A)= 0
A=r-:; -: -:J
.\ + 11 5 3
det(.\I- A)= -12 .\- 7 -2 =(A- 1)(.\- 2)(,\ + 3) = 0.
-12 -!i .\ -·1
Thus the eigenvalues of A are
.\a= -a.
Now, for Pach eigenvalue ,\ WC' dPt('rtniue linearly independent eigenvectors
namely fundamental solutions of hotnogeneous equatious (.\I- A)X::: 0. For
.\ 1 = 1, the coefficient matrix is
12 ,) 3]
.\1!- A= [ -12 -6 -2
-12 -5 -3
200
and is row-equivalent to
[ 120 -15 3]
1 .
0 0 0
So we have y = z and x = -~z which yield the fundamental solution
p1 = [-2/3] 1
1
A2I- A= [ -12 -5 -2
13
5 3] [ -12 -50 -21]
-+ '
1
-+
-12 -5 -2 0 0 0
1 1]
[ 0 -5 10
0 -+
[1
0 1
0 -21]
0 0 0 0 0 0
[
-4 -.5
0 -.5
0
-12
10 -10
l[ l
-5
1
5
-7
_,.
-4 -5
0
1
1 -1
0
-12 -5 -7
0 0
~l l
yield
P, =[- =d P, = [ -:
201
The matrix Dis the diagonal matrix whose diagonal entries are the eigenvalues
corresponding to the eigenvectors P1 , P..J and P3 respectively:
D=[~ ~ ~]·
0 0 -3
These matrices satisfy th<' cqualit;.·
3 - 2 0
ft ;::::
n
-
<;
0
t- 5 0 -·1
(i t-1 =(/-l)(t 2 -1)=(/-1) 3 (t+l).
6 0 t+G
Therefore its cltaracterist ic val uPs arc
Aa = -1.
For A = 1 we find
1-A= [ 6
.\ 1
- I
(j
0
0
()
-·1
1
(j
l[ l 1 0 1
0
()
0
()
l
0
202
· which invoh·cs the dependent variable x, the independent varia hle t and the
d.r
derivative - 1-. To obtain its nonzero solutions we write
d
dx 1 ,
-=Act
X
or
where Cis the constant of int<'~?;ration. Note that when C ranges over all real
numbers, =fee will range over all non-zero reals. Including the zero solution
also we see that all solutions of ( I) are functions of the form
(2)
dx 1
ell
dx 2
dt
clxn
(3)
dt
where the A;i are real numbers and .r 1{1), .. . ,xn(t) are functions to be deter
mined. Letting
!k.l.
dt
~
dX dt
..tl= and
dt
4:!:....
dt
(5)
204
or
p-1 d;~ = np-1 x.
Now the substitution
y = p-l X
namely
J(le>.,t
K2e>.2t
p-lx =Y =
--------------------------------------------------------=2Q2
A"1e>.,t
](2e>.,t
X=P (o)
dx 1
lit
dx 2
dt
A= [-4 6]
-3 5
.A+4 -6 ., )( .A+l)=O.
'
det{.AJ- A)= =.A~-.A-2=().-2
3 .A-5 '
namely we get
6x- 6y = 0
3x- 3y 0
wl1ich is equivalent to
X- y 0
0 = 0.
[: =: l~ [~ -~ ]·
p = [ : ~ l
207
d:r 1
dt
(L!:2
dt
d:r 3
dt
A= [ -~ ~
-G 0 -5
_: ]·
[ :: ]· = [P = [
X3
~ ~ -2/~ ;:~,]·
1 0 1
]· [
/~3e
We can use the met hod <'Xpla.ined above in solving higher order linear dif
ferential equations. To this <'nd, let
dnx dn-lx dx
-d + a,_~-1--1 + ·· ·+ a1-dt + a 0x =0
t" <tn-
Xt X
dx dxt
x2 dt = dt
d2 x dx 2
:r 3 dt 2 =
dt
d.rn-1
x,
dt
d.rn
dl
and it has the coc•ffici<'lll ma.trix
0 0 0
0 0 1 0
A=
0 () 0 1
-ao -(/1 -a2 -a,._ I
Th<' characteristic equation of A IS
). -1 0 0
0 ). -1 0
= ao +alA+···+ a 11 _ 1>."- 1 + ;.n =0
0 0 0 -1
flo a! a2 A +an-I
.i\ot<' 1 hat this equation is obtained from the given difTerential equation by
suhstit tiling ).k for~:~. Tht:' si~nifiranre of this characteristic equation is illns
t ra1<'d in the following example.
a: 1 .r
d.r d:t 1
:z·~ dt = --;[[
d:r., lP.r
d(- = d!~ =x=:z·l·
210
). -1
= ). 2 - 1 = 0.
-1 ).
This shows also that the solutions form a vector space which is spanned hy
eA 1 t = e1 and eA 2 t = e-t. Note that here )q = 1 and .-\ 2 = -1 are roots of the
characteristic equation.
CAYLEY-HAMILTON TI!F:OilE~I
Proof. (Optional) Let,\ lw an~· n X 11 matrix. Note that the entries of the
matrix .AI - A can be viewed I o IH' polynomials of degree ::; 1 in .A, so that
this matrix can be considered as a matrix over a field containing polynomials
(such a field do exist!). The ent riPs of its adjoint are cofactors of entries of
.AI- A and as such they are polynomials of degree ::; n- 1 and therefore we
can write
and using
(AT- A)adj(Al- A)= det(>.J- A)I
we get
-ABo
Eo- AB1
Bn-1 = I.
Now multiply the first equality by A 0 =I, the second by A,··· the last one
by An and add them all to get 0
as asserted.
A=[~~ ~]·
1 -1 0
Solution: We have
A -1 -1 -1
b.A(X) = -1 ). -1 = A3 - A2 - A- 1.
-1 1 ),
213
So we have
nanwly
215
~II EXERCISES
[~ ~ =:
0 1 -c
l
(a)
[ l [110]
2 k
k 2 (b) ~ ~ ~ (c) [ (2) 20. 2]
2 2 0
2
l
.) . What must be kif the matrix
[
~ ~1 ~;
0 1 -1
[: :l
has real eigenvalues. Prove that this condition is satisfied by 2 X 2 real sym
metric matrices.
2 -2 -2 1
~2 ~3]
-1
(a) [ ~2 :l (b) [
-2 -2
2 (c)
-1
1
0
2
-3
0
-2
1
0
2
-1
4
10. Given
:J/7
Use this to prove that cigem;C'ctors lwlonging to distinct t>ig<>nvahws are lin
('arly independent and that an n xu matrix which has n distinct ('ig<'nva]uC's
is diagonalizahk.
218
It follows that
that is to say
either A2 - At = 0 or P[ P2 = 0.
21.9
Now, we give a name to matrices whose columns (and rows) form an or
thonorma.l set.
Note that a matrix is orthogonal iff Q-t =QT. Considering the partitioned
form
Q?:
we see that the equality QTQ =I is equivalent to the conditions
for which SQ 1 = .\ 1Q 1 • Now, the matrix QTSQ is symmetric and the parti
tioned product
shows that
In particular
(QTSQ)il = Q[SQl
= Q[AtQ1
= .\Q[Ql = At6il.
Therefore
At 0 ... 0
0
QTSQ =
s
0
[ ~ ~ ] 7 Q'l' sQ [ ~~ l
which is a diagonal matrix D. Il<'nn' W<' have
pT s p = D
where
I 0 ]
P=Q [ 0 Q .
Here P is an orthogonal matrix sine<'
pTp 0] [1 QT0][10]
(J Q
[1 QT0]
Q
0 0 0
s=
2 -1
[ -1
-1
1
2 -1
-1 2
l
by means of an orthogonal matrix.
We have
.X-2 1 1
b.s= 1 .X-2
1 A-2
222
[~I l= ~ [ -J
[v3
1
.J:l
l
.J:l
1
-vz
0
1
,f'i
1
-~
J6
1
J6
l
cliagonalizPs S :
X=
or X=QY
l
reducl's tlH' form to
we may bring the left hand side to the form XT AX namely we get
Eigenvalues of
2 v'2]
[ v'2 3
are found to be A1 = 1 and A2 = 4 and thus the substitution Y = QT X gives
us
225
y~ + .Jy~ = 16
or
Y¥ + y~ = 1.
1(j ·1
\Ye identify the conic as an ellipsC' with principal axes a = 4 and b = 2.
2 .1,21 + ·l ,2
,;.I~ + ,;J.a-
•l,2 ;) , .,
,;.l1.1~- 2·. .
.l2::Z3- 2 .1.a:r1-
• • - 2~
1.
.Y'~' .. tX
with
2
A= [ -I
-1
we can use Exampl<' 1.2.2 to writ<' th<' standard <'quation as
or
which is a circular rylindN. Its ax<'s ran lw tak<'n to he lines paralkl to the
l
unit vectors
V}
I [ :I l U·>
-
= v2
Ir.:, [ -011 and u, = ~ -~ [ ]
G!fJ EXERCISES
(a) [ ~! n :l [ 0 2
(b) 2 0
2 2
(c) [ ~!
-2
2
-4
3
~2]
1 2 0 0
~J
3
2 1 0 0
(d) [ i6 -2
...:.a
(e)
0 0 1 2
0 0 2 1
227
(c) - a: 2 + y 2 + 2VJxy + 3y = 0
5. \Vrite the standard equations of the quadric surfaces
r
L.
-2
1
1
1
-2
1
1
1
-2
l
(a.) Find eigenvalues and eigenvectors of A .
(b) Determine whether A is invertible or not.
(c) Find, if possible, an ivertible matrix P for which p-l AP is diagonal.
\Vrite this diagonal matrix.
(d) Find the genNal solution of the system of differential equations:
x; = -2x 1 + X2 + xa
x; = x 1 - 2x2 + xa
x; = x1+ x2 - 2xa
~Chapter I
5
LINEAR TRANSFORMATIONS
In every branch of mat hemal irs 0110 introduces some objects, investi~?;atf's
To test whether it is a. Jirwar transformation or not we check the con <lit ions
(a) and (b) of the definition.
L(u+v) = L(xt+Yt,X2+Y2,xa+Ya)
((.1't + yi) + (x2 + Y2), (x2 + Y2) + (xa + YJ)
(;z·1 + X2 + Y1 + Y2, x2 + X3 + Y2 + Ya)
L(u) + L(v) L(xh x2, X3) + L(yh Y2, Ya)
(.rt + x2, X2 + x3) + (Yt + Y2, Y2 + Y3)
(:z·t + x2 + Yt + Y2, X2 + X3 + Y2 + Y3).
L(cu) = L( cx 17 cx2, ex a)
L( c(Xt, x2, x3))
as n'qnired. Therefore L satisfies the requirem('nts (a) and (b) and conse
qn('nt ly L is a linear transformation.
wll<.'r('as
Thus
L(u +v) i- L(u) + L(r).
Theorem 5.1.4. Let V and H' be vector spaces and let L : V -+ lV be a
mapping. Then Lis a linear transformation iff
and consequently
Conversely if
L(au + bv) = aL(u) + bL(v)
is satisfied then
is not linear. Dy Theorem 5.1.'1. 1n SP(' this it is enough to find a specific vector
v for which
and they arc not equal. Note th<-lt ])('r<' J,(O,O,O) = (0,0,0) but Lis not linear.
Examples of linear transformations are abundant. Now, we shall exhibit
some of them.
GEOMETRIC EXAMPLES
X!
Further for each u == (;1· 1 ,x 2 ,x3 ) in R 3 and for each scalar c we have
cu = (cx 1 ,cx 2 ,cx 3 ) and hence
is called the reflection with respect to the x 1 x 2 -plane and it is a linear trans
formation.
2.15
In fact for vectors u = (x 1..r:J.a· 3), v = (yhy2,y3) and for any scalar we
have:
= R(u) + R(v)
given by
is obta.ined by rotating each wc1or ·z· through a fixed angle (J because the angle,
between any v =(xi. x 2 ) and RH( r·) = R 8 (x 1 , x 2 ) is computed from
to be fJ.
Examples 5.1.7. (a) The' mapping D from the vector space of differen
tiable functions in an interval (a. b) into the set of all functions defined in (a, b)
236
which maps each function f to its derivative f' is a linear transformation since
](x) = lx f(x)dx
is obviously linear.
CANONICAL TRANSFORMATION
L(X) =AX.
0
0
0
Ei = - j - th
1
0
l
will be a column matrix say
.r\ IJ
[ : = Ai
rl~,j
and we get
L(X) !,( [
;l~j
.tn
.:
l
)= L('f:, :ri Ei)
[.-\J"'Ai"'An] [
XJ
.:
:In
l
= :IX.
In the next section we shall try to express all linear transformations of finite
dimensional vector spaces in this canonical form.
238
and hence
we obtain a map from ll to liV. Note that since c1 , ••• , Cn are uniquely deter
mined for v there is only one vector c1 w 1 + · · · + Cn Wn to take as L( v ). This
map is a linear map because for
we have
L(u + v)
2.3.9
L(u) + L(v)
[:
1 1
0 0 a
0 0 1 0 b
1 0 0 1 ('
l [~ -1
1
-1
1 0 0
1 0
0 -1 0 1 c- a
b~ a]
B {(1,1,1),(L0.1),(1,0,0)}
/-.,
rx2p, 1,1)+ (,7:3- x2)(1,0, 1)+ (xl- x3)(1,0,0)
and
w 1 =(-2,3), w2 =(1,1) and w 3 =(a,b)
where P'(t) stands for the <!Privative of P(t), is a linear operator on the
space of polynomial functions.
(a) (1, -1, 1) = (2, -1), T( I. I ,0) = ( 1, -1), 1'(3, -1, 2) = (5, -3)
1'(1,1,1)= (a+b+c)(l.l)'?
T( [ 1 0
0 0
l[ l [ l [ l[ l
)= 2 -1
1 1
, T( 1 1
0 0
) = T( 1 1
1 0
)= -1
-1
1
-1
and
T( [ : : l) ~ ~ l·
= [
Compute
T( [
(I
r·
b
d
l )
2.f2
T( 1 - x + x2) = [ a a
-b 2
l
T( -1 + x + 3x 2 ) = [ be -22]
T(x 2 ) = 0 1
[ 1 d ]
ThE>orem and Definition 5.2.1. L('t \' and 11' IH· fiuit<• dinH•usioual
V<'ct or spaces with bases B = { 1' 1 , •••• 1' 11 } and C = { 11• 1 , ••• , w,,} r<•:,qwct in•ly.
For !'acl, litH'ar trausforlllatiou {,: \ ' - 11.· thPn' <•xists a uuiqu<' 111atrix .-\
such t !tat
fin· all r 111 \ •• This nr;tl rix :I is ral!Pd /}!( nwlriJ· of/, n la/i('( to /ftc pair B.C
of IHisrs. and is d!'Jt011•d h.v [rJt'· It is tl1!' 111 X 11 rnatrix ~i\'Pil by
/_( ,.'2)
\\'hl'll \ · = 11· and 8 =C W(' sp1•ak oft h(' 1//(/ff'i.r of t/1( /inmr OJ}( mlnr /,
N lulit·r loB.
(I)
for allt·tn 1·. Sinn• [t·]L, is art 11 X l tllatrix a11d [!_(t·)],· is an 111 x l rnatrix \\'('
first rtoff' that .-ltnnst IH• all 111 X. II IJI<tlrix. l.1•t .-llt<t\'(' r()llltlllls .-1 1.. 1'2 ..... .-1,.
In (1) taking v = v; we get
Thus A must be the matrix whose ith column A, is the coordinate matrix of
L( v;) relative to C namely the entries of the ith column are coefficients of WJ.
's when L( v;) is expressed in terms of WJ. 's.
Existence: Now take A to be the matrix whose columns are the coordinate
matrices [L( v1))c. Then for each v in V we have a unique expression
giving
L(v) = c1L(v1)+···+cnL(vn)
[L( v )]c cl[L( vl)]c + · · ·+ Cn[L('Vn)]c
[(L(v.)]c · · ·[L(v.)]c] [ :~ ]
= A[v]s
[ = {(1,0,0.0).(0.1.0.0),(0,0,1,0),(0,0,0, 1)}
c = {(I. 1). ( 1' -1 )}.
To this <'nd we exprPss T(t'd = (1,0), T(11~) = (0,-1), T(v 3 )
(1,0), T(v.d = (0,-1) in t('rms of u· 1 = (1,1) and w~ = (1,-1):
[
1 1 I
L -1 0
0
-1
l
0 I /'2
1/2
Thus we obtain
I 1
T(t't) (1.0)= -(1,1)+ -(1.-1)
2 2
I 1
T( I'~) (0.-1)= --(1.1)+-(1,-1)
:2 2
I 1
T(z·:d (1.0)= -(1.1)+-(l.-1)
:2 2
I I
T(1•.1 ) = ((l. - I ) = - :{1. 1 ) + 2( I. - 1).
Taking th<' nwfrici<•nts to IH'
;1
<'111
= [
riPs of columns we gd tlH' matrix of Las
:!l
&
- :!l
& &
2l - 2I
~
l.
.!:]
l
1. _1. 1. _1.
[/,(I' )]c = .I[ 1']t = [ :!
1.
".!
:!
1.
2
:!
l
2
:!
1.
2
:r2
;1:3
;r4
-246
L(xt. x2, xa, x4) = (x1 + xa, -x2- x4) = ~(x1 - x2 + xa- x4)(1, 1)
1
+2(x1 + X2 + Xa + x4)(1, -1)
0 0 0
L( 1 ) = .t1 1 = [A1 ···A;·· ·An] 1 =A;.
0 0 0
0 0 0
Thus this matrix is precisely equal to A. Actually the aim of representing
linear transformations by matrices is to describe the linear transformation in
the form
L(X) =AX
. by means of coordinate matrices of vectors. If L is the transformation from
R.n to R.m given by
namely in the form L(X) = AX and its matrix relative to the standard ha.ses
is A.
le-J:.J EXERCISES
1. Let L : JR3 -+lR 2 be the linear transformation for which
L(1, 0, 0) = (-2, -3), L(O, 1, 0) = (3, 2), L(O, 0, 1) = (1, -1).
a) Find L(l, -1, 2).
a) Find L(2- t + t 2 - n.
b) Compute the matrix of/, r<'lative to the basis B = { 1, t, t 2 , t 3 } (namely
3. Let
L(X) =AX- XA
4. LC't V be a finite dinwnsional vector space with bases B and B' and let
P, be the B to B' chan~" of basis matrix. The operator I: V __,. V givC'n
by J( v) = v for all v in l" is ra ll<'d the identity operator. Find the matrix
of I
a) relative toC
b) relative to B'
250
6. Find the matrix of the reflection (ofiR 3 ) with respect to the x 1 xTplane
relative to the standard basis.
7. Find the matrix of the reflection (of R 2) with respect to the x 2-axis
relative to the standard basis.
8. Find the matrix of the rotation (of 1R 2 ) through the angle e relative to
the standard basis.
L(v) = .\v.
Find a basis for R 3 consiting of v's satisfying this condition and deter
mine the matrix of L relative to this basis.
2/'il
The aim of this appendix is to gin' a rigorous proof of the fact that funclions
of the form xneax cos bx and xmeu sin d.1: constitute a basis for UC-functions.
For this purpose we make use of rompl<'x valued fnnrtions. All functions that.
we are going to consider will be dPfilli'd in the intNval (-oo,oo) and as su('h
this domain of definition will not IH' specified.
If u and v are differentiable real valued functions then the derivative oft h<'
complex-valued function <p given hy
where a 0,ab ···,an and a are complex numbers and p(x) = p(x) + ap(x) is
a polynomial which has the same dPgree as p( x) when a -:/= 0.
and
is linearly independent.
Proof: Suppose on the contrary t.l1at the set under consideration is linearly
dependent. Then there exists real numbers Aii and Bkl such that
where some coefficient is non-zero. Tl1a.t is, a sum of non-zero terms of the
form Axnea"' cos bx and Bxmeex sitl da: will be equal to zero. Collecting terms
involving the same ea"' cos bx or the same ec"' sin dx such an equality can be
brought to the form
2::: Pk(x )eakx cos bkx + 2::: Q,(x )ec,x sin b,x = 0
k I
25.'3
where Pd .r) and Q1(x) arc non-z0ro polynomials. By making the substitutions
ei8 + e-i&
cosO=---
'2
deriwd from r.i 8 = cos(} + -i sin B and £ -iB = cos() - i sin(} we obtain a rcla tion
of the form
11
Lfl(.t)("kX =0
A'=l
where the pk( x) are non-zero polynomials with complex coeffici<'nts and
o 1 ,o 2 ,···,on are distinct compi<'X numbers.
Thus, if the set under consid('!'il1 iou is linearly dependent there exists non
zero polynomials p 1 ( x ), · · ·, p,( :r) with complex coefficients and distinct com
plex numhei·s oh · · ·, o,. so that
Among all such relations chaos<.' the one for which n is minimal.
Multiplying this equality through E-o.x and letting f3k = ok - o,. the
relation becomes
n.-1
L pk(.r)rl'k·r + Pn(x) = 0.
k=l
Differentiating l+deg (p,.(.r)) tinH'S and using (iii) above the last relation
yields
n-1
LPk(x)e.6•x = 0, with deg(f3k(x)) = deg(pk(x)).
'k=l
So there is a relation with n-1 non-zeo tNms which contradicts the minimality
of n. This contradiction comes from 1he assumption that the set given in the
theorem is linearly dependent. So 1his set must be linearly independent. 0
254
LAGRANGE'S INTERPOLATION
Now, we exhibit another basis for this space associated to any n + 1 distinct
sclars a 0 , a 1 , ···,an. Consider the polynomials
(x - ao) · · · ( x - ak_l)( x - ak+r) · · · ( x - an)
Pk(x) = (ak- ) .
ao · · ·(ak- ak_t)(ak- ak+I) · · ·(ak- an)
;k = o, .. ·,n
Hen\e, therelation
n
l::.:CkPk(x) =0
k=O
implies that for each j = 0, · · ·, n we have
n n
0 = L ckPk(aj) = L ck8kj = Cj.
k=O k=O
Thus the polynomials P0 (x ), · · ·, Pn(x) are linearly independent polynomials
of degree n. Since the dimension of the space of polynomials of degree :S n
is 11 + 1 these n + 1 linearly independent polynomials form a basis. Namely,
every polynomial P(x) is a linear combination of P0 (x),P1 (x),·· ·,Pn(x):
n
P(x) = I:ckPk(x).
k=O
255
k:O k=O
P( -1) = 2. r((l) = 2, P( 1) = -1
IS
where
So
P( X) = .2 . :r( .1· 2- I) ,
- 2 . ( .!' + I )( ./' - 1)
(a: + 1 ).r = - -X
3 2
-
3
-X + 2.
2 2 2
256
additive inverse 2
Cramer's rule 90
adjoint 88
alternating 79
d0gree 105
augmented matrix 53
df'terminant 76
diagonal 2.5
basic variable 55
differential equations 202
. basis 121
dimension 126
echelon matrix 28
cof'ffident matrix 53
free variahlf' 1R5
cofactor 88
Fourier roefficif'nts 191
consistent 53
GAUSSIAN Plimination 55
generator 109
linearly depend<'n t 119
Gram-Scdmidt 181
linearly indPrwnt 11 D
LU-decomposition ()()
homog0neous system 60
identity elenwnt 2
magic sq u arc !)
identity matrix 7
main diagonal (j
matrix 6
inconsist<'nt 5:l
inwrtihlc ,12
minor /.1
multiplicativ<' invcrs<' :l
Kron<'ckN <klt.a 7
n-lin<>ar function 7S
leading <'ntry 28
orthogonal 1GI
kngt h 1(i7
258
partitioned form 18
subspace 108
'
polynomial .function 99
symmetric matrix 28
projection 233
system of linear equations 53
reflection 234
transpose 8
right inverse 42
triangle inequality 169
rotation 235
trivial subspace 108
row operation 33
Sarrus' rule 75
Vandermoncle determinant 98
scalar 3
vector space 100
scalar matrix 26
vector 100
skew-symmetric 28
zero element 2
solution .53