An Introduction To Linear Algebra by Krishnamurthy Mainra Arora PDF
An Introduction To Linear Algebra by Krishnamurthy Mainra Arora PDF
LINEAR ALGEBRA
V. KRISHNAMURTHY
Professor of Mathematics
\
V. P. MAINRA
Assistant Professor of Mathematics
and
J. L. ARORA
.Assistant Professor of Mathematics
All at the Birla Institute of Technology and Science,
Pilani
lm
AFFILIATED EAST-Wl!ST PRESS PVT LTD
NEW DELHI · MADRAS
Priae Rs 22.50
PUFACI ii
1 SETS AND FUNCTIONS 1
1.1 Sets 1
1.2 Operations on Sets 5
1.3 Relations 8
1.4 Functions 10
1.S Binary Operations 11
1.6 Algebraic Structures 20
1.7 Operatio~s on Functions 24
2 VECTORS 33
2.1 Space Coordinates 33
2.2 Vectors-Addition and Scalar Multiplication 31
2.3 Dot Product of Vectors 48
2.4 Applications to Geometry 55
3 VECTOR SPACES 63
3.1 Vector Spaces 63
3.2 Subspaces 69
3.3 Span of a Set 14
3.4 More about Subspaces 11
3.5 Linear Dependence, Independence BS
3.6 Dimension and Basis 95
4 LINEAR TRANSFORMATIONS 107
4.1 Definition and Examples 101
4.2 Range and Kernel of a Linear Map 114
4.3 Rank and Nullity 118
4.4 Inverse of a Linear Transformation 122
4.5 Consequences of Rank-Nullity Theorem 111
4.6 The Space LC,U, Y) 129
4.7 Composition of Linear Maps 133
4.8 Operator Equations 1J8
4.9 Applications to the Theory of Ordinary U1111r .
Ditrcrential Equations 141
Viii / CONTl!NTS
5 MATRICES 141
5.1 Matrix As4iociat~d "ith a Linear Map 148
5.2 Linear Map Associated with a Matrix 154
S.3 Linear Operations in M,,., ,. 158
5.4 Matrix Multiplication 16.3
s.s Rank and Nullity of a Matrix 174
S.6 Transpo.,e of a Matrix and Spec ·al Type4i of Matrices 180
S.1 Elementary Row Operations 185
S.8 Sy~tem~ of Linear Equation, 195
S.9 Matrix lmer~ion :00
6 DETERMINANT5 203
6.1 Definition 203
6.2 Fundamental Properties of Oelerminanb 207
6.3 Proof, of Theorems 212
6.4 Cofactors 2 /6
6.5 Determinant Minor., and Rank of a Matrix 214
6.6 Product of Determinanb 227
6.7 Applications to Linear Equations 230
6.8 Eigenvalues, Eigenvector& 233
6.9 Wrom,kians 238
6.10 Cro<i, Product in VJ 241
7 MORE MATRIX THEORY 250
7.1 Similarity of Matrice!-. 250
7.2 Inner Product Space& 258
7.3 Orthogonal and Unitary Matrices 264
7.4 Application to Reduction of Quadrics 266
APPBNDJX Ordinary Linear Differential Equations 276
ANSWERS TO PROBLEMS 297
INDEX 325
Preface
V. Kri&hnamurthy
Pilani V. P. Mainra
September 1938 J. L. Arora
Chapter 1
I.I SETS
The meaning of 'set' is intuitively understood as a well-determined
collection of objects, called its 'members' or 'elements'. The objects
(members or elements) are said to 'belong to' the set or to be 1in' the set.
Here all the words in quotation marks are taken to be undefined terms.
To illustrate the meaning of •set', let us consider some examples.
Example 1.1 The collection of the lhree boys : Arun, Mohan, and Ram.
Example 1.2 The collection of the three symbols : 6., □ and a.
Example 1.3 The collection N of all natural numben.
E,ca,nple 1.4 The collection Z of all integen.
1?%t1111Ple 1.S The collection Q of all rational numben.
&am,le J.6 The collection R of alt nal numbers.
Eample 1.1 The collection of all the past presidents of Dldia.
B:u,mple 1.8 The collection of all the tint year 1tudent1 of Birla Institute
of Teohnoloo and Science (BITS).
l I sm AND FUNCTIONS
Example 1.9 The aggregate of the livina men in the world whose height
exceeds 2 metres.
Example 1.10 The aggregate of the roots of the equation x'' - 1 - 0.
Example 1.11 The aggregate of the integers that leave a remainder 2
when divided by S.
Example 1.12 The group of cricketers who were out for 99 runs in a
test match.
Example l.JJ The coJJection of all positive prime~.
Example 1./4 The collection of derivatives of all ord,rs of the function
et-.
Example 1.15 The collection C of all complex numbers.
All these are examples of sets. We can construct i-everal such exam-
ples. Let us now consider two collecuons which are not sets :
. (i) The collection of some natural numbers.
(it) The collection of the politicians of India.
In (1) it is not clear which numbers are included in the collection.
Whether the number 2 is in the collection or not cannot be answered
without first explaining the word 'some'. Again, in (1i) the question
whether a specific person is a politician or not would get d1trerent responses
from different persons. Thus, collections (i) and (ii) arc not 'weJl deter-
mined'.
In contrast to these two examples it is worthwhile to analyse Examples
1.12 and 1.13. In Example 1.12 we note that every cricketer is either in
the group or he is not in the group. We do not have to check any records
to say this. In Example 1.13, again, either a number is a prime or it is not
a prime. Here it matters little whether it is known that a particular number
ia a prime or not. In other words, it is immaterial whether we can
answer the question : Is this particular object in the given collection or
not ? What matters for specifying a set is to know unambiguously that
only one of the two answers is t,ossiblc : The object' in question belongs
to the aiven collection or it does not bclona to it.
Thus, we can elaborately describe a set as a collection of objects
which is well determined in the sense that, for every object, there should
be only two possibilities available unambiguously: either it belonas or it
docs not belona to the coJlection.
If .4. is a set and" is an element of .4., then we write x E .4. (E is read
as 'belongs to' or 'ia in' or 'ia an eleme11t of' or 'ia a member of'). The
noption of tbia ia denoted by x tie .4. ( te i1 read aa 'docs not belong to' or
'i• not in' or 'is not an element or or 'ia not a member of').
J.n•tead orauch a detailed description of aeta, two kinda of symbolic
descriptiOJla are very oftn used. OU is by listina, if possible, all the
1.1 sm / 3
-
elements of the set within braces, e.g. sets of Ex11mples 1.1 and J.2 aro
respectively written as
{Arun, Mohan, Ram}
and {.t.i.. □. n} .
The other is by using a dummy element, say x, and uiting the character-
istic properties of x which precisely make it an element of the set. Thus,
the set of elements characterised by the prorerties, say P, (!, ... , is written
{A I x satisfies P, Q, ... }
or {.'t : x satisfies P, Q, ... } .
(' I • and ·:'are reeJ as 'such that'.) Thus, the sets in Examples I.II and
I. 12 may respectively be "ritten a~
{x I xis an integer and x -= Sk + 2 for some integer k}
and { X .• x• is a cricketerh who was out for just 99 runs} •
1n a test mate
Note that in \the first method the order in which the elements arc listed
is immaterial. Thus, {Aron, Mohan, Ram} and {Mohan, Aron, Ram} are
the same sets. We shall make this precise by defining the equality of
sets.
1.1.t Deftnltion Two sets A and Bare said to be equal if every member of
A is a member of B, and every member of Bis a member of A. In
such a case we write A = B. ·
For example, {O, 1, 2, 3} = {2, 1, 0, 3}. Also Z = {x Ix is an in-
teger} = {... , -3, -2, -1, 0, I, 2, 3, ... }.
1.1.2 Convention All our definitions are 'if and only if' (ijf) statements.
For example, if a definition reads 'A triangle is said to be equilateral
if all its sides are equal', we actually mean that 'I\ triangle is said to
be equilateral (ff all its sides are equal'. . .
Jo, view of_ this convention, Definition I. l. I means that two sets A and
B are said to be equal iff every member of .4 is a member of B, and every
member of Bis a member of A ..
SUBSETS
Let A and B be two sets such that every member of A is alao a member
o( B. Then .4. is said to be a subset of B. In symbols we write A C B
('C' is read a■ 'is a subset of' or 'is contained in'). Whenever .4. is. a
subset of B, we also say that Bis a superset of .4.. In symbols we write
B ::::> ..4. (1 :J' is read as 'is a superset of' or 'contain■'). •
Obviously, every set ia a subset (superset) of itself.
4 / lffl AND FUNCTIONS
EMPTY SET
Consider the set of women presidents of India elected before
31st December 1974. There was no such woman president. So this set
bu no members in it. Such a set, i.i,. a set containing no elements in it,
is called an empty set or a null ,et. It is denoted by +.
It may be noted here that there is only one empty set. As a clarifica•
tion of this, note that the set {x : x is a real number satisfyins x• + 1 = O}
is also empty and we can write.
The set of all women presidents of India elected before
31st December 1974 = ♦
== {x : x is a real number satisfying x 2 + 1 - O}.
+
The set {x : x is a real number satisfying x1 x - O} consists of only
one member, namely, zero. So this set can be written as {O}. Note that this
is no, the empty set.
Nonempty sets that consist of only a single member are called singleton,.
The set in Example 1.14 is a singleton as itconsists of only one element e".
It would be interesting (at least for cricket fans) to find out whether the
set in Example 1.12 is empty or a singleton or bas more· members than
one.
+
Finally, we note that C A for all sets A. Thus, given a set A, it has
two extteme subsets. One is A itself and the other is +. Any subset
+
of A other than A and is called a proper mbset of A.
(d) z i + 2z + 2z -
S == O
6. Determine all the subsets of the set :
(a) {O, 1, 2} (b) {«, ~. y, 8}.
7. Let X be a set containing n elements. What is the total number of
distinct subsets of X ?
8. Determine whether each of the following statements is true or false
(N, Z, Q, and Rare defined in Examples 1.3, 1.4, 1 5, and 1.6, res•
pectively):
(a) NC Q. (b) {N, Q, Z} C R. (c) NC {N}.
(d) N E {N, Q, Zt (e) R C {N, Q, R}.
9. Given .4 = =
{O, 1, 2, 3, 4}, B -= {x E NI x ~ IO}, C {x E R Ix~ 1O},
determine whether each of the following stateu,ents is true or false :
(a) A C B. (b) BC C. (c) CC .4.
(d) 1 ~ B. (e) -1 E B. (f) -1 EC.
(g) 3 e: B. (h) C C B. (i) 8 = C.
,1.2 OPERATIONS ON SETS
The standard operations on sets, which yield new sets from old ones,
are (i) union, (ii) mtersection, and (iii) complementation.
Given two sets A and B, the union of A and B, written as A U B, is
defined as
A U B = {x I x e: A. or x E B} .
Here •x E A or x E B' means xis in A or in B or in both. Throughout
this book the word 'or' will be used in the inclusive sense as here, ex\!ept
when otherwise stated.
The intersection of two sets A. and B, written as A n B, is defined as
A. n B == {x I x E A and x E B} .
The complement of B in A, written as A '-. B, is defined as
A '-. B =- {x I x E A and x ~ B} .
A.'-. Bis also denoted by c,.(B).
Example 1.16 Let A - {I, 2, 3, 4, 5, 6} and B = {O, -1, 2, -3, 4, -5, 6}.
Then
A U B = {-5, -3, -1, 0, 1, 2, 3, 4, 5, 6}
An~=- {2, 4. 6}
A'\. B = {1. 31 S} and B'\.A. - {O, -1, -:3, -5}.
The geometrical iUustrations in Figure I.I, called Venn diaarams, 'are
helpful in undemanding theso operations.
6 / SETS AND FUNCTIONS
CARTESIAN PRODUCT
Let A and B be two sets. Consider the set of all ordereJ pairs (x, y),
where :x E .4 and y E B. This set is called the ,:artesian p1wJuct of sets
A and B, written as A x B. In symbols we write
A x B - {(x, y) I x E A, y E B} .
Example 1.17 Let A = {l, 2} and B = {x, y, z}. Then
A x B = {(1, x), (2, x), (l, y), (2, y), (1, z), (2, z)}.
Example 1.18 Let R be the set of real numbers. Then R x R (also written
as R') is the set or all ordered pairs (x, y}, where x and y are real num:
bers, that is,
R x R == {(x, y) Ix, y are real numbers}.
Geometrically, R represents the set of points or a straipt line (called
the real line), and R x R represents the set or points of a plane.
(a) A u B An B
(b) (c) AnC
(d) BL1 C (el A" B (f) B" C
(g) C'., A (h) A < B.
2. let .1 =-- (O, 1, 2, 3, 4}, B -= l~ E N I x -~- 20},
C - t.\ E: NI~ 3}, D = t~ E N ! x 1" d1\IS1ble hy 7f.
Determmc
(a) A I J B (h) AuD (c) Br IC
(d) 8(1 D (<') A I IC (fl A n D
(g) A'-. B (h) A.'\./> lll B'.,C
(J) B '"\. D (k C ' . D t I) AX B.
3. If the c;eti, A, B, C, and Dare a, defined in Problem 2. then deter-
mme ,,hethcr each of the following citatcments 1s true or fal!ie:
(a) B I C r _ D. \b) A .... B. (c) A C C.
(d) A I 1 /J - +. (e) A { ( r: B (f) C''\. A C D.
4. Let A - [u, ~. ·r}, B = f~. ,.,, O}, C ~", ~-. e:}, Determine
(a) A '< R (b) A y C' (c) B ;< C.
5. Dc~cr1be }he follo\\mg ,;ubsets of R
(ct) { ~ I , > 7 J l J [A I x < OJ (b) fY I x .. 1} U {x I x < I}
tc) 'x Ix - l} J t~ I\ l} (d) (r Ix >-1: n {x Ix< 1}
M {x : x < - 11 r) {x !x > I} (f) {x I x > 11 n tx I x ), O}.
6. If A, /J, C, and D are any seh, prove that
(a) (A X B) ("I (C X D) == (A. (' C) X (B ft D)
(b) (A ,,, B) .J (C X D) 1...: (A U C) x (BU D).
7. Give an e>.ample in which
(A X B) U (C X D) :;c (A U C) X (B U D)
8. lf A, B, and C are an) sets, prove that
(a) A U (B n C) = (A U B) n (A U C)
(b) A n (B l.J C) == (A n B) \) (A n C)
(c) AU (BU C) = (AUB) UC
(d) A n (B n C) =-= (A n B) n C.
9. Prove DeMorgan's Theorem:
(a)
(b)
s'
If A and B are both subsets of a set S, then
(A u B) = (S '- A)
S'- (A n B) = (S '\. A) U (S ' . B).
n (S' B)
1O. True or false ?
(a) If .4 and B are two sets, A n B is the largest subset of A U B1
which is contained in both A and B. '
(b) If A and B are two sets, f U B is the smallest superset or .4 n B.
'Which contaias both .4 and B.
8 / SBTS AND FUNCTIONS
1.3 RELATIONS
Let ,4 be a nonempty set. A subset )If of Ax A is called a relation
on .4. If (x, y) E JI, we say x 1& related to y by the relation )If and
symbolically we write x )l y. For example, JI = {(I, 2), (l, 3), (2, 3)}
defines a relation on the set ..t = {1, 2, 3}. Here I JI 2, 1 :7f 3, 2 :rt 3.
Obviou1ly, it is the usual relation'<', because 1 < 2, l < 3, 2 < 3. On the
same set A, the relation•~• is described by the set {(l, l), (l, 2), (1, 3),
(2, 2), (2, 3), (3, 3)}.
We note that'<', '=', '>',•~•.etc., are relations on R, N, Z, and Q.
'Is the mother of', 'is the brother of', 'is married to' are relations on the set
of all human beings.
1.3.t Definition Let :7f b~ a relation on a set ..t.
(a) If x Jlf x (1 e. (x, x) E Jlf ), for every x E A, )If is said to be a
r,jlexi,e relation.
(b) If, whenever x )I y, it ,s also true that y JI x (i.e. whenever
(x, y) E JI, (y, x) also belongs to JI), 1l is said to be a symmetric
relation.
(c) If, whenever x Jf y and y :JI z, it is also true that x 9f z (i.e. if
(x, y) E 9'l and (y, z) E 911, then (x, z) E 911), 911 is said to be a
transiti,e relation.
(d) A relation 911 on A that is reflexive, symmetric, and transitive is
called an epl,alence relation on A..
We shall consider several examples of relations.
Example 1.19 Let ..t ... Z, the set of all integers Consider the subset 91
of Z X Z defined by
911 = {(x, y) : x - y is divisible by 3} •
Here x 91 y llf (x-y) is divisible by 3.
(a) DI is reflexive, becauso x-y is divisible by 3 for every integer x.
(b) x-y is divisible by 3 clearly means y-x is also divisible by 3.
Hence, :,c II y implies y 91 x. So JI is symmetric.
(c) If x-y is divisible by 3 and y-z is divisible by 3, it is certainly
+
true that x - s - (~ - y) (y - s) is aleo divisible by 3. Thus, x 91 y
and y I! • Imply x 91 •· Hence, 91 la transitive.
1.3 RILAnONS / 9
f.4 FUNCTIONS
The reader D'lusl have already been exposed, in his earlier training, to
a proper definition of functions and operations with functions. Hoy.ever,
as the concept of a function is very fundamental to mathematics and its
applications. we ,hall now give the basic ideas about functions relevant
to our subject.
1.4.1 De&nitioo Let A and B be two nonempty ,ets. A function (map)
f from A to B 1s a rule whlch, to each element of A, associates a
unique element of B. Symbolically, Y.e \\rite
f:A ➔ B.
FIGURI! 1.2
We shall set up a number of con,entions on usage, by means of a
simple example of a function : Con&ider the function /: R ➔ R di=fined by
the rule, which, to each x e: R, associates the value )."~. So /(x) = xJ for
all x E R. By abuse of language, Y.e sometimes say that the function is x 1 ,
and we even use the symbol /(x) for the function. Very often, we use the
more precise language that the· function is given by /(x) -= x 2 , x E R.
Alternatively, we denote this function as the set {(x, x 1) I x E ~ of or-
dered pairs. Thia set is also called the graph of the function; because, if
we plot the points (x, x2 ) in a plane, we will get the geometrical curve
nspresentina the function.
We shall now illustrate the foregoing ideas by listing si=ven ways of
npressina this x1-function. The first four methods arc technically perfect,
aad. methods (v) to (~ii) are sanctified by custom and Ulage.
(i) /: R-=+- R defined by x t-+ x•.
(ii) •/: x ..,... x• for all x E R.
u FUNCffONS I 1J
(iii) /: R-1 ➔ R defined by /(x) a= x•.
(iv) / == {(x, x 1) I x E R}.
(v) The function x= defined on R.
(\'i) The graph of the function is {(x, x 2 ) I _x E R}.
(vii) The graph of the function is as in Figure 1.3.
y
fJGURB 1.3
Note that the symbol ·t--►' used in (i) and {ii) is different from the
symbol •-• for a function. The symbol '-+' is used when we want to
specify the domain and target set off, \\ hereas 'f-+' is used when we want
to indicate the image of a single element of the domain. Thus, x t-➔ x•
means that the image of x is x~. On the other hand, when we write
/: A -+ B, we are not specifying the image of a single element.
Though we shall allow oursel\'es the liberty of using any of these
methods to sp1:cify a function, in the sequel, more often than not, we shall
adhere to methods (i), (ii), and (iii). In case there is doubt while using the
other methods, the reader should go back to the first three methods for
clarification.
It may be noted that, if it is possible to list, without ambiguity, all the
values of /(x) as x varies in the domain, we do so, and we also- ri.ay that
this list is the function.
Now we emphasise the word 'unique' occurring in Jhe definition of a
function. for each x E A, f(x) should be unique. ln other words, f
cannot h.ave t\\o or more values at a given x. For example, if A - {1 1 21
3, 4, S} and B - (-1, 2, -3, 4, -5, 6, O}, then the association 1 ...... -1;
12 / SliTS AND 'FUNCTIONS
• of all naJ polynomials of delfw not areater than II i1 dlnotecl l,y g.,.
Note tbat tbe zero polynomial, 0, belonp to botb !I and g.. Tbe delftl8
of tbe ~ polynomial'i1, by con,eatioo, uaipod to be -ai,. Similarly,
1: C .... C aiven by JI(~) - a.+ 11sz +...
+~.where a.. ai, .•.• a. ue
complex numbers. ii ca.Jled a complex polynomial on C.
&.nple 1.40 Define/: U.➔ Jt by /{J,) -p(O) for all p e fl. For inl,.
taace, ir the polynomial is p: x ...., 2r + I, then ft.I,) - I. Note tbal
tbia is not a coastant function. (Why ?)
Bzam,le 1.41 Let N bo the set of natural· numben. Then any function
/: N ➔ R can be written u the set of ordered pain {(1,/(1)). (2. /(2)).
(3,/(3)), ... }. Such a functioa i1 called a Ufl/MIU of real numben. '11111
11quence ii alBO written u •
.f(l), J{2), /(3), ...
Por instance, 2. 4, 6, 81 ••• i1 a aequenoe. It i1 just the function .x-t-+ 1x
onN.
In tbe dilCUUion of any function /, there are four fundamental quea-
tioaa: \
{i) What i1 tbe domain of/ 7
(ii) What i1 th1t raaae of/ 7
(iii) 11 the function •onto' or not ?
(iv) 11 the funetion •oae•one' or not ?
We have already introduced the domain of/. We 1ball now introduce
the conceptl related to other queatiom.
1.4.2 Delaltloa Let /: .4 ➔ B be a function and , e B. The pre-tmap
of y i1 tbe aet {.x E A. I Jtx) - y}.
t.4.3 Rellluk An accepted notation for the pre-imaae of y under/ ia
/·1(1). However, we shall not 111e tbil notation in thi1 IIDII.
Bza,,,ple 1.42 Define /: R ➔ Jt by tbe rule /(x) - x•.
The pre-imap
of 4 ia {2, -2}, and the pre-imqo of O i1 {O}. The pie-imap of-1 i1
the empty set +.
1.4.4 Delllltlea Let/: A. ➔ B be .a funetion. The aet { /(,c) : ,c e .4} ii
called the ,_,. of/ and ii denoted by RCn.
It ii in fact tho let of all/-imapa.
l.4.5 Delaltla A map / : A.. ➔ B ia said to be 01llo (1~ive) if
Jt(/) - B.
Clearly, tbe followina statement ii true.
1.4.6 Fact A •Pl: A ➔ B ii onto f6 either one of the followina proper-
tia bolda: •
(a) For every 1EB, tbae 1ai111 at leut one ~EA IIICh tbu
/(,c)-1. (I)
14 / 51TS AND FUNCTIONS
X1 + Cl = X1 + 11•X1 == X1,
The maps of Examples 1.36 and 1.37 are clearly not one-one, because
all members of A. are mapped into one member of B. In fact, this is tho
extreme case of not being one-one. On the other hand, the map of
Example 1.38 is one-one.
In Example 1.39 whether or not the function is one-one depends upon
the polynomial, because the polynomial p : :,c I-+ x + l is one-one, where.
u the polynomial q : x I-+ x• + 1 is not one-one. (Why ?)
The map of Example 1.40 is not one-one, because (x+l) and {:ic1 +1)
have the same imaae,
In Example 1.41 also the one-oneness depends upon tho sequence.
.... tho sequence 1, 1/2, 1/3, 1/4, ... is one-one, whereas the, equence
1, l, 1, ... is not one-one.
L4.18 Delaldoa Two aetl A. and B are said to be in ope-on, corr,qond-
.,,_ if there nilts a . . . , : A. ➔ B that ia one-one and onto.
IA.II D1l11dia If a Ill .4. ii iD one-one COfflllpOJldenc with the let
n. 2. .....11l. for 10me DOlitive inteaer 11 • ..4 ii said to be a Jlntt, 111.
16 I sm AND l'UNCTICJNS
•
X X
0 0
Ca) Cb)
V V
~
X
0 0
Cc) Cd)
V y
-ioi---------x
0
m
1.5 BINARY OPERATIONS
We shall now introduce a very important class of functions which in•
eludes the familiar operations of arithmetic, namely, addition and multi-
plication. If we look back carefully at the opeiation of addition or
numbers, we find it is essentially a process of combining two number■ and
obtaining another number in an unambiguoua way. Thf: aame ia true of
multiplication. A pneralisation of this idea is the followina definition.
1.5.1 Delllldaa Given a nonempty Mt A., any function from A x A. to A.
is called t. b"""1 opm,tlon OD A..
18 / IBTI AND PUNCTIONS
+ Gt ~ y X Gt ~ y
« « ~ y GC « « GC
~ ~ y Gt ~ Gt ~ y
y y ~ y y
'
Gt Gt
GROUPS
Now we take up another structure in algebra, which is simpler than a
field. Suppose we restrict our attention to only one binary operation on
a nonempty set X and denote it by•+•. If this satisfies axioms Fl, F2,
and F3, we say that the system is a group. If, in addition, F4 is satis-
fied, we call it a commutative group. We shall now give the formal defi-
nition.
1.6.2 Oe&nition Let G be a nonempty set on which a binary operation •••
is defined. Then G is said to be a group, under the operation '•', if
the followiog axioms are satio;fied :
G 1 ••• is associative.
G2 There exists an element e E G such that e • a = a = a • e for
all a E G. e is called an idtntity for •••.
G3 For each a E G, there exists an element x E G sudl that
a• x = e ... x • a. x is said to be an inverse of a for the optta-
tion •••.
If, in addition,
04 The operation ••• is commutative,
then the aroup ii said to be commutatlH or ahllan.
22 / SETS AND FUNCTIONS
ltINGS
Another important structure is that or a ring. Let us once apin
consider the nine field axioms. The first four axioms-Fl, F2, P3, and
F4-are pertinent to addition, whereas the next four axioms-FS, F6, F7,
and F8-are pertinent to multiplication. The ninth axiom, F9, pertains to
both addition and multiplication.
If there is one binary operation on a set X and the four axioms Fl to
F4 are satisfied, then the system is a commutative group. If a set X bas
two binary operations and' all the nine axioms are satisfied, we say tlle
system is a field. Id between these two situations, we llave a ring
structure.
Suppose a set X has two binary operations. Let, for the first operation,
the system 'be a commutative group. The minimum that we require of the
second operation, in order to get a meaningful system with two operations,
is the associativity FS and the distributivity F9.
So whenever· there are two operations, say • +' and •• •, on a set X, and
Fl to F4, and FS and F9 are satisfied, we call such a system a ring.
We shall not give a formal treatment of rings in this book. It suffices
to note that
(i) A set X with two operations •+• and •.' such that (a) it is a
commutative group for •+• and (b) FS and F9 are satisfied is called a
ring.
(ii) A rin1 in which F6 bolds is called a rl11g with Identity (unity).
(iii) A ring in which F8 holds is calletl a commutative ring.
For example, C, R, .Q, and Z are rings under the usual operations of
addition and multiplication. All these are commutative rings and have
unity. The set E of even integers is a commutative rin1 under the opera-
tions of addition and multiplication. But it bas no unity.
Watch for further examples of rings as we go along.
v•ainx
FIGUU 1.4
26 / SITS AND FUNCflONS
'I a cos x
PIGUltB l .S
J-4-~+~--¥-l-.f:-----1Hn:--,..-J--r--.x
= sinx
FIGURE 1.6
.-"".A'l~:--1~-:;;IJ.:"-+---:::E--+-:~--+--7'-::_~x
' ,...~ y
slnx
FIGURE 1.7
A special case of scalar multiplication when 'A = -1 gives ( -1) / as
the function
((-1)/)(x) - (-1) (/(x)) - ---/(x) for all x ER.
( -1 )/ is also denoted by - /.
1.7.5 Definition Let/: ..4 ➔ R and g: .A.-+ R be two real-valued func-
tions. Then /g, called the pointwise product of/ and g, is defined
as a real-valued function
Jg : .A. ➔ R such that x ....,.. /(x) g(x) •
In other words, (/g)(x) = / (x) g(x) for all x E ..4.
Example 1 57 Let/: R ➔ R be defined by x· f-t> x• and g: R ➔ R be
defined by x t-+ x + 1. Then the function Jg : R -+ R is defined by
x I-➔ x• (x + 1).
Example I .58 The araph of the product of sine and cosine is aiven
in Fiaure 1.8,
.v
y • 11nx cosx
1., R•ark
1. We have constructed functions / + g, Af. and /g.' The
construction of / + g from / and g is nothing but an operation
which. to each pair of functions/and g of :1(.A.). aaociatet • func-
tion / + r in f(..4). Tb111, thi• operation is a binary operation on
.:f(A). It ia called addition. Tile acoond operation, that of Iormina
A/ rrom Aand J. called acalar multiplication, ii not a binary opera•
28 / Slff'I AND FUNC'T~S
...
1..
Vectors
This point O is~called the oriain and the lines are called x-axis, y-axis, and
z-a1ii. respectively. These three mutually perpendicular lines, called the
FIGURE 2.1
FIOURI! 2.4
36 / VECTORS
= v ((RQ)• + (P R) + (P Q)
1 2 2 2
FIGURE 2.5
FIGURE 2.6
-
PQ is the distance between P and Q and is denoted by I PQ I . A vector
of unit lenatb is called a u11it ,ector.
Two vccton o; and o•J• are said to be parallel if the line segments
OP and O 'P' are parallel. Two parallel vectors may have the same' or
opposite direotions. In aeometry two parallel vectors are considered to
be equivalent if they have the same len9th and the same direction. In
Pi1ure 2.7 o1. O'I, and i1 are pan.Del wc:ton. and o1 and i'J aro
equivalent vecton, HoWftll', in mechanics the need may arile to eonsider
38/ VECTORS
P'
flOURil 2.7
-+ ~
them as different. For example, OP and RS may represent two forces
acting at two different points O and R. In our study of vectors we shall
ignore this distinction between initial points. We shall assume that there
-+
is always a coordinate system with an origin 0. Given any vector AB,
-+
unless otherwise stated. we shalJ always think of a parallel vector OP
-+ ~
with an initial point O and a terminal point P such that OP and AB have
--+ -+
the same directions and I OP I ==- I AB I , as in Figure 2.8.
y z
B
lal X (bl
-----...v
FIGURE 2.8
-+ -+
Note that, given AB, we arrive at only one OP in this way. Thus, all
our vectors shall start from a fixed point, namely, the origin, in the chosen
coordinate system. Such vectors arc also called free vectors.
y y
o'-----------x
lal lb)
FIGUBI 2.9
2.2 Vt!C~--ADDITION AND SCALAR MULTIPLICATION/ 39
~
Consider the vector OP in a plane, where P is the point with coordi•
nates (a1 , aJ relative to the chosen coordinate system referred to O as
origin (Figure 2.9).
The sum of two vectors -+ -+ is given by OR,
OP and OQ - the diagonal of
the parallelogram whose adjacent sides are OP and OQ (figure 2.10). If
P and Q have coordinates (,1 1, a 2) and (h1 , b1 ), respectively, tbeo elementary
geometry tells us that the coordinates of R are (a1 + bu a1 + b1). (Wby ?)
Suppressing the directed 1ine segments and writing only the coordinates
of tbe vertices, we have Figure 2.10 (b) ,
V V
R "
R<a,•b1,a2+b2>
I
I
•P<•1••2>
I
I
I
---1
I
I
I
-
Q(b1,D2)
I I
X0 l(
<a> lb)
flGUllE 2.10
--+
Geometrically, AOP, with A> O. is defined as the vector whose length
-+-
is ~-times the length of OP, i.e. ~ I OP I , and whose direction is the same
as that of ----- --+ denotes a vector of length _-,... I ~
OP. If ,\ < 0, then .\OP OP l in
-+
a direction opposite to that of OP. As before we get the pairs of diagrams
as shown in Figures 2.1 I (a) and (b), and 2.1·2 (a) .ind (b}. (In these dia-
grams I A I > I.)
An analogous situation arises in space, where the point P bas the
coordinates (ai, a,, a 1 ).
A>O. OP'=AOP V
V
p•
I
I
----1II
I
I
0 )( 0 ' X
Cb)
fIGVRI 2.11
40 / VECTOllB
7'<0, 6P'=A~
V
Vt
p • P(a1'"2>
,------- X 0 X
: 0
I
I
I
p' •Pha,,A•z>
<a> (bJ
FIGURE 2.12
2.2.2 Deftaltloa (a) The set of all plane vectors (i.e. the set of all ordered
pairs of real numbers) is denoted by J"1 •
(b) The set of all space vectors (i.e. the set of all ordered triples of
real nun1bers) is denoted by J"1 •
Note that J"1 is the cartesian product R x R. For this reason J"1 may
also be denoted by Ra. Similarly, Y8 - R X R X R - ~.
2,2,3 Definition (Equality) (a) Two plane vectors (01, a 1) and (bi, b1)
are said to be equal if 0 1 = b 1 and a 1 - b1, i.e. (ai, a.) = (bi, b.)
if a, - b,, I -= 1, 2.
(b) Two space vectors (ah a., aa) and (bi, b1 , b3) are said to be
= =
equal if a1 == "1, a, b1, and aa ba, i.e. (ai, a1, a3) = (bi, b1, ba)
if a. = b,, i = l, 2, 3.
The direction of a nonzero vector in V1 is the radian measure 8,
O ~8 < 2 ,c, of the angle from the positive direction of the .x-axis to
-+ -+
the vector OP measured counter-clockwise. If • == (a1, a.) = OP, then
the direction e of • ls given by
sin O - va, 2 a.
~ a.1 ' cos 8 = v0 1.., + a.•
01
02 Existence of identity
For any • = (a,. a 1 ) in V1 , we have
(0, 0) + (a1, a,.) = (a11 a1 ) -= (ai, a 1 ) + {0, 0) •
So the zero vector (0, O) plays the role of identity. The zero vector
usually denoted by O.
03 Existenc.e of inverse
Given a vector 11 - (a1, a,), the vector z - (-ai, -a.) satisfies
z+•== 0 == ■ +.¥.
This vcctur " is called the additive in"lerse of • or the negati,e of • and
ii denoted by-•·
04 Vector addition is commutative
Let u - (a1, a.) and • =
(61 , b1) be two vectors in Y1 • Then
• + , ==
(01 +
61, a1 + b1)
aad •+• -
(61 +
a1, b1 +
a.) .
But we know that for real numbers a 1 , a 1 , 61, b8 (cf§ 1.6)
+
a, + b, === b1 a1, I - 1, _2.
Therefore, • +, - , + • for all •• , E Y1•
This completes the proof that v,.
under the operation •+\ ii a
commutative group. I
It may be noted that the additi'vo identity of the group Y• ii &ho vector
(0, 0, 0). We shall use the same symbol O for tbil vector al10. la "• if
• • (Oa, ._, a.), tben -• - -(Oa, a., Ga) - (-11i,-0p -Ila).
1M D tJlfha (Scalar ""'1tt,Ntot1on) Mvltipliation or vecton ia Y1 by
2.2 VICl'ORS-ADDITION AND SCALAR WLTIPLICATION J 43
a real number is defined u
i\(a1, a1) = (Aa1, ,\a.)
for every (alt a.) E Y1 and every real number A- Similarly, in Y11
A (a1, a.. a 1) -. (A. a1, .\ Os, .\ a,)
for every (a1, a 1 , a 8 ) E J/8 . and every real number A.
Example 2.10 If the vector • = (3, -]). then the vector 2u - (6, -2)
and Ju = (1, -1/3).
Note that scalar multiplication is not a binary operation. (Why?)
However, it does satisfy certain natural properties which we shall list and
prow.
2.2.7 Properties of Scalu Multiplication
For all 1•e,·tors 11. " in V2 ( or in J/3) and real nw11bers at, (i, we have
(a) a(u + v) =-= ar1 + a,.
(h) (.x + ~ II = all + ~u.
(c) a(~u\ = (11~)u = ~(all).
(d) lu = u.
(e) Ou == 0.
Proof: We shall prove (a) as an illustration of the method of proving
these properties. Further, we shall pro,e this m Va only. The rest are
left to the reader.
Let • = (a1 , a 1 , a 3) and , --= (b 1, bi, b.) be two vectors in Y3 , and a.
be a real number. Then
• +, ::::a (a1+ bi, a2 + b1 , aa + ba)
and a(u +- •) = +
(a(a1 +
b1), a(a1 +- b1). a(a1 b1))
= +
(aa1 + ab1• aa1 ab1, ixa3 + ab,) (1)
by properties of real numbers (cf§ 1.6). Further, we have
a■ + a,
=-' a(a1 , "2, 0 0) + a(bi, b,, ba)
== (aa1, «tis, aa:i) +
(arb1, a.b11t abs)
- (C&Di + 0W1, fl"t + @1, Ma + f&ba)
= + «<• •> (by (1)). I
The last two properties. namely, td) and (c), may look trivial to the
reader. In fact, they just say
l(a.. a.. aJ - (lai, la., laa} (Definition 2.2.6)
- (Oi,O.. a.)
and
0(4a, ... -.) - (0.a. Ga.., Goa) (De&nitioa 2.2.6)
- (0, 0. 0).
44 /VEctOU
Let the reader reserve his opinion until be comes to the discussion or
an abstract vector space in Chapter 3. Then he will see the sipiftcanc~
or properties (d) and (e) and also the delicate difference between them.
2.2.8 Deflllitlon The difference of two vectors• and,, written as• - ,, is
defined by• - , - • + {-,).
Exampl, 2.11 u,t • = (3, 4), , = (-1, '3). ~hen• - , = • + (-,)
- (3, 4) + (1, -3) - (4. 1).
-+
We sbaU now find the vector • - , geometrically. Let • = OP
-+
- (ai, a.) and , =
OQ == (b1 , b2). The negative of ,, i.e. -,, ia
-+
(-b,, -61). Geometrically, it is the vector QO which is the same as
-+
OQ' (see Figure 2.13), where Q' is the point whose cartesian coordinates
v·
fIGUlll! 2.13
alO ( -bi, -b1). The vector • - , is the diagonal of the parallelogram
-+
formed by • and ,. It is the vector OR = -QP -= lD1 - b1, a, - b1).
Analogo111ly1 we can find • - , in V1 •
For every nonzero vector ■, the vector r¾i- • is a unit vector in the
direction of•· If • == (a1, a.) :I= 0, the unit vector in the direction of• is
( a1 a1 )
va,• +a.•. vo,• + a.• .
If• ("1, a.. a1) :¢= 0, the unit vector in the direction of• is
( a, a. a1
'"...; "11 + a.1 + 0»1 • v'iil" + a.• + a.• • 'Va,•+ a.•+ a,i).
Example 2.12 If• - (3, 4), the unit vector in the direction of • it:
(3/S, 4/S).
lb«lln/lle 2.13 If • - (1, 1, 1), the unit vector in the direction of • la
(1/4/3. 1/4/3. 1/4/3).
2 2 verroas-ADDfflON AND SCALAll MULTIPLICAnON / 45
In a plane the umt vector in the direction of an angle 8 is (cos e. sin 8),
as shown in Fi1ure 2.14. It is also calJed the unit vector at an angle 8 with
y
P <cos e. s,n e>
IOP1=1
f-lGURE 2.14
the positwe direction of the x-axis. The umt vectors 10 the positive
directions of the x-axis and the y-axii are (). 0) and (O. 1), respectively.
They are denoted by i andj, respectively, 1.e i = (l, 0) and J - (0, I),
as in Figure 2 I S.
V
(0,1>
j
<1,0)
X X
0 0
z z
(0,0,1)
k
X
•
FIGURE 2.15
In apace the unit vectors along tbe positive directions of the x-axis,
the 1-axi1 1 and the r-axis are respectively (1, 0, 0), (0, l, 0), and (0, 0, 1).
They aro denoted b>- i,, J, and i, respectively, i.e. i - (1. d. O),
J - (0, I, 0), and i ""' (0, O, 1). Though we use the same letters I and
j hero aa in the case of plane vectors, thero cannot be any confusion,
becau• tbe context will always ■bow whether we are talkioa of plane
vectors or apaoe vccton. Tbo importance of unit veeton is brouabt out
by the tollowina theorem.
~ 46 /VICTOU
2.Z.9 'J'lleore• (a) ENry pl0111 NCtor Is of th, form a11 + aJ and nay
r,ctor of this form Is a plan, Hctor.
(b} EHry space rector is of the form aif + a. J + a,> and ner:, Hctor
of this form 18 a space ,ector.
Proof: We shall prove (b). Tho proof or (a) is analogous. Let
(ai, "-• aa) be a vector in Y1 • Then
(Os, a., a,) - (a1, 0. 0) + (0, "a, 0) + (0, 0, a 1)
== 01(1, 0, 0) + a.CO, I , O) + a1(0, 0, I)
== 41i + aJ + a,.k .
Thus, every vector in Y1 is of the form a11 + aJ + a1k. On the other
hand, a vector of the form a 1i + aJ + ask is the sum of vectors a11,
aJ, and aJi of Y1 • Since wctor addition is an associative binary opera-
tion, a11 + aJ + a1t is a vector in Y1 , i.e. a space vector, I
Geometrically, Theorem 2.2.9 means that every vector can be represent-
ed as a sum of scalar multiples of unit vectors along the coonimate axes,
as shown in Figure 2.16.
V Z
<•>
flGURB 2.16
The numbers ai, a 1 are called the component, of the plane vector
• - (a1, a.) == a'J.I + a.J. a1 is called the I-component (or x-component)
and a. is called the }-component (or 7-component). We also say that
a.
ai and are the coordinates of • with respect to I and J.
Similarly, a1, a.,
and a. are called tbo co,npoMnts of tho space vector
• - (Os, a., + +
a.> - "1,I aJ ¥• a1 is the I-component, a. is the
}-component, and a. the I-component. We tllo aay that "1, "9, aa4 "a aro
coordinates of• with respect to l,J, and I. •
-+ -+
J1. Let • == OA· and • - OB be two vectors. Then prove that
-+
•- • = B.A.
12. Prove the properties of scalar multiplication (b) and (c) stated
in§ 2.2.7.
13. Prove that a•
= 0 iff « = 0 or • = O.
14. Let • be a nonzero vector in Y1• Then prove that the set
S = {au I « E R} is a group under vector addition.
15. Let u and , be two nonzero vectors in V1 • Then prove that the set
S == {rxu + ~• \ «. ~ E.: R} is a group under vector addition.
16. True or false?
(a) i(2, 0, 4) = (1, 0, 4).
(b) (a + b, b + + + (c, a, b) =
c, e a) (a -+ b + c) (1, 1, 1).
(c) 21 + 3j = 3j + 2i.
(d) ; + J + k is a unit vector.
(e) 0 - (a, b)== (-a, b).
-+ -+ -+
(f) OP - OQ = PQ.
(g) If u and • are vectors, then u bas not been defined so far.
(h) «b = O.
(i) Ou = o.
(j) The zeros on both sides of (e} in § 2.2.7 are different.
-+ -+ -+
(k} Let .ABC be a triangle. Then AB + BC + CA == O.
)(
FIGURE 2.17
or 21
--+
We have QP
--+
OP
- - = -
OP -
- OQ. From geometry, we have
.....
I OP 12 + I OQ lz - 21 OP I I OQ I cos 6 = I QP 11
--+
I I OQ I cos 8 = I -OP I 2 + I OQ
- --+
12 - I QP j2
- -
I (ai, a2, a 3) I2 + (bi, b2, b3) 12
- I (a1 - bi, aa- b2, a 3 - b 3) 1 1
= (a? + al' + a + (b/ + bl + bf)
3 2)
Therefore, - -
I OP I I OQ I cos 8
= 2a1b1
== a 1b1
+ 2a b + 2a b
+ab +ab
Now let 6 = 0 or O = n. In these cases we have OP = ~OQ,
2 2
2 2
3 3,
3 3•
--
where ~ > 0, if~ = 0, and ~ < 0, if 6 = 1t, i.e. ~ = I ~ I cos 6. Thus,
(a 1, a2 , a 3) = ~(b 1, b 2, b3), i.e. a, = ~b,, i = 1, 2, 3.
Hence I~
OP- -
I I OQ I cos 6 == --+ 2
I ~ I I OQ I cos 8
= I ~ I (cos 8)(b/ + b/ + b32 )
= ~(bi'+ bl + bs2)
= a 1b1 + a2b1 + a 3b3 ,
because a, = ~b;, i = l, 2, 3.
A similar result can be obtained for plane vectors. We have only to
(2)
where 6 is the angle between the vectors • and ,.
so I VICJ'ORS
Example 2.14 let • = (3, 2) and , = (2, S). Then the angle 8 between
the vectors• and , is given by
I (3, 2) I I (2, 5) I cos 8 == 6 + 10 .
So 8 16 16
cos = v'l3 v'29 """ v'377
Example 2.15 Let • == (1, -1, 3) and , = (3, I, I). Then the angle
8 between the vectors u and , is given by I (1, ~I, 3) I I (3, 1, I) I
cos 8 = 3 - 1 + 3. So cos 8 = S/11.
Now we shall define the dot product of two vectors. Actually, there
are two different ways of multiplying two vectors. In one the product is
a real number, whereas in the other the product is a vector. In this article
we shall consider the first kind of product. The second kind will be dealt
with in Chapter 6.
2.3.3 Definition Let u = -OP and .., = OQ -+ be two nonzero vectors and
6 be the angle between them. Then the dot product (also called scalar
product or inner product), written as " • ,,, is the real number defined
by
-+- -+-
,, · • = I OP I I OQ I cos 8 = I r, I I ,, I cos 8 . (3)
If either" or Y is the zero vector, then r, • • = 0.
This definition is based on the geometric concept of the angle between
two vectors. However, Theorem 2.3.2 gives an algebraic expression for
the dot product of" and ,, in terms of their components. Hence, Definition
2.3.3 may be equivalently written as follows.
2.3.4 Definition Let " = (a1, a 2) and v = (h1, h1) be two vectors of Y1 •
Then the dot product 11 • Y is defined by
11 • • = a1b1 + a,.b 8 • (4)
If u = (ai, a8 , a 3 ) and , = (b1 , b1 , b3) are vectors of V3 , then 11 • ,
is defined by •
(S)
Note that the formation of the dot product is not a binary operation.
(Why?) However, it satisfies the following properties.
2.3.5 Properties of the Dot Product
Let "• •• w be vectors in Y3 ( or In Y1 ) and«, ~ be real numbers. Then
(a) u · • = I• I 2; hence 11 • • ;> 0. (6)
(b) u · 11 =- 0 i/J• = 0. (7)
(c) • • , = , • 11. (8)
(d) • • (, +
w) =r • • , + ." · w. (9)
(e) (a) • , • cz <• •,) - • . («,). (10)
2.3 DOT PllODUCT OF VECTORS / 51
cose ==
..
So we assume that r, and • are nonzero vectors. Let 8 be the an&le
between• and ,. Then, by Definition 2.3.3, \\e have
l•I l•l
,
But I cos 8 I ~ I, Therefore,
l•·•l~l•I 1•1,
i.e. <• · •>"
~ I • I " I , I 1 = <• · •> (t
· I •>·
Exampk 2.16 The dot product of the vectors (2, 1) and (-1, 6J ia
(2. 1) • (-1, 6) == -2 + 6 == 4.
Example 2.11 The dot product of the vectors (I, -1, 3) and (0, l, -3)
is O + (--1) + (-9) - -10.
Example 2.18 For any real numbers "1, a..
a1 , b1, b1, and b1, wo have
(a,,b1 + a,J,1 + a,J,1)'- <;. ("1.1 + a.• +
a.■) (bi' b11 + b11). +
(Hint : Apply Schwarz inequality to the vectors • - "lai, "-• a.) aud.
, - (bi, ht, ba).)
52 / VECTORS
Therefore,
•·•
"1- ,-:-, , and "• - • - "1 ... • -
.., ,
- ., , (Fi,ure 2.18).
2.3 DOT PRODUcr OF VECTORS / 53
p
.. ..
the vector projection of • along , is
" " "
;:-;•-~r,r
== (scalar projection of • on ,) times tho unit vector
in the direction of ,.
Thus, the magnitude of the vector projection of • along " ;s the absolute
value of the scalar projection of • on " and the direction is the same as
that of ,, if 8 < 1r/2, and opposite to that of"• if 8 > 1r/2. If 8 .... TC/2,
the vector projection of r, along , is O.
Example 2.21 Let ■ == (I, 2, 3) - i + 2} + 3k and , == (-2, 3, 0)
= -21 + 3} + Ok. Then
scalar projection of • on • is
■ ·,
T,T" = -2 +6
v' 13 == v'
4
13
. • .• 4
vector projection of• along , 1s ,.--:-, • - 13 (-2, 3, 0)
8 • , 12 •
= - 13 I ;- 13J ...,.. ua.
I ftL
Abo
8 12 .
u = (1, 2, 3) == i + 2} + 3i =- (- 13 1+ 13 J+Oi)
+ ( f~ I + ~: J + 3k)"
54 / VICTOllS
This is the resolution of • along and perpendicular to ,. The reader
8 12 . ( 21 I
can check that ( - U; + 7"f J + Ok ) 11 along , and. 13 +
fIGURB 2.19
with the positive direction of the x-axis are respectively
cos 8i + sin 6} and cos cj,i + sm ♦J.
The angle 8 - +between the vectors OQ
-+ and OP
- 1s given by
-+ ~
COS (8 - ♦) = ~
OP• OQ
- = COS 8 COS cj, + SID. OSin. ♦ ,
IOPI I OQI
Example 2.23 Prove, by using vector algebra, that the three altitudes of
a triangle are concurrent.
Proof : Let AD and BE be the two altitudes of a triangle ABC, meoting
at the point O (see Figure 2.20). It 1s sufficient to prove that CO i's
perpendicular to AB.
Clearly.
-+ -+
AO• BC - 0,
and
_,. ....
BO, CA= O. (1)
56 / VBCTOllS
A
B D C
FIGURB 2.20
Note that we are working with vectors that do not have the same
initial point. Now
-+ -+ --+ -+ -.
CO • AB = CO • (AC + CB)
= -+
CO
-+ - -+
• AC + CO • CB
-+ --+ -+ -+ -+ -+
- (CB+ BO)• AC+ (CA+ AO)· CB
-+
-
-+
CB · AC + -+
BO
--+
· AC + -CA · CB
--+
+
-+ -+
AO • CB
-+ --+ - _.
= CB• AC+ CA • CB (by (I))
-+ -+ -+--+ -+ -+
= CB • AC - AC • CB (because CA = -AC)
-+-+
= CB • AC - -
CB ·-AC = 0
{by 2.3.S {c)).
We shall now take up the standard applications of vectors to equations
of lines and planes in analytic geometry. First, we note that if P (x1 , Y1)
-+
and Q(x1, Y1 ) are two points in a plane, then the vector PQ is given by
--+ -+ --+
PQ == OQ - OP - (Xa - x1~y1 -Yi) = (X1 - X1)i +(Ya -Y1U •
Similarly, if P(x1, Yu z1) and Q(x1, y1, z1) are two points in space, then
-+
the vector PQ is (x1 - x1 )i + (y1 - y1)J + (z1 - z1)k.
2.4. t The Equation of a Straight Line
Consider a straight line L parallel to a vector ■ and passing through
a point P (Figure 2.21).
FlGURB 2.21
2.4 APPLICATIONS TO GIDMl'l'&Y J 57
·-+
Let Q be a point OD £. The vector r - 0Q (0 ii the oripn) i■ called
the position vector of point Q. !:et • be the po,ition vector ot point l'.
Then
~ -+ -+
r-OQ-OP+PQ
... • + a vector parallel to •
-,+r■, (2)
where, is a real number. This equation is satisfied by all point■ Q on the
line and by no points off the line. (Why?) It is called the vector equa-
tion of a straight line through a pven point P and parallel to a given
vector ■• fbe followina special cues are worth notins-
2.4.2 Line I■ a Pine
Let P be the point (x1, 1 1) and • be tho vector (a, b). Then , = -+
OP
=- (xu 11) and r = OQ - (x, 1). Then the equation of the line takes
the form
(x, y) = {x1, y 1) + l(a, b) = (x1 + ta, Yi + tb) .
This equation can be written as a pair of equations :
x = Xi+ ta, 1 = y1 + tb,
i.e. (x - x1) == ta, (y - y 1) = tb.
These are called the parametric equations of the straight line. Eliminating
t, we get the cartesian equation
a(y - y 1} = b(x - X1) • (3)
If a = 0 or b == 0, vector • becomes a vector along one of the coordi-
nate axes and the line bas the equation x == x1 or y = Yi•
If a ¢ 0 and b #- 0, Equation (3) may be written in the symmetric
form as
(4)
Crue 2 n -= o.
(7)
provided a, b, and c are nonzero. Thus, we get another form of the equation
of a straight line through P(x19 Yi, z1) and parallel to the vector
•=-al+ b} + ck.
If any two of the numbers a, b, c are zero, say a == b = O, then the
vector• == (0, 0, c) becomes a vector along the z-axis. The equation of
the straight line, from Equation (6), is
X - Yi•
Xi, y - (8)
Note that these two together represent the straight line.
2,4 APPLICATIONS TO OIIOIIIITRY / 59
y
•
FIGURE 2.22
the vector • == ai + bJ + ck. These are also called the direction cosines
of a line parallel to • and in the same direction as •· The ordered set
{cos«. cos~. cosy} is called the set of direction cosines of the line re-
presented by vector•· Clearly,
" .i a
I. I I I I -
cos « = (10)
va + b' + c3
1
" .J b
I• I Ii I = va1 .+b"+c1
cos~= (11)
cosy= •·•
I• I Iii
==
va'-+h'+c•
C
(12)
FIGURE 2.23
--.
be a point in the plane. Then the vector QP =
(x - x1)i +(
y - y1)j
+ (z - z1)k lies in tho plane and is perpendicular to the vector ■ = ai
+ bj + ck. Therefore,
which gives
a(x - X1) T b(y - Y1) , e(z - Z1) = 0. (13)
This rs satisfied by all pomts P on the plane and by no pomts oft the
plane. Equallon (13) 1s called the equation of the plane through Q(xu Yi, z1)
and perpendicular to the vector• = (a, b, c).
Equation (13) of the plane can further be simplified to the form
ax + by + cz + d = 0•
where d = -ax1 -by1 -cz1• Thus, the equation of a plane 10 space 1s
a lmear equation in x, y, and z. Conversely, consider a hnear equation
A.x+ By+ Cz+ D = 0, (14)
where A, B, and C are not ,dl zero. Let A. =I: 0. Then we can write
Equation (14) as
D
A(x+ y) + B(y-0) + C(z-0)-= 0,
which 1S the equation of a plane through the point (-D/..4., 0, 0) and per-
pendicular to the vector A.i + BJ +Ck.
Two planes, if not parallel, intersect 1n a straight line. Thus, a straight
line may also bo considered as the intersection of two planes, and hence
2.4 APPUCATIONS 'ro OIONBTI.Y / 61
where 11 +
J - 1.
62 / VECl'OJrS
6. Prove that the perpendicuJar distance from the point P to the line
-+
r - , + t• is f,
+ (OP-,)·"• -
••II
DP/·
7. Find the angles between the Jines
x- a y - b z -c x - a' y - b' z - c'
I = m 11 ' /' = ~ -= n'
8. Find the angles that the vector • = 2i - 2} + k makes with the
coordinate axes.
9. Find the equation of the plane perpendicular to the vector 11 - 2i
+ 3} + 6k and passing through the terminal point of the vector
i + Sj + 3k.
10. Determine a vector perpendicular to the plane containing the vecb>rs
r, = 2i - 6} - 3k and , = 4i +
3J -- k.
11. Find the angle between the planes
(a) 2x - y + 2z = 1 and x - y = 2
(b) 3x + 4y - Sz = 9 and 2x + 6y + 6z = 1.
12. Prove that tile distance of the point P 0(x0, Yo, : 0) f1om the plane
ax + by + ez + d = O is
I ax0 + by0 + ez0 + d I .
va'+ll-+c•
13. Prove that the equation of the line of intersection of the plants
4x + 4y - Sz = 12 and 8x + 12y - 13z =- 32 can be written in
the form
x-1 y-2 z
-r-= 3 =4·
14. Prove that the equation of the sphere of radius r with centre
P(x11 Yu Z1) is (x - x1)1 + (y - ,Y1)z + (z - z1)2 = r 2•
Chapter 3
Vector Spaces
In Chapter 2 we saw that the set of all plane (space) vectors forms a
commutative group relative to addition and, further, relative to scalar
multipJication it satisfies certain properties (Properties 2.2.7). All these
properties of vectors are so fundamental in mathematics that whenever
any system satisfies them we give it a special name, namely, vector space.
The precise definition of a vector space follows.
Example 3.2 Let :1(1) be the set of all real-valued functions defined on
the interval l With pointwise addition and scalar multiplication (cf§ J. 7),
:1(1) becomes a real vector space. The zero of this space is the function
0 given by O(x) =
0 for all " E /.
Example 3.3 If, instead of the real-valued functions in Example 3.2, we
use complex valued functions defined on T and pointwise addition and
scalar multiplication, then we get a complex vector space {using complex
scalars). We denote this complex vector space by ~ c(/).
Example 3.4 Let 9(1) denote the set of all polynomials p with real
coefficients defined on the interval I. Recall (cf Example 1.39) p is a
function whose value at x is
p(x) = ,; + a1x + .. . +11.X" for all " E /,
where «,'s are real numbers and n is a nonnegative integer. Using point-
wise addition and scalar multiplication as for functions, we find that 9(1)
is a real vector space. If we take complex coefficients for the polynomials
and use complex scalars, then we get the complex vector space 9c(l). In
both cai.es the vector O of the space is the zero polynomial given by
O(x) == 0 for all x E I.
NOTATIONS
Problem Set 3 1
1. Let u1= (1, 3, 2, 7), Us - (0, 2, -3, 5), u3 =- (-1, 3, 1, --4), and
"• = (- 2, 16, -1, S) be vectors of v.. Then evaluate
(a) u1 + Ui (b) 2ui + 3u8 - 1u• (c) u1 + 2u8 + 3u8 - u,
(d) 3u1 + (u,-u1) (e) (au 1-bu2 ) + au3 ,
2. Consider the set VS of all ordered n-tuples of complex numbers. By
defining equality, addition, and 'iCalar mult1phcat1on (with complex
scalars) as in Jlf;, prove that V;
is a complex vector space. Is it a real
vector space ?
3. Let R+ be the set of all positive real numbers. Define the operations
of addition and scalar mult1plicat1on as follows :
u + v =- u•"for all u, v E R+
a.u = 1P for all u E R+ and real scalar a..
Prove that R"' 1s a real vector space.
4. Which of the follo\\ing subsets of V, are \ector spaces for coordinate-
wise addition and scalar multiplication ?
The set of all vectors (Xi, x8, x3, .x,) E V, &uch that
(a) x, = 0 (b) x1 == 1 (c) x 2 > 0
(d) xi :>
0 (e) .xf < 0 (f) 2x1 + 3x2 =- 0
(g) X 1 + fx 8 - 3x3 + x, = 1.
S. In any vector space prove that rxu = 0 iff either rx = 0 or u = O.
6. Which of the following sub,ets of f/l are vector spaces ?
The set of all polynomials p such that
(a) degree of p ~ " (b) degree of p = 3
(c) degree of p :> 4 (d) p(l) = 0
(e) p(2) = I (f) p'(l) - 0
(g) p has integral coefficients.
7. Which of the following subsets of'(; (0, 1] are vector spaces?
The set of all functions / E W [0, l] such that
(a) /(1/2) =- 0 (b) /(3/4) 1 =
(c) /'(x) =- xf(x) (d) /(0) - /(1)
(e) /(x) =- 0 at a finite number of points in [O, l]
(f) / has a local maxima at x = 1/2
(a) / bu a local extrema at x - 1/2.
8. Let Y be a real vector apace and X an arbitrary set. Let y.r be the set
of all functiona /: X-+ Y. Prove that y.r ia a real vector apace for
pointwile addition and scalar multiplication. where definitiona are
analogous to tboae for real-valued fu6ctioo1.
3.2 SUBSPAC!S / 69
9. True or false?
(a) In a vector space V, (-1) <-u) - u for all u E V.
(b) In a vector space V, --u - v ::a -v -.u for all u, v e: Y.
(c} In a vector space V, -u - v == -(u ..1.. v) for all u, v E V.
(d) In a vector space V, -(-u) = u for all u E. V.
(e) R x R 1s a vector space.
(f) If the scalars are complex. numbers, then ~(/) 1s a complex vector
space.
(g) In C, the set of complex numbers considered as a complex vector
+ ~•
space, 1f ot • 1 = 0, then « == O == ~-
3.2 SUBSPACES
3.2.l Deftnitlon Let S be a nonempty subi.et ot a vector space V. S 1s
said to be a subspace of V ,r S 1s a vector space under the same
operations of add1t1on .ind i.c.c1l,tr mult1phcauon as m V.
In order to unq_erstand this defimuon as well as the future concepts 10
the subJect, we shall repeatedly consider the concrete cases of V1 and V3 •
Y1 1s the Buchdean plane. Take r1n) ,;tra1ght hne S through the origin
O. Any point P on this straight bne can be considered a& a ,ector OP of
-
~ -,.
V1 in S. The sum of two such vectors OP and OQ, \\here P and Q both
he in S, is agam a vector OR, where R hes m S. S1mllarly, a scalar multi·
pie of any vector in S 1s agam a vector m S. All other axioms of a vector
space are automatically satisfied 1n S. So S 1s a vector space under the
same operations as an Y1 • Thus, S is a subspace of V1 • In other words, C\er1
line through the origin is a subspace of V,.
In the same manner, in V3 we find that any plane S through the origin
1s a subs1'ace of V3 • Also, every straight hne L through the OJt81n 1s a sub-
space of V1 • The followmg example~ further illustrate the concept of a
subspace.
Example 3.7 Let L be the set of all vectors of the form (x, 2x, -3x, x)
in Y,. Then L is a subspace of v,.
Proof: If u == (x, 2x, -3x, x) and v = (y, 2y. -3y, y) are two elements
of L, then clearly
u +, =- (x+ y, 2(x + y), -3(x + y), x + y)
ucl «u == (cxx, 2(ocx), -3(cu), atx)
allo beloq to L. The zero element (0, o. 0, 0) 11 also of this form and
_ heace belonp to L. The negative of u is -u - (-x•.-2x, 3x, -x),
which ia apin of the same form and bcooe belon,p to L. The other law,
of auociatMty and commutativity for additioa, clitvibutivo laws, and the
ftnal uiotn lu • u are all true in L, becauae elem.Inti of£ are elemcnta
ot V,. and in V, a\\ these \aws are true. Thus, Lis a subspace or Y1 , Tn
fact, L is the 'line' through the point (1, 2, -3, 1) and the or1gm
(0, o, o, 0) in r•.
... .
Example J 8 Generalising Example 3. 7, we can say that all scalar multi-
ples of a givea element u0 of a vector space V form a subspace of Y.
Proof: Let the set of all scalar multiples of u0 be denoted by [uo]. If u and
v are two vectors in {UoJ, then u = cxu0 for some ex and v = ~U0 for
some~-
u +, = cxu0 + ~u0 = (cx + ~)u0 , i.e. u + , is a scalar multiple of
Uo- Hence, u + v E [uJ. Again, Au = A(cxu0 ) = ().cx)U0 , i.e. Au is also
a scalar multiple of u0 • Hence, Au E [u.,] for all scalars;. and u E [u0].
0 E [u0], because O = 011o by Theorem 3.1.7. If u E [u0 ], then
-u = (-l)u by Theorem 3.1.7, and so -u E [uJ. The other axioms,
which are only interrelations, are true for all elements of Y and so are
true for all elements of [u0]. Hence, [u0] is a subspace of V.
Example 3. 7 is a special case of Example 3.8. The subspace of V,
considered in Ex.ample 3.7 is just [(1, 2, -3, 1)).
In these examples we note that to prove that S is a subspace of V we
explicitly checked only the following :
(i) The sum of any two vectors in S is again in S, i.e. addition is
closed in S.
(ii) The scalar multiple of any vector in S is again in S, i.e. scalar
multiplication is closed in S.
(iii) The existence of O in Sand the existerice of a negative for each
element in S.
The other axioms were not explicitly checked, because this, as the
following theorem shows, was not necessary. In fact, the theorem says
that even (iiy need not have been checked.
3.2.2 Theorem A nonempty subset S of a vector space V is a subspace of
V iff thefollowi11g conditions are satisfied:
(a) If u, v E S, then u + v E S.
(b) /Ju E Sand ar. a scalar, then cxu E S.
In other words, a subset S of a vector space V is a subspace of V IJf it
is closed under addition and scalar multiplication defined in V.
Proof: Let S be a subspace of V. Then S is a vector space under
the same operations as those of V. Hence, S satisfies (a) and (b). Con-
versely, if (a) and (b) are satisfied, then we have to prove that S satisfies
all the axioms of a vector space. VSl and VS2 are satisfied for S,
because this i1 exactly what (a) and (b) 1ay. We shall now show that
VS3 for S follows from VSl and VS2 for S.
3 .2 SlJBSPACIS / 71
Ou -= 0 for any " E: J' and therefore for any u0 E S. Takin1 any
Uo E: S, we see that, from (b), 0 - Ou0 E S. Similarly, (-l)u == -u
for any u E J' and therefore also for a siven "u E S. Hence, -r,0 e S
for every Uo E S. Thus, 02 and G3 hold in S.
Axioms GI, G4, and (b), (c), and (d) of VS3 are automaticaUy satisfied
for elements of S, since these laws already bold for elements of Y. Note
that this needs no checking, because these axioms are interrelations of the
elements. In the cases of G2 and 03 the checking was necessary, because
we had to show the existence of certam elements in S. I
3.2.3 Remark In view of Theorem 3.2.2, whenever we want to prove
that a set S for certain operations 1s a vector space, we try to look
at S as a subset of Y, where Y is a vector space and the operations of
addition and scalar multiplication in S are the same as the opera-
tions in V. Once we have done this, it is enough to prove that the
operation& are closed in S.
Example 3.9 Take <'6...>[a, b] of Example 3.5. <lJ [a, b] is a subset of
:J'[a, bJ, which is a vector space by Example 3.2. So to prove that
W[a, b] is a vector space (which we have done in Example 3.5), it is
enough to prove that 6:7 [a, b] is a subspace of :J'[a, b]. Therefore, by
Theorem 3.2.2, we only need to prove that 'G[a, b] 1s closed for addition
and scalar multiphcation.
Since the sum of two continuous functions is continuous and any
scalar multiple of a continuous function is again continuous, we find that
addition and scalar multiplication are closed in Cf! [a, b]. This observa-
tion not only provt:s that W [a, b] is a vector space, but also that it is a
subspace of :1 [a, b].
3.2.4 Remark The spaces y,/[a, b], ~< 1 >[a, b], '2:'(n>[a, b], and B>[a, b]
are subspaces of :f[a, bJ.
Further, note that
(a) tl[a, b) is a subspace of W[a, b].
(b) <Ct1>[a, b] is a subspace of 1G'La, b].
(c) W<n>[a, b] is a subspace of g:::'[a, b] for every positive integer n.
(d) W<n>[a, b] is a subspace of 'ii;-'<m>[a, b] for every m < n.
(e) 9[a, b) is a subspace of W <n>[a, b) for every positive integer n.
(f) Similar results are true for functions defined on (a, b).
Example 3.10 The set S of all polynomials p E 9>, which vanishes at a
fixed point .¥0, is a subspace of 9>.
We have S - {p E 9> I p(Xo) - O} •
If p and q are two members of S, then p(xo) - 0 and q(x.) - o. So
(p + q)(Xo} - ,Cx.,) + 'l(x0) =- 0, which means that the polynomial
p + 'l also vanishes at .¥t and ao p + tJ E S. So addition is cl01ed in S.
12 J VECTOR SPACES
«u + ~v E
S for an u, v E S and all scalars ar:, ~-
2. Let W - {(xi, x., ... , x,.) E Y,. I x 1 == O}. Prove that Wis a s'Ub-
space of Y,..
3. Prove that
w= {(Xi, X2, .. ' Xn) E v~ I Gt1X1 + «1XJ + ... + lly,,Xn = 0,
«,•s are given constants}
is a subspace of~-
4. Which of the following sets are subspaces of Y3 ?
(a) {(x1, x 1 , x 3 )I x 1x 1 = OJ
(b) {(Xi. X2, X3) I Xz = v2}
-'"1
(c) {(X1, Xi, Xs) I v2X1 y'3X2} =
(d) {fa1, x2, X3} I x 3 1s an integer}
(e) {(Xi, x., x,f) I xi + Xi + X1 <: n
(f) {(xi, x 2 , x3) l X 1 + X 2 + xa '.> O}
(g) {(xi, x 1 , x 3) I X1 =- y2l'1 and xt = 3x2}
{h) {(xi, X 1 , X 3) I X1 = 2x: or X3 = 3x2}
But these equations cannot hold"at the same time. So (3., 7) ~ ((I, 2).
(2, 4)J.
Example 3.13 In the complex vector space Vf show that (1 t, 1 - I) +
belonp to ((1 + I, 1), (l, 1 - i)].
[(1 + i, 1), (1, 1 - I)] -= {«(I + ;, ]) + ~(l, I - i) I rx, ~ complex scalars}
== {(« +
~ + a.i, « + ~ - ~i) I «,~complex scalars}.
(1 + f, J - I) belongs to [(I + t, 1), (1, 1 - I)] if
(1 + i, 1 - I) = (at + ~ + «f, IX + ~ - ~i)9
for aome scalars «, ~- Thus,
1 + i ... « , ~ + «i = at(l + t) + ~
1 - i = Gt ~ + -
~i = a + J(l - i) .
Solving for a, ~. we get cc = 1 + i, ~ = 1 - i, show mg that (1 + i,
1 - i) belongs to [(1 + #, 1), (1, 1 - I)].
1. Let S = {(l, 2, 1), (1, 1, -1), (4, 5, -2)}. Determine which of the
following vecton are in [SJ :
(a) (O, 0, 0) (b) (1, 1, 0) (c) (2, -1, -8)
(d) (-i, -¼, !) (e) (1, 0, 1) (f) {1, -3, S)
2. Let S ={x', x2 + 2x, x' + 2, 1 - x}. Determine which of the follow-
ing polynomials are in [S'J :
(a) 2x' + 3x' + 3l. , 7 (b) l.' t 7x + 2
X
FIGUllB 3.1
Example 3.11 In Ya Jet .4 -= {c(l, 2, 0) l « a scalar}, B - {P(O, 1, 2) I fl
a scalar}. Then
.4 + B -= {«(1, 2, 0) +11(0, 1, 2) I cx, Pscalars}
- {(«, 2«, +~, 2fl) I «. fl scalars} •
Geometrically• .« avd B ate lines throu&b the oriain in Ya and.« B +
is a plane containing these lillOI and puaina through the origin (Fipre,3.2).
3,4 MORI! ABOUT SUBSPACU / 79
In this case ..4 ..J.. B = {(I, 2, 0), (0, 1, 2)).
FIGURI! 3.2
DIRECT SUM
We have just seen that if U and W are subspaces of a vector space Y,
then the sum U + W is also a subspace of Y. If, in addition, U fl W
= Y0 = {O}, the sum U + W is called a direct sum. The direct sum of
U and W is written as U EB W.
Example !J.19 We can check the following additions in Y8 :
xy-plane + yz-plane = Ya (6)
xy-plane+ z-axis = V3 • (7)
The sum in Equation (7) is a direct sum, because (xy-plane) n z-axis
= {O}. So Equation (7) can be rewritten as
(xy-plane) EB (z-axis) = Ya • (8)
The sum in Equation (6) is not a direct sum, because
(xy-plane) n (yz-plaoe) =- (y-axis) =#: {0} •
Any vector (a, b, c) E V8 can be written as
{a, b, c) = (a, b, 0) + (0, 0, c) , (9)
where (a, b, c) E xy-plane, (0, 0, c) E z-axis. Thus, (a. b, c) is the
sum of two vectors, one in the xy-plane and the other in the z-axis. The
advantage of the direct sum lies in the fact that the representation (9) is
unique. That is, we cannot find two other vectors, one in the xy-plane and
the other in z-axis, whose sum is (a, b, c). The reader can check this for
himself.
On the other hand, in Equation (6) any vector {a, b, c) can be
written as the sum of two vectors, one in the xy-plane and the other in
the yz-plaoe, in more than one way, for example,
(a, b, c) = (a, b, 0) + (0, 0, c) (10)
and
(a, b, c) - (a, 0, 0) + (0, b, c) • (11)
The fact asserted in this example is generalised in the following
theorem.
3.4.4 neorem Let U and W be two subspace, of a vector space Y and
Z - U + W. Then Z = U Ea W iJf the following condition Is
satilfied:
.4ny vector z E Z can be expressed uniquely a the sum
z == u + w, u E U, w E W. (12)
Proof: Let Z - U EB W. Since Z = U + W, any vector z E Z
can be written aa
z=u+w, (13)
for some II e U and w E W.
Suppose it is possible to have another representation
z-u•+w•, (14)
S.4 MOU ABOUT SUBSPACES / 81
y y
/
JC 0 v•ff.0) X 0
Ca) Cb>
FJouu3.3
Aa a second illustration, take the plane y = 0 in Y1 and call it U. Con-
sider the point, - (1, 1, 1) e: Y1 • (1, 1, 1) + U is the set or all pointl of
82 / VEC10R SPAC!S
+
V1 given by (I, 1, )) u, where II E U. Geometrically, it is the plane
.1,1arallel to y = 0 through the point (1, I, 1) (Figure 3.4).
z
Part (a) of Theorem 3.4.6 says that to describe this set we could take
instead of (l, I, I), any other vector from P. Let us take vector (1, O, OJ,
which is also in P. The theorem says that everv vector
(1, 1, I)+ (0, ~. y)
can also be written in the form
(J, O, 0) + (0, ~•. y')
for suitable~• arid y'. Clearly, ~• = I t ~ and y' ~ 1 -t y.
To continue the illustration, both (1, I, 1) and (1, 0, 0) are an P. The
difference (0, 1, 1) is obviously m U. Part (b) of the theorem says that
whenever the difference of two vectors belong11 to U, then they both belong
to the same parallel and conversely.
3.4. 7 Remark Starting from a subi.pace U, we can obtain many parallels
of U. This i11 the same as starring from one straight line throuah
the origin and drawing many parallel straight lines through different
points in the plane. Given a vector v E JI, we get a parallel v + U.
If we take another vector v' E V, the parallel v' + U obtain.:d from
+
,• will be d1fferenl from v U iff v' 1s not in V U. +
Problem Set 3.4
1. Let U 1, U1, '"i Un be n subspaces of a vector space Y. Then prove
n
that U1 n U1 n ... n U,. = n U, is al~o a i.ubspdce of V.
i=I
2. If U and Ware subspaces of a vector space V, prove that
(a) U n
W is a subspace of W
(b) U U W is a subspace of V iff U C W or W C U.
3. If S is a nonempty subset of V, prove that" [SJ is the intersection of all
subspaces of Y containing S.
4. Prove that the set of vectors (x1, x1 , ... , x,.) E Y,. satisfying the follow-
ing m conditions is a subspace of 'Y• :
«11X1 + «11X1 + •.. + «1n-Xn = 0
«11X1 + C11.ztX1
... + ...
+ «1nXn = 0
... ...
«1a1X1 + «m1'\"1 + ... + «111nXn = 0'
«,,'s are fixed scalars.
5. Find the intenection of the given sets U and W and determine
whether it is a subspace.
(a) U ,_ {(xi, x.) E v. >
t x, O}, W ==- {(X1, X1) E t x, <; O} y•
(b) U = {/E W(-2, 2} IJt-1) - O},
W ... {/E ~(-2, 2) 1./U) - O}
84 / VECTOR SPACES
3.5 J DelJaltJon If Utt "2, ... , u,. arc n vectors of a vector space Y, then
the Hnear combination
«1111 + «alls + ... + «1111,. (I)
is called a tri,ia/ /i11ear c~ mbina1ion if all the scalars «i, «,.... , «11
1
Example 3.22 Let (I, 1, 1), (1, I, 0), and (1, 0, 1) be three vectors. As
befort, if
a.(l, 1, 1) + ~(l, 1, 0) + y(l, 0, 1) = (0, 0, 0),
then {« + ~ + 'Y, « + ~. « + y) = (0, 0, 0) .
Therefore, ex + ~ + y = 0, « + ~ = 0, and « + ·r = 0. These give 11 =
0 ='= 'Y·
It is clear that these three examples fall into two categories. Examples
3.21 and 3.22 are of one kind, whereas Example 3.20 is of another. In
Example 3.20, (1, 0, 0) is a linear combination of (2, 0, 0) and (0, 0, 1):
(1, o, 0) == 1(2. o, 0) + 0(0, 0, 1) •
In other words, (1 1 0, 0) dopends on (2, 0, 0) and (0, 0, 1). We say that
these three vecton are linearly dependent. On the other hand, in Examples
3.5 UNl!AR DBPEND£NCI!, IHD.EPBNDINCJI / 81
3.21 and 3.22 there is no such dependence of one vector on the others. We
say that they are Uoearly independent.
We formalise these concepts of linear dependence and linear indepen-
dence in the following definitions.
3.5.2 De8Ditlon A set {ui, Us, ... , u.} of vectors is said to be linearly
dependent (LD) if there exists a nontrivial linear L"ombin1ition of
Ui, Us, ... , u. that equals the zero vector.
Example 3.23 Prove that the vtctors (1, 0, l), (1, l, 0), and (-1, 0, --1)
are J..D.
To prove linear dependence we must try to find scalars et, (i, y such that
«(l, 0, 1) + ~(1, 1, 0) +
y(--1, 0, -1) = 0 -= (0, 0, 0) .
Therefore, IX + ~ - y = 0, ~ = 0, and IX - y = 0. This can happen iJf
« = y. Any nonzero value for IX, say 1, will do. Thus
+
1(1, 0, 1) 0(1, 1, 0) + 1(-1, 0, -1) = 0.
So there exists a nontrivial linear combination of the given vectors, which
equals the zero vector. Hence, the vectors are LD.
Once we find a nontrivial linear combination equal to the zero vector,
linear dependence is proved. Very often it is possible to guess such a
linear combination.
3.5,3 Definition A set {u1 , Ui, ••• , un} of vectors i1 said to be linearly
independent (LI) if no nontrivial linear combination of ui, "~· ... , u,.
equals the zero vector.
'No nontrivial linear combination equals the zero vector' means the
following:
If at all there is a linear combination that equals the zero vector, then
it must be the trivial linear combination. Now recall that the trivial
linear combination i& always zero. So the statement withm quotation
marks means 'The only linear combination that equals the zero vrctor is
the trivial linear combination'. In view of thi1, Definirion 3.S.3 can
be reworded as follows.
3.5.4 Detialtioa (Reworded) A set {ui, Us, ... , 1111} of vectors is said to
be linearly independent (LI) if tho only linear combination of
u1, U., ... , "• that equals the zero vector is the trivial linear
combination.
By convention, the empty set is considered to be LI. Note that linear
dependence and linear independence are opposite concepts. ·
Example 3.24 Prove that the vecton (J, 0, 1), (1, I, 0), and (l, 1, -1)
are LI,
If possible, let
«(l, 0, l) + 11(1, 1, 0) + y(I, 1, -l) .. 0 == (0, 0, 0) . (2)
This gives • + + I) +
y - 0, It y =- 0, and 11 - y - o. Thercfore, IX = 0
88 / VJfC'J'OJl SPACB.9
2
«
+ 213 = 0 and -
« 13
2 + -2 = 0.
Solving these two equations, we get at = 0 = 13. Hence, the set is u
over (--1, I).
We shall now take up the geometrical me,10ing of linear dependence.
J'or this we need the following definitions in a vector space V.
3.5.5 Definition Given' a vector v -/= 0, the set of all scalar multiples of
v i~ called tire line thl'ough v.
Geometrically, in the cases of V1, V1 , and V3 it is nothing but the
Hraight line through the origin and ,.
3.5.6 Definition 1 wo \'ectors v1 and v1 are collinear if one of them lies in
the 'line' through the other.
Clearly, 0 is collinear with any nonzero vector v.
3.5.7 Definition Gh-en two vectors v1 and V1, which are not collinear,
their span, namely, [v1, , 1] 1 is called the plane through r1 and v1•
Geometrically, in the cases of V1 and J/8 it is nothing but the plane
passing through the origin, , 1 and v1 •
3.5.B Definition Three vectors vi, v1 , and v8 are coplanar if one of them
lies in the 'plane' through the other two.
Clearly, 0 is coplanar with every pair of noncollinear wctors.
Example 3.28 The vectors , and 2, of ·a vector space V are collinear,
because 2v lies in the 'line' through v, i.e. 2v is a scalar multiple of v. In
particular, sin x and 2 sin x are collinear in :1(1).
Example 3 29 The functions sin x and cos JC in :1(1) are nC't collinear,
because neither of the two lies in the 'line' through the other, i.e. one is
not a scalar multiple of the other. Their span, namely,
[sin x, cos x] =- {« sin x+ ~ cos JC I «, ~ any scalars}
is the plane through the vectors sin JC and cos JC,
Examp/1 3.30 The functions sin x, cos x, tan x in :1(/) are obviously not
coplanar, because none of them tica in the 'plane' through the other two.
90 / VBCl'OR SPACES
Example 3.31 The functions sin9x, cos:x, cos hare coplanar. cos 2x lies
in the plane through sin1x and cos•x, since cos 2x is a linear combination
of cos1 x and sin1x.
3.S.!> Theorem Let V be any ,ector 1pace. Then
(a) The 1el {11} i8 LO i.ff I' = o.
(b) The set {111, , 1} 1, LD i.ff 111 and'• are collinear, I.e. one of them;, a
scalar multiple of the other.
(c) The set {,1 , '•• v8} is co iff Iii, , 1 , and'• are coplanar, i.e. one of
them is a linear combination of the oth,r two.
Proof: (a) {v} is LD iff there exists a nonzero scalar cx such that
cz,.• = O. This is possible (Problem 5, Problem Set 3.1) ,j/ v = O.
(b) Suppose {v11 v1} is LD. Then there exist scalars«,~ with at least
one of them, say « -1- O, such that «•·1 + ~"• = O. Therefore, v1
= (-~/cx, 11,, which means v1 is a scalar multiple of , 1, i.e. , 1 lies in the Jine
through"•· Hence, 'i, "•
are collinear.
Conversely, if v1 and v1 are collinear, then, by definition, one of them,
say"" hes in the line through v1 • Therefore, v1 is a scalar multiple of"•·
So ,[ = °'"••
i.e. I • v1 - «v l -=- O. Hence, 1·1 and "• are LD.
(c) Suppoi.e {v1 , Va, v3} is LD, Then there exist scalars ex, ~. and y
with at least one of them, say ex al= O, such that
GtV1 + ~"• + Y"■ = 0
or "1 = (-(i,'«) "•+ (-y/«) Va •
This means v1 E [v1 , v3]. Hence, 111 lies in the plane tb1ough "• and v3 • So
v1 , "•• and v3 are coplanar.
Conversely, if , 1, "•• and "• are coplanar, then one of them, say
, 1 E [v2, v:J, i.e. 111 = «1• 1 + ex1v1 for suitable scalars ex1 and ex 8 • Therefore,
1v1 - ex8111 - ex3v8 = O. Hence, , 11 111, and 118 are LD.
As illustration, consider the three vectors (1, I, ]), (1, -1, I), and
(3, -1, 3). They are LD, for
1(1, 1, 1) + 2(1, -1, 1) - 1(3, -1, 3) = 0.
Hence, by Theorem 3.5.9, the plane through (1, 1, 1) and (3, -1, 3) con-
tains the point (1, -1, 1). Let us verify this.
The plane through (1, 1, 1) and (3, - 1, 3) is
((1, J, 1), (3, -1, 3)) == {(« + 3~, « - ~. • + 3~) I•,~ any scalars}.
This set contains (I, -1, 1), for we have only to choose cx and I! such that
ex + 3~ = 1, Gt - ~ = -1, 11 + 315 == 1.
This gives et = -1 /2 and ~ == 1/2.
Before we proceed, we shall record some simple facts about linear
dependence.
3.9.10 fact In a vector space Y any set of vecton containiq the zero
vector is LD.
3.5 LINEAR DEPENDENCE, INDa,BNDINCB / 9J
For, if {•u "•· ... , .,.,) is a set and,, = 01
Lhen
Ov1 + Ov1 + ... + Ov,_1 + 1,,
+ Ovf+1 + ... + Ov.,
is a nontrivial linear combination resulting in the zero vector.
3.5.11 Fact In a vector space JI', if " is a linear combination of
,., 111, ••• , 11., i.e. v E [,1, 111, ••• , vJ, then {v, v10 ... , ,·.} is LD.
For, " E [vi, "l• ... , v11] means
"= «1V1 + Cls"a + •·· + «.,11. ,
i.e. 1v - « 1v1 - ... - «,.11 11 = o.
3.5.12 Fact In a vector space JI', if the set {vi, 111 , ... , 11,.} is LI and
v ~ [v11 v1 , ••• , v,.], then the set {v, 111, 111 , ... , v,.} is LI.
The proof is lefL to the reader.
In the case where the set consists of just two or three vectors, we can
look at these fctcts from the 'geometric' point of view of Theorem 3.5.9.
For the purpose of understanding, imagine that JI' = V8 (though this is not
necessary for the argument). Since O lies in the lme through v, {O, 11} is
always LD by Theorem 3.5.9 (b). Similarly, 0 always hes in the plane
through any two vectors 1111 v1 • Hence, {O, vi, v11} is always LD by Theore10
3.5.9 (c).
The idea contained in Theorem 3.5.9 does not stop with just three
vector&. We <.an in fctct work out a general result on the same lines for
a linearly dependent set of n vectors. Before we do that, let 01 look at the
theorem once again. Actually, in the proof, we have indicated the
following:
(1) {v1 , v1} is LD iff one of the vectors is a scalar multiple of the
other and (ii) {v1, v11 , v1 } is LD i.ff one of them is a linear combination of
the others.
In general, we shall now prove in Theorem 3.5.14 that if {v11 v1 , ... , 11,.}
is LD, then one of the vectors is a linear combination of the others, the
converse of which is Fact 3.S.11. Before that, we state the followina&
simple theorem and leave its proof to the reader.
3.5.13 Tbeorea (a) // a set 11 LI, then any subset of it is also LI, and
(b) If a set is LD, then any 1uperset of it i• also LD.
3.5.14 Theorem In a ,ector space V 1uppo1e {, 1, '•• ... , 11,.} 11 an ordered set
of vectors with , 1 :/=, 0. The set is LD I.ff one of the 'Vettori
1111 , 8, ... , ,,., ,ay 111:, belongs to the span 0/111, ... , '►1o i.e. 111c E (1111
, 1, .. , Vt_ 1] for some k = 2, 3, ... , n.
Proof: Spppose v., E (,1, v., ... , "•-1). Then 111: is a linear combi-
nation of 111• 111, ... , 111-•• Thus, the set {v1, ,., ... , v,r-1o v.} is LD, Hence,
by Theorem 3.S. 13 (b). {v1, v1 , ... , v,.} is LD
92 / VIICT<m SPACES
Conversely, suppose that {,1, ,,, ... , , 11} is LD. Consider the sets
Si= {vi}
S1 = (v.a, , 1 }
with at least one of the CA.·s not zero. Surely «1. ::fo 0, for if «,., = 0, then
S,r,-1 would become a linearly dependent set contradicting our statement
that S,., is the first linearly dependent set. Therefore, we can write
V1: ~) V1 + (-..!!,_)v1
= (- 11,t «1:
+ ,,. + (- Clt-1) l'.t-1 •
«.
which means v11 E [111, v2, .... v,1: _J.
3.5,15 Corollary .4.,finite Jubset S = {vi, v1 , ... , vn} of a vector space JI
containing a nonzero vector has a linearly Independent Jubset A such
that [A] = [SJ.
Proof: We assume v1 ::fo O. If S is u, then there is nothing to
prove, as we have A = S. If not, then, by Theorem 3.5.14, we have a
vector 11: such that v,, E [111, •••• V,1:_ 1]. Discard ,,,. The remaining
set S1 = {v1, ... , Vt-1, V1:+1, .... vn} has the same span as that of S (see
Problem 4 (c), Problem Set 3.3). If S 1 i11 u. then we are done. If not, then
repeat the foregoing process. This must finally lead to a linearly indepen•
dent subset A such that [A]= [SJ. (Why?)
Example 3.32 Show that the ordered set {(l, 1, 0), (0, 1, 1), (1, 0, -1),
(1, 1, 1)} is LD and locate one of the vectors that belongs to the span of
the previous ones.
Consider the sets
S1 =
{(1, 1, 0)}
S1 = {(1, 1, 0), (0, 1, I)}
S1 = {(I, I, 0), (0, 1, 1), (1, 0, -1)}
S• = {(1, 1, 0), (0, I, 1), (I, 0, -1), (1, 1, I)}.
Obviously, S 1 is LI. Sa is also LI, because neither of the two vecton in S1
is a scalar multiple of the other. S 1 is LD, because
1(0, 1, 1) - 1(1, 0, -1) - (0, 0).
1(1, 1, 0), - o.
Hence. (1, o, -1) E [(I, I, 0), (0, I, I)), s.
is LD, because of Theorem
3.S.13 (b).
J,5 UNBAR DIIPIINDBNCE, INDIPIINDINCI / 93
This means 11 + r = O, 11 + ~ + y = 0, ~ +
y = 0. Solving these equa-
tions, we get only one solution, i.e. 11 = 0 = ~ = y. Hence, the set is LI.
Therefore, A = {( 1, 1, 0), (0, I, 1), (I, J, 1)} is the largest linearly inde-
pendent subset and [A) = [SJ.
Finally, let us extend the concept of linear dependence and linear
independence to intmite sets.
3.5.16 Deflaitloa An i,ifinite subset S of a vector space Y is said to be
linearly inden_endent (LI) if every finite subset of S is LI.
Sis said to be linearly dependent (LD, if it is not LI.
Example 3.34 The subset
S = {I, x, x•, x•, ... }
of fl' is LI.
For, suppose a 1x"1 + o,xk• + ...
+ anit'" = 0 with k 1, k1 , ••• , k. being
distinct nonnegative integers. Note that the equality is an algebraic
identity, since the right-hand side is the zero polynomial. So, either
by giving various values to x or by repeated differentiation of both
sides of the identity, we get
a1 = 0 = a1 = ... = a,,. •
(d) S = {() /2, 2/3, 3/4, 4), (0, 01 2, 0) 1 (1. 2 1 O, l), (1/2, 2/3, 3/41 4)}
(el S = {(1 1 21 3, 0), (-1. 7, 3, 3), (I, -1, I, -1)}.
3. Which of lhe following subsets S of fl are LI ?
(a) S = {r - 1, X + 1, X ·- I}
(b) S={l,x+r,x-x',3x}
(c) S = {x, x8 - x, x' + x'l, x + x• + x' + 1/2}
(d) S = {x1 - 4, X + 2, X - 2, .r/3}.
4. Which of the following subsets S of ('7'(0, ooJ are 1.1?
(a) S = {x, sin x, cos x}
(b) S = {sin1 x, cos 2x, I}
(c) S = {sin x, cos x, sin (x + I)}
(d) S = {In x, In r, In r}
(e) s = {re2 , xeZ, (x' + .'"C - l)eZ}.
S. Prove that the vectors (a, b), (c. d) are LD iff ad= be.
6. Prove that
(a) 1 and I arc LD in the set C of complex numbers considered as
a complex vector space
(b) 1 and i are LI in the set C of complex numbers considered as a
real vector space.
7. Show that the set S = {sin x, sin 2x, ... , sin nx} is a linearly indepen·
<lent subset of 'i; [-n:, n:] for every positive integer n.
8. Prove Theorem 3.5.13.
9. In the proof of Corollary 3.5.15 answer the question 'why'.
10. If u, v, and ware three linearly independent vectors of a vector space
V, then prove that u + v. v + w, and w +
u are also LI.
1I. Find a linearly independent subset .4 of S in Problem I such that [.4)
= [SJ.
12. Proceed as in Problem 11 for the sets S in Problem 2.
13. Proceed as in Problem 1 I for the sets S in Problem 3.
Jt Proceed as in Problem 11 for the sets Sin Problem 4.
15. In Problems 1, 2, 3, and 4 1 whenever a set is LD, locate one of the
YC'Ctors that is in tbe span of the others.
J6. True or false ?
(a) {sin x, cos x, 1} is LD in 7,'(0, 1).
(b) {i +J. i-J. 2i + 3j} is LD in v•.
(c)· Jf u1• u1, u8 are LD, then
«1U1 + a,u1 + «1111 = 0 ~ one of the «·s is not zero.
Cd) Every set of three ~ectors in J/1 is LD.
(e) A set of two vectors in J/1 is always LI.
(f) If a vector 11 is a linear combination of u10 u., ...• u11• then
{uh "■• ... , u.} is LI.
3.6 DIMENSION AND BASIS / 95
which equals the zero vector. But 0, on the other hand, is already equal
(3)
(10)
Linear Transformations
The significance of vector spaces arises from the fc1ct that we can pass
from one vector space to another by means of functions that possess
a certain property called linearity. Thtse functions are called linear
transformations.
r•····
0 I
I
I
! •
--
,
"I
>- cx1,-x,>
FIGURE 41
Example 4.7 Define D: ~ CU(a, b) - > t (a, b) by the rule
D(f) =I'•
where/' is the derivative off. The facts that D(f + g) = (( + g)' = f' + ~•
and that D(a.f) =- (a./)' = r,f' are elementary but important rei.ults in
calculus. In view of these results, we see that D is linear. This trans-
formation D is called the Jijferelllial operator.
Example 4 8 Define J : r (o, b) - R by the rule
b
c'l(f) = J/(x) dx.
a
Again, by the properties of integral, we find that J is a linear map.
Example 4.9 Suppose Ne -f O is a fixed vector of a \'ector space U.
Define T : U -➔ U by the rule
T(x) = x + u0
for all x E U. This map Tis not linear, because
T(x + y) = (x + y) + u0
=I= (x + 110) + (y + u0) = T(x) + T(y) •
This map is called traMlatlon by the vector u0• •
plane is a straiaht Hile. But it is not a linear map from the vector
space V1 to itself, in the sense of Definition 4.1.l.
4.1.4 neorem Let T : U -> Y be a linear map. Then
(a) T(Ou) = Oy, (bl T(-u) =
-T(u), and
(c) 1T«1U1 + ... + «11u,.) = «17'(.u1) + ... + «.T(uJ.
In other words, a linear map T transforms the zero of U into the zero
of Y and the ne,ative of every u E U into the negative of T(u) in V.
The proofs of (a) and (b) are left to the reader. The proof of (c) is
obtained by induction on 'n', starting from the fact that T(ixu) == «T(u) and
using the property
T(«u 1 + ~u2 ) --= T(«u1) +
T(~u1 ) = ixT(u1) + ~T(u1) •
In viev. of (c), we get a standard technique of defining a linear trans-
for.nation Ton a finite-dimensional vector space. Suprose B == {u11 u1 , ... ,
u,.} is a basis for U Then any vector u e: U can be expres!led uniquely in
the form
u = cz 1 u1 + «.uJ + ... -1- ix,.u,. .
So, if T : U - V is a linear map, then
T(u) = T(«1 u1 + «1 u11 + ... + «.u,.)
= «1 T(111) + «1 T(u1 ) + ... + a:,. T(u,.) •
Thus, T(u) is known as soon as T(u1); T(112), ... , T(u,.) are known. This
is formalised in the following theorem.
4.1.5 Theorem A linear transformation T is compe/tely determined by its
values on the elements of a basis. Precisely, if B = {ui, u11 , ... , Un} is
"n
a basis for U and v11 v8 , .. , be n vectors (not n;,cessarily distinct) in
Y, then there exists a unique linear transformation
T: U-~ V
such that
T(u,) = v, for i = I, 2, .. , n. (3)
Proof: let u e: U. Then u can be expressed uniquely in the form
= ac 1u1 + «1u1 + ... + a:,.u,. .
u
We define
T(u) = x1v1 + at1 v8 + ... + at,.11,. • (4)
We now claim that this tran!lformation T is the required transformation.
To prove our claim, we have to show that (i) T is linear, (ii) T satisfies
Equation (3), and (iii) Tis unique.
(ii) is obvious, since u, = Ou1 + ... + Ou,- 1 + Ju, + ... + Ou., and so
T(u,) = h, = v, for all i.
(iii) follows, because if there were another such linear map S with
S(uJ ==,,,then
S(u) = S(at1u1 + at1U1 + ...
+ at,.u.)
= «1S(u1) + izsS(ua) + ... + «.S(u,.) (S is linear)
- 111'1 + Gt~"• + ... + «-'•
- T(u).
4,1 DEPINITION AND.EXAMPLES/ 111
This is true fo, every u E. U So S = T.
It only remains to prove (IJ, whicb is just a verification of the two
relatfons
T(u + v) = T(u) + T(v) and T(atu) == cxT(u)
for arbitrary u, v E U and all scalars ex. Let u, v be two vectors of U.
Then
U =- IJ.1U1 + ... + «,.U,u V = ~1U1 -t ... -1- ~,.u,.
and we have
U +V = («1 -j- ~1)U1 + ,,, -t- («11 -J- ~,.)U11 •
Hence, by the definition of T, we have
T(u + v) = («1 + ~1)V1 + ... + (1J.11 + ~11)1'11 •
Also, T(u) -1- T(v) = (« 1v1 + ... + o:nv,.) + (~1l'1 + ... + ~-,l'11)
== («1 -j- ~1)111 + • • I- («,. + ~ ..),•11 •
Therefore, T(u + 11) = T(u) + T(v). Again,
T(«u) = «« 111 1 + .. . 1· «1,.11,.
= «(«11'1 + •· -1-- 11 P,.) 11
= «T(u).
Thus, T is linear aad the theorem is proved. I
4.1.6 Remark Theorem 4.1.5 will be used in the following way : To de-
fine a linear map, very often we shall be content with listing
T(u1), Ttu1 ), •• , T(u,). This m1;ans the value of Ton a general u is
to be obtained by the prr ce&s shown in the theorem, namely, if
then
T(u) = 1J.1T(u1) + tJ.2T(u9) t- ... + tJ.aT(u,.) •
This proce-,s of defining Ton the basis and extending it to the re-
maining elements is called linear t."(tension </ T. So we shall simply
say •Define T on a basis and extend it linearly' or, equivalently,
'Define the linear map T by specifying its values on a basis'.
Example 4.10 Suppose we want to define a linear map T : V1 - Y,.
Take a basis for Y1 , say {(I, 1), (1, -]}}. We have only to fix T(l, l}
and T( I, --1) in v,. In fact, every ordered pair of vectors in Y, will give
us one linear map T. We shall cite_ a few in Table 4.1.
TABLE 4.1
------- - - - - --- ---- ------
Linear map-
- Ti Ta Ta T,
Value at
...
(0, o. o. 0) (0, 1, o. 0) (I, I, l, 1)
(l, 1)
. (1 1 l, 0, O)
=O.
Therefore, T1 is the zero map.
X - y
T1(x,y> = X----~2 }'
T2(1, I)+ -- 2-- T1(1, -1)
x+y x-y
= -2- (0, ], 0, O) + --T (I, 0, 0, 0)
x-y x+y
=-=<--2. -2--,0,0).
't' -I- )' X -- )'
T8 (x, y)-=- - 2 - T3 (1, lJ -t - 2 - T3 (1, -1)
= x+J'
--f - (I,
X-}'
I, 1, I) +(-2-) (-1, -1, -1, -1)
=- (y, y, J', y} .
T,(x, y) = -X -f-y
2 -Til,
X - y
I) -t - 2 - T.(1, --1)
= --x+y
2 (I, 1, 0, 0) + x-y
- 2 (0, 0, 0. O)
xi..y x+v
- ( 2 • -2-• 0, O) .
D(/) ~= /'. In thi,; case R(D} -.:: '( (a,/,), ,;ince every continuous function
K on (a, b) pos!tesses an antideri,ative and hence D is an onto map.
N(D) is the set of all cunstant functions m 'f, lll(a, h).
Example 4.17 In Example 4.8 /J : f, (n' b) - , R ii defined by J(/)
= J'a' f(x)dx. Here the range is the whole of R, since every real number can
be obtained as the algebraic area under some curve )' =- f(x) from a to b.
Therefore, it is an onto map. The kernel is the set of all those functic.ns
'/' for which the area under the curve}' = f(x) from a to b is zero. Jt is
difficult to say anything more than this about the kernel.
We shall now check whether the linear tram,formations discussed in
Examples 4.11 to 4.17 are one-one lcf I .4. 7 for the definition of one-one).
In Example 4.1 l N(n is the x3 ~axis. So all points on the Xa·axis go
into (I\ 0, 0). So this map is not one-one.
In Example 4.12 N<n =[(I, I, -1)]. So, many poiats go into
(0, 0, 0). This again means r is not one-one.
[n Example 4.13 N(T) = U. So the zero map is not one-one. kcause
all elements go into the zero of V.
In -Example 4.14 the identity map is one-one, because, if x =I= y. then
certaialy /(x) ef:. l(y). It may be noted that Nm i"' the zero subspace of
u,
116 / UNbAlt 1KANSFOlUIATlONS
u1, "•• ... , u. are vectors of U sue/, that T(u1) = v1 , T(u1) = "•• ... ,
T(u.) = "•• then U.1, u., ... , 11. are LI. •
Proof: (a) Let T be one-one and llu 111, ••• , u,. be linearly indepen-
dent vectors in U. To prove that T(uJ, T(u1), ••• , T(u11) are LI, we
assume
«1T(u1) + «1 T(u1) + ... + «.T(u..) = 0
or T(«1u1 + «1u1 + ... + «.u,.) = 0,
since Tis linear. So «1111 + «.au. + ... + «..11. = 0, since T is one-one.
But u1 , "•• ••• , u.. are u. So «1 = 0 = «1 = ... = «11•
Thus, «1Tlu1) + «1 T\u1) + ... + «..T(u.) = 0 implies that each oi:, is
zero. Hence, T(uJ, T(11.), .•• , T(u.) are LI.
(b) Let u1, 111, ••. ,"•and v1 , ••• ... , v.. be as stated in Theorem 4.3.1.
To prove that 1110 111, ••• , "• are LI, suppose
«1111 + «1U1 + ... + otnU11 = 0 .
Since T is linear, we have
= +
Ov = T(Ou) T(«1u1 «1111 + ... + «,.u.)
or «1 T(111 ) + oi:1 T(u1) + ... + x.T(r,,.) = Ov
or «1v1 + «1v1 + ... + cz11v,. = Ov .
But ".1• ,·1 , ••• , v11 are LI. Therefore, at1 = 0 = at1 = ... = at,..
Thus, «1111 + r;U. + ... + «.u,. = 0 implies that «1 = 0 = a, = ...
= «,.. Hence, 111, 111, ••• , "" are LI. I
Example 4.18 Prove that the linear map T: Y1 - Y1 defined by T(e1)
= e1 - e1 , T(e1) = 2e1 + e1 , T{e1 ) = e1 + e1 + e1 is neither one-one nor
onto.
Since [eh e1, eJ = V1 , by Theorem 4.2.2 (d),
R(n = [T(e1), T(e1 ), T(e8)] = [e1 - e2 , 2e1 + e3 , e, + e1 + e8]
= (e1 - e1 , 2e2 + e1] ,
because e 1 + e1 + e1 is a linear combination of e1 - e1 and 2e, +
e1 •
Now we see that e1 -- e1 and 2e1 + e1 are LI. So dim R\T) = 2.
Therefore, R(T) is a proper subset of Y1 • Hence, T is not onto.
To prove that T is not one-one, we check N(T). N(7) consists of
those vectors (x1, Xu x1 ) in Y1 for which
T(xi, Xe, x8) = 0
or T(x1e1 + x1e1 + x1e1 ) = 0
or XiT(ei) + x1 T(e1) + ..c1 T(e1 ) = 0,
because Tis linear. Thus,
(X1 + Xa, -Xi + 2x1 + X1, X1 + X1) = (0, 0, 0) ,
i.e. x1 + x1 = O, x1 + x1 = 0, and -x1 + 2x1 + x1 == 0. · Solving these,
We get X1 = Xa == -Xa, nerefore,
N(T) = {(x 1, Xi, -xJ I Xi an arbitrary scalar} = [(I, I, -1)) .
Hence, by Theorem 4.2.2, T it not one-one.
120 / UNBAR TRANSFORMATIONS
In ~pie 4.18 the linearly independent vectors ei, '•• and e, span the
domain space Ya, but lheir images e1 - e1 , 2e1 + ea. and e1 e, + +
e, are
LD and dim R(T) = 2.
Thus, we find that the effect of the linear map T on U = V, is to
shrink V1 to a 2-dimensional subspace R(T) of V = V1 • What happens
to the remaining 'one' dimension ?
Observe, in this case, that the kernel of T is [(I, I, -1)) and so
dim N(T) = I. Thus, it appears that in this example
dim R(T) + dim N(T) = dim V8 •
This is not an accident. A general result of the same kind is true for
all linear maps whose domain space is finite-dimensi.onal. We shall now
record this as a major theorem.
4.3.2 Theorem (Rank-Nullity Theorem) Let T: U - V be a linear map
and U a.finite-dimensional vector space. Then
dim R(T) +
dim N(T) == dim U. (IJ
In other words,
r(T) + n{T) = dim U (2)
or rank + nullity == dimension of the domain space.
Proof: N(T) is a subspace of a finite-dimensional vector space U.
Therefore, N(T) is itself a finite-dimensional vector space. Let dim
=
N(T) n(n n and dim U p (p ~ n). Let B {u1, u. •... , u.} be a
= = =
basis for N(T). Since 11, E N{T), T(11,) = 0 for each i == l, 2, ... , n. B
is u in N(T) and therefore in U. Extend this linearly independent set of
U to a basis for U. Let B1 == {111, 111 , ••• , 11., ll11+u ••• , 1111} be a basis for
u.
Consider the set
A = {T{Un+J), T(11n+t), ••• , T(u11)}.
We shall now prove that A. is a basis fol' R(T). Observe that, if this is
proved. the proof of the theorem is over; for. this means
dim R(T) = p - n = dim U - dim N(T) ,
which is the samo as Equation (1).
It is therefore enough to prove
(i) (A.) - R(T), and
(ii) A. is LI.
To prove (i) we proceed as follows: Since [B1] == U, it follows from
Theorem 4.2.2 (d) that R(T) = [7\uJ, T(u1), ••• , T(1111) 1 T(u..+1), ••• , T(u.)].
But T(u,) = Ofor I - 1, 2, ... , n. Hence,
R(T) = (7'(11,a+1), T(u..+a), ••. , T(1111)] •
To prove (ii), consider
«n+1TCu.+1> + .. . +«.71:u.,) = o • (3)
Using the fact that Tis linear, we 1et
TCc.+-1"-+ + .... + c.,u.) == o,
4.3 aANE AND NULLITY / l 21
which means that «11+11111+1 + ... + «,,u,, E N(T). Therefore. «n+1un+1
+ ... + «,.M,, is a unique linear combination of the basis B for N(T).
Thus,
«n+1Un+1 + •, · + «,,U,, = ~1U1 + ••• + ~nlln ,
i e. ~1111 + ••• + ~nUn - «n+1U +1 - ••• - «,,Ur, = 0 .
11
Example 4.23 Let Ube the set of all infinite sequences {x1, x1, ... , Xn, ... }
of real numbers. Define the operations of addition and scalar multipli-
cation coordinatewise as in Y., i.e. if
X == {Xi, X1, ••. , X,., ... }
and Y == {.Yi, Y1, ... , Yn, ... } ,
then X + ,Y == {x1 +
,Yi, x. + Y1, .. , Xn + .Yn, ••• }
and
The set U with these two operation, becomes a real vector space (check!).
Note that the sequence {0, 0, ...,o, ... } is the zero of U and {-Xi,
-x1 , ... , -x,u ... } is the negative of x.
Let T : U -+ U be defined by
T(x) = T({xi, x~, ... , x., ... })
= {Xa, X3, ... , Xn, ... } ,
It is easy to check that T is linear. Here R(T) = U. For, take the
sequence {.Yi, y 2 , ••• , y,., ... }. Its pre-image by T can be any sequence of
the form {z, ,Yi, .Y2, ... , .Yn, ... }, where z ran be a real number. In parti-
cular, z can be zero. Hence, T 1s onto.
But T is not one-one, because all the &equences {z, .Yi, y 1, ... , y.., ... }
map into {y1 , y 1 , ... , y,., ... }. Further, N(T) is the set of all sequences of
the form {z, 0, 0, ... , 0, ... }.
Thus, though T 1s onto it does not have an inverse, since T is not
one-one.
Example 4. 24 Let U be the vector space of Example 4.23. And let
T : U - U be defined by
T({Xi, x1, ... , X,., .. }) = {0, Xi, Xa, ... , Xn, ... } •
Obviously, Tis linear (check!). Now Tis one-one, for N(T) is the set of
all sequences of the form {O, 0, 0, ... , 0, ... }. There is only one such
sequence, namely, the zero element of the space U. So N(T) = {Ou}.
But Tis not onto, because the element {J, 1, 1, ... , 1, ... } has no pre-
r
image in U. So R(T) '=I: U. Hence, does not have an inverse.
In Examples 4.23 and 4.24 we prod11ced situations where just one of the
two conditions, namely, (i) T is one-one, (ii) T is onto, holds and the other
does not hold. Further, note that in these examples the space U involved
is not finite-dimensional. (Why?) When U is finite-dimensional, we may
not be able to produce such an example, because of the following
theorem, which essentially says that if T : U ➔ Y is a linear map and
dim U == dim Y, then the two conditions (i) T is one-one and (ii) T is onto
are implications of each other.
4.4.5 Theorem // U and Y are finl/e-dimen,lonal vector space, of the same
dimension, then a linear map T : U ➔ Y is one-one iff It l1 onto.
4.4 INVDSB OF A LINEAR TRANSPOlllfATION / 125
Linear maps
J.
R e, I eJ e1 - "• + f';i :e. ;- 4e.l
s e1 -- e_ e2 t'1 ,_ el - 7e3
IC
• II
II
S(u)+T(u)
FIGURE 4.2
FIGURB 4.3
4.7.1 neorem Let T., T1 be linear maps from U to V. Let S 11 S1 be
linear maps from V to W. Let P be a linear map from W to Z, wl,ere
z
U, V, W, and are vector spaces over the same field of scalars. Then
(a) S1(T1 + T1) = S1T1 + S1T1 .
(b) (S1 + s.)T1 = S1T1 + s.T1 •
(c) P(S1T1) = (PS1)T1 •
(d) («S1)T1 = «(S1T1) = Si(«T1), where « is a scalar.
(e) lvT1 = T1 and T 1 lv = T1 •
Proof: (a) T1 + T1 : U - V and S1 : V ➔ W are linear. The pro-
ducts S1(T1 + T1) and S1 T1, S1 T1 are defined. So both sides of (a) make
sense. Now, if u is a vector of U, then
(S1(T1 + T,))(u) == S1((T1 +
T1)(u)) (definition of product)
. == S1(T1(u) + T1(u)) (definition of addition
in L(U, J'))
== S1(T1(u)) + S1(7i(u)) (S1 is linear)
-= (S1T1)(u) + (SJTJ(u) (definition of product)
== (S1T1 + S,T,)(11) or
(deftnition additi.on
in L(U, JI)).
Hence,
s.c1i. + r,> - s,Ti + s,r, .
The proofs of tile remaining parts are Jef't to tbe reader. I
Note •~Y the diapam for part (e) of Theorem 4.7.1 :
4 7 COMPOSlTION ClP UNPAR MAPS/ 135
1i. J., Ti
U--+ V-- ► V= l'-- ► V
Iv T1 Ti
U - .. C,-~V=-U-+V.
Let T: U _.,. V be a non~mgular hnear m.ip. 1 e T I\ linear, one-one.
an,;1 onto. Then we know that T-1 : V -• U exists and is lmear (Theorem
4.4.3). Further. TI 1 = /r and T- 1T -- Ju (ct§ 1.7). In fact, th,s
cbaracteru,e1,, nonsingular1ty a1,, shown in the followmg theorem
4,7.2 Theorem A. lmear map T: U - V ,s nonsingu/ur if/ there exl\ts a
linear map S : V ~ U .,uch tl,at TS = Iv and ST = lu,
In such a case S = T-1 anti T = s-1 •
Proof: Let T be nomingular. The existence of S, namely, T-1,
follows from the observation• made immediately before the theorem.
Conversely, let S and T exist with the properties stated, i e. TS= Iv
and ST= Ju.
Let u e: N(T). Then T(u) ~ 0 and consequently S(T(u)) =- O But
S(T(u)) = lu(u) = u. Therefore, u = O. So N('T) 1s the ztro subspace
of U. Hence, T is one-one.
Now let v E V. Then T(S(v}) -= (TS)(v) = /r,(v) = ,, Therefore,
there exists an element u = S(v) in U such that 7l.u) = v. Hence, T11
onto. Thus, T H nonsingular. Therefore, T-1 exists by Fact 44.2.
Further, using Theorem 4.7.1, we get
T-1 = T-1(/v) = T-1(TS) = (T- 1T)S = luS = S .
Similarly, T = S-1. I
Finally, we prove the following interesting theorem.
4.7.31beorem Let T: U-+ V and S: V-+ W be two linear mapr Then
(a) If S and Tare nonringular, then ST Is nonsingular lllld
(ST)-1 = T-1S-l.
(b) If ST Is one-one, then T Is one-OM.
(c) If ST is onto, then S 18 onto.
(d) If ST is nonsingular, then T Is one-one and S is onto.
(e) q U, Y, War, of the same finite dl~nslon and ST Is nonsingular,
then both S and Tare nonsln,ular.
Proof: (a) Since Sis nonsingular, S-1 is defined and SS-1 = Iw
and S-1S - Iv. '
Since Tia nonsingular, T-1 ia defined and TI'-'= Iv and T-1T = Iu.
By virtue of(c) and(e) of Theorem 4.7.1, we have
(STX.T-18-1) - S(7tJ'-IS-l)) .. S((TT-1)S-')
- S(I..s-1) • Ss-1 - 'fr· , /
136 / LINEAR 1RANSFOllMATIONS
Similarly,
(T-1s 1>csn = T-1cs-1csn> = r-1«s-1s,n
= T- 1(1vT) = T- 1T == Ju.
=
Hence, by Theorem 4.7.2, ST is nonsingular and (SD-1 T- 1s-1.
(b) Let u € N(T). Then T(u) = Ov. So S(T(u)) == Ow, i.e. (ST)(u)
== Ow. But ST is one-one. Therefore, u = Ou. i.e. N(T) = {Ou}. Thus,
T is one-one.
(c) Let w E W. Since ST is onto, there exists a vector u E U such
that (ST)(u) = w. Therefore, S(T(u)) = w. Hence, there exists a vector
v = T(u) € Y such that S(v) = w. Thus, S is onto.
The proofs of (d) and (e) are left lo the reader. I
Notations (a) The set of all linear operators on U is denoted by L(U).
(b) If Tis a linear transformation on U, then the composition 1T is also
denoted by TJ. Similarly, TL = T"- 1T for any positive integer k. By
convention, ro = I.
Problem Set 4. 7
1. Let T: Y3 - Ya be defined by T(x1, x 1, x 3) = (x1 -t Xa + x 3 , x 1)
l'nd S: V1 ➔ ¥:a be defined by S(x1, X2) ~ (x11, x 1). Then determine
ST.
2. Let S and T be as in Problem I,§ 4.6. Determine
(a) ST (b) TS (c) st (d) -rs.
3. Let S and T be as in Problem 2, § 4.6. Determine
(a) ST (b) TS (c) STS (d) TST.
4. Let R, S, T be as in Problem I, § 4.4. Determine
(a) ST (b) RT (c) RST (d) R(S + T) (e) r
(0 'PST. Also verify that R(S + T) = RS + RT.
S. Determine two linear transformations S and T on V1 such that ST
== 0 L(f'a) and TS =I= 0£(Va).
6. Let S and T be two linear maps on U such that ST == TS. Then
prove that
(a) (S + 1')1 = S2 + 2(ST) + T8
{b) (S +TY,= S" + nc1sn-1T + ... + nc.T-.
(Hint : Use induction.)
7. Let Y be a 1-dimensional vector space and S, T two linear maps on
Y. Then prove that SI' = TS.
8. Let T be a linear map on a I-dimensional vector space Y. Then
prove that ra
== •T for some fixed scalar •·
9. Find the ranae, kernel, rank, and nullity of the followins linear
4.7 COMPOSlTlON Of Ll'NEU MA.PS / 137
maps, where R, S, Tare as in Problem 1 of Problem Set 4.4 :
(a) RS (b) RT (c) RST.
10. Let T be a linear map on a finite-dimensional vector space V. Then
prove that
(a) R(T) n N(T) = {O} iff T 1x = 0 ~ Tx ::, 0
(b) If r(TJ) = r(T), then R(T) n
N(T) = {O}.
1J. If a linear transformntion Ton V satisfies the condition T'- + I= T,
then prove that T-1 exi&ts.
12. Let T be a linear map on V:1 defined by T{e 1) = e1 , T(e1 ) = e11
T(ea) = e2• Then rrove that TJ =- T 1 •
13. Let T : U ·➔ V and S : V -► W be two linear maps. Then prove that
(a) If Tis onto, then r(ST) :;;: r(SJ
(b) If Sis one-one, then r(ST) = r(T).
J4. A linear transformation T on a vector space V 1& said to be idempo-
tent if T 2 = T. For example, the zero transformation and the
identity tran&formation are idempotent.
(a) Let S and T be two linear maps on V3 defined as
T(x1 , x 2 , xJ) = (0, x2 , Xs)
S(x1, Xz, x 3) - {x1, 0, O) .
Then prove that both S and T are idempotent.
{b) If S1 and s~ are idempotent on a vector space V, then find the
conditions under which S/;1 and S 1 + Si are idempotent.
(c) If S 1s idempotent on a vector space V, then I - S is alro
idempotent on V.
{d) Determine two idempotent transformations Sand Ton a vector
space V such that ST = TS = 0, but neither S =-- 0 nor T -= O.
(e) If T 1s an idempotent transformation on V, then pro\-e that
N(T) = R(I - T) and R(T) = N(I - T) .
I 5. A linear transformation T on a vector space Y is said to be nilpotent
on Y if T" = O for aome integer n > l, and the smallest such integer
•n• is called the degree of n;/potence of T.
(a) Prove that the differential operator D is nilpotent on £1' •· What
is the degree of nilpotence of D 'l
(b) Is D 1 + D nilpotent on £1',? If yes, find its degree or nilpo•
tence.
(c) Let T : Y1 -► V, be defined by
7lx1, X:, X1 , Xe) = (0, X1, X1, Xa) •
Is T nilpotent ?
138 / LINEAR TRANSPORMATIONS
The assertion is proved if we can show that Equation (14) holds only
when c1 = c, = ... = c,. 0. =
Differentiating Equation (14) successively (n - 1) times, which is
justified because Y1(x), y 1(x), .•. , Yn(x) belong to 1( / 1•>(/), we get the
following equations :
C1Y1(x) +
CsYs(X) I- ... + Cn)',.(X)
c1y;(x)
. +
c8y~(x) + ... + c,.y~(x) .. ..
~1Y1 1"-11(x) + C2Y2'"-11 (x) + .. + C,J',. 1"-l>(x) = 0,
for all x E /. Thus, for x 0 E /, we obtain
C1Y1(x0) + CaY1(x0) + ... + c,.y,.(x0) =0,
CJ.Y~((x0) + CzY~(x0 ) + ... + c,.v~(x0 ) = 0,
C1Y1'"-U(xo) + C2Y2'"-ll(xo) + .. + c,.y,.'" ll(xo) = 0,
which 1s equivalent to
C1(Yt(Xo), y'1(Xo), ... , Y1 1"- 11 (x0)) t ciy1(x0), J'~(x0), ... , Y21"-0 (x.,)) +
.., t- c,.(y,.(x0 ), y~(x0 ), ... , y,. 1"- 11 (x0)) = 0.
In other words, c1,·1(x0) ~ c,1v 2(xo) + ... + c,.v,.(x0) = 0. Since Y,(x0),
i - 1, 2, , n are LI in V,., all the scalars Ci, c2 , , c,. are zero. Hence,
Y1t Ya, .. , Yn are LI o-ver J. I
Proof of Theorem 4.9.3 By Lemma 4.9.4, the solution space- of
Equation (13), denoted by K, i,; a subspace of l 1"1(/). We shall now prove
that dim K -= n.
Let x 0 be a fixed point in I. Then, by the existence and uniqueness
theorem, there exists a unique solution Yt satisfying the initial conditions
y;
Y1(Xo) = 1, Y~ (x0 ) = 0, (Xo) = 0, , , , Y1 <n-ll(x0) = O.
Similarly, there exist unique solutions y 2 , y 3 , •• , y,. ,;atisfymg respectively
the initial conditions
y 8(X0 ) = O,y~(x0) = 1, y;(x0 ) = 0, ... ,y21"-11(x0) = O;
y 3(X0) =- 0, y;(x0) = 0, y;(x0) = 1, •·, y 31"-ll(x0) = 0;
Matrices
r«.
form
«1 7
I
L «,, .J
where coordinates of the vector are written in a vertical column read from
top to bottom. It may be noted that so far we have been writing this as
a row vector, i.e. in a horizontal row from left to right. To save space,
column vectors are also written as («1, «2, ... , «,,)r, where the letter T
denotes that the vector is a column vector. (The appropriateness of the
letter Twill become clear later.)
For example, the vectors e1, e1 , and e1 of V1 may be written as
S.t MATRIX ASSOCIATED WllH A LlNEA'll MAP/ 149
r17 ro7
107
l:JLJ··l:J
or as (I, 0, O)T, (0, 1, O)T, and (0, 0, J)T. If 11 = 2e1 - 3e1 + e3 , then
the coordinate vector (2, - 3, I) of II relative to the basi~ {ei, ei, e3} can be
written as
r 2 -1
I -3 or (2, -3, J)T.
L l _J
Before we discuss the general definition of a m~trix, let us take a simple
example : Let B 1 = {ui, u8 , u3} and B 3 -- {111, 111, 113 , 11,} be ordered bases
of Ya and v., respectively. Let T: V3 - v. be a linear map defined by
T(u1 ) =- 111 - 2111 +
113 - 11•
T(u1 ) -= 111 111 t 211,
, T(u3 ) = 2112 +
3118 - 11, •
Then the coordinate vectors (relative to B 1 ) of T(u1), T(u2 ), ar.d T(u.),
written as column vectors, are respectively
r
L-l_J
17
-2
1
r
L
-1
0
17
2_j
, and I 07
2
3
L -1 _J
Everything about the linear map T is contained in these 3 x 4 = 12
numbers, written as above. The arrangement of these 12 numbers in the
form
r I
I -2 -1
L-1
is called the matrix of T relative to B1 and B1 •
We shall now give the general definition of a matrix.
5.1.1 Deflaltlon Let U and V be vector spaces of dimensions n and m,
respectively. Let B1 = {u1, ... , u,.} and B1 -== {111, ••• , 11,,.} be ordered
bases of U and V, respectively. Let T : U - V be •a linear map
defined by T(u,) = «,,111 + «11v1 + ... + ac.,11,,., j = 1, 2, •· , n so
that the coordinate vector of T(u,) written as a column vector is
ISO I MATlUCliS
I «11
«11
L «...,
7
_J
1·
I
follows :
«u «11 ... «u
I:l
Lo _J
and [ - :
7
J7.
5.t MA1RIX ASSOCJATEI> WITH A UNIIAR MAP / 151
L0
r: _:1-
7_j
Exampie 5.2 Let a linear map T : Va ➔ Ya be defined by T(x11 x1, Xa)
= (X1 - "• + Xa, 2x1 + 3xa - ix,, X1 + x 1 - 2x3).
If B1 = {e1, e2 , e8}, the standc1rd basis, and B 1 = {(I, I, 0), (1, 2, 3),
(-1, 0, I)}, then T(e1 ) = (], 2, 1) = 2(1, 1, 0) -l- 0(1, 2, 3) + 1(-1, 0, I).
So the coordinate vector of T(ei) relative to B1 is (2, 0, I). Similarly, the
coordinate vectors of T(e1 ) and T(e3 ) relative to B,. are (6, -j/2, 11/2) and
(0, -1/4, -5/4), respectively. Writing these successively as column
vectors, we get the matrix of T relative to B1 and B~ as
Ir: _: _:7J·
LI ¥ -t
Example 5.3 Let a:1inear map T: V3 ➔ V1 be C:cfined b} T(e1) = 2/1
+
- f., T(e1 ) == / 1 + 2/a, T(e3) == 0fi 01,., where {e1 , e1 , eJ} and {.'1, fa}
are standard bases in V3 and V1 , respectively. Then the matrix of T
relative to these bases is
r 2 1 07
I I•
L-1 2 0_j
Let us find the matrix of the same linear map T relative to some other
bases, say B1 == {(I, I, 0), (I, 0, 1), (0, 1, 1)} and B1 = {(I, 1\, (l, -lJ}.
We have
T(x1, x 1, x 3 ) == x1T(e1) + x 1 T(,1 ) + xJT(e3)
== (2x1 + x,) Ii + (2xa - X1) h
= (2X1 + l'1, 2X1 - X1) ,
Therefore, T(l, 1, 0) = (3, I)= 2(1, 1) + 1(1, -1)
T(l, 0, I) = (2, -1) = i(l, 1) + f(I, -1)
T(O, 1, 1) = (1, 2) == 1(1, 1) - i(l, -1).
Hence, the matrix of T relative to B1 and Ba is
r2 i 17
I I.
Ll f -l_j
Noto that the matrix of T changes when we change the bases. Wo
shall pursue this matter in Chapter 7.
Example 5.4 Let a linear map T: Ela ➔ !J1 be defined by TC«o + «ix +
V + «1 x') == oc1 + («. + «3)x + (oc0 + «1).x1•Let us calculate the
152 / MA11UCl!S
r: ~ :_:1
Hence, the matrix of T relative to B1 and B1 is
Ll O -l 2_J
Example 5.5 let D: 9', - 9'1 be the differential map D(p) = p'. let
us calculate the matrix of D relative to the &tandard bases {I, x, x•, x 3} and
{l, x, x1}:
D( I) --= 0 = 0 , l -t Ox Ox 1 +
D(x) = 1 ~ 1 • 1 + Ox t Ox1
D(x11) = 2x = 0 • l l- 2x + Ox8
D(x3 ) = 3x2 = 0 • I +Ox+ 3x1 •
Hence, the matrix of D relative to the standard bases is
ro 1 o 07
I o o 2 o
LO O O 3_j
Example 5.6 We can easily check that the matrix of the identity map
T: U ➔ U (dim U = n) relative to a basis B(B1 = B,. = B) is the n x n
matrix
0 0 07
1 0 0
(I)
0 0 l_J
nis matrix is called the identity matrix and is denoted by 1,. (or simply
1. if n is understood).
The ij-th entry of this matrix is usualJy denoted by 611 , where 611, called
tho Kronecker delta, is defined by
6,, -_{I,0 , if i = J
jf i #: J • (2)
So 1. is the matrix (8, 1).x.,
Example $.1 Usina tho same procedure as in Example 5.6, we can check
5 1 MATRIX AS!OCIATE~ WITH A UNEAR MAP / I 53
that, whatever the bases used, the zero map O Y, Y"' will have the
matrix
ro 0 07
0 0 0
(3)
I
LO O •• O_J
Each entry 10 this matrix 1s zero. Th11i, matrix 1s called the zero matr,x
or order m x n and 1s denoted by O,,.,..n or simply O, 1r the order 1s under-
stood.
Problem Set 5 1
In Problems I through 7 dete1mme the matrix (T. B" BA) for the given
hnear transfo1 mallon T c1nd the bases B 1 and Bl
1. T: v~
- ► v 1 , T(Y, >) = (x, -y)
(J) B1 -= {e1 , el}, B11 -= {(I, I), (1, -1)}
(b) B 1 - {(I, I), (I, 0)}, B2 = {(2, 3), (4, 5)}
2 T V3 ► v~. T(x, y, z) -=- (x .J J, y-+ z).
(t) B1 =--{(I, I, 1),(1,0,0),(1, 1,0)},~= {ei,e2 }
(b) B1 = {(I, I, t), ( ~I, 2, - l), (2, 3, I)}. BA-= {( '3), (I, l)}.
3. T: V, - V5, T('- 11 X2, x3, x,) = (2.\ 1 + x2, '\1 -x3, x3 + x,, Yi,
, 1 + X 11 , 3.\3 -t- x,).
LO l O J
by e1 = T(e3), e1 = T(e 1 ), e8 = T(e1 ) relative to the standard
bases.
(f) To every linear transformatior, there corresponds a unique
matrix.
:
~fl ~,. ~-, ~, .
.
L~.1
~-- ~-, ~--.
5,2 LINIIAR MAP ASSOCIATIID WITH A MATRIX/ 155
is called an III x n matrix. If m = n, the matrix is called a square
matrix.
Two matrices A. = <«.,).,)(,. and B = (~ 11 ).,x,. are said to be eq11al,
written as A. = B, if «, 1 = ~., for every ; and j.
All the statements in Rei;nark S.1.2 also apply to a general matrix as
defined in Definition S.2.1.
A matrix whose entries are all real (complex) numbers is called a real
( omplex) matrix.
As already mentioned, our discussions m this chapter deal with real
vector sraces, i e. Vn and V"'. Also, the matrices used will be real.
However, all the d1scumons hold for complex matrices if the real vector
spaces V" and V111 are repl.tced by the complex vector spaces Vf and V~.
Let us start with a matrix B = (~11) 11u<n and ask whether we can find a
lmear transformation T: U-+ V, where U and V are suitable vector
spaces with ordered ba~es, say B1 and BJ, so that B = (T: Bi, B8).
The answer to this quei.tion is yes, because the process of obtaining the
matrix from a hne~r transformation (see Definition S.1.1) 1s reversible.
Let us g1> through thl." steps of this reverse process.
Let the rn,ilnx B = (~, 1)mxn be given. Ui,ing lhts matrix B, we shall
define a linear map S : U -.. V, where U and V are vector spaces of
d1menMoni, n and m, respectively.
Let B1 -= {111, "i, .. , un} and B1 =- {v1 , v1 , ... , v.,} be ordered bases for
U and V, rei.pecuvely Then we define a linear map S : U -+ V by pres-
cribing \alue!!i of Son the vectors of B1 as follows:
f (u1) = + ~11Va + ,.. + ~m1Vm ,
f11V1
r 2 -3
I 0
L I
-2
2 -2_J
I -:1-
This matrix will give rise to a Jinear map S : V3 ➔ v,. Suppose
B1 = {ui, Ila, ua} and B1 -= {vi, v1, v3 , v,} are ordered bases for V1 and v,,
respectively. Write
S(u1) = 2v1 + v1 - 2v3 + v,
S(ua) = -3v1 + v3 + 21•,
S(u3 ) = 4v1 - 1·1 - 2v,.
The linear extension of this definition of S on the basis elements is the
required map S. If BJ and 8 1 are the standard bases {e 1, e1 , e3 } and
{/11/1./1,/,} for Va and V,, respectively, then we get a linear transfor-
mation T : V1 ➔ V, given by
T(e1 ) = (2, 1, -2, I), T(e1 ) = (-3, 0, I, 2), T(e3 ) = (4, -1, 0, 2),
i.e. T(Xi, Xz, Xa) = (2X1 - 3x. + 4xa, Xi - Xa, -2X1 + X1,
X1 + 2xa + 2x3) •
Let M.,,.. ,. denote the set of all m x n real matrices. Let U and V be
real vector spaces of dimensions n and m, respectively. Fix ordered bases
B1 for U and Ba for Y. Then the process of determining the matrix. of a
linear map and the linear map corresponding to a matrix shows that the
map
't': L(U, V) ➔ M..,.,,. (1)
defined by -r(T) = (T : Bf, B 2) is one-one and onto.
This result merely says that, if bases B1 and Ba are fixed, then to each
linear map T : U - V there exists a unique m x n matrix and to each
m x n matrix there exists a unique linear map from U to Y. This enables
us to pass from linear transformation to matrht and vice versa.
When U = v. and Y = V.,. and the bases B1 and B1 are the standard
bases in the spaces, then the matrix associated with T: Y. ➔ Y.. is called
its natural matrix.
Note (cf Example 5. 7) that -r(O) = O.xa• Further, note (cf Example
5.6) that, if
't' : L(U) - t.\,., " (2)
then -r(I) = I,., where I is the identity transformation on V, and r. is the
n x n identity matrix.
I. A= r: ::
Ll. 2 O O__I
(a) B1 and Ba are standard bases for V, and V8 , respectively
(b) B1 = {(l, 1, 1, 2~, (1, 1, 0, 0), (0, 0, I, 1), (0, I, 0, O)},
B1 -c.. {(1, 2, 3), (I, -J, I), (2, I, I)}.
r1 o o 1
2. A= 0 1 0 I·
LO O 1 _J
(a) B1 = BA - {e1, e2, e3}
(b) B1 - {(I, l. 1), (I, 0, Ol, (0, l, O)},
H2 =-- {(), 2, 3), (l, 1, 1), (2, 1, I)}
(c) B 1 -- {(l, 2, 3), (), -1, I), (2, I, J)},
B 2 =- lO, I, I), (), 0, 0), (0, 1, O)}.
r 1 1 2-1
3. A:- I I.
I 1 1 0 I
(a) B1 and B2 are the btandard bases for V3 and V2, respectively
(b) B1 :- {(I, I, I), (I, 2, 3) (I, 0, O)},
Bi · {(1, 1), (I, - l)}
(c) B1 -- {(I, 1, 1), (I, 2, 0), (0, - I, O)},
B1 -- {(J, 0), (2, I)}.
4. A= r~ : :7
L- I 3__1
(a) B1 and B2 are standard bases for Va and VJ, respectively
(b) B1 - {(l, I), (-1, l)},
B, -- {(1, J, 1), (1, -1, 1), (0, 0, 1)}
(c) B1 = {(1, 2), ( - 2, 1)},
B2 - {(I,-· l, -1), (I, 2, 3), (- 1,0, 2)}.
I
5• If Lsincos 8 - cos 87 1s the matrix o f a Jmear map T : V.11 ➔ V.1
sin 0_J
8
relative to the btandard bases, then find the matrix of T-1 relath e to
the standard bases. '
6. True or false ?
(a) The matrix of the linear map T : U ➔ V is square if/ dim U =
dim V.
(b) An identity matrix is a square matrix.
(c) A zero matrix is a square matrix.
158 / MATRICES
r1 0 07
{d)
LO
0 0 1
1 O..J
Iis the matrix of the identity transformation
(e) I; :1
LI 1..J
determines a linear transformation from Y, to Y,
(7)
0 ..•
0 ...
...
0
0
..
l 11 / 161
L IX1111 1Xma O •• 0 _j
Hence, Gt,, = O for all ; = I , ... , m and j -== J, ... , n. Thus, B is u.
Again, to pro,e that [BJ =- M. 111 , ' " let us take a general matrix A ==
(Gt 11) m M. 111 , 11 • Then Jt can be eas1ly verified that
+ IX1,.E1t, + Gt~1E11 + ••• + «.nE111 + Gt1111Em1
+ ... t Gt11111E11111 •
Hence, B J& a bas,-.. Hence the theorem. I
The bac;1s {E, 1 I , I, ... , m; j = 1, ... , n} is called the .,tandard basis
for M.111 • ,..
Comhmmg Theorems 5.3.S and S 3.6, \\e have the foUowing corollary.
S.3 7 Corollary 1/ U and V are fin,te-dimensiJnal vector spaces, then
dim L(U, V) dim U x dim V.
3. A.=
(a) Gt
r
I
=
-1
3, ~
1 2 3 47
1 1 I
= S
I·
(b)
B =
Gt
I -= 2,
3 -1
1
~
s ,
= -3.
2 07
3 I·
L 3 1 2 0.J L2 1 C 1.J
(a) II= 2,f3 = -6 (b) ar. == 3, f3 == S
(c) ar. - -7, f3 - 3.
4. A=
r1
j2
1-3
-1
-3 -1
2
17
OJ
I, B
r1
==\ 0 l
L3 1
0
-1
:1
162 / MA TRICES
[
«11 «11] [«11 + «11 o «11 + «s1] .
0
T «11 «:11 = «11 «11 + «11
Prove that T is linear and determine its matrix relative to the
standard bases for t.\1 , 1 and t.\1, a•
10. Repeat Problem 9 for T: M,, 8
T[
«11 «11 «117
- [«i1
«.1 + «.a
defined as
+ «11 + «11
1111
1111 J .
11. Let V be the subspace of ~c•>(-00,00) spanned by the functions
sin x, cos x, sin x cos x, sin1 x, cos' :c. Determine lhe dimension of
V, and prove that the differential operator D• maps Y into itself for
every positive integer n < m.
Determine the matrix of (a) 2D 3, (b) 3D' - D +
4 relative +
to the basis of Y obtained from the aiven 1pannin1 set or Y.
5.4 MA1RIX MULTIPLICATION/ 163
12. True or false 'l
12 3 17 14 67 16 9 17
(al I I+ I 1=1 I,
L4 5 7_j LS -1 _J L9 4 7.J
r « 3«7 ,1 37
(b) I I = IX~ I I.
L2~ 4~_J L2 4_j
(c) OA = Om.<n for any m X n matrix A
(d) The set of all square matrices 1s a vector space.
(e) Every matrix can be written as the sum of two matrices, one of
which 1s two times the other.
(t') The set of all 3 x 3 matncei. of the form
ro
I
IX
Iy 0
L6 p
1s a group under the operation of add1t1on.
A=
0
1
-1
0 -2
1
L-2
I
,B
J_j
=I
r 1
0 -:1.
-2
L2 O_j
Y81
:::1
'Y111
•
LYn Y,z_
where
y11 = (1st row of .A) • (1st column of B)
= 3 X 1 + l X O + 2 X (-2) == - I
y11 = (1st row of A.) • (2nd column af B)
= 3 X (-1) + 1 X 1 + 2 X 3 = 4
y11 = (2nd row of A.) • (ht column of B)
= 0 X l + (-)) X O + 1 X (-2) = -2,
and so on. Completing the calculations for all the y's, we find that
AB-r=: _:7J·
L 2 -4
Bzo,nple S.12 Let
rt+, _, 27 .
r o l -17
A=-1 landB== 1 2
. L-4 + I 1 - I O...J
LI - I 1 _j
Here ~ is a 2 x 3 matrix and B i• a 3 x 2 matrix. So AB is the 2 x 2
matrix
166 / MATRICES
r«u «i17
I I .,
L«11 G&u.J
where «11 = (1 + i)O - i(l) + 2(1 - i) .,. 2 - 31
at11 = (1 + i)(l - i) - 1(2) + 21 == 2
r= = = = 7 7
l ~. ~.. · · ~-.. ~- I
Li : : i -1
••• ~ltl ••• Yu ...
: .J
1·
Here attention is f 1.,cussed on the f.. ct that the ij-th element of AB is (I-th
row of A} • (j-th column of B). But recall that we can form the inner
product of two vectors only if both have the same number of coordinates.
In other words, m order that matrix multiplicJtion be possible. we
should have the number of entiies in the i·lh row of A equal to the
number of entries in thej-th column of B, i.e. the number of columns of
A should be the same as the number of rows of B. Thus, an m x n
matrix A can be multiplied on the right by a q x p matrix B tJI 11 IJ· =
If n =I=, tJ, then the product AB is not defined.
A 1 x n matrix A is simply a row vector cif the form (•1, «., ... , ••>·
An n x I matrix B ia simply a column vector of the form (P1, Ji., ... , p.)'.
The product AB of these two matrices is a 1 x 1 matrix [«1Pt + «.Ji.
+ .•. + •.JJ, whose only entry is nothins b11t a 1calar, which is the
inner product of vectors ..t and B.
In seneral, matrix multiplication is not commutative. For, if
5.4 MA11llX MULTJIUCAflON / 167
Then
0
1 0
01 = B.A..
0 l_J
Therefore, B is an inverse of .A. and .A. is an inverse of B.
5.4.7 Tlaeorem If an Inverse of a matrl-c .A. exiat.r, then it i.r unique.
Proof: Suppose that B and C are two inverse~ of .A.. Then ,U1 == I
= BA. and A.C == I = C.A.. Thus, we have
C = CI = C(.A.B)
= (C.A.)B (Theorem 5.4.S(a))
== IB =- B.
Hence, B == c.1
168 / MATRICES
AX==
«11X1 + «11X1 + ••• + Gt■11X11
or X - A 1y.
l:
r1
I I,
-1
,.> 0
0
1
O_j L -1.J
2
(b)
LO 2 1
3 1
0
4.J
I 1
L 2_j
r 17
I
r 17 '
2 2
(c) [2 -1 I 3) (d) (3 1· 2)
-1 -1
l I _J
L O_j L
\10 f MATlUCl!S
rt 17 r 2 7
(e) I I I I·
Li -l_J L I ...J
3. For the given matrices A. and B, determine the product AB or BA,
whichever exists.
ro 17
(a) A=
rt
I
L2
3
:J, B = 13 21
I :_:7
LI 5_1
, I 3 07
(b) A= I I, B =
1_2 1 3_j -S 1
5 77
-1 0 •B
L
=
r·I3
O • 2
2 37
4_1
I
2 l_j L3 7 O_I
s7 r1 2 3 4 57
l,B= I I
l_j LO 1 3 0 l_J
r o 2 17
r1 2 3 0 17
3 S 7
(e) A == I O 1 0 1 0
-1 1 0
l,B=
L3 O 1 l S_j
L O 0 -2_j
r1 17 r-1 17
CO A=I l,B=I I-
LI -LJ L-i - l_j
4. Determine A.· 1 for the given matrix A. in each of the following :
ro 17 r1 27 r« ~7
(a) A =I I (b) A =I I (t') A =I I , ~ -:/- 0
LI O_I LO l_l L~ OJ
r1 o -27
r« 07
(d) A == I
L~ y_l
I , aty :/= 0 (e) A ==
~1 0
2 I
0
2..J
I·
S. Compute A1 , .41, and A' for each given square matrix A.
5.4 MAntx MULTIPLICATION f 171
(a)
r1
l 0
L3 -1 0.....1
2 37
2 I (b)
r1
I
L.3 1-J
27
I (c> IL
3 2
-1
1
0
s
17
0
7_j
I·
6. Determine all 2 x 2 matrices that commute with
r 1 27 ro 17 r1 07
(a) I I (b) I I (c) I I
L-1 0.....1 Ll -LJ LO 2_j
r 0 17
(d) I I.
L-1 0.....1
ro 17 r3 17 r1 07
7. If A= I I ,B = I 1.c = I I.
L2 l_j L2 -l_J L2 l_j
r 1 27
D=
r1
I
2
37I ,E = l 1 0 I•solve the rollowing matrix
LO -I 1J I
L-1 l_J
equation41:
(a) 3X- AB= C (b) (BXD + AXD)E =- C 1
(c) EXD = I.
rA1 0 07 r1L1 0 0 7
I
8. .A.=
0 A
- 0
,B =
0 I-ta 0
LO 0 A11_J LO 0 !-t11.....I
Compute (a) AB (b) BA. (c) A.- 1, if it exists.
9. An n x n matrix ..4. is said to be nilpotent if A11 = O for some positive
integer n. , The smallest positive integer n, for which .A.11 0, is =
called the degree o/nl/potence of A.
Check whether the following matrices are nilpotent. In the case
or nilpotent matrices find the degree of nilpotence.
ro 0 0 l 2 -17
(a)
•
~ ~ 0
Ly, Y, Y1 0...J
0 0
:1 ~, r: - LO 0 0
0 I
0 0
2
1
0.....1
172 / MATRICES
5 -27 r 1 -3 _4-7
I
r1
(c) I 1 2
I -1 I (dj -1 3 --41,
L3 6 -3.J L. t -3 -4_1
10. If A. and Bare square matrices of the same order, then p,ove that
(a) A.:1 - /J2 -=-= (A -- B)(A .L B) i/J AB = BA.
(b) A.1 ±- 2.AB t Bl -= (A .±. B)2 ijf AB = BA.
f - IJtu IXu !Xia. 1
11. Denote the matrix A = I oc 21 11 11 «za I
L_«,u 'lt3z 11aa - I
r-An Au7
as A =-=- I I , where A 11 --= the matrix [«iJ, A11 =- [«11 «18],
L_Azi A12 I
r-«u -i r«iii «u7
Au= I I, A 21 =I I . Similarly, define B, Bu, •·· by
L«u J Lot.12 «J3 •. i
replacing «" by (i,,. Then pro"e that
r An+ Bu A12 -1- Bu7
A+B==I I
LA11 -f Bu A11 + B22_I
r-A.uBu + AuBu A B12 + Ai:1B117
11
and AB= I 1-
I_AinBu + AaB21 A11 Bu + A21B22_J
12. If«, ~ are any scalars, then prove that A2 - (« + ~).A + a.~/
= (A. - rxl)(A - ~I). where A is any square matrix of order n and
I= I,..
13. If «, ~ are scalars such that A = a.B + ~I. then prove that
AB= BA.
14. A square matrix A is said to be involutory if A.1 =- /. Prove that the
r1 «7 1 07 r
matrices I I and I I are involutory for all scalan
LO -l_J L« -l_J
GC,
D:a= I
r -i 07
I, Da
,i
= I
07
I,
L 0 j_J L _O -i_J
f 0 17 r o -17
D, I I ,D,.Jt.-- I I'
L-1 0 -
I
I LI O_J
I- 0 -r1 ro ;7
D, = I I, D7 = I I.
[_-i O_J L_i O_j
Write the mult1rhc 1110n tab 1e of the group.
21. True orf albe?
ra b7 ,- -a b7 r-a• +bd ac + bd7
(a) I I I I= I I-
Le d_j L c d_J L - ac + bd c9 + d8 _J
(b) If AB = I, then .A. 1s invertible.
(c) If A and B are square matrices of the same order, then
AB= BA.
(d) lhe system of t\\o equations 2x + Jy == land 4x + 6y, - 2
cannot be solved by the method discussed in this article.
(e) Let A be 2 x 3 and B be 3 x 2 matrices. Further, let C bo
2 ,1(, 2 and D be 3 x 3 matrices. Then
(i) AB is defined but not BA.
(ii) AB and B~ arc both dctlnc4.
J14 / JIATRICBJ
(2)
(l)
r:::
I =
+ «1n.:\n 7
f Gt1"X,. I
I•
L«,.11X1 -I «.,.zXa + •·• 1 cx,,.,.Xn J
r«1/7 r«u -, 1«111 -1
I
I O'n
I e. + x., : + ... -l x,,
The range of A is [(3, 2, l, O), (1, - J, J. I), (2, 0, l, 2)). The rank of A
is 3, since the three vectors listed are u.
Example 5.15 Let
r3 I 4 07
A= 0 2 2 0
l_l - I 0 0 _J
We ha\e R(A) - ((3, 0, I), (I, 1, -1), (4, 2, 0), (0, 0, 0)) = ((3, 0, I),
(1, 2, -1), (4, 2, O)J = ((3, 0, I), (I, 2, - l)J, since (4, 2, O) = (3, O, I)
+ (I, 2, - I). The rank of A is 2, because (3, 0, I) and (1, 2, -1) are LI.
Kernel of A The kernel of A is the set of all vectors x E V,. such that
Ax= 0,.,,<1.
/1.'ullity of A The nullity of A is the dimension of the kernel of A.
Example 5.16 Let
r 2 17
A == I -I 2 I.I
L 3 O_l
This is a linear transformation from V2 to V8 • We have
[ - 2 17 I X1 7 I 2x1 + X2 7 r- 0 7
I -I 2 I I
l_ 3 O_J Lx2
J"-' -xi
L 3x1
+ 2x,
_j
= I
L O
O
_j
I·
Thus, Ax = 03 x 1 implies
2x1 + x 8 = (l, -x1 + 2x1 = 0, 3x1 = O. .
This gives x 1 = 0 = x 2 • Therefore, the kernel of A is {O} and the nullity
of A is 0.
Example 5 .17 Consider the matrix of Example 5.15. We have
rx17 13x1 + Xa + 4xa7
I 4 07
2 2 o I Xa \ =:
LI -1 0 O_l x, I I
Lx,-1L X1 - X1 _J
Thus, Ax.= 0 implies 3x1 + x 1 + 4x8 == 0, 1x1 + 2x3 = .0, x 1 - x 1 == 0.
This gives x1 == x1 == -x8• So the kernel of A is the set of all vectors
of the form (xi, Xu -x1 , x,) = x 1{1, J • -1, 0) + x.(0, 0, 0, 1). It is
therefore the span [(I, 1, -J, O), (0, 0, 0, 1)). So the nullity of A is 2.
In Examples 5.14-5. l 7 rank + nullity = dimension o: V11 • This
theorem, which is true for linear transformations between finite-dimen-
sional vector spaces, is also true for matrfoes, because of the identification
9f matri~ as linear transformations in the above disc:ussion.
.5•.5 llANK. AND NULLllY OP A MATUX / 177
Recall that a linear transformation from v. to v. is one-one lffit is
onto. So an n x n matrix A will be a one-one hnear transformation lff
its range is the entire space, 1.e. iff its rank is n. This means that the
maximum number of linearly independent column vectors 1s n. il other
words, the column vectors of A are LI. Since the foregoing argument is
reversible, we have the following theorem.
5.5.2 neorem A square matri'lc is nonsingular iff its column vector,
are LI,
Example 5.18 Let
l 0
L -1 1
The range of A 1s [(I, I, - I), (0, 2, 1), (I, 3, O)]. For the kernel of A,
consider
A-[: _: ~:]
The ranse of A is [(2, I, 1), (I, 2, -1), (--1, 0, I)]. The rank of A is 3,
because we can easily check that these three vectors are u. Hence, A is
nonsingular. This can also be arrived at by showing that the kernel of A
is {O}.
r1 27
Example 5 20 Prove that A == I I is nonsingular and find its
LO 1.J '
inverse.
The column vectors (I. O)T and (2, J)r arc u. Tberdore, tho matrii
178 / MATRICES
rat f37_
Then
is nonsingular. Assume that the inverse of A. is I
LY BJ
,,., r1 27 rGt f37 r1 07
I I I I= I I,
LO l_l Ly 3_j LO l_j
r ix + 2y ~ + 2s 7 r1 07
i.e. I I= I I-
L y 3_J LO I_!
So at I- 2y = I, f3 + 28 == 0, y == 0, 3 = I, which gives « = I,
f3 = - 2, y =- 0, 8 = I. It is easily seen that
r 1 -- i- 1 r 1 27 r 1 2 7 r 1 -27 r1 07
I I I l=I 11 l=I l-
LO l_J LO l_J LO l_j LO 1_1 LO 1.J
Hence,
rt
A.-1 =I
LO
Example 5 21 Find the inverse of the matrix
I
I 2 3-1
A =1 0 I 2
L-1 1 1.J
The reader can check that the column vecton, of A. arc
.
and so A. 1s LI
.
nonsingular. To find the inverse, we soh,e the matrix equation
r 1 2 37 r«1 «z «a-1 r-1 0 07
(d)
0
) -3
3 ·- 1 2 I
2 4 3
II
I
(e)
2
l
3 -1
5 2
1 l
0 \
I
L2 3 0 3 OJ L_O 0 1 I _I
r-1 1 17
I
(f) 3 1 -1 I
IL 2 2 1J
2. Prove thc1t the following mdtrtce-. are nonsingular and find their
mverses:
( 0 0 ,-1 -1 17
r I 27 I
(a) I
L-1 O.J
I (b) I2
I
I
0
I
(c)
I3 0 1
I
L.3 4 ">
.. ..J
I
LO 1 1
r-1 2 37 2 2 I
(d) I0 2 1
(e)
0 ~ 1
II 3 -I 0 0 l
LI -I O_J LO () 0
3 Fmd the values of oi: and ~ for wtuch the followmg matrix 1s inverti-
ble. Fmd the inverse when 1t ex1st'i
101: ~ 07
0 «_J
0
L~
°' ~ I·
4. Prove the following :
(d) If two rows of a matrix are interrhanged, then the rank doe,
not change
(b) If a row of a matrix 1s mlllt1phed by a nonzero scdlar, then the
rank does not change.
S. True or false ?
(a) The range and kernel of a square matrix are of the same
dimension.
(b) There exists a 7 x 12 matrix whose rank is 10.
(c) If two columns of a matrix are interchanged, then the rank does
not change. ·
(d) If a column of a matrix is multiplied by a ponzero scalar, then
tbe rank does not change,
J 80 / MATRICl:S
I
L«,..1 «,..1 ... ar:,.., a.,.n_l Lo-111 «.,. ... «,,. ... «,..,.J
Clearly, AT is an n x m matrix whose ij-th element 1s ar:,.. the ji-th
element of A. It follows immediately from the definition of transpose that
P. = I,. and O~x. = O,.x,... •
5..6.t Theorem If A and B are two m X II matrkes, then
(a) (..4 + B)T = AT + BT,
(b) (aiA)T = aiAT, and
(c) (.A.T)T = .A..
The proof is straightforward and left to the reader.
5.6.2 Theorem If A is an m x n marrix and B Is an n x p matrix, then
(.AB)T == BTA.T.
Pre of : First, note that AB is defined and is an m X p matrix. So
(.A.B) 1 is a p x n matrix'. Again, Ar is an n x m matrix and BT is a
p x n matrix. Therefore, we have p x m matrices on both sides of the
octuality in the theorem and so the equality is mcaninsful.
S.6 TRANSPOSE OF A MATID: / 111
To prove that the equality ltolds1 it suffices to prove that the ij-th
element or (AB)T is equal to the ij-th element or BTA.T. Let A.= («.,>..x.
and B := (~,,)11Xt1• Then AT = c«:1)11)(111 and BT = (~;,).,x11, where «~
= "" and = ~~, = ~,,. The ij-th element of BTAT i:1
n n
t ~~Glflll :::. .& Gl1t~k1 ,
k=I k ... l
which is the ji-th element of AB and hence the ij-th element of (AB)T.
Thus, (.AB)T =- BT.AT. I
S.6.3 Theorem If A is a nonsingular square matrix, then AT is also no,r-
slngular and (A')-1 == (A.-1)T. ·
Proof: Since A is nonsingular, there exists a matrix B such that
AB =I= BA. Therefore,
(AB)T = JT = (B.A)T .
This gives BTAT = I--= A BT. Therefore, .A has the mverse, namely,
BT, and (.A.1)· 1 = BT ,... (.A-1)T, because B =- ,4-1 . I
5.6.4 Corollary The· columns of a SCJUare matrix A. are LI if/ its rows
are LI.
Proof· Columns of A are LI .A is nonsingular (Theorem S.5.2)
A is nonsingular (Theorem 5.6.3) -e:> columns of AT are LI (Theorem
-e:>
5.5.2) -e:> rows of .A are LI (definition of AT). I
We shaU prove in Theorem 5.7.5 that in any matrix, not necessarily
square, the maximum number of linearly independent columns is the same
as the maximum number of linearly independent rows.
Before we conclude this article let us familiarize ounclves with certain
special types of matrices.
A square matrix of the form
rA1 0 0 ... 0 0 7
0 A.0 ... 00
0 0 A. .:. 0 0
o o o ... o~.
is called a diagonal matrix. In other words, A = (1111) 11x• is a diagonal
matrix if «u = 0 whenever i :;: J, i.e. «u == .,.._Bu,
A square matrix of the form
182 / IIATRICBS
r >. 0 ... 07
0 ·A ... 0
:': : =='Al
LO O ... A_J
is called a scalar matrix. In other words, A = (at")11x11 is a scalar matrix
if 11.11 == >.811 for a fixed scalar 'A.
A square matrix of the form
r«u at11 at11 ... Gt111 7
LO 0 0 ... «1111_J
ro 17
and I I is a skew-Hermitian matrix.
LI O_j
From these t!efinitions of special matrices, we get the following results :
(1) The product of two diagonal matrices of the same order n is again
a diagonal matrix of order n and
r"lo O •. • 0
>-. ..• o
711-'ol O •• • 0
1,1,a ••• o
7I 0
I=
. • . . I
• : : : • I
I
LO O >..,._J LO O 11,11 _J
(ii) Any two diagonal matrices of the same order commute with each
other.
(iii) A diagonal matrix is nonsmgular iff none of tl:e diagonal entries 1s
zero.
(iv) The in~erse of the diagonal matrix
0 ... O 7 r-1r>-1 0 .. 0 7
[ >-a
0
.
LO
0
0
A,._J
A, /:- 0, i = 1, 2, ... , n.
l'i
1/>n_J
0 l/'>-1. ..
..
0
(c)
r2
I •-
+ 3; 2- ; 7
2i 2 - 2i (d)
r- 1
I 2 - i
1- ;
; 2i=/l
L3 + 4i 2-ti_J LI +i 2 +; 3 + 2i_J
1a h g-7
6 .. If .A = Ih b f I and X -=--a (x, y, z), evaluate XAXT.
Lg f c_J
7. If A is an m x n matrix with complex entries such that AA• = 0,
prove that A = A = o.
8. If A and B are square matri~s of the same order and A. is symmetric,
prove that BTAB is also symmetric.
9. (a) Prove that a triangular matrix whose leading diagonal entries are
all zeros is nilpotent.
(b) Determine a matrix that is both upper and lower triangular.
10. Let A be a square matrix. Then prove that
(a) A + A• is Hermitian.
(b) .4 - ...4.• is skew-Hermitian.
11. (a) Prove that every square matria: can be expressed as the sum of
a Hermitian and a skew-Hermitian matrix.
5.7 ELEMENTARY llOW OPERATIONS/ 185
(b) Prove that every square matrix ca11 be expressed as the sum of
a symmetric and a skew-symmetric matrix.
12. P,ove that the set of all n X n diagonal matrices is a subspace of
tA. •• •·
13. Prove that the set of all n x n invertible diagonal matricei. is a group
under multiplication.
14. True or fab,e ?
(a) The set of all n x n diagonal matrices 1s a group under multi-
plicat1on.
(b) If A 1s mvolutory, then its transpose is also involutory. (A is
mvolutory means A 2 = /.)
rO I 27
(c) The matrix - 1 0 3 1s skew-symmetric.
'- -2 -3 l_J
(d) Every !tyhlmelnc matrix is Hermitian.
(e) A nonLero idempotent matrix 1s not nilpotent.
(f) An idempotent matnx A 1& singular unless A = I.
0
(E7): X + 2y - 2z == -1 (BIO)
Ox+ y- ¼z = -t
.
(ES): (Ell)
(El) X (7) + (E9) Ox + Oy - 1/z == -Y (El2)
(EIO): x+ 2y- 2z == -1 (B13)
(Ell) Ox+ y .... h=-t (B14)
(El2) X (- i\) Ox+Oy+ Z= 3 (EJS)
0
(1:\13) : Ix+ 2y- 2z = -1 (EJ6)
(EIS) x U> + (El4): Ox+ ly + Oz= 2 (E17)
(EIS): Ox -f- Oy + lz= 3 (El8)
0
(El7) X (-2) + (El6) Ix+ Oy - 2z = -S (EJ9)
(E17): Ox -t ly + Oz= 2 (E2O)
{EIS) Ox+ Oy + lz = 3 (E21)
0
(Ell) x (2) + (J::19) 'Jx + Oy + Oz= I
(E2O) Ox + ly + Oz = 2 (B)
(E21) Ox + Ov + lz = 3
which gives x = J • y = 2. and z = 3.
It can be seen that throughout we have worked with only the coeffi-
cients in the equations and the numbers on the right-hand side. The pre-
sence of the symbols x, y, and z and the sign of equality does not at all
atrect the working. Now, deletmg the symbols x, y, and z and the sign of
equality from the pattern, which otherwise remains unchanged, we shall
write only the numbers involved in the form of a matrix A. Repeating
the foregoing sequence of steps, we get the following sequence of matrices,
where the symbol - is an analogue of the symbol o, and the letters r1• r1 ,
and r3 stand respectively for row 1, row 2. and row 3. Further. r 1 + kr1
means 'add k times the j-th row vector to the I-th row vector' • We have
r2 -3 1 -17 11 2 -2 -17
matrix A = 1' 3 0 1 6 Iinter;ange
r1 and , 1
3 0 1 6 I
r
LI 2 -2 -l_J L2 -3 1 -l_J
l
,,+£ o
-2 3i,, -6
2
7
-17 [l
91 •• x(-t) o
2-2 I I -i -I
-17
,, + Io
7, 1 1 -i· -t j ,, x C-1") lo 1 -f -t I
LO O -¥ -¥ _J O O 1 3..J
::, 1 ELEMENTARY ROW OPERATIONS/ 187
r1 2 -2 -17 r1 0 -2 57
r8 ,. ., lra I 0
+ 1 0 2 ! ,...,
rt+ (-2)r2 :
Io 1 0
2 I
LO 0 3..J LO 0 1 3..J
rt 0 0 17
0 , 2 0 2 0 2 0 0 0 0
0 0 0 1 -1 0 3 a a o -2
0 0 0 0 0 1 0 0 0 0 2
0 0 0 0 0 0 a 0 0 0
a 0 0 a 0 0 0 1 0 ~
0 0 0 0 0 0 0 0 0 .._
1 __ 2
0 0 0 0 0 0 0 0 0 0 0
188 / MATIUCl!S
we shall prove that for any matrix A. the column rank is equal to the
row rank; so we can afford to call both of them simply 'rank'-. Till
then,' we shall have to distinguish between the row rank defined in
5.7.2 and the rank(= column rank) defined in§ S.S.
The aforesaid process of obtaining matrix B from matr;x A is called
the process of row reduction. In practice \\ e usually aim at getting B in a
standard form called row-reduced echelon form.
Consider the following 7 X 11 matrix and the 'stairs' or 'steps' marked
therein:
Such a matrix is said to be in row-reduced echelon form. We shall
now give its precise definition.
5.7.4 Definition A matrix is said to be in roH"-reduced echelon form if it
satisfies the following criteria :
(a) The fir!i.t nonzero entry in each nonzero row is 1.
(b) If a column contains the first nonzero entry of any row, then
every other entry in that column is zero.
(c) The zero rows (i.e. rows containing only zeros) occur below all
the nonzero rows.
(d) Let there be r nonzero rows. If the first nonzero entry of the
i-th row occurs in column k, (1 =-= 1, 2, ... , ,·), then k 1 < k,.
< ... < k,. .
Draw horizontal and vertical partition lines such that below and left
of these lines there are only zeros and such that at the point where the
vertical line 1s followed by a horizontal line there are 1's, namely, the first
nonzero entries of nonzero rows. These turnmg po111ts are called steps.
See the foregoing 7 x 11 matrix, where there are s•x steps.
We now state the main theorem of this article.
5.7.5 'fteorem (al The ,ow rank and column rank of a matrix A are the
~e,me. In other words, the m:iximum number of /im,arly independent
row vectors is equal to the maximum number of linearly independent
column vectors and is equal to the rank of the matrix.
(b} The rank of a matrix A. la the number of nonzero row, In its row-
reduced echelon form.
5.7 ELEMENTARY BOW OPERATIONS/ 169
We shaU prove this theorem by establishing a 1.uccession of lemmas.
The sequence of proof can be easily understood from p:gure 5.1 :
II I• ..J
.:
Row rank of 8 Column rank of B
• 4-
by Lcll'm!l 5 710
I\~
<.,~~
Q
d'.,.
.> '0~
.~
~C)(,:,
-..;
~umber of Number of
nonzero rows steps
of B ,n B
A. Given matrix
,.,, ,.,
B: Row reduced echelon form of A <Lemma 5. 7 &,
FIGURE 5.1
5.7.6 J.emma Ei·ery matrix A i$ roM•equivalent to a row-~educed echelon
matrix.
In other words, every matrix can be reduced to the row-reduced
echelon form by a finite sequence of elementary row operations.
We shall omit the proof of this lemma, as careful scrutiny of the
process of row reduction will convince the reacer that this lemma is true.
5.7.7 Lemma // a matrix is in tl,e row-reduced echelon form, its row rank
is the 11umber of nonzero rows in it.
The proof is left to the reader.
5.7.8 Lemma 7he row rank of a matrix A is equal to the row rank of the
row-reduced echelon matrix B, obtained from .4.
Proof: B has been obtained from A by a finite sequence of
elementary row operat:ons. We .shall show that these row operations do
not affect the row rank of A. It is clear that the row operations of type
I and type Udo not alter the row rank (see•Problem 4, Problem Set 5.5).
It is therefore enough to prove that an elementary row operation of type
III does not change the max:mum number of linearly independent row
vectors. Supposo we Jdd «-times a row vector •i to rnot)l~r row vector
I 90 /MATRICES
v1 • Let the other row vectors be , 9, ,,, •· , v,... Examine the two sets of
vectors
P = {111 , 112 , ••• , v,,,} and Q = {v., v8+ix1'i, V3 , ... , v,..} •
It can be easily checked that (1) ,f P is u, I hen Q 1s also LI and (ii) if P is
LD, then Q js LD. This shows that a type Ill oper,uion does not affect the
maximum number of UnearJy independent row vectors, and so the row
rank is unaffected by such an operation (sec Problem 9). I
5. 7.9 Lemma // a malrbc is in the row-reduced echelon form, the,, its
.
(column) rank is the number of ·steps' in it .
rroof: Let the number of steps hep. Every column that occurs
hefore the first step is a zero vector, and every row that occurs after the
last step is also a zero vector. So the nonzero column vectors can be
con,;idered vectors in V,. Therefore, the column rank is less tha1i or equal
to p. On the other hand each step contributes one nonzero column
vector. The set of these column vectors is LI. This means the column
rank is greater than or equal to p. Thus, the column rank is p. I
5.7.10 Lemma If a matrix is In the row-reduced t!chelon form, then its
(column) ro,rk is equal to its ro,v rank.
Proof: Note that the number of steps in such a matrix is equal to
the number of nonzero rows. Using lemmas 5.7.7 and 5.7.9, we find that
its (column) rank is the number of step~ m it, i.e. the number of nonzero
rows, i.e. itc; row rank. I
5.7.11 Lemma Tlie (column) rank of a matrix A i, equal to the (column)
.
rank of the row-reduced <che/012 matrix B, obtained from A .
Proof: To facilitate understanding cf this lemma, wc shall prove
it for a matrix lf 01der 3 x 4. By using proper notations, the same
proof can be extenc!ed to any m X n matrix.
We have only to prove that the th1ee types of elementary row
oper.ations do not affect the column I ank. The fact that types I and II
Ol}lralions do not affect the column rank is easy to prove and is left to the
reader. To prove that a type Ill operation does not affect the column
rank, com,ider the matrix
: :: : :~l
L.a.31 11.az «aa «u.J
Let us pe1form a type III operation on P1 namely, let us add ix-times row
3 to row 2. The resulting matrix Q 9,ifl be
r «n
I
IXJS
a= + CI.IX31
ixa1 IXaa + ««11 «.a + ix«aa
1.- «.1
5.7 ELEMENTARY ROW OPERATIONS / 191
Suppose the column 1ank of Pis 3 (any other value of the rank can be
handled analogously). Then in P there exist three linearly independent
column vectors and every four column vectors are LD. We shall prove
that the same phenomenon occurs in Q also.
Let the three linearly independent column vectors of P be the first three
(1f they are any other three, they can be brought to the first three positions
without affecting tl•e argument). Let us call them C1, C1 , and C3 •
If Fi, F2 , and F 3 are the first three column vectors of Q, then the vector
equation ~1 F1 + ~2F2 +~ 3 F3 = 0 gives
~1 1111 + ~21Xu + ~3IXJ3 = 0
81(IX21 + IXIX31) + f32(IX22 +. 0'0'32) + ~3(0'23 ,- IXIX33) = 0
A=
-1
2
3
I
0
3 -2
-41
I
L 1 J I - I.J
by reducing it to row-reduced echelon form
.
S -I
-3
-1
S -2
07
-4
r,
1 1 I -I_J LO -1 2 -I_J
r1 2 -t 2 -1 07
,...,
inter•
changing
, 1 and r,
0 -1
0 -3
: -~
7
j,,x(-1) 1
:
1 -2
-3 5 -2
I
LO S -I -4 LO S -:• -4.J
2 -1 07 r1 2 -I 07
,..., 0 1 -2 1 ,._, 0 -2 I
r1 + Jr1, r, t 9r3
r , - fr1 0 0 -1 1 0 0 r-1 1
LC 0 9 -9.J LO 0 0 0.J
r1 2 -1 07 r1 2 0 -1,
0 -2 0 0 -1
r1 X (-1) 0 0 1 -1 0 0 l -1
LO O O U,.J 0 0 0 O
5,7 ELIIMENTAllY llO\\' OPBllATIONS / 193
1 0 0 17
....., 0 l 0 -1 I
'1 - 2,. 0 0 1 -1 I
0 0 0 O_J
Hence, the rank of A is 3.
Recall the explanation at the beginning of this article. The row reduc-
tion process is just an abstract imitation of the elimination method we
adopt in solving simultaneous linear equations, except that we now do it in
an organised way. The row reduction worked out in Example 5.22 is there-
fore also the method of solvmg the system
X -j 2y - - Z = 0
-x + 3y = -4
2x I- y + 3z - -2
X Z = -1, + )I+
giving the solution x =- I, y -- -1, z -=- -1.
This takes us to the subject of the solution of a syi.tem of linear equa-
tions, which we st-all deal with in § 5.8. '
Problem Set 5. 7
,~
1. Reduce the following matrices to the row-reduced echelon form :
-1 17 rJ 2 3 17
(a)
L3
-1
1
2
1 _J
I (b)
I4
L2
3
1
5
I
2
O_J
I
r1 3 2 07
r1 2 3 4 -17
2 0 1 I
{c) I 4
1
0
s
2
--6
-2
101
6_J
(d)
I
2 3
0 I
L2
L3 -1 0 l_J
r-1 1 1 07
r1 2 0 07
I I I
l:
1 -1
(e)
0 I 2 -1 II
I
(f) 1
2
-1
1
_;J
L 1 l 0 - 2_J
'r3
I
1 -17 r o 6 6 17
I 2 3 -8 .7 2
(8) (h) •
b.1
4 0
s
I
3.J b. 1
-3 2
l -1
l
:J
JH / IIATRICD
ri o 1 97
5 2 2 JO
(i)
3 -2 3 11
L2 -1 3 8_J
2. Determine the column rank of the matrices in Problem I by two diffe-
rent mMhods.
3. Determine the row rank of the matrices in Problem 1 by two different
methods.
4. Determine which or the square matrices in Problem 1 are nonsingular
and in each case find the inverse.
S. Solve the following systems of linear equations by using the row•
reduction method :
(a) 'J X - Jy - l (b) y- 2z = 3
2x- Y+ Z= 2 3x + z=4
3x+ y-2z= 1 x+y+ z-=-1
(c) x- y+ Z= 0 (d) x - y + 3z = 1
2x+ y- 3z= 1 2x+y- Z= 2
-x+ y + 2z = -1 3x - y + 2z -== 2
(e) x+ y-2z= 3
3x+ y- Z= 8
2x- y+ Z= 0.
6. Prove that •,...,• is an equivalence relation.
7. In the 7 x 11 matrix on p. 187 find the numbers k1, k1, •• , k, or
Definition S.7.4.
8. Prove Lemma S.7.7.
9. Write down the details of the proof that the row operation of type Ill
does not affect the row rank of the matrix.
10. True or false 1
(a) For any matrix .tC, the ranks of ..4. and .A.1 are the same.
(b) The row-reduced echelon form of a diagonal matrix ..4. is .4 itself.
(c) Every upper triangular matrix is row equivalent to a diagonal
matrix.
(d) There exists a 12 x 7 matrix of rank 10.
(e) ne row-reduced echelon form of a symmetric matrix is also
symmetric.
•
(f) The type DI operation performed on the columns of a matriJ
~ves its rank unchan,ed.
S.8 SYSTEMS Of LINEAR EQUATIONS / 1'5
5.8 SYSTEMS OF LINEAR EQUATIONS
In this article we shall study the general system of m linear equations in
n unknowns.
«11X1 + at11X1 + ,. • + «111X11 = b1 }
where A=
This theorem and Theorem 4.8. I, properly interpreted for our system,.
of equations (NH) and (HJ. give the following comprehensive theorem.
5.8.2 Theorem Consider the systems of equations (NH) and (H). Let the
rank of A be r. Then
ta) (NH) has a solution iff (A, b) has rank r.
(b) If r - m, then (NH) always has a solution, whatever may be b
E ~- •
{c) If r - m =-= n, then (NH) has a unique solution, whatever may
be b E V,,.; and further (H) has a unique .•olution, namely, the trivial
solution.
(d) If r •- m < n, whatever may be b E V.,., (NH) as well as (H)
ha,•e an infinite number of solution:r. Tn fact, r of tht: unknowns can
be determined in terms of the remaining (11 - r) unknowns, whose
,,afues can be arbitrarily chosen.
(e) In the cases (i) r < m =--= n, (1i) r. < m < n, and (hi) r < n <
m, if tNH) ha, a solution, then there is an infinite number of solutions.
In fact, r of the unknowns can be determined in terms of the remain-
ing (n - r) unknowns, whose values con be arbitrarily chosen. Further,
(H) hos 011 infi,ute number of solutions.
(f) In the case r -- n < m, (H) has a unique solution, namely, the
trMal s lution, and if (NH) has a sol11tion, then that solution is unique.
(g) If m :..... n, (H) hos a nontrivial solution iff A is singular.
Proof: Parts (a) and (f} are only restatements of Theorem 5.8. I.
They arc included here for completeness. Part (b) follows, since r = m
implies the range of A is V,,.. Part (c) follows from part (b) and Theorem
5.8. I (b). Now we ha\e to prove only parts (d), (e), and (g).
To prove part (d), first note that, since r = m, the range of A is V,..
and so every b E V,.. has an A pre-image in V,.. The kernel of A has the
dimension n -- r > 0. So the kernel K, being a subspace, has an infinite
number of \ectors in it. By Theorem 4. 8.1, it follows that the solution set
of (NH) is a trans1atc of K. So it also has an infinite number of vectors.
The fact that r unknowns can be determined in terms of the remaining
(n - r) unknowns can be seen from the row reduction process of A (see
Example S.23).
To pro\e part (e), note that the proof is the same as that of part (d),
except that, first, we should know whether there exii.ts a solution. Once
the solution exists, the rest of tte argument is tJ-e same.
Finally, part (g) follows from parts (e(i)) ar.d (c) once we recall that an
n x n square matrix is singular iffits rank is less than n. I
The result of this theorem cln be presented as in Table S. I.
Let the rank of matrix A be r and that of the augmented matrix (.4, b)
be , 1• Further, Jct the rank of the kernel be k = n - r. Obviously, r <:
r1, r <: m , r <; n. (Numbers within brackets refer to the pa'11 of
Theorem S.8.2.)
$ 8 SYSTEMS OP UNBAR EQUATIONS/ 197
TABLB 5.1
=I 1 0 l -1 1 1
l ,. ., l
'1 - 2,••
I
1 0 1 -1 1 1
'a - 2ra
L12 2 8 0 2 l2_J LO 2 -4 12 -10 oJ
,...,
r1 0 1 -1 1 17
interchange
r 1•and , 1 ,
ra Xi
I 0
LO
0
1 -2
-1 1 -1
6 -S
ol
O_J
I
,...,
r1 0 0 0 0 17
'ra1 +- '•·
2ra I0 0 -] 1 -1
0 I
LO I 0 4 -3 O_J
11 0 0 0 0 17
,...,
interchange
r 1 and r3 I0 1 0 4 -3
0 I
LO 0 -1 -J • O_J
,...,
rt 0 0 0 0 17
ra X (-1)
I0
LO
I
0
0
I -1
4 -3
1
0
O_J
I·
This shows
Xi = I
Xa + 4x, - 3x = 0 6
Xa - x, + X = 0. 1
So Xi= I
x11= -4x, + 3x1
X8 = x, - x1 •
Rank of the coefficient matrix A = 3 = rank of the
augmented matrix.
S -3 - 2 unknowns can be arbitrarily chosen. They are x, and x1 • Three
unknowns x1, x 1, and x 1 are determined in terms of x, and x1 • There is an
infinite number of solutions. This is case (d) of Theorem S.8.2.
The set of solutions can be written as
{{I, -4x, + 3x1, x, -x" x,, x.> I x,, x 6 arbitrary scalars} are
-= {(I, 0, 0, 0, 0) + x,(O, -4, 1, 1, 0) + x1(0, 3, -1, O, 1) Ix,, x1
are arbitrary scalan}
5.8 SYSTl:.MS OP LINEAB. EQUATIONS / 199
r~ _: =: -: :7 rt 47
2 4 1
,...,
r11 - 2r1, 0 -4 -9 -S -4
= 'a - ri,
r, - 3r1 0 -4 -S -1 -4
,..,, r1 o -I -I 27
r1 - 2,1, 0 1 t t 1
r, + 4r 1,
r, + 5r1 0 0 4 4 0
LO O -1 -1 -2.J
r1 o -l -f 27
,..,, 0 1 t t
r8 X l,
r, X (-f)
LO
0 0
0
1
1 1
1
:J
-,,
,.'1 +- ,,....
r1
0
0
1
0 ;-1
0 -1
27
1
,,- '• 0 0 1 1 0
"LO O f..J
O O
The last row shows that Ox1 + Ox1 + Ox. + Ox, ,.. 8/7, which is
-200 / MATRICES
absurd. Otherwise also it sliows that the rank of A is 3 and the rank of
(A, b) is 4. (Why?) So "the system cannot have a solution. In other
words, the equations are inconsistent.
Ax= b.
Therefore, ..4· 1(Ax) = .4.- 1b
or (..4-1 A).l. =-- ..4-1b
or l.x = .i4-1b
or x-= ,4- 1b.
So, solving the equation Ax = bis equivalent to finding .i4- 1b. But the
solution of equations can be done by the process of row reduction. There-
fore, the method of row reduction wlll also be adaptable to find A-1b and
consequently A-1 • We can write the equation ..4x = bas
A1e = 1,.b . (1)
Therefore A-1Ax = .4-•1"b,
i.e. l"x -=-- .4-1b . (2)
In the process of reduction we always work from A and arrive at I,.
(see Coro lary 5.7.12). Equations (1) and (2) md1cate that, ,r the same
row reduction 1s applied to I,., the 1dentaty matnx, then we end up with the
matrix .i4-t on the right-hand side of (1). Thus, we have the followmg
method of mvertmg A..
Perform a sequence of elementary row operations on A so that it
reduces to I,. Perform the same sequence of elementary row operations
on the matrix 1,.. The resulting matnx 1s ,4- 1 •
In prart,ce, we perform the sequence of elementary row operations on
A and In simultaneously, keeping them side by side as illustrated JD the
followmg example.
Example 5.25 Invert the matnx
1 r
1 07
A= j 1 -1
LI -1
1
2_j
I·
Note Whether the matrix is nonsingular or not w,11 also be clear by the
process of row reduction. If the final rnatnx has all its rows nonzero, then
the original matrix should be nonsmgular. (Why ?)
We write the 3 x 3 matrix A and the matrix la side by side and
obtain a 3 x 6 matrix with a veruc:al line separating the entries of the
matrices A and / 1 as shown below. Then we perform the elementary row
operations on this 3 x 6 matrix in r,uch a way that the entries on the left-
hand side of the vertical line ultimately form la, The entries on the right-
hand side of the vertical bne then give the required inverse of A. The 3 X 6
matrix used here 1s called the enlarged matrix of A. '
r1 1 o 1 o
ILI
l - 1
-1
1
2 0
0 l
0 1..J
202 / MATRICES
r1
I
1 0 1 0 07
lLO 0
-2
--2 2
1 -1
-1
1 0
0 l_J
ra
'a
,...,,
X
X
(-½),
(-½)
I~
LO
1
1 --¼
1 -1
0 1
l
0
o -l_J
07
0 I
,...,, r1 1 0 l O 07
I 0
1 --l i -l OJ
LO o -I o l -l
11 l 0 0 0-7
,...,,
r3 X (-2)
I
Io
LO 0 1
l -l
0 -1
0
l_J
I
1 I 1 0 1 0 07
,..._,
I
I
I 0
1 0 -1
l
LO 0 1 0 -1 l_J
.....,
1 l 0 0 t 1 -½7
Io 1 0 I -1 ½ I.
I
LO 1 0 -1 l_J
1 -½7
I
Hence,
Determinants
6.1 DEFINITION
We start with some preliminaries on permutations of the set {I, 2, ... ,
n}.
6.1.1 De&nltion An ordered pair (p, q) of distinct poc;itive integers p
and q is said to be an im·ersion if p > q.
For eilample, (5, 3) 1s an mvers1on but (3, 5) 1s not.
6.1.2 Definition Given a permutation P =- U1, ii, .. , in) of the set
{1, 2, ... , n}, we define the bet
cilp-= {(j1,j2), (ji,j3), ... , (j1,jn),
Uz,ia), •··• Ul,jn),
(jn-1,inlJ·
The number of inversions in <l>p ib called the number of inversions in
the permutatio11 P.
Example 6.1 Let P be the permutation (3, 4, 1, S, 2). Then <l>p =
{(3, 4), (3, 1), (3, 5), (3, 2), (4, 1), (4, 5), (4, 2), (1, 5), (J, 2), (S, 2)}. Count·
ing the inversions, we see that the number of inversions in Pis 5.
6.t.3 De8nition A permutation (ji, j 2, ••• , in) of the set {l, 2, .. ,, n} is
said to be an even (odd; permutation if the number of inversions in it
is even (odd).
Examp!~ 6.2 The permutation (3, 4, 1, S, 2) in Exampre 6.1 is an odd per•
mutation, whereas (3, 1, 4, 5, 2) is an even permutation.
Now let us take up the definition of a determinant. · To each n x n
matrix A. of numbers we associate a number called the determinant of A. in
the following manner. Let
204 / DETERMINANTS
This last symbol is itself called a determinant of order n. The precise defi-
nition is as follows. ·
6.1.4 Deftaltion If
r«u «11 «1..7
A= !123 1.
l_«,i e,dJ 17 J1
Here the six possible products are « 1111 11 or,u, cxua i.1:>'u, «u11u,- 31 , «11:x11«88,
«11«11«a■, «1a«11«a1• The &ign, to be attached to these products are
determined ac; follows : for example, t.tke «u«u!X11 The order of the
columns chosen here I!, (2, 1, 3). This i!, an c,dd permutation. So the
sign, to be .:1.ttached to th1c; product 1'1 'mmus' Domg this for every pro-
duct, we finally get \
I A I = «11«1211,, 1n2i1'a2 -t txu:t2,u,1 - O'u0121«aa
+ rL13!X910CJ2 - 1i,OC211IX31 • (3)
As a numencal examplt, we ma) take
2 -1 3
=2 X 4 X 2 - 2 X OX I (-1) XO X +
1 4 0 (-]) - (- 1) X 1 X 1 + 3 X 1 X 1
3 X 4 X ( -1) = 32 .
-1 1 2.
A careful arrangement of the six term1, m Equc1t1on (3) gl\,es
«n
Clearly,
~1 ~.
if A. is a 1 X 1 matrix [«11, then det .A. = «1,
When it comes to considering determinants of the fourth and ,higher
orders, the calculation invoJved in writing out all the terms becomes prohi-
bitive. So we study further properties of determinants and arrive at
shorter methods of calculating det .4. However, certain special detcrmi•
nants ~ be eval •.&ated directly from the definition, for example,
206 / DETERMINANTS
A1 0 0 0
0 >-1 0 0
0 0 0 >-.
In particular, det I. = 1 and det O.x,. = 0.
r: : :]
3. Find det A for the given matrix A :
•
(a) A = (b) A = [: : :] .
Li o o 1 1 o
4. Evaluate the following determinants :
I x x2 I X yz a b
1• cl
I:
(a) 1 )' y• (b) y zx (C) • C a b,.
I
1 z z'I. z X)' I b C a1
s. Prove that the solution of the system
«1X1 + «1X1 = 81
~1X1 + ~zX1 = 8~
I81
8a
"a
fia
I I«1
~1
81
s.
I
is = = ,
"~•- I
X1 X2
I I
"1
~1
provided that «1~ -I= «1~1-
"1
~1
•
:1
6. True or false ?
(a) If (a, b, c) is an even permutation, then (b, a, c) is odd.
(b) The sign attached to the product or:11a.u«,1 of a third order de-
terminant is 'plus'.
6.2 FUNDAMPNTAL PROPU.nas o, DBTDMINAND / 207
0 1 31
(c:) 2 0 -4 1 =0
-5 5
ol .
0 1 1
(d) 2 0 0 = 0.
-5 s s
12 - ).
(e) The equation 3
1 - ).
j -= 0 has no real root.
4 I
(f) The determinant of a t1iangular matri,c is the product of its
diagonal entries.
r1 21
(g) I I -= -2.
L3 4_j
(h) I 1
-I -1
; I= 0.
Examp/e6.J 2 3 -l 4
4 0 -I
- 0
1 2 -3
4 0 -- I
without any further calculation.
6.2.3 Theorem If A has a row of zeros, then det A - 0.
Proof: det A = >; E • • ) (1!I/ «2 .... « .. If the k-th row is the
lh,12, , ,,. 1 11 n1,.
zero vector, then ,xk. is zero for every choice of it• Therefore, each
"JI:
product is zero and thi'I gives det A - 0. I
Example 6 6 2 1 3,
b1 b8 b3 b, c1 C'z c3 c,
II 31 3a 33 3, 31 31 83 3,
Proof of Theorem 6.2.4 Let b11 = (b11u b111, ... , b11J and ca: = (cu,
c111, ••• , c11J, Then r1.kJ1, = blrJ,, + ck/11 for every k. This shows that each
product "JJi «211 ••• «k/11 ••• «n1,. can be written as (r1.1/a «211 ••• bk/1: ••• r1.ttJ)
+ (atl/1 "1.J•... Olk/• ... Gt"'·>·
6.2 FUNDAMENTAL HOPl!P.TJES OP DlTBIUONANTS / 209
The alaebraic sum of the first type of products with the appropriate
signs gives det B, and the algebraic sum of the second type of products
with the appropriate signs gives det C. Hence the result. I
6.2.5 Theorem If the matrix B is obtained from A. by an elementary
operation of type II (i e. multiplying any row ,·ector by a nonzero
1calar c). then 'det B = c det A.
Example 6.8 a1 · h1 c1 I n1 b1 C1 1
a3 b3 c8 a9 b3 c3 I
Proof of Theorem 6.2.5 Let the k-th row of A. be multiplied by c.
Then the k-th row of B is (c«u, c2., 2 •••• , cxtn). So, in the expansion of
det Beach i,roduct is of the form :it 11 «1'1 .•. (c«k.) ... « .• This shows that
I I Jlt 11.Jn
every term of the expan,ion of det B i'i a multiple of the corresponding
term by a factor c, in the expansion of det A. Hence. det B -== c detA. I
Note that Theorem 6.2.5 holds even if c -:.- 0.
6.2.6 Theorem // tf,e matrix B is obtained from A. by an elementary row
operation of tJpe Ill (i.e. adding c times the s-th row to the k-th row),
then det B --= der A.
Examplc·6.9
<"1 C1. Cs
Example 6 10 2 3 1 2 3 I 2 3
4 4 4 4 4 4 - 4 4 4 =0,
S 6 7 5-1 6-27-3 4 4 4
by Theorem 6.2.2.
Proof of Theorem 6.2.6 Let the Jc-th row of B be ,,. + er., where r.,
and ,. are the k-th and .r-th rows of A. So, by Theorem 6.2.4,
at11 at11 at1,. ~11 «ui «1..
.:
at11 «11 «1.. c«,1 COl,z c«.,. -k-th row.
I
detB = ·+
Giel 111•• 111...
I «.1 «,. "•• ➔ 1-th row.
. I!
11.., «., ... ix,.• • «111
"•• at,."
210 / DETERMINANTS
I :
I •
I a.1
I
det B = det A ~ c I:
I
I
1 «•1
I
.
:
I r, nl 1Xr,J 1Xnn
l 6 62
0 1 01. +~I
-= (et - ~) (fi - 6) (6 - r,.) 0 0 1 I
1 6 52 I
s. Without cxpandinf, prove that
l 1 3 b +C l a
(a) 2 9 1 =0 (b) l -1- a 1 b =0
4 11 7 b ta 1 cl
a b C y b q X y z'I
(c) X y z - X a p = p q rl
I
I
p q r z C r I a b cl
I:,I :;i y
z I 1 1
x2 ,-2 z2 1
1 I
(d) ya
lyz
= xa ya :za'
:zx
a 1-- b b+c e +a a b C
x+3 x+6 4 7 10
(0
X
3a 3b 3c e f e + r
a,+ a1 P1 + Pa 11 -t 1. 01 P1 11 I Oa Pa I,
(e) b1 -1- b1 ql + ql m1 t- ma - b1 ql ml + I
b1 q, m1
I
C1 + Ca r, + '2 t- nz
n1 C1 T1 n1 I c1
'• n1
a a' a' a a' a'
(0 b b' b' = - b b' b'
, ,
'c c' C C c' C
a 0 01 Id 0 0
(g) b 0
0 I=I leIo I
0 0
0 d e I
a b
I01
a,. Oa i a1 - b1 a1 -b1 a:s -ba
(b)
(1
C1
ba ba•
Ca Ca j
= h1 -
I C1 -
t'1
01
b1 - Ca ba -
c1 - a 1 Ca - aa
Ca
t t mterchanges t
(jk1' i1, ... , i,._1-I, ik1+l, (k1, 1, 2, ., k1 -- ], k1 + I, ... , n)
... ,i..)
( ik 1' j1t .. , 11.,; ·• ,j.. ) (ki, 1, 2, ... , k 2 , • , n)
s interchanges
\
The tin,t line ID T.ible 6.1 md1cdtes the order of rows (--PJ and the
order of columns (ndtural order) in product (2). The last line md1cates
the order of rows (natural order) and the order of columns (=Q) in
product (3).
The passage from product (2) to product (3) presented in Table 6.1 can
be realised by successive interchanges of adjjcer.t entries in the respective
permutations. Since P = (j1 ,j2 , ••• ,i,.), look forJi<1 = 1 and brmg it to
the first position by interchanging 1t successively with all lhe entries
preceding it. Thu, needs a certain number of interchanges, say t. The
effect of these interchanges is to reduce the number of inversions in P by t.
If we make the sami kind and number of mterchanges in the order of
columns, namely, the natural order 1, 2, 3, ... , n, we get (k1 , 1, 2, ... ,
k1 - 1, k 1 + I, ... , n), the effect of which is to increase the nut'nber of
inversions in (I, 2, ... , n) by t.
Now we look for ik = 2 and bring 1t to the se~ond position by
making successive interchanges. Suppose this requires s interchuges.
The same s interchanges carri~ out on the order of columns will bring k 1
216 / DBTBRMINANTI
6.4 COFACTORS
The expansion of an n-th order determinant is the sum of all the n!
terms, each of which is a product obtained by taking one (and only one)
element from each row and each column. Let A. = (ac,1) 11 x11 , Suppose we
collect all the terms in the expansion of det A., in which the fixed element
«u appears as a factor, and write down their sum in the factored form
•uA". Here A." denotes the factor remaining when the element ac11 is fac-
tored out. We call A." the cofactor of «u in det A..
Example 6.13 «u «11 «11 == «11(«..«aa - ataa«aa)
A11 = I «11
ataa
at1•1
«1:a
and so on.
Example 6.15 If .4. = (a.11).x,., then
«a, n-1
Gtn-1, 1 «n-i, 1
= A,.,..
Now we take up the general case and reduce it to the special case. We
have
p Q R
.......
detA. = ·······
s . a.u { T
...... : t
·······
u V . w
where P, Q, R, S, T, U, V, and Ware mere abbreviations-for the remainiq
entries. U, Y, and W have (n -- I) rows; R; T, and W have (n - J)
2J8 / DETERMINANTS
columns; S and T have only one rcw, and Q, V have only one column.
Clearly,
(2)
S ex11 T
Again, by another series of interchanges of the j-th column with (n - j)
columns of R, W, and T and by apply mg 1 heorems 6.2.1 and 6.2.9,
we get
p R Q
a I cl
The cofactor of h occurring in the second row, first column, is
(- J)l+I I: ; I= gh - qf.
64 COFACTORS/ 219
Since this determinant is symmetric, the cofactors of elements,
which are symmetrically placed with respect to the diagonal, are the aame.
So, for example, we need not distinguish between ac11 -= h and «ti == h.
Both h1Ve the same cofactor. (Check!) Denoting the cofactors by
capital letter-i, we have
A -= be - /2, B = ca - g1 , C = ab -- h1 ,
F = gh - af, G = hf - bg, H = fg - ch.
The n! terms th.it occur m the expansion of llet A can be separated
into n groups ai, follows :
•
(I) all tenns contaimng «11
(2) all terms containing ac 11
1 0 -1 2 1 0 -1 2
fl = =
7 0 -s 8 r, - 4r1 7 0 -s 8
-1 4 2 1 -9 0 14 -ts
Expanding fl in terms of the second column, we get
- 1 -1 2 1 0 0
fl== 7 -S 8 == 7 2 -6 •
c,_+ C1o
-9 14 -IS c1 - 2e1 -9 S 3
where c, denotes the i-th column. Now expansion in terms of the fint
This determinant is zero, because two rows are 1dent1cal. On the other
hand, we can expand det B m terms of its j-th row. Smee the deletion of
~heJ-th row and the k-th column of det B gives the same 111-,, as th"
deletion of the J-th row and the k•th column of det ..4., we get
0 = det B = cz, 1A,1 + cz,1A,1 + ... + «111 A, 11
= r, · R1• I
6.4.5 Remark A similar result as in Theorem 6.4.4 is true for columns.
6 .•.6 Corollary Let A be the matrix (cz.,)11x11 and B the matrix (~11).x.,
wMre ~., = A1,. Then
fl 0 0 07
AB= BA=
0 t,. 0
U=M.,
0 0 0
where t,. == det A.. Consequently, det (AB) = t,,.• == det (B..4.).
'lbe proof is obvious when we carry out the multiplication of .4 and
B and repeatedly use Theorem 6.4.4.
222 / DETERMINANTS
A=l0 l 1/.
LI 2 l_j
The cofactors of the different entries in det A are
A11 = -1, A11 = I, A13 = -1,
..411 = 1, A11 = 1, Asa =- -3,
A11 = -1, A 31 = -1, Aaa = I.
Therefore,
r-1 1 -17
adj A= I
I
-I I·
L-1 -3 ]_J
The adjoint is of sigo.ificance in finding the inverse of a nonsingular
matrix, as will b;: shown at the end of§ 6.S.
(a)
2 I 0
0 3
II 2 4
(b) 12 3
3
(c) 6 I; I
0 3 2 13 4 2 10 20
I
3 2
15 29
41
2 141
I 3 8 7 6
I
7 4 10 2
I 1
I 2 3 4
1 I
s
(d> I (e) (fl I 3 6 10 15
I
I
16 19 3 17 6 8 s 8
I 4 10 20 3S
: 33 3~ 8 38 19 s 3 9
I SIS 35 69
4. Without expanding, prove th,tt
I yz y + z 1 I I
1 ZX Z +X - X Y Z
I X)' X +y
6 4 COFACTORS / 22J
S. Prove that
1 X x1
(a) 1 y y• = (x - y) (y - z) (z - x)
z z1
X JZ
Ix, t'3
x. "•
X1 l.:a -
3
JI (X1 t Xie.>, -f X461!)
I
I I
I XA Xa Y1
This determm,int 1s called a c,rcula111 ot the third order Write
dow,1 a circulant of order n Write al!.o its value
7. (a) Prove that the equation of a ui cle throurh three pomts (xi, y 1),
(x1, l 1), and (Y3 , h) 1s g1~en by
x~ + y' x y l I
xf t- yf X1 Yi I
0
l.: I+ y: 1 11 __
XA >'•
YI + J'i 'ta Ya l I
(b) Determine the equation of the sphere passmg through the points
(Y1,) t• z1), (l.,. Yz, Za), (xa, Ya, Z1), and (x., Y,, z,) 1n the deter•
mmant form
8 Solve the equation
X t- a b C
C x+b a == 0.
a b x+c
221 / DBrBR-.1 NANTS
A== ls 9
6
10 11
1
12.J
ThenB=
r•
I
2 37
I is a submatrix of A and I119 .1131 is a minor of
L9 10 lt_j
A and also of B.
6.5.2 Lemma (a) Let A be an m x n matrix. Apply som? elementary row
operations 10 obtain a ne,v matrix Ai, Delete some columns of Ai to
get a submatrix B1 of Ai, The same submatrlx Bi could also luwe
been obtained if we l1ad first deleted the same column., from A Itself
to obtal11 a aubmatrlx B r,nd then applied the same elementary row ope•
ration, to B,
6.5 DETERMINANT MINORS AND RANE OF A MATRIX / 225
(b) I/ Bi la a square matrix (and therefore B also), then det B ta tf/Ulll
to some nonzero scalar multiple of det B1 :
The proo! is obvious, because of the way elementary row operations
work. Deletion of a column or columns will not affect the result of the
elementary row operations on the remaining submatrix.
Part (b) is qnly a restatement of Remark 6.2.7.
6.5 3 Theorem If A is an m X n matrix, then the rows of A are £1 tJf A has
a nonzero minor of order m.
Proof: Let A1 be the row-reduced echelon form of A. Suppose the
rows of A are LI. Then .A.1 would contain m nonzero rows. The m
columns containing the 1's at the steps will aiso be nonzero. Delete the
remaining columns of .A.1 and obtam an m x m submatrix B1 of A 1 • Ob-
viously, B1 = I,. and det B1 = det T.,. = 1. By Lemma 6.5.2 (b), the cor-
responding square submatrix B of the matrix A will have the property that
det B is a nonzero scalar multiple of det Bi, which is 1. So det B =I= o.
Thus, A has a nonzero minor of order m.
Conversely, suppose A has a nonzero minor of order m. Let B be the
corresponding submatrix. We shall now prove that A1 contains m nonzero
rows. Suppose, if possible, A1 has a zero row. The same operations
that reduced .A. to A1 will reduce the submatrix B to B1 • Since B1 is of
order m x m and at least one of the m rows of .A.1 is zero, it follows that one
of the r.)WS of B1 is zero. So det B1 = 0. But, by Lemma 6.5.2 (b), det B
is equal to some nonzero scalar multiple of det Bi, which is zero. This is
a contradiction. Thus, all the rows of A1 are nonzero. In other words,
the rows of A are LI. I
6.5.4 neorem The rank r of an m x n matrix A is the order of the largest
nonzero minor of A.
In other words, the matrix A is of rank r ijJ there exists a minor of
order r which is nonzero, and every minor of order s > r is zero.
Proof: Let p be the integer such that there exists , minor of order
p, which is nonzero, and every larger minor is zero. We shall lk>w prove
that r = p.
Since .4 has rank r, r rows of A are u. Choose such r rows. Let B be
the submatrix of .A obtained by deleting the remaining rows. B is an r x n
matrix, all of whose rows are LI. Therefore, by Theorem 6.5.3, B contains
a nonzero minor of order,. Hence, r < p. ,
On the other hand, look at the nonzero minor of order p. This gives
rise to a p x p submatrix B. This ht turn gives rise to a p x n submatrix
C bf~- C contains a p x p nonzero minor. Hence, by Theorem 6.5.3, the
rows of C are u. This meaos that .4 co,tains p LI rows and so p < r.
Thus,,== p. I
226 / DE'IERMJNANTS
I -3 0 1 -3 =- 0.
6 -1 1 0 1 -3
A(~B)~ t!B)A=l.
~
T her.,ore, A-1 = 1 B
'ii: = fll a d'J A
= det
1
A ~~ 7J·
LA1,. A1,. A,.,.
Example 6 22 Find the inverse of the matrix of Example 6.19.
r
We have
1 -1 07 --1
A==
[
0 1lj and adj.if== 1
l 2 1 L-1
lf~re det ..4 = -2. Therefore,
6 6 PRODUCT OF DETBDIINANTS / 227
L
3 6
1 s
1
3_j
I·
0 -1 I O -1
1bmddA =:; 3
2
4
l
=
r1 - 2r1,
0 3
0 2
6
2
'• - '1
-1: :1- -1 2 0
-6.
and det B- 3 6 I -=
I ,,'• +
+ 3r1,
,,
' 3
6·6 nooucr OP DBl'llltlalWm / 229
det (AB) =
-2 -3 -3
11 42 IS =
-2 -3\ --3
11 42 15 =
-2 -3 -3\
1 27 0
6 19 5
'•+Sri
0 10 -4 0 10 -4
l
0 51 -3
(b) Z O X == xy yz
y X 0
2}'1' - :JI:' zl
zx
,. yz "'+1•
X y I 1
(c)
,.
z' 2zx - y'
:JI:' 2xy - z'
,:I = Y
Z X y
Z X •
230 / DETERMINANTS
7. Prove that
«11 «u oc1a i"·~
I
-
6 7 APPLICATIONS TO LINEAR EQUATIONS/ 231
The last equality can be seen by exp,.inding this new determinant in terms
of its first column :
} ~J IX29 IXu
So Xi= det A
In the same manner, if v.e multiply the fir!>t equ 1t1on by Au, the
second by A2, and so on, the last by A 111 , and add, we get
IXu 0'12 ••• IXi, r l 131 IXi, ,tt °'111
: ..• L..f3• .J
x• .J t..A1• A,. ... A.. .J
Therefore,
1
"' .... det A (~1A11 + ~A11 + ... + ~.A"')
232 / DBTBRMINANTS
0 1 -1 = 1. JI= !1 0 -1 = -1,
1 1 -1 1
2 -3 I
and z =--r1 I 1 0 = -4.
-1 -1 I
Note that if det A = 0, then A. has a rank less than n and so Ax == 0
bas nontrivial solutions. In t.bis case Ax = b may or may not have a solu-
tion. Even if it has a solution the solution will not be unique. Converse-
ly, if Ax == 0 has nontrivial solutions, then rank A is less than n. So
det A. == 0. Thus, we have proved the following theorem.
6. 7.1 Theorem q A. == (a:u)11X•• then A.x =0 has nontrMal solution, i/1
det A =- o.
6.7.2 R•art When det A = 0, Cramer's rule cannot be used. (Why?)
Problem Set 6. 7
I. Use Cramer's rule (if applicable) to find the solutions of the following
systems of linear equations :
(a) 3x + y = 1 (b) 2x - 3y = 1 (c) X + 2y + 31' - 3
5x+2y==3 x+4y-l 2x - s-4
4x+2y+2z=5
6.8 DGBNVALUl81 mGl!NVBCTOU / 233
(d) x - y + 2z = 1 (e) X + y + 2z = 3
2x +2z=3 2x+2y+2z==1
3x + y + 3z = 7 3x + 4y + 3z = 2 •
2. Repeat Problem l for the systems "r linear equations of Problem 1,
Problem Set 5.8.
3. Find the inverse of the coefficient matrix for the systems of linear
equations of Problem I by any method, and use Equation (4) to solve
these systems.
4. If t'he coordinate axes in a plane are rotated through an angle «, then
the old coordinates (x, y) are expressed in terms of the new coordi-
nates (x', y') as
x=x'cos« -y'sin «
y = x' sin « + J/ cos « •
Use Cramer's rule to express (x'. y') in terms of (x, y).
lI
«11 «11
i.e.
I
Gt11 «n - ).. Gt13 X1 1-•o.
L «11 «.1 «u-l...l Lxa...l
234 / DETl!RMINANTS
A== I 0
Ll ::J
To find the eigenvalues of .4, we solve
6 8 EIGENVALUES, EIGENVECTORS / 235
I - 'A -1 2
0 1 -A
1 2
This gives
(1 - 'A)((l - >.) 2 -I (--2)) = 0
or (I - >.)(>-2 - 2>. - - 1) =-- 0 , •
i.e. ). = 1, 1 + v2, 1 - y2.
So the eigenvalues of A are
11 == 1, >-2 =-= I -I v2, >-a =- I - v'2.
To find E(A 1) = E(l}, we write
r-1 - l 27 r-x 1 7 1 -1 fl 0 r-x
I O O II Xi !~ I I I:~ l I O
x~ 1
LI 2 I _J L_x3 J L xJ _l L._ 0 0
r-o -1 r1 1x1-1 107
i.e. I o, 0 0 j I Xi : - ! 0
I_ I 2 O_J I_ ,J _J I _0
or - .\J + 2x 3 = 0, ., 1 -I 2xl =- 0.
This gives x 1 =- 2.:\ 2 nnd AJ = 2xJ, The eigenvectors corresponding to
the eigenvalue I a1c Lherefore of the form ( 4xJ, :ZxJ, x 3), 'whe1e x 8 =t= 0.
Therefo·e,
l,(I) = [(--4, 2, 1)1\{0}.
To find .E(i.2) - ~ E( I ~ \/2), we wtite
r- 1 -1 2 -1 •'"1 7 r- r-.,.1 7
IO 1 0 I ! ·'"2 I = (I !- y'2) X2 I,
I
0 0 >.a 0
LO O O >..
P100/: The set {1111 v1, ••• , v.} is u by Lemma 6.8.3. Since the
domain space is n-dimensional, it follows. by Theorem 3.6.7, that the set
6 8 EIGBNVALUU, EIOENVBCl'ORS / 237
{vi, , 1 , ... , v"} is a basis for the domain space of A. For the second part
of the tbe'lrem, note that Av, = >.,v, = Ov1 +
Ov1 + ... + 0v,_1 + >.,v, +
Ov,+1 + ... + Ov". Thus, the i-th column of the required matrix is
(0, 0, ... , ).., .. 1 O)T •
t
,-th place
Hence the theorem. I
l11 such a case the matrix is said to have been diagonalised.
(a)
r3
I
17
I (b) I
1-0 -1-1
I (c)
I 0
I I
3 1
(d)
13 l2
2
o2
47
L6 2_J L_l O_f L2 - I _J
L4 2 3_J
r 3 2 l 0"'1
:1 0 07 r-o ; I ]
I
I
0 l 0
(e}
·l2
1
0
(f)
Ii 0 l
I
(g)
0 2 -l 0
L3 2 O_J Li i O_j
LO 0 0 i.J
2. Diagonalise the matrices of (b), (cJ, and (g) m Problem I.
3. Find the characteristic polynomial of the matrix
r l 0 0 07
-1 i 0 0
2 ½ -I 0
'
L t -i '11: -lJ
Diagonalise this matrix, if possible.
4. If). is an eigenvalue of the matrix A, prove that
(a) ).1 is an eigenvalue of A 1
(b) )." is an eigenvalue of A"
(c) «>. is an eigenvalue of «A, whe1 e « is a scalar
(d) gQ.) is an eigenvalue of g(A), where g is a polynomial.
S. If x is an eigenvector of A corresponding to the eigenvall•e ).1 prove
that
(a) x is an eigenvector of A" corresponding to the eigenvalue >.•
(b) ~ i$ an eigenvector of g(A) corresponding to the eigenvalue I(>.)
238 / DBTEllMJNANTS
6.9 WRONSKIANS
In this article we shall use determinants to discuss dcpendeuce anb
independence of functions. These ideas help in an understar.ding ot the
theory of ordinary linear differential equations. First, we shall c!efine the
Wronskian of functions.
6.9.t Definition If Yi, Yh • , y,, are n functions in •(/ '"- 11 (/), tl--en the
determinant
Y1(x) Y2(X) y,.(x)
-sinx 1
= -x sin " -- cos x .
(a) 1/ there exists x0 E I such that W(x0) =I= o. then Y1t >'•• ... , Yn are
LI over I. •
(b) _J/ there exi,ts X 0 E I S&U'h that W(x0) = 0 and if ,Yi, y1 , •• , Yn
are solutions of the dijferentlal equation
a0ylnl + ai11t•-11 + ... + Dn-i y' + DJ = 0 , ( 1)
where "o• Di, ••·• a,. e: '"(( (1), then Yu y 1 , .... y,. are LD over I.
(c) If Yu Ya, .... y,,. are solutions of Equation (1) and are LI over 1,
then W(x} neh r fJanishes on I.
(d) // Y1t Y1, ••• , Yn are solutions of Equation (1), then either W(x)
= 0 over I or W(x) has no zeros in I.
Proof: (a) Since W(x0 ) /- 0, W(x0) considered as a square matrix
is nonsingular by Corollary b.5.5. Hence, by Theorem 5.5.2, the column
vectors
,.(xo) = (y,(Xo), ,;cxu), ...• y, Cn- 11 (xu)). i = I, 2, .. , n (2)
arc LI in V,,.. Therefore, y1, y., ... , y,. are LI over/, by Lemma 4.9.5.
(b) If W(x0) = 0, the columns or W(x0 ), namely, ,,(x0) =--= (y,(x0),
y;(xJ, ... , y 1«n-1>(x0)), i = I, ... , n, are LD in V,., by Theorem S.S.2. There-
fore, there exist seal us C1, C1, ••• , C,,, not all .zero, such that
• C1v1(x0) + C1 v11(x0) t- ••. + C,.1 ,.(.T0 ) = 0.
1
This means
CiY1'1:1(xo) + c.y.li:l(xo) + ... ~ C,.y,." 1(Xo) =0, k - 0, I, 2, ... , (n - I),
(3)
for at least one scalar C,. not zero. Consider the linear combination y ==
C1y 1 + C1y 1 + ... + C,.y,.. Since y 1 , y~, •· , y,. are solutions of Equation
(I), y is also a solution of (I). But by Equation (3), we have
y(x0) = y'(x0) = ... = y«n 11 (x0) = 0. (4)
Thus, y is a solution of Equ.ition ( I) and satisfies the initial conditions (4).
But the zero function also satisfies Equation (I) and the initial conditions
(4). Hence, by Theore111 4.9.2 (existence and uniqueness), y(x) = 0 for all
X in/.
Thus, C1y 1(x) + C1 y 1 (x) + ... + C"yti(x) = 0 for all x E / and at
least one C., =p 0. Therefore, Yi, )"1, .. , .Vn are LD over /.
(c) This is the contrapositive statement of (b).
(d) It follows from (b) and (c). I
6 9.3 '11leorem If y 1 , are solutions of
)'1
"9Y + 01)' + D1Y = 0,
11 1
(5)
where a., a1, and a1 belong to fjj' (1), and a0 is noM here zero on I, "then
_ J a1{xl dx.
W(y1(x), y 1(»))'- Ce t1t(x) (6)
Proo/: ,, and 1, •re soh1tione of (S), Therefore,
240 J DBTERMINANTS
• Qa)'~ + 01Y; + Os.Yi = 0
and ao1; + Ol)'~ + a1y8 = 0.
Also WC,,i, Y1J = I1; y;
Y1 Ya 1 ,
I -= Y1Y1 - Y1Y1 •
I
,
dW ,· • • N , , • •
So dx = Yi>'• + YiYa - Y1Yz - Y1Ya = Y1Y1 - >'i.Y•
= Yi ( -- a1
Ya, -- -a, Ya ) - Ya { -
Oo Oo
01 (
= -- ,
Y1Y1
')
- Y1J'~ = -- -01 W•
Oo Oo
Hence, on integration, we obtain
. - l Oi(x) Jx
W[y1(x), y 2{x)] --= Ce a,(x) ,
where C is a:1 arbitrary constant. I
6.9.4 Remark The foregoing result can be extended to a differential
equation of the n-th order. Here "'e state the general result. If
y., Yx, .. , Yn are n solution& of the n-th order normal hnear differen-
tial equation ao1'"1 -t a 1y 1" ·u + ... + an1· -= 0, on a'h interval I,
- I a1~xl Jx
then W(y1(x), J2(,t), , Jn(x)] - Ce a.(x) (7)
Equations (6) and (7) are known as Abel's/ormulae.
i
b3 1
• b1 b,. bJ
To obtain the cross product we expand (I} in terms of the first row as I/
it were a determinant. Note that (I) is not a 'pukka' determinant.
From Definition 6.10. l it follows immediately that
i X j -=-= k, j X k = i. k X i --= i.
j X i = --k, k X j = -i, i ,< k-= -j,
; x i = 0 = J x j = k x k.
.
and
Note that the cros~ prodcct is a binary operation on Va.
Example 6.27 (1, 2, -1) x (0, -1, 3) .
- (I -1
2 -1
3
I· 1-l I· 11
3 0·
1
0 -1
21)
== (S, -3, -1).
242 / DBTIIMlNAN'II
6.10.2 Theorem (Properties of cross product) ut u, ,, w be any three
•ectors In Y1 • Then
(a) u x , = -(, x u).
(b) u X (• + w) = u X v w.+U X
(C) For (lt E R, GtU X ' = «(u X v) = " X GtY,
(d) u • (u x v) = 0, and v • (u X v) = 0, i.e. u X 11 11 orthogonal
to both II and 11.
(e) I II x v 1== I u I 2 I II I 11 - I 11 • 11 I 11-
Since the proofs are straightforward, we shall prove only (e).
Proof o/(e): Let 11 = (alt a 1 , a3 ) and v = (b1t b1 , ha). Then ,
lul 1 l11l 1 -l11•vl 1
=- (~ + a: + a!)(b! + b: + b:l - (a1b1 + aaba + aaba)1
= E a~(b: + b:) - 2 I a 1b1asb1
= I (~b= + a:b!) - 2 I a1b1asb1
= E (a1b1 - a-61)1
== lux,1•. I
6.10.3 Corollary 1/8 is the angle between two nonzero vectors u and II in
Y1, then
I u x II I = I u I I , I sin 8 •
Proof: We have
u • v = Iu I Iv I cos 0 (cf Definition 2.3.3).
Hcnc:c, from (e) of Theorem 6.10.2 it follows that
I u x II I 11 = I u I I I v I 1 - I 11 11 I II I I cos11 8
= I II I I I v I a sin1 8 .
So I u X II 8 <: 1t. I
I = I II I I II I sin 8, since O <
Note that I u I I v I sin 8 is the area of the parallelogram formed by
vectors u and ,. Hence, I u x v I represents the area of the parallelo-
aram formed by u and v.
DIRECTION OF u X v
Given two vectors u and v, we have just proved that
(i) I u X. v I = I u I I v I sin 8, where 8 is the angle between u and v,
and
(ii) u X • is perpendicular to both II and v and therefore to the plane
containing u and,. ·
(i) specifies the magnitude of II x v, but (ii) leaves us to decide which
of the two directions perpendicular to the plane of u and , is the direction
of u X •· We shall now settle this question. •
We start from the ordered triple of vectors (i.J, i). Any ordered
triple of linearly independent vectors (u, .,, w) is said to be a poritiN triple
if, b7 f8dually,changing the directions of u, .,, w b11t without making them
6.10 CllOII HGDIJCI' DI f'1 / 243
linearly dependent (i.e. without making any of the veetora cross the plane
of the other two), we can bring them into coincidence with the directions
of I,}, and le. Otherwise it is said to be a negatlie triple.
Examples of positive triples arc
(i, }, k), (}1 k, i), (k, i, }).
Examples of negative triples arc
(}, i, k), (k,J, i), (i, lc,J).
6.10.4 Lemma Three linearly ind,pendent vectors u = (ai, a1, a1),
• = (bi, b1, ha), w = (ci, c11 c3) form a positive triple iJf
a1 a2 o1 l
b1 h1 b81 I> 0.
Ci Ca Ca •
, c1 Ca Ca
"'"'01
'ha b1
I C1 Ca
II+ a. Iba
t'1
b1
Ci
I + a, I
I b1
Ci
b1
c.
I
a1 a. aa I
I
= bi b1 ba ! (by expansion of the determinant in
terms of the first row).
C1 C1 c. I
ai a1 Di,
(b) U•l'XW= bi b. b,
£'1 C1 Ca
bi ba ba
= c1 c1 c8 ( by two interchanges of rows)
a1 a 1 a 8
=1'·WXU.
Similarly, 1' • W X 11 =
u X 1'■
W •
Thus, 11 • , x w = v • w x 11 = w • u x v •
Since the dot product is commutative, these are also equal to , x w • u
== W )( U • V = r, X 1' • w.
(c) If one of the \!eCtors is zero or , x w ... O, the proof is obvious.
We theref'orc usume that none of these vectors is zero and so also
,xw~o.
6.10 CkOSS PllQpUCT IN Ya/ 245
Hence, the vector v x w is perpendicular to the plane detemuned by
the vectors v and w (cf Theorem 6. I0.2 (d)). So either " x w or -(v x w)
..
makes an angle ; < n/2 with u, as shown in Figure 6.2.
,,
,, ,,
,,,,,,,.. __
f
------·,,,
,,.,,,
,, ,,, I
.,, ' .,, I
.,,.,, f ,,," I
,,, I ,,,, I
_,, I / I
- - - - - -,,- - - - - - - - 7 I
I
f ' f
I ' f
f '
f I I
I I I
I ,' f
------ -t- -----~
I ,,
I , .,,
I _,,
I
I .,,
I /
,.,,,,,
"I
FIGURE 6.2
So lu•vxwl= lul lvxwlcos,f,
= ( I u I cos ,f,) I V X w I
= (height of the parallelepiped formed by u, v, and
w) x (area of the parallelogram formed by v
and w)
= volume of the parallelepiped formed by u, v,
and w. I
6.10.7 Corollary u • v X w = 0 if! u, v, and ware LD.
Proof: From the proof of rart (c) of Theorem 6.10.6, it follows
that
Iu . V X wI = I u I I V X w I cos ,f,
= I u I I v I I w I sin 8 cos ,f, •
Hence, u • v x w = 0 means one of the three vectors is O or sin 8 = 0
or cos ,f, = O. If one of the vectors is 0, then u, v and w are LD, If
sin 8 = 0 or cos ,f, = 0, then the three vectors are coplanar and hence LD.
Sou • v x w = 0 implies that the three vectors are LD. The converse is
obvious. I
We shall now consider two important applications of the cross product
to geometry.
(i) Vector equation of a plane through three glvtn points P1 , P1 , Pa in
space : Let Q be a point in the plane P 1 P1P1 (see Figure 6.3). The vectors
2# / 1>61DIII.NANTS
--+- --+
P1P1 and P,P1 lie in the plane. The problem now reduces to finding the
FIGURE 6.3
equation of a plane through P 1 and perpendicular to the vector
- -- -
~ - -➔
N = P1P1 x P1P3 • Hence, the required equation is P1Q · N = 0,
i.e. P1Q · (P1P1 X P.,P1) = 0 • (2)
This equation is satisfied by all points Q on the required plane and by no
points oft' the plane.
If Pu Pa, and P 3 are (x1 , Yu z1), (x1 , Ya, z1 ), and (xJ, Ya, z3) respectively,
then Equation (2) becomes
X - X1 y - Y1 Z - Z1
Ya - Ya Zs - Z1
Xa - Xa
Example 6.28 The equation of the plane through P 1(0, 0, O), P 1 (1, -1, 1),
and P 1(2, 1, 2) is
X y Z
1 -1 1 = 0,
1 2 1
i.e. x - z = o.
(ii) Shortest distance between two llne.r : If we have two lines in a
plane, then either they are parallel or they intenect. On the other band,
two lines in space can behave in any one of three ways : (a) they are
6.10 CROSS f'RODUCI' IN Ya / 247
parallel or (ltl they inter1iect or (c) neither are they parallel nor do they
intersect. In case (c) they are called &kew lines. Let L1 and La be two
skew lines (see Figure 6.4).
FICiURE 6.4
Let Pi, Q1 be two points on L,,, and P 1, Q1 two points on La- 'I'he
-- -+
vgctor P 1Q1 x P1 Q1 is perpendicular to both L 1 and La- The scalar pro-
--+
jection of P 1P1 qn this perpend1culur is the required shortest distance
between L1 and £.. It is given by
J -+
P1P1 -
• P1Q1 X -P.Q. I
- --➔
I P1Q1 X P.Q. I
Example 6.29 Let the two lines be
x-1 y-2 z-1 x-3 y-2 z+4
L1 3-
: -- = - 2 = S- and La: -,- = - 2- = -r.
Since L 1 is parallel to the vector (3, 2, S), and La is parallel tc the
vector (1, 2, 2), the vector N = (3, 2, S) x (J, 2, 2) is perpendicular to
both L1 and La- We can take P1 as (1, 2, I) and Pa as (3, 2, -4).
Hence, the shortest distance between L,, and La is
2, 0, -5) • (3, 2, 5) X (1, 2, 2)
I (3, 2, SJ X (I, 2, 2)
2 0 -51
= 3 2 S I+ \/'36 + 1 + 16
I 2 2 j
= I -12 - 20 I + v'53
32
=- --=- •
\/'53
,.18.I Remark The method in Eu.mple 6.29 fails in ~ the lines L 1 and
La are parallel. (Why?) In this case let P 1 and (21 be two Points
248 / Dll'IDJIINAN18
--
--
d = J P1P1 P1Q1 I
1 P1Q1 I
-----.,,---....-•------L1
.,,,, ,,
01
.,,,,
,,,,,,
FIGURE 6.5
Hence, the (shortest) distance p between £ 1 and La is given by
P = JI Pi~ I 2 --~ 2
I =....- -
= JI I --+
P1P,. 2 -
I P1P1 · __
__ --:...::.+.
P,Q1_I 11
I PlQl I 1
--
---+ -~
I P1P1 X P,Q, I
= (cf Theorem 6.10.2 (e))
1 P1Q1 I
Problem Set 6.10
1. Evaluate u x v for the vectors u and " 10 the following :
(a) u = i + +
J k, v = 2i - 3} k +
(b) u = i - 2k, v = 2j + 1k
(c) u = (1, 3, O), 11 = (2, -1, -1)
(d) u == (-1, 2, 3), v = (I, I, 2).
2. Prove that O x " == 0 for every vector v in V1 •
3. Prove the distributive law u x (11 + w) == (u x 11) + (u > w) and
verify it for 1he vectors u, v, and win the following:
(a) u == i - 2} + k, v == 3i - j, w = 'li + } - 4k
(b) u = (1, 2, 3), v = (3, 0, 1), w = (2, 4, 5)
(c) u = (-1, 1, 2), ,.. = {I, -1, -1), w = (41 2, -3).
4. In the discussion of direction of the vectOl' u x 11 we have taken
O < 8 < ff, What happens when 8 == 0 or 8 == ff ?
5. Determine whether the vectors u x II and w lie on the same side of
6 JO CROSS l'lt.ODUCT IN f,'1 / 249
~he plane containing the vectors u and v for the vectors u, v, and w
m Problem 3.
6. Fmd the equation of the plane passing through the three pomts
(a) A(l, 2, -3J, B(J, -1, I), and C(2, I, I)
(b) A(l, 2, 3), B{J, 0, 1), and C{2, 4, 5)
(c) A(l/2, 1, -1), B(l/2, -1, -1), and Cl4, 1/3, -3).
7. Prove that for vectors u, v, and w of V,1
(a) (u x v) x w = (u w)v - (v • w}u
(b) u X (v X w) - (u • w)v - (u • v)w
8 Evaluate the product
(a X b) X (c X d)
of the four vectors a, b, c, and d of V8 • P1 ove that this 1s a vector in
the plane of a and bas well as of c and d.
9 Find the shortest distance b,!tween the skew Imes
X-2 L-t-1
3 -= ., 2 =- I - z and x = y - z .
i.e. f~11-
L«n-1
(•u) I;~l
L~,. J
' This pro,es the theorem. I
7.J.2 Theorem Let T he a /i11ear ope1ator on V. Let (T: F, F) he the
matrix B _:,. (~1, ) and (T: G, GJ he the maim: C ~ (y.,). Then, with
(«, 1 ) defined as i'n Equation (I),
B(oc,1) = («u)C • (3)
In other words,
B = (oc,1)C(«.,)-1 • (4)
Proof; We have
T(u,) = I ~.,u., and T(v,) = I r.sv., j = 1, 2, ... , n.
i '
Now, for eachj = 1, 2, .•. , n, \.\e have
T(v1) = T(A(u,)) , (\\here A is defined by (I))
= TlUtkJU1:) • (again by (1))
" =
= Iocr..,T(u.,) I«A:,(I~,11:u,)
k k l
= I(I~ 1:ot1c,)u, .
1 (S)
I k
On the other hand, for each j = 1, 2, •··, n,
T(v,) = Iy,.,i•,. = Iy1: 1A(u1:)
k "k
= Iy.,,Iot,11:u, (by (1))
k I
= I(Iot,.1:Yt1)U, • (6)
I k
Equating (S) and (6) and observing that th u,s are u, we get, for each
=
J I, 2, ... , n,
I,
I~,1:ot1:, = Iot..tYld •
k k
This gives Equation (3) and proves the theorem. I
7 .1.3 Theorem If («1 , «a, ... , oc,.) is a given coordinate vector and » =
Iot,u., y = I«,v., then the two vectors x and y are connected by the
I I
relation
.Ax= Y, (7)
where A Is defined by Equation (1).
Proof: Since A(u,) = v. for all l = I. 2•.•.• n. it r~uows that
7.1 SIMIUltlTY OP MATRICIIS / 253
Ax = A(I«,u,) = I«,A(u,) = I«,v, = y. I
I I I
7.1.,t neorem Let (c11) be a given n X n matrix. Define linear operators
Rand S by •
R(u,) = Ic,1u, and S(v1) = Ic, 1v., J = 1, 2, ... , n.
I I
Then
SA. =AR. (8)
In other words,
S = ARA-1 , (9)
where A. Is defined by ()).
Proof: We have, for eachj = I, 2, :.. , n,
(SA)(u1) = S(A(u 1)) = S(11.1)
= Ic, 1v, = !c,,A(u,) = A(u.,u,)
I ' I
=
A(R(u1)) =- (AR)(u1) •
Hence SA =-= AR. I
Equations (3), (4), (8), and (9) provide the motivation for the f, llow-
ing definition.
7.1.5 Definition (a) Two n X n matrices Band Care said to be similar
if there exists a nonsingular matrix P such that BP = PC or
B= PCP- 1•
(b) Two linear operators Sand Ron V are said to be similar if then,
exists a nonsingular linear operator A on V such that SA = AR or
S = ARA.-1•
In the light of this definition the answers to Questions 2 and 4 can
now be briefly stated as : The matrices or the linear transformations in
question should be similar •
let U\ now look at a comprehensive numerical example, which illus-
trates Theorems 7. 1.1 to 7.1.4.
Example 7.1 let V = V3 • Consider the ordered bases
F = {(0, I, I), (1, 0, 1), (1, I, 0)}
and G = {(-1, -2, I), (l, 3, -1), (0, -1, 2)}.
(a) Given x = (1, 1, I), verify I esttlt (2).
(b) Given a linear operator T: Ya - V8 defined by T(x1, x,, x 1) =
(x1 + Xa, x 1 + 2x1 + Ka, x1 - x 1), verify result (3).
(c) O,ven the coordinate vector (4, S, -1), verify result (7).
(d) Given the matrix
r
1 -1 07
(c:,> _ I o 1 2 1·.
L-1 -2 -l_J
254 / MOU MATILD[ TBBORY
ro
The matrix for these three systems is
1 1 -1 I 07
II
Lt 1 0
O I -2
-1
3 -1
2_J
I·
Applying the row reduction process to this matrix, we get
rt o o o l 17
I o 1 o
LO O 1·
1 _,
-2 I -t.J
, I·
, o l 17
So (cr.u) = 1 -I f I,
I
L-2 t -f_j
(a) Let [x), be (s11 &a, 11a)r. Then
(1, 1, 1) == s 1(0, 1, 1) + «.(I, 0, 1) + s 1(1, 1, 0).
From this we get s 1 == 1/2 == &a .. &a·
Let [x]a == CP1, ~ ~r. Then
(I, 1, 1) == ~1(-1, -2, 1) + ~(I, 3, -1) + '-CO, -1, 2),
which 1ives ~ - -1, 13. - 0, P, == I. Thus!
. [x), = Cl, l, ur
and (x]a == (-1, 0, l)T.
We have now only to check
•
7.1 SIIIILAlll'lY OP IIATIUCl!S / 255
i7 r-17
-2ct + 313 - r= -4 6 0
at - f3 + 2y = -1 2 -1.
The matrix of this system is
r-1 1 0 0 0 27
I-2 3 -1
L 1 -1 2
-4 6
-1 i -1
oJ'
256 / MOlll! MATRIX TRl!OllY
= .4(4(0, 1. l) + 5(1, 0,
(0, --1, 2) = (I
I) - 1(1, t, O))
..
8 -3)
.
= 4.4(0, I, I)+ SA(l, 0, I) - A(I, I, 0)
= 4(-1, -1, I)+ 5(1, 3, -1) - (0, -1, 2)
== (I, 8, -3) == y.
(d) We have (SA)(O, I. l) = S(A(O. 1, I))
= S(-1. -2, 1) = Ic,111,
i
= 1(-l, -2, 1) - 1(0, -1, 2}
= (-1, -1, -1)
and f.4R)(0, l, 1) = A(R(0, 1, I)) == .4.(Ic0 u,)
I
= A(l(O, 1, 1) - 1(1, 1, 0))
= .4.(0, 1, I) - .4.(1, I, 0)
= (-1, -2, I) - (0, -1, 2) = (-1, -1, -1).
Slmilarly. we could ,check that S.4. and AR coincide on the other basis
elements Ila and Ila also. Hence, S.4. = .4. R.
1 1 1
singular matrix v'l - ~ - \/'3 .
1 1 1
Lv'l ~6 v'3...J
5....... tllatj _:
L-10
-1 -107
10
7 10
-2_J
I· olmilar to\
16
LO
o. -12
0
:1
0 18..J
258 / MORE_MATalX THEORY
1 1 17
v12 v'-6 v3
1 1
via the nonsingul1r matrix
v'2 y6 - v3
2 1
LO v6 - y3_J
6. Prove that o;imilar matrices have the same eigenvalues.
I· g = 10l /(t)g(t)dt.
-
= x2 - i J~ 1 t 2dt - ~x ( 1 t8dt = x 2 - i.
v, = "3 _ ~~ • _! 1_~~•_xx_ x• • (x~ - IJ3\.r .:_ i)
2 2/3 2/5
= X 3 - fx,
Thus, the orthogonal basi:,, 1s
{I, x, x 2 - l, X 3 - fx} .
To get the corresponding orthonormal basis, we divide these by the res-
pective norms and get
I
{ v2' v'3 3v'S 2 Sv'7 3 I }
,/2 x, 2,/2 (x - l), 2y'2 (x - x) '
L u! _J
where u, - (uu, u8,, u,.,)T and u,1 = (,, 11, aia., ••• , u,.,).
r iiI 7
So
L urII _J
ufu,.7
li{u,.
-~_J
1
[x
,a
This can be written in matrix form as
yJ I
h7 rx7
I I I= k .
Lh b_J LY J
In 3-dimensional analytic geometry a similar equation represents what
is called a central quadric. It is of the form
[x Y z] 1: : ;1 1:1
L_g I c_J Lz J
d ·
U=
Lx,._J
and A. is a real symmetric matrix of order n.
In discu\sing a quadric in V,. we shall use geometrical ideas like
tangent and normal. A tangent to a quadric is a line that meets the
quadric at two or more coincident points. If a straight line is perpendi-
cular to the set of all tangents at a point, we say it is a normal at the
point. We assume the natural inner product on Vn and begin with a
fundamental lemma.
7.4.t Lemma /fu1 fa a poi11t on the quadric 11TA.11 = k, then the normal
at "• is in the direction of A11e.
268 / MOllll llAl"RIJ( THEORY
Proof: Figure 7.1 represents
the situation in the 2-dimens.ional
case. Let L = {u0 + tv I t E R}
be an arbitrary tangent, in the
direction of "• at 11o on the surface.
£ clearly passes throuBh the point
u0• Any other point, say u0 + tv, 0
where this line L meets the quadric, ~
will be given by those values of t
that are roots of (u0 + t,)TA
(Uo + tv) = k. This is a scalar 'V
equation, though written in matrix FIGURE 7.1
form. It simplifies to
utAUo + 111T AUg + u{A(tv) + t•vrAv = k •
But u{Au0 = k, since 110 lies on the quadric. So the equation becomes
t(vT .4Uo + u{Av) + t1 11TA11 = 0.
One root of this is t = 0, which corresponds to the point 110 itself. In
order that L may be a tangent to the surface, we require the other root
also to be zero. This gives
vTAUo + u{AY = 0.
But v1 Au1 and u[Av represent the same real scalar. Hence,
vTAu0 = 0 = ufAv.
The first equation means AUg is perpendicular to v, i.e. to 110 + tv, an
arbitrary tangent at u0 • Thus, Au0 is perpendicular to all tangents at llo·
In other words, A.u0 gives the direction of the normal at u0• I
Let u0 be a point on a quadric. Then the vector u0 is said to deter-
mine the direction of a principal axis if it is normal to the quadric at the
point u0• By Lemma 7,4.1, it follows that the principal axes are in the
directions 110 , where A11o = Nl0 for some real scalars >.. Thus, we have
established the following theorem.
7.4.21heorc• In the IJUQdlic uTAu = k the principal axes of the flllldric
are In the direction of the eigenvector, of A.
Example 1.6 Find the directions of the principal axes of the conic
10.x' + 4xy + 7y1 == 100 •
In matrix form this is written as
r10 27 r.x7
[x 11 I I I I == 100 •
L 2 7.J LY.J
r10 27
Jbe eiaenvalues of A = I I are aiven by
L 2 7.J
7.4 APPUCATION TO llEDUCllON OP QUADIU<:11 / 269
110 2- ). 7 -2 ). 1-- O,
which gives >. = 11, 6.
Corresponding to the eigenvalue ). = 11, the eigenvectors are given by
r-1 27 rx7 ro7
I I I I= I 1.
L 2 -4_J Ly_J LO_J
r27
This gives an eigenvector I I.
Ll_J
Again, corresponding to the eigenvc1lue ). = 6, the eigenvectors are
given by
r4 27 rx7
I 11 I= I
L2 l_J Ly_J
r 17
This gives an eigenvector I
I . Therefore, the axes of the conic are
L-2_J
+
along the dtrections (2i j) and (i - 2j).
In Example 7.6 we have determined two eigenvectors corresponding to
two distinct eigenvalues. Naturally they are orthogonal. Let us normal-
ise these vectors and construct the matrix
r-2 17
vS vs = H (say).
] 2
Lv'S y'S_J
Observe that this 1s an orthogonal matrix (since its columns are ortho-
normal). Use this as a similarity transformation on the given symmetric
matrix
r10 27
A= I I-
L 2 7_J
Calculate HTA.H. We obtain the matrix
ru 07
D:; I I,
LO 6_J
Note that this matrix is a diagonal matrix whose diagonal elements are the
r10 27
eigenvalues of A.. Thus, A =I I is orthogonally similar to the
L 2 7_J
r11 07
diagonal matrix D =I I• The passage from ..4 to D is called
LO 6_j
27,0 / MORI MATIIX TIIIIORY
"= II vs I
Lv5
Usu,g this, we get
rx7
I I=
Ly.J
r
lvS 2
1
17
v'S
2
Irx'7
:
Ly'.J
I.
v3 v"1-1
Write this as u == Hv. Then UTAU = k beoomes vTHTAHv = k. Thia
simplifies to
rn 07 rx'7
[x' y'l I I I I == 100.
LO 6.J L.>"-1
i.,. 1b'1 + 6y'1 = I 00
x'' ,••
er c101v11>1 + c10/v6)1 == 1 •
nia ia Ill ellipse with Jcnaths of semiaxes lO/v 11 and 10/../6, Thia
81181,.a shews that
7.4 APPLICATION T9 I.EDUCTION 01' QUADRICS / 271
(i) the eigenvectors of A give the directions of the principal axes of the
conic;
(ii) the eigenvalues >., and ).1 of A specify the lengths of the semiaxes
vk/">.1 and v'kf>.2;
(iii) the orthogonal matrix l/ constructed out cf the normaHsed eigen-
vectors of A is the matrix that reduces A to the diagonal form; and
(iv) the matrix A is orthogonally similar to a d agonal matrix, whose
diagonal elements are the eigcnvatues of A.
We shall now prove that this situauon is true in the n-dimensional
case also, at least for the ca!>e where the matrix has distinct eigenvalues.
The result of the theorem i\ true even for general real 1,ymmctric matrices.
But we lea,e the 1heoretical clevelopmcnt of this case to more advanced
textbooks.
7.4.3 Theorem (Diagonalisation of .l real i,ymmetric matrix) Let A be a
real symmetric rnatrh: with distinct dgenm/ues. Let the normaliud
eigenvectors of A be ll'ritten as column ve1·tors of a matrix H. Then
(a) H is orth<?gonal.
(b) HTAH(~-D) is a diagonal matrix, whose diagonal entries are
the eigenvalues of A.
(c) A is tJrthogonally similar to the d1a!;o11a/ matrix D.
Proof: (a) First, we prove that the eigenvectors of .A corresponding
to distinct eigenvalues are orthogond. Let u1 and 112 be e genvectors
corresponding to distinct eigenvalues A1 and )2, respectively Note that
both are real (see Problem 7, Problem Set 7 2). Then
(A1 - A2)(u1 • U2) = ),1(U1 • Ue) - Az(U1 · U2)
=- (A1U1 U2) - (u1 • AzU2)
rlz) - (u1 · Au2)
=-= (Au 1 •
= (Au1)Tua - uf(Aua)
= uf A 1ua - ufAu2
=0 (beca~se A 1 = A).
Since >.1 -=f:: >.a, it follows that u1 • u1 = 0. lhis coupled ·, ith Theorem
7.3.2 shows that His an orthogonal. matrix.
(b) H = £u1 Ua •.. Un) ,
where "' are the eigenvectors of A written as column vectors.
,.
LuT ...J ,.
L~TA
212 / MOU MATRIX THBORY
r<Aui)T -7
to its principal axes. Note that when this equation is written in full it
takes the form •
7x + 7y - 2z + 20yz - 2<'zx - 2xy = 36 .
2 2 2
-r _; -: -::1
According to the preceding discussion we need to reduce the matrix
L-10 10 --2.J
to the diagonal form. The eigenvalues of A are 6, -12, and 18. The
corresponding eigenvectors are (1, 1,0), (1, -1,2), ard (1, -1, -1).
Hence, the normalised eigenvectors are
(l/y'2, l/v'2, 0), (1/y'6, -1/y'6, 2/\/6), and (l/y3, -l/v'3, -1/y'J).
These three vectors give the following orthogonal matrix
I r1 I 7
v'l y-6 v'3
I l I
H= v1 - '1ti -- v3
Lo
7.4 APPLICATION TO l\BDUCTION OF QUADI.IC:S / 273
Hence, HT AH
r l 1 0 7 r 1 l 17
~i. J2 v'6 yJ
r 7 -1 -107 1 V' 2
==
I
1
\16
l
- vo
l
2
v'6
l
IL-10 -l 7
10 -2
to J'l
I
..;1
I
l
-,J6
1
---
v3
vJ --
2 1
Lv'3 y3_J LO y6 - v'l.J
0
-12
LO
is the required diagonal matrix.
S.o the equation of the quadric referred to its axes is
6X1 - l2yl + 18~ = 36 I
x3 y• z•
i.e. T - T -t T = 1 .
This is a hyperboloi,J of one sheet.
We shall conclude this article with another example where the matrix
does not pos&e<1ci distinct eigenvalues. Though we have not developed the-
necessary theory for this, the following example will show lhat a reduction
to diagonal form is possible even m ciuch a ca!le,
Example 7 8 Reduce the matrix
r 1 2 -27
A= I 2 2 I
L- 2 2 l_J
to the diagonal form, and hence reduce the quadric
x• + y1 + r 1- 4yz - - 4zx 4xy '"" 21 + (2)
to its principal axes.
The eigenvalues of the matrix A are 3, 3, and - 3. Note that in this
case 3 is a repeated eigenvalue. So Theorem 7.4.3, which depends on
the fact that the matrix has distinct eiaenvalues, cannot be applied. How-
ever, an orthogonal matrix can be obtained by looking at the eigenvectors
corres?()nding to 3 and -3.
H..,,,,
I
1
v'2 -
1
y'6 -,/3 1
.
I _1 1 ] I
Lv'2 y6 v'3.J
Hence, HTA.H
0
1
y2 v2
17 ro v2
-v3
17
v'3
r 1 2 -27
- --
v'2
v'3
l
- v'6 v'6
l
IL-2
2 1 2 I 1
v'2
--1
"6
--v'31
2 J.J
L
1
vJ -- I
v'3 0-1
1
Lv'2
1 l
v6
1
:J)_j
rJ 0 07
- I0
LO 0 -3.J
0
3
I·
Thus, the quadric (2) reduces to
3x1 + 3y' - 3z' == 27 ,
~ ,.. z'
i.o. -+----1.
9 9 9
7.4 APPLICATION TO I.EDUCTION Of QUADRICS ( 275
It may be noted here that Equation (2) can be written as
1 1 2 -27 rx7
(x y z] I 2
L- 2 2
2 \ \ y \ - 21 .
1_J Lz _I
Ordinary Linear
Differential Equations
In§ 4.9 we saw that the general solution of the n-th order normal
linear differential equation
Jny ti" ly
Oo(X) iJxn + 01(X) dX"- 1 + ••• t a,.(x)y = g(x) (J)
is of the form y = J'o + yp, where Ye is the solution of the associated
homogeneous equation
dny d•-ly *
a..(x) dA" + a1(x) dx" 1 -t •.. + a,.(x)y = 0, (2)
and }'p is one particular solution <'f (1). y,. is called the complementary
function and is the kernel of the linear differential operator
d• d-1
L = Oo(X) d£• + 01(X) dx"-1 + ... -I On(X) • (3)
Thus. the method of solving Equation (I) involves two steps. namely.
finding Ye and YP• Further, it may be noted that the solution space of
Equation (2) is n-dimensional (cf Theorem 4.9.3).
The solution of the first order normal linear differential equation has
thus been completely discussed in§ 4.9. In this appendix we shall develop
methods of finding Yo and YP for differential equations of arbitrary order.
This gives
(a,,m2 + a 1m + a2)e"''" =-- 0 .
But e""'' is never zero. So
a0 m' + a1m + a2 = 0 . (7)
Equation (7) ts called the indicial equation or the auxiliary equation. Tlh:
values of m for which Equation (7) holds will lead to solutions e"'" of the
differential equation (5). Let the roots of Equation (7) be m1 and m_.
Then e""-x and e~ are two solutions. Their Wronskian is
emlx em.x
W[em,x, em.x ] = = (mJ - m1)e(m1 + 111a1.r •
m1emtx m~em.x
The different roots of the auxiliary equation (7) give rise to tbe follow-
ing three cases.
Case 1 m1 =I= m. In this case the Wronskian is never zero (cf Theorem
4.9.2). So em,.r and e"'-.r are two linearly independent solution~ of
Equation (S). The solution space has dimension 2. Therefore, ,m,.r and
e"'-" form a basis for the solution space. Hence, the gener~ solution is
Yo - C1e'"l" + c,e"'-" ' •
ffl / UIINDII
01# 1 • 1 - .,,,,. la du, ~ it foJlowl from Equdma (7) daat
2-i - -tlJ11e and one ,aJutlon of Equation t,) II ,i - .-.-,
111J - -11J2t,,. Let:,, be tJJe otber solution auch that ~1 and Ya are u. TJlea
I••
W[e•11t , y.J - Ce- •• tbt .., ci1111" ,
Dy Abel"s formula (cf Remark 6.9.4). This, on simplification, pvea
e"'11ty; - mi,at"'"' .. Ce2111,." •
Heaoe, y; -
mi11 . . Ce-.;& •
Solvina th11 fint order linear difl'enntial equation, we obtain a puci-
cular solution Ya - ex,•"' . This aives us Y■ -= '""" as a aecond
10lutlon (taking C == 1) of Equation (5). Obviously, y1 - ~ and
71 - "'""" are LI. (Why ?) Hence, the general solution ii
Ye == C1 ,•.;t
+ C,,c ,l"Jl' •
Thu, we have obtained Yo•
&,,,,,,1, A..l Obtain Yo, the complementary function for the differential
equation
(D' - 2D - l)y - sin x •
Tbe auxiliary equation of the associated homogeneous equation ii
m•-2m-1-o.
Its roots are m1 • 1 +./2, "'• • 1 - ./2. So ,u +vZ).1 and ,<1-v~
an two linearly independent solutiona of the auociated homopaeou1
equation. Honce,
Yo ... c11U+v'2 z + cil-v2)r .
._,,,,,, A..2 Consider the differential equation
(4D' + 12D + 9)y ,... ~ , . .
The auiliary equation of the associated homopneou equation ii
4m' + 12m + 9 - 0.
Its IOOtl are m, =- -3/2 - ,,,_, Al the roots are equal, the two lilleady
indlpmdent solutiom are e-■-'8 and. x.-..,., 1hu, the complementary
funotion it
Y• - Ci,-,. 11 + C1,crlafl •
.,_,,,,. A..J Consider the diff'erential equation
(D' - 4D + 13)y ... ~ + cos 2x •
The auiliarY equation of the auociated homopneous cciuation ii
,,.a - 4ffl + 13 - 0 •
Its fOl)tl are ffla - 2 + 31, m. - 2 - 31. Sm ms rt/: -.. two linearly
lndepmlclnt solutiom are ,,....,,. and r._..,.. nu., tbe ccaplemenlary
fUDOliollla
A I BOIIIDGINIOUI DlnllDmAL tlQuA'IIGNI / 2'19
Tbll ii a OQllpla aolutlon. We •hall now di1e11■ tbe method of lltdDa
1111 IDlutlolll from complex tolutiou.
Al.t LteM I/ u(.x) + 1¥(.x) II • compl,,c '°"'tlon of a homo,...,
,,,,_, tllff,rntlal ,.,.,Ion Ly - 0 wld& nol co,Jficien11, then llotlt
1(.¥) 11114 ,(.x) 11,a,11t1ly 1111#1/y the quatlon Ly ... O.
Proof : Since u(z) + l,(z) is a solution or Ly - O, we have
L(u(z)-+ l,(x)) - 0.
£ it a liaeu operator. Therefore,
L(l(x)) +
IL(,(x)) - O •
Equadna the nal and imapnary parts, we get
L(r,{x)) = 0 and L(,(x)) = 0 •
HIDCI tho lemma. I
Ca.r, J Auxlllrlry lfllllllon with complex roots Jr « + ,, is a root of the
auxiliary equation (with real coefficients), then 11 - , , is another root.
lince m, ::p "'-• two linearly independent solutions are .C•+IJ)1t ud
,<•-•> 11• These are complex solutions.
\
As
,<•+IIS)1t = •" cos Jx + l,u un ll.x
ii a solution of.the d.ift"erential equation (5), it follows from Lemma Al.I
tbat ,■" ms ~% and ,- liD ~ are two real solutiom of Equation (5).
Further. they are linearly indepeadent, because
W[e" cos~. e" sin jlx] - ,,"
Problem Set A 1
l. Solve the following homogeneous linear dift'erential equations :
(a) y'' + 4y' + 4y = 0 (b) y" + 4y == 0
(c) y"' + y' == 0 (d) 3y'' + y' -2y == 0
(e) y'" - y == 0 (f) y" - 2y' - 3y = 0
(I) y" + 2y' - y 0 = (b) y'" - y" - y' + y - 0
(i) (D' + 4D1 -- SD!)y = 0 (j) (D1 - D' - 6D)y = 0
(t) (D' + 4D + 11)1y O = (I) (D' - l)'y == 0
(m) (D' - 3D + 2)1y - 0 (n) {D' + 3D + l)y - 0
(o) (2D' + SD' - 3D)y =- 0 (p) (D' - 4D' + 4D)y - O.
A2 VARIATION OF PAI.AMl1ERS / 281
Ua(X) ==
I O y,,(x)
,<x)fa, y~(x) =_ y,(x),C.x) (6)
IY1\X) Y,(x) a0 W1Y1lx), Y1(x))
Jy;(x) y;(x)
and
(7)
lla(x) = ix g(tlY1rdt ) •
a,W'11lt , y,(t)
Thus, a particular solution of Equation (I) is
_ _ ( ) I" g(t)y,(t)dt ( ) I" g(t)}'1(t)dt
.)'p{X) - 11 X a.W[y,(t), Y,V)l + >'• X a.W[11V). Y1ll))
=- I" Y■lXlY1(t) - Y1(x)y.(t) g(t)dt • (8)
a,W(y,(tJ, Y,V»-
Hcnce, the complete solution of Equation ( l) is
Yi = Ci1,(x) + ca1.(.x) + 1" Y,(.x)y,(t) - y,(x)y.(t) g(t)dl •
a,WIY1(t),y,(t))
Al.I Remark The same· method is applicable to cquationa of arbitrar,
order. The only modification is that in the case of the n-th order
equation we shall need n arbitrary functions 11i(x)1 u,(x), ... , u.(~) in
Jp, since there are n arbitrary constants C1, C., ... , c. in the comple-
mentary function. We shall impose on it (n - I) conditions, one at
each sua:essive differentiation, in addition to the flnt condition that
YP """ "1.)'1 + ""• + ... + "-"• satisfies the nonhomogeneous dift'er-
ential equation.
Instead of carryioy out these details in the aeneraJ case, we ahall
illustrate the method by wortiDa out some suitable eumples.
Eza,npl, .4.6 (D1 - 2D - l)y - sin x.
Tlae complementary function ia (cf Example A.I)
Y• - c,,U+\1'2l¥ + c_.U-\l'Z)ll.
A2 VAllATION OF itAliMBTliS / 21.1
Sowoauumo
1, - 11i(x)e(l+v'l).r + lla(x)e(l-vi)x.
We ~ now _obtain u1(x) and u.~x) by going through the procesa already
described. Since we have done this for the general equation of the second
order, we shall merely plus in the functions in Equation (8). We have
WCY1(x), Y1(x)J = W[e(H v'l)x, e'1-v'l)x) = -2y2e2x .
Hence
yp(x) -= r ----e(I-.J2)x /1+v'2)1
-
(l-v'2)t (I+v'2)x
-----,,,...--=-e=-_ _e____ sin t dt
-2y2ell
== - { - (sin x - co& x) .
A2.2 Remark The reader is advised not to use result (8) for findina ,,.
He should carry out all the steps needed to amvc at Equation (8) as
illustrated again in Example A.7.
= C1 + D1 cos x + C1 sin x + + x1
Note that in this exJmple we needed three conditions, because the
- 4x - 2x sin x .
Problem Set A2
1. Using the method of variation of parameters, find the general solution
for the following differential equations:
(a) (D1 + 4)y = x sin 2x (b) (D1 - 4D + 4)y = xe-•
(c) (D' + 4)y = sec 2x, (d) (308 + D - 2)y == 3,-
0 < < n/4
X
(e) (4D' + l)y = sec9 x/2, (f) (D1 + l)y == sin x
O<x<n
(s) (D1 + 4D + 4)y = xe-•• (h) (6D' + D - 2)y = xr"
(i) (J>I + D)y == x• (j) (D' - D)y == x cos x
(k) y"' - 3y" - y' + 3y == eh (I) 2y"' + y" - y' = e• 11•
2. Let y1(x), y1(x), ... , y ..(x) be n linearly independent solut1\lns of the
associated homogeneous equation of
(a.»" + a 1.l)a-1 + ... + a,.)y B(x) • =
A3 METHOD OF UNDBTEllMiNED COEFFICIENTS / 285
Use the method of variation of parameters to prove that
yp(x) = J.x K(x, t)_g(!l._ dt, where
Oo
J'1(t)
y;(o
.
Ya(t)
'
y,.(t)
Yz(t) y~(t)
K(x, t) = 1
W(Yi(t), Yz(t), ,. , y,.(t)] :
Y1 <n-2>(t) Ya'n-2l(t) y,. In -ll(t)
(e) (.D3 - D1 - D +
l)y = sin x + 1 +
xe •
(0 (D + 9D)y = x
1
A4 OPERATIONAL METHODS
The two method<1 discussed in § A2 and A3, namely, the method of
variation of parameters and the method of undetermined coefficients,
suffice to take care of all problems of finding a particular solution. How•
ever, in this article we add to the reader's armour another tool. which
often comes 1n very useful: Consider the equation
f(D)y = g(x) , (1)
where /(D) = 00 D11 f- a 1/)fl- 1 + .. . + a._1D + a,,, and Do, ai, .•. , a,. are
constants. To obtain one particular solution of this equation. we are
tempted to write formally
1
y = 7(D'f g(x) . (2)
This is only a formal step, which would be useful only if it enables us to
calculate y. First, we have to give a meaning to Equation (2). Second,
we have to relate Equation (2) to Equation (1). We achieve both purposes
simultaneously by defining Equation (2) as follows.
A4.t Definition If /(D) == aJ)" + a1D•-1 + ... + a,., where Do, a1 ... , a,.
are constants, then
1
/{D)g(x)
is defined as a function y for whichf(D)y = g(x).
This definition gives meaning to Equation (2) and also relates it to
Equation (lJ. •
We shill now show that /tb) g(x) can indeed be cal~lated easily in
certain favourable situations.
290 / APPENDIX
The simplest case occ1Jrs ,, hen /(D) = D. Now ~ g(x) is that func•
tion y for which Dy - g(x). This means
dy
dx = g(x).
So y -= J g(x)dx. Smee we are mterested :n only one particular solution,
we shall ignore the arbitrary constant. Thus, we have
1
D g(x) =I g(x)dx . (3)
as follows : Let
I
y =D -=-«g(x) .
E:x:imple A.14 Calculate D -1 - e1z sin ~x. Using Equation (5), we get
- at
== 1 1 (.
2 D _ 1 •un x - cos x
)
= - 21 .
s1nx.
To solve such problems we can also use partial fractions u follows :
1 . 1 .
"JI _ l SID X == (l> _ l)(D + l) SID JC
1 ' j • •
- -(-- - -)SID~
2 D-1 D+l
292 / APPBNDIX
1 I . 1 I .
= -r D- l SJDX- 2 ff+ 1 SIDX
= i; e9 I r' sin x dx - f e _. I el' sin x dx •
This, on simplification, gives
1 . 1 .
g_ 1 smx=- 2 s1nx.
Actually, there should be two arbitrary constants, but we omit them since
we are interested in only a particular integ1al.
In such operator methC1ds the follo~ ing theorem is useful.
A4,4 Theorem If f(D) is the operator polynomial a,,D" + a1D"- 1 + ... + a.
with constant cotfficients, then
(a) f(D)e&- ==- f(a.)e«-.
(b) /(D)(eOg(x)) = ecr/(D + a.)g(x).
(c) /(D') sin a.x =-= f( -a.1) sin a.x.
(d) /(D') cos a.x = /(-a.1) cos a.x.
The proof of this theorem is left to the reader. Using Theorem A4.4
and Definition A4.1, the next theorem immediately follows.
A4.5 Theorem Let f(D) = a0D" + a1 D"-1 + ... + a,. be an operator
polynomial with constant cotfficients. Then
I eP
(a) /(D) eP = /(a.) , /(«) :I 0.
= {- (cos x - ~in x) .
It may be noted that, as a working rule, we can use also the following
method:
I .
YP = 1)1 _ 2D _ 1 sm x
= _1_ 1 _
20 1 sin ~ (by writing -1 for 1)1)
1 1 .
= -T»+l smx
= _.1
I
D
DI - 1
~ sin x = _.1(D - 1) ( -1 - sin x)
I D2 -- 1
= ¼(D - 1) sin x (by Theorem A4.5}
=
-¼(sin x - cos x).
Note that in this method we replaced Dz by -1. In general, to
evaluate ~(~) sin ax or /(~) cos ax, we replace D2 by -a2 ia1 /(D) pro
vided (D1 +a 1) is not a factor of /(D).
1 --1:1 ,s 1 k /ii
(b) Y, = 4D1 + J2D + 9 xe = (2D + 3)11 xe-
1
= e--la/1 -40- X (by Theorem A.4.5)
x'e-k/1
=74•
1
(c) YP == oz - 4D + 13 (r + cos 2x)
l 1
== D' - 4D + 13 xi+ D' - 4D + 13 cos 2x.
Taking the second part, we have
1 l
D1 - 4D + 13 cos 2x = 9-.= 4D cos 2x
2'4 / APPIINDJX
9 +4D
== 81 - 16.0- cos 2.x
D1
1
4D+ 3 .c'= ~• 4D x2
- l 13(1 + -=--)
13
1 D1 - 4D (D' - 4D)1
== 1T (l- 13 + 169 + ...)r
(by a formal long division)
= _1_ (x11 _ !-=- 8x + .E_)
JJ 13 169
1 Sx 6
= 13 Cr + 7T + 16!Y •
Hence,
Y,=E. + 8x + _!_+9cos2x-8sin2x
13 169 2197 145
A4.6 Remark The working of n•· -1D + jj x', though not justifiable
at this Mage, can be very powerful. The reader is Wied to use it
with the full knowledge that be is using only a formal method, the
validity of which is beyond the scope of this appendix.
In any case the reader will have already noted that wherever the
operational method can be used it is really powerful.
(d) YP = jji ~ D (4 sin x + 2r)
-= D ~ D ( 4 sin x) +
1 .& ~D (2r)
-YPI +Y.,.·
_ 23 DJ +
-, 11 ~
1
= 2f (l - DI + D' -- .. )x1
=- 3
i
(r - 6x) = 3 -
2r 4x .
Thus,
.YP = -cos x - 2x sin x --1- 2; - 4x.
It may be noted that the term (-cos x) can be omitted in view of the
fact that cos x is a part of Ye•
Problem Set A4
1. Factorise the operator in each of the following cases and hence find
a particular integral :
(a) (D9 - l)y == e9
(b) (J)I - JD + 2).)' == x + el-
(c) (J)I + 4D9 - SD)y == x + sin x
(d) (SD' + lW - 2D - 3)y = x + xe •.
2. Usina operational methods, determine a particular integral for the
rollowiq differential equations :
(a) (Ji - 3D' + 1D + S)y == ,_..
(b) (D' + 9D)y - cos 3x
(c) (.DI+ 3D +
l)y == r' sin 2x
296 / APPENDIX
(d) (IJI +
4D)y = s:n 2x
(e) (IJI + 3D - 4)y = e"'(3x2 + Sx - 1)
(f) (D - 2)2y = e-u cos x
(g) (2.D8 + SD3 - 3D)y = e-az
(h) (D1 - 4D + 4)2y = x3e2z.
3. Prove Theorem A4.2.
4. Prove Theorem A4.4 .
•
S. Prove Theorem A4.5.
Answers to Problems
Problem Set 1. 7
2. go/: x ~ sin• x, -1t < x < 1t, f o g is not defined since R(g)
- D(/). If D(g) = [- vn,
\/1t], then/ o g is defined and
Jog: x ..._.. sin xS, -v'n < x < y'n.
3. g of: x t-+ v'(l + x 1), -1 < x < I. /o g is not defined since
R(g) ~ D(f). If D(g} = (0, 1), then / o g is defined and
+
IO g : X ~ l x, 0 < X < I.
6. (/ + g)(x) = sin x + x 1 ; ( / - g)(x) = sin x - r;
{/g)(x) = x• sin x.
+
1. (/ g)(x) = 1 + x• + y'x; ( f - g)(x) = 1 x1 - y'x; +
(/g)(x) = {l + .r)\/'x.
10. No.
S. T(x, y) = (3x-2+Y
--, y).
(a) R f- 2S
(b) 2R + ST
(c) S - T -e3
(b) T(X1, Tz, Xa, x,) = (x2 r x,, Xi -f- X3 + Xo -x. + Xa, Xa),
S(x1, x1, x8, x,) -= (xi - 2x1 - 4x3, 2x3 + 2x,, Xi+ Xa - x,,
x~ + ~xJ
(c) Tlxi, x 1 , x 8, x,) = (x2, x 2 + x3 - x,, 2x1 + Xs, Xa + 6x,),
S(x1> x1 , xJ, x,) = (Xi + xd/2 + 3x,/2, -x,/2 - Xa/2 + xJ2,
x1/2 + 3xJ2, 2x1 +· X 1 + xJ2)
(d) T = S.
306 / ANIWDI TO PJtOBUllfl
11
'• '•
- - ----
{a) ST le, - '• - 1,. -2e1 - 8e1 + 35e1 e1 - Sr, + 14e1
(b) RT 3e1 + 2r1 + 3r1 -J2e, + 3,. -le1 + le1 - &.1
- 20e.
(c) RST -20e1 + le1 95e1 + 6e1 38e1 + 6e1 + 5le1
- 29ea + l32e,
(d) R(S + T) 3e1 + 4,1 + 2,1 -1Je1 + 2e1 -2211 + lea
- 19e1 - 35,,
(e) 7" e, - '• + 4e1 -12,, - 3,. -3e1 - 3e1 + 7e1
+ 13e1
co rsT 35e1 + 22e1 -lle1 - 79e1 19e1 - 28e1
- 54,. + 133,, + 37,.
5. S(.x1, ~) = (x1 - x 1, x1 - x 1), T(x10 x1) == (x1, x.).
9. (a), (b), (c:) Range : Y1 ; kernel,: J/1 ; rank: 3 ; nullity : 0.
14. (b) S1S1 is idempotent if S1S1 -= S 1S1 ; S1 + S 1 is idempotent if
s,s. + s.s, == •
(4) 7{x1, x1 , .x.> = (x1, x1 , 0), S(x1 , x1 , xJ = (0, 0, x1 ).
15. {b) Yes; 5 (c) Yea (d) No
A.NSWIU TO ftOILl!MI / 30'1
1 I I 0
13 2 2 5..J
ro 1 o o 7
0 0 2 0
4.
0 0 0 n
LO O O O _J
ro -1 -1 -1 -17 0 1 0 1 07
0 0 0 0 0 0 -1 0 -1
5. (a) 0 O 1 O O (b) 0 O O 1 -1
0 0 0 1 0 0 0 0 1 0
LO 0 0 0 l_J
-1 -37
1 3
1 -1
o -l..J
, ANswns TO PllOBLIIMS 1 309
rt l7
7. I I.
LI -I_J
r1 07 r1 I 17
r1
8. (a) I
LI
17
oJ (b)
ILO l_JI
1
1 (c) I 1. - , o I
LO O l...J
r1
fi ~l
0 0 0 -17
I
(d) 0 (e)
I0 1 1 o II
LO 1 O_J LO 0 l -1.J
0 07 r3 2 0 07
(f)
r:
LO 0
1 0
1 l
l_j
(g)
l 0
LO l
0 -I
0
0
-1
0
0.J
I. (a)
r
I
9 4
Lil 9
137
I
-S..J
(b)
r
I
L--7 I 5.J
I 6 77
,.
rs 67 r-1 47
2. (a) I I (b) I 1.
Ll9 -2..J L O 5.J
r-16 10 -6 87 rt8 1 19 127
3. (a)
I
L -6
-8 -28 -40 -161 (b)
-4 4
2
ILl9 28 38 18 I
-6..J 8 6 5.J
2 -17 -15 -287
(c)
I
L-15
r 2
10 8
-4 -14
-7 -37
14 2
3.J
r:
I
I.
I
-l 1.;-7
4. (a)
114
L6
-26
9 -•:J (b) I · -t i I
Ll l --1..J
(c) rt: -J
-i
ff7
l-. .
L\' u -t.J
-S 07
1-•-1 -2
6.
L-1 -4
0
-1 -17 -33
-9
_:J 147
s s S -4
7. 0 10 44 -6 .
0 0 -14 -2
L O 0 0 36
ANSWERS TO Pll08LEMS / 311
1 1 0 07
0 0 0 0
0 1 0 1
9.
0 1 0 0
0 0 1 1
ILO 0 0 0 _J
r1 1 l 0 0 07
I
10.
Io
I
0 0 l 0 0
.
I
I
0 0 0 0 0
I
LO 0 0 0 l_J
11. dim V .,_- 5. The matrices with respect to the ordered bai,is {sin x,
co:. :x, sin x ~os x, sin1 x, cos• x} are
rJ -2 0 0 07
I
I 2 3 0 0 0
(a) 0 0 3 4 -4
0 0 -2 3 0
I
LO 0 2 0 3 _J
- I I 0 0 07
-1 l 0 0 0
(b) 0 0 -8 -2 2
I 0 0 1 -2 6
I -2
L 0 0 -1 6
(b) BA=
r_:
I _; 9
-3
67
-8 21
L 11
I
r22 56
(c) AB= 3 5
L3 9
r3 11
(d) AB - I
L7 IS 24 28
r 3 2 4
24 11 16 19
(c) BA -
-1 -1 -3 1
L -6 O -2 -4
ro 07 r-2
(f) AB == I
I ' BA. = I
LO O_J L-21
ro 17 r1 -27 r o 11~ 1J
4 • (a) L1 oJ (b) Lo 1J (c) I I
r-: :
L J/~ -«/~• _J
17
(d) j l/« O 7 (c) -1 I.
r:
L-~/«y J/y_J ,. I
L-¼ O ,_j
13_j
r124 59 967
"'-I 4S 26
L 66 53 87_j
371
ANSWIRS TO PllOBLIIMS / 313
(b) .4• = j7
L6 7-J
47 , ..t• = j 19
L27
187 ,A'= 173 567
l9_j L84 73_j
r 10 5 107 r 35 60 107
~ 75 2455 2505j ,•
rlX -2~ 7 JIX ~ 7 r« 07
6. (a) I I (b) I I (c) I I
L~ IX+ ~_J L~ X - [i_J LO ~_j
r IX ~7
(d) I I , where «and ~ are arbitrary scalars.
L-~ IX_J
7. (a)
r 1 -i7
I I (b)
r
I
. \· -l7
I (c) Solution does not exist.
L1/- f _j L-io LJ
rA11L1 0 0 7 r!A-1>-1 0 0 7
0 >-21Lt 0 0 (,l,d>-2 0
8. (a) (b)
LO 0 >.,.(L,._J LO 0 1-'n>.,._J
r-11>-, 0 0 7
I
(c) O i,i0.,-ft0,i=l,2,.,.,n.
I
L O O 1/>.,._j
9. (a) Nilpotent; 4 (b) Nilpotent; 4
(c) Nilpotent; 3 (d) Not nilpotent.
r 1 - IX1 7 IX P7 r r1 07
14. IL~ IX p-
-IX _J
I. I ·-=-~
L
at• -
at
I· ± I
_J LO
I,
1_J
r-1 07 .
±I I , where IX and ~ are arbitrary scalars with P'I: 0.
L O l_J
r -5 so7 r-21 607 r 22· -247
16. (a) I I (b) I I (c) I I
L-" ~_J L-~ ~_J L-~ %_j
314 / ANSWt.u 10 PllOBLl!MS
D, D, D, Dr, D, ». Da I »1
r 2 0 07 •
ro
2. (a) i I
LI
-27
I {b)
l.J
•I L
-4
5
2
-4
0
l_l
I
r-1 I --17
I
(c) i ! 3 -1 s ·I
I
r: - :
L 3 --1 3.J
r 7 -13 8 0 07
-3 9 0 :~J. Cc) l
-1 o I
4 -4 8 -12 0 0 o 2 I
I_ 2 10 -8 6 LO 0 2 -2.J
r- ~17
I
ex 11 --ex~
2. (a)
L3
I
I
I
2 -2J
1 2 (b) 1
'
L-1
0
2.J
(c) I -3 0
L s 4.J
I
-• I I
I
r 4 -2 77
I r-3 -4 s-1
(d) -1i I -3 (c) I
L 8 4 - 4_J
I
L I -3 8.J
r-2 --3 S7 r 3 --27 r 2 2 27
(f) I I (g) I I (h) I I
L 7 0 4.J L-2 1,0.J L2 --1 4.J
r2 3 17
(i) I I.
LO -I S.J
+;
r1 2-i7
s. (a)
r t - I -1+17
I
L3 + I 2 + i.J
I (b) ·12 + 31 I-4ij
Ll-i 3 + 2ii
r2 - 3; 1 + 2; 3 - 417
(c) I I
L2 +; 2 + 21 2 - '.J
316 / ANSWllts TO fl.Olt.md
r 1 2+1 1-17
(d) I 1 +I -i 2 - i I·
L-1 - 21 I 3 - 2/__J
6. ax• + by + 1 er
+ 2hxy + 2/yz 2gzx. +
9. (b) A diagonal matrix.
Problem Set 5. 7
1. (a)
r1 o
Io
I
LO O
1 ; -: -:1
0 0 O_j
1 0 0 07
0 1 0
0 1 0 0
1 1 0
(c) [ : - : ] (d) 0 0 1 0
0 0 1
0 0 0 l_j
r1 o o o 1 r1 o 07
0 0 -27
Io
(e)
0 1 0 0
0 0 1 0
0
1 0
I
1j
-3
(s)
I
I
1
O O
o
t,
LO O O l_j LO O O__J
(b)
rt
0 1 0
0 0
-1.l
1\·
(i)
r1
0 1 0
0 0 27
-I
0 0 1 -ii. 0 0 1 1
LOO O 0 _J LO o ·o O_J
2. (a) 2 (b) 2 (c) 3 (d) 4 (e) 4 (0 3 <s> 3
(h) 3 (i) 3.
3. Same as the answer to Problem 2.
r 1 -s 1 57
4. (d) Nonsingular; t
4
-3
1 -3
1 4 -1 -IJ
L 1 16 -6 -9
ANSWIIU TO PROBLEMS / 317
r 3 3 -3 37
7 3 -S 4
(e) Nonsingular; l
-1 0 2 -1
L
5 3 -4 2-1
5. (a) X = 14/19, y = 3/19, Z = )3/19
(b) x = 1/4, y = 1/2, z = -S/4
(c) x = 1/9,y = -2/9,z = -1/3
(d) x = 5/7, y = 1, z = 3/7
(e) X = 8/5, y = 5, Z = 9/5.
7. k1 = 2, k1 == 4, k3 == 6, k, = 8, k, = 9, k, = 10.
-7
-1
3
-4J
10
l
r-3 I
(c) Nonsingular;
¼ I
1 _:1
10 -2 -2.J
318 / ANS'WBU TO PI.OBLltNS
r 16 7 -1 -77
-6 I 4 -1
(e) Nonsingular; i,
1 -6 5 6
L-7 6 -5 23.J
X1 Xa Xt Xi
where ,.,1t fa>a, ... , c.,• arc n n-th roots of unity.
ANSWBI.S TO PI.OILIMS I 3\9
1x'-+ Y1
I
+ z2 X y z
1xf + y~ + zf X1 Y1 Z1 :1
7. (b) Ix!+ Y! t- z1 Xa y, z, 1 = 0.
Ix:+ y: -1- z: Xa Ya Za 1
I
Ix! 1- v: + z, x, Y, z, 1
8. x = 0, -a - b - c.
10. (a}
r-o
I
Ll
-27
I
l_J
(b)
r
I
I
L
-4
2
s
0 07
-4
2
:-1
(c) r-: :J
L 3
-l
-1
1 -17
r-1 13 - 8 --97 -2 2 0 07
I
3 -9 0 -3 0 -2 2 0
(d) (e)
-4 4 -8 12 0 0 0 -4
L--2 -10 8 -6_J L 0 0 -4 4_j
Problem Set 6. 6
5. 0.
Problem Set 6. 7
1. (a) x-= -1,y = 4 (b) x = 31/11,Y-= -5/11
(c) x - - 2, y ==-- 29/2, z = -8 (d) x = -1/2, y = 5/2, z == 2
(e) x-= 2S/2, y = -17/2, z = -1/2.
2. (a) X1 = 5/16, X2 = 1/16, Xa = 3/16, x, = 1/8
(b} x1 = 2x2 = 2, X,1 = x, = 0
(h) X1 =- 7/6, x 2 = 41/6, x3 = 1/3, x, = -S9/6.
3. Inverser. of the coefficient matrices :
r,-1 -1 17
r 2 -17 r 4 37
(a) I
L-5
I (b) ·l'r I
3.J
I
L-1 2.J
(c)
I 4
L-2 -3
s -1
2~
I
320 / ANSWDI TO P&OBLIMI
L
0 0
O O 0
r1 0
I
:J 0 07
Io ' 0 0
.
3. >,.• - 1;
lo 0
_, 0
I
LO 0 0 -LJ
Problem Set 6. 10
2. 2(b) {( 1
v6' - v6' v6'
_1 2 o>, c-L. _ ~-•
v3o • v3o v3o
~-.o>.
( 12 0 _ 6 s > <-2=. 0• __ ! --, _ --!-n
v12os • • J20s' -v26s • v41 • v4°f v4I
1 2 1
2(d) {(0, 0, 1, 0), (v'6' v6' 0, v'6)'
17 46 O 109 )}
<- vt4286' - vI4286' •v142s6 ·
- - S ...s + x - l-
3. 3(a) { \/'15 _. _ v' 151 _
4 ""- 4 4 ""- 4.
S I 1
- 2y'2 x' + v'2 " + 2v72}
3(c) { '\1'3 5'\1'7 x' _ 3'\1'7 X 3'\1'35 (xi + x•),
y2 X, 2 v'2 2 v'2 ' 2 v'94
v'S (47 - 42(.xi + .x'))}.
\111562
322 / ANSWERS TO PROBLEMS
5. (a) 0 (b) 0.
8. If v, is the first vector in the sp.1n of the earlier ones, then the
corresponding vector in the proce<;s of orthogonalisation will be zero.
Problem Set A 1
1. (a) )' c,e-M + C3xe 2 ' (b) y =- C1 cm, 2.'- -t- C1 sin 2x
(l') J' - c, + cl co, .\ + Ca 'iin A {d) J' = • C1e2-' 11 -j- c,.e-..
(e) Y "- l\<'-' -t- e "' 1 (C2 coi, (\, 3x,2) l C, sin (y3t/2))
(0 y - C1e3r -t CJe ., (g) y C,e- (l+v2 1ll' -1- C1e- (I- v2)x
(h) -. C1e
y C2 x~
1 + CJe., +
(i) C1 -t ('4x -l C 3t.. I- C4e
l' -- 61
Problem Set A3
l. (a} y -= C 1 e3,. + Cle .. - e2J /3
(b) y =- C1 coi, 2x 1 cl '!tin 2:c (.t/4) cos 2t
X 4
{c) y ---= (C, co,; x -1-- CJ ~in .:\)e' -t - er - - ~
~ l~
l 2 I
+-y..\-\+:?
(d) y - C1e 1 - 1 I , '21 '" -I c~/ -I \ 2~ t- r.mhx
(e) y =-- C 1e'" I- Cixe' I- CJe-" l- (2 cor. x -t 2 -;in " l- he-..
+ ..\ie ' I- 8)/8
(f) y = C1 t- cl CO'J 3t -f CJ l>in 3x +- "t 2/l8
(g) y -.., (( 1 c1>s (y'3x/.?) l- C, .,m (\'h/2))e--· 4 !- C 1e"' + C,e-
x~ +h - 5 t- ftHh t· 2x2 ) coi,h ~ - (6x + xl)
':.inh x}
(h) y -- C 1 I- Ca cos X + C3 !.Ill X - X cos X
27. t-1 3 .
- 169 smx 13 xcos 2
x t 26 X SIO X
Problem Set A4