Notes On Linear Equations
Notes On Linear Equations
In this chapter we explore the important connections between differential equations and
linear algebra. We seek a common language and set of tools that will help unify our discussion of
these two subjects. The key players in our study are vectors and matrices.
Most of the remaining text concerns objects that engage in repetitive or periodic motion.
A typical example is the simple harmonic model (SHM), given by the second-order equation
We may describe any process as linear if there is an operator L that obeys the following
properties for any two quantities
(1)
(2)
The derivative is a linear operator for any two functions ƒ and g since
Many embarrassing algebra mistakes occur when students don’t recognize that other operations
are not linear, which explains why (x + y)² x² + y².
has a solution point or vector x 0 ú6, given by (x1, x2,..., xn). If b = 0, we have the homogeneous
equation L(x) = 0 and the superposition properties
which are direct results of the linear operator L. Consider the equation
The linear nth-order differential equation L(y) = ƒ(t) is formed when the operator
is applied to some function y(t). We had the first-order equation L(y) = yN + p(t)y = q(t). The
2.1 Second-Order Equations
is also a solution, since the linear properties this operator permit us to confirm that
When we impose the initial conditions y(a) = b0 and yN(a) = b1 on the equation
The values of the c’s will determine the nature of any possible solutions y(x).
The second-order equation yO + p(x)yN + q(x)y = 0 has a unique solution for any
conditions y(a) = b0 and yN(a) = b1 only if the functions {y1,y2} are linearly independent and not
multiples of each other, where y1/y2 = k (constant). The functions sin x and cos x are independent
since (sin x)/(cos x) = tan x does not have a constant value. On the other hand the functions
are both solutions of the equation yO- 2yN+ y = 0. The general solution
is any function of the form (graphs shown)
and the initial conditions y(0) = 3 and yN(0) = 1 produce the system
Notice that we can reduce the order if we let v = yN and obtain the first-
order equation vN = -2v, which by inspection has a solution
where the new constant B = -C/2. The initial values y(0) = A + B and
yN(0) = -2B form a simple linear system, where each solution (A,B)
2.1 Second-Order Equations
clearly has one trivial solution (0,0), but it may also have other non-trivial solutions depending on
the values of the coefficients.
PS2.1
2.2 Linear Systems
where the superposition principle is applied to n solutions {y1,y2,...,yn}. There is a unique solution
determined by n initial conditions that form a linear system
For example, we can verify that each one of the following functions
The conditions y(0) = 0, yN(0) = 5 and yO(0) = -39 correspond to the unique solution (graph)
A linear system can have any number of unknowns and equations, for example
The second version is the augmented matrix [A b] with a coefficient matrix A. The size of any
matrix is specified by the number of rows × columns, thus the matrices
The solution of a linear system is found by reducing the augmented matrix using
elementary row operations: multiply any row by some constant and add the result to another
row. For example, consider this simple case:
Here we have used these operations to create row equivalent matrices. This method is called
Gaussian elimination, where we seek to write the matrix in a row-echelon form E such as
Notice that the leading entry or pivot in each row is 1 with only zeros below.
2.2 Linear Systems
Example 1 Solve:
Our strategy is to simplify the augmented matrix using the “leading 1"
or pivot in the upper left corner to create zeros below it and express the
matrix in row-echelon form:
Since there are 5 unknowns, but only 3 equations, there are many
solutions, which are described by two free parameters (choices), say
An extension of this method is Gauss-Jordan elimination where the goal is to write the
matrix in reduced row-echelon form (RREF), where there are zeros above and below each pivot.
The RREF immediately gives the solution as (5, -2, 3), without the
need for any back substitution.
# For practice, let (1,4,-2)
Observe in this example that a unique solution exists and the coefficient matrix is row
equivalent to the identity matrix
We know that any homogeneous system [A 0] is always consistent since it has a trivial or
zero solution. Consider the system
which is the homogeneous version of the system solved in the previous example. Since the
coefficient matrix is row equivalent to the identity matrix, there is only one solution (0,0,0).
Example 4 Solve:
2.2 Linear Systems
This system also has more unknowns (4) than equations (3), so we
expect many solutions, which is confirmed by row reduction:
The reduced system has two equations and four unknowns, so the
solution is described by two parameters, say x3 = s and x4 = t, so
Each pair of values {s,t} yields a particular solution (x1, x2, x3, x4).
Example 5 We had the third-order equation y(3) + 3yO + 4yN + 12y = 0 with a
general solution
The initial conditions y(0) = 0, yN(0) = 5 and yO(0) = -39 produce the
following linear system and its RREF version:
which is the transpose of a column (proper) vector. The two basic vector operations, scalar
multiplication and addition each follow expected patterns. For example, consider the vectors
We have more unknowns (4) than equations (2), so this system has
many solutions given by (say) the parametric form
If two matrices have the same size, their sum is found by adding the corresponding row
(or column) vectors.
(A)
(B)
The third important vector operation involves the row and column vectors
The rules of matrix algebra were originally developed to study linear systems, where we
recognize, for example, the following equivalent forms:
2.3 Matrix Operations
In the second version we “factor out” the common vector x, so that we may write the system as
the simple matrix equation Ax = b, where we apply the row × column inner product rule.
The matrix product AB is formed by computing the all inner products of the rows of
matrix A with the columns of matrix B. The matrices must therefore be compatible: each row
vector of A(m × p) must have the same number of components as each column vector of B(p ×
n), so that each entry or location of AB(m × n) is the product
(A)
(B)
Some of the typical algebra rules also work with matrices, while others do not. We just
observed the failure of the commutative rule, AB BA, while the associative law ABC = (AB)C
= A(BC) and distributive law A(B + C) = AB + AC still hold.
Example 4 Consider
The cancellation rule failed since there is no division of matrices. Instead we require the
existence of the inverse matrix defined by the matrix products
the identity matrix. For example, we have the pair of inverse matrices given by
2.3 Matrix Operations
The transpose of any matrix is formed if we interchange rows and columns, so that
A matrix is symmetric across the main diagonal if the matrix and its transpose are equal A = AT .
The product of a matrix and its transpose is always symmetric.
Matrices have other uses besides their role in working with systems of linear equations.
The linear system Ax = b can also be interpreted as the mapping of one vector x into another b,
an operation T: ú6÷ ú6 given by the linear function T(x) = Ax. One important mapping that we
use later in the course is the result of the rotation matrix in ú², given by
PS2.3
2.4 Determinants
provided that the inverse matrix exists. We may verify for the simplest case, the 2 × 2 matrix
which exists only if the determinant *A* / ad - bc 0. We could also solve the corresponding
linear system
Observe that we have the ratio of two determinants: the bottom number is the determinant of the
coefficient matrix, while the top determinant is found by replacing the first column of A with the
vector b. The second solution will follow a similar pattern:
We have found an equivalent unique solution of the system Ax = b, called Cramer’s rule with
the same restriction *A* 0.
Example 1 Solve:
We know that the second-order homogeneous equation yO + p(x)yN + q(x)y = 0 has only
two independent solutions {y1,y2} that are chosen to satisfy the linear system
For any constants c1 and c2 we can apply Cramer’s rule to show that these functions should satisfy
the Wronskian condition
(A)
(B)
There is only one non-zero minor to compute down the first column
of this triangular matrix, the product along the main diagonal.
Example 3 Solve:
Our work with the linear system Ax = b is nicely summarized by the following list of
equivalent statements, where A is any n × n matrix:
(1) A is row-equivalent to In;
(2) Ax = b has a unique solution for every vector b;
(3) Ax = 0 has only the trivial (zero) solution;
(4) det(A) 0;
(5) A is invertible.
Linear Independence
Recall that two functions y1 and y2 are dependent if they are scalar multiples of each other,
so y1 = ky2 or y2 = ky1. The same definition applies for two vectors, such as v1 = (2,3) and v2 =
(4,6) since 2v1 = v2, which is the linear combination
so v3 = -4v1 + 3v2. There are many other non-trivial combinations. So, vectors are linearly
independent only if c1 = c2 = ... = ck = 0.
The condition
Row operations also reveal that *A* 0, so the system has only the
trivial solution c1 = c2 = c3 = 0, hence these vectors are linearly
independent.
# For practice, try (1,1,1), (-1,0,1) and (-2,1,4). {dependent}
The number of vectors will determine whether they are independent. It is easy to show
that if the number of vectors k from ú6 are linearly dependent if k > n since the system Ac = 0 has
more unknowns (n+1) than equations (n). So, adding one more vector (say) w 0 ú6 implies that
for some non-zero coefficient ci. If c 0 and the n vectors {v1,v2,...,vn} are independent, then w
is a linear combination of this spanning set that forms a basis in ú6. The standard basis in ú³ is
the set of special vectors
is a linear combination by the usual rules of matrix algebra. They form the linear system Ic = 0 so
c = 0, confirming that they are linearly independent. Any set of n independent vectors is a basis in
ú6. In the previous example the vectors (1,1,1), (1,2,-1) and (1,3,2) also form a basis for ú³.
If we treat the set of functions {ƒ1,ƒ2,...,ƒk} as one-dimensional vectors, then the definition
of linear independence can be extended so that
is spanned by the fundamental set {1, t, t²}. We know from basic algebra
is only valid for all t when each coefficient is zero. The set {1,t,t²} is the standard basis for P2
and forms a general solution of the third-order equation yNNN = 0.
This matrix has a determinant *A* = -4, so this system will have a trivial
solution. For any system Ac = b there is a unique These functions are
not only independent, but will also span every polynomial P2 and form
another basis.
# For practice, try {basis for P3}
We may extend the Wronskian determinant to test for the linear independence given a set
of n differentiable functions, where
of the equation
PS2.4