Differential Forms PDF
Differential Forms PDF
Differential Forms
Alexia E. Schulz
and
William C. Schulz
2 Mathematical Theory 73
2.1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . 74
2.2 Manifolds Mappings and Pullbacks . . . . . . . . . . . . . . . . . 74
2.3 Proofs of Stokes Theorem . . . . . . . . . . . . . . . . . . . . . . 74
2.4 Proofs of the Converse of the Poincare Lemma . . . . . . . . . . 74
2.5 Permutations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
2.6 The operator Φ . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
iii
iv CONTENTS
Chapter 1
1
2 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
1.1 INTRODUCTION
We want to give a real world example here but remember it must be inex-
act since real world objects are granular (atomic) in constitution, so can only
approximate the perfect mathematical objects. Some people prefer to eat the
closed peach (with fuzzy skin), some people prefer the open peach (fuzzy skin
removed, peach◦ ) and the boundary of the peach, ∂peach, is the fuzzy skin.
Perhaps this will help you remember. Deeper knowledge of these matters can
be found in the wonderful book [2] and also [4].
4 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
For functions we will use a slightly augmented variant of the physics conven-
tion. When we write f : S → R we mean a function whose input is a point p ∈ S
and whose output is a real number. This is theoretically useful but not suitable
for calculation. When we wish to calculate, we need to introduce coordinates. If
we are dealing with the upper half of the unit sphere (set of points in R3 whose
distance from the origin is exactly one and for which z ≥ 0) then we might write
f (x, y) if we choose to represent points in the x, y coordinate system. Notice,
and this is an important point, that the coordinate x takes as input p ∈ S and
outputs a real number, it’s x coordinate. Hence the coordinates x and y are
functions just like f . If S is the upper half of the unit sphere in R3 then x
and y are not really good coordinates. It would be be better to use longitude
and colatitude for my coordinates and then we would write f (φ, θ). 1 Note
use of the same letter f no matter what the coordinate system, because the f
represents a quantity in physics, whereas in math it represents a functional re-
lationship and we would not use the same letter for different coordinates. Note
also that f (.5, .5) is ambiguous in physics unless you have already specified the
coordinate system. Not so with the math convention.
Finally, we will almost always use the letters f, g, h for functions on A, M, S, K.
Mostly these will occur in coordinate form, for example f (x, y, z) for a function
on M .
and you recall the Green, divergence and Stokes theorems, which I list here for
convenience:
Green’s theorem
Z Z
∂g ∂f
f (x, y) dx + g(x, y) dy = − dxdy
∂A A ∂x ∂y
and Z Z
v · dℓ = curl v · dS
∂S S
There are some conventions on integrals that we will mention now. In for-
mer
R R Rtimes when integrating over a three dimensional object we would write
M div v dV This is now completely antiquated, and we will not do it.
On the other hand, there is a convention that when integrating around curves
or surfaces that have no boundary we put a small circle on the integral, so that
we write I Z
v · dS for v · dS
∂M ∂M
Since this is favored by the physics community we will mostly use it. Notice that
if a geometric object is the boundary of something, then it itself has no boundary,
and so we will use the circled integral almost exclusively with boundaries.
For our purposes we will define a differential form to be an object like
which we find as integrands in the written out forms of the Green, divergence
and Stokes theorem above. If ω is a sum of such objects it turns out that the
three theorems collapse to one mighty theorem, called the generalized Stokes
theorem, which is valid for all dimensions:
I Z
ω= dω
∂S S
If you are concerned about what differential forms ARE, the answer is a
little tricky and we am going to put it off for the moment. Later we will
discuss the surprisingly dull answer to this question. Incidentally, the difficulty
in explaining what they really are is one reason they have not become more
common in elementary textbooks despite their extreme usefulness.
Just to give a tiny hint of the geometrical interpretation of differential forms.
A two form measures the density of lines of force of a field, as introduced by
James Faraday a century and a half ago. For more on this subject see [1] or [8].
We will discuss it a bit more when we have more equipment.
2 bad pun
1.5. THE ALGEBRA OF DIFFERENTIAL FORMS 7
form df :
∂f ∂f ∂f
df = dx + dy + dz
∂x ∂y ∂z
Recalling that, like f , the coordinate x is also a function on R3 the previous
formula writes the differential of f in terms of the differentials of the three special
functions x, y, z. So we note that there is not much difference between df and
dx; they are the same kind of object. All objects of this type (differentials of
functions) are collected together in the set
Λ1 (R3 )
f dx ∧ g dy = f g dx ∧ dy ∈ Λ2 (R3 )
Note that functions f commute with dx: f dx = dx f ; see below. Linear combi-
nations of such objects are called 2-forms. And of course there are 3-forms
f dx ∧ dy ∧ dz ∈ Λ3 (R3 )
To complete the system we will place the functions in the basement of the
building: f ∈ Λ0 (R3 ). It is customary to omit the wedge when multiplying by
a function; we write
f dx ∧ dy for f ∧ dx ∧ dy
There is no significance to this; it is just convention.
The algebra of these objects is just like ordinary algebra except for the
changes caused by the rule
dg ∧ df = − df ∧ dg (anti-commutativity)
for the 1-forms df and dg. An algebra satisfying this rule is called an exte-
rior or Grassmann algebra. This algebra was invented by Hermann Guenther
Grassmann about 1840 in an attempt to find an algebra mirroring elementary
geometry. It is sufficient to postulate this rule for the coordinate differentials
only,
dy ∧ dx = − dx ∧ dy etc.
since the general rule will follow by linearity.
Thus the exterior algebra is not commutative. Our rule is often called anti-
commutative and is the simplest generalization of commutative, but it has many
consequences which seem strange to the beginner. For example, if we substitute
f for g in the rule dg ∧ df = − df ∧ dg we get
df ∧ df = − df ∧ df
8 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
so moving the right side to the left side by the usual algebraic processes which
are all valid here we get
df ∧ df + df ∧ df = 0
2 df ∧ df = 0
df ∧ df = 0
Thus the product of a one form with itself is 0, which is very important. Let’s
look at another example
(f dx + g dy) ∧ (f dx + g dy) = f f dx ∧ dx + f g dx ∧ dy + gf dy ∧ dx + gg dy ∧ dy
= 0 + f g(dx ∧ dy + dy ∧ dx) + 0
= fg 0 = 0
as promised.
We also see from this that there are no four forms in 3-space, since if we
multiply four of the objects dx, dy, dz together there will be a repetition which
will kill the form:
dx ∧ dy ∧ dz ∧ dx = − dx ∧ dy ∧ dx ∧ dz = dx ∧ dx ∧ dy ∧ dz = 0 ∧ dy ∧ dz = 0
(dy ∧ dz) ∧ dx = dy ∧ dz ∧ dx
= − dy ∧ dx ∧ dz
= dx ∧ dy ∧ dz
and if you look at how this special case works you will see why the general case
works.
Note that nothing prevents us from mixing things up as in
2 dx + 3 dx ∧ dy
but such things do not occur in practise. Forms where each term has the same
number of differentials (forms of the same degree) are called homogeneous, and
we almost always use homogeneous expressions.
1.6. THE OPERATOR D 9
1. d is a linear operator
∂f ∂f
df = du1 + . . . + dun
∂u1 ∂un
d(ω ∧ η) = dω ∧ η + (−1)j ω ∧ dη
we also used rule 4, ddx = 0, in the last equation. This derives the practical
rule in a special case, and the general case (see problems) will be the same.
This is the practical rule for d and the one you will use for almost everything
you do, so learn it well.
Now let ω = f dx and η = g dy. Then we have
1.7 Orientation
At this point we must deal with one of the less pleasant aspects of elementary
geometry which is orientation. An orientation, to speak loosely, is a sense of
twist in a space. For example, in R2 our standard sense of twist is counter-
clockwise; we measure angles from th x axis in the direction of the y axis. If we
reverse either axis, we get the opposite twist. If the y axis went down instead
of up then we would measure angles in a clockwise direction and R2 would have
the opposite orientation. If we reverse both axes then the sense of twist returns
to counterclockwise.
In R3 the standard orientation is given by the following rule: if you place
the fingers of your RIGHT hand so that the fingers curl from the x to the y
coordinate axes then your thumb points in the direction of the z axis. This
is called the right hand rule. It has become standard to use this in Calculus
books only since the 1940’s, and the opposite convention is still found in Italian
books, so it is wise to check. To appreciate the subtlety of this concept, think
of trying to communicate it to inhabitants of a planet in another galaxy. Since
nature is almost symmetric in terms of left and right, the only way we know to
clarify this is certain non-symmetric aspects of beta decay. This is referred to
in physics as parity. Hopefully parity properties remain the same from galaxy
to galaxy.
This idea of orientation, though subtle, is strongly coupled with differential
forms and is the reason for the anticommutativity. It also is a prime source of
mistakes, and great care must be taken to keep things in correct order. In R2
(and it’s subspaces) the correct order, which expresses the orientation properly,
is dx ∧ dy. Thus dy ∧ dx is in incorrect order as indicated by the minus sign in
dy ∧ dx = − dx ∧ dy
it is critical that the direction around the boundary of the left integral be coun-
terclockwise. If it is taken clockwise then the two sides of the equation will have
opposite signs. This is again due to the orientation which is built into R2 but
which we seldom notice explicitly. There are similar worries in the use of the
divergence theorem and Stokes theorem.
In applications, the principal place where orientation occurs in Rn is in n-
forms and (n − 1)-forms. We will first tell you the the general formula and
then give you practical methods to make orientation (relatively) easy to deal
with. Let the variables used in Rn be u1 , u2 , . . . , un . (We have switched from
x1 , x2 , . . . , xn to u1 , u2 , . . . , un to emphasize that the variables can be general;
not necessarily rectangular or orthogonal.) The use of superscripts to number
the variables is to conform to tensor analysis standards and we don’t need to go
into the reasons for it here; just do it! And remember u3 is the third variable,
1.7. ORIENTATION 11
not the cube of u. If we choose an order for the variables, which we did by
numbering them, this chooses one of the two orientations. Then
because
du2 ∧ du1 ∧ . . . ∧ dun = − du1 ∧ du2 ∧ . . . ∧ dun
As you can see with a little practice, interchanging any two of the dui reverses
the sign and changes correct to incorrect order or incorrect to correct order. If
you are familiar with permutations, odd permutations of du1 ∧ du2 ∧ . . . ∧ dun
give incorrect order and even permutations give correct order.
That part is easy. The tricky part is the (n − 1)-forms. Here the correct
order is (with dui missing from the list)
which is correct because the dui must hop over the n−1 elements du1 , . . . , dui−1
in order to get back into correct order and each hop contributes a minus sign.
So much for theory. In R3 correct order is
dx ∧ dy ∧ dz correct order
dxdydzdxdydz
and noting that the order of a wedge of two differentials is correct if it occurs
in this list, for example dz ∧ dx is correct but dx ∧ dz is incorrect since dxdz
does not occur in the list. Other incorrects are dy ∧ dx and dz ∧ dy. The use of
differential forms in R3 relies critically on writing things with correct order.
3 Cyclic order is a 3-space concept and does not generalize to n-space at all well.
12 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
u × (v × w) = (u · w)v − (u · v)w
It is often said that the cross product cannot be generalized to higher dimensions
but this is not true; what is true is that the analog of the cross product in n
dimensions involves not two but n − 1 vectors. Thus the elementary geometric
applications of the cross product can often be reproduced, but the physical
applications not so much, which is the reason that for relativity (4 dimensions
and space-time rather than just space) we must fall back on Tensor Analysis.
In 3 dimensions there are many formal analogies between differential forms
and vector analysis. That is, differential forms will act like vectors in many ways.
It is a little difficult to find good mathematical reasons for these analogies and
we will not concern ourselves here with what these reasons might be, although
we will return to the matter later. The practical consideration is that things
will work well if we keep in mind the following rules. For 1-forms we have the
correspondences
dx ↔ î dy ↔ ĵ dz ↔ k̂
For 2-forms we have the correspondences
dy ∧ dz ↔ î dz ∧ dx ↔ ĵ dx ∧ dy ↔ k̂
Note that in the second case we have been careful to place the 2-forms in proper
order. This is critical. If you screw this up you will get the wrong sign. Watch
dz ∧ dx term particularly carefully.
With these correspondences kept in mind, we can easily derive many for-
mulas of vector analysis in simple ways. Many things which appear different in
vector analysis can be treated in a unified way with differential forms.
d(f dx + g dy + h dz) = df ∧ dx + dg ∧ dy + dh ∧ dz
∂f ∂f ∂f
= dx + dy + dz ∧ dx
∂x ∂y ∂z
∂g ∂g ∂g
+ dx + dy + dz ∧ dy
∂x ∂y ∂z
∂h ∂h ∂h
+ dx + dy + dz ∧ dz
∂x ∂y ∂z
∂f ∂f
= dy ∧ dx + dz ∧ dx
∂y ∂z
∂g ∂g
+ dx ∧ dy + dz ∧ dy
∂x ∂z
∂h ∂h
+ dx ∧ dz + dy ∧ dz
∂x ∂y
∂h ∂g
= − dy ∧ dz
∂y ∂z
∂f ∂h
+ − dz ∧ dx
∂z ∂x
∂g ∂f
+ − dx ∧ dy
∂x ∂y
Now recall that if v = f i + gj + hk then
i j k
∂
curl v = ∂x ∂y ∂ ∂
∂z
f g h
∂h ∂g ∂f ∂h ∂g ∂f
= − i+ − j+ − k
∂y ∂z ∂z ∂x ∂x ∂y
Thus we see we have counterfeited the curl on vectors provided we keep the
2-forms in correct order (so the signs come out right) and we use the correspon-
dence
dy ∧ dz ↔ i dz ∧ dx ↔ j dx ∧ dy ↔ k
Finally, we want to see the differential forms turn up the divergence. Let
once again v = f i + gj + hk and using the above correspondence let us apply d
to the form f dy ∧ dz + g dz ∧ dx + h dx ∧ dy. Then we get
d(f dy ∧ dz + g dz ∧ dx + h dx ∧ dy) = df ∧ dy ∧ dz + dg ∧ dz ∧ dx + dh ∧ dx ∧ dy
14 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
∂f ∂g
= dx ∧ dy ∧ dz + dy ∧ dz ∧ dx
∂x ∂y
∂h
+ dz ∧ dx ∧ dy
∂z
∂f ∂g ∂h
= + + dx ∧ dy ∧ dz
∂x ∂y ∂z
Def A region K is simply connected if and only if any simple closed curve in K
can be shrunk to a point in K.
A few more examples, this time in R3 . A ball (solid sphere) is simply con-
nected, as is also a ball with the center removed. However, a ball around the
origin with the z axis removed is not simply connected and neither is the whole
of R3 with the z axis removed. The unit sphere (points in R3 at unit distance
from origin) is simply connected. (Note the contrast to the unit circle in R2
which is not simply connected.
These kinds of considerations are called topological, and there is a wonderful
branch of mathematics which studies such things called topology. Due to the
enormous amount of material needed in an engineering curriculum, it is not
customary for topology to be included, but if you wish to learn more we recom-
mend the books [2] and [4], which are specifically written with the engineer and
physicist in mind.
Now that we know some topology, we can state the theorem:
Theorem (Converse of Poincaré Lemma) Let K be a simply connected region
and ω ∈ Λr (K) and dω = 0. Then there is an α ∈ Λr−1 (K) such that dα = ω.
It should be mentioned that the α is highly non-unique; there are many α’s
that will work.
In the problems we will see examples where the region is not simply con-
nected and, though dω = 0 there is no α for which dα = ω.√ Practically speaking,
this usually takes the form of α being multi-valued, like x, and thus not being
a proper function. In this case, we can often manage to get some use out of the
multi-valued function provided we restrict our attention to a simply connected
subregion. The vector potential of the magnetic field of an infinite vertical
current carrying wire is the paradigm example and we will look at it in the
problems.
Next we will derive two popular theorems of vector analysis (in R3 ) from
the converse of the Poincaré lemma.
For the first we again let v = f i + gj + hk and we suppose that curl v = 0.
We then form the corresponding differential 1-form ω = f dx+g dy +h dz. From
the results of the last section, we see that the condition curl v = 0 translates
into the condition dω = 0. Assuming the region K ⊆ R3 is simply connected
16 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
∂k ∂k ∂k
dk = dx + dy + dz = f dx + g dy + h dz
∂x ∂y ∂z
∂k ∂k ∂k
i+ j+ k = f i + gj + hk
∂x ∂y ∂z
grad f = v
curl grad k = 0
div curl v = 0
These are both consequences of dd = 0 which you probably already figured out.
Also available are results about path independent line integrals which we will
look at in the problems.
1.11 Boundaries
Before we can deal with Stokes theorem we must talk a bit about boundaries
and their orientation, and also how to compute integrals of differential forms.
The boundary of a area or surface is simply its edge. We will be satisfied
with an intuitive understanding of this and not go into the topological details.
The symbol of the boundary of a region or surface K is ∂K. For example the
boundary of the unit disk in R2 is the unit circle, and the boundary of the upper
half of the unit sphere in R3 is the unit circle in the x, y-plane. The boundary of
the unit ball in R3 is the unit sphere. It is common for there to be no boundary;
the boundary of the unit sphere in R3 is empty, which we write ∂S(0, 1) = ∅
where ∅ is the empty set.
For Stokes theorem it is important that the boundary ∂K be oriented cor-
rectly relative to K itself. We will do this with examples, but the methods are
1.11. BOUNDARIES 17
is the unit tangent vector. There are two directions one may go along a curve,
and the parameterization must be chosen so that n̂, T̂ has the same orientation
as x, y. This amounts to, first, the angle from n̂ to T̂ is a positive right angle,
and second that as t increases we trace the boundary counterclockwise around
A. So this comes down to just going around A counterclockwise.
For M the situation is more difficult to see since the boundary will be two
dimensional. Suppose the parameters are u and v. Let the surface be given
by r(u, v). Then when we increase u and hold v fixed, a curve will be traced
∂r
out in certain direction and it will have tangent vector ru = ∂u pointing in
that direction. Similarly for rv . Then the requirement is that n̂, ru , rv have the
same orientation as x, y, z. If the orientation is opposite, reverse the order of
the parameters u and v.
On a practical note, one usually finds the n̂ for ∂M by forming
ru × rv
n̂ =
|ru × rv |
so that one really has only to check whether this n̂ is pointing in or out of M
and if it is pointing in change the order of u and v. Also, one does not actually
have to calculate ru × rv , only figure out its direction.
As an example, take the upper half of the unit sphere and let us use for
coordinates longitude φ and colatitude θ. A tangent vector to a phi curve goes
to the fight (looking at the sphere from outside) and a tangent vector to the
θ curve goes down. Crossing these gives a vector pointing in so we have the
wrong order: the order should be θ, φ not φ, θ. Any 2-forms that occur should
thus be presented as f (θ, φ) dθ ∧ dφ. This will be critically important when we
form integrals.
Also as a practical matter, for ∂M ⊂ R3 one often uses two of x, y, z as
parameters. If this is the case, the proper orders are dy ∧ dz, dz ∧ dx, dx ∧ dy
as one can determine by using the above rules.
The last case we are going to discuss is S, the surface in R3 . If the surface has
no boundary then the situation is just that of ∂M above (practically speaking)
and we have already covered the situation; an example would be the unit sphere.
18 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
However, if the surface has a boundary, then there is no obvious way to decide
which way the normal should point. Mathematically one just has to choose
oneself either a direction for the normal n̂ or, equivalently, and order for the
coordinates u, v. Physically there may be an obvious way to make the choice.
Make sure that your n̂ is consistent with your choice of u, v so that n̂, ru , rv is
oriented like x, y, z. Once this is settled it only remains to orient the boundary
∂S. Since ∂S is one dimensional, it is only a question of choosing which way to
go around it. Here is the rule.
Walk around the boundary of the surface S with your body pointing the same
way as n̂ and your LEFT hand pointing toward the surface. The direction your
are walking is the correct orientation of the boundary.
This is the orientation you must use for the correct application of Stokes
theorem. If you mess it up the sign will be wrong.
Of course in physics the sign often comes out wrong, and one just reverses
it at the end. Still, it is reassuring when it comes out right without human
intervention.
To calculate the integral we recall that in a multiple integral the order of the
differentials doesn’t matter but for forms the order matters, as the sign will
change if two differentials are interchanged. The calculation of an integral is
reduced to three steps.
1. Rearrange the integral so that the differentials are in correct order.
2. Remove the wedges to get an ordinary multiple integral
3. Calculate the multiple integral by the appropriate iterated integral
This should become clear from an example. We wish to calculate the integral
Z
x dy ∧ dx
A
2
where A is the unit disk in R . Step 1 is to realize that the differentials are not
in the right order. We fix.
Z Z
x dy ∧ dx = − x dx ∧ dy
A A
1.13. VARIABLE CHANGES 19
Now the differentials under the integral are in the correct order. Step 2 is to
throw out the wedges
Z Z
− x dx ∧ dy = − x dxdy
A A
where the last integral is an ordinary double integral from your advanced cal-
culus course. Step 3 is then to go over to the an iterated integral
Z Z 1 Z √
1−x2
− x dxdy = − √ x dxdy
A −1 − 1−x2
Step 4, which we did not mention above, is then to plug it into your fancy cal-
culator and get the answer. Another method is to calculate the double integral
by switching to polar coordinates
Z Z 2π Z 1
− x dxdy = − (r cos θ)r drdθ
A 0 0
x = x(u, v)
y = y(u, v)
which is called the Jacobian matrix4 of the variable change. Thus we can write
the above equation, with det(·) signifying the determinant, as
Z Z
∂(x, y)
f (x, y) dx ∧ dy = f˜(u, v) det( )du ∧ dv
A A ∂(u, v)
Z ∂x ∂x
= f˜(u, v) ∂u
∂y
∂v
∂y du ∧ dv
A ∂u ∂v
theorem. We will work it out for surfaces in R3 but the techniques applied will
work just as well for (n − 1)-surfaces in Rn . This is one of the places in the
theory which, while not difficult, is a little tricky.
While it may not be possible to cover the surface with a single coordinate
patch, it is always possible to break up the integral into integrals over each
patch and add them. Hence we only need to work out the formula for a single
coordinate patch.
The standard thing to have to work out for surface integrals is
Z
P dy ∧ dz + Q dz ∧ dx + R dy ∧ dz
S
and so the bit of oriented area can be obtained from the cross product
∂r ∂r
∆S = × ∆u∆v
∂u ∂v
Now remembering that
r(u, v) = x(u, v) î + y(u, v) ĵ + z(u, v) k̂
we have
∂r ∂x ∂y ∂z
= î + ĵ + k̂
∂u ∂u ∂u ∂u
∂r ∂x ∂y ∂z
= î + ĵ + k̂
∂v ∂v
∂v ∂v
∂r ∂r ∂y ∂z ∂z ∂y ∂z ∂x ∂x ∂z
× = − î + − ĵ
∂u ∂v ∂u ∂v ∂u ∂v ∂u ∂v ∂u ∂v
∂x ∂y ∂y ∂x
+ − k̂
∂u ∂v ∂u ∂v
∂(y, z) ∂(z, x) ∂(x, y)
= î + ĵ + k̂
∂(u, v) ∂(u, v) ∂(u, v)
22 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
Finally we have
Z
P dy ∧ dz + P dz ∧ dx + P dx ∧ dy
S
Z
∂(y, z) ∂(z, x) ∂(x, y)
= P du ∧ dv + Q du ∧ dv + R du ∧ dv
S ∂(u, v) ∂(u, v) ∂(u, v)
Z
∂(y, z) ∂(z, x) ∂(x, y)
= (P i + Q j + R k) · î + ĵ + k̂ du ∧ dv
S ∂(u, v) ∂(u, v) ∂(u, v)
Z
∂(y, z) ∂(z, x) ∂(x, y)
= (P i + Q j + R k) · î + ĵ + k̂ dudv
∂(u, v) ∂(u, v) ∂(u, v)
ZS
= v · dS
S
we have
∂r ∂r ∂r ∂r
dS = × dudv = n̂ ×
∂u ∂v ∂u ∂v dudv = n̂ dS
we can write Z Z
v · dS = v · n̂ dS
S S
It would be wise to note that the integrals we have been working with in
this section are often called flux integrals. In contrast, there is another kind of
1.15. THE GENERALIZED STOKES THEOREM 23
surface integral which has little to do with what we are doing. This kind of
integral is not vectorial in nature. They look like
Z
f (x, y, z) dS
S
A problem that utilized this kind of integral would be finding the center of mass
of a hemispherical shell. We have developed the necessary tools to evaluate
these integrals, although for us this is a side issue and we only mention it since
students should be aware of the two distinct kinds of surface integral.
To evaluate this kind of integral, choose parameter u, v for the surface, (or
a portion of the surface,) and then use the above above formula for dS to get
Z
f (x, y, z) dS
S
s 2 2 2
Z
∂(y, z) ∂(z, x) ∂(x, y)
= f (x(u, v), y(u, v), z(u, v)) + + dudv
∂(u, v) ∂(u, v) ∂(u, v)
These kinds of integrals are very sensitive to the choice of parameters, and an
integral that might be extremely difficult with parameters x, y (so that z =
z(x, y)) might be quite easy with, say, θ, φ as parameters.
Proof Let ω = P (x, y) dx + Q(x, y) dy. Applying the general Stokes’ theorem
we have
I Z
P (x, y) dx + Q(x, y) dy = ω
∂A
I∂A
= dω
ZK
= d(P (x, y) dx + Q(x, y) dy)
K
Z
∂P ∂Q
= dy ∧ dx + dx ∧ dy
K ∂y ∂x
Z
∂Q ∂P
= − dx ∧ dy
K ∂x ∂y
Z
∂Q ∂P
= − dxdy
K ∂x ∂y
where we used dy ∧ dx = − dx ∧ dy and in the last line we used the rule from
converting from oriented integrals to ordinary double integrals.
Notice that it was not necessary for us to remember where the sign goes in
Green’s theorem; our methodology automatically puts it in the correct place.
We will do Stokes’ theorem next since it looks so similar in many ways to
Green’s theorem.
Stokes’ Theorem Let S be a surface in R3 with definite choice of normal n̂
and correctly oriented boundary ∂S. Let v = P î + Q ĵ + R k̂ Then
I
P dx + Q dy + R dz =
∂S
Z Z Z
∂R ∂Q ∂P ∂R ∂Q ∂P
= − dydz + − dzdx + − dxdy
S ∂y ∂z S ∂z ∂x S ∂x ∂y
or in vector notation I Z
v · dl = curl v · n̂ dS
∂S S
Recall that the boundary of S is oriented so that when you walk around the
boundary with your body in the direction of n̂ and your left hand reaching
in toward the surface then you are walking in the direction of the orientation.
(This can be reformulated in terms of fingers and thumb but we find this less
confusing.)
Proof: Let ω be the 1-form
ω = P dx + Q dy + R dz
Then applying Stokes’ theorem
I
P dx + Q dy + R dz =
∂S
1.15. THE GENERALIZED STOKES THEOREM 25
I
= ω
∂S
Z
= dω
ZS
= d(P dx + Q dy + R dz)
ZS Z Z
∂R ∂Q ∂P ∂R ∂Q ∂P
= − dy ∧ dz + − dz ∧ dx + − dx ∧ dy
S ∂y ∂z S ∂z ∂x S ∂x ∂y
Z Z Z
∂R ∂Q ∂P ∂R ∂Q ∂P
= − dydz + − dzdx + − dxdy
∂y ∂z ∂z ∂x ∂x ∂y
ZS S S
= curl v · dS
S
where we have used the rules governing the transition from oriented to unori-
ented (ordinary) integrals. Recall that in unoriented integrals the order of the
differentials does not matter, but we have left them in cyclic form as a matter
of good practise. The transition to vector form was discussed in the previous
section.
Now to the divergence theorem. This is quite easy.
The Divergence theorem (Gauss’s theorem) Let M be a region of R3 and
∂M the surface which is its correctly oriented boundary. Then
I
P dydz + Qdzdx + P dxdy =
∂M
Z
∂P ∂Q ∂R
= + + dxdydz
M ∂x ∂y ∂z
or in vector notation
I Z
v · n̂ dS = div vdxdydz
∂M M
The other two terms are handled similarly. Notice that we have proved slightly
more than the theorem states; we have proved the analog of the theorem for
each term. This can occasionally be useful.
r = r(u1 , . . . , un )
∂r
ei =
∂ui
gij = (ei , ej ) = ei · ej
(g ij ) = (gij )−1
1.16. CURVILINEAR COORDINATES I: PRELIMINARY FORMULAS 27
The sum sign is often omitted when the same letter appears both up and down;
since invented by Einstein it is probably a good idea, but you do need to keep
it in mind. This simplifies the notation considerably.
Next we need a function where you input the vector v and it outputs the ith
coordinate v i . You have known this function for years, but never knew what it
was. It is dui . Thus
dui (v) = v i
This is indeed what dui really is, but with the current organization of the math
curriculum it is not convenient to explain this at the entry level. However, now
you know! Incidentally, it is possible to make this consistent with the idea that
dui is a little bit of ui . It works like this. Nearby points are connected by
vectors which are very short. If v is such a short vector, then v i = dui (v) is
also small. So if we abbreviate dui (v) by dui it will behave as a small number,
sort of. It’s surprising that this works as well as it does.
Since dui : V → R (and is linear) it is a linear functional and thus dui ∈ V ∗ ,
where V ∗ is the vector space of linear functionals on V. A wedge product of
three dui , for example du1 ∧ du3 ∧ du5 would then be in Λ3 (V ∗ ).
The du1 , . . . , dun are the dual basis to e1 , . . . , en and form a basis for V ∗ .
Now we discuss Φ : V → V ∗ and Φ−1 : V ∗ → V . We have
Here ω is an arbitrary element of V ∗ and the λi are its coordinates in the basis
du1 , . . . , dun .
Recall that the gij give an inner (dot) product for V and in Chapter two we
show that this may be ”lifted” to V ∗ . If
λ = λi dui λ∈ V∗ λi ∈ R
µ = µi dui µ∈V∗ µi ∈ R
then
(λ, µ) = g ij λi µj
28 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
just like
(u, v) = gij ui v j
Let {k1 , . . . , kn−r } = {1, 2, . . . , n} − {i1, . . . , ir } so that we can form the permu-
tation
1 . . . r r + 1 . . . n
i1 . . . ir k1 . . . kn−r
of {1, 2, . . . , n}. The sign of this permutation is (−1)s where s is the number
of interchanges necessary to rearrange the bottom line into the top line (or vice
versa). The vertical line is just for the convenience of the reader; it divides the
first part of the permutation from the second. Then
hk1 · · · hkn−r k1
∗(dui1 ∧ dui2 ∧ . . . ∧ duir ) = (−1)s du ∧ duk2 ∧ . . . ∧ dukn−r
hi1 · · · hir
which looks a lot like curl. Now if we are starting with a vector v = v 1 e1 +
v 2 e2 + v 3 e3 we can get to ω by using Φ. Indeed, using the formulas from the
last section,
Then dω gives us the above 2-form. However, we cannot use Φ−1 to descend
again to vectors because Φ−1 eats only 1-forms. However ∗ gets us from 2-forms
to 1-forms, and we have
3 1 2
∂v ∂v 2 ∂v ∂v 3 ∂v ∂v 1
∗dω = − dx + − dy + − dz
∂y ∂z ∂z ∂x ∂x ∂y
3 1 2
∂v ∂v 2 ∂v ∂v 3 ∂v ∂v 1
Φ−1 (∗dω) = − ~e1 + − ~e2 + − ~e3
∂y ∂z ∂z ∂x ∂x ∂y
Φ−1 (∗dΦ(v)) = curl v
curl = Φ−1 ◦ ∗ ◦ d ◦ Φ
Since the coordinate system is orthogonal, we have gij = 0 for i 6= j and gii = h2i .
Thus we have
v = v 1 e1 + v 1 e2 + v 1 e3
= v 1 h1 ê1 + v 2 h2 ê2 + v 3 h3 ê3
= ṽ 1 ê1 + ṽ 2 ê2 + ṽ 3 ê3
where
ṽ1 = h1 v 1 ṽ2 = h2 v 2 ṽ3 = h3 v 3
In terms of the ṽ i and the êi the formula for curl becomes
1 h ∂(h23 v 3 ) ∂(h22 v 2 )
curl v = − h1 ê1 + etc.
h1 h2 h3 ∂u2 ∂u3
1 h ∂(h3 ṽ 3 ) ∂(h2 ṽ 2 )
= − h1 ê1 + etc.
h1 h2 h3 ∂u2 ∂u3
h1 ê1 h2 ê2 h3 ê3
1 ∂
= ∂ ∂
h1 h2 h3 ∂u 1 ∂u 2 ∂u 3
1 1 1
h1 ṽ h1 ṽ h1 ṽ
As an example, let us write the formula for Spherical Coordinates using the
material for Spherical coordinates at the end of the previous section. Recall
that h1 = 1, h2 = ρ, h3 = ρ sin θ. Then
ê1 ρ ê2 ρ sin θ ê3
1 ∂
curl v = ∂ ∂
ρ2 sin θ ∂ρ1 ∂θ
2
∂φ
3
ṽ ρṽ ρ sin θṽ
While the curl will only work in 3 dimensions, we can get formulas for grad
and div in n dimensions with no extra effort at all. This is important since
we might need either of them in 2 as well as 3 dimensions, and maybe even 4
dimensions for relativity. So we shall do them all at once.
There is no difficulty at all with grad. It inputs a function and outputs a
vector so the immediate candidate for grad is Φ−1 ◦ d. This obviously works in
rectangular coordinates since then Φ(ei ) = dui . and thus
∂f
grad f = Φ−1 ( dui ) (sum on i understood)
∂ui
Xn
∂f
= e
i i
i=1
∂u
The last is our old friend the advanced calculus gradient. Notice that we could
not use the summation convention here since both the i’s in the last term count
as low indices. (Recall that high index in a denominator counts as low.) Watch
carefully in the next developments to see why this has happened.
Next recall that for general (not necessarily orthogonal) coordinates we have
Φ(ei ) = gij duj Φ−1 (dui ) = g ij ej
and so
grad f Φ−1 (df )
=
∂f
= Φ−1 ( i dui )
∂u
∂f ij
= g ej
∂ui
which is nice and simple. For orthogonal coordinates we have
ij 0 if i 6= j
g = 1
= 1 if i = j
gii h2i
and thus, and thus with natural basis vectors ei and physical (normalized) basis
vectors êi = h1i ei we have
Xn
1 ∂f
grad f = 2 ∂ui ei
h
i=1 i
Xn
1 ∂f
= êi
h
i=1 i
∂ui
1.17. CURVILINEAR COORDINATES II: THE CALCULATIONS 33
The second formula, with physical basis vectors êi , is what one generally sees
in reference books.
Our next task is the divergence. This is harder, since we must use the ∗
operator, but we can break it down into easy steps. The first observation we
make is important for many purposes. Let
n
X
ω= (−1)i−1 v i du1 ∧ · · · ∧ dui−1 ∧ dui+1 ∧ · · · ∧ dun
i=1
where the hat on the dûi indicates that it is NOT THERE. This is just notation;
it tends to make everything look cleaner, but you must keep your wits about
you and watch for it. Now when we find dω it comes out very nice:
n
X ∂v i j
dω = (−1)i−1 du ∧ du1 ∧ · · · ∧ dûi ∧ · · · ∧ dun
i,j=1
∂uj
Now note that if j 6= i then the differential contain a repetition and thus give
0. Hence only the terms with j = i need be retained and we have
n
X ∂v i j
dω = (−1)i−1 du ∧ du1 ∧ · · · ∧ dûi ∧ · · · ∧ dun
i,j=1
∂uj
X n
∂v i i
= (−1)i−1 du ∧ du1 ∧ · · · ∧ dûi ∧ · · · ∧ dun
i=1
∂ui
Xn
∂v i 1
= i
du ∧ · · · ∧ dui ∧ · · · ∧ dun
i=1
∂u
n
!
X ∂v i
= i
du1 ∧ · · · ∧ dun
i=1
∂u
where in one step the dui had to hop over i − 1 other differentials to find its
proper slot. You will notice has this has the look of divergence about it.
To utilize this formula we must get from our vector to an (n − 1)-form. We
can get from the vector to a 1-form using Φ and then to an (n − 1)-form using
∗. From there, d will take us to an n-form and then another ∗ will get us back
to a 0-form, or scalar, so ∗ ◦ d ◦ ∗ ◦ Φ will get us from a vector to a scalar as we
wish for divergence. Thus ∗ ◦ d ◦ ∗ ◦ Φ is a good candidate for divergence. Let’s
try it in rectangular coordinates. In rectangular coordinates the natural basis
e1 , . . . en is just i, j, k when n = 3
v = v i ei
34 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
Φ(v) = v i dui
Xn
∗Φ(v) = v i (−1)i−1 du1 ∧ · · · ∧ dûi ∧ · · · ∧ dun
i=1
n
!
X ∂v i
d ∗ Φ(v) = du1 ∧ · · · ∧ dun
i=1
∂ui
Xn
∂v i
∗ d ∗ Φ(v) = = div v
i=1
∂ui
which shows us that indeed the invariant form of div is ∗ ◦ d ◦ ∗ ◦ Φ. If you are
worried about the expressions for ∗, they follow immediately from the general
expressions below.
Before we launch into the final calculation for div, let’s recall the formulas
for the ∗ operator. For ω = ωi dui we have
n
X √
∗ω = (−1)k−1 g ik g ωi du1 ∧ . . . ∧ dûk ∧ . . . ∧ dun
i,k=1
where, as before, the hat on dûj indicates the term is NOT THERE. This is
derived in the Chapter 2. Also derived there are
√
Ω0 = g du1 ∧ . . . ∧ dun
∗ Ω0 = 1
∗ 1 = Ω0
div v = ∗ d ∗ Φ(v)
= ∗ d ∗ (gij v i duj )
n
!
X
k−1 jk √ i 1 k n
= ∗d (−1) g g (gij v )du ∧ . . . ∧ dû ∧ . . . ∧ du
k=1
Since g jk gij = δik , the only non-zero terms are when k = i, to the above simplifies
to
n
!
X
i−1 √ i 1 i n
div v = ∗ d (−1) g v du ∧ . . . ∧ dû ∧ . . . ∧ du
i=1
1.17. CURVILINEAR COORDINATES II: THE CALCULATIONS 35
n
X √ i
∂( g v )
= ∗ (−1)i−1 j
duj ∧ du1 ∧ . . . ∧ dûi ∧ . . . ∧ dun
i,j=1
∂u
n √ !
X ∂( g v i ) 1 i n
= ∗ ∧ du ∧ . . . ∧ du ∧ . . . ∧ du
i=1
∂ui
n √ !
1 X ∂( g v i ) √
= ∗ √ g du1 ∧ . . . ∧ dun
g i=1 ∂ui
n √ !
1 X ∂( g v i )
= √
g i=1 ∂ui
and this is the formula for div when v is expressed in natural coordinates v =
v i ei . For orthogonal coordinates we have as usual ei = hi êi and
n
X
v = v i ei
i=1
n
X
= v i hi êi
i=1
n
X
= ṽ i êi
i=1
we have
Xn
1 ∂
div v = (h1 h2 · · · hn v i )
h1 h2 · · · hn i=1 ∂ui
Xn
1 ∂
= (h1 · · · hi−1 hi+1 · · · hn ṽ i )
h1 h2 · · · hn i=1 ∂ui
Lastly we want to give a formula for the Laplacian. In physics the formula
for the Laplacian in rectangular coordinates is
Xn
∂f
△f =
p
i
i=1
∂u
The normal notation for the Laplacian is △f . However, in recent years in math-
ematics there has been a big tendency to use the symbol △f for the negative
36 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
of the sum above. There are very good reasons for this5 . So to make sure there
will be no confusion, we will use the symbol △ p f for the Laplacian as custom-
ary in physics. It will be awhile before we need to deal with the mathematical
Laplacian.
To find the Laplacian of functions in any coordinate system is now absurdly
easy, because we have
△p f = div grad f
Thus we need only recall the two previously derived formulas for grad and div
∂f ij
grad f = g ej
∂ui
n √
1 X ∂( g v j )
div v = √
g j=1 ∂uj
∂f ij
vj = g
∂ui
into the formula for div v to get
n
1 X ∂ √ ij ∂f
△
p f = √ g g
g i,j=1 ∂uj ∂ui
Xn
1 ∂ ∂f
△f =
p h1 · · · hi−1 hi+1 · · · hn i
h1 · · · hn i=1 ∂ui ∂u
The critical thing here is that these functions have many derivatives. Many
means, for physics purposes, at least three continuous derivatives. The manifold
is then called a C3 -manifold. In mathematics it is customary to avoid the boring
details and require the functions to have infinitely many continuous derivatives
so it is C∞ -manifold. This seemingly boring requirement is the key to the whole
subject; it is what makes the manifold smooth like a ball rather than unsmooth
like a cube. It takes some contemplation to see this but it’s true.
One more comment; in classical tensor analysis they are continually talking
about coordinate changes as if they made some big difference. They do; the
talk of coordinate changes in tensor analysis is just how that subject deals with
the material in the previous paragraph6.
6 This is not a well known fact even among the tensor users. Reveal it only to trusted
friends
38 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
There is a little more to the story, because, after all, we have to put the
manifold inside Rn and this requires a little more to be put into the definition.
The requirement is easy to state. Since the manifold is inside Rn and p will
thus have coordinates x1 , . . . , xn . These will be functions of p’s disk coordinates
u1 , . . . , um ; that is
xi = xi (u1 , . . . , um ) i = 1, . . . , n
have rank m. Since m ≤ n, m is the largest rank it could have, so we often say
the matrix has maximal rank. The reason this is important is that we could set
up a coordinate system around any p for Rn that used u1 , . . . , um for the first
m coordinates and then n − m other coordinates to finish the job, but we do
not need to pursue this.
We should also mention that while in theory we have set the system up to
use disks in Rm to make coordinates, you can actually do it with any regions
you like rather than disks. We only did it with disks to make it precise and
visual. Things are rather looser in practise.
The test to see that this condition is satisfied is to check that all the m × m
minors of the matrix are non-zero. It is now time for an example. We will
use (as usual) the sphere in 3-space, which illustrates the concepts nicely. The
coordinates will be θ, φ where φ is the longitude and θ is the angle off the z-axis7.
The R3 coordinates of the point with coordinates θ, φ are
x = sin θ cos φ
y = sin θ sin φ
z = cos θ
R = (sin θ cos φ, sin θ sin φ, cos θ)
There is a lot that can be learned from this example. Notice first that the minors
are all 0 at the poles where θ = 0 or θ = π. This is reflected also in e2 which is
0 at the poles and thus not much of a basis vector for the tangent space. Note
that at the poles φ is not defined either. Note that in n̂ = (e1 × e2 )/|e1 × e2 |
both numerator and denominator are 0, so n̂ is not well defined though it seems
to be after the 0’s disappear because of the cancellation of sin θ.
The proper response to this is to select a second coordinate system, perhaps
one with east and west poles, to cover the two offending points with proper
coordinates. Needless to say, this is hardly ever done in this and many other
cases, because with a little sympathetic treatment it is possible to get around
the troubles at the poles. However, it is really important to notice such bad
points and to make sure whatever you are doing makes sense at such bad points.
Often it does, sometimes it doesn’t. Beware.
Now on to other things. Notice in the example how easy it was to get the
basis of tangent vectors e1 , e2 , one for each coordinate, at each point of the
manifold. The vector space spanned by e1 , e2 is called the Tangent plane, and
is denoted by Tp (S) (where S is the sphere). In the more general case of an
embedded m-dimensional manifold S in n-dimensional space, we will have m
tangent vectors at each point, e1 , . . . , em , and they will be a basis for an m-
dimensional space Tp (S) called the Tangent Space to the manifold S at p. The
condition on the minors guarantees that the ei will be linearly independent.
1 m
Consider an m-dimensional manifold S with coordinates Pu i , . . . , u and basis
i
vectors ei = ∂R/∂u . Now let v ∈ Tp (S). Then v = v ei and we need a
function whose input is v and whose output is v i . This function is dui . Although
the idea is new to you, this is what dui actually is. Thus the defining equation
is
dui (v) = v i
(We have seen this before; I’m just reminding you in this new context.) If
you are familiar with the the space of linear functionals of a vector space, then
du1 , . . . , dum form a basis for this space of linear functionals (called the dual
space and denoted by Tp∗ (S).
40 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
When you follow this path for awhile, you become aware of what a small role
the embedding space Rn plays in the theory. Riemann also wondered about this,
and eventually showed how to get around it. Let us ask what the important
contributions of the embedding space Rn are. There are two most important
contributions. First, the embedding space provides an inner product for the
Tp (S) which live inside it. Second, it provides us with a convenient way (visible
in the sphere example) of finding tangent vectors and thus the tangent space
Tp (S). The first is a critical idea; the second just requires technical trickery. For
the first, Riemann proposed attaching an inner product in the form of an m × m
positive definite matrix to each point of S. In doing this Riemann invented
the Riemannian Manifold. For the second, the tangent vectors, Tensor Analysis
defined a vector as an object that changes in certain ways under coordinate
changes. This works but is a bit unsatisfying. The more modern way is to define
tangent vectors as directional derivatives on the manifold. We will look at this
further in chapter 2 where we will define differentiable manifolds, although most
of the work has already been done in this section. For the remainder of this
chapter we will stay with embedded manifolds.
We must also mention that the theory we develop subsequently must be
modified if the manifold is not orientable. Examples of non-orientable manifolds
are the Möbius strip and the Klein bottle. The problem is that when a right
handed orthonormal basis of the tangent space is slid around the median line
of the Möbius strip it returns as a left handed orthonormal basis. The easiest
definition of orientable is that an n-dimensional manifold K (embedded or not)
is orientable if an only if it has a never vanishing form ω ∈ Λn (K). Such a form
is called a topform. From it one can always manufacture a volume form if K
has an inner product (e.g. is a Riemannian Manifold). We will not consider
non-orientable manifolds in what follows and some of the material will not work
on them. See Chapter 2 for more discussion of this matter.
and
r=1 ∗ : Λ1 (R) → Λn−1 (R)
r =n−1 ∗ : Λn−1 (R) → Λ1 (R)
In each case it is only necessary to derive one of the two formulas; the other is
then obtained from
∗ ∗ ω = (−1)r(n−r) for ω ∈ Λr (R)
To define ∗ it is necessary first to put an inner product on each Λr . This is
easily done; We define the inner product with
ω = ωi dui and η = ηi dui
by
(ω, η) = g ij ωi ηj
We can now put an inner product on each Λr as follows
(ω1 , η1 ) . . . (ω1 , ηr )
.. .. ..
(ω1 ∧ . . . ∧ ωr , η1 ∧ . . . ∧ ηr ) = . . .
(ωr , η1 ) . . . (ωn , ηr )
Thus
√ √
( g du1 ∧ . . . ∧ dun , gdu1 ∧ . . . ∧ dun ) = 1
We will define
def √
Ω0 = g du1 ∧ . . . ∧ dun
and refer to it as the normalized topform. The normalized topform is unique up
to sign; it will change sign if two variables interchange their numbers. Choice of
a sign is the same as choice of an orientation. The reason it is essentially unique
is that Λn is one-dimensional, so there are only two elements of size 1 and they
are negatives of one another.
If M is a n-dimensional region than the volume of M is
Z
vol(M ) = Ω0
M
There are of course subtleties here; for example the sign could come out wrong.
And of course you have to prove this works using whatever your definition of
volume is. We will ignore these problems. Also we mention that if f is the
density of something, electrical charge for example, or gas, then the amount of
the stuff will be Z
Amount of stuff in M = f Ω0
M
ω ∧ α = k Ω0
ω ∧ ∗η = (ω, η) Ω0
This is the most important equation involving ∗; if you remember this you
can derive everything else. The existence and uniqueness of ∗η are derived in
Chapter 2, but it is not a difficult matter; it comes down to the representation
of a linear functional in an inner product space.
Some equations readily follow from the basic equation. For example, we
have
which has important uses, for example in the next section on the codifferential.
Other important equations like
∗ ∗ ω = (−1)r(n−r)ω for ω ∈ Λr
1.19. THE DUALIZING OPERATOR ∗ 43
do not follow so readily from the basic equation. We will have more to say about
this equation later.
Next, using the basic equation ω ∧ ∗η = (ω, η) Ω0 we want to derive the
formulas for ∗ for general coordinates and 1-forms. Recall that a hat over a
term means that it is missing. We have {du1 , . . . , dun } is a basis for Λ1 . A
basis for Λn−1 is
ci ∧ . . . ∧ dun
du1 ∧ . . . ∧ du 1≤i≤n
Thus, since dui ∈ Λ1 , we have ∗dui ∈ Λn−1 and then we can express ∗dui as
n
X
∗dui = dj ∧ . . . ∧ dun
(−1)j−1 aj du1 ∧ . . . ∧ du
j=1
(−1)j−1 is inserted for convenience. We now compute both sides of the basic
equation and setting them equal will give us the aj . The basic equation is
duk ∧ ∗dui = (duk , dui )Ω0 . Computing the left side of the basic equation we
have
n
X
duk ∧ ∗dui = duk ∧ (−1)j−1 aj du1 ∧ . . . ∧ du dj ∧ . . . ∧ dun
j=1
n
X
= dj ∧ . . . ∧ dun
(−1)j−1 aj duk ∧ du1 ∧ . . . ∧ du
j=1
All terms on the right side in the sum will be 0 except the one where j = k
because if j 6= k there will be repetition in the differentials killing the term.
Thus
duk ∧ ∗dui = ak du1 ∧ . . . ∧ duk ∧ . . . ∧ dun
Notice how the (−1)j−1 was used to return duk to its proper place in the product
of differentials. This is a good trick to remember, though it is never important,
just convenient. Next we compute the right side of the of the basic equation.
We have
√
(duk , dui )Ω0 = g ki g du1 ∧ . . . ∧ dun
Comparing the two expressions we see that
√
ak = g ki g
and thus
n
X
∗dui = dk ∧ . . . ∧ dun
(−1)k−1 ak du1 ∧ . . . ∧ du
k=1
n
X √ dk ∧ . . . ∧ dun
= (−1)k−1 g ki gdu1 ∧ . . . ∧ du
k=1
44 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
This equation
Pn k−1 ki √ dk ∧ . . . ∧ dun
∗dui = k=1 (−1) g gdu1 ∧ . . . ∧ du
This is not quite sufficient for our needs. We also need formulas for ∗ when
r = 0 and when r = n. These are easy. Recall that Λ0 is just the scalars R and
that the inner product is just the ordinary product in R and that a basis for R
is just the single scalar 1. Then
∗1 = Ω0
1.19. THE DUALIZING OPERATOR ∗ 45
since
1 ∧ ∗1 = 1 ∧ Ω0 = Ω0 = (1, 1)Ω0
as required. Then
∗Ω0 = ∗ ∗ 1 = (−1)0(n−0) 1 = 1
∂R
ej =
∂ui
hi = (ei , ei )
1
êi = ei
hi
1
(dui , duj ) = g ii =
h2i
(hi dui , hi duj ) = 1
Ei = hi dui
so that {h1 du1 , . . . , hn dun } form an orthonormal set. For convenience let us
set Ei = hi dui . Then (Ei , Ej ) = δij . Now from the basic equation ω ∧ ∗η =
(ω, η)Ω0 we see immediately that
Here the convention is 1 ≤ i1 < . . . < ir ≤ n and 1 ≤ j1 < . . . < jn−r ≤ n but in
fact the formula will work just as well with the i’s and j’s in any order. Indeed
it suffices to check this for ω running through a basis {Ek1 ∧ · · · ∧ Ekr ) 1 ≤
i1 , . . . , ≤ kr } of Λr .
Now if {k1 , . . . , kr } 6= {i1 , . . . , ir } there will be repetitions on the right side and
the result will be 0. Hence we will take {k1 , . . . , kr } = {i1 , . . . , ir } and since
46 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
Now if {k1 , . . . , kr } =
6 {i1 , . . . , ir } then there will be a row in the determinant
which is entirely 0. Hence we take {k1 , . . . , kr } = {i1 , . . . , ir } and since both
are in increasing order we have iℓ = kℓ , the determinant has 1’s on the main
diagonal and 0’s elsewhere, so the determinant is 1 and the result is
Hence if {k1 , . . . , kr } =
6 {i1 , . . . , ir } both sides of the fundamental equation are
0 and if {k1 , . . . , kr } = {i1 , . . . , ir } then both sides are Ω0 , proving that
∗ ∗ ω = (−1)r(n−r) ω
1.20. THE CODIFFERENTIAL δ 47
Now if we move the r i’s each past the n − r j’s there will be a total of r(n − r)
hops to get from σ̃ back to σ. Hence if it takes s hops to return σ to the identity,
sgn(σ) = (−1)s and
Since this is true for the elements of a basis, it will be true for all ω by linearity.
Some points to remember. The inner product (ω, η) is an inner product that
lives in Tp (M ) for each p ∈ M . It and ∗ are strictly local, being algebraic con-
structions on each tangent space Tp (M ) separately. On the other hand, ((ω, η))
depends on integrating the information given at each p over the manifold. It is
thus a global object. Second, although we express it in our formulas for a single
coordinate system, in many cases it will be necessary to cut M into separate
pieces each of which lives inside the domain of a coordinate system, and to use
that coordinate system when integrating over that piece. Sometimes, as for the
sphere, we can use a single coordinate system even though that system is bad
at specific points, like the North and South pole. This is best considered dumb
48 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
luck, and one should always be careful when doing it. But it does work often in
applications as long as nothing is discontinuous at the bad points.
Now that we have the inner product on forms, we can consider ((ω, dη))
where ω ∈ Λr and η ∈ Λr−1 . We will consider two possible scenarios
a The manifold M has no boundary; ∂M = ∅ Such a manifold is often referred
to as closed. The idea ”no boundary” includes ”no boundary at infinity”
so that in this case the manifold M is compact.
b Either ω or η vanishes off a compact subset M0 of M , so the integral will
always be finite.
With one of these situations in place, we can do an integration by parts in
the way you have seen done in differential equations, and come up with formal
adjoint δ for d. After all this preamble, the actual calculation is quite short.
We need to recall that, with ω ∈ Λr−1 and η ∈ Λr
ω ∧ ∗η = η ∧ ∗ω
d(ω ∧ ∗η) = dω ∧ ∗η + (−1)r−1 ω ∧ d ∗ η
k2 ≡ k (mod 2)
−k ≡ k
2k ≡ 0
k(k − 1) ≡ 0
Thus
≡ (n − r)(r − 1) − 1
≡ n(r − 1) − r(r − 1) − 1
≡ n(r − 1) − 1 + 2
≡ n(r − 1) + 1
Hence
Z
n(r−1)+1
((dω, η)) = (−1) ω∧∗∗d∗η
K
Z
= ω ∧ ∗ (−1)n(r−1)+1 ∗ d ∗ η
K
Thus if we set
δ : Λr → Λr−1
defined by
δη = (−1)n(r−1)+1 ∗ d ∗ η η ∈ Λr
we have
((dω, η)) = ((ω, δη)) ω ∈ Λr−1 , η ∈ Λr
n odd: δω = (−1)r ∗ d ∗ ω
n even: δω = − ∗ d ∗ ω
There are some useful identities connecting ∗, d, and δ which are conse-
quences of ∗ being almost an involution, which we now derive. First we have
∗ δω = (−1)n(r−1)+1 ∗ ∗ d ∗ ω ω ∈ Λr
= (−1)n(r−1)+1 (−1)(n−r+1)(r−1) d ∗ ω ω ∈ Λr
= (−1)r d ∗ ω ω ∈ Λr
since
∗δ ∗ω = (−1)n−r d ∗ ∗ ω
= (−1)n−r+r(n−r) d ω
= (−1)(n−r)(r+1) d ω
∗∗ δ∗ω = (−1)(n−r)(r+1) ∗ d ω
(n−r−1)(r+1)
(−1) δ∗ω = (−1)(n−r)(r+1) ∗ d ω
δ∗ω = (−1)(n−r)(r+1)+(n−r−1)(r+1) ∗ d ω
δ∗ω = (−1)(r+1)(n−r+n−r−1) ∗ d ω
δ∗ω = (−1)r+1 ∗ d ω
The r in these formulas always refers to the degree of ω, and this must be
carefully remembered when applying the formulas. It is easy to make mistakes
applying these formulas.
i=1
∂xi2
Note also the extremely important point that by it’s definition the Laplacian
is independent of the coordinate system. It is even independent of the choice of
orientation, since reversing the orientation reverses the sign of each of the two
∗’s found in each term of it’s definition and since there are two minus signs each
term will come out the same.
Most annoyingly, the mathematical operator △ does not quite coincide with
the operator △ p used in classical mathematical physics and we must be a bit
clear which Laplacian we are using I have invented the symbol △ p for the familiar
Physics Laplacian.
There are a couple of things that can be said in favor of the mathematics
notation. First, the eigenvalues λ of △ are positive (or 0) while those of △ p
are negative (or 0). Connected with this is the fact that ((△ω, ω)) ≥ 0 while
((△p ω, ω)) ≤ 0. The wave equations are
1 ∂2u 1 ∂ 2u
+ △ω = 0 −△
p ω = 0
c2 ∂t2 c2 ∂t2
and the Schrödinger equations
∂Ψ ~2 ∂Ψ ~2 p
i~ = △Ψ+VΨ i~ =− △Ψ + V Ψ
∂t 2m ∂t 2m
and Poisson’s equation for the potential of charge distribution
ρ ρ
△φ = △
p φ = −
ǫ ǫ
with the math forms perhaps marginally more elegant. There may be other
reasons too for suspecting that Laplace chose the wrong sign. All we can do at
the present time is to be careful about which one we are using.
Notice that d : Λr → Λr+1 and δ : Λr → Λr−1 so that the combined
△ : Λr → Λr . Notice also that the definition of the Laplacian defines △ on
forms of all degrees, not just on functions. This is very important.
Another point of interest is that △ is a sort of square. Recall that d2 = dd = 0
and δ 2 = δδ = 0. Thus we have
(d + δ)2 = (d + δ)(d + δ) = dd + dδ + δd + δδ
= dδ + δd = △
d + δ : Λr → Λr+1 ⊕ Λr−1
which is just what we were expecting. A very similar calculation will get △f
for general coordinates, a result which we already have found by other methods,
and which you can do in the problems.
For r = 1 the computation is a lot more complicated and we will only do it
for n = 3, which avoids unpleasantness and suffices for our applications. (The
big advantage of doing it for n = 3 is that we can exploit cyclic order.) We will
first compute δdω, then compute dδω, and then put them together to get △ω.
Although there are some interesting features, you might want to just skim this
calculation.
ω = ω1 dx + ω2 dy + ω3 dz
∂ω3 ∂ω2 ∂ω1 ∂ω3 ∂ω2 ∂ω1
dω = − dy ∧ dz + − dz ∧ dx + − dx ∧ dy
∂y ∂z ∂z ∂x ∂x ∂y
∂ω3 ∂ω2 ∂ω1 ∂ω3 ∂ω2 ∂ω1
∗ dω = − dx + − dy + − dz
∂y ∂z ∂z ∂x ∂x ∂y
∂ ∂ω2 ∂ω1 ∂ ∂ω1 ∂ω3
d ∗ dω = − − − dy ∧ dz
∂y ∂x ∂y ∂z ∂z ∂x
∂ ∂ω3 ∂ω2 ∂ ∂ω2 ∂ω1
+ − − − dz ∧ dx
∂z ∂y ∂z ∂x ∂x ∂y
1.21. THE LAPLACIAN 53
∂ ∂ω1 ∂ω3 ∂ ∂ω3 ∂ω2
+ − − − dx ∧ dy
∂x ∂z ∂x ∂y ∂y ∂z
2
∂ ω2 ∂ 2 ω1 ∂ 2 ω1 ∂ 2 ω3
∗ d ∗ dω = − − + dx
∂y∂x ∂y 2 ∂z 2 ∂z∂x
2
∂ ω3 ∂ 2 ω2 ∂ 2 ω2 ∂ 2 ω1
+ − − + dx
∂z∂y ∂z 2 ∂x2 ∂x∂y
2
∂ ω1 ∂ 2 ω3 ∂ 2 ω3 ∂ 2 ω2
+ − − + dx
∂x∂y ∂x2 ∂y 2 ∂y∂z
OK, ∗d ∗ d is done. Now we need d ∗ d∗, and then we can put them together.
This one is easier.
ω = ω1 dx + ω2 dy + ω3 dz
∗ω = ω1 dy ∧ dz + ω2 dz ∧ dx + ω3 dx ∧ dy
∂ω1 ∂ω2 ∂ω3
d∗ ω = + + dx ∧ dy ∧ dz
∂x ∂y ∂z
∂ω1 ∂ω2 ∂ω3
∗d∗ ω = + +
∂x ∂y ∂z
2 2
2
∂ ω1 ∂ ω2 ∂ 2 ω3 ∂ ω1 ∂ 2 ω2 ∂ 2 ω3
d∗ d∗ ω = + + dx + + + dy
∂x2 ∂x∂y ∂x∂z ∂y∂x ∂y 2 ∂y∂z
2
∂ ω1 ∂ 2 ω2 ∂ 2 ω3
+ + + dz
∂z∂x ∂z∂y ∂z 2
How general is this result? It certainly cannot work for general coordinates, but
it does work for rectangular coordinates on Λr (Rn ). Unfortunately we do not
know any better way to prove this than by brute force. For our applications
what we have done suffices.
54 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
There is a vector form of the above calculation based on a well known formula
in three dimensions
which we rewrite as
△
p (v) = grad div(v) − curl curl(v)
In general coordinates and three dimensions we can use the above formula as a
definition of △
p (v) since we have formulas for grad div(v) and curl curl(v). If
you examine the differential form calculation above carefully, you will be able
to see the dδω corresponds to grad div(v) and δdω corresponds to curl curl(v).
Soon we will show that ∗ △ ω = △ ∗ ω in all circumstances. From this we
can easily derive that in Λ2 (R3 )
and in Λ3 (R3 )
△(f dxdydz) = △0 (f ) dxdydz
which tells us all we need to know about Λr (R3 ).
Now let’s develop a bit more theory. First, recall that the global inner
product (on all of K) is
Z Z
((ω, η)) = (ω, η)Ω0 = ω ∧ ∗η ω, η ∈ Λr (K)
K K
This shows that if ω has coefficients with two continuous derivatives that
where we are still assuming that supp(ω) or supp(η) is compact. The equation
with the condition on the supports means that △ is a formally self adjoint
operator. It is standard in elementary mathematical physics to leave things here.
To get real self adjointness requires consideration of the boundary conditions
for the operator, and to deal properly with this requires functional analysis and
Sobolev spaces, which is a can of worms we wish to keep in the can here.
Our final Laplacian duty is the equation ∗ △ ω = △ ∗ ω which we can do with
a lovely little calculation. Recall from the chapter on the codifferential that for
ω ∈ Λr (K)
d∗ω = (−1)r ∗ δω
δ∗ω = (−1)r+1 ∗ dω
Then we have
= d(δ ∗ ω) + δ(d ∗ ω)
= d((−1)r+1 ∗ dω) + δ((−1)r ∗ δω)
= (−1)r+1 (d ∗ dω) + (−1)r (δ ∗ δω)
= (−1)r+1 (−1)r+1 (∗δdω) + (−1)r (−1)r (∗dδω)
= ∗(δdω + dδω)
= ∗△ω
Notice that in going from step 4 to step 5 that dω ∈ Λr+1 and δω ∈ Λr−1 which
is what makes the signs come out right.
1 ∂B
dD = P dE = −
c ∂t
1 ∂D 1
dB = 0 dH = + j
c ∂t c
The equations of most significance derived from these are the equation of con-
tinuity and the potential equations. The equation of continuity is easy:
1 ∂D 1
dH = + j
c ∂t c
1 ∂ 1
0 = ddH = dD + dj
c ∂t c
1 ∂ 1
0 = P + dj
c ∂t c
∂P
0 = + dj
∂t
This is the form that the equation of continuity takes in our treatment. We note
however that if we just consider the coefficient of dxdydz from the equation it
becomes
∂ρ
0= + div j
∂t
Now we start on the potential equations.
dB = 0 =⇒ B = dA
(by the converse of the Poincaré lemma), where A is a 1-form. A is not uniquely
determined; we can add dG to A for any G ∈ Λ0 (that is, a function), since
We can use this G to modify A and this is called changing the gauge. We will
return to this matter later on. Next we have
1 ∂
dE = − B
c ∂t
1 ∂ 1 ∂A
= − dA = − d
c ∂t c ∂t
1 ∂A
d E+ = 0
c ∂t
1 ∂A
Since E + c ∂t is a 1-form, there must be a 0-form −φ for which
1 ∂A
E+ = −dφ
c ∂t
58 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
A′ = A + dG
′ 1 ∂
φ = φ− G
c ∂t
where we have set C = 0 as customary.
Our next job is to derive the potential equations. However, there is so much
slop in A that we cannot reasonably expect nice equations without putting some
extra conditions on A. We could simply pull the condition out of the air, but it
will be more fun to see it appear in context.
Recall that the physics Laplacian △ p is the negative of the mathematical
δω = (−1)r ∗ d ∗ ω ω ∈ Λr
Now we have
ǫµ ∂ 2 A µ ǫµ ∂φ
− △A =
p ∗ j + d δA −
c2 ∂t2 c c ∂t
ǫµ ∂ 2 A µ
A = 2 2
−△p A = ∗j
c ∂t c
with velocity √cǫµ if not for the term d δA − ǫµ ∂φ
c ∂t . Hence, using the slop in
A, we will set this to 0.
ǫµ ∂φ
δA − =0 Condition of Lorenz
c ∂t
Of course, we need to know that adding a suitable dG will force the Condition
of Lorenz to be true. We will look at this later. (Usually this step is neglected!)
It is also useful to decode the condition of Lorenz. We have
ǫµ ∂φ
(−1)1 ∗ d ∗ A − = 0
c ∂t
ǫµ ∂φ
− ∗ d (A1 dydz + A1 dzdx + A1 dxdy) − = 0
c ∂t
∂A1 ∂A2 ∂A3 ǫµ ∂φ
−∗ + + dxdydz − = 0
∂x ∂y ∂z c ∂t
∂A1 ∂A2 ∂A3 ǫµ ∂φ
− + + + = 0
∂x ∂y ∂z c ∂t
We must also have the potential equation for φ. This is derived in a similar
manner, but easier. It is interesting that it throws up the same extra terms which
we eliminate with the Condition of Lorenz. This suggests that the time and
space variables might have more of a connection than we expect á priori. This
was likely one of the things that induced Lorenz develop the Lorenz-Einstein
transformation equations. We imitate the previous calculation, noting that
60 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
δφ = 0 since δ is 0 on Λ0 ,
1 ǫµ ∂ 2 φ
△φ = ∗P − 2 2
ǫ c ∂t
ǫµ ∂ 2 φ 1 1
−△
p φ = ∗P = ∗ (ρΩ0 )
c2 ∂t2 ǫ ǫ
2
ǫµ ∂ φ 1
−△
p φ = ρ
c2 ∂t2 ǫ
which is the familiar wave equation for φ in coordinate independent form (except
for orientation considerations and provided that ρ really is the physical charge
density and has not been modified to fit into some special coordinate system).
Note that we used △ p φ = − △φ again. Note also that the equation is identical in
form to the Dalembertian equation for A. And finally note that the Condition of
Lorenz did not pop up so naturally in this derivation as it did in the calculation
for A.
Our next job is a look at the Condition of Lorenz. A close examination of
the derivation of the wave equation for A will show that if A is a solution of
the equation then the Condition of Lorenz must hold. The question is, can we
always force the Condition of Lorenz to hold by choosing an appropriate G in
1 ∂G
A′ = A + dG φ′ = φ −
c ∂t
The answer, as we will show, is yes. We need
ǫµ ∂φ′
δA′ − =0
c ∂t
How do we find the G?. Substituting into this equation we have
ǫµ ∂ 1 ∂G
δ(A + dG) − φ− = 0
c ∂t c ∂t
1.23. INDEFINITE INNER PRODUCT SPACES 61
ǫµ ∂ 2 G ǫµ ∂φ
(δd + dδ)G + + δA − = 0
c2 ∂t2 c ∂t
ǫµ ∂ 2 G ǫµ ∂φ
+ △G = − δA −
c2 ∂t2 c ∂t
ǫµ ∂ 2 G ǫµ ∂φ
−△ p G = − δA −
c2 ∂t2 c ∂t
(recalling that △
p = −△) and once again we have a wave equation. We solve for
G, correct the A and φ to A′ and φ′ , and then solve the wave equations for A′
and φ′ . Actually we don’t have to make the correction, since solving the wave
equations for A and φ will work just fine; the A and φ we find will automatically
satisfy the Condition of Lorenz. The importance of the above is to show that
A and φ we seek actually exist.
Notice the c. replaces the usual condition (v, v) > 0 for v 6= 0. In indeffinite
inner product spaces there will certainly be vectors for which (v, v) = 0. In
Relativity these are called null vectors and in math isotropic vectors.
Using the usual methods of linear algebra and being careful to avoid vectors
v for which (v, v) = 0 one can construct without difficulty an orthonormal basis
e1 , . . . , en for V . Renumbering if necessary they can be arranged in a sequence
62 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
(ei , ei ) = 1 for i = 1, . . . , n − s
(ei , ei ) = −1 for i = n − s + 1, . . . , n
Here, orthonormal means (ei , ei ) = ±1. Those elements of the basis with
(ei , ei ) = +1 will be called positive basis elements and those elements with
(ei , ei ) = −1 will be called negative basis elements. J. J. Sylverster proved early
in the history of linear algebra (1852) that the number s of negative basis ele-
ments does not depend on the choice of basis; all orthonormal bases will have
the same s. This is called Sylvester’s law of inertia. This s will metastisize
through all the formulas.
It is worth mentioning that since we have assumed that the inner product
is non-degenerate there will indeed by n basis vectors in the orthonormal basis.
For the orthonormal basis defined above we get the usual matrix of metric
coefficients gij = (ei , ej ) and it and its inverse will be
1 0 ··· 0 0 0 ··· 0
0 1 ··· 0 0 0 ··· 0
.. .. .. .. .. .. .. ..
. . . . . . . .
0 0 ··· 1 0 0 ··· 0
E=
0 0 · · · 0 −1 0 ··· 0
0 0 · · · 0 0 −1 ··· 0
.. .. .. . .. .. .. ..
. . . .. . . . .
0 0 ··· 0 0 0 · · · −1
We know that det(αij ) > 0 because the orientations match. Then we have
E = (αij )⊺ (dui , duj ) (αij )
and thus
det E = det (αij )⊺ (dui , duj ) (αij ) = det(αij )⊺ det (dui , duj ) det(αij )
2
(−1)s = det(αij ) det(g ij )
2
Since [det(α)] is positive (−1)s and det(g ij ) have the same sign and we have
(−1)s det(g ij ) > 0. Thus to normalize the topform du1 ∧ . . . ∧ dun it is natural
to take
1
Ω0 = p du1 ∧ . . . ∧ dun
(−1) det(g ij )
s
Our next job is to revise the formula for ∗ ∗ ω. We take E 1 , . . . , E n−s be the
normalized positive basis elements, ((E i , E i ) = +1), and E n−s+1 , . . . , E n to be
the normalized negative basis elements, ((E i , E i ) = −1). We now take a typical
basis element E i1 ∧ . . . ∧ E ir with i1 < i2 < · · · < ir . To this basis element
corresponds a permutation
1 2 . . . r r + 1 r + 2 . . . n
σ=
i1 i2 . . . ir k1 k2 . . . kn−r
with k1 < k2 < · · · < kn−r . We suspect from our previous work that
∗ E i1 ∧ . . . ∧ E ir = a E k1 ∧ . . . ∧ E kn−r
Let s1 of E i1 , . . . , E ir have (E i , E i ) = −1
k1 kn−r
Let s2 of E , . . . , E have (E i , E i ) = −1
where of course s1 + s2 = s. Using the definition of ∗ and our guess above about
the form of ∗ E i1 ∧ . . . ∧ E ir we have
E i1 ∧ . . . ∧ E ir ∧ ∗(E i1 ∧ . . . ∧ E ir ) = det (E ij , E iℓ ) Ω0
64 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
E i1 ∧ . . . ∧ E ir ∧ a E k1 ∧ . . . ∧ E in−r = (−1)s1 Ω0
a sgn(σ)E 1 ∧ . . . ∧ E n = (−1)s1 Ω0
a sgn(σ)Ω0 = (−1)s1 Ω0
a sgn(σ) = (−1)s1
a = (−1)s1 sgn(σ)
and thus
Hence
∗ ∗ ω = (−1)r(n−r)+s ω for ω ∈ Λr
Then
Z
((dω, η)) = dω ∧ ∗η
K
1.24. MAXWELL’S EQUATIONS IN SPACE-TIME 65
Z Z
= d(ω ∧ ∗η) − (−1)r−1 ω∧d∗η
K K
Z Z
= ω ∧ ∗η − (−1)r−1 ω∧d∗η
∂K K
Z
= 0 + (−1)r ω∧d∗η
K
by Stoke’s theorem, because we are assuming boundary conditions that kill off
the boundary intergral, as is usual in this kind of calculation. Continuing we
have
Z
r (n−r+1)(r−1)+s
((dω, η)) = (−1) (−1) ω∧∗∗d∗η
K
Z
= (−1)n(r−1)+s+1 ω∧∗∗d∗η
K
Z h i
= ω ∧ ∗ (−1)n(r−1)+s+1 ∗ d ∗ η
K
= ((ω, (−1)n(r−1)+s+1 ∗ d ∗ η))
So if we define
δη = (−1)n(r−1)+s+1 ∗ d ∗ η
we have
((dω, η)) = ((ω, δη))
Finally we want to introduce the analog of the Laplacian. Since the situation
is so different from that of positive definite inner products, it seems reasonable
to use a different notation, especially since it (more or less) coincides with that
usual in physics. We will use the notation ω for the new operator, which is
defined exactly as was the Laplacian in the positive defintate case, namely
Def ω = (dδ + δd)ω
The notation is meant to suggest a relationship with the classical Dalember-
tian
1 ∂2f ∂2f ∂2f ∂2f
p
f= 2
− 2
− 2 − 2
c ∂t ∂x ∂y ∂z
For Special Relativity, the coordinate system cdt, dx, dy, dz forms an oth-
onormal coordinate system, with that order. Thinking of it as dx0 , dx1 , dx0 , dx1
for convenience, we have
(cdt, cdt) = +1, (dx, dx) = −1, (dy, dy) = −1, (dz, dz) = −1,
and
(dxi , dxj ) = 0 for i 6= j
The matrix for this coordinate system is
1 0 0 0
0 −1 0 0
ij i j
(g ) = (dx , dx ) =
0 0 −1 0
0 0 0 −1
We can perform the ∗ operation using the following permutations, all of
which are even permutations (have sgn(σ) = +1).
0 1 2 3 0 1 2 3 0 1 2 3
0 1 2 3 0 2 3 1 0 3 1 2
0 1 2 3 0 1 2 3 0 1 2 3
2 3 0 1 3 1 0 2 1 2 0 3
Using these and the formula
where
1 2 ··· r r+1 r + 2 ··· n
σ=
i1 i2 · · · ir j1 j2 · · · jn−r
and s1 is the number of negative basis elements (that is dx, dy, dz) among
dxi1 , . . . , dxir , we have the following formulas
∗1 = cdt ∧ dx ∧ dy ∧ dz
∗ cdt = dx ∧ dy ∧ dz ∗ dx ∧ dx ∧ dx = cdt
∗ dx = cdt ∧ dy ∧ dz ∗ cdt ∧ dy ∧ dz = dx
∗ dy = cdt ∧ dz ∧ dx ∗ cdt ∧ dz ∧ dx = dy
∗ dz = cdt ∧ dx ∧ dy ∗ cdt ∧ dx ∧ dy = dz
∗ cdt ∧ dx = −dy ∧ dz ∗ dy ∧ dz = cdt ∧ dx
∗ cdt ∧ dy = −dz ∧ dx ∗ dz ∧ dx = cdt ∧ dy
∗ cdt ∧ dz = −dx ∧ dy ∗ dy ∧ dz = cdt ∧ dx
∗ cdt ∧ dx ∧ dy ∧ dz = −1
For example, for the 2nd entry we have the permutation
0 1 2 3
σ=
0 1 2 3
1.24. MAXWELL’S EQUATIONS IN SPACE-TIME 67
with s1 = 0 and sgn(σ) = +1, whereas for the fourth entry we have
0 1 2 3
σ=
2 0 3 1
with s1 = 1 and sgn(σ) = −1, this being the the fourth permatation in the
above list with the second and third entries swapped. For the seventh entry we
have
0 1 2 3
σ=
0 2 3 1
so s1 = 1 and sgn(σ) = 1, this being the second permutation in the list. The
entries in the second column can be derived from those in the first, but beware
since the permutation is reversed. The fourth row second column has
0 1 2 3
σ=
0 3 1 2
and here s1 = 2 and σ is the third entry in the list of permatations so sgn(σ) = 1.
It is far easier to derive the second column using ∗ ∗ ω = (−1)r(4−r)+3 ω.
Next we are interested in the codifferential δ. We have, as usual, with wedges
omitted for ease of reading,
A = −φ cdt + A1 dx + A2 dy + A3 dz
Then
The minus signs come from swapping differentials around and sometimes from
the action of ∗. This is the negative of the classical d’Albertian
1 ∂2f ∂2f ∂2f ∂2f
p f =
− − −
c2 ∂t2 ∂x2 ∂y 2 ∂z 2
An obnoxious calculation shows that if
A = A0 cdt + A1 dx + A2 dy + A3 dz
then
A = ( A0 )cdt + ( A1 )dx + ( A2 )dy + ( A3 )dz
We know of no intelligent way to prove this but in view of its importance we
will do the calculation in an appendix to this chapter, which should be skimmed
from a distance.
We now want to apply this equipment to the problem of potential equations
in four dimensional Space-Time. We first consider the case when ǫ = µ = 1
where everything is simple. When these are not 1, there are certain complica-
tions which we will discuss later.
For ease of reference we repeat here Maxwell’s equations.
div D = ρ curl E = − 1c ∂B
∂t D = ǫE
1 ∂D 1
div B = 0 curl H = c ∂t + c j B = µH
along with
1 ∂A
curl A = B dφ = −E −
c ∂t
Let us set as is usual F = dA. We then have
A = −φcdt + A1 dx + A2 dy + A3 dz
1 ∂A1 ∂φ 1 ∂A2 ∂φ 1 ∂A3 ∂φ
F = dA = + cdtdx + + cdtdy + + cdtdz
c ∂t ∂x c ∂t ∂y c ∂t ∂z
∂A3 ∂A2 ∂A1 ∂A3 ∂A2 ∂A1
+ − dydz + − dzdx + − dxdy
∂y ∂z ∂z ∂x ∂x ∂y
= −E1 cdtdx − E2 cdtdy − E3 cdtdz
+B1 dydz + B2 dzdx + B3 dxdy
Next we take the ∗ of both sides. We remember that here we have ǫ = 1 and
µ = 1 so that E = D and B = H. Then
1 1 1
= ρ dxdydz − j1 cdtdydz − j2 cdtdzdx − j3 cdtdxdy
c c c
1 1 1
δF = ∗ d ∗ F = ρ cdt − j1 dx − j2 dy − − j3 dz
c c c
Now the coup; recall that we may take δA = 0 (the condition of Lorenz) so that
we have
1 ∂ 2 Ai ∂ 2 Ai ∂ 2 Ai ∂ 2 Ai
p A =
i − − − = − Ai
c2 ∂t2 ∂x2 ∂y 2 ∂z 2
we have the potential equations in four space
p φ = ρ
p A =
1
i ji
c
which is the same result we got in the section on Maxwell’s equations in three
dimensional space.
We now want to extend this result to the case where ǫ or µ are not one.
This is surprisingly difficult. However, we suspect it can be done because the
corresponding equations in three space have the same form for the potential φ
and the vector potential components Ai . Recall that the form of the equation
is
ǫµ ∂ 2 Ai ∂ 2 Ai ∂ 2 Ai ∂ 2 Ai µ
− − − = ji
c2 ∂t2 ∂x2 ∂y 2 ∂z 2 c
This suggests an electromagnetic wave moving at a speed of k = √cǫµ . Now we
must steer the boat, like Odysseus, between Charybdis and Skilla. Charybdis
refers to the fact that for the d’Alembertian to come out with k12 = ǫµ c2 in it,
we are going to have to modify the ∗ operator by changing c to k. Skilla refers
to the fact that Maxwell’s equations have c not k in them. Thus replacing all
the c’s by k’s won’t work; we must steer more subtlely. Nevertheless the fact
that the equations for φ and Ai have the same form in 3-space suggests that it
is possible to navigate successfully, which we now do.
In the following calculations the ∗ operator uses the same equations as before
but the constant c in those equations is replaced by the constant k. Except for
70 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
some trivial algebra the calculation goes as before; the potential form A is, as
√
before, recalling that k = c/ ǫµ,
A = −φ cdt + A1 dx + A2 dx + A3 dx
c
A = − φ kdt + A1 dx + A2 dx + A3 dx
k
√
A = − ǫµ φ kdt + A1 dx + A2 dx + A3 dx
F = dA
1 ∂A1 ∂φ 1 ∂A2 ∂φ 1 ∂A3 ∂φ
= + cdtdx + + cdtdy + + cdtdy
c ∂t ∂x c ∂t ∂y c ∂t ∂z
∂A3 ∂A2 ∂A1 ∂A3 ∂A2 ∂A1
+ − dydz + − dzdx + − dxdy
∂y ∂z ∂z ∂x ∂x ∂y
= −E1 cdtdx − E2 cdtdy − E3 cdtdz
+B1 dydz + B2 dzdx + B3 dxdy
√
= ǫµ (−E1 kdtdx − E2 kdtdy − E3 kdtdz)
+B1 dydz + B2 dzdx + B3 dxdy
√
∗F = ǫµ (E1 dydz + E2 dzdx + E3 dxdy)
+B1 kdtdx + B2 kdtdy + B3 kdtdz
√
ǫµ
= (D1 dydz + D2 dzdx + D3 dxdy)
ǫ
µ
+ √ (H1 cdtdx + H2 cdtdy + H3 cdtdz)
ǫµ
r
µ
= F̃
ǫ
where
1.25. ENERGY FLOW AND ENERGY DENSITY 71
r
ǫ
F̃ = ∗F
µ
= H1 cdtdx + H2 cdtdy + H3 cdtdz
+D1 dydz + D2 dzdx + D3 dxdy
Now we find dF̃ ,
1 ∂D1 ∂H3 ∂H2 1 ∂D2 ∂H1 ∂H3
dF̃ = − − cdtdydz + − − cdtdzdx
c ∂t ∂y ∂z c ∂t ∂z ∂x
1 ∂D3 ∂H2 ∂H1 ∂D1 ∂D2 ∂D3
+ − − cdtdxdy + + + dxdydz
c ∂t ∂x ∂y ∂x ∂y ∂z
1
= ρ dxdydz − (j1 cdtdydz + j2 cdtdzdx + j3 cdtdxdy)
c
√
ǫµ
= ρ dxdydz − (j1 kdtdydz + j2 kdtdzdx + j3 kdtdxdy)
c
Finally, we have
A = (δd + dδ)A
= δdA + 0 (conditon of Lorenz δ A = 0)
= δF r = ∗ d ∗ Fr
µ µ
= ∗d F̃ = ∗ dF̃
ǫ ǫ
r √
µ ǫµ
= ∗ ρ dxdydz − (j1 kdtdydz + j2 kdtdzdx + j3 kdtdxdy)
ǫ c
r √
µ ǫµ
= ρ kdt − (j1 dx + j2 dy + j3 dz)
ǫ c
r √
µ 1 ǫµ
= √ ρ cdt − (j1 dx + j2 dy + j3 dz)
ǫ ǫµ c
ρ µ
= cdt − (j1 dx + j2 dy + j3 dz)
ǫ c
p = −, to
This decodes in the usual way, with
1 ∂ 2 φ ∂ 2 φ ∂ 2φ ∂ 2 φ ρ
p φ
= 2 2
− 2
− 2 − 2 =
c ∂t ∂x ∂y ∂z ǫ
1 ∂ 2 Ai ∂ 2 Ai ∂ 2 Ai ∂ 2 Ai µ
p A
i = 2 2
− 2
− 2
− 2
= ji
c ∂t ∂x ∂y ∂z c
~ + d~ℓ is
~ to R
Thus the work done in going from R
Thus
dW ~ · ~v
= eE
dt
Now let ρ be a continuous charge distribution. Then, as usual, we replace e by
ρ dt and integrate to find dW/dt. We a also need ~j = ρ~v . The we have
Z
dW ~ · ~v
= (ρ dτ )E
dt
ZM
= ~ · (ρ~v ) dτ
E
M
Z
= ~ · ~j dτ
E
M
∂E
= ǫE ∧ ∗
∂t
∂E
= ǫ ∧ ∗E
∂t
∂E
= ∧D
∂t
Thus
1 ∂ 1 ∂E ∂D
(E ∧ D) = ∧D+E∧
2 ∂t 2 ∂t ∂t
1 ∂D ∂D
= E∧ +E∧
2 ∂t ∂t
∂D
= E∧
∂t
In an exactly similar calculation we have
∂B 1 ∂
∧H = (B ∧ H)
∂t 2 ∂t
Resuming the previous calculation we have
Z
dW 1 ∂D
= c E∧ − + dH
dt c ∂t
ZM
1 ∂D
= c −E ∧ + E ∧ dH
M c ∂t
Z
1 ∂D
= c −E ∧ + dE ∧ H − d(E ∧ H)
c ∂t
ZM
1 ∂D 1 ∂B
= c −E ∧ + − ∧ H − d(E ∧ H)
c ∂t c ∂t
ZM
1 ∂ 1 ∂
= c − (E ∧ D) − (B ∧ H) − d(E ∧ H)
M 2c ∂t 2c ∂t
Z Z
∂ 1 1
= − (E ∧ D) − (B ∧ H) − c E∧H
∂t M 2 2 ∂M
we can interpret the previous equation as the change in the Energy as due to
the decrease in the field energy less the flux of energy through the boundary.
74 CHAPTER 1. INTRODUCTION AND BASIC APPLICATIONS
Chapter 2
Mathematical Theory
75
76 CHAPTER 2. MATHEMATICAL THEORY
2.1 INTRODUCTION
This chapter has two purposes. First we must prove the two major theorems
of elementary differential forms, and second we want to extend the theory with
some deeper methods. For example, we would like to have formulas for the ∗
operator on manifolds, where orthonormal coordinates systems are usually not
possible, so we need some more flexible methods for that.
We also want discuss the mathematical theories that underlie the develop-
ment in the first chapter. Also, we will look at the theory in a more general
setting; that of a differentiable manifold rather than simply n-space. A dif-
ferentiable manifold is like a surface in n-space but with more dimensions and
without the n-space. If the manifold is actually inside an n-space the manifold
is called an embedded manifold. In point of fact, all manifolds can be embed-
ded in an n-space of sufficiently high dimension 1 , but it is no longer customary
to do this. Also it may not be natural to do this; few people think that the
four dimensional Space-Time of our universe is embedded in a five (or more)
dimensional Euclidean space Rn .
circumstances
2.5. PERMUTATIONS 77
2.5 Permutations
One cannot get very far using differential forms without the dualizing operator
∗ which we saw used in the previous section for deriving formulas for curvilinear
coordinates. In this section we will give an introduction to the ∗ operator which
will be more of a tour than a treatment; the full treatment will be found in
Chapter 2. Thus this section will present the basic ideas and the formulas the
physicist will find most useful. Formulas for more sophisticated uses will be
found in Chapter 2.
A critical and not particularly well liked part of this is some of the lore of
permutations. A permutation is a rearrangement of the set {1, 2, . . . , n}; for ex-
ample {3, 5, 6, 1, 2, 4} is a rearrangement of {1, 2, 3, 4, 5, 6}. By interchanging two
elements, adjacent or not, a permutation may be brought back to {1, 2, . . . , n}.
There are many ways to do this of course, but it turns out that the number
of interchanges to do this is always the same mod 2. Thus no matter how the
interchanges are done, (−1)s will always have the same value, and that value
is the sign of the permutation. It is convenient for us (though somewhat non-
standard) to write a permutation σ as a function with inputs at the top and
outputs at the bottom. Thus the permutation above would be written
1 2 3 4 5 6
σ =
3 5 6 1 2 4
Here are a possible sequence of interchanges to get back to the identity.
1 2 3 4 5 6 1 2 3 4 5 6
−→ −→
3 5 4 1 2 6 3 2 4 1 5 6
1 2 3 4 5 6 1 2 3 4 5 6
−→
1 2 4 3 5 6 1 2 3 4 5 6
We used four interchanges; s = 4. Clearly this can be done in many ways but
the number of interchanges s will always be an even number, and thus
sgn(σ) = (−1)s = +1
is well defined.
There is one other useful observation. Suppose that a permutation has the
property that it compounded of two increasing subsequences of the numbers
{1, 2, . . . , n}; for example
1 2 3 4 5 6 7 8 9
3 5 7 8 1 2 4 6 9
or more generally with i1 < . . . < ir and ir+1 < . . . < in
1 . . . r r + 1 . . . n
σ=
i1 . . . ir ir+1 . . . in
78 CHAPTER 2. MATHEMATICAL THEORY
Note that the elements after the vertical line are determines by those before it.
Now the interesting thing here is that the sign of such a permutation can be
found as follows. First set
r(r + 1)
Tr = 1 + 2 + · · · + r =
2
Then for such a σ Pr
sgn(σ) = (−1) j=1 ij −Tr
Since this kind of permutation is the most common type in differential forms,
this is quite handy.
gij = (ei , ej )
P i
Then v = i v ei for any v ∈ V but, following Einstein, we will omit the
sum sign when an index is repeated, one up one down, and we will write this
v = v i ei . Similarly w = wi ei and we have
(v, w) = (v i ei , wj ej )
= v i wj (ei , ej )
= gij v i wj
2.6. THE OPERATOR Φ 79
This connects (v, w) whose value does not depend on a choice of basis with
quantities gij , v i and wj whose values do depend upon the basis. It is important
to be able to go back and forth between coordinate free notation and indexed
notation where the objects depend upon the choice of basis. Tensor notation is
an attempt to live in both worlds, and it does it quite well, at the expense of
rather complex looking equations.
Because an inner product is non-degenerate, (which means that if (v, w) = 0
for all w ∈ V then v = 0) we must have
det(gij ) 6= 0
This is most important.
A linear functional ℓ is defined by the requirement that
ℓ(αv + βw) = α ℓ(v) + β ℓ(w)
The set of all linear functionals is a vector space and is called V ∗ .
An important theorem is
Theorem Representation of linear functional.
Let ℓ be a linear functional on a vector space with an inner product. There
there exists a unique u ∈ V for which
ℓ(v) = (u, v)
We will prove this shortly. First we give an important example. If v = v i ei
the we can define a linear functional whose value on v is v i . This functional is
called Ei (note upper index). It should be clear that
1 if i = j
Ei (ej ) = δji =
0 if i 6= j
Now let ℓ ∈ V ∗ be a linear functional and let λi = ℓ(ei ). Then I claim that
ℓ = λi Ei . Indeed
λi Ei (v) = λi Ei (v j ej )
= λi v j Ei (ej )
= λi v j δji
= λj v j
= v j ℓ(ej )
= ℓ(v j ej )
= ℓ(v)
Then it is easy to prove that the Ei are linearly independent and thus form a
basis for V ∗ .
Now back to Φ : V → V ∗ . Let ℓ ∈ V ∗ be a linear functional and let it be
represented by u ∈ V . Then the theorem says that for all v ∈ V we have
ℓ(v) = (u, v)
80 CHAPTER 2. MATHEMATICAL THEORY
We define Φ by
Φ(u) = ℓ
Since this has been defined in terms of the inner product, which does not depend
upon the basis, Φ is independent of the basis.
For our next task we need to define
(g ij ) = (gij )−1
When these matrices are written out the inverse relationship becomes
gij g jk = δik
ij
g gjk = δki
Also note that gij = (ei , ej ) = (ej , ei ) = gji the order of the indices in the
above equations is not important.
It is now time to prove the theorem. Since we would like to know the
connection between ℓ and u in coordinates, it makes sense to prove the theorem
using coordinates. Let ℓ = λj Ej and define u by
ui = g ij λj and u = u i ei
(u, v) = (ui ei , v k ek )
= gik ui v k
= gik g ij λj v k
= gki g ij λj v k
= δkj λj v k
= λk v k
= λk Ek (v)
= ℓ(v)
and of course
Our next job is to put an inner product on V ∗ and we will do this by exporting
the inner product on V to V ∗ in such a way as to make Φ an isometry. That is
we define, for λ, µ ∈ V ∗ ,
Def
(λ, µ) = (Φ−1 (λ), Φ−1 (µ))
2.6. THE OPERATOR Φ 81
(λ, µ) = gij g ik λk g jm µm
= δjk λk g jm µm
= λj g jm µm
= g jm λj µm
83