Chapter04 PDF
Chapter04 PDF
Lecturer: McGreevy
4-1
4.1 Basic principles and definitions
The machinery we are about to develop will take us from simple laws of mechanics (what’s
the energy of a configuration of atoms) to results of great generality and predictive power.
Ensemble we recognize, at least. This name means: counting states of an isolated system.
Isolated means that we hold fixed
V, the volume
(walls can’t move and do work on unspecified entities outside the room.)
Sometimes other things are held fixed, sometimes implicitly: magnetization, polarization,
charge.
4-2
For later convenience, we adopt the following lawyerly posture. Assume that the energy
lies within a narrow range:
E ≤ energy ≤ E + ∆, ∆ E .
Think of ∆ as the imprecision in our measurements of the energy – our primitive tools cannot
measure the energy of the system to arbitrary accuracy. This is a technicality, in that all of
our results will be independent of ∆. We don’t need this lawyering for V .
Consider an ensemble of copies of the system all with the same macrostate, but in different
microstates.
Now we’re going to think about what we mean by equilibrium, from a microscopic point
of view. 1
Suppose that at t = 0 the probability of finding any accessible microstate is equal. That
is, every accessible microstate occurs with equal probability.
Now let the laws of mechanics evolve each state in the ensemble forward in time.
Claim: At later times, every accessible microstate still occurs in the ensemble with equal
probability.
Starting with this distribution, it stays that way. This is what we mean by thermal equi-
librium.
It’s possible to prove this in both classical mechanics and quantum mechanics with some
heavy machinery (Liouville’s Theorem). 2
1
Everything I say between here and The Fundamental Principle of Statistical Mechanics is in-
tended as motivation for that Principle. Much of it (the second half) is not rigorous – actually it’s worse:
it is rigorous in a way which is not physically relevant. If you are unhappy about it, you can ignore it and
take the Principle on its own terms as an axiom. Understanding the correct basis of Statistical Mechanics
(that is, the actual reason that it successfully describes much of the world) in fact remains a frontier research
problem in theoretical physics.
2
(Actually: We haven’t shown that there aren’t other distributions with this property.)
4-3
What about the converse?
Suppose that at t = 0 the ensemble contains only a subset of accessible states. For example,
suppose all the air molecules are found in half the room.
Evolve each member of this ensemble forward in time.
It is highly unlikely that a system in this ensemble stays in the specified subset of states.
An unnecessarily strong claim is: if we wait long enough, each member of the ensemble
will pass through all the accessible states. This statement is demonstrably true (and goes
by names like “ergodic theorem”, “Boltzmann H theorem”) for some systems, under some
assumptions; however, the fact that systems equilibrate is more general.3
It is not fully understood, but it is a fact of experience: things equilibrate. We take this as
Postulate 2:
• If an isolated system is not found with with equal probability for each accessible mi-
crostate, it is not in equilibrium.
The first of the above statements is the contrapositive of the converse of our Posulate 1.
It is equivalent to:
If an isolated system is in equilibrium, it is found with equal prob in each of its accessible
microstates.
3
Remark which can be ignored: The specific point I am complaining about regarding the assumptions of
these theorems is that an estimate of the time it takes for a system to explore all of the accessible microstates
(in the classical case, one asks how long it takes to come within a fixed small distance from any given point
in the phase space) is much longer than it takes in practice for systems to be well-described by equilibrium
thermodynamics.
4-4
Combining these two statements, we get:
An isolated system is in equilibrium iff all accessible microstates are equally probable.
Claim: All the rest of 8.044 follows from this box. We could have begun the class here if
we wanted to be deductive (not always the best way to learn). From it we can derive:
• relationships among macroscopic variables e.g. P V = N kT, U = 32 N kT 2nd Law, Curie
law for paramagnets...
• probability densities for microscopic quantities (e.g. velocity distribution for an ideal gas
in equilibrium.)
This will be our starting point for deriving thermodynamics and other things. Whether it
is an accurate description of, say, a series of snapshots of a real system which we observe to
be an an equilibrium macrostate is not something we are prepared to address. Rather, it is
a useful starting point for encoding our partial knowledge about such a system.
A more Bayesian perspective on what we are doing is: we want to encode our partial
knowledge about the microstate in a probability distribution on the phase space (this is a
name for the space of microstates). All we know is the macroscopic variables E, V, N . The
distribution which is consistent with this information but introduces the smallest amount
of extraneous, wrong information is the uniform distribution on the accessible states. This
perspective has been forcefully advocated by E. Jaynes; we will come back to it again later.
4-5
Now, in equations:
Note: Ω is dimensionless.
In classical mechanics, this counting becomes integration; we get a probability density for
xi , pi .
4
A tricky point which you should ignore on a first pass (and which you should ignore completely if you
haven’t taken QM):
You might object to this statement that p is a probability, and not a probability density. What about
superposition? That is: how do we determine the probability of being in a state a|k1 i + b|k2 i? a, b here are
continuously variable, so to construct probabilities we would need a probability density.
I haven’t told you how to answer this question; it’s harder. The formula above is the answer to a more
restricted question. We’ve assumed here that we have picked a basis of the Hilbert space labelled by the
quantum number k, and we know that the state is one of these basis vectors. The answer to the more general
question, which alas we will not discuss in 8.044 requires density matrices.
4-6
Classical version:
This p is now a probability density for 6N variables. Those variables are {p, q}, a specific set
of N positions (qs) and N momenta (ps) in three dimensions.5 The space of {p, q} is called
phase space.
Note that the integrand of these integrals is very boring: 1. All of the physics is in the
limits of integration.
Yes, we are using p to denote both probability density and momentum. There are only so
many letters.
5
Some of you may note a signficant ambiguity here: if I change coordinates on phase space, the probability
distribution transforms like a distribution (recall ‘functions of a random variable’); if p is uniform with respect
to one choice, it will in general not be uniform with respect to others. This problem simply doesn’t exist in
the quantum version. The resolution is that p is uniform in certain special sets of coordinates, namely when
{p, q} are canonically conjugate variables. This is the case which arises when classical mechanics arises as a
(large-quantum-numbers) limit of quantum mechanics, as indeed it always does.
4-7
Squashing
How do we extract information from this formal distribution? First microscopic data, then
thermodynamics.
Suppose we are interested in what one of the particles is doing, e.g. in its momentum
distribution in the x direction:
Z Z
1
p(p1x ) = p({p, q}) = 1
all variables except p1x Ω accessible values of all variables with p1x , V, E, N fixed
The integrand is still 1; all information is in the limits of integration.
More generally: if we pick X to label a special set of states of the system in which some
subset of {p, q} are fixed to satisfy some condition (e.g. all the x coordinates are fixed to be
in the northern half of the room), then
Z
1
p(X) = 1
Ω accessible values of all un-fixed coordinates
Ω0 (consistent with X)
=
Ω
volume of accessible phase space consistent with X
=
volume of accessible phase space
The prime on Ω0 is to distinguish this quantity from the total Ω. In cases (QM) where the
states are discrete, this is:
how many states satisfy the given condition X
p(X) =
how many states altogether
Now you see why I asked to you to count ways to get various poker hands.
Then the cumulative probability and the probability density are related in the usual way, by
differentiating. Our lawyering with the energy window ∆ makes an appearance here. If we
define the density of microstates at energy E to be ω:
ω(E, V, N ) ≡ ∂E Φ(E, V, N )
Then
Ω(E, V, N ) = ω(E, V, N )∆ .
Ω is a probability (that the energy is in the window of width ∆), and ω is a probability
density.
4-8
4.2 Temperature and entropy
Consider a situation where two systems separated by a diathermal wall are isolated from the
rest of the universe:
In the second step here we enumerated the states of the whole system as a product of the
number of states of the two systems.
Maximize ln p(E1 ):
∂E2
=⇒ 0 = ∂E1 ln Ω1 (E1 ) + ∂E2 ln Ω2 (E2 )
∂E
|{z}1
=−1
4-9
At the E1 at which p(E1 ) is extremized:
This is a property of the systems which is equal when 1 + 2 are in thermal equilibrium,
and which depends only on 1 or 2 . This must therefore be a function of T .
~.
d¯W = 0 means (as above) we’re not doing work on the system: constant V , M
Why 1/T ? Because this definition gives the ideal gas law, that is, it is the same as ther-
modynamic temperature, defined above. Soon.
Call the equilibrium value of E1 E1eq . This is the value of E1 at which the system has the
most microstates available to it.
4-10
Entropy and the 2nd Law
p(E1 ) ≤ p(E1eq )
Ω1 (E1eq ) Ω2 (E − E1eq )
rearrange: 1 ≤
Ω1 (E1 ) Ω2 (E − E1 )
4-11
Entropy and the 2nd Law, cont’d.
S(E, V, N ) ≡ kB ln Ω(E, V, N )
With this definition of entropy, we have (log is a monotonic function so the inequalities are
preserved):
0 = kB ln 1 ≤ S1 (E1eq ) − S1 (E1 ) + S2 (E − E1eq ) − S2 (E − E1 )
=⇒ ∆S1 + ∆S2 ≥ 0
(where ∆S1,2 are the changes in entropy during the equilibration process)
4-12
Basic facts about entropy:
• Entropy measures (the logarithm of) the number of accessible microstates. Information
theory point of view: it’s a measure of our ignorance of the microstate given the macrostate.
Ωtotal = Ω1 Ω2 =⇒ Stotal = S1 + S2
Note that when two bodies at different temperatures are brought into contact, our statistical
notions show that heat flows from the hotter body to the colder body. This follows since the
change in entropy is
!
∂S1 ∂S2 1 1
0 ≤ ∆S = − ∆E1 = − ∆E1
∂E1 i1 ∂E2 i2 T1i T2i
i
where T1,2 are the initial temperatures.
4-13
Thermodynamic entropy and statistical entropy are the same.
dE2
dS2 = Now note: V is fixed. No work is done. dE2 = d¯Q2
Treservoir
d¯Q2
= ← d¯Q2 ≡ heat entering 2
Treservoir
d¯Q1
= − d¯Q1 = −d¯Q2 (since the whole system is isolated)
Treservoir
2nd Law: dS = dS1 + dS2 ≥ 0
d¯Q1
=⇒ dS1 ≥
Treservoir
This is the form of the 2nd Law valid when exchanging heat with a reservoir.
Equality occurs if the heat is exchanged quasistatically:
quasistatic d¯Q1
dS1 = |quasistatic
T
Note: for quasistatic processes, T1 = Treservoir = T.
4-14
Now we can rewrite
First Law of Thermodynamics for Quasistatic Processes as:
dU = d¯Q + d¯W this is the general version
If there are other ways of doing work on the system, there are more terms here.
4-15
Program for the rest of 8.044:
• Pick a system
4-16
4.3 Classical monatomic ideal gas
Here we will see how the formal structure introduced above actually works for a classical
monatomic ideal gas. This means N ∼ 1024 1 featureless classical particles (which do not
interact with each other) in a box with sides Lx , Ly , Lz . So the volume is V = Lx Ly Lz .
Comments:
• Notation: p~2i ≡ p2xi + p2yi + p2zi . I may sometimes omit the little arrow. So the sum in
H is over all 3N momenta.
• The energy doesn’t care about where the atoms are – H is independent of the qi . In a
non-ideal gas there would be extra terms depending on qi − qj – that is, interactions.
• Here, there is only translational kinetic energy. No contribution from e.g. molecular
vibrations or rotations. Hence, monatomic gas. Later we will consider diatomic gases
which have more microscopic variables.
4-17
Z
Ω = Ω(E, V, N ) = dq1 dq2 ...dq3N dp1 dp2 ...dp3N · 1
accessible
| microstates}
{z
all the physics is hidden in here
Z Lx N Z Ly N Z Ly N Z
= dx dy dz dp1 ...dp3N · 1
0 0 0 E≤H({p})≤E+∆
N
Z
= Lx Ly Lz dp1 ...dp3N · 1
| {z } E≤H({p})≤E+∆
V
1
P3N
Now recall that H({p}) = 2m i=1 p2i , so the condition H({p}) ≤ E defines a
X
3N -dimensional ball : {p| p2i ≤ 2mE}
i
√
(the interior of a 3N − 1-dimensional sphere) of radius R = 2mE.
What I mean by x! here is defined by the recursion relation x! = x(x − 1)! with the initial
conditions
1 1√
1! = 1 and ! = π.
2 2
6
Note that Stirling’s formula still works for non-integer x. Check that the formula for v3N
works for N = 1.
6
R ∞ Forx x−swhich are not half-integers the right generalization is the Gamma function x! = Γ(x + 1) ≡
0
dss e . We won’t need this in 8.044.
4-18
V N π 3N/2
Φ(E, V, N ) = 3N
(2mE)3N/2
2
!
3N/2
N 4πemE
Φ=V
3N
3N/2
dΦ 3N 1 N 4πemE
ω= = V
dE 2 E 3N
3N/2
3N ∆ N 4πemE
Ω(E, V, N ) = ω∆ = V
2 E 3N
This result is wrong for two reasons, one clerical and one important.
• The clerical reason is the convention about the normalization of Ω. The Ω we wrote
above is dimensionful. To make Ω dimensionless, we divide by h3N = (2π~)3N . This
factor had better not matter in any thermodynamic observable. (Notice that our
lawyering factor ∆ also multiplies the answer for Ω.) This will provide a small check
on our answers.
• The important thing we left out is that the atoms are identical. The counting we have
done above treats the atoms as if they are wearing nametags which we could use to
distinguish which atom is which, and then we could say things like “Oh yeah, that’s
Bob over there. His kinetic energy is 2eV.” This is not an accurate account of the
microphysics. The 1st atom here and the 2nd atom there is the same microstate as
the 2nd atom here and the 1st atom there:
4-19
This is actually a result from quantum mechanics. In classical mechanics it is possible
to make a gas out of particles with nametags. (Actually we could just label them by
their histories.) That’s just not the kind of gas we get from atoms and molecules, which
are really described by quantum mechanics, in a way which sometimes (conveniently
for us) reduces to classical mechanics. Any two atoms of the same kind (same isotope,
same state of nucleus, same state of electrons) are identical. This is a deep fact7 .
We’ve over-counted the configurations of the atoms by a factor of N !. If we continue
to leave out this factor of 1/N ! from the correct answer we will run headfirst into a
problem with the extensivity of our thermodynamic variables which are supposed to
be extensive, which is called Gibbs’ Paradox (on the next pset). Gibbs recognized that
this factor of 1/N ! had to be there in the late 19th century, by this reasoning. That is
bad-ass. In some ways this can therefore be considered the first QM result.
N 3N/2
V E 3N ∆ em 3N/2 N
= e (2)
N N 2E 3π~2
| {z }
this is the important bit for thermo
Now that we’ve done the counting, let’s follow the rest of the 8.044 prescription. Find S
then T then P . The physical results are about to fall out here fast and easily like few other
times your experience of physics.
V 3 E 3 em 3N ∆
ln N + 2 ln N +
S = kB ln Ω = kB N 1 + ln 2
+kB ln
| 2 {z 3π~ }
2E
| {z }
only differences in entropy matter, classically N. ignore.
V 3 E
=⇒ S = kB N ln + ln + const
N 2 N
Note that S is extensive: cut the box in half and see what happens to the entropy. If we
divide E, V, N by 2, we can see from the expression above that S is divided by 2.
You can see that if we had not put in the 1/N !, S would not have been extensive.
7
It is actually a result of quantum field theory: e.g. the electrons in the atoms are all identical because
they are all quanta of the same field.
4-20
Internal energy of monatomic ideal gas, from scratch
Rearrange this to find an expression for the internal energy of a monatomic ideal gas:
3
E = N kB T for a monatomic ideal gas.
2
5 5
=⇒ CP = CV + N k B = N k B , =⇒ γ = .
2 3
4-21
Ideal gas law from scratch
We have derived this statement, starting from the basic postulate of statistical mechanics.
Be impressed.
where c1,2,3 are constants, independent of the thermodynamic state (these constants drop out
when we take derivatives, as they did above). So we see that along a path where the entropy
is constant, P T 5/2 is constant. A constant entropy, reversible path has ∆S = ∆QT
= 0, no
heat in or out. This determines the shape of an adiabat. It agrees with the answer we got
before.
4-22
Microscopic information from the stat mech postulate
So far we got thermodynamic relations out, now let’s think about microscopic probability
densities for ideal gas. We’ll do two examples:
1) What is p(xi )? p(xi ) ≡ the probability density for the x-position of the ith particle, in
equilibrium.
Q: WHICH ONE IS THE ith ONE? Here I am imagining that I can label one of the
particles. The factor of (N − 1)! for the remaining particles will drop out of the normalized
probabilities.
−1 N N
LN
x Ly Lz · · ·
=
LN N N
x Ly Lz · · ·
1
= for 0 ≤ xi ≤ Lx
Lx
This was to be expected. Since the positions of the atoms in an ideal gas are statistically
independent, we infer that the probability that all the atoms are in the North half of the
N 24
room = 21 ∼ 2−10 .
4-23
2) What is p(pxi )? p(pxi ) ≡ the probability density for the x-component of the momentum
of the ith particle.
Ω0 (pxi )
Z
1
p(pxi ) = = dq1 ...dq3N dp1 ...dp3N
Ω Ω E≤H( {p, q} )≤E+∆ | {z }
| {z } except pxi
list includes pxi
In this expression, we integrate only over values of the other pi where the energy ε ≡ p2xi /(2m)
(the KEx of atom 1) plus the energy of all the other pi adds up to E. So we have energy
E − ε to distribute amongst the other 3N −√ 1 pi . The cumulative probability is the volume
of a ball in 3N − 1 dimensions, of radius 2m E − ε. So Ω0 is the same as Ω in Eqn. (2) but
with 3N → 3N − 1 ps and E → E − ε:
N 3N2−1
(3N − 1)∆ em 3N2−1 N
0 V E−ε
Ω = e .
N 3N − 1 2(E − ε) π~2
E−ε
3N2−1 3N −1 N −1
!
Ω0 VN h3N 3N −1 (4πem) 2
E−ε
p(pxi ) = = 3N 3N N
Ω VN h3N E 2 (4πem) 2
E
3N
|{z} |{z}
range of accessible positions same. same normalization factor.
Note that the units work: [Ω] = [1], [Ω0 ] = [p1xi ] . The last factor approaches 1 in the thermo-
dynamic limit N → ∞, ε E. This leaves:
1/2 3N −1
N 3N/2
3 (E − ε) 2
p(pxi ) = 3N/2 3N −1
4πem | E {z } |(N − 1/3) 2
≡A
{z }
≡B
Use E = 32 N kB T :
23 N
ε
1 kB T 1 − k εT
A' q 1 −
3
'q e B
3
N kB T 2
N 3
N kB T
2 2
1 N
(For the second step, recall: limN →∞ 1 + N
= e, the base of the natural log.) Similarly,
! 23 N
√ 1 √ 1/2
B' N 1
' Ne .
1 − 3N
4-24
Put back together the probability distribution for the x-momentum of one particle:
1/2
√
3 1 − k εT 1 2
p(pxi ) ' q N e1/2 e B =√ e−pxi /(2mkB T )
4πem 3
N kB T 2πmkB T
2
Correctly normalized, correct units. To find the velocity distribution from the momentum
distribution, recall vxi = pxi /m and our discussion of functions of a random variable:
1/2 −1 2
m 2 mvxi
p(vxi ) = e kB T
(3)
2πkB T
This is the distribution (notice that the units are right and it’s correctly normalized) that
I presented out of nowhere in Chapter 2. It’s called the Maxwell velocity distribution,
and we just derived it from the Fundamental Postulate – equal probabilities for accessible
microstates.
− KEx
Note that the probability in (3) is proportional to e kB T – the quantity in the exponent is
the energy in question divided by the energy scale associated with the temperature. This is
a Boltzmann factor, with which we will have much truck in the coming weeks.
With this victory in hand we conclude our discussion of the classical, monatomic, ideal gas.
Later we will remove and modify these modifiers.
Some words about the bigger picture: this straight-ahead method of just doing the integral
definining Ω(E) doesn’t usually work. (One more system next for which it works, a few more
on pset 6.)
To do better, we need a more powerful method (the Canonical Ensemble, which is the sub-
ject of Chapter 6). Before that, in Chapter 5 we finish up our discussion of Thermodynamics
– in particular with this understanding in terms of counting accessible microstates, we are
better prepared to discuss entropy as a state variable.
4-25
4.4 Two-state systems, revisited
•
N = N0 + N1 E = 0 + εN1
|{z} |{z}
# OFF # ON
4-26
How many microstates?
S ≡ kB ln Ω
(Stirling)
=⇒
= kB N ln N − N1 ln N1 − (N − N1 ) ln(N − N1 ) −N + N1 + (N − N1 )
| {z }
=0
E
S(E)/kB = N ln N − N1 ln N1 − (N − N1 ) ln(N − N1 ) N1 =
ε
Now we can get all the stuff:
1 ∂S 1 ∂S kB
= = = (− ln N1 − 1 + ln(N − N1 ) + 1)
T ∂E ε ∂N1 ε
ε N − N1 N0
= ln = ln
kB T N1 N1
ε N
e kB T = −1
N1
N εN
N1 = ε =⇒ E= ε (4)
1+e kB T
1 + e kB T
Plots:
∂S 1
∂E
= T
4-27
From the plot we see that N1 → 0 at low T , N1 → N/2 at high T .
At high T , ON and OFF are equally probable; the fact that ON has higher energy becomes
irrelevant for kB T ε.
4-28
Heat capacity:
2 ε
dE ε e kB T
C≡ |ε = N kB 2
dT kB T ε
1 + e kB T
2
ε − k εT
low T : C ∼ N kB e B
kB T
2
ε 1
hight T : C∼
kB T 4
In the formulae above we’ve assumed that all the 2-state systems have the same level
spacing. A generalization of the above formula for the energy to the case where the level
spacings are differnet (which may be useful for a certain extra credit problem) is
N
X εi
E= εi .
i=1 1 + e kB T
This formula is actually quite difficult to derive using the microcanonical ensemble, though
you can see that it gives back the formula for the energy (4) above. It will be simple using
the canonical ensemble, in Chapter 6.
4-29
Probability that a bit is ON
Ω0 (N → N − 1, N1 → N1 − n)
p(n) =
Ω
(N − 1)! N1 ! (N − N1 )!
=
| N{z! } (N1 − n)! (N − 1 − N1 + n)!
1
| {z } | {z }
=N
1 if n = 0
N − N1 if n = 0
= =
N if n = 1
1 if n = 1
1
N1
1 −
N
for n = 0
p(n) =
N1
N
for n = 1
This is a check on the formalism.
N1 1
p(1) = = ε
N 1 + e kB T
4-30
Recap of logic
Model the system. What is the configuration space? What is the Hamiltonian?
S = kB ln Ω
1 ∂S P ∂S
= , =
T ∂E V,N... T ∂V E,N,...
Ω0 (X)
p(specific property X) =
Ω
The counting is the hard part. We can only do this for a few simple cases.
4-31