0% found this document useful (0 votes)
81 views

Physics 115A: Statistical Physics: Prof. Clare Yu Email: Cyu@uci - Edu Phone: 949-824-6216 Office: 210E RH Fall 2013

This document provides an introduction to the statistical physics course Physics 115A taught by Professor Clare Yu at UCI. It discusses how statistical physics is needed to describe systems with large numbers of interacting particles, since directly solving equations of motion is intractable. There are two main approaches: thermodynamics uses general assumptions to deduce relationships between macroscopic properties, while statistical mechanics induces these from microscopic interactions. Probability and binomial distributions are introduced as tools to relate microscopic states to macroscopic observables like temperature and pressure.

Uploaded by

Mahendra Sankhua
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
81 views

Physics 115A: Statistical Physics: Prof. Clare Yu Email: Cyu@uci - Edu Phone: 949-824-6216 Office: 210E RH Fall 2013

This document provides an introduction to the statistical physics course Physics 115A taught by Professor Clare Yu at UCI. It discusses how statistical physics is needed to describe systems with large numbers of interacting particles, since directly solving equations of motion is intractable. There are two main approaches: thermodynamics uses general assumptions to deduce relationships between macroscopic properties, while statistical mechanics induces these from microscopic interactions. Probability and binomial distributions are introduced as tools to relate microscopic states to macroscopic observables like temperature and pressure.

Uploaded by

Mahendra Sankhua
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 12

Physics 115A: Statistical Physics

Prof. Clare Yu
email: [email protected]
phone: 949-824-6216
Office: 210E RH
Fall 2013

LECTURE 1
Introduction
So far your physics courses have concentrated on what happens to one object or a
few objects given an external potential and perhaps the interactions between objects.
For example, Newton’s second law F = ma refers to the mass of the object and its
acceleration. In quantum mechanics, one starts with Schroedinger’s equation Hψ = Eψ
and solves it to find the wavefunction ψ which describes a particle. But if you look
around you, the world has more than a few particles and objects. The air you breathe
and the coffee you drink has lots and lots of atoms and molecules. Now you might
think that if we can describe each atom or molecule with what we know from classical
mechanics, quantum mechanics, and electromagnetism, we can just scale up and describe
1023 particles. That’s like saying that if you can cook dinner for 3 people, then just scale
up the recipe and feed the world. The reason we can’t just take our solution for the single
particle problem and multiply by the number of particles in a liquid or a gas is that the
particles interact with one another, i.e., they apply a force on each other. They see the
potential produced by other particles. This makes things really complicated. Suppose
we have a system of N interacting particles. Using Newton’s equations, we would write:
d~pi d2~xi X ~
=m 2 = Fij = −∇Vi (~x1 , ~x2 , ..., ~xN ) (1)
dt dt j6=i

where F~ij is the force on the ith particle produced by the jth particle. Vi is the total
potential energy of the ith particle assuming that all the forces are conservative. We
must also specify the initial conditions, i.e., the initial positions and velocities of all the
particles. Then we must solve the N coupled second order partial differential equations.
This is quite a daunting task. It’s true that we now have powerful computers and
molecular dynamics simulations carry out such tasks, but they can only handle a few
thousand particles and track their motions for perhaps a few times 10−8 seconds. So we
throw up our hands. This is not the way to go.
Fortunately, there is a better way. We can use the fact that there are large numbers
of particles to apply a statistical analysis to the situation. We are usually not interested
in the detailed microscopics of a liquid or a gas. Instead we are usually interested in
certain macroscopic quantities like

• pressure

• temperature

• entropy

• specific heat

• electrical conductivity

• magnetic and electric susceptibilities

• etc.

These concepts don’t make sense for one atom or even a few particles. For example,
what is the temperature of one atom? That doesn’t make sense. But if we have a whole
bucket of atoms, then temperature makes sense. Since we are not concerned with the
detailed behavior of each and every particle, we can use statistical methods to extract
information about a system with N ≫ 1 interacting particles. For example, we may be
interested in the average energy of the system, rather than the energy of each particle.
Statistical methods work best when there are large numbers of particles. (Think of
polling.) Typically we have on the order of a mole’s worth of particles. A mole is the
number of atoms contained in 12 grams of carbon 12. (A carbon 12 atom has 6 protons
and 6 neutrons for a total 12 nucleons. Its atomic weight is 12.) So a mole has 6.022×1023
atoms. This is called Avogadro’s number.
There are two basic approaches to describing the properties of a large number of
particles:

Thermodynamics: This is essentially a postulational approach, or a deductive approach.


On the basis of a few general assumptions, general relationships between and among
macroscopic parameters are deduced. No reference is made to the microscopic
details of the system.

Statistical Mechanics: This is an inductive approach. With only a few assumptions,


relations between the macroscopic parameters of a system are induced from a con-
sideration of the microscopic interactions. This approach yields all of the thermo-
dynamic results plus a good deal more information about the system of interacting
particles.

2
Probability
We will talk more about probability in the discussion section. If we flip a coin, the
chance of heads is 1/2 and the chance of tails is 1/2. The probability of an outcome is
the number of ways of getting that outcome divided by the total number of outcomes.
We often write P (xi ) for the probability that an outcome is xi . So if we flip 2 coins,
then P(head,head)=1/4, and P(head,tail)=1/2 because there are a total of 4 possible
outcomes and 2 ways of getting one coin heads and one coin tails.
Binary Model
Let’s work out a simple example of a binary model. Rather than heads or tails, think
of a line of spins which can be up (↑) or down (↓). The spins correspond to magnetic
moments ±m. +m corresponds to ↑ and −m corresponds to ↓. Suppose we have N
sites, each with a spin. On each site the spin can be up or down. The total number of
arrangements is 2 × 2 × 2 × 2...2 = 2N . So the total number of states or outcomes is 2N .
We can denote a state by
↑↑↑↓↓↑↓↓ ... (2)
Now suppose we want the number of configurations with n up sites, regardless of what
order they’re in. The number of down sites will be N-n, since the total number of sites
is N. The number of such configurations is
N!
(3)
n!(N − n)!
To see where this comes from, suppose we have N spins of which n are up and (N-n) are
down. How many ways are there to arrange them? Recall that the number of ways to
arrange N objects with one object on each site (or in each box) is
• the first place can be occupied by any one of the N objects
• the second place can be occupied by any one of the N-1 remaining objects
etc. until the last place can be occupied by the last 1 object. So there are
N (N − 1)(N − 2)...1 = N ! (4)
configurations. In these configurations the n up spins can be arranged in any of n! ways.
Also the remaining (N-n) sites have down spins which can be arranged in any of (N − n)!
ways. The spins are regarded as distinguishable. So there are 2 ways to have ↑↑: ↑↑
and ↑↑. They look the same, but if one were red and the other blue, you would see the
difference (red, blue) and (blue, red). We are dividing out this overcounting; that’s what
the denominator n!(N − n)! is for.
So the probability of a state with n up spins and (N-n) down spins is
N!
P (n ↑, N ) = ÷ (total number of states)
n!(N − n)!
N! 1
= (5)
n!(N − n)! 2N

3
Now suppose that the probability of a site getting ↑ is p and the probability of
getting ↓ is q = 1 − p. So far we have considered up and down to be equally probable,
so p = q = 1/2. But what if p 6= q? This might be caused by an external magnetic field
which biases the spins one way or another. If we have 2 spins, the probability of getting
2 up spins is p2 . (Recall that the probability of flipping a coin twice and getting heads
both times is 1/4.) If we want n up spins and (N-n) down spins, then the probability is

P (n ↑, N ) = (number of arrangements) × pn q N −n
N!
= pn q N −n (6)
n!(N − n)!

When p = q = 1/2, this reduces to our previous result (5). Equation (6) is called
the binomial distribution because the prefactor N !/n!(N − n)! is the coefficient in the
binomial expansion:
N
N!
(a + b)N = aN −n bn
X
(7)
n=0 n!(N − n)!
Normalization, Averages, Second Moment
Let’s go back to our general considerations of probability distribution functions. Let
P (xi ) be the probability distribution for the probability that an outcome is xi . If we sum
over all possible outcomes, one of them is bound to happen, so
X
P (xi ) = 1 (8)
xi

In other words, the probability distribution function P (xi ) is normalized to one. One
can also have a probability distribution function P (x) of a continuous variable x. For
example, suppose you leave a meter stick outside on the ground. P (x)dx could be the
probability that the first raindrop to hit the stick will strike between position x and
x + dx. P (x) is called the probability density. The normalization condition is
Z
dxP (x) = 1 (9)

Note that dxP (x) is the area under the curve P (x). The normalization sets this area
R

equal to 1.
The average of some quantity x is

dxP (x)x
R
< x >= x = R (10)
dxP (x)

Averages can be denoted by < x > or x. For a discrete quantity

P (xi )xi
P
x = Pi (11)
i P (xi )

4
If the probability is normalized, then the denominator is 1. For example, suppose N (Ai )
is the number of people with age Ai . Then the average age is
P∞
N (Ai ) Ai
A =0
A = Pi∞
Ai =0 N (Ai )
dA N (A) A
R
→ R (12)
dA N (A)
The probability that a person is A old is
N (A)
P (A) = R (13)
dA N (A)

Notice that this satisfies the normalization condition. More generally, if f (x) is a function
of x, then the average value of f (x) is

dxP (x)f (x)


R
f (x) = (14)
dxP (x)
R

One of the more useful functions concerns the deviation of x from the mean x:

∆x = x − x (15)

Then

∆x = x − x
= x−x
= 0 (16)

A more useful quantity is the square of the deviation from the mean

(∆x)2 = (x − x)2
= (x2 − 2xx + x2 )
= x2 − 2xx + x2
= x2 − x2 (17)

This is known as the second moment of x about its mean. The first moment is just x.
In general the nth moment of x about its mean is given by
dx(x − x)n P (x)
R
(x − x)n = (18)
dxP (x)
R

Two other terms that are often used are the most probable value of x and the median.
The most probable value of x is the maximum of P (x). The median xmed is the value of
x such that half the values of x are greater than xmed , i.e., x > xmed , and half the values

5
of x are less than xmed , i.e., x < xmed . In terms of the area under the curve, the median
is the vertical dividing line such that half the area lies to the left of the median and half
to the right of the median.
Gaussian Distribution and Central Limit Theorem
One of most useful distributions is the Gaussian distribution. This is sometimes called
the bell curve which is well known as the ideal grade distribution.

P(x)

µ x
The formula for a Gaussian distribution is
1 2 2
P (x)dx = √ e−(x−µ) /2σ dx (19)
2πσ
where µ = x is the mean. The coefficient is set so that the normalization condition is
satisfied. σ 2 = (∆x)2 = (x − x)2 . 2σ is the width of the distribution. There is a 68%
chance that −σ ≤ x ≤ σ. One obtains this by integrating P (x) from −σ to +σ. σ is
sometimes called the root–mean–square (rms) deviation or the standard deviation. You
will have the opportunity to check some of these assertions in your homework.
For a Gaussian distribution, all the higher moments can be expressed in terms of the
first and second moments. For example (x − x)4 = 3(x − x)2 .
There is one other interesting aspect of Gaussian distributions. It’s called the central
limit theorem. It can be stated in terms of a random walk in one dimension. A drunk
starts at a lamp post and staggers back and forth along the one dimensional sidewalk.
Let w(s)ds be the probability that a step length lies in the range between s and s + ds.
No matter what the probability distribution w(s) for each step may be, as long as the
steps are statistically independent and w(s) falls off rapidly enough as |s| → ∞, the total
displacement x will be distributed according to the Gaussian law if the number of steps
N is sufficiently large. This is called the central limit theorem. It is probably the most
famous theorem in mathematical probability theory. The generality of the result also
accounts for the fact that so many phenomena in nature (e.g., errors in measurement)
obey approximately a Gaussian distribution.

6
In fact, if we go back to the binomial distribution given in eq. (6), then in the limit of
large N, this becomes the Gaussian distribution (see Reif 1.5). In this case, x represents
the net magnetization, i.e., the difference between the number of up and down spins:

x = N↑ − N↓ (20)

The average magnetization


µ = (p − q)N (21)
and the mean square deviation is
σ 2 = 4N pq (22)
More on the Random Walk
(Reference: Howard C. Berg, Random Walks in Biology.)
The random walk is an important concept. Let us go back to the drunk starting at
the lamp post and stumbling either to the right or the left. Consider an ensemble of M
drunks. Let xi (N ) be the position of the ith drunk after N steps. Suppose everyone
starts at the origin so that xi (N = 0) = 0 for all i. Let the step size be of fixed length s.
Then
xi (N ) = xi (N − 1) ± s (23)
The average displacement from the origin is
M
1 X
x(N ) = xi (N )
M i=1
M
1 X
= [xi (N ) = xi (N − 1) ± s]
M i=1
M
1 X
= xi (N − 1)
M i=1
= x(N − 1)
= 0 (24)

So the mean displacement is zero because on average, half of the displacements are
negative and half are positive. But this does not mean that all the drunks are sitting at
the origin after N steps.
It is useful to look at the spread in positions.
q A convenient measure of spreading is
the root-mean-square (rms) displacement x2 (N ). Here we average the square of the
displacement rather than the displacement itself. Since the square of a negative or a
positive number is positive, the result will be positive.

x2i (N ) = x2i (N − 1) ± 2sxi (N − 1) + s2 (25)

7
Then we compute the mean,
M
1 X
x2i (N ) = x2 (N )
M i=1 i
M h
1 X i
= x2i (N − 1) ± 2sxi (N − 1) + s2
M i=1
= x2i (N − 1) + s2 (26)

So x2 (0) = 0, x2 (1) = s2 , x2 (2) = 2s2 ,..., and x2 (N ) = N s2 . So the rms displacement


q √
x2 (N ) = Ns (27)

The rms displacement is a good measure of the typical distance that the drunk reaches.
Notice that it goes as the square root of the number of steps. If τ is the time between
successive steps, then it takes a time t = N τ to take N steps. So
s
q
t √
x2 (N ) = s ∼ ts (28)
τ
Notice that the rms displacement goes as square root of the time t.
A random walk is what particles execute when they diffuse. Think of a opening a
bottle of perfume. The perfume molecules diffuse through the air. We can define a
diffusion coefficient D by
s2
D= (29)

(The factor of 2 is by convention.) Then in one dimension

x2 (t) = 2Dt
q √
x2 (t) = 2Dt (30)

For 2 dimensions, if the x and y directions are independent,

r2 (t) = x2 (t) + y 2 (t) = 4Dt (31)

Similarly for 3 dimensions

r2 (t) = x2 (t) + y 2 (t) + z 2 (t) = 6Dt (32)

Notice that in all dimensions, the mean square displacement is linear in the number of
steps N and in the time t. The diffusion coefficient D characterizes the migration of
particles in a given kind of medium at a given temperature. In general, it depends on
the size of the particle, the structure of the medium, and the absolute temperature. For
a small molecule in water at room temperature, D ≃ 10−5 cm2 /sec.

8
If we go back to the drunk who is taking steps of size s to the right or to the left,
then we can use the binomial distribution. Let p be the probability to take a step to the
right and q = 1 − p be the probability to take a step to the left. After taking N steps,
the mean number n1 of steps to the right is (see Reif Eq. (1.4.4)):
N
X
n1 = P (n1 , N )n1 = N p (33)
n1 =0

The mean number of steps to the left is

n2 = N q (34)

Let the displacement be


∆x = ms = (n1 − n2 )s (35)
The mean displacement is (Reif Eq. (1.4.6)):

∆x = ms = (n1 − n2 )s = N (p − q)s (36)

So if p = q = 1/2, then the mean displacement ∆x = 0.


The displacement squared is

(∆x)2 = (∆m)2 s2 = (m − m)2 s2 (37)

The dispersion of the net displacement or the mean square displacement is

(∆x)2 = (∆m)2 s2 = 4N pqs2 (38)

where we used Reif Eq. (1.4.12). If p = q = 1/2, then the mean square displacement
(∆x)2 = N s2 which is what we got before. The rms displacement is
q q √
(∆x)2 = 2 N pqs ∼ Ns (39)
q √
If p = q = 1/2, then the rms displacement is (∆x)2 = N s which is what we got
before. Notice again that the rms displacement goes as the square root
q of the √ number of
steps. Again the characteristic or typical displacement is given by (∆x)2 ∼ N s.
Poisson Distribution
(Ref.: C. Kittel, Thermal Physics) It is worth mentioning another widely used dis-
tribution, namely, the Poisson distribution. This distribution is concerned with the
occurrence of small numbers of objects in random sampling processes. For example, if on
average there is one bad penny in 1000, what is the probability that no bad pennies will
be found in a given sample of 100 pennies? The problem was first considered and solved
by Poisson (1837) in a study of the role of luck in criminal and civil trials in France.

9
So suppose that there are N trials and the probability for a certain outcome is p. The
probability that there will be n outcomes of a certain type is

hnin e−hni
P (n) = (40)
n!
where the mean number of desired events is hni = N p. You will derive the Poisson
distribution in your homework starting from the binomial distribution.
Example: Let us consider the problem mentioned above. If on average there is one
bad penny in 1000, what is the probability that no bad pennies will be found in a given
sample of 100 pennies?
Solution: We are interested in the outcome that there is a bad penny. The probability
of getting a bad penny is p = 1/1000 = 0.001. The number of trials is N = 100. So
the average number hni of bad pennies in 100 is hni = N p = 100/1000 = 0.1. The
probability of finding no bad pennies is P (n = 0) which is given by

hni0 −hni
P (n = 0) = e = e−hni = e−0.1 = 0.905 (41)
0!
Alternative solution: The probability of getting a bad penny is p = 1/1000 = 0.001.
The probability of not getting a bad penny is q = 1 − p = 0.999. So the probability of
no bad pennies in 100 is P (n = 0) = q N = (0.999)100 = 0.905 which agrees with our
previous result.
Quantum Mechanics
Let’s review the quantum mechanics that you need to know. Mostly it’s just notation
and jargon. Particles or systems of particles are described by a wavefunction ψ. The
wavefunction is a function of the coordinates of the particles or of their momenta, but
not both. You can’t specify both the momentum and position of a particle because of
the Heisenberg uncertainty principle
>
∆x∆px ∼ h̄ (42)

So the wavefunction of a particle is written ψ(~r). The wavefunction is determined by


Schroedinger’s equation
Hψ = Eψ (43)
where H is the Hamiltonian and E is the energy. The Hamiltonian is the sum of the
kinetic and potential energies. You solve Schroedinger’s equation to determine the wave-
function. Schroedinger’s equation can be written as a matrix equation or as a second
order differential equation. Often the wavefunction solutions are labeled by definite val-
ues of energy, momentum, etc. These values can only have discrete values, i.e., they are
quantized. They are called quantum numbers. We can label the different wavefunction
solutions by their quantum numbers. Quantum numbers correspond to conserved quanti-
ties like energy, momentum, angular momentum, spin angular momentum, etc. Perhaps
the most familiar example of this is the electronic orbitals in the hydrogen atom. The

10
electron orbitals, like the 1s and 2p states, have definite discrete energies and definite
values of the orbital angular momentum. For example, the energy of the 1s state in
hydrogen is 1 Rydberg or 13.6 eV. The electron in this state can be described by a
wavefunction ψ1s (r). The quantum numbers of this state are n = 1 and L = 0. n = 1
means that it has energy E1 = 13.6 eV and L is the orbital angular momentum. (s is the
spectroscopic notation for L = 0.) Solving the Schroedinger equation for the hydrogen
atom yields many wavefunctions or orbitals (1s, 2s, 2px , 2py , 2pz , 3s, etc.).
Another example is a particle in a box.

V = infinity

V=0
x=0 x=a
The energy eigenvalues
p2 n2 π 2 h̄2
En = = (44)
2m 2ma2
Each value of n corresponds to a different eigenvalue of the energy En . Notice that the
energy levels are not equally spaced; they get farther apart as you go up in energy. Each
value of n also corresponds to a different wavefunction
nπx
 
ψn (x) = A sin(px/h̄) = A sin (45)
a
Notice that the more nodes there are, the more wiggles there are, and the higher the
energy is.
A harmonic oscillator is another example. This is just the quantum mechanical case
of a mass attached to a spring. In this case the potential is a parabola rather than being
a square well. A particle of mass m in this potential oscillates with frequency ω. The
Hamiltonian is
p̂2 1
Ĥ = + mω 2 x̂2 (46)
2m 2
You will learn how to solve this in your quantum mechanics course. Let me just write
down the result. The energy eigenvalues are
1
En = (n + )h̄ω (47)
2
where n = 0, 1, 2, 3, .... Notice once again that the energy levels are quantized. In this
case they are evenly spaced by an amount ∆E = h̄ω. The harmonic oscillator has
Gaussian wavefunctions (Hermite polynomials).

11
Wavefunctions can describe more than one particle. You can have a wavefunction
associated with a state with many particles. We call this a many body wavefunction. In
this case we can describe a state with N particles by ψ{n} (~r1 , ~r2 , ..., ~rN ). The subscript
{n} denotes the quantum numbers that label the wavefunction.

12

You might also like